commit stringlengths 40 40 | old_file stringlengths 4 184 | new_file stringlengths 4 184 | old_contents stringlengths 1 3.6k | new_contents stringlengths 5 3.38k | subject stringlengths 15 778 | message stringlengths 16 6.74k | lang stringclasses 201 values | license stringclasses 13 values | repos stringlengths 6 116k | config stringclasses 201 values | content stringlengths 137 7.24k | diff stringlengths 26 5.55k | diff_length int64 1 123 | relative_diff_length float64 0.01 89 | n_lines_added int64 0 108 | n_lines_deleted int64 0 106 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
8e762699b830015ab3907788370d1d1baf7fd2ce | IPython/nbconvert/templates/python.tpl | IPython/nbconvert/templates/python.tpl | {%- extends 'null.tpl' -%}
{% block in_prompt %}
# In[{{ cell.prompt_number if cell.prompt_number else ' ' }}]:
{% endblock in_prompt %}
{% block input %}
{{ cell.input | ipython2python }}
{% endblock input %}
{% block markdowncell scoped %}
{{ cell.source | comment_lines }}
{% endblock markdowncell %}
{% block headingcell scoped %}
{{ '#' * cell.level }}{{ cell.source | replace('\n', ' ') | comment_lines }}
{% endblock headingcell %}
| {%- extends 'null.tpl' -%}
{% block header %}
# coding: utf-8
{% endblock header %}
{% block in_prompt %}
# In[{{ cell.prompt_number if cell.prompt_number else ' ' }}]:
{% endblock in_prompt %}
{% block input %}
{{ cell.input | ipython2python }}
{% endblock input %}
{% block markdowncell scoped %}
{{ cell.source | comment_lines }}
{% endblock markdowncell %}
{% block headingcell scoped %}
{{ '#' * cell.level }}{{ cell.source | replace('\n', ' ') | comment_lines }}
{% endblock headingcell %}
| Add coding magic comment to nbconvert Python template | Add coding magic comment to nbconvert Python template
Closes gh-5470
| Smarty | bsd-3-clause | ipython/ipython,ipython/ipython | smarty | ## Code Before:
{%- extends 'null.tpl' -%}
{% block in_prompt %}
# In[{{ cell.prompt_number if cell.prompt_number else ' ' }}]:
{% endblock in_prompt %}
{% block input %}
{{ cell.input | ipython2python }}
{% endblock input %}
{% block markdowncell scoped %}
{{ cell.source | comment_lines }}
{% endblock markdowncell %}
{% block headingcell scoped %}
{{ '#' * cell.level }}{{ cell.source | replace('\n', ' ') | comment_lines }}
{% endblock headingcell %}
## Instruction:
Add coding magic comment to nbconvert Python template
Closes gh-5470
## Code After:
{%- extends 'null.tpl' -%}
{% block header %}
# coding: utf-8
{% endblock header %}
{% block in_prompt %}
# In[{{ cell.prompt_number if cell.prompt_number else ' ' }}]:
{% endblock in_prompt %}
{% block input %}
{{ cell.input | ipython2python }}
{% endblock input %}
{% block markdowncell scoped %}
{{ cell.source | comment_lines }}
{% endblock markdowncell %}
{% block headingcell scoped %}
{{ '#' * cell.level }}{{ cell.source | replace('\n', ' ') | comment_lines }}
{% endblock headingcell %}
| {%- extends 'null.tpl' -%}
+
+ {% block header %}
+ # coding: utf-8
+ {% endblock header %}
{% block in_prompt %}
# In[{{ cell.prompt_number if cell.prompt_number else ' ' }}]:
{% endblock in_prompt %}
{% block input %}
{{ cell.input | ipython2python }}
{% endblock input %}
{% block markdowncell scoped %}
{{ cell.source | comment_lines }}
{% endblock markdowncell %}
{% block headingcell scoped %}
{{ '#' * cell.level }}{{ cell.source | replace('\n', ' ') | comment_lines }}
{% endblock headingcell %} | 4 | 0.235294 | 4 | 0 |
57cb8d7924195c770499e69ab90eb3b978035f46 | actionmailbox/app/controllers/rails/conductor/action_mailbox/inbound_emails_controller.rb | actionmailbox/app/controllers/rails/conductor/action_mailbox/inbound_emails_controller.rb |
module Rails
class Conductor::ActionMailbox::InboundEmailsController < Rails::Conductor::BaseController
def index
@inbound_emails = ActionMailbox::InboundEmail.order(created_at: :desc)
end
def new
end
def show
@inbound_email = ActionMailbox::InboundEmail.find(params[:id])
end
def create
inbound_email = create_inbound_email(new_mail)
redirect_to main_app.rails_conductor_inbound_email_url(inbound_email)
end
private
def new_mail
Mail.new(params.require(:mail).permit(:from, :to, :cc, :bcc, :in_reply_to, :subject, :body).to_h).tap do |mail|
params[:mail][:attachments].to_a.each do |attachment|
mail.attachments[attachment.original_filename] = { filename: attachment.path, content_type: attachment.content_type }
end
end
end
def create_inbound_email(mail)
ActionMailbox::InboundEmail.create! raw_email: \
{ io: StringIO.new(mail.to_s), filename: "inbound.eml", content_type: "message/rfc822" }
end
end
end
|
module Rails
class Conductor::ActionMailbox::InboundEmailsController < Rails::Conductor::BaseController
def index
@inbound_emails = ActionMailbox::InboundEmail.order(created_at: :desc)
end
def new
end
def show
@inbound_email = ActionMailbox::InboundEmail.find(params[:id])
end
def create
inbound_email = create_inbound_email(new_mail)
redirect_to main_app.rails_conductor_inbound_email_url(inbound_email)
end
private
def new_mail
Mail.new(params.require(:mail).permit(:from, :to, :cc, :bcc, :in_reply_to, :subject, :body).to_h).tap do |mail|
params[:mail][:attachments].to_a.each do |attachment|
mail.attachments[attachment.original_filename] = { filename: attachment.path, content_type: attachment.content_type }
end
end
end
def create_inbound_email(mail)
ActionMailbox::InboundEmail.create_and_extract_message_id!(mail.to_s)
end
end
end
| Use create_and_extract_message_id! to create an inbound email. | Use create_and_extract_message_id! to create an inbound email.
This makes sure the created email has checksum and message id columns set.
| Ruby | mit | yawboakye/rails,georgeclaghorn/rails,MSP-Greg/rails,iainbeeston/rails,MSP-Greg/rails,tgxworld/rails,fabianoleittes/rails,gauravtiwari/rails,arunagw/rails,flanger001/rails,repinel/rails,palkan/rails,utilum/rails,Erol/rails,arunagw/rails,schuetzm/rails,BlakeWilliams/rails,Erol/rails,EmmaB/rails-1,prathamesh-sonpatki/rails,yawboakye/rails,notapatch/rails,tjschuck/rails,yhirano55/rails,Envek/rails,Envek/rails,tgxworld/rails,rails/rails,iainbeeston/rails,yahonda/rails,lcreid/rails,MSP-Greg/rails,prathamesh-sonpatki/rails,repinel/rails,betesh/rails,fabianoleittes/rails,utilum/rails,BlakeWilliams/rails,bogdanvlviv/rails,yhirano55/rails,palkan/rails,Vasfed/rails,notapatch/rails,Envek/rails,MSP-Greg/rails,Envek/rails,rafaelfranca/omg-rails,vipulnsward/rails,bogdanvlviv/rails,tgxworld/rails,mechanicles/rails,bogdanvlviv/rails,kddeisz/rails,shioyama/rails,utilum/rails,esparta/rails,yalab/rails,kmcphillips/rails,vipulnsward/rails,jeremy/rails,yhirano55/rails,arunagw/rails,Stellenticket/rails,Stellenticket/rails,yahonda/rails,utilum/rails,yhirano55/rails,flanger001/rails,lcreid/rails,esparta/rails,rails/rails,tjschuck/rails,rails/rails,eileencodes/rails,flanger001/rails,prathamesh-sonpatki/rails,georgeclaghorn/rails,kmcphillips/rails,eileencodes/rails,Edouard-chin/rails,iainbeeston/rails,EmmaB/rails-1,flanger001/rails,shioyama/rails,Edouard-chin/rails,yalab/rails,Edouard-chin/rails,yawboakye/rails,jeremy/rails,Stellenticket/rails,mechanicles/rails,esparta/rails,shioyama/rails,Vasfed/rails,starknx/rails,betesh/rails,starknx/rails,Edouard-chin/rails,prathamesh-sonpatki/rails,betesh/rails,joonyou/rails,georgeclaghorn/rails,EmmaB/rails-1,jeremy/rails,jeremy/rails,yalab/rails,fabianoleittes/rails,vipulnsward/rails,lcreid/rails,tjschuck/rails,tgxworld/rails,tjschuck/rails,yawboakye/rails,kddeisz/rails,rails/rails,bogdanvlviv/rails,schuetzm/rails,gauravtiwari/rails,kmcphillips/rails,BlakeWilliams/rails,georgeclaghorn/rails,kddeisz/rails,notapatch/rails,iainbeeston/rails,Stellenticket/rails,starknx/rails,joonyou/rails,esparta/rails,yahonda/rails,BlakeWilliams/rails,yalab/rails,Erol/rails,joonyou/rails,eileencodes/rails,kmcphillips/rails,schuetzm/rails,joonyou/rails,shioyama/rails,vipulnsward/rails,palkan/rails,rafaelfranca/omg-rails,notapatch/rails,lcreid/rails,rafaelfranca/omg-rails,Vasfed/rails,mechanicles/rails,Vasfed/rails,mechanicles/rails,Erol/rails,kddeisz/rails,betesh/rails,yahonda/rails,palkan/rails,fabianoleittes/rails,repinel/rails,gauravtiwari/rails,arunagw/rails,eileencodes/rails,schuetzm/rails,repinel/rails | ruby | ## Code Before:
module Rails
class Conductor::ActionMailbox::InboundEmailsController < Rails::Conductor::BaseController
def index
@inbound_emails = ActionMailbox::InboundEmail.order(created_at: :desc)
end
def new
end
def show
@inbound_email = ActionMailbox::InboundEmail.find(params[:id])
end
def create
inbound_email = create_inbound_email(new_mail)
redirect_to main_app.rails_conductor_inbound_email_url(inbound_email)
end
private
def new_mail
Mail.new(params.require(:mail).permit(:from, :to, :cc, :bcc, :in_reply_to, :subject, :body).to_h).tap do |mail|
params[:mail][:attachments].to_a.each do |attachment|
mail.attachments[attachment.original_filename] = { filename: attachment.path, content_type: attachment.content_type }
end
end
end
def create_inbound_email(mail)
ActionMailbox::InboundEmail.create! raw_email: \
{ io: StringIO.new(mail.to_s), filename: "inbound.eml", content_type: "message/rfc822" }
end
end
end
## Instruction:
Use create_and_extract_message_id! to create an inbound email.
This makes sure the created email has checksum and message id columns set.
## Code After:
module Rails
class Conductor::ActionMailbox::InboundEmailsController < Rails::Conductor::BaseController
def index
@inbound_emails = ActionMailbox::InboundEmail.order(created_at: :desc)
end
def new
end
def show
@inbound_email = ActionMailbox::InboundEmail.find(params[:id])
end
def create
inbound_email = create_inbound_email(new_mail)
redirect_to main_app.rails_conductor_inbound_email_url(inbound_email)
end
private
def new_mail
Mail.new(params.require(:mail).permit(:from, :to, :cc, :bcc, :in_reply_to, :subject, :body).to_h).tap do |mail|
params[:mail][:attachments].to_a.each do |attachment|
mail.attachments[attachment.original_filename] = { filename: attachment.path, content_type: attachment.content_type }
end
end
end
def create_inbound_email(mail)
ActionMailbox::InboundEmail.create_and_extract_message_id!(mail.to_s)
end
end
end
|
module Rails
class Conductor::ActionMailbox::InboundEmailsController < Rails::Conductor::BaseController
def index
@inbound_emails = ActionMailbox::InboundEmail.order(created_at: :desc)
end
def new
end
def show
@inbound_email = ActionMailbox::InboundEmail.find(params[:id])
end
def create
inbound_email = create_inbound_email(new_mail)
redirect_to main_app.rails_conductor_inbound_email_url(inbound_email)
end
private
def new_mail
Mail.new(params.require(:mail).permit(:from, :to, :cc, :bcc, :in_reply_to, :subject, :body).to_h).tap do |mail|
params[:mail][:attachments].to_a.each do |attachment|
mail.attachments[attachment.original_filename] = { filename: attachment.path, content_type: attachment.content_type }
end
end
end
def create_inbound_email(mail)
- ActionMailbox::InboundEmail.create! raw_email: \
? ^^ ^ ^^^
+ ActionMailbox::InboundEmail.create_and_extract_message_id!(mail.to_s)
? ^^^^^^^^ ^^ + ++++++++++ ^^^^^^
- { io: StringIO.new(mail.to_s), filename: "inbound.eml", content_type: "message/rfc822" }
end
end
end | 3 | 0.088235 | 1 | 2 |
885cb5a82bc4c1c632a976e0694889f161c35d6f | static/css/main.css | static/css/main.css | .pure-menu.pure-menu-fixed.main-menu {
border-bottom: none;
}
.pure-button.signup-button {
margin: 2px;
}
.splash-container {
background: #521F8F;
width: 100%;
height: 100%;
top: 0;
left: 0;
position: fixed;
}
.splash {
width: 50%;
height: 50%;
margin: auto;
position: absolute;
top: 100px;
left: 0;
bottom: 0;
right: 0;
text-align: center;
color: white;
}
.splash h1 {
font-size: 250%;
}
strong.highlight {
color: #2B0F4D;
}
| .pure-menu.pure-menu-fixed.main-menu {
background: #E6E6E6;
border-bottom: none;
}
.pure-button.signup-button {
margin: 2px;
}
.splash-container {
background: #521F8F;
width: 100%;
height: 100%;
top: 0;
left: 0;
position: fixed;
}
.splash {
width: 50%;
height: 50%;
margin: auto;
position: absolute;
top: 100px;
left: 0;
bottom: 0;
right: 0;
text-align: center;
color: white;
}
.splash h1 {
font-size: 250%;
}
strong.highlight {
color: #2B0F4D;
}
| Change the look of top banner | Change the look of top banner
| CSS | mit | malea/formaldatefinder,malea/formaldatefinder,malea/formaldatefinder | css | ## Code Before:
.pure-menu.pure-menu-fixed.main-menu {
border-bottom: none;
}
.pure-button.signup-button {
margin: 2px;
}
.splash-container {
background: #521F8F;
width: 100%;
height: 100%;
top: 0;
left: 0;
position: fixed;
}
.splash {
width: 50%;
height: 50%;
margin: auto;
position: absolute;
top: 100px;
left: 0;
bottom: 0;
right: 0;
text-align: center;
color: white;
}
.splash h1 {
font-size: 250%;
}
strong.highlight {
color: #2B0F4D;
}
## Instruction:
Change the look of top banner
## Code After:
.pure-menu.pure-menu-fixed.main-menu {
background: #E6E6E6;
border-bottom: none;
}
.pure-button.signup-button {
margin: 2px;
}
.splash-container {
background: #521F8F;
width: 100%;
height: 100%;
top: 0;
left: 0;
position: fixed;
}
.splash {
width: 50%;
height: 50%;
margin: auto;
position: absolute;
top: 100px;
left: 0;
bottom: 0;
right: 0;
text-align: center;
color: white;
}
.splash h1 {
font-size: 250%;
}
strong.highlight {
color: #2B0F4D;
}
| .pure-menu.pure-menu-fixed.main-menu {
+ background: #E6E6E6;
border-bottom: none;
}
.pure-button.signup-button {
margin: 2px;
}
.splash-container {
background: #521F8F;
width: 100%;
height: 100%;
top: 0;
left: 0;
position: fixed;
}
.splash {
width: 50%;
height: 50%;
margin: auto;
position: absolute;
top: 100px;
left: 0;
bottom: 0;
right: 0;
text-align: center;
color: white;
}
.splash h1 {
font-size: 250%;
}
strong.highlight {
color: #2B0F4D;
} | 1 | 0.027027 | 1 | 0 |
ffe7899cd8154882bbdc6b331dcdbbaf893795f4 | README.md | README.md | [](http://www.koara.io)
[](https://travis-ci.org/koara/koara-swift)
[](https://coveralls.io/github/koara/koara-swift?branch=master)
[](https://img.shields.io/cocoapods/v/Koara.svg)
[](https://github.com/koara/koara-java/blob/master/LICENSE)
# Koara-swift
[Koara](http://www.koara.io) is a modular lightweight markup language. This project is the core koara parser written in Swift.
If you are interested in converting koara to a specific outputFormat, please look the [Related Projects](#related-projects) section.
## Getting started
- CocoaPods
```ruby
use_frameworks!
target 'MyApp' do
pod 'Koara'
end
```
- Carthage
_TODO_
- Swift Package Manager
_TODO_
| [](http://www.koara.io)
[](https://travis-ci.org/koara/koara-swift)
[](https://coveralls.io/github/koara/koara-swift?branch=master)
[](https://img.shields.io/cocoapods/v/Koara.svg)
[](https://github.com/koara/koara-java/blob/master/LICENSE)
# Koara-swift
[Koara](http://www.koara.io) is a modular lightweight markup language. This project is the core koara parser written in Swift.
If you are interested in converting koara to a specific outputFormat, please look the [Related Projects](#related-projects) section.
## Getting started
- CocoaPods
```ruby
use_frameworks!
target 'MyApp' do
pod 'Koara'
end
```
- Carthage
- Swift Package Manager
| Test commit github (was not working) | Test commit github (was not working)
| Markdown | apache-2.0 | koara/koara-swift,koara/koara-swift | markdown | ## Code Before:
[](http://www.koara.io)
[](https://travis-ci.org/koara/koara-swift)
[](https://coveralls.io/github/koara/koara-swift?branch=master)
[](https://img.shields.io/cocoapods/v/Koara.svg)
[](https://github.com/koara/koara-java/blob/master/LICENSE)
# Koara-swift
[Koara](http://www.koara.io) is a modular lightweight markup language. This project is the core koara parser written in Swift.
If you are interested in converting koara to a specific outputFormat, please look the [Related Projects](#related-projects) section.
## Getting started
- CocoaPods
```ruby
use_frameworks!
target 'MyApp' do
pod 'Koara'
end
```
- Carthage
_TODO_
- Swift Package Manager
_TODO_
## Instruction:
Test commit github (was not working)
## Code After:
[](http://www.koara.io)
[](https://travis-ci.org/koara/koara-swift)
[](https://coveralls.io/github/koara/koara-swift?branch=master)
[](https://img.shields.io/cocoapods/v/Koara.svg)
[](https://github.com/koara/koara-java/blob/master/LICENSE)
# Koara-swift
[Koara](http://www.koara.io) is a modular lightweight markup language. This project is the core koara parser written in Swift.
If you are interested in converting koara to a specific outputFormat, please look the [Related Projects](#related-projects) section.
## Getting started
- CocoaPods
```ruby
use_frameworks!
target 'MyApp' do
pod 'Koara'
end
```
- Carthage
- Swift Package Manager
| [](http://www.koara.io)
[](https://travis-ci.org/koara/koara-swift)
[](https://coveralls.io/github/koara/koara-swift?branch=master)
[](https://img.shields.io/cocoapods/v/Koara.svg)
[](https://github.com/koara/koara-java/blob/master/LICENSE)
# Koara-swift
[Koara](http://www.koara.io) is a modular lightweight markup language. This project is the core koara parser written in Swift.
If you are interested in converting koara to a specific outputFormat, please look the [Related Projects](#related-projects) section.
## Getting started
- CocoaPods
```ruby
use_frameworks!
target 'MyApp' do
pod 'Koara'
end
```
- Carthage
- _TODO_
+
- Swift Package Manager
- _TODO_
| 3 | 0.096774 | 1 | 2 |
21c477b9815bc0d3bf2315568ed1e48082384adb | src/Tools/Analyzer/SecurityChecker.php | src/Tools/Analyzer/SecurityChecker.php | <?php
namespace Edge\QA\Tools\Analyzer;
use Edge\QA\OutputMode;
class SecurityChecker extends \Edge\QA\Tools\Tool
{
public static $SETTINGS = array(
'optionSeparator' => '=',
'internalClass' => 'SensioLabs\Security\SecurityChecker',
'outputMode' => OutputMode::RAW_CONSOLE_OUTPUT,
'composer' => 'sensiolabs/security-checker',
);
public function __invoke()
{
$composerLock = getcwd() . "/composer.lock";
foreach ($this->options->getAnalyzedDirs() as $escapedDir) {
$dir = rtrim(trim($escapedDir, '"'), '/');
$path = "{$dir}/composer.lock";
if (file_exists($path)) {
$composerLock = $path;
break;
}
}
$args = [
'security:check',
$composerLock,
];
if (getenv('TRAVIS_COMMIT')) {
$args['end-point'] = 'http://security.sensiolabs.org/check_lock';
}
return $args;
}
}
| <?php
namespace Edge\QA\Tools\Analyzer;
use Edge\QA\OutputMode;
class SecurityChecker extends \Edge\QA\Tools\Tool
{
public static $SETTINGS = array(
'optionSeparator' => '=',
'internalClass' => 'SensioLabs\Security\SecurityChecker',
'outputMode' => OutputMode::RAW_CONSOLE_OUTPUT,
'composer' => 'sensiolabs/security-checker',
);
public function __invoke()
{
$composerLock = getcwd() . "/composer.lock";
foreach ($this->options->getAnalyzedDirs() as $escapedDir) {
$dir = rtrim(trim($escapedDir, '"'), '/');
$path = "{$dir}/composer.lock";
if (file_exists($path)) {
$composerLock = $path;
break;
}
}
return [
'security:check',
$composerLock,
];
}
}
| Revert "Security-checker - fix tls issue at travis.ci" | Revert "Security-checker - fix tls issue at travis.ci"
This reverts commit 4c483c163b8e9c49eaca5cc34a9e74dc866b1a45.
| PHP | mit | EdgedesignCZ/phpqa,EdgedesignCZ/phpqa | php | ## Code Before:
<?php
namespace Edge\QA\Tools\Analyzer;
use Edge\QA\OutputMode;
class SecurityChecker extends \Edge\QA\Tools\Tool
{
public static $SETTINGS = array(
'optionSeparator' => '=',
'internalClass' => 'SensioLabs\Security\SecurityChecker',
'outputMode' => OutputMode::RAW_CONSOLE_OUTPUT,
'composer' => 'sensiolabs/security-checker',
);
public function __invoke()
{
$composerLock = getcwd() . "/composer.lock";
foreach ($this->options->getAnalyzedDirs() as $escapedDir) {
$dir = rtrim(trim($escapedDir, '"'), '/');
$path = "{$dir}/composer.lock";
if (file_exists($path)) {
$composerLock = $path;
break;
}
}
$args = [
'security:check',
$composerLock,
];
if (getenv('TRAVIS_COMMIT')) {
$args['end-point'] = 'http://security.sensiolabs.org/check_lock';
}
return $args;
}
}
## Instruction:
Revert "Security-checker - fix tls issue at travis.ci"
This reverts commit 4c483c163b8e9c49eaca5cc34a9e74dc866b1a45.
## Code After:
<?php
namespace Edge\QA\Tools\Analyzer;
use Edge\QA\OutputMode;
class SecurityChecker extends \Edge\QA\Tools\Tool
{
public static $SETTINGS = array(
'optionSeparator' => '=',
'internalClass' => 'SensioLabs\Security\SecurityChecker',
'outputMode' => OutputMode::RAW_CONSOLE_OUTPUT,
'composer' => 'sensiolabs/security-checker',
);
public function __invoke()
{
$composerLock = getcwd() . "/composer.lock";
foreach ($this->options->getAnalyzedDirs() as $escapedDir) {
$dir = rtrim(trim($escapedDir, '"'), '/');
$path = "{$dir}/composer.lock";
if (file_exists($path)) {
$composerLock = $path;
break;
}
}
return [
'security:check',
$composerLock,
];
}
}
| <?php
namespace Edge\QA\Tools\Analyzer;
use Edge\QA\OutputMode;
class SecurityChecker extends \Edge\QA\Tools\Tool
{
public static $SETTINGS = array(
'optionSeparator' => '=',
'internalClass' => 'SensioLabs\Security\SecurityChecker',
'outputMode' => OutputMode::RAW_CONSOLE_OUTPUT,
'composer' => 'sensiolabs/security-checker',
);
public function __invoke()
{
$composerLock = getcwd() . "/composer.lock";
foreach ($this->options->getAnalyzedDirs() as $escapedDir) {
$dir = rtrim(trim($escapedDir, '"'), '/');
$path = "{$dir}/composer.lock";
if (file_exists($path)) {
$composerLock = $path;
break;
}
}
- $args = [
+ return [
'security:check',
$composerLock,
];
- if (getenv('TRAVIS_COMMIT')) {
- $args['end-point'] = 'http://security.sensiolabs.org/check_lock';
- }
- return $args;
}
} | 6 | 0.166667 | 1 | 5 |
70ae942364907ce12566c16355017d1d2f706e6b | .buildkite/steps/publish-docker-images.sh | .buildkite/steps/publish-docker-images.sh | set -euo pipefail
dry_run() {
if [[ "${DRY_RUN:-}" == "false" ]] ; then
"$@"
else
echo "[dry-run] $*"
fi
}
if [[ "$CODENAME" == "" ]]; then
echo "Error: Missing \$CODENAME (stable, experimental or unstable)"
exit 1
fi
echo "--- Logging in to Docker Hub"
dockerhub_user="$(aws ssm get-parameter --name /pipelines/agent/DOCKER_HUB_USER --with-decryption --output text --query Parameter.Value --region us-east-1)"
aws ssm get-parameter --name /pipelines/agent/DOCKER_HUB_PASSWORD --with-decryption --output text --query Parameter.Value --region us-east-1 | docker login --username="${dockerhub_user}" --password-stdin
version=$(buildkite-agent meta-data get "agent-version")
build=$(buildkite-agent meta-data get "agent-version-build")
for variant in "alpine" "ubuntu" "centos" ; do
echo "--- Getting docker image tag for $variant from build meta data"
source_image=$(buildkite-agent meta-data get "agent-docker-image-$variant")
echo "Docker Image Tag for $variant: $source_image"
echo "--- :docker: Pulling prebuilt image"
docker pull "$source_image"
echo "--- :docker: Publishing images for $variant"
.buildkite/steps/publish-docker-image.sh "$variant" "$source_image" "$CODENAME" "$version" "$build"
done
| set -euo pipefail
dry_run() {
if [[ "${DRY_RUN:-}" == "false" ]] ; then
"$@"
else
echo "[dry-run] $*"
fi
}
if [[ "$CODENAME" == "" ]]; then
echo "Error: Missing \$CODENAME (stable, experimental or unstable)"
exit 1
fi
echo "--- Logging in to Docker Hub"
dockerhub_user="$(aws ssm get-parameter --name /pipelines/agent/DOCKER_HUB_USER --with-decryption --output text --query Parameter.Value --region us-east-1)"
aws ssm get-parameter --name /pipelines/agent/DOCKER_HUB_PASSWORD --with-decryption --output text --query Parameter.Value --region us-east-1 | docker login --username="${dockerhub_user}" --password-stdin
version=$(buildkite-agent meta-data get "agent-version")
build=$(buildkite-agent meta-data get "agent-version-build")
for variant in "alpine" "ubuntu" "centos" "sidecar" ; do
echo "--- Getting docker image tag for $variant from build meta data"
source_image=$(buildkite-agent meta-data get "agent-docker-image-$variant")
echo "Docker Image Tag for $variant: $source_image"
echo "--- :docker: Pulling prebuilt image"
docker pull "$source_image"
echo "--- :docker: Publishing images for $variant"
.buildkite/steps/publish-docker-image.sh "$variant" "$source_image" "$CODENAME" "$version" "$build"
done
| Add the sidecar variant to the docker hub publishing | Add the sidecar variant to the docker hub publishing
| Shell | mit | buildkite/agent,buildkite/agent,buildkite/agent | shell | ## Code Before:
set -euo pipefail
dry_run() {
if [[ "${DRY_RUN:-}" == "false" ]] ; then
"$@"
else
echo "[dry-run] $*"
fi
}
if [[ "$CODENAME" == "" ]]; then
echo "Error: Missing \$CODENAME (stable, experimental or unstable)"
exit 1
fi
echo "--- Logging in to Docker Hub"
dockerhub_user="$(aws ssm get-parameter --name /pipelines/agent/DOCKER_HUB_USER --with-decryption --output text --query Parameter.Value --region us-east-1)"
aws ssm get-parameter --name /pipelines/agent/DOCKER_HUB_PASSWORD --with-decryption --output text --query Parameter.Value --region us-east-1 | docker login --username="${dockerhub_user}" --password-stdin
version=$(buildkite-agent meta-data get "agent-version")
build=$(buildkite-agent meta-data get "agent-version-build")
for variant in "alpine" "ubuntu" "centos" ; do
echo "--- Getting docker image tag for $variant from build meta data"
source_image=$(buildkite-agent meta-data get "agent-docker-image-$variant")
echo "Docker Image Tag for $variant: $source_image"
echo "--- :docker: Pulling prebuilt image"
docker pull "$source_image"
echo "--- :docker: Publishing images for $variant"
.buildkite/steps/publish-docker-image.sh "$variant" "$source_image" "$CODENAME" "$version" "$build"
done
## Instruction:
Add the sidecar variant to the docker hub publishing
## Code After:
set -euo pipefail
dry_run() {
if [[ "${DRY_RUN:-}" == "false" ]] ; then
"$@"
else
echo "[dry-run] $*"
fi
}
if [[ "$CODENAME" == "" ]]; then
echo "Error: Missing \$CODENAME (stable, experimental or unstable)"
exit 1
fi
echo "--- Logging in to Docker Hub"
dockerhub_user="$(aws ssm get-parameter --name /pipelines/agent/DOCKER_HUB_USER --with-decryption --output text --query Parameter.Value --region us-east-1)"
aws ssm get-parameter --name /pipelines/agent/DOCKER_HUB_PASSWORD --with-decryption --output text --query Parameter.Value --region us-east-1 | docker login --username="${dockerhub_user}" --password-stdin
version=$(buildkite-agent meta-data get "agent-version")
build=$(buildkite-agent meta-data get "agent-version-build")
for variant in "alpine" "ubuntu" "centos" "sidecar" ; do
echo "--- Getting docker image tag for $variant from build meta data"
source_image=$(buildkite-agent meta-data get "agent-docker-image-$variant")
echo "Docker Image Tag for $variant: $source_image"
echo "--- :docker: Pulling prebuilt image"
docker pull "$source_image"
echo "--- :docker: Publishing images for $variant"
.buildkite/steps/publish-docker-image.sh "$variant" "$source_image" "$CODENAME" "$version" "$build"
done
| set -euo pipefail
dry_run() {
if [[ "${DRY_RUN:-}" == "false" ]] ; then
"$@"
else
echo "[dry-run] $*"
fi
}
if [[ "$CODENAME" == "" ]]; then
echo "Error: Missing \$CODENAME (stable, experimental or unstable)"
exit 1
fi
echo "--- Logging in to Docker Hub"
dockerhub_user="$(aws ssm get-parameter --name /pipelines/agent/DOCKER_HUB_USER --with-decryption --output text --query Parameter.Value --region us-east-1)"
aws ssm get-parameter --name /pipelines/agent/DOCKER_HUB_PASSWORD --with-decryption --output text --query Parameter.Value --region us-east-1 | docker login --username="${dockerhub_user}" --password-stdin
version=$(buildkite-agent meta-data get "agent-version")
build=$(buildkite-agent meta-data get "agent-version-build")
- for variant in "alpine" "ubuntu" "centos" ; do
+ for variant in "alpine" "ubuntu" "centos" "sidecar" ; do
? ++++++++++
echo "--- Getting docker image tag for $variant from build meta data"
source_image=$(buildkite-agent meta-data get "agent-docker-image-$variant")
echo "Docker Image Tag for $variant: $source_image"
echo "--- :docker: Pulling prebuilt image"
docker pull "$source_image"
echo "--- :docker: Publishing images for $variant"
.buildkite/steps/publish-docker-image.sh "$variant" "$source_image" "$CODENAME" "$version" "$build"
done
| 2 | 0.055556 | 1 | 1 |
99a46910d22d39bcd3499ecfab473991df26f83c | lib/optparse/pathname.rb | lib/optparse/pathname.rb | module Optparse
module Pathname
# Your code goes here...
end
end
| require 'optparse'
require 'pathname'
OptionParser.accept Pathname do |path|
Pathname.new path if path
end
| Make Pathname acceptable to OptionParser | Make Pathname acceptable to OptionParser
| Ruby | mit | KitaitiMakoto/optparse-pathname | ruby | ## Code Before:
module Optparse
module Pathname
# Your code goes here...
end
end
## Instruction:
Make Pathname acceptable to OptionParser
## Code After:
require 'optparse'
require 'pathname'
OptionParser.accept Pathname do |path|
Pathname.new path if path
end
| - module Optparse
- module Pathname
- # Your code goes here...
- end
+ require 'optparse'
+ require 'pathname'
+
+ OptionParser.accept Pathname do |path|
+ Pathname.new path if path
end | 9 | 1.8 | 5 | 4 |
82c1210d720ed7d0aa72e71339d9594d25cd0c2f | piece.rb | piece.rb | require "./chess_consts"
class Piece
attr_reader :type
attr_accessor :row, :col
def initialize(type = Chess::W_KING, row = 1, col = "e")
@type = type
@row = row
@col = col
end
def to_s
"#{@type} - #{col}#{row}"
end
end
| require "./chess_consts"
class Piece
attr_reader :type
attr_accessor :row, :col
def initialize(type = Chess::W_KING, row = 1, col = "e")
@type = type
@row = row
@col = col
end
def to_s
"#{@type} - #{col}#{row}"
end
def is_pawn?
if @type.eql? Chess::B_PAWN or @type.eql? Chess::W_PAWN then
return true
else
return false
end
end
def is_rook?
if @type.eql? Chess::B_ROOK or @type.eql? Chess::W_ROOK then
return true
else
return false
end
end
def is_knight?
if @type.eql? Chess::B_KNIGHT or @type.eql? Chess::W_KNIGHT then
return true
else
return false
end
end
def is_bishop?
if @type.eql? Chess::B_BISHOP or @type.eql? Chess::W_BISHOP then
return true
else
return false
end
end
def is_queen?
if @type.eql? Chess::B_QUEEN or @type.eql? Chess::W_QUEEN then
return true
else
return false
end
end
def is_king?
if @type.eql? Chess::B_KING or @type.eql? Chess::W_KING then
return true
else
return false
end
end
end
| Add identity methods to Piece | Add identity methods to Piece
| Ruby | mit | mahimahi42/chess | ruby | ## Code Before:
require "./chess_consts"
class Piece
attr_reader :type
attr_accessor :row, :col
def initialize(type = Chess::W_KING, row = 1, col = "e")
@type = type
@row = row
@col = col
end
def to_s
"#{@type} - #{col}#{row}"
end
end
## Instruction:
Add identity methods to Piece
## Code After:
require "./chess_consts"
class Piece
attr_reader :type
attr_accessor :row, :col
def initialize(type = Chess::W_KING, row = 1, col = "e")
@type = type
@row = row
@col = col
end
def to_s
"#{@type} - #{col}#{row}"
end
def is_pawn?
if @type.eql? Chess::B_PAWN or @type.eql? Chess::W_PAWN then
return true
else
return false
end
end
def is_rook?
if @type.eql? Chess::B_ROOK or @type.eql? Chess::W_ROOK then
return true
else
return false
end
end
def is_knight?
if @type.eql? Chess::B_KNIGHT or @type.eql? Chess::W_KNIGHT then
return true
else
return false
end
end
def is_bishop?
if @type.eql? Chess::B_BISHOP or @type.eql? Chess::W_BISHOP then
return true
else
return false
end
end
def is_queen?
if @type.eql? Chess::B_QUEEN or @type.eql? Chess::W_QUEEN then
return true
else
return false
end
end
def is_king?
if @type.eql? Chess::B_KING or @type.eql? Chess::W_KING then
return true
else
return false
end
end
end
| require "./chess_consts"
class Piece
attr_reader :type
attr_accessor :row, :col
def initialize(type = Chess::W_KING, row = 1, col = "e")
@type = type
@row = row
@col = col
end
def to_s
"#{@type} - #{col}#{row}"
end
+
+ def is_pawn?
+ if @type.eql? Chess::B_PAWN or @type.eql? Chess::W_PAWN then
+ return true
+ else
+ return false
+ end
+ end
+
+ def is_rook?
+ if @type.eql? Chess::B_ROOK or @type.eql? Chess::W_ROOK then
+ return true
+ else
+ return false
+ end
+ end
+
+ def is_knight?
+ if @type.eql? Chess::B_KNIGHT or @type.eql? Chess::W_KNIGHT then
+ return true
+ else
+ return false
+ end
+ end
+
+ def is_bishop?
+ if @type.eql? Chess::B_BISHOP or @type.eql? Chess::W_BISHOP then
+ return true
+ else
+ return false
+ end
+ end
+
+ def is_queen?
+ if @type.eql? Chess::B_QUEEN or @type.eql? Chess::W_QUEEN then
+ return true
+ else
+ return false
+ end
+ end
+
+ def is_king?
+ if @type.eql? Chess::B_KING or @type.eql? Chess::W_KING then
+ return true
+ else
+ return false
+ end
+ end
end | 48 | 3 | 48 | 0 |
fa737f3ceb041d152a5fb53faa66b3456ffcd10f | src/App.js | src/App.js | import React from "react";
import { RoutingApp } from "./modules/RoutingApp";
import store from "./store";
import { Provider } from "react-redux";
import { STUY_SPEC_API_URL } from "./constants";
import { ApolloProvider } from "react-apollo";
import { ApolloClient } from "apollo-client";
import { HttpLink } from "apollo-link-http";
import { InMemoryCache } from "apollo-cache-inmemory";
import { objectFilter } from "./utils";
import './index.scss';
const apolloClient = new ApolloClient({
link: new HttpLink({ uri: `${STUY_SPEC_API_URL}/graphql` }),
cache: new InMemoryCache(),
connectToDevTools: true
});
Object.filter = objectFilter;
function App () {
return(
<Provider store={store}>
<ApolloProvider client={apolloClient}>
<RoutingApp />
</ApolloProvider>
</Provider>
)
}
export default App;
| import React, { useEffect } from "react";
import { RoutingApp } from "./modules/RoutingApp";
import store from "./store";
import { Provider } from "react-redux";
import { STUY_SPEC_API_URL } from "./constants";
import { ApolloProvider } from "react-apollo";
import { ApolloClient } from "apollo-client";
import { HttpLink } from "apollo-link-http";
import { InMemoryCache } from "apollo-cache-inmemory";
import { objectFilter } from "./utils";
import useDarkMode from "use-dark-mode";
import "./index.scss";
const apolloClient = new ApolloClient({
link: new HttpLink({ uri: `${STUY_SPEC_API_URL}/graphql` }),
cache: new InMemoryCache(),
connectToDevTools: true,
});
Object.filter = objectFilter;
function App() {
const darkMode = useDarkMode(false);
useEffect(() => {
if (localStorage.getItem("darkMode") == null) {
darkMode.disable();
}
});
return (
<Provider store={store}>
<ApolloProvider client={apolloClient}>
<RoutingApp />
</ApolloProvider>
</Provider>
);
}
export default App;
| Set default to be light mode | Set default to be light mode
| JavaScript | mit | stuyspec/client-app | javascript | ## Code Before:
import React from "react";
import { RoutingApp } from "./modules/RoutingApp";
import store from "./store";
import { Provider } from "react-redux";
import { STUY_SPEC_API_URL } from "./constants";
import { ApolloProvider } from "react-apollo";
import { ApolloClient } from "apollo-client";
import { HttpLink } from "apollo-link-http";
import { InMemoryCache } from "apollo-cache-inmemory";
import { objectFilter } from "./utils";
import './index.scss';
const apolloClient = new ApolloClient({
link: new HttpLink({ uri: `${STUY_SPEC_API_URL}/graphql` }),
cache: new InMemoryCache(),
connectToDevTools: true
});
Object.filter = objectFilter;
function App () {
return(
<Provider store={store}>
<ApolloProvider client={apolloClient}>
<RoutingApp />
</ApolloProvider>
</Provider>
)
}
export default App;
## Instruction:
Set default to be light mode
## Code After:
import React, { useEffect } from "react";
import { RoutingApp } from "./modules/RoutingApp";
import store from "./store";
import { Provider } from "react-redux";
import { STUY_SPEC_API_URL } from "./constants";
import { ApolloProvider } from "react-apollo";
import { ApolloClient } from "apollo-client";
import { HttpLink } from "apollo-link-http";
import { InMemoryCache } from "apollo-cache-inmemory";
import { objectFilter } from "./utils";
import useDarkMode from "use-dark-mode";
import "./index.scss";
const apolloClient = new ApolloClient({
link: new HttpLink({ uri: `${STUY_SPEC_API_URL}/graphql` }),
cache: new InMemoryCache(),
connectToDevTools: true,
});
Object.filter = objectFilter;
function App() {
const darkMode = useDarkMode(false);
useEffect(() => {
if (localStorage.getItem("darkMode") == null) {
darkMode.disable();
}
});
return (
<Provider store={store}>
<ApolloProvider client={apolloClient}>
<RoutingApp />
</ApolloProvider>
</Provider>
);
}
export default App;
| - import React from "react";
+ import React, { useEffect } from "react";
? +++++++++++++++
import { RoutingApp } from "./modules/RoutingApp";
import store from "./store";
import { Provider } from "react-redux";
import { STUY_SPEC_API_URL } from "./constants";
import { ApolloProvider } from "react-apollo";
import { ApolloClient } from "apollo-client";
import { HttpLink } from "apollo-link-http";
import { InMemoryCache } from "apollo-cache-inmemory";
import { objectFilter } from "./utils";
+ import useDarkMode from "use-dark-mode";
- import './index.scss';
? ^ ^
+ import "./index.scss";
? ^ ^
const apolloClient = new ApolloClient({
link: new HttpLink({ uri: `${STUY_SPEC_API_URL}/graphql` }),
cache: new InMemoryCache(),
- connectToDevTools: true
+ connectToDevTools: true,
? +
});
Object.filter = objectFilter;
- function App () {
? -
+ function App() {
+ const darkMode = useDarkMode(false);
+
+ useEffect(() => {
+ if (localStorage.getItem("darkMode") == null) {
+ darkMode.disable();
+ }
+ });
- return(
+ return (
? +
<Provider store={store}>
<ApolloProvider client={apolloClient}>
<RoutingApp />
</ApolloProvider>
</Provider>
- )
+ );
? +
}
export default App; | 20 | 0.606061 | 14 | 6 |
8f4ec53cc6452d1ad7cbe9790976367545837a30 | app/component/popup.vue | app/component/popup.vue | <template lang="html">
<div>
<button @click="closeTabs">
Tidy
</button>
</div>
</template>
<script>
export default {
methods: {
closeTabs () {
chrome.storage.sync.set({'foo': { name: 'bar' }});
chrome.storage.sync.set({'baz': { name: 'bar' }});
chrome.storage.sync.get(null, function(items) {
const allKeys = Object.keys(items);
});
chrome.windows.getCurrent(null, window => {
chrome.tabs.query({}, tabs => {
tabs.forEach(t => {
if (t.windowId === window.id) {
console.log(`Closing tab with url of ${t.url}`);
chrome.tabs.remove(t.id)
}
})
});
});
const dashboardURL = chrome.extension.getURL('pages/dashboard.html')
chrome.tabs.create({ url: dashboardURL });
}
}
}
</script>
<style scoped="true" lang="sass">
@import '../styles/settings';
button {
width: 100px;
background: $color-primary;
color: white;
}
</style>
| <template lang="html">
<div>
<button @click="closeTabs">
Tidy
</button>
</div>
</template>
<script>
export default {
methods: {
saveTabData (tabs){
chrome.windows.getCurrent(null, window => {
chrome.tabs.query({}, tabs => {
let data = {};
tabs.forEach(t => {
if (t.windowId === window.id) {
data[t.url] = {
'url': t.url,
'icon': t.favIconUrl,
'title': t.title,
}
}
})
chrome.storage.sync.get('tidy_storage', function(items) {
let tabData = items['tidy_storage'] || {};
const timeStamp = Date.now();
tabData[timeStamp] = data;
chrome.extension.getBackgroundPage().console.log(tabData);
chrome.storage.sync.set({'tidy_storage': tabData});
});
});
});
},
closeTabs () {
this.saveTabData()
chrome.windows.getCurrent(null, window => {
chrome.tabs.query({}, tabs => {
let windowTabs = tabs.filter(t => t.windowId === window.id);
let tabsClosed = 0
windowTabs.forEach(t => {
tabsClosed++;
if(tabsClosed == windowTabs.length){
const dashboardURL = chrome.extension.getURL('pages/dashboard.html')
chrome.tabs.create({ url: dashboardURL });
}
chrome.tabs.remove(t.id);
})
});
});
}
}
}
</script>
<style scoped="true" lang="sass">
@import '../styles/settings';
button {
width: 100px;
background: $color-primary;
color: white;
}
</style>
| Save Tabs to Chrome Storage | Save Tabs to Chrome Storage
| Vue | mit | eggplanetio/tidytab,eggplanetio/tidytab | vue | ## Code Before:
<template lang="html">
<div>
<button @click="closeTabs">
Tidy
</button>
</div>
</template>
<script>
export default {
methods: {
closeTabs () {
chrome.storage.sync.set({'foo': { name: 'bar' }});
chrome.storage.sync.set({'baz': { name: 'bar' }});
chrome.storage.sync.get(null, function(items) {
const allKeys = Object.keys(items);
});
chrome.windows.getCurrent(null, window => {
chrome.tabs.query({}, tabs => {
tabs.forEach(t => {
if (t.windowId === window.id) {
console.log(`Closing tab with url of ${t.url}`);
chrome.tabs.remove(t.id)
}
})
});
});
const dashboardURL = chrome.extension.getURL('pages/dashboard.html')
chrome.tabs.create({ url: dashboardURL });
}
}
}
</script>
<style scoped="true" lang="sass">
@import '../styles/settings';
button {
width: 100px;
background: $color-primary;
color: white;
}
</style>
## Instruction:
Save Tabs to Chrome Storage
## Code After:
<template lang="html">
<div>
<button @click="closeTabs">
Tidy
</button>
</div>
</template>
<script>
export default {
methods: {
saveTabData (tabs){
chrome.windows.getCurrent(null, window => {
chrome.tabs.query({}, tabs => {
let data = {};
tabs.forEach(t => {
if (t.windowId === window.id) {
data[t.url] = {
'url': t.url,
'icon': t.favIconUrl,
'title': t.title,
}
}
})
chrome.storage.sync.get('tidy_storage', function(items) {
let tabData = items['tidy_storage'] || {};
const timeStamp = Date.now();
tabData[timeStamp] = data;
chrome.extension.getBackgroundPage().console.log(tabData);
chrome.storage.sync.set({'tidy_storage': tabData});
});
});
});
},
closeTabs () {
this.saveTabData()
chrome.windows.getCurrent(null, window => {
chrome.tabs.query({}, tabs => {
let windowTabs = tabs.filter(t => t.windowId === window.id);
let tabsClosed = 0
windowTabs.forEach(t => {
tabsClosed++;
if(tabsClosed == windowTabs.length){
const dashboardURL = chrome.extension.getURL('pages/dashboard.html')
chrome.tabs.create({ url: dashboardURL });
}
chrome.tabs.remove(t.id);
})
});
});
}
}
}
</script>
<style scoped="true" lang="sass">
@import '../styles/settings';
button {
width: 100px;
background: $color-primary;
color: white;
}
</style>
| <template lang="html">
<div>
<button @click="closeTabs">
Tidy
</button>
</div>
</template>
<script>
export default {
methods: {
+ saveTabData (tabs){
+ chrome.windows.getCurrent(null, window => {
+ chrome.tabs.query({}, tabs => {
+ let data = {};
+
+ tabs.forEach(t => {
+ if (t.windowId === window.id) {
+ data[t.url] = {
+ 'url': t.url,
+ 'icon': t.favIconUrl,
+ 'title': t.title,
+ }
+ }
+ })
+
+ chrome.storage.sync.get('tidy_storage', function(items) {
+ let tabData = items['tidy_storage'] || {};
+
+ const timeStamp = Date.now();
+
+ tabData[timeStamp] = data;
+
+ chrome.extension.getBackgroundPage().console.log(tabData);
+
+ chrome.storage.sync.set({'tidy_storage': tabData});
+
+ });
+ });
+ });
+ },
closeTabs () {
+ this.saveTabData()
- chrome.storage.sync.set({'foo': { name: 'bar' }});
- chrome.storage.sync.set({'baz': { name: 'bar' }});
- chrome.storage.sync.get(null, function(items) {
- const allKeys = Object.keys(items);
- });
chrome.windows.getCurrent(null, window => {
chrome.tabs.query({}, tabs => {
+
+ let windowTabs = tabs.filter(t => t.windowId === window.id);
+
+ let tabsClosed = 0
+
- tabs.forEach(t => {
? ^
+ windowTabs.forEach(t => {
? ^^^^^^^
- if (t.windowId === window.id) {
- console.log(`Closing tab with url of ${t.url}`);
- chrome.tabs.remove(t.id)
+ tabsClosed++;
+ if(tabsClosed == windowTabs.length){
+ const dashboardURL = chrome.extension.getURL('pages/dashboard.html')
+ chrome.tabs.create({ url: dashboardURL });
}
+ chrome.tabs.remove(t.id);
})
});
});
-
- const dashboardURL = chrome.extension.getURL('pages/dashboard.html')
- chrome.tabs.create({ url: dashboardURL });
}
}
}
</script>
<style scoped="true" lang="sass">
@import '../styles/settings';
button {
width: 100px;
background: $color-primary;
color: white;
}
</style> | 54 | 1.2 | 42 | 12 |
d31b31299c0dbd0aafcee52ae95e1e2386dc1cb1 | Snow/_posts/2016-06-02-github-cvs-resumes.md | Snow/_posts/2016-06-02-github-cvs-resumes.md | ---
layout: post
title: GitHub CVs / Resumes
categories: General
date: 2016-06-02 10:00:00
comments: true
sharing: true
footer: true
permalink: /github-cvs-resumes/
---
I just discovered that Github can automatically generate a CV/resume for its users based on the public information in their GitHub account. You have to opt-in by going to their [GitHub project](https://github.com/resume/resume.github.com) page and starring the project. Check out mine at [http://resume.github.com/?mwhelan](http://resume.github.com/?mwhelan).
<!--excerpt-->
| ---
layout: post
title: GitHub CVs / Resumes
categories: General
date: 2016-06-02 08:00:00
comments: true
sharing: true
footer: true
permalink: /github-cvs-resumes/
---
I just discovered that Github can automatically generate a CV/resume for its users based on the public information in their GitHub account. You have to opt-in by going to their [GitHub project](https://github.com/resume/resume.github.com) page and starring the project. Check out mine at [http://resume.github.com/?mwhelan](http://resume.github.com/?mwhelan). What a clever idea!
<!--excerpt-->
| Set post time earlier to facilitate publishing | Set post time earlier to facilitate publishing
| Markdown | mit | mwhelan/mwhelan.github.io,mwhelan/mwhelan.github.io | markdown | ## Code Before:
---
layout: post
title: GitHub CVs / Resumes
categories: General
date: 2016-06-02 10:00:00
comments: true
sharing: true
footer: true
permalink: /github-cvs-resumes/
---
I just discovered that Github can automatically generate a CV/resume for its users based on the public information in their GitHub account. You have to opt-in by going to their [GitHub project](https://github.com/resume/resume.github.com) page and starring the project. Check out mine at [http://resume.github.com/?mwhelan](http://resume.github.com/?mwhelan).
<!--excerpt-->
## Instruction:
Set post time earlier to facilitate publishing
## Code After:
---
layout: post
title: GitHub CVs / Resumes
categories: General
date: 2016-06-02 08:00:00
comments: true
sharing: true
footer: true
permalink: /github-cvs-resumes/
---
I just discovered that Github can automatically generate a CV/resume for its users based on the public information in their GitHub account. You have to opt-in by going to their [GitHub project](https://github.com/resume/resume.github.com) page and starring the project. Check out mine at [http://resume.github.com/?mwhelan](http://resume.github.com/?mwhelan). What a clever idea!
<!--excerpt-->
| ---
layout: post
title: GitHub CVs / Resumes
categories: General
- date: 2016-06-02 10:00:00
? -
+ date: 2016-06-02 08:00:00
? +
comments: true
sharing: true
footer: true
permalink: /github-cvs-resumes/
---
- I just discovered that Github can automatically generate a CV/resume for its users based on the public information in their GitHub account. You have to opt-in by going to their [GitHub project](https://github.com/resume/resume.github.com) page and starring the project. Check out mine at [http://resume.github.com/?mwhelan](http://resume.github.com/?mwhelan).
+ I just discovered that Github can automatically generate a CV/resume for its users based on the public information in their GitHub account. You have to opt-in by going to their [GitHub project](https://github.com/resume/resume.github.com) page and starring the project. Check out mine at [http://resume.github.com/?mwhelan](http://resume.github.com/?mwhelan). What a clever idea!
? ++++++++++++++++++++
<!--excerpt-->
| 4 | 0.285714 | 2 | 2 |
0b21de3d7f5be04d818f807d37a425f737c662fb | apps/chloe/src/chloe.erl | apps/chloe/src/chloe.erl | -module(chloe).
%% API
-export([
start/0,
stop/0
]).
%%--------------------------------------------------------------------
%% API
%%--------------------------------------------------------------------
start() ->
ok = application:start(inets),
application:start(chloe).
stop() ->
ok = application:stop(chloe),
application:stop(inets).
| -module(chloe).
%% API
-export([
start/0,
stop/0
]).
%%--------------------------------------------------------------------
%% API
%%--------------------------------------------------------------------
start() ->
ensure_started(inets),
application:start(chloe).
stop() ->
application:stop(chloe).
%%--------------------------------------------------------------------
%% Internal functions
%%--------------------------------------------------------------------
ensure_started(App) ->
case application:start(App) of
ok ->
ok;
{error, {already_started, App}} ->
ok
end.
| Use ensure_started for safer starting | Use ensure_started for safer starting
| Erlang | mit | mashion/chloe,mashion/chloe,mashion/chloe | erlang | ## Code Before:
-module(chloe).
%% API
-export([
start/0,
stop/0
]).
%%--------------------------------------------------------------------
%% API
%%--------------------------------------------------------------------
start() ->
ok = application:start(inets),
application:start(chloe).
stop() ->
ok = application:stop(chloe),
application:stop(inets).
## Instruction:
Use ensure_started for safer starting
## Code After:
-module(chloe).
%% API
-export([
start/0,
stop/0
]).
%%--------------------------------------------------------------------
%% API
%%--------------------------------------------------------------------
start() ->
ensure_started(inets),
application:start(chloe).
stop() ->
application:stop(chloe).
%%--------------------------------------------------------------------
%% Internal functions
%%--------------------------------------------------------------------
ensure_started(App) ->
case application:start(App) of
ok ->
ok;
{error, {already_started, App}} ->
ok
end.
| -module(chloe).
%% API
-export([
start/0,
stop/0
]).
%%--------------------------------------------------------------------
%% API
%%--------------------------------------------------------------------
start() ->
- ok = application:start(inets),
+ ensure_started(inets),
application:start(chloe).
stop() ->
- ok = application:stop(chloe),
? ----- ^
+ application:stop(chloe).
? ^
- application:stop(inets).
+
+ %%--------------------------------------------------------------------
+ %% Internal functions
+ %%--------------------------------------------------------------------
+
+ ensure_started(App) ->
+ case application:start(App) of
+ ok ->
+ ok;
+ {error, {already_started, App}} ->
+ ok
+ end. | 17 | 0.894737 | 14 | 3 |
6ecc67abc3b42eb862afb1e886c866d748fb60ec | spec/ruby/core/file/absolute_path_spec.rb | spec/ruby/core/file/absolute_path_spec.rb | require File.expand_path('../../../spec_helper', __FILE__)
ruby_version_is "1.9" do
describe "File.absolute_path" do
before(:each) do
@abs = File.expand_path(__FILE__)
end
it "returns the argument if it's an absolute pathname" do
File.absolute_path(@abs).should == @abs
end
it "resolves paths relative to the current working directory" do
path = File.dirname(@abs)
Dir.chdir(path) do
File.absolute_path('hello.txt').should == File.join(path, 'hello.txt')
end
end
it "doesn't expand '~'" do
File.absolute_path('~').should_not == File.expand_path('~')
end
it "accepts a second argument of a directory from which to resolve the path" do
File.absolute_path(__FILE__, File.dirname(__FILE__)).should == @abs
end
it "calls #to_path on its argument" do
File.absolute_path(mock_to_path(@abs)).should == @abs
end
end
end
| require File.expand_path('../../../spec_helper', __FILE__)
ruby_version_is "1.9" do
describe "File.absolute_path" do
before(:each) do
@abs = File.expand_path(__FILE__)
end
it "returns the argument if it's an absolute pathname" do
File.absolute_path(@abs).should == @abs
end
it "resolves paths relative to the current working directory" do
path = File.dirname(@abs)
Dir.chdir(path) do
File.absolute_path('hello.txt').should == File.join(path, 'hello.txt')
end
end
it "does not expand '~' to a home directory." do
File.absolute_path('~').should_not == File.expand_path('~')
end
it "does not expand '~user' to a home directory." do
path = File.dirname(@abs)
Dir.chdir(path) do
File.absolute_path('~user').should == File.join(path, '~user')
end
end
it "accepts a second argument of a directory from which to resolve the path" do
File.absolute_path(__FILE__, File.dirname(__FILE__)).should == @abs
end
it "calls #to_path on its argument" do
File.absolute_path(mock_to_path(@abs)).should == @abs
end
end
end
| Add spec for receiving '~user' in File.absolute_path. | Add spec for receiving '~user' in File.absolute_path.
| Ruby | mpl-2.0 | ngpestelos/rubinius,mlarraz/rubinius,digitalextremist/rubinius,jsyeo/rubinius,sferik/rubinius,digitalextremist/rubinius,digitalextremist/rubinius,jemc/rubinius,ruipserra/rubinius,pH14/rubinius,ngpestelos/rubinius,mlarraz/rubinius,pH14/rubinius,ngpestelos/rubinius,ruipserra/rubinius,benlovell/rubinius,ruipserra/rubinius,benlovell/rubinius,Wirachmat/rubinius,mlarraz/rubinius,benlovell/rubinius,mlarraz/rubinius,Azizou/rubinius,ruipserra/rubinius,sferik/rubinius,jemc/rubinius,Wirachmat/rubinius,dblock/rubinius,dblock/rubinius,jemc/rubinius,benlovell/rubinius,kachick/rubinius,heftig/rubinius,digitalextremist/rubinius,kachick/rubinius,benlovell/rubinius,jemc/rubinius,jsyeo/rubinius,dblock/rubinius,Wirachmat/rubinius,Wirachmat/rubinius,Wirachmat/rubinius,ruipserra/rubinius,ruipserra/rubinius,ngpestelos/rubinius,sferik/rubinius,Wirachmat/rubinius,mlarraz/rubinius,jsyeo/rubinius,Azizou/rubinius,sferik/rubinius,ngpestelos/rubinius,heftig/rubinius,kachick/rubinius,ngpestelos/rubinius,ngpestelos/rubinius,jsyeo/rubinius,heftig/rubinius,pH14/rubinius,digitalextremist/rubinius,mlarraz/rubinius,pH14/rubinius,dblock/rubinius,pH14/rubinius,Azizou/rubinius,jsyeo/rubinius,kachick/rubinius,jemc/rubinius,heftig/rubinius,Azizou/rubinius,benlovell/rubinius,kachick/rubinius,jsyeo/rubinius,Azizou/rubinius,pH14/rubinius,dblock/rubinius,heftig/rubinius,sferik/rubinius,dblock/rubinius,pH14/rubinius,heftig/rubinius,jemc/rubinius,digitalextremist/rubinius,mlarraz/rubinius,Wirachmat/rubinius,digitalextremist/rubinius,sferik/rubinius,sferik/rubinius,ruipserra/rubinius,dblock/rubinius,kachick/rubinius,jsyeo/rubinius,benlovell/rubinius,Azizou/rubinius,jemc/rubinius,Azizou/rubinius,kachick/rubinius,kachick/rubinius,heftig/rubinius | ruby | ## Code Before:
require File.expand_path('../../../spec_helper', __FILE__)
ruby_version_is "1.9" do
describe "File.absolute_path" do
before(:each) do
@abs = File.expand_path(__FILE__)
end
it "returns the argument if it's an absolute pathname" do
File.absolute_path(@abs).should == @abs
end
it "resolves paths relative to the current working directory" do
path = File.dirname(@abs)
Dir.chdir(path) do
File.absolute_path('hello.txt').should == File.join(path, 'hello.txt')
end
end
it "doesn't expand '~'" do
File.absolute_path('~').should_not == File.expand_path('~')
end
it "accepts a second argument of a directory from which to resolve the path" do
File.absolute_path(__FILE__, File.dirname(__FILE__)).should == @abs
end
it "calls #to_path on its argument" do
File.absolute_path(mock_to_path(@abs)).should == @abs
end
end
end
## Instruction:
Add spec for receiving '~user' in File.absolute_path.
## Code After:
require File.expand_path('../../../spec_helper', __FILE__)
ruby_version_is "1.9" do
describe "File.absolute_path" do
before(:each) do
@abs = File.expand_path(__FILE__)
end
it "returns the argument if it's an absolute pathname" do
File.absolute_path(@abs).should == @abs
end
it "resolves paths relative to the current working directory" do
path = File.dirname(@abs)
Dir.chdir(path) do
File.absolute_path('hello.txt').should == File.join(path, 'hello.txt')
end
end
it "does not expand '~' to a home directory." do
File.absolute_path('~').should_not == File.expand_path('~')
end
it "does not expand '~user' to a home directory." do
path = File.dirname(@abs)
Dir.chdir(path) do
File.absolute_path('~user').should == File.join(path, '~user')
end
end
it "accepts a second argument of a directory from which to resolve the path" do
File.absolute_path(__FILE__, File.dirname(__FILE__)).should == @abs
end
it "calls #to_path on its argument" do
File.absolute_path(mock_to_path(@abs)).should == @abs
end
end
end
| require File.expand_path('../../../spec_helper', __FILE__)
ruby_version_is "1.9" do
describe "File.absolute_path" do
before(:each) do
@abs = File.expand_path(__FILE__)
end
it "returns the argument if it's an absolute pathname" do
File.absolute_path(@abs).should == @abs
end
it "resolves paths relative to the current working directory" do
path = File.dirname(@abs)
Dir.chdir(path) do
File.absolute_path('hello.txt').should == File.join(path, 'hello.txt')
end
end
- it "doesn't expand '~'" do
+ it "does not expand '~' to a home directory." do
File.absolute_path('~').should_not == File.expand_path('~')
+ end
+
+ it "does not expand '~user' to a home directory." do
+ path = File.dirname(@abs)
+ Dir.chdir(path) do
+ File.absolute_path('~user').should == File.join(path, '~user')
+ end
end
it "accepts a second argument of a directory from which to resolve the path" do
File.absolute_path(__FILE__, File.dirname(__FILE__)).should == @abs
end
it "calls #to_path on its argument" do
File.absolute_path(mock_to_path(@abs)).should == @abs
end
end
end | 9 | 0.28125 | 8 | 1 |
54458daf55f76a395f532bfb54985de10a542814 | crispy_forms/templates/bootstrap3/whole_uni_form.html | crispy_forms/templates/bootstrap3/whole_uni_form.html | {% load crispy_forms_utils %}
{% specialspaceless %}
{% if form_tag %}<form {{ flat_attrs|safe }} method="{{ form_method }}" {% if form.is_multipart %} enctype="multipart/form-data"{% endif %}>{% endif %}
{% if form_method|lower == 'post' and not disable_csrf %}
{% csrf_token %}
{% endif %}
{% include "bootstrap3/display_form.html" %}
{% if inputs %}
<div class="form-group form-actions">
{% if label_class %}
<div class="controls col-lg-offset-{{ label_size }} {{ label_class }}"></div>
{% endif %}
<div class="controls {{ field_class }}">
{% for input in inputs %}
{% include "bootstrap3/layout/baseinput.html" %}
{% endfor %}
</div>
</div>
{% endif %}
{% if form_tag %}</form>{% endif %}
{% endspecialspaceless %}
| {% load crispy_forms_utils %}
{% specialspaceless %}
{% if form_tag %}<form {{ flat_attrs|safe }} method="{{ form_method }}" {% if form.is_multipart %} enctype="multipart/form-data"{% endif %}>{% endif %}
{% if form_method|lower == 'post' and not disable_csrf %}
{% csrf_token %}
{% endif %}
{% include "bootstrap3/display_form.html" %}
{% if inputs %}
<div class="form-group form-actions">
{% if label_class %}
<div class="aab controls {{ label_class }}"></div>
{% endif %}
<div class="controls {{ field_class }}">
{% for input in inputs %}
{% include "bootstrap3/layout/baseinput.html" %}
{% endfor %}
</div>
</div>
{% endif %}
{% if form_tag %}</form>{% endif %}
{% endspecialspaceless %}
| Fix for button alignment for non-lg bootstrap classes | Fix for button alignment for non-lg bootstrap classes
| HTML | mit | scuml/django-crispy-forms,IanLee1521/django-crispy-forms,zixan/django-crispy-forms,impulse-cloud/django-crispy-forms,maraujop/django-crispy-forms,davidszotten/django-crispy-forms,damienjones/django-crispy-forms,jtyoung/django-crispy-forms,uranusjr/django-crispy-forms-ng,spectras/django-crispy-forms,maraujop/django-crispy-forms,treyhunner/django-crispy-forms,smirolo/django-crispy-forms,bouttier/django-crispy-forms,dessibelle/django-crispy-forms,RamezIssac/django-crispy-forms,dessibelle/django-crispy-forms,ngenovictor/django-crispy-forms,iedparis8/django-crispy-forms,iris-edu/django-crispy-forms,smirolo/django-crispy-forms,davidszotten/django-crispy-forms,django-crispy-forms/django-crispy-forms,rfleschenberg/django-crispy-forms,iedparis8/django-crispy-forms,dzhuang/django-crispy-forms,VishvajitP/django-crispy-forms,saydulk/django-crispy-forms,iris-edu-int/django-crispy-forms,rfleschenberg/django-crispy-forms,jtyoung/django-crispy-forms,alanwj/django-crispy-forms,RamezIssac/django-crispy-forms,spectras/django-crispy-forms,scuml/django-crispy-forms,agepoly/django-crispy-forms,alanwj/django-crispy-forms,schrd/django-crispy-forms,jcomeauictx/django-crispy-forms,IanLee1521/django-crispy-forms,dzhuang/django-crispy-forms,eykanal/django-crispy-forms,Stranger6667/django-crispy-forms,ngenovictor/django-crispy-forms,impulse-cloud/django-crispy-forms,jcomeauictx/django-crispy-forms,carltongibson/django-crispy-forms,uranusjr/django-crispy-forms-ng,tarunlnmiit/django-crispy-forms,schrd/django-crispy-forms,avsd/django-crispy-forms,bouttier/django-crispy-forms,damienjones/django-crispy-forms,Stranger6667/django-crispy-forms,eykanal/django-crispy-forms,tarunlnmiit/django-crispy-forms,django-crispy-forms/django-crispy-forms,VishvajitP/django-crispy-forms,zixan/django-crispy-forms,iris-edu/django-crispy-forms,saydulk/django-crispy-forms,iris-edu-int/django-crispy-forms,agepoly/django-crispy-forms,treyhunner/django-crispy-forms,carltongibson/django-crispy-forms,avsd/django-crispy-forms | html | ## Code Before:
{% load crispy_forms_utils %}
{% specialspaceless %}
{% if form_tag %}<form {{ flat_attrs|safe }} method="{{ form_method }}" {% if form.is_multipart %} enctype="multipart/form-data"{% endif %}>{% endif %}
{% if form_method|lower == 'post' and not disable_csrf %}
{% csrf_token %}
{% endif %}
{% include "bootstrap3/display_form.html" %}
{% if inputs %}
<div class="form-group form-actions">
{% if label_class %}
<div class="controls col-lg-offset-{{ label_size }} {{ label_class }}"></div>
{% endif %}
<div class="controls {{ field_class }}">
{% for input in inputs %}
{% include "bootstrap3/layout/baseinput.html" %}
{% endfor %}
</div>
</div>
{% endif %}
{% if form_tag %}</form>{% endif %}
{% endspecialspaceless %}
## Instruction:
Fix for button alignment for non-lg bootstrap classes
## Code After:
{% load crispy_forms_utils %}
{% specialspaceless %}
{% if form_tag %}<form {{ flat_attrs|safe }} method="{{ form_method }}" {% if form.is_multipart %} enctype="multipart/form-data"{% endif %}>{% endif %}
{% if form_method|lower == 'post' and not disable_csrf %}
{% csrf_token %}
{% endif %}
{% include "bootstrap3/display_form.html" %}
{% if inputs %}
<div class="form-group form-actions">
{% if label_class %}
<div class="aab controls {{ label_class }}"></div>
{% endif %}
<div class="controls {{ field_class }}">
{% for input in inputs %}
{% include "bootstrap3/layout/baseinput.html" %}
{% endfor %}
</div>
</div>
{% endif %}
{% if form_tag %}</form>{% endif %}
{% endspecialspaceless %}
| {% load crispy_forms_utils %}
{% specialspaceless %}
{% if form_tag %}<form {{ flat_attrs|safe }} method="{{ form_method }}" {% if form.is_multipart %} enctype="multipart/form-data"{% endif %}>{% endif %}
{% if form_method|lower == 'post' and not disable_csrf %}
{% csrf_token %}
{% endif %}
{% include "bootstrap3/display_form.html" %}
{% if inputs %}
<div class="form-group form-actions">
{% if label_class %}
- <div class="controls col-lg-offset-{{ label_size }} {{ label_class }}"></div>
? -------------------------------
+ <div class="aab controls {{ label_class }}"></div>
? ++++
{% endif %}
<div class="controls {{ field_class }}">
{% for input in inputs %}
{% include "bootstrap3/layout/baseinput.html" %}
{% endfor %}
</div>
</div>
{% endif %}
{% if form_tag %}</form>{% endif %}
{% endspecialspaceless %} | 2 | 0.076923 | 1 | 1 |
925a345dd2e2202b362c02ab5e381039cd6c5cb6 | ansible/roles/gcp_packages/tasks/main.yml | ansible/roles/gcp_packages/tasks/main.yml | ---
- name: install packages
yum:
name: "{{ item }}"
state: installed
with_items:
- python-gcloud
| ---
- name: install packages
yum:
name: "{{ item }}"
state: installed
with_items:
- python-gcloud
- gcloud
| Install the gcloud package which has gcloud cli | Install the gcloud package which has gcloud cli
| YAML | apache-2.0 | themurph/openshift-tools,openshift/openshift-tools,ivanhorvath/openshift-tools,blrm/openshift-tools,blrm/openshift-tools,ivanhorvath/openshift-tools,drewandersonnz/openshift-tools,joelsmith/openshift-tools,andrewklau/openshift-tools,blrm/openshift-tools,themurph/openshift-tools,andrewklau/openshift-tools,joelddiaz/openshift-tools,jupierce/openshift-tools,ivanhorvath/openshift-tools,joelddiaz/openshift-tools,rhdedgar/openshift-tools,jupierce/openshift-tools,blrm/openshift-tools,blrm/openshift-tools,drewandersonnz/openshift-tools,themurph/openshift-tools,twiest/openshift-tools,tiwillia/openshift-tools,ivanhorvath/openshift-tools,drewandersonnz/openshift-tools,jupierce/openshift-tools,ivanhorvath/openshift-tools,rhdedgar/openshift-tools,openshift/openshift-tools,jupierce/openshift-tools,drewandersonnz/openshift-tools,joelddiaz/openshift-tools,twiest/openshift-tools,twiest/openshift-tools,andrewklau/openshift-tools,openshift/openshift-tools,drewandersonnz/openshift-tools,openshift/openshift-tools,andrewklau/openshift-tools,andrewklau/openshift-tools,twiest/openshift-tools,joelsmith/openshift-tools,twiest/openshift-tools,blrm/openshift-tools,tiwillia/openshift-tools,tiwillia/openshift-tools,twiest/openshift-tools,joelddiaz/openshift-tools,themurph/openshift-tools,rhdedgar/openshift-tools,joelsmith/openshift-tools,tiwillia/openshift-tools,openshift/openshift-tools,ivanhorvath/openshift-tools,rhdedgar/openshift-tools,jupierce/openshift-tools,openshift/openshift-tools,joelddiaz/openshift-tools,joelsmith/openshift-tools,themurph/openshift-tools,drewandersonnz/openshift-tools,tiwillia/openshift-tools,rhdedgar/openshift-tools,joelddiaz/openshift-tools | yaml | ## Code Before:
---
- name: install packages
yum:
name: "{{ item }}"
state: installed
with_items:
- python-gcloud
## Instruction:
Install the gcloud package which has gcloud cli
## Code After:
---
- name: install packages
yum:
name: "{{ item }}"
state: installed
with_items:
- python-gcloud
- gcloud
| ---
- name: install packages
yum:
name: "{{ item }}"
state: installed
with_items:
- python-gcloud
+ - gcloud | 1 | 0.142857 | 1 | 0 |
d87643e8c1a944098b021c0f70f3d5531c7d871d | l10n/zh-TW/chrome.properties | l10n/zh-TW/chrome.properties | unsupported_feature=此 PDF 文件可能並未正確顯示。
open_with_different_viewer=使用其他檢視器開啟
open_with_different_viewer.accessKey=o
| unsupported_feature=此 PDF 文件可能無法正確顯示。
unsupported_feature_forms=此 PDF 文件包含表單,不支持表單欄位的填寫。
open_with_different_viewer=使用其他檢視器開啟
open_with_different_viewer.accessKey=o
| Add form warning and minor translation update for zh-TW | Add form warning and minor translation update for zh-TW
| INI | apache-2.0 | sitexa/pdf.js,selique/pdf.js,petercpg/pdf.js,douglasvegas/pdf.js,declara/pdf.js,joelkuiper/pdf.js,iamkhush/pdf.js,thejdeep/pdf.js,fashionsun/pdf.js,Flekyno/pdfjs,xialei/pdf.js,bobwol/pdf.js,erikdejonge/pdf.js,yp2/pdf.js,ShotaArai/gwitsch-pdf.js,humphd/pdf.js,fkaelberer/pdf.js,sitexa/pdf.js,parthiban019/pdf.js,KlausRaynor/pdf.js,iesl/pdf.js,shaowei-su/pdf.js,savinn/pdf.js,CodingFabian/pdf.js,sunilomrey/pdf.js,viveksjain/pdf.js,1and1/pdf.js,jibaro/pdf.js,zyfran/pdf.js,wuhuizuo/pdf.js,kalley/pdf.js,bhanu475/pdf.js,youprofit/pdf.js,WoundCentrics/pdf.js,nkpgardose/pdf.js,vivin/pdf.js,ynteng/pdf.js,microcom/pdf.js,KlausRaynor/pdf.js,zyfran/pdf.js,brendandahl/pdf.js,Lhuihui/pdf.js,sunilomrey/pdf.js,mcanthony/pdf.js,NitroLabs/pdf.js,ebsco/pdf.js,WoundCentrics/pdf.js,DavidPrevot/pdf.js,a0preetham/pdf.js,sdmbuild/pdf.js,barrytielkes/pdf.js,allenmo/pdf.js,reggersusa/pdf.js,ShiYw/pdf.js,xialei/pdf.js,pramodhkp/pdf.js-gsoc-2014,thepulkitagarwal/pdf.js,allenmo/pdf.js,ShotaArai/gwitsch-pdf.js,dirkliu/pdf.js,Rob--W/pdf.js,JasonMPE/pdf.js,Ricardh522/pdf.js,humphd/pdf.js,mdamt/pdf.js,mysterlune/pdf.js,adlerweb/pdf.js-Touch,redanium/pdf.js,tinderboxhq/pdf.js,drosi94/pdf.js,ynteng/pdf.js,yp2/pdf.js,ebsco/pdf.js,mauricionr/pdf.js,bhanu475/pdf.js,CodingFabian/pdf.js,THausherr/pdf.js,mukulmishra18/pdf.js,sdmbuild/pdf.js,dannydes/WebSeniorReader,NitroLabs/pdf.js,calexandrepcjr/pdf.js,elcorders/pdf.js,youprofit/pdf.js,happybelly/pdf.js,zwily/pdf.js,joelkuiper/pdf.js,CodingFabian/pdf.js,kalley/pdf.js,yurydelendik/pdf.js,caremerge/pdf.js,mweimerskirch/pdf.js,jlegewie/pdf.js,timvandermeij/pdf.js,vivin/pdf.js,mcanthony/pdf.js,selique/pdf.js,eddyzhuo/pdf.js,zwily/pdf.js,ajrulez/pdf.js,Mendeley/pdf.js,gigaga/pdf.js,zhaosichao/pdf.js,bobwol/pdf.js,viveksjain/pdf.js,Flipkart/pdf.js,ynteng/pdf.js,mcanthony/pdf.js,wuhuizuo/pdf.js,viveksjain/pdf.js,yuriyua/pdf.js,sdmbuild/pdf.js,petercpg/pdf.js,selique/pdf.js,harshavardhana/pdf.js,MASmedios/pdf.js,sunilomrey/pdf.js,reggersusa/pdf.js,Thunderforge/pdf.js,mozilla/pdf.js,Snuffleupagus/pdf.js,mcanthony/pdf.js,dsprenkels/pdf.js,ynteng/pdf.js,DORNINEM/pdf.js,OHIF/pdf.js,luiseduardohdbackup/pdf.js,ajrulez/pdf.js,iamkhush/pdf.js,JasonMPE/pdf.js,yp2/pdf.js,mweimerskirch/pdf.js,douglasvegas/pdf.js,caremerge/pdf.js,mauricionr/pdf.js,xpati/pdf.js,redanium/pdf.js,bobwol/pdf.js,DORNINEM/pdf.js,mweimerskirch/pdf.js,allenmo/pdf.js,MASmedios/pdf.js,Ju2ender/pdf.js,allenmo/pdf.js,shaowei-su/pdf.js,iesl/pdf.js,kalley/pdf.js,ShiYw/pdf.js,sitexa/pdf.js,Ju2ender/pdf.js,Snuffleupagus/pdf.js,eddyzhuo/pdf.js,sunilomrey/pdf.js,parthiban019/pdf.js,tmarrinan/pdf.js,THausherr/pdf.js,1and1/pdf.js,mauricionr/pdf.js,Dawnflying/pdf.js,sdmbuild/pdf.js,douglasvegas/pdf.js,KlausRaynor/pdf.js,calexandrepcjr/pdf.js,ShiYw/pdf.js,gmcdowell/pdf.js,xpati/pdf.js,maragh/pdf.js,jokamjohn/pdf.js,bobwol/pdf.js,DORNINEM/pdf.js,Ju2ender/pdf.js,HongwuLin/Pdf.js,MultivitaminLLC/pdf.js,Datacom/pdf.js,Ricardh522/pdf.js,deenjohn/pdf.js,DORNINEM/pdf.js,tmarrinan/pdf.js,zwily/pdf.js,yuriyua/pdf.js,Ricardh522/pdf.js,THausherr/pdf.js,mauricionr/pdf.js,barrytielkes/pdf.js,jazzy-em/pdf.js,zhaosichao/pdf.js,calexandrepcjr/pdf.js,zyfran/pdf.js,KamiHQ/pdf.js,Ricardh522/pdf.js,haojiejie/pdf.js,xialei/pdf.js,Shoobx/pdf.js,shaowei-su/pdf.js,bhanu475/pdf.js,sdmbuild/pdf.js,WhitesteinTechnologies/wt-vaadin-pdf.js,HongwuLin/Pdf.js,sitexa/pdf.js,eddyzhuo/pdf.js,elcorders/pdf.js,tinderboxhq/pdf.js,bh213/pdf.js,dirkliu/pdf.js,a0preetham/pdf.js,viveksjain/pdf.js,shaowei-su/pdf.js,KamiHQ/pdf.js,erikdejonge/pdf.js,pairyo/pdf.js,MASmedios/pdf.js,joelkuiper/pdf.js,NitroLabs/pdf.js,luiseduardohdbackup/pdf.js,Mendeley/pdf.js,JasonMPE/pdf.js,Ricardh522/pdf.js,petercpg/pdf.js,elcorders/pdf.js,Flipkart/pdf.js,calexandrepcjr/pdf.js,Mendeley/pdf.js,ShotaArai/gwitsch-pdf.js,ajrulez/pdf.js,Snuffleupagus/pdf.js,reggersusa/pdf.js,petercpg/pdf.js,nawawi/pdf.js,msarti/pdf.js,xavier114fch/pdf.js,selique/pdf.js,fashionsun/pdf.js,macroplant/pdf.js,anoopelias/pdf.js,zyfran/pdf.js,nkpgardose/pdf.js,maragh/pdf.js,brendandahl/pdf.js,HongwuLin/Pdf.js,Workiva/pdf.js,mainegreen/pdf.js,ShiYw/pdf.js,gmcdowell/pdf.js,EvilTrev/pdf.js,ydfzgyj/pdf.js,mbbaig/pdf.js,nawawi/pdf.js,nirdoshyadav/pdf.js,zhaosichao/pdf.js,douglasvegas/pdf.js,DavidPrevot/pdf.js,jokamjohn/pdf.js,reggersusa/pdf.js,parthiban019/pdf.js,drosi94/pdf.js,Thunderforge/pdf.js,Dawnflying/pdf.js,mmaroti/pdf.js,existentialism/pdf.js,showpad/mozilla-pdf.js,xavier114fch/pdf.js,yp2/pdf.js,Dawnflying/pdf.js,msarti/pdf.js,datalanche/pdf.js,skalnik/pdf.js,shaowei-su/pdf.js,Flekyno/pdfjs,allenmo/pdf.js,viveksjain/pdf.js,dirkliu/pdf.js,vivin/pdf.js,elcorders/pdf.js,Workiva/pdf.js,yp2/pdf.js,macroplant/pdf.js,jokamjohn/pdf.js,ynteng/pdf.js,mdamt/pdf.js,gmcdowell/pdf.js,ydfzgyj/pdf.js,yurydelendik/pdf.js,msarti/pdf.js,youprofit/pdf.js,THausherr/pdf.js,zwily/pdf.js,HongwuLin/Pdf.js,mweimerskirch/pdf.js,Mendeley/pdf.js,sitexa/pdf.js,Dawnflying/pdf.js,happybelly/pdf.js,gorcz/pdf.js,joelkuiper/pdf.js,datalanche/pdf.js,mozilla/pdf.js,MultivitaminLLC/pdf.js,Lhuihui/pdf.js,parthiban019/pdf.js,ajrulez/pdf.js,deenjohn/pdf.js,THausherr/pdf.js,Dawnflying/pdf.js,ShotaArai/gwitsch-pdf.js,pairyo/pdf.js,dsprenkels/pdf.js,mcanthony/pdf.js,youprofit/pdf.js,haojiejie/pdf.js,declara/pdf.js,mmaroti/pdf.js,zwily/pdf.js,elcorders/pdf.js,ebsco/pdf.js,happybelly/pdf.js,fashionsun/pdf.js,pairyo/pdf.js,kalley/pdf.js,jokamjohn/pdf.js,reggersusa/pdf.js,Ju2ender/pdf.js,CodingFabian/pdf.js,luiseduardohdbackup/pdf.js,existentialism/pdf.js,dannydes/WebSeniorReader,mysterlune/pdf.js,fashionsun/pdf.js,EvilTrev/pdf.js,deenjohn/pdf.js,pairyo/pdf.js,bhanu475/pdf.js,yuriyua/pdf.js,luiseduardohdbackup/pdf.js,mdamt/pdf.js,jasonjensen/pdf.js,tathata/pdf.js,microcom/pdf.js,redanium/pdf.js,KlausRaynor/pdf.js,luiseduardohdbackup/pdf.js,sunilomrey/pdf.js,happybelly/pdf.js,bhanu475/pdf.js,showpad/mozilla-pdf.js,Thunderforge/pdf.js,mdamt/pdf.js,savinn/pdf.js,JasonMPE/pdf.js,jibaro/pdf.js,DavidPrevot/pdf.js,MASmedios/pdf.js,DORNINEM/pdf.js,joelkuiper/pdf.js,bobwol/pdf.js,Shoobx/pdf.js,Datacom/pdf.js,mbbaig/pdf.js,maragh/pdf.js,eddyzhuo/pdf.js,nkpgardose/pdf.js,pairyo/pdf.js,gmcdowell/pdf.js,Mendeley/pdf.js,mdamt/pdf.js,selique/pdf.js,xialei/pdf.js,ajrulez/pdf.js,Rob--W/pdf.js,Ju2ender/pdf.js,haojiejie/pdf.js,declara/pdf.js,KlausRaynor/pdf.js,HongwuLin/Pdf.js,haojiejie/pdf.js,yuriyua/pdf.js,erikdejonge/pdf.js,websirnik/pdf.js,deenjohn/pdf.js,Lhuihui/pdf.js,humphd/pdf.js,mauricionr/pdf.js,tathata/pdf.js,wuhuizuo/pdf.js,anoopelias/pdf.js,Datacom/pdf.js,Thunderforge/pdf.js,nkpgardose/pdf.js,savinn/pdf.js,Lhuihui/pdf.js,redanium/pdf.js,jibaro/pdf.js,macroplant/pdf.js,sdmbuild/pdf.js,nawawi/pdf.js,gmcdowell/pdf.js,humphd/pdf.js,xpati/pdf.js,humphd/pdf.js,msarti/pdf.js,xialei/pdf.js,wuhuizuo/pdf.js,jibaro/pdf.js,Lhuihui/pdf.js,xpati/pdf.js,mweimerskirch/pdf.js,WoundCentrics/pdf.js,mmaroti/pdf.js,gorcz/pdf.js,xpati/pdf.js,msarti/pdf.js,haojiejie/pdf.js,petercpg/pdf.js,yuriyua/pdf.js,WoundCentrics/pdf.js,Datacom/pdf.js,thepulkitagarwal/pdf.js,timvandermeij/pdf.js,showpad/mozilla-pdf.js,ebsco/pdf.js,NitroLabs/pdf.js,gorcz/pdf.js,jazzy-em/pdf.js,erikdejonge/pdf.js,jokamjohn/pdf.js,JasonMPE/pdf.js,Shoobx/pdf.js,timvandermeij/pdf.js,ShiYw/pdf.js,zyfran/pdf.js,deenjohn/pdf.js,zhaosichao/pdf.js,barrytielkes/pdf.js,tmarrinan/pdf.js,mukulmishra18/pdf.js,DavidPrevot/pdf.js,barrytielkes/pdf.js,mainegreen/pdf.js,gigaga/pdf.js,MASmedios/pdf.js,erikdejonge/pdf.js,WoundCentrics/pdf.js,existentialism/pdf.js,kalley/pdf.js,maragh/pdf.js,jibaro/pdf.js,nirdoshyadav/pdf.js,parthiban019/pdf.js,barrytielkes/pdf.js,reggersusa/pdf.js,douglasvegas/pdf.js,calexandrepcjr/pdf.js,fkaelberer/pdf.js,maragh/pdf.js,erikdejonge/pdf.js,Flekyno/pdfjs,operasoftware/pdf.js,CodingFabian/pdf.js,savinn/pdf.js,mmaroti/pdf.js,ebsco/pdf.js,DavidPrevot/pdf.js,Datacom/pdf.js,wuhuizuo/pdf.js,WhitesteinTechnologies/wt-vaadin-pdf.js,websirnik/pdf.js,redanium/pdf.js,adlerweb/pdf.js-Touch,fashionsun/pdf.js,eddyzhuo/pdf.js,jasonjensen/pdf.js,nkpgardose/pdf.js,ShotaArai/gwitsch-pdf.js,bh213/pdf.js,savinn/pdf.js,mbbaig/pdf.js,Thunderforge/pdf.js,zhaosichao/pdf.js,skalnik/pdf.js,OHIF/pdf.js,youprofit/pdf.js,mozilla/pdf.js,happybelly/pdf.js,ShotaArai/gwitsch-pdf.js,mmaroti/pdf.js | ini | ## Code Before:
unsupported_feature=此 PDF 文件可能並未正確顯示。
open_with_different_viewer=使用其他檢視器開啟
open_with_different_viewer.accessKey=o
## Instruction:
Add form warning and minor translation update for zh-TW
## Code After:
unsupported_feature=此 PDF 文件可能無法正確顯示。
unsupported_feature_forms=此 PDF 文件包含表單,不支持表單欄位的填寫。
open_with_different_viewer=使用其他檢視器開啟
open_with_different_viewer.accessKey=o
| - unsupported_feature=此 PDF 文件可能並未正確顯示。
? ^^
+ unsupported_feature=此 PDF 文件可能無法正確顯示。
? ^^
+ unsupported_feature_forms=此 PDF 文件包含表單,不支持表單欄位的填寫。
open_with_different_viewer=使用其他檢視器開啟
open_with_different_viewer.accessKey=o
| 3 | 0.75 | 2 | 1 |
c1c897078398353af66402da20c0e5f23c9c21e4 | redux/articleList.js | redux/articleList.js | import { createDuck } from 'redux-duck'
import { fromJS } from 'immutable'
import gql from '../util/GraphQL'
const articleList = createDuck('articleList');
// Action Types
//
const LOAD = articleList.defineType('LOAD');
// Action creators
//
export const load = () => dispatch =>
gql`{
ListArticles {
id
text
}
}`().then((resp) => {
dispatch(articleList.createAction(LOAD)(resp.getIn(['data', 'ListArticles'])));
})
// Reducer
//
const initialState = fromJS({
state: {isLoading: false},
data: null,
});
export default articleList.createReducer({
[LOAD]: (state, {payload}) => state.set('data', payload),
}, initialState)
| import { createDuck } from 'redux-duck'
import { fromJS, List } from 'immutable'
import gql from '../util/GraphQL'
const articleList = createDuck('articleList');
// Action Types
//
const LOAD = articleList.defineType('LOAD');
// Action creators
//
export const load = () => dispatch =>
gql`{
ListArticles {
edges {
node {
id
text
}
}
}
}`().then((resp) => {
dispatch(articleList.createAction(LOAD)(resp.getIn(['data', 'ListArticles', 'edges'], List()).map(edge => edge.get('node'))));
})
// Reducer
//
const initialState = fromJS({
state: {isLoading: false},
data: null,
});
export default articleList.createReducer({
[LOAD]: (state, {payload}) => state.set('data', payload),
}, initialState)
| Fix article list API call | Fix article list API call | JavaScript | mit | cofacts/rumors-site,cofacts/rumors-site | javascript | ## Code Before:
import { createDuck } from 'redux-duck'
import { fromJS } from 'immutable'
import gql from '../util/GraphQL'
const articleList = createDuck('articleList');
// Action Types
//
const LOAD = articleList.defineType('LOAD');
// Action creators
//
export const load = () => dispatch =>
gql`{
ListArticles {
id
text
}
}`().then((resp) => {
dispatch(articleList.createAction(LOAD)(resp.getIn(['data', 'ListArticles'])));
})
// Reducer
//
const initialState = fromJS({
state: {isLoading: false},
data: null,
});
export default articleList.createReducer({
[LOAD]: (state, {payload}) => state.set('data', payload),
}, initialState)
## Instruction:
Fix article list API call
## Code After:
import { createDuck } from 'redux-duck'
import { fromJS, List } from 'immutable'
import gql from '../util/GraphQL'
const articleList = createDuck('articleList');
// Action Types
//
const LOAD = articleList.defineType('LOAD');
// Action creators
//
export const load = () => dispatch =>
gql`{
ListArticles {
edges {
node {
id
text
}
}
}
}`().then((resp) => {
dispatch(articleList.createAction(LOAD)(resp.getIn(['data', 'ListArticles', 'edges'], List()).map(edge => edge.get('node'))));
})
// Reducer
//
const initialState = fromJS({
state: {isLoading: false},
data: null,
});
export default articleList.createReducer({
[LOAD]: (state, {payload}) => state.set('data', payload),
}, initialState)
| import { createDuck } from 'redux-duck'
- import { fromJS } from 'immutable'
+ import { fromJS, List } from 'immutable'
? ++++++
import gql from '../util/GraphQL'
const articleList = createDuck('articleList');
// Action Types
//
const LOAD = articleList.defineType('LOAD');
// Action creators
//
export const load = () => dispatch =>
gql`{
ListArticles {
+ edges {
+ node {
- id
+ id
? ++++
- text
+ text
? ++++
+ }
+ }
}
}`().then((resp) => {
- dispatch(articleList.createAction(LOAD)(resp.getIn(['data', 'ListArticles'])));
+ dispatch(articleList.createAction(LOAD)(resp.getIn(['data', 'ListArticles', 'edges'], List()).map(edge => edge.get('node'))));
? +++++++++ ++++++++++++++++++++++++++++++++++++++
})
// Reducer
//
const initialState = fromJS({
state: {isLoading: false},
data: null,
});
export default articleList.createReducer({
[LOAD]: (state, {payload}) => state.set('data', payload),
}, initialState) | 12 | 0.342857 | 8 | 4 |
4424e10e32d69c3ed93e6b432c6bb1d235ea146f | src/js/reducers/index.js | src/js/reducers/index.js | import { createStore, combineReducers } from 'redux';
import header from './header';
import device from './device';
const types = {
ACTIVE: 'module_active',
LOGIN: 'app_login'
};
export { types }; // MUST put before reducers define
const reducers = combineReducers({
header,
device
});
export default createStore(reducers);
| import { createStore, combineReducers, applyMiddleware } from 'redux';
import thunkMiddleware from 'redux-thunk';
import header from './header';
import device from './device';
const types = {
ACTIVE: 'module_active',
LOGIN: 'app_login'
};
export { types }; // MUST put before reducers define
const reducers = combineReducers({
header,
device
});
export default createStore(reducers, applyMiddleware(thunkMiddleware));
| Introduce redux-thunk for async actions | Introduce redux-thunk for async actions
| JavaScript | mit | masogit/dashboard,masogit/dashboard | javascript | ## Code Before:
import { createStore, combineReducers } from 'redux';
import header from './header';
import device from './device';
const types = {
ACTIVE: 'module_active',
LOGIN: 'app_login'
};
export { types }; // MUST put before reducers define
const reducers = combineReducers({
header,
device
});
export default createStore(reducers);
## Instruction:
Introduce redux-thunk for async actions
## Code After:
import { createStore, combineReducers, applyMiddleware } from 'redux';
import thunkMiddleware from 'redux-thunk';
import header from './header';
import device from './device';
const types = {
ACTIVE: 'module_active',
LOGIN: 'app_login'
};
export { types }; // MUST put before reducers define
const reducers = combineReducers({
header,
device
});
export default createStore(reducers, applyMiddleware(thunkMiddleware));
| - import { createStore, combineReducers } from 'redux';
+ import { createStore, combineReducers, applyMiddleware } from 'redux';
? +++++++++++++++++
+ import thunkMiddleware from 'redux-thunk';
import header from './header';
import device from './device';
const types = {
ACTIVE: 'module_active',
LOGIN: 'app_login'
};
export { types }; // MUST put before reducers define
const reducers = combineReducers({
header,
device
});
- export default createStore(reducers);
+ export default createStore(reducers, applyMiddleware(thunkMiddleware)); | 5 | 0.294118 | 3 | 2 |
c7bec2a16f80b5b4f6a2de2cc41ad36dc2b5d492 | app/widgets/Presence/presence.js | app/widgets/Presence/presence.js | var Presence = {
clearQuick : function() {
localStorage.removeItem('quickDeviceId');
localStorage.removeItem('quickLogin');
localStorage.removeItem('quickHost');
localStorage.removeItem('quickKey');
}
}
MovimWebsocket.initiate(() => Presence_ajaxHttpGetPresence());
| var Presence = {
clearQuick : function() {
localStorage.removeItem('quickDeviceId');
localStorage.removeItem('quickLogin');
localStorage.removeItem('quickHost');
localStorage.removeItem('quickKey');
},
setFirebaseToken : function(token) {
Presence_ajaxSetFireBaseToken(token);
Android.clearFirebaseToken();
}
}
MovimWebsocket.initiate(() => Presence_ajaxHttpGetPresence());
| Clear the app token once it is sent to Movim | Clear the app token once it is sent to Movim
| JavaScript | agpl-3.0 | movim/movim,movim/movim,movim/movim | javascript | ## Code Before:
var Presence = {
clearQuick : function() {
localStorage.removeItem('quickDeviceId');
localStorage.removeItem('quickLogin');
localStorage.removeItem('quickHost');
localStorage.removeItem('quickKey');
}
}
MovimWebsocket.initiate(() => Presence_ajaxHttpGetPresence());
## Instruction:
Clear the app token once it is sent to Movim
## Code After:
var Presence = {
clearQuick : function() {
localStorage.removeItem('quickDeviceId');
localStorage.removeItem('quickLogin');
localStorage.removeItem('quickHost');
localStorage.removeItem('quickKey');
},
setFirebaseToken : function(token) {
Presence_ajaxSetFireBaseToken(token);
Android.clearFirebaseToken();
}
}
MovimWebsocket.initiate(() => Presence_ajaxHttpGetPresence());
| var Presence = {
clearQuick : function() {
localStorage.removeItem('quickDeviceId');
localStorage.removeItem('quickLogin');
localStorage.removeItem('quickHost');
localStorage.removeItem('quickKey');
+ },
+ setFirebaseToken : function(token) {
+ Presence_ajaxSetFireBaseToken(token);
+ Android.clearFirebaseToken();
}
}
MovimWebsocket.initiate(() => Presence_ajaxHttpGetPresence()); | 4 | 0.4 | 4 | 0 |
dc376a44e2b7dcc61575e2a5b4797cb4b6063622 | README.md | README.md |
Jordan Suchow's portfolio of research and other work.
[](https://magnum.travis-ci.com/suchow/suchow.io)
|
Jordan Suchow's portfolio of research and other work.
[](https://travis-ci.org/suchow/suchow.io)
| Use Travis CI .org, not .com for build status | Use Travis CI .org, not .com for build status
| Markdown | mit | suchow/suchow.io,suchow/suchow.io,suchow/suchow.io | markdown | ## Code Before:
Jordan Suchow's portfolio of research and other work.
[](https://magnum.travis-ci.com/suchow/suchow.io)
## Instruction:
Use Travis CI .org, not .com for build status
## Code After:
Jordan Suchow's portfolio of research and other work.
[](https://travis-ci.org/suchow/suchow.io)
|
Jordan Suchow's portfolio of research and other work.
- [](https://magnum.travis-ci.com/suchow/suchow.io)
? ------- - ^ --------------------------- ------- - ^
+ [](https://travis-ci.org/suchow/suchow.io)
? ^^ ^^
| 2 | 0.5 | 1 | 1 |
2ccf34093dee4d35f7299b3542ab33e47ab112ec | tests/publisher_structures.txt | tests/publisher_structures.txt | dfid does not use transaction at activity hierarchy 1
dfid does not use description at activity hierarchy 2
dfid does not use document-link at activity hierarchy 2
dfid does not use sector at activity hierarchy 1
dfid does not use transaction at activity hierarchy 1
dfid does not use recipient-country at activity hierarchy 1
dfid does not use recipient-region at activity hierarchy 1
| GB-1 does not use transaction at activity hierarchy 1
GB-1 does not use description at activity hierarchy 2
GB-1 does not use document-link at activity hierarchy 2
GB-1 does not use sector at activity hierarchy 1
GB-1 does not use transaction at activity hierarchy 1
GB-1 does not use recipient-country at activity hierarchy 1
GB-1 does not use recipient-region at activity hierarchy 1
| Change publisher identifier to organisation identifiers in publisher structures | Change publisher identifier to organisation identifiers in publisher structures
| Text | agpl-3.0 | pwyf/IATI-Data-Quality,pwyf/IATI-Data-Quality,pwyf/IATI-Data-Quality,pwyf/IATI-Data-Quality | text | ## Code Before:
dfid does not use transaction at activity hierarchy 1
dfid does not use description at activity hierarchy 2
dfid does not use document-link at activity hierarchy 2
dfid does not use sector at activity hierarchy 1
dfid does not use transaction at activity hierarchy 1
dfid does not use recipient-country at activity hierarchy 1
dfid does not use recipient-region at activity hierarchy 1
## Instruction:
Change publisher identifier to organisation identifiers in publisher structures
## Code After:
GB-1 does not use transaction at activity hierarchy 1
GB-1 does not use description at activity hierarchy 2
GB-1 does not use document-link at activity hierarchy 2
GB-1 does not use sector at activity hierarchy 1
GB-1 does not use transaction at activity hierarchy 1
GB-1 does not use recipient-country at activity hierarchy 1
GB-1 does not use recipient-region at activity hierarchy 1
| - dfid does not use transaction at activity hierarchy 1
? ^^^^
+ GB-1 does not use transaction at activity hierarchy 1
? ^^^^
- dfid does not use description at activity hierarchy 2
? ^^^^
+ GB-1 does not use description at activity hierarchy 2
? ^^^^
- dfid does not use document-link at activity hierarchy 2
? ^^^^
+ GB-1 does not use document-link at activity hierarchy 2
? ^^^^
- dfid does not use sector at activity hierarchy 1
? ^^^^
+ GB-1 does not use sector at activity hierarchy 1
? ^^^^
- dfid does not use transaction at activity hierarchy 1
? ^^^^
+ GB-1 does not use transaction at activity hierarchy 1
? ^^^^
- dfid does not use recipient-country at activity hierarchy 1
? ^^^^
+ GB-1 does not use recipient-country at activity hierarchy 1
? ^^^^
- dfid does not use recipient-region at activity hierarchy 1
? ^^^^
+ GB-1 does not use recipient-region at activity hierarchy 1
? ^^^^
| 14 | 2 | 7 | 7 |
03efa19812ec144ba2941d6d471c0369fe5533ab | packer_templates/windows/cookbooks/packer/recipes/vm_tools.rb | packer_templates/windows/cookbooks/packer/recipes/vm_tools.rb | if vbox?
directory 'C:/Windows/Temp/virtualbox' do
recursive true
end
powershell_script 'install vbox guest additions' do
code <<-EOH
Get-ChildItem E:/cert/ -Filter vbox*.cer | ForEach-Object {
E:/cert/VBoxCertUtil.exe add-trusted-publisher $_.FullName --root $_.FullName
}
Start-Process -FilePath "e:/VBoxWindowsAdditions.exe" -ArgumentList "/S" -WorkingDirectory "C:/Windows/Temp/virtualbox" -Wait
EOH
ignore_failure true
end
directory 'C:/Windows/Temp/virtualbox' do
action :delete
end
end
| if vbox?
directory 'C:/Windows/Temp/virtualbox' do
recursive true
end
powershell_script 'install vbox guest additions' do
code <<-EOH
Get-ChildItem E:/cert/ -Filter vbox*.cer | ForEach-Object {
E:/cert/VBoxCertUtil.exe add-trusted-publisher $_.FullName --root $_.FullName
}
Start-Process -FilePath "e:/VBoxWindowsAdditions.exe" -ArgumentList "/S" -WorkingDirectory "C:/Windows/Temp/virtualbox" -Wait
EOH
ignore_failure true
end
directory 'C:/Windows/Temp/virtualbox' do
action :delete
end
end
# install vmware tools on vmware guests
# This is from https://github.com/luciusbono/Packer-Windows10/blob/master/install-guest-tools.ps1
if vmware?
powershell_script 'install vbox guest additions' do
code <<-EOH
$isopath = "C:\\Windows\\Temp\\vmware.iso"
Mount-DiskImage -ImagePath $isopath
$exe = ((Get-DiskImage -ImagePath $isopath | Get-Volume).Driveletter + ':\setup.exe')
$parameters = '/S /v "/qr REBOOT=R"'
Dismount-DiskImage -ImagePath $isopath
Remove-Item $isopath
EOH
end
end
| Install vmware tools on windows vmware-iso boxes | Install vmware tools on windows vmware-iso boxes
Signed-off-by: Tim Smith <764ef62106582a09ed09dfa0b6bff7c05fd7d1e4@chef.io>
| Ruby | apache-2.0 | norcams/bento,chef/bento,chef/bento,norcams/bento,norcams/bento,chef/bento | ruby | ## Code Before:
if vbox?
directory 'C:/Windows/Temp/virtualbox' do
recursive true
end
powershell_script 'install vbox guest additions' do
code <<-EOH
Get-ChildItem E:/cert/ -Filter vbox*.cer | ForEach-Object {
E:/cert/VBoxCertUtil.exe add-trusted-publisher $_.FullName --root $_.FullName
}
Start-Process -FilePath "e:/VBoxWindowsAdditions.exe" -ArgumentList "/S" -WorkingDirectory "C:/Windows/Temp/virtualbox" -Wait
EOH
ignore_failure true
end
directory 'C:/Windows/Temp/virtualbox' do
action :delete
end
end
## Instruction:
Install vmware tools on windows vmware-iso boxes
Signed-off-by: Tim Smith <764ef62106582a09ed09dfa0b6bff7c05fd7d1e4@chef.io>
## Code After:
if vbox?
directory 'C:/Windows/Temp/virtualbox' do
recursive true
end
powershell_script 'install vbox guest additions' do
code <<-EOH
Get-ChildItem E:/cert/ -Filter vbox*.cer | ForEach-Object {
E:/cert/VBoxCertUtil.exe add-trusted-publisher $_.FullName --root $_.FullName
}
Start-Process -FilePath "e:/VBoxWindowsAdditions.exe" -ArgumentList "/S" -WorkingDirectory "C:/Windows/Temp/virtualbox" -Wait
EOH
ignore_failure true
end
directory 'C:/Windows/Temp/virtualbox' do
action :delete
end
end
# install vmware tools on vmware guests
# This is from https://github.com/luciusbono/Packer-Windows10/blob/master/install-guest-tools.ps1
if vmware?
powershell_script 'install vbox guest additions' do
code <<-EOH
$isopath = "C:\\Windows\\Temp\\vmware.iso"
Mount-DiskImage -ImagePath $isopath
$exe = ((Get-DiskImage -ImagePath $isopath | Get-Volume).Driveletter + ':\setup.exe')
$parameters = '/S /v "/qr REBOOT=R"'
Dismount-DiskImage -ImagePath $isopath
Remove-Item $isopath
EOH
end
end
| if vbox?
directory 'C:/Windows/Temp/virtualbox' do
recursive true
end
powershell_script 'install vbox guest additions' do
code <<-EOH
Get-ChildItem E:/cert/ -Filter vbox*.cer | ForEach-Object {
E:/cert/VBoxCertUtil.exe add-trusted-publisher $_.FullName --root $_.FullName
}
Start-Process -FilePath "e:/VBoxWindowsAdditions.exe" -ArgumentList "/S" -WorkingDirectory "C:/Windows/Temp/virtualbox" -Wait
EOH
ignore_failure true
end
directory 'C:/Windows/Temp/virtualbox' do
action :delete
end
end
+
+ # install vmware tools on vmware guests
+ # This is from https://github.com/luciusbono/Packer-Windows10/blob/master/install-guest-tools.ps1
+ if vmware?
+ powershell_script 'install vbox guest additions' do
+ code <<-EOH
+ $isopath = "C:\\Windows\\Temp\\vmware.iso"
+ Mount-DiskImage -ImagePath $isopath
+ $exe = ((Get-DiskImage -ImagePath $isopath | Get-Volume).Driveletter + ':\setup.exe')
+ $parameters = '/S /v "/qr REBOOT=R"'
+ Dismount-DiskImage -ImagePath $isopath
+ Remove-Item $isopath
+ EOH
+ end
+ end | 15 | 0.75 | 15 | 0 |
5aa05bc289fac9e942cf19f1b5912aa3ddaf5ad5 | CHANGELOG.md | CHANGELOG.md |
* Update dependencies (no user-facing changes).
## 0.6.1
### New Features
* Added #siblings scope to HasTree concern.
|
* Add Orderable concern.
## 0.6.2
* Update dependencies (no user-facing changes).
## 0.6.1
### New Features
* Added #siblings scope to HasTree concern.
| Update change log for release. | Update change log for release.
| Markdown | mit | sleepingkingstudios/mongoid-sleeping_king_studios | markdown | ## Code Before:
* Update dependencies (no user-facing changes).
## 0.6.1
### New Features
* Added #siblings scope to HasTree concern.
## Instruction:
Update change log for release.
## Code After:
* Add Orderable concern.
## 0.6.2
* Update dependencies (no user-facing changes).
## 0.6.1
### New Features
* Added #siblings scope to HasTree concern.
| +
+ * Add Orderable concern.
+
+ ## 0.6.2
* Update dependencies (no user-facing changes).
## 0.6.1
### New Features
* Added #siblings scope to HasTree concern. | 4 | 0.5 | 4 | 0 |
1631aa8093c4fbdff82b9365318500c5510439f4 | Casks/nzbget.rb | Casks/nzbget.rb | cask :v1 => 'nzbget' do
version '15.0'
sha256 'd4e6fa16d8404a2d1c1e6fe28e4bf4ca4f10d80cf53c724879867d82335167af'
url "https://downloads.sourceforge.net/project/nzbget/nzbget-stable/#{version}/nzbget-#{version}-bin-osx.zip"
name 'NZBGet'
homepage 'https://nzbget.net'
license :gpl
app 'NZBGet.app'
zap :delete => [
'~/Library/Application Support/NZBGet',
'~/Library/Preferences/net.sourceforge.nzbget.plist'
]
end
| cask :v1 => 'nzbget' do
version '16.1'
sha256 '4ec71fa2dcd94c69520eda5c7f43d36e1ab49149bd8b79339da9d5e08f740c8e'
# github.com is the official download host per the vendor homepage
url "https://github.com/nzbget/nzbget/releases/download/v#{version}/nzbget-#{version}-bin-osx.zip"
appcast 'https://github.com/nzbget/nzbget/releases.atom'
name 'NZBGet'
homepage 'http://nzbget.net'
license :gpl
app 'NZBGet.app'
zap :delete => [
'~/Library/Application Support/NZBGet',
'~/Library/Preferences/net.sourceforge.nzbget.plist'
]
end
| Update NZBGet to 16.1 and fix download url. | Update NZBGet to 16.1 and fix download url.
| Ruby | bsd-2-clause | fanquake/homebrew-cask,squid314/homebrew-cask,devmynd/homebrew-cask,dcondrey/homebrew-cask,feigaochn/homebrew-cask,sscotth/homebrew-cask,gerrypower/homebrew-cask,0xadada/homebrew-cask,riyad/homebrew-cask,farmerchris/homebrew-cask,kongslund/homebrew-cask,ksato9700/homebrew-cask,colindunn/homebrew-cask,pacav69/homebrew-cask,hanxue/caskroom,williamboman/homebrew-cask,miguelfrde/homebrew-cask,adrianchia/homebrew-cask,mahori/homebrew-cask,toonetown/homebrew-cask,andrewdisley/homebrew-cask,seanorama/homebrew-cask,mauricerkelly/homebrew-cask,mathbunnyru/homebrew-cask,colindean/homebrew-cask,scottsuch/homebrew-cask,Amorymeltzer/homebrew-cask,rajiv/homebrew-cask,imgarylai/homebrew-cask,crzrcn/homebrew-cask,joshka/homebrew-cask,My2ndAngelic/homebrew-cask,mindriot101/homebrew-cask,FredLackeyOfficial/homebrew-cask,linc01n/homebrew-cask,nathanielvarona/homebrew-cask,deiga/homebrew-cask,forevergenin/homebrew-cask,okket/homebrew-cask,jhowtan/homebrew-cask,jeroenseegers/homebrew-cask,diogodamiani/homebrew-cask,tmoreira2020/homebrew,sscotth/homebrew-cask,helloIAmPau/homebrew-cask,stephenwade/homebrew-cask,danielbayley/homebrew-cask,nrlquaker/homebrew-cask,leipert/homebrew-cask,shorshe/homebrew-cask,dustinblackman/homebrew-cask,bric3/homebrew-cask,stevehedrick/homebrew-cask,elnappo/homebrew-cask,tangestani/homebrew-cask,Gasol/homebrew-cask,andrewdisley/homebrew-cask,ywfwj2008/homebrew-cask,alebcay/homebrew-cask,MerelyAPseudonym/homebrew-cask,hyuna917/homebrew-cask,hanxue/caskroom,syscrusher/homebrew-cask,yumitsu/homebrew-cask,ebraminio/homebrew-cask,sosedoff/homebrew-cask,winkelsdorf/homebrew-cask,mrmachine/homebrew-cask,inz/homebrew-cask,toonetown/homebrew-cask,n8henrie/homebrew-cask,chadcatlett/caskroom-homebrew-cask,samnung/homebrew-cask,tangestani/homebrew-cask,xakraz/homebrew-cask,hovancik/homebrew-cask,mindriot101/homebrew-cask,6uclz1/homebrew-cask,colindean/homebrew-cask,lantrix/homebrew-cask,ianyh/homebrew-cask,mhubig/homebrew-cask,zerrot/homebrew-cask,ksylvan/homebrew-cask,claui/homebrew-cask,miccal/homebrew-cask,jawshooah/homebrew-cask,jalaziz/homebrew-cask,cprecioso/homebrew-cask,riyad/homebrew-cask,renaudguerin/homebrew-cask,kongslund/homebrew-cask,mingzhi22/homebrew-cask,stigkj/homebrew-caskroom-cask,leipert/homebrew-cask,kkdd/homebrew-cask,moimikey/homebrew-cask,neverfox/homebrew-cask,wmorin/homebrew-cask,dwkns/homebrew-cask,hristozov/homebrew-cask,Ibuprofen/homebrew-cask,sjackman/homebrew-cask,scribblemaniac/homebrew-cask,flaviocamilo/homebrew-cask,mlocher/homebrew-cask,xyb/homebrew-cask,exherb/homebrew-cask,markthetech/homebrew-cask,kassi/homebrew-cask,johndbritton/homebrew-cask,forevergenin/homebrew-cask,patresi/homebrew-cask,ninjahoahong/homebrew-cask,phpwutz/homebrew-cask,samdoran/homebrew-cask,hyuna917/homebrew-cask,jmeridth/homebrew-cask,sebcode/homebrew-cask,axodys/homebrew-cask,slack4u/homebrew-cask,lukasbestle/homebrew-cask,brianshumate/homebrew-cask,elyscape/homebrew-cask,alexg0/homebrew-cask,JosephViolago/homebrew-cask,dictcp/homebrew-cask,rogeriopradoj/homebrew-cask,scottsuch/homebrew-cask,xight/homebrew-cask,RJHsiao/homebrew-cask,vitorgalvao/homebrew-cask,tjt263/homebrew-cask,cblecker/homebrew-cask,howie/homebrew-cask,jeroenj/homebrew-cask,miku/homebrew-cask,mchlrmrz/homebrew-cask,gmkey/homebrew-cask,gabrielizaias/homebrew-cask,malob/homebrew-cask,tarwich/homebrew-cask,tedski/homebrew-cask,dwihn0r/homebrew-cask,jonathanwiesel/homebrew-cask,buo/homebrew-cask,inta/homebrew-cask,decrement/homebrew-cask,JosephViolago/homebrew-cask,xcezx/homebrew-cask,schneidmaster/homebrew-cask,xight/homebrew-cask,nathanielvarona/homebrew-cask,n0ts/homebrew-cask,jedahan/homebrew-cask,yutarody/homebrew-cask,samdoran/homebrew-cask,tyage/homebrew-cask,giannitm/homebrew-cask,jellyfishcoder/homebrew-cask,kTitan/homebrew-cask,jbeagley52/homebrew-cask,mingzhi22/homebrew-cask,coeligena/homebrew-customized,josa42/homebrew-cask,KosherBacon/homebrew-cask,buo/homebrew-cask,lcasey001/homebrew-cask,cblecker/homebrew-cask,jalaziz/homebrew-cask,rogeriopradoj/homebrew-cask,CameronGarrett/homebrew-cask,williamboman/homebrew-cask,faun/homebrew-cask,AnastasiaSulyagina/homebrew-cask,bcomnes/homebrew-cask,samshadwell/homebrew-cask,howie/homebrew-cask,anbotero/homebrew-cask,mikem/homebrew-cask,cblecker/homebrew-cask,larseggert/homebrew-cask,MoOx/homebrew-cask,tjnycum/homebrew-cask,daften/homebrew-cask,cobyism/homebrew-cask,0xadada/homebrew-cask,fanquake/homebrew-cask,mikem/homebrew-cask,lvicentesanchez/homebrew-cask,chrisfinazzo/homebrew-cask,puffdad/homebrew-cask,hovancik/homebrew-cask,rickychilcott/homebrew-cask,ptb/homebrew-cask,gurghet/homebrew-cask,goxberry/homebrew-cask,ninjahoahong/homebrew-cask,julionc/homebrew-cask,FinalDes/homebrew-cask,uetchy/homebrew-cask,miccal/homebrew-cask,shorshe/homebrew-cask,cfillion/homebrew-cask,Cottser/homebrew-cask,JacopKane/homebrew-cask,xtian/homebrew-cask,Ngrd/homebrew-cask,lucasmezencio/homebrew-cask,miguelfrde/homebrew-cask,moimikey/homebrew-cask,miccal/homebrew-cask,okket/homebrew-cask,bdhess/homebrew-cask,psibre/homebrew-cask,nathancahill/homebrew-cask,kingthorin/homebrew-cask,markhuber/homebrew-cask,casidiablo/homebrew-cask,victorpopkov/homebrew-cask,amatos/homebrew-cask,blogabe/homebrew-cask,danielbayley/homebrew-cask,janlugt/homebrew-cask,morganestes/homebrew-cask,JacopKane/homebrew-cask,reitermarkus/homebrew-cask,esebastian/homebrew-cask,nshemonsky/homebrew-cask,mjdescy/homebrew-cask,sanyer/homebrew-cask,kassi/homebrew-cask,jeroenseegers/homebrew-cask,lukeadams/homebrew-cask,jawshooah/homebrew-cask,mathbunnyru/homebrew-cask,gabrielizaias/homebrew-cask,maxnordlund/homebrew-cask,Ephemera/homebrew-cask,sgnh/homebrew-cask,franklouwers/homebrew-cask,julionc/homebrew-cask,markhuber/homebrew-cask,kesara/homebrew-cask,Ngrd/homebrew-cask,zmwangx/homebrew-cask,kpearson/homebrew-cask,samshadwell/homebrew-cask,stephenwade/homebrew-cask,albertico/homebrew-cask,yuhki50/homebrew-cask,mazehall/homebrew-cask,blogabe/homebrew-cask,codeurge/homebrew-cask,uetchy/homebrew-cask,BenjaminHCCarr/homebrew-cask,anbotero/homebrew-cask,kamilboratynski/homebrew-cask,dictcp/homebrew-cask,FranklinChen/homebrew-cask,helloIAmPau/homebrew-cask,sanchezm/homebrew-cask,MircoT/homebrew-cask,mattrobenolt/homebrew-cask,caskroom/homebrew-cask,nshemonsky/homebrew-cask,deanmorin/homebrew-cask,a1russell/homebrew-cask,mattrobenolt/homebrew-cask,kesara/homebrew-cask,faun/homebrew-cask,tjt263/homebrew-cask,shoichiaizawa/homebrew-cask,mjgardner/homebrew-cask,mchlrmrz/homebrew-cask,jeanregisser/homebrew-cask,Dremora/homebrew-cask,johnjelinek/homebrew-cask,singingwolfboy/homebrew-cask,xyb/homebrew-cask,jpmat296/homebrew-cask,Saklad5/homebrew-cask,puffdad/homebrew-cask,troyxmccall/homebrew-cask,bric3/homebrew-cask,johndbritton/homebrew-cask,scottsuch/homebrew-cask,gerrypower/homebrew-cask,retbrown/homebrew-cask,lumaxis/homebrew-cask,pkq/homebrew-cask,janlugt/homebrew-cask,michelegera/homebrew-cask,ericbn/homebrew-cask,Keloran/homebrew-cask,joshka/homebrew-cask,nrlquaker/homebrew-cask,xyb/homebrew-cask,usami-k/homebrew-cask,shoichiaizawa/homebrew-cask,MoOx/homebrew-cask,m3nu/homebrew-cask,antogg/homebrew-cask,zerrot/homebrew-cask,alebcay/homebrew-cask,scribblemaniac/homebrew-cask,farmerchris/homebrew-cask,bosr/homebrew-cask,lumaxis/homebrew-cask,timsutton/homebrew-cask,xight/homebrew-cask,m3nu/homebrew-cask,doits/homebrew-cask,julionc/homebrew-cask,athrunsun/homebrew-cask,shonjir/homebrew-cask,afh/homebrew-cask,arronmabrey/homebrew-cask,goxberry/homebrew-cask,skatsuta/homebrew-cask,dvdoliveira/homebrew-cask,arronmabrey/homebrew-cask,stevehedrick/homebrew-cask,JosephViolago/homebrew-cask,kronicd/homebrew-cask,retrography/homebrew-cask,dcondrey/homebrew-cask,mauricerkelly/homebrew-cask,JacopKane/homebrew-cask,ksylvan/homebrew-cask,mrmachine/homebrew-cask,boecko/homebrew-cask,asins/homebrew-cask,optikfluffel/homebrew-cask,hristozov/homebrew-cask,scribblemaniac/homebrew-cask,joschi/homebrew-cask,retrography/homebrew-cask,BenjaminHCCarr/homebrew-cask,mishari/homebrew-cask,lantrix/homebrew-cask,joschi/homebrew-cask,haha1903/homebrew-cask,koenrh/homebrew-cask,neverfox/homebrew-cask,jhowtan/homebrew-cask,adrianchia/homebrew-cask,codeurge/homebrew-cask,paour/homebrew-cask,lukasbestle/homebrew-cask,jgarber623/homebrew-cask,markthetech/homebrew-cask,jiashuw/homebrew-cask,bric3/homebrew-cask,corbt/homebrew-cask,tedski/homebrew-cask,boecko/homebrew-cask,blogabe/homebrew-cask,tan9/homebrew-cask,wickles/homebrew-cask,zmwangx/homebrew-cask,Amorymeltzer/homebrew-cask,Ketouem/homebrew-cask,cliffcotino/homebrew-cask,pacav69/homebrew-cask,MerelyAPseudonym/homebrew-cask,mlocher/homebrew-cask,wickedsp1d3r/homebrew-cask,amatos/homebrew-cask,sanyer/homebrew-cask,jaredsampson/homebrew-cask,elyscape/homebrew-cask,artdevjs/homebrew-cask,asbachb/homebrew-cask,kingthorin/homebrew-cask,chrisfinazzo/homebrew-cask,sgnh/homebrew-cask,y00rb/homebrew-cask,6uclz1/homebrew-cask,cedwardsmedia/homebrew-cask,wmorin/homebrew-cask,yumitsu/homebrew-cask,Ketouem/homebrew-cask,casidiablo/homebrew-cask,cobyism/homebrew-cask,jgarber623/homebrew-cask,xcezx/homebrew-cask,0rax/homebrew-cask,robbiethegeek/homebrew-cask,johnjelinek/homebrew-cask,neverfox/homebrew-cask,tangestani/homebrew-cask,flaviocamilo/homebrew-cask,antogg/homebrew-cask,mhubig/homebrew-cask,chuanxd/homebrew-cask,Ephemera/homebrew-cask,shonjir/homebrew-cask,jangalinski/homebrew-cask,adrianchia/homebrew-cask,SentinelWarren/homebrew-cask,malford/homebrew-cask,onlynone/homebrew-cask,cobyism/homebrew-cask,hellosky806/homebrew-cask,joshka/homebrew-cask,kesara/homebrew-cask,michelegera/homebrew-cask,stephenwade/homebrew-cask,renard/homebrew-cask,fharbe/homebrew-cask,inz/homebrew-cask,feigaochn/homebrew-cask,slack4u/homebrew-cask,0rax/homebrew-cask,nightscape/homebrew-cask,imgarylai/homebrew-cask,gyndav/homebrew-cask,mwean/homebrew-cask,danielbayley/homebrew-cask,tjnycum/homebrew-cask,schneidmaster/homebrew-cask,KosherBacon/homebrew-cask,rajiv/homebrew-cask,sanchezm/homebrew-cask,tedbundyjr/homebrew-cask,giannitm/homebrew-cask,wmorin/homebrew-cask,nathansgreen/homebrew-cask,xtian/homebrew-cask,Dremora/homebrew-cask,seanorama/homebrew-cask,FredLackeyOfficial/homebrew-cask,mjgardner/homebrew-cask,malob/homebrew-cask,vin047/homebrew-cask,jasmas/homebrew-cask,moimikey/homebrew-cask,miku/homebrew-cask,singingwolfboy/homebrew-cask,cfillion/homebrew-cask,mathbunnyru/homebrew-cask,ianyh/homebrew-cask,cliffcotino/homebrew-cask,diogodamiani/homebrew-cask,greg5green/homebrew-cask,mjgardner/homebrew-cask,ptb/homebrew-cask,coeligena/homebrew-customized,tsparber/homebrew-cask,ldong/homebrew-cask,greg5green/homebrew-cask,jgarber623/homebrew-cask,jalaziz/homebrew-cask,Labutin/homebrew-cask,wickles/homebrew-cask,antogg/homebrew-cask,My2ndAngelic/homebrew-cask,yurikoles/homebrew-cask,claui/homebrew-cask,samnung/homebrew-cask,jedahan/homebrew-cask,caskroom/homebrew-cask,jbeagley52/homebrew-cask,nightscape/homebrew-cask,optikfluffel/homebrew-cask,MichaelPei/homebrew-cask,esebastian/homebrew-cask,seanzxx/homebrew-cask,guerrero/homebrew-cask,robbiethegeek/homebrew-cask,JikkuJose/homebrew-cask,opsdev-ws/homebrew-cask,asins/homebrew-cask,mattrobenolt/homebrew-cask,jacobbednarz/homebrew-cask,mjdescy/homebrew-cask,decrement/homebrew-cask,doits/homebrew-cask,esebastian/homebrew-cask,tmoreira2020/homebrew,jppelteret/homebrew-cask,thehunmonkgroup/homebrew-cask,yutarody/homebrew-cask,timsutton/homebrew-cask,Keloran/homebrew-cask,jonathanwiesel/homebrew-cask,josa42/homebrew-cask,wastrachan/homebrew-cask,kteru/homebrew-cask,Fedalto/homebrew-cask,yutarody/homebrew-cask,mgryszko/homebrew-cask,y00rb/homebrew-cask,tjnycum/homebrew-cask,phpwutz/homebrew-cask,corbt/homebrew-cask,perfide/homebrew-cask,stonehippo/homebrew-cask,tyage/homebrew-cask,shonjir/homebrew-cask,chrisfinazzo/homebrew-cask,winkelsdorf/homebrew-cask,deanmorin/homebrew-cask,AnastasiaSulyagina/homebrew-cask,n0ts/homebrew-cask,rajiv/homebrew-cask,deiga/homebrew-cask,kkdd/homebrew-cask,reitermarkus/homebrew-cask,mahori/homebrew-cask,exherb/homebrew-cask,pkq/homebrew-cask,shoichiaizawa/homebrew-cask,lucasmezencio/homebrew-cask,daften/homebrew-cask,lvicentesanchez/homebrew-cask,ebraminio/homebrew-cask,squid314/homebrew-cask,ddm/homebrew-cask,uetchy/homebrew-cask,andyli/homebrew-cask,jppelteret/homebrew-cask,vin047/homebrew-cask,kamilboratynski/homebrew-cask,klane/homebrew-cask,moogar0880/homebrew-cask,jconley/homebrew-cask,lcasey001/homebrew-cask,hellosky806/homebrew-cask,afh/homebrew-cask,jmeridth/homebrew-cask,theoriginalgri/homebrew-cask,patresi/homebrew-cask,thehunmonkgroup/homebrew-cask,haha1903/homebrew-cask,gilesdring/homebrew-cask,pkq/homebrew-cask,gyndav/homebrew-cask,robertgzr/homebrew-cask,gilesdring/homebrew-cask,13k/homebrew-cask,elnappo/homebrew-cask,renaudguerin/homebrew-cask,perfide/homebrew-cask,n8henrie/homebrew-cask,vigosan/homebrew-cask,dictcp/homebrew-cask,santoshsahoo/homebrew-cask,otaran/homebrew-cask,Bombenleger/homebrew-cask,muan/homebrew-cask,skatsuta/homebrew-cask,theoriginalgri/homebrew-cask,tedbundyjr/homebrew-cask,Fedalto/homebrew-cask,lifepillar/homebrew-cask,winkelsdorf/homebrew-cask,larseggert/homebrew-cask,yurikoles/homebrew-cask,napaxton/homebrew-cask,sosedoff/homebrew-cask,mishari/homebrew-cask,chuanxd/homebrew-cask,wickedsp1d3r/homebrew-cask,diguage/homebrew-cask,sebcode/homebrew-cask,axodys/homebrew-cask,cprecioso/homebrew-cask,sanyer/homebrew-cask,kpearson/homebrew-cask,jconley/homebrew-cask,sscotth/homebrew-cask,kingthorin/homebrew-cask,hanxue/caskroom,nathansgreen/homebrew-cask,dwihn0r/homebrew-cask,coeligena/homebrew-customized,gmkey/homebrew-cask,malford/homebrew-cask,bdhess/homebrew-cask,deiga/homebrew-cask,guerrero/homebrew-cask,maxnordlund/homebrew-cask,ericbn/homebrew-cask,sjackman/homebrew-cask,vigosan/homebrew-cask,stonehippo/homebrew-cask,wastrachan/homebrew-cask,troyxmccall/homebrew-cask,klane/homebrew-cask,kTitan/homebrew-cask,dwkns/homebrew-cask,alebcay/homebrew-cask,alexg0/homebrew-cask,seanzxx/homebrew-cask,FranklinChen/homebrew-cask,13k/homebrew-cask,napaxton/homebrew-cask,blainesch/homebrew-cask,sohtsuka/homebrew-cask,yurikoles/homebrew-cask,sohtsuka/homebrew-cask,Labutin/homebrew-cask,albertico/homebrew-cask,kteru/homebrew-cask,gyndav/homebrew-cask,ywfwj2008/homebrew-cask,m3nu/homebrew-cask,brianshumate/homebrew-cask,hakamadare/homebrew-cask,wKovacs64/homebrew-cask,jellyfishcoder/homebrew-cask,Ibuprofen/homebrew-cask,Amorymeltzer/homebrew-cask,ksato9700/homebrew-cask,crzrcn/homebrew-cask,jaredsampson/homebrew-cask,bosr/homebrew-cask,gibsjose/homebrew-cask,fharbe/homebrew-cask,tolbkni/homebrew-cask,thii/homebrew-cask,Cottser/homebrew-cask,opsdev-ws/homebrew-cask,nathancahill/homebrew-cask,mchlrmrz/homebrew-cask,diguage/homebrew-cask,tan9/homebrew-cask,nathanielvarona/homebrew-cask,rogeriopradoj/homebrew-cask,ayohrling/homebrew-cask,rickychilcott/homebrew-cask,robertgzr/homebrew-cask,gurghet/homebrew-cask,vitorgalvao/homebrew-cask,JikkuJose/homebrew-cask,malob/homebrew-cask,colindunn/homebrew-cask,asbachb/homebrew-cask,kronicd/homebrew-cask,tolbkni/homebrew-cask,CameronGarrett/homebrew-cask,timsutton/homebrew-cask,franklouwers/homebrew-cask,victorpopkov/homebrew-cask,psibre/homebrew-cask,mgryszko/homebrew-cask,jasmas/homebrew-cask,aguynamedryan/homebrew-cask,MichaelPei/homebrew-cask,linc01n/homebrew-cask,thomanq/homebrew-cask,MircoT/homebrew-cask,artdevjs/homebrew-cask,devmynd/homebrew-cask,jangalinski/homebrew-cask,ddm/homebrew-cask,cedwardsmedia/homebrew-cask,lukeadams/homebrew-cask,BenjaminHCCarr/homebrew-cask,morganestes/homebrew-cask,ericbn/homebrew-cask,usami-k/homebrew-cask,josa42/homebrew-cask,athrunsun/homebrew-cask,muan/homebrew-cask,dvdoliveira/homebrew-cask,dustinblackman/homebrew-cask,mwean/homebrew-cask,ldong/homebrew-cask,stigkj/homebrew-caskroom-cask,jpmat296/homebrew-cask,thomanq/homebrew-cask,mazehall/homebrew-cask,inta/homebrew-cask,kiliankoe/homebrew-cask,lifepillar/homebrew-cask,a1russell/homebrew-cask,mahori/homebrew-cask,yuhki50/homebrew-cask,reitermarkus/homebrew-cask,kiliankoe/homebrew-cask,stonehippo/homebrew-cask,thii/homebrew-cask,imgarylai/homebrew-cask,renard/homebrew-cask,santoshsahoo/homebrew-cask,joschi/homebrew-cask,reelsense/homebrew-cask,wKovacs64/homebrew-cask,jacobbednarz/homebrew-cask,FinalDes/homebrew-cask,retbrown/homebrew-cask,SentinelWarren/homebrew-cask,jeroenj/homebrew-cask,koenrh/homebrew-cask,Bombenleger/homebrew-cask,tsparber/homebrew-cask,blainesch/homebrew-cask,Ephemera/homebrew-cask,optikfluffel/homebrew-cask,paour/homebrew-cask,aguynamedryan/homebrew-cask,andyli/homebrew-cask,jiashuw/homebrew-cask,jeanregisser/homebrew-cask,a1russell/homebrew-cask,Saklad5/homebrew-cask,paour/homebrew-cask,alexg0/homebrew-cask,ayohrling/homebrew-cask,onlynone/homebrew-cask,claui/homebrew-cask,andrewdisley/homebrew-cask,moogar0880/homebrew-cask,bcomnes/homebrew-cask,gibsjose/homebrew-cask,otaran/homebrew-cask,reelsense/homebrew-cask,singingwolfboy/homebrew-cask,chadcatlett/caskroom-homebrew-cask,syscrusher/homebrew-cask,RJHsiao/homebrew-cask,xakraz/homebrew-cask,nrlquaker/homebrew-cask,tarwich/homebrew-cask,hakamadare/homebrew-cask,Gasol/homebrew-cask | ruby | ## Code Before:
cask :v1 => 'nzbget' do
version '15.0'
sha256 'd4e6fa16d8404a2d1c1e6fe28e4bf4ca4f10d80cf53c724879867d82335167af'
url "https://downloads.sourceforge.net/project/nzbget/nzbget-stable/#{version}/nzbget-#{version}-bin-osx.zip"
name 'NZBGet'
homepage 'https://nzbget.net'
license :gpl
app 'NZBGet.app'
zap :delete => [
'~/Library/Application Support/NZBGet',
'~/Library/Preferences/net.sourceforge.nzbget.plist'
]
end
## Instruction:
Update NZBGet to 16.1 and fix download url.
## Code After:
cask :v1 => 'nzbget' do
version '16.1'
sha256 '4ec71fa2dcd94c69520eda5c7f43d36e1ab49149bd8b79339da9d5e08f740c8e'
# github.com is the official download host per the vendor homepage
url "https://github.com/nzbget/nzbget/releases/download/v#{version}/nzbget-#{version}-bin-osx.zip"
appcast 'https://github.com/nzbget/nzbget/releases.atom'
name 'NZBGet'
homepage 'http://nzbget.net'
license :gpl
app 'NZBGet.app'
zap :delete => [
'~/Library/Application Support/NZBGet',
'~/Library/Preferences/net.sourceforge.nzbget.plist'
]
end
| cask :v1 => 'nzbget' do
- version '15.0'
? ^ ^
+ version '16.1'
? ^ ^
- sha256 'd4e6fa16d8404a2d1c1e6fe28e4bf4ca4f10d80cf53c724879867d82335167af'
+ sha256 '4ec71fa2dcd94c69520eda5c7f43d36e1ab49149bd8b79339da9d5e08f740c8e'
- url "https://downloads.sourceforge.net/project/nzbget/nzbget-stable/#{version}/nzbget-#{version}-bin-osx.zip"
+ # github.com is the official download host per the vendor homepage
+ url "https://github.com/nzbget/nzbget/releases/download/v#{version}/nzbget-#{version}-bin-osx.zip"
+ appcast 'https://github.com/nzbget/nzbget/releases.atom'
name 'NZBGet'
- homepage 'https://nzbget.net'
? -
+ homepage 'http://nzbget.net'
license :gpl
app 'NZBGet.app'
zap :delete => [
'~/Library/Application Support/NZBGet',
'~/Library/Preferences/net.sourceforge.nzbget.plist'
]
end | 10 | 0.625 | 6 | 4 |
834a6a65f144e17f22851230d2baf3524f5e98c0 | flexget/plugins/est_released.py | flexget/plugins/est_released.py | import logging
from flexget.plugin import get_plugins_by_group, register_plugin
log = logging.getLogger('est_released')
class EstimateRelease(object):
"""
Front-end for estimator plugins that estimate release times
for various things (series, movies).
"""
def estimate(self, entry):
"""
Estimate release schedule for Entry
:param entry:
:return: estimated date of released for the entry, None if it can't figure it out
"""
log.info(entry['title'])
estimators = get_plugins_by_group('estimate_release')
for estimator in estimators:
return estimator.instance.estimate(entry)
register_plugin(EstimateRelease, 'estimate_release', api_ver=2)
| import logging
from flexget.plugin import get_plugins_by_group, register_plugin
log = logging.getLogger('est_released')
class EstimateRelease(object):
"""
Front-end for estimator plugins that estimate release times
for various things (series, movies).
"""
def estimate(self, entry):
"""
Estimate release schedule for Entry
:param entry:
:return: estimated date of released for the entry, None if it can't figure it out
"""
log.info(entry['title'])
estimators = get_plugins_by_group('estimate_release')
for estimator in estimators:
estimate = estimator.instance.estimate(entry)
# return first successful estimation
if estimate is not None:
return estimate
register_plugin(EstimateRelease, 'estimate_release', api_ver=2)
| Fix estimator loop, consider rest plugins as well. | Fix estimator loop, consider rest plugins as well.
| Python | mit | ianstalk/Flexget,qk4l/Flexget,malkavi/Flexget,crawln45/Flexget,tobinjt/Flexget,xfouloux/Flexget,asm0dey/Flexget,Flexget/Flexget,ibrahimkarahan/Flexget,vfrc2/Flexget,jacobmetrick/Flexget,crawln45/Flexget,tarzasai/Flexget,Pretagonist/Flexget,xfouloux/Flexget,spencerjanssen/Flexget,lildadou/Flexget,JorisDeRieck/Flexget,jawilson/Flexget,tsnoam/Flexget,Danfocus/Flexget,ratoaq2/Flexget,tsnoam/Flexget,ibrahimkarahan/Flexget,offbyone/Flexget,OmgOhnoes/Flexget,camon/Flexget,X-dark/Flexget,Flexget/Flexget,asm0dey/Flexget,v17al/Flexget,ianstalk/Flexget,voriux/Flexget,ratoaq2/Flexget,X-dark/Flexget,tsnoam/Flexget,jawilson/Flexget,Flexget/Flexget,grrr2/Flexget,patsissons/Flexget,poulpito/Flexget,jacobmetrick/Flexget,camon/Flexget,malkavi/Flexget,lildadou/Flexget,poulpito/Flexget,qk4l/Flexget,tarzasai/Flexget,thalamus/Flexget,vfrc2/Flexget,Danfocus/Flexget,cvium/Flexget,tvcsantos/Flexget,ZefQ/Flexget,jacobmetrick/Flexget,dsemi/Flexget,tobinjt/Flexget,ibrahimkarahan/Flexget,qk4l/Flexget,OmgOhnoes/Flexget,drwyrm/Flexget,gazpachoking/Flexget,dsemi/Flexget,malkavi/Flexget,oxc/Flexget,malkavi/Flexget,grrr2/Flexget,vfrc2/Flexget,drwyrm/Flexget,drwyrm/Flexget,asm0dey/Flexget,poulpito/Flexget,xfouloux/Flexget,LynxyssCZ/Flexget,spencerjanssen/Flexget,JorisDeRieck/Flexget,jawilson/Flexget,grrr2/Flexget,tobinjt/Flexget,antivirtel/Flexget,Danfocus/Flexget,patsissons/Flexget,ianstalk/Flexget,lildadou/Flexget,Flexget/Flexget,ZefQ/Flexget,LynxyssCZ/Flexget,dsemi/Flexget,tobinjt/Flexget,oxc/Flexget,oxc/Flexget,voriux/Flexget,thalamus/Flexget,qvazzler/Flexget,JorisDeRieck/Flexget,antivirtel/Flexget,OmgOhnoes/Flexget,cvium/Flexget,v17al/Flexget,ZefQ/Flexget,sean797/Flexget,offbyone/Flexget,sean797/Flexget,crawln45/Flexget,Pretagonist/Flexget,qvazzler/Flexget,offbyone/Flexget,patsissons/Flexget,tvcsantos/Flexget,Pretagonist/Flexget,ratoaq2/Flexget,LynxyssCZ/Flexget,v17al/Flexget,qvazzler/Flexget,gazpachoking/Flexget,crawln45/Flexget,tarzasai/Flexget,cvium/Flexget,LynxyssCZ/Flexget,thalamus/Flexget,X-dark/Flexget,JorisDeRieck/Flexget,antivirtel/Flexget,jawilson/Flexget,Danfocus/Flexget,sean797/Flexget,spencerjanssen/Flexget | python | ## Code Before:
import logging
from flexget.plugin import get_plugins_by_group, register_plugin
log = logging.getLogger('est_released')
class EstimateRelease(object):
"""
Front-end for estimator plugins that estimate release times
for various things (series, movies).
"""
def estimate(self, entry):
"""
Estimate release schedule for Entry
:param entry:
:return: estimated date of released for the entry, None if it can't figure it out
"""
log.info(entry['title'])
estimators = get_plugins_by_group('estimate_release')
for estimator in estimators:
return estimator.instance.estimate(entry)
register_plugin(EstimateRelease, 'estimate_release', api_ver=2)
## Instruction:
Fix estimator loop, consider rest plugins as well.
## Code After:
import logging
from flexget.plugin import get_plugins_by_group, register_plugin
log = logging.getLogger('est_released')
class EstimateRelease(object):
"""
Front-end for estimator plugins that estimate release times
for various things (series, movies).
"""
def estimate(self, entry):
"""
Estimate release schedule for Entry
:param entry:
:return: estimated date of released for the entry, None if it can't figure it out
"""
log.info(entry['title'])
estimators = get_plugins_by_group('estimate_release')
for estimator in estimators:
estimate = estimator.instance.estimate(entry)
# return first successful estimation
if estimate is not None:
return estimate
register_plugin(EstimateRelease, 'estimate_release', api_ver=2)
| import logging
from flexget.plugin import get_plugins_by_group, register_plugin
log = logging.getLogger('est_released')
class EstimateRelease(object):
"""
Front-end for estimator plugins that estimate release times
for various things (series, movies).
"""
def estimate(self, entry):
"""
Estimate release schedule for Entry
:param entry:
:return: estimated date of released for the entry, None if it can't figure it out
"""
log.info(entry['title'])
estimators = get_plugins_by_group('estimate_release')
for estimator in estimators:
- return estimator.instance.estimate(entry)
? - ^^^
+ estimate = estimator.instance.estimate(entry)
? + ^^^^^^^
+ # return first successful estimation
+ if estimate is not None:
+ return estimate
register_plugin(EstimateRelease, 'estimate_release', api_ver=2) | 5 | 0.185185 | 4 | 1 |
4d359aea5dd4f780dbc374794958066de619fb30 | doc/periodictask/eventcounter.rst | doc/periodictask/eventcounter.rst | .. _eventcounter:
EventCounter
------------
TODO | .. _eventcounter:
EventCounter
------------
The Event Counter task module can be used with the :ref:`periodic_tasks` to create time series of certain events.
An event could be a failed authentication request. Using the Event Counter privacyIDEA can create graphs that display
the development of failed authentication requests over time.
To do this, the Event Counter task module reads a counter value from the database table ``EventCounter`` and writes this
current value in the time series in the database table ``MonitoringStats``.
As the administrator can use the event handler :ref:`counterhandler` to record any arbitrary event under any condition,
this task module can be used to graph any metrics in privacyIDEA, be it failed authentication requests per time unit,
the number of token delete requests or the number of PIN resets per month.
Options
~~~~~~~
The Event Counter task module provides the following options:
**event_counter**
This is the name of the event counter key, that was defined in a :ref:`counterhandler` definition and that is
read from the database table ``EventCounter``.
**stats_key**
This is the name of the statistics key that is written to the ``MonitoringStats`` database table.
The event counter key store the current number of counted events, the ``stats_key`` takes the current number
and stores it with the timestamp as a time series.
**reset_event_counter**
This is a boolean value. If it is set to true (the checkbox is checked), then the event counter will be reset to zero,
after the task module has read the key.
Resetting the the event counter results in a time series of "events per time interval". The time intervall is
specified by the time intervall in which the Event Counter task module is called.
If ``reset_event_counter`` is not checked, then the event handler will continue to increase the counter value.
Use this, if you want to create a time series, that display the absolute number of events.
| Add documentation for Event Counter task module | Add documentation for Event Counter task module
Working on #1183
| reStructuredText | agpl-3.0 | privacyidea/privacyidea,privacyidea/privacyidea,privacyidea/privacyidea,privacyidea/privacyidea,privacyidea/privacyidea,privacyidea/privacyidea | restructuredtext | ## Code Before:
.. _eventcounter:
EventCounter
------------
TODO
## Instruction:
Add documentation for Event Counter task module
Working on #1183
## Code After:
.. _eventcounter:
EventCounter
------------
The Event Counter task module can be used with the :ref:`periodic_tasks` to create time series of certain events.
An event could be a failed authentication request. Using the Event Counter privacyIDEA can create graphs that display
the development of failed authentication requests over time.
To do this, the Event Counter task module reads a counter value from the database table ``EventCounter`` and writes this
current value in the time series in the database table ``MonitoringStats``.
As the administrator can use the event handler :ref:`counterhandler` to record any arbitrary event under any condition,
this task module can be used to graph any metrics in privacyIDEA, be it failed authentication requests per time unit,
the number of token delete requests or the number of PIN resets per month.
Options
~~~~~~~
The Event Counter task module provides the following options:
**event_counter**
This is the name of the event counter key, that was defined in a :ref:`counterhandler` definition and that is
read from the database table ``EventCounter``.
**stats_key**
This is the name of the statistics key that is written to the ``MonitoringStats`` database table.
The event counter key store the current number of counted events, the ``stats_key`` takes the current number
and stores it with the timestamp as a time series.
**reset_event_counter**
This is a boolean value. If it is set to true (the checkbox is checked), then the event counter will be reset to zero,
after the task module has read the key.
Resetting the the event counter results in a time series of "events per time interval". The time intervall is
specified by the time intervall in which the Event Counter task module is called.
If ``reset_event_counter`` is not checked, then the event handler will continue to increase the counter value.
Use this, if you want to create a time series, that display the absolute number of events.
| .. _eventcounter:
EventCounter
------------
- TODO
+ The Event Counter task module can be used with the :ref:`periodic_tasks` to create time series of certain events.
+ An event could be a failed authentication request. Using the Event Counter privacyIDEA can create graphs that display
+ the development of failed authentication requests over time.
+
+ To do this, the Event Counter task module reads a counter value from the database table ``EventCounter`` and writes this
+ current value in the time series in the database table ``MonitoringStats``.
+ As the administrator can use the event handler :ref:`counterhandler` to record any arbitrary event under any condition,
+ this task module can be used to graph any metrics in privacyIDEA, be it failed authentication requests per time unit,
+ the number of token delete requests or the number of PIN resets per month.
+
+ Options
+ ~~~~~~~
+
+ The Event Counter task module provides the following options:
+
+ **event_counter**
+
+ This is the name of the event counter key, that was defined in a :ref:`counterhandler` definition and that is
+ read from the database table ``EventCounter``.
+
+ **stats_key**
+
+ This is the name of the statistics key that is written to the ``MonitoringStats`` database table.
+ The event counter key store the current number of counted events, the ``stats_key`` takes the current number
+ and stores it with the timestamp as a time series.
+
+ **reset_event_counter**
+
+ This is a boolean value. If it is set to true (the checkbox is checked), then the event counter will be reset to zero,
+ after the task module has read the key.
+
+ Resetting the the event counter results in a time series of "events per time interval". The time intervall is
+ specified by the time intervall in which the Event Counter task module is called.
+ If ``reset_event_counter`` is not checked, then the event handler will continue to increase the counter value.
+ Use this, if you want to create a time series, that display the absolute number of events.
+ | 37 | 6.166667 | 36 | 1 |
06dfe2dbf49e126a55632c2c39a1625be140c9ee | installed/ubuntu-12.10/Dockerfile | installed/ubuntu-12.10/Dockerfile | from saltstack/ubuntu-12.10-minimal
MAINTAINER SaltStack, Inc.
# Install Dependencies
RUN apt-get install -y -o DPkg::Options::=--force-confold python-setuptools
# Install Salt
RUN easy_install https://github.com/saltstack/salt/archive/develop.tar.gz
| from saltstack/ubuntu-12.10-minimal
MAINTAINER SaltStack, Inc.
# Upgrade System and Install dependencies
RUN apt-get update && \
apt-get upgrade -y -o DPkg::Options::=--force-confold && \
apt-get install -y -o DPkg::Options::=--force-confold curl
# Install Latest Salt from the Develop Branch
RUN curl -L http://bootstrap.saltstack.org | sh -s -- git develop
| Upgrade system, install deps and bootstrap salt | Upgrade system, install deps and bootstrap salt
| unknown | apache-2.0 | saltstack/docker-containers | unknown | ## Code Before:
from saltstack/ubuntu-12.10-minimal
MAINTAINER SaltStack, Inc.
# Install Dependencies
RUN apt-get install -y -o DPkg::Options::=--force-confold python-setuptools
# Install Salt
RUN easy_install https://github.com/saltstack/salt/archive/develop.tar.gz
## Instruction:
Upgrade system, install deps and bootstrap salt
## Code After:
from saltstack/ubuntu-12.10-minimal
MAINTAINER SaltStack, Inc.
# Upgrade System and Install dependencies
RUN apt-get update && \
apt-get upgrade -y -o DPkg::Options::=--force-confold && \
apt-get install -y -o DPkg::Options::=--force-confold curl
# Install Latest Salt from the Develop Branch
RUN curl -L http://bootstrap.saltstack.org | sh -s -- git develop
| from saltstack/ubuntu-12.10-minimal
MAINTAINER SaltStack, Inc.
- # Install Dependencies
+ # Upgrade System and Install dependencies
+ RUN apt-get update && \
+ apt-get upgrade -y -o DPkg::Options::=--force-confold && \
- RUN apt-get install -y -o DPkg::Options::=--force-confold python-setuptools
? ^^^ ^^^^^^^^^^ ^^^^ -
+ apt-get install -y -o DPkg::Options::=--force-confold curl
? ^ ^ ^
- # Install Salt
- RUN easy_install https://github.com/saltstack/salt/archive/develop.tar.gz
+ # Install Latest Salt from the Develop Branch
+ RUN curl -L http://bootstrap.saltstack.org | sh -s -- git develop | 10 | 1.25 | 6 | 4 |
0b392ecff52ad56cec684203bf13cd9bdeff23cb | src/Project/AppBundle/Resources/views/Promotion/index.html.twig | src/Project/AppBundle/Resources/views/Promotion/index.html.twig | {% extends "ProjectAppBundle::layout.html.twig" %}
{% block content %}
<h2>Promotions</h2>
<div class="bloc-top">
<a class="btn btn-second" href="{{ path('promotion_new') }}">Ajouter une promotion</a>
</div>
<div class="panel-group">
{% if promotions is empty %}
<div class="panel-item">
<div class="panel-content">
Aucune promotion n'est actuellement disponible.
</div>
</div>
{% else %}
{% for promotion in promotions %}
<div class="panel-item">
<div class="panel-content">
<div class="panel-header">
<h3 class="pull-left">{{ promotion.startDate|date('Y') }} - {{ promotion.endDate|date('Y') }}</h3>
</div>
<div class="panel-content">
Formation : {{ promotion.formation.name }}
<div class="pull-right">
<a class="btn btn-second" href="{{ path('promotion_show', { 'id': promotion.id }) }}">Gérer</a>
<a class="btn btn-second" href="{{ path('promotion_show', { 'id': promotion.id }) }}">Archiver</a>
</div>
</div>
</div>
</div>
{% endfor %}
{% endif %}
</div>
<div class="bloc-bottom">
<a class="btn btn-second" href="{{ path('promotion_new') }}">Ajouter une promotion</a>
</div>
{% endblock %}
| {% extends "ProjectAppBundle::layout.html.twig" %}
{% block content %}
<div class="bloc-top">
<a class="btn btn-second" href="{{ path('promotion_new') }}">Ajouter une promotion</a>
</div>
<div class="panel-group">
{% if promotions is empty %}
<div class="panel-item">
<div class="panel-content">
Aucune promotion n'est actuellement disponible.
</div>
</div>
{% else %}
{% for promotion in promotions %}
<div class="panel-item">
<div class="panel-content">
<div class="panel-header">
<h3 class="pull-left">{{ promotion.startDate|date('Y') }} - {{ promotion.endDate|date('Y') }}</h3>
</div>
<div class="panel-content">
Formation : {{ promotion.formation.name }}
<div class="pull-right">
<a class="btn btn-second" href="{{ path('promotion_show', { 'id': promotion.id }) }}">Gérer</a>
<a class="btn btn-second" href="{{ path('promotion_show', { 'id': promotion.id }) }}">Archiver</a>
</div>
</div>
</div>
</div>
{% endfor %}
{% endif %}
</div>
{% endblock %}
| Fix promotion template to fit with wireframes | Fix promotion template to fit with wireframes
| Twig | mit | vrap/project-b,vrap/project-b,vrap/project-b | twig | ## Code Before:
{% extends "ProjectAppBundle::layout.html.twig" %}
{% block content %}
<h2>Promotions</h2>
<div class="bloc-top">
<a class="btn btn-second" href="{{ path('promotion_new') }}">Ajouter une promotion</a>
</div>
<div class="panel-group">
{% if promotions is empty %}
<div class="panel-item">
<div class="panel-content">
Aucune promotion n'est actuellement disponible.
</div>
</div>
{% else %}
{% for promotion in promotions %}
<div class="panel-item">
<div class="panel-content">
<div class="panel-header">
<h3 class="pull-left">{{ promotion.startDate|date('Y') }} - {{ promotion.endDate|date('Y') }}</h3>
</div>
<div class="panel-content">
Formation : {{ promotion.formation.name }}
<div class="pull-right">
<a class="btn btn-second" href="{{ path('promotion_show', { 'id': promotion.id }) }}">Gérer</a>
<a class="btn btn-second" href="{{ path('promotion_show', { 'id': promotion.id }) }}">Archiver</a>
</div>
</div>
</div>
</div>
{% endfor %}
{% endif %}
</div>
<div class="bloc-bottom">
<a class="btn btn-second" href="{{ path('promotion_new') }}">Ajouter une promotion</a>
</div>
{% endblock %}
## Instruction:
Fix promotion template to fit with wireframes
## Code After:
{% extends "ProjectAppBundle::layout.html.twig" %}
{% block content %}
<div class="bloc-top">
<a class="btn btn-second" href="{{ path('promotion_new') }}">Ajouter une promotion</a>
</div>
<div class="panel-group">
{% if promotions is empty %}
<div class="panel-item">
<div class="panel-content">
Aucune promotion n'est actuellement disponible.
</div>
</div>
{% else %}
{% for promotion in promotions %}
<div class="panel-item">
<div class="panel-content">
<div class="panel-header">
<h3 class="pull-left">{{ promotion.startDate|date('Y') }} - {{ promotion.endDate|date('Y') }}</h3>
</div>
<div class="panel-content">
Formation : {{ promotion.formation.name }}
<div class="pull-right">
<a class="btn btn-second" href="{{ path('promotion_show', { 'id': promotion.id }) }}">Gérer</a>
<a class="btn btn-second" href="{{ path('promotion_show', { 'id': promotion.id }) }}">Archiver</a>
</div>
</div>
</div>
</div>
{% endfor %}
{% endif %}
</div>
{% endblock %}
| {% extends "ProjectAppBundle::layout.html.twig" %}
{% block content %}
- <h2>Promotions</h2>
-
<div class="bloc-top">
<a class="btn btn-second" href="{{ path('promotion_new') }}">Ajouter une promotion</a>
</div>
<div class="panel-group">
{% if promotions is empty %}
<div class="panel-item">
<div class="panel-content">
Aucune promotion n'est actuellement disponible.
</div>
</div>
{% else %}
{% for promotion in promotions %}
<div class="panel-item">
<div class="panel-content">
<div class="panel-header">
<h3 class="pull-left">{{ promotion.startDate|date('Y') }} - {{ promotion.endDate|date('Y') }}</h3>
</div>
<div class="panel-content">
Formation : {{ promotion.formation.name }}
<div class="pull-right">
<a class="btn btn-second" href="{{ path('promotion_show', { 'id': promotion.id }) }}">Gérer</a>
<a class="btn btn-second" href="{{ path('promotion_show', { 'id': promotion.id }) }}">Archiver</a>
</div>
</div>
</div>
</div>
{% endfor %}
{% endif %}
</div>
-
- <div class="bloc-bottom">
- <a class="btn btn-second" href="{{ path('promotion_new') }}">Ajouter une promotion</a>
- </div>
{% endblock %} | 6 | 0.15 | 0 | 6 |
0d6d4e461a8930d63d708f64387ece7bb83bffe2 | app/templates/views/agreement.html | app/templates/views/agreement.html | {% extends "withoutnav_template.html" %}
{% from "components/sub-navigation.html" import sub_navigation %}
{% block per_page_title %}
GOV.UK Notify data sharing and financial agreement
{% endblock %}
{% block maincolumn_content %}
<div class="grid-row">
<div class="column-one-third">
{{ sub_navigation(navigation_links) }}
</div>
<div class="column-two-thirds">
<h1 class="heading-large">
GOV.UK Notify data sharing and financial agreement
</h1>
<p>
<a href="{{ url_for('main.download_agreement') }}">Download</a>.
</p>
</div>
</div>
{% endblock %}
| {% extends "withoutnav_template.html" %}
{% from "components/sub-navigation.html" import sub_navigation %}
{% block per_page_title %}
Download the GOV.UK Notify data sharing and financial agreement
{% endblock %}
{% block maincolumn_content %}
<div class="grid-row">
<div class="column-one-third">
{{ sub_navigation(navigation_links) }}
</div>
<div class="column-two-thirds">
<h1 class="heading-large">
Download the GOV.UK Notify data sharing and financial agreement
</h1>
<p>
This agreement needs to be signed by someone who has the authority to do so on behalf of your whole organisation. Typically this is a director of digital or head of finance.
</p>
<p>
Return a signed copy to <span style="white-space: nowrap">notify-support@digital.cabinet-office.gov.uk</span>
</p>
<p class="panel panel-border-wide bottom-gutter-2-3">
The agreement contains commercially sensitive information.<br>
Do not share it more widely than you need to.
</p>
<p>
<a href="{{ url_for('main.download_agreement') }}">Download the agreement</a>.
</p>
</div>
</div>
{% endblock %}
| Add wording to download page | Add wording to download page
| HTML | mit | alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin | html | ## Code Before:
{% extends "withoutnav_template.html" %}
{% from "components/sub-navigation.html" import sub_navigation %}
{% block per_page_title %}
GOV.UK Notify data sharing and financial agreement
{% endblock %}
{% block maincolumn_content %}
<div class="grid-row">
<div class="column-one-third">
{{ sub_navigation(navigation_links) }}
</div>
<div class="column-two-thirds">
<h1 class="heading-large">
GOV.UK Notify data sharing and financial agreement
</h1>
<p>
<a href="{{ url_for('main.download_agreement') }}">Download</a>.
</p>
</div>
</div>
{% endblock %}
## Instruction:
Add wording to download page
## Code After:
{% extends "withoutnav_template.html" %}
{% from "components/sub-navigation.html" import sub_navigation %}
{% block per_page_title %}
Download the GOV.UK Notify data sharing and financial agreement
{% endblock %}
{% block maincolumn_content %}
<div class="grid-row">
<div class="column-one-third">
{{ sub_navigation(navigation_links) }}
</div>
<div class="column-two-thirds">
<h1 class="heading-large">
Download the GOV.UK Notify data sharing and financial agreement
</h1>
<p>
This agreement needs to be signed by someone who has the authority to do so on behalf of your whole organisation. Typically this is a director of digital or head of finance.
</p>
<p>
Return a signed copy to <span style="white-space: nowrap">notify-support@digital.cabinet-office.gov.uk</span>
</p>
<p class="panel panel-border-wide bottom-gutter-2-3">
The agreement contains commercially sensitive information.<br>
Do not share it more widely than you need to.
</p>
<p>
<a href="{{ url_for('main.download_agreement') }}">Download the agreement</a>.
</p>
</div>
</div>
{% endblock %}
| {% extends "withoutnav_template.html" %}
{% from "components/sub-navigation.html" import sub_navigation %}
{% block per_page_title %}
- GOV.UK Notify data sharing and financial agreement
+ Download the GOV.UK Notify data sharing and financial agreement
? +++++++++++++
{% endblock %}
{% block maincolumn_content %}
<div class="grid-row">
<div class="column-one-third">
{{ sub_navigation(navigation_links) }}
</div>
<div class="column-two-thirds">
<h1 class="heading-large">
- GOV.UK Notify data sharing and financial agreement
+ Download the GOV.UK Notify data sharing and financial agreement
? +++++++++++++
</h1>
<p>
+ This agreement needs to be signed by someone who has the authority to do so on behalf of your whole organisation. Typically this is a director of digital or head of finance.
+ </p>
+ <p>
+ Return a signed copy to <span style="white-space: nowrap">notify-support@digital.cabinet-office.gov.uk</span>
+ </p>
+ <p class="panel panel-border-wide bottom-gutter-2-3">
+ The agreement contains commercially sensitive information.<br>
+ Do not share it more widely than you need to.
+ </p>
+ <p>
- <a href="{{ url_for('main.download_agreement') }}">Download</a>.
+ <a href="{{ url_for('main.download_agreement') }}">Download the agreement</a>.
? ++++++++++++++
</p>
</div>
</div>
{% endblock %} | 16 | 0.592593 | 13 | 3 |
2a0494dc273fd77d06d9c7b1ea7c590b86c62bd6 | lib/proiel/visualization/graphviz/modern.dot.erb | lib/proiel/visualization/graphviz/modern.dot.erb | digraph G {
charset="UTF-8";
graph [truecolor=true,bgcolor=transparent];
root [label="",shape=circle];
rankdir="<%= @direction -%>";
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= node token.id, token.relation.to_s.upcase, shape: :box -%>
<%- if token.relation -%>
<%= edge (token.head ? token.head.id : :root), token.id, '', weight: 1.0, color: :orange -%>
<%- end -%>
<%- token.slashes.each do |(relation, target)| -%>
<%= edge token.id, target, relation.to_s.upcase, weight: 0.0, color: :blue, style: :dashed %>
<%- end -%>
<%- end -%>
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= edge token.id, "T#{token.id}", nil, weight: 10, arrowhead: :none -%>
<%- end -%>
{
rank="same";
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= node "T#{token.id}", token.form, shape: :none -%>
<%- end -%>
}
}
| digraph G {
charset="UTF-8";
graph [truecolor=true,bgcolor=transparent];
root [label="",shape=circle];
rankdir="<%= @direction -%>";
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= node token.id, token.relation.to_s.upcase, shape: :box -%>
<%- if token.relation -%>
<%= edge (token.head ? token.head.id : :root), token.id, '', weight: 1.0, color: :orange -%>
<%- end -%>
<%- token.slashes.each do |(relation, target)| -%>
<%= edge token.id, target, relation.to_s.upcase, weight: 0.0, color: :blue, style: :dashed %>
<%- end -%>
<%- end -%>
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= edge token.id, "T#{token.id}", nil, weight: 10, arrowhead: :none -%>
<%- end -%>
{
rank="same";
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= node "T#{token.id}", token.form, shape: :none -%>
<%- end -%>
<%= @graph.tokens.select { |t| t.empty_token_sort != 'P' }.map { |token| "T#{token.id}" }.join('->') -%> [style="invis"];
}
}
| Enforce linear order in "modern" viz | Enforce linear order in "modern" viz
| HTML+ERB | mit | proiel/proiel,proiel/proiel | html+erb | ## Code Before:
digraph G {
charset="UTF-8";
graph [truecolor=true,bgcolor=transparent];
root [label="",shape=circle];
rankdir="<%= @direction -%>";
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= node token.id, token.relation.to_s.upcase, shape: :box -%>
<%- if token.relation -%>
<%= edge (token.head ? token.head.id : :root), token.id, '', weight: 1.0, color: :orange -%>
<%- end -%>
<%- token.slashes.each do |(relation, target)| -%>
<%= edge token.id, target, relation.to_s.upcase, weight: 0.0, color: :blue, style: :dashed %>
<%- end -%>
<%- end -%>
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= edge token.id, "T#{token.id}", nil, weight: 10, arrowhead: :none -%>
<%- end -%>
{
rank="same";
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= node "T#{token.id}", token.form, shape: :none -%>
<%- end -%>
}
}
## Instruction:
Enforce linear order in "modern" viz
## Code After:
digraph G {
charset="UTF-8";
graph [truecolor=true,bgcolor=transparent];
root [label="",shape=circle];
rankdir="<%= @direction -%>";
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= node token.id, token.relation.to_s.upcase, shape: :box -%>
<%- if token.relation -%>
<%= edge (token.head ? token.head.id : :root), token.id, '', weight: 1.0, color: :orange -%>
<%- end -%>
<%- token.slashes.each do |(relation, target)| -%>
<%= edge token.id, target, relation.to_s.upcase, weight: 0.0, color: :blue, style: :dashed %>
<%- end -%>
<%- end -%>
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= edge token.id, "T#{token.id}", nil, weight: 10, arrowhead: :none -%>
<%- end -%>
{
rank="same";
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= node "T#{token.id}", token.form, shape: :none -%>
<%- end -%>
<%= @graph.tokens.select { |t| t.empty_token_sort != 'P' }.map { |token| "T#{token.id}" }.join('->') -%> [style="invis"];
}
}
| digraph G {
charset="UTF-8";
graph [truecolor=true,bgcolor=transparent];
root [label="",shape=circle];
rankdir="<%= @direction -%>";
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= node token.id, token.relation.to_s.upcase, shape: :box -%>
<%- if token.relation -%>
<%= edge (token.head ? token.head.id : :root), token.id, '', weight: 1.0, color: :orange -%>
<%- end -%>
<%- token.slashes.each do |(relation, target)| -%>
<%= edge token.id, target, relation.to_s.upcase, weight: 0.0, color: :blue, style: :dashed %>
<%- end -%>
<%- end -%>
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= edge token.id, "T#{token.id}", nil, weight: 10, arrowhead: :none -%>
<%- end -%>
{
rank="same";
<%- @graph.tokens.select { |t| t.empty_token_sort != 'P' }.each do |token| -%>
<%= node "T#{token.id}", token.form, shape: :none -%>
<%- end -%>
+
+ <%= @graph.tokens.select { |t| t.empty_token_sort != 'P' }.map { |token| "T#{token.id}" }.join('->') -%> [style="invis"];
}
} | 2 | 0.066667 | 2 | 0 |
2a6f75e5e87e1ee780cabdd131988fbcc418dce6 | spec/client/spec.client_bootstrap.js | spec/client/spec.client_bootstrap.js | define([
'client/client_bootstrap',
'client/controllers/dashboard'
],
function (bootstrap, DashboardController) {
describe('client bootstrap', function () {
var config;
beforeEach(function () {
spyOn(DashboardController.prototype, 'render');
config = {
'page-type': 'dashboard'
};
});
it('instantiates a controller from config data', function () {
var controller = bootstrap(config);
expect(controller instanceof DashboardController).toBe(true);
expect(controller.render).toHaveBeenCalled();
});
it('executes page preprocessors', function () {
var originalPreprocessors = bootstrap.preprocessors;
bootstrap.preprocessors = [
jasmine.createSpy(),
jasmine.createSpy()
];
bootstrap(config);
expect(bootstrap.preprocessors[0]).toHaveBeenCalled();
expect(bootstrap.preprocessors[1]).toHaveBeenCalled();
bootstrap.preprocessors = originalPreprocessors;
});
});
});
| define([
'client/client_bootstrap',
'client/controllers/dashboard'
],
function (bootstrap, DashboardController) {
describe('client bootstrap', function () {
var config;
beforeEach(function () {
spyOn(DashboardController.prototype, 'render');
config = {
'page-type': 'dashboard'
};
$('body').removeClass('ready loaded');
});
it('instantiates a controller from config data', function () {
var controller = bootstrap(config);
expect(controller instanceof DashboardController).toBe(true);
expect(controller.render).toHaveBeenCalled();
});
it('executes page preprocessors on controller ready', function () {
var originalPreprocessors = bootstrap.preprocessors;
bootstrap.preprocessors = [
jasmine.createSpy(),
jasmine.createSpy()
];
var controller = bootstrap(config);
expect(bootstrap.preprocessors[0]).not.toHaveBeenCalled();
expect(bootstrap.preprocessors[1]).not.toHaveBeenCalled();
controller.trigger('ready');
expect(bootstrap.preprocessors[0]).toHaveBeenCalled();
expect(bootstrap.preprocessors[1]).toHaveBeenCalled();
bootstrap.preprocessors = originalPreprocessors;
});
it('executes page preprocessors immediately if no controller defined', function () {
var originalPreprocessors = bootstrap.preprocessors;
bootstrap.preprocessors = [
jasmine.createSpy(),
jasmine.createSpy()
];
bootstrap({
'page-type': 'no-contoller'
});
expect(bootstrap.preprocessors[0]).toHaveBeenCalled();
expect(bootstrap.preprocessors[1]).toHaveBeenCalled();
bootstrap.preprocessors = originalPreprocessors;
});
it('adds a ready class to the body on controller ready', function () {
var controller = bootstrap(config);
expect($('body').hasClass('ready')).toBe(false);
controller.trigger('ready');
expect($('body').hasClass('ready')).toBe(true);
});
it('adds ready class to body immediately if no controller defined', function () {
bootstrap({
'page-type': 'no-contoller'
});
expect($('body').hasClass('ready')).toBe(true);
});
});
});
| Fix tests for client bootstrap | Fix tests for client bootstrap
Add tests for ready classes
| JavaScript | mit | tijmenb/spotlight,tijmenb/spotlight,alphagov/spotlight,keithiopia/spotlight,tijmenb/spotlight,alphagov/spotlight,keithiopia/spotlight,keithiopia/spotlight,alphagov/spotlight | javascript | ## Code Before:
define([
'client/client_bootstrap',
'client/controllers/dashboard'
],
function (bootstrap, DashboardController) {
describe('client bootstrap', function () {
var config;
beforeEach(function () {
spyOn(DashboardController.prototype, 'render');
config = {
'page-type': 'dashboard'
};
});
it('instantiates a controller from config data', function () {
var controller = bootstrap(config);
expect(controller instanceof DashboardController).toBe(true);
expect(controller.render).toHaveBeenCalled();
});
it('executes page preprocessors', function () {
var originalPreprocessors = bootstrap.preprocessors;
bootstrap.preprocessors = [
jasmine.createSpy(),
jasmine.createSpy()
];
bootstrap(config);
expect(bootstrap.preprocessors[0]).toHaveBeenCalled();
expect(bootstrap.preprocessors[1]).toHaveBeenCalled();
bootstrap.preprocessors = originalPreprocessors;
});
});
});
## Instruction:
Fix tests for client bootstrap
Add tests for ready classes
## Code After:
define([
'client/client_bootstrap',
'client/controllers/dashboard'
],
function (bootstrap, DashboardController) {
describe('client bootstrap', function () {
var config;
beforeEach(function () {
spyOn(DashboardController.prototype, 'render');
config = {
'page-type': 'dashboard'
};
$('body').removeClass('ready loaded');
});
it('instantiates a controller from config data', function () {
var controller = bootstrap(config);
expect(controller instanceof DashboardController).toBe(true);
expect(controller.render).toHaveBeenCalled();
});
it('executes page preprocessors on controller ready', function () {
var originalPreprocessors = bootstrap.preprocessors;
bootstrap.preprocessors = [
jasmine.createSpy(),
jasmine.createSpy()
];
var controller = bootstrap(config);
expect(bootstrap.preprocessors[0]).not.toHaveBeenCalled();
expect(bootstrap.preprocessors[1]).not.toHaveBeenCalled();
controller.trigger('ready');
expect(bootstrap.preprocessors[0]).toHaveBeenCalled();
expect(bootstrap.preprocessors[1]).toHaveBeenCalled();
bootstrap.preprocessors = originalPreprocessors;
});
it('executes page preprocessors immediately if no controller defined', function () {
var originalPreprocessors = bootstrap.preprocessors;
bootstrap.preprocessors = [
jasmine.createSpy(),
jasmine.createSpy()
];
bootstrap({
'page-type': 'no-contoller'
});
expect(bootstrap.preprocessors[0]).toHaveBeenCalled();
expect(bootstrap.preprocessors[1]).toHaveBeenCalled();
bootstrap.preprocessors = originalPreprocessors;
});
it('adds a ready class to the body on controller ready', function () {
var controller = bootstrap(config);
expect($('body').hasClass('ready')).toBe(false);
controller.trigger('ready');
expect($('body').hasClass('ready')).toBe(true);
});
it('adds ready class to body immediately if no controller defined', function () {
bootstrap({
'page-type': 'no-contoller'
});
expect($('body').hasClass('ready')).toBe(true);
});
});
});
| define([
'client/client_bootstrap',
'client/controllers/dashboard'
],
function (bootstrap, DashboardController) {
describe('client bootstrap', function () {
var config;
beforeEach(function () {
spyOn(DashboardController.prototype, 'render');
config = {
'page-type': 'dashboard'
};
+ $('body').removeClass('ready loaded');
});
it('instantiates a controller from config data', function () {
var controller = bootstrap(config);
expect(controller instanceof DashboardController).toBe(true);
expect(controller.render).toHaveBeenCalled();
});
- it('executes page preprocessors', function () {
+ it('executes page preprocessors on controller ready', function () {
? ++++++++++++++++++++
var originalPreprocessors = bootstrap.preprocessors;
bootstrap.preprocessors = [
jasmine.createSpy(),
jasmine.createSpy()
];
- bootstrap(config);
+ var controller = bootstrap(config);
+
+ expect(bootstrap.preprocessors[0]).not.toHaveBeenCalled();
+ expect(bootstrap.preprocessors[1]).not.toHaveBeenCalled();
+ controller.trigger('ready');
expect(bootstrap.preprocessors[0]).toHaveBeenCalled();
expect(bootstrap.preprocessors[1]).toHaveBeenCalled();
bootstrap.preprocessors = originalPreprocessors;
});
+
+ it('executes page preprocessors immediately if no controller defined', function () {
+ var originalPreprocessors = bootstrap.preprocessors;
+ bootstrap.preprocessors = [
+ jasmine.createSpy(),
+ jasmine.createSpy()
+ ];
+
+ bootstrap({
+ 'page-type': 'no-contoller'
+ });
+
+ expect(bootstrap.preprocessors[0]).toHaveBeenCalled();
+ expect(bootstrap.preprocessors[1]).toHaveBeenCalled();
+
+ bootstrap.preprocessors = originalPreprocessors;
+ });
+
+ it('adds a ready class to the body on controller ready', function () {
+ var controller = bootstrap(config);
+
+ expect($('body').hasClass('ready')).toBe(false);
+ controller.trigger('ready');
+ expect($('body').hasClass('ready')).toBe(true);
+ });
+
+ it('adds ready class to body immediately if no controller defined', function () {
+ bootstrap({
+ 'page-type': 'no-contoller'
+ });
+
+ expect($('body').hasClass('ready')).toBe(true);
+ });
});
}); | 42 | 1.135135 | 40 | 2 |
304a888a8e96c2e54281bb7af045d4a311636e97 | packages/an/animate-sdl2.yaml | packages/an/animate-sdl2.yaml | homepage: https://github.com/jxv/animate-sdl2#readme
changelog-type: ''
hash: e26f1306fb29f678de88ff7a470804c73e4c1ca0d999a97d36148f46ace284cb
test-bench-deps: {}
maintainer: Joe Vargas
synopsis: sdl2 + animate auxiliary library
changelog: ''
basic-deps:
base: ! '>=4.7 && <5'
sdl2: ! '>=2.4.0.1 && <2.5'
animate: ! '>=0.6.1 && <1'
sdl2-image: ! '>=2.0.0 && <3'
aeson: ! '>=0.11 && <2'
all-versions:
- '0.0.0'
author: ''
latest: '0.0.0'
description-type: markdown
description: ! '# animate-sdl2
[`sdl2`](https://github.com/haskell-game/sdl2) is a commonly used media library.
[`animate`](https://github.com/jxv/animate) is a general animation library.
Combining `sdl2` and `animate`, `animate-sdl2` provides accessible glue-code to
load and draw sprites.'
license-name: BSD3
| homepage: https://github.com/jxv/animate-sdl2#readme
changelog-type: ''
hash: b917411ff96a1f12f62d8adc5a006a027de3ba1d50a49981128a5f4f330f9102
test-bench-deps: {}
maintainer: Joe Vargas
synopsis: sdl2 + animate auxiliary library
changelog: ''
basic-deps:
base: ! '>=4.7 && <5'
sdl2: ! '>=2.4.0.1 && <2.5'
animate: ! '>=0.6.1 && <1'
sdl2-image: ! '>=2.0.0 && <3'
aeson: ! '>=0.11 && <2'
all-versions:
- '0.0.0'
- '0.1.0'
author: ''
latest: '0.1.0'
description-type: markdown
description: ! '# animate-sdl2
[`sdl2`](https://github.com/haskell-game/sdl2) is a commonly used media library.
[`animate`](https://github.com/jxv/animate) is a general animation library.
Combining `sdl2` and `animate`, `animate-sdl2` provides accessible glue-code to
load and draw sprites.'
license-name: BSD3
| Update from Hackage at 2018-05-12T22:59:54Z | Update from Hackage at 2018-05-12T22:59:54Z
| YAML | mit | commercialhaskell/all-cabal-metadata | yaml | ## Code Before:
homepage: https://github.com/jxv/animate-sdl2#readme
changelog-type: ''
hash: e26f1306fb29f678de88ff7a470804c73e4c1ca0d999a97d36148f46ace284cb
test-bench-deps: {}
maintainer: Joe Vargas
synopsis: sdl2 + animate auxiliary library
changelog: ''
basic-deps:
base: ! '>=4.7 && <5'
sdl2: ! '>=2.4.0.1 && <2.5'
animate: ! '>=0.6.1 && <1'
sdl2-image: ! '>=2.0.0 && <3'
aeson: ! '>=0.11 && <2'
all-versions:
- '0.0.0'
author: ''
latest: '0.0.0'
description-type: markdown
description: ! '# animate-sdl2
[`sdl2`](https://github.com/haskell-game/sdl2) is a commonly used media library.
[`animate`](https://github.com/jxv/animate) is a general animation library.
Combining `sdl2` and `animate`, `animate-sdl2` provides accessible glue-code to
load and draw sprites.'
license-name: BSD3
## Instruction:
Update from Hackage at 2018-05-12T22:59:54Z
## Code After:
homepage: https://github.com/jxv/animate-sdl2#readme
changelog-type: ''
hash: b917411ff96a1f12f62d8adc5a006a027de3ba1d50a49981128a5f4f330f9102
test-bench-deps: {}
maintainer: Joe Vargas
synopsis: sdl2 + animate auxiliary library
changelog: ''
basic-deps:
base: ! '>=4.7 && <5'
sdl2: ! '>=2.4.0.1 && <2.5'
animate: ! '>=0.6.1 && <1'
sdl2-image: ! '>=2.0.0 && <3'
aeson: ! '>=0.11 && <2'
all-versions:
- '0.0.0'
- '0.1.0'
author: ''
latest: '0.1.0'
description-type: markdown
description: ! '# animate-sdl2
[`sdl2`](https://github.com/haskell-game/sdl2) is a commonly used media library.
[`animate`](https://github.com/jxv/animate) is a general animation library.
Combining `sdl2` and `animate`, `animate-sdl2` provides accessible glue-code to
load and draw sprites.'
license-name: BSD3
| homepage: https://github.com/jxv/animate-sdl2#readme
changelog-type: ''
- hash: e26f1306fb29f678de88ff7a470804c73e4c1ca0d999a97d36148f46ace284cb
+ hash: b917411ff96a1f12f62d8adc5a006a027de3ba1d50a49981128a5f4f330f9102
test-bench-deps: {}
maintainer: Joe Vargas
synopsis: sdl2 + animate auxiliary library
changelog: ''
basic-deps:
base: ! '>=4.7 && <5'
sdl2: ! '>=2.4.0.1 && <2.5'
animate: ! '>=0.6.1 && <1'
sdl2-image: ! '>=2.0.0 && <3'
aeson: ! '>=0.11 && <2'
all-versions:
- '0.0.0'
+ - '0.1.0'
author: ''
- latest: '0.0.0'
? ^
+ latest: '0.1.0'
? ^
description-type: markdown
description: ! '# animate-sdl2
[`sdl2`](https://github.com/haskell-game/sdl2) is a commonly used media library.
[`animate`](https://github.com/jxv/animate) is a general animation library.
Combining `sdl2` and `animate`, `animate-sdl2` provides accessible glue-code to
load and draw sprites.'
license-name: BSD3 | 5 | 0.178571 | 3 | 2 |
ed8dd70918fddfe9928ddb43cb7c4b1fa3ad899b | README.rst | README.rst | ===============================
Nose-Watcher
===============================
.. image:: https://badge.fury.io/py/nose-watcher.png
:target: http://badge.fury.io/py/nose-watcher
.. image:: https://travis-ci.org/solarnz/nose-watcher.png?branch=master
:target: https://travis-ci.org/solarnz/nose-watcher
.. image:: https://pypip.in/d/nose-watcher/badge.png
:target: https://pypi.python.org/pypi/nose-watcher
.. image:: https://coveralls.io/repos/solarnz/nose-watcher/badge.png?branch=master
:target: https://coveralls.io/r/solarnz/nose-watcher?branch=master
A nose plugin to watch for changes within the local directory.
* Free software: BSD license
* Documentation: http://nose-watcher.readthedocs.org.
Inspired by the `nose-watch <https://github.com/lukaszb/nose-watch>`_ nose
plugin.
Note: nose-watcher will only run on linux, due to the depenency on `python-inotify` and `inotify`.
Features
--------
* Watches for changes in the local directory, then runs nosetests with the
specified command line options.
* Doesn't run the tests multiple times if you're using vim, Unlike the similar
plugin `nose-watch`.
Installation_
------------
.. code-block:: shell
pip install nose-watcher
Usage
-----
.. code-block:: shell
nosetests --with-watcher
| ===============================
Nose-Watcher
===============================
.. image:: https://badge.fury.io/py/nose-watcher.png
:target: http://badge.fury.io/py/nose-watcher
.. image:: https://travis-ci.org/solarnz/nose-watcher.png?branch=develop
:target: https://travis-ci.org/solarnz/nose-watcher
.. image:: https://pypip.in/d/nose-watcher/badge.png
:target: https://pypi.python.org/pypi/nose-watcher
.. image:: https://coveralls.io/repos/solarnz/nose-watcher/badge.png?branch=develop
:target: https://coveralls.io/r/solarnz/nose-watcher?branch=develop
A nose plugin to watch for changes within the local directory.
* Free software: BSD license
* Documentation: http://nose-watcher.readthedocs.org.
Inspired by the `nose-watch <https://github.com/lukaszb/nose-watch>`_ nose
plugin.
Note: nose-watcher will only run on linux, due to the depenency on `python-inotify` and `inotify`.
Features
--------
* Watches for changes in the local directory, then runs nosetests with the
specified command line options.
* Doesn't run the tests multiple times if you're using vim, Unlike the similar
plugin `nose-watch`.
Installation_
------------
.. code-block:: shell
pip install nose-watcher
Usage
-----
.. code-block:: shell
nosetests --with-watcher
| Switch the badge branches to develop | Switch the badge branches to develop
| reStructuredText | bsd-3-clause | solarnz/nose-watcher,solarnz/nose-watcher | restructuredtext | ## Code Before:
===============================
Nose-Watcher
===============================
.. image:: https://badge.fury.io/py/nose-watcher.png
:target: http://badge.fury.io/py/nose-watcher
.. image:: https://travis-ci.org/solarnz/nose-watcher.png?branch=master
:target: https://travis-ci.org/solarnz/nose-watcher
.. image:: https://pypip.in/d/nose-watcher/badge.png
:target: https://pypi.python.org/pypi/nose-watcher
.. image:: https://coveralls.io/repos/solarnz/nose-watcher/badge.png?branch=master
:target: https://coveralls.io/r/solarnz/nose-watcher?branch=master
A nose plugin to watch for changes within the local directory.
* Free software: BSD license
* Documentation: http://nose-watcher.readthedocs.org.
Inspired by the `nose-watch <https://github.com/lukaszb/nose-watch>`_ nose
plugin.
Note: nose-watcher will only run on linux, due to the depenency on `python-inotify` and `inotify`.
Features
--------
* Watches for changes in the local directory, then runs nosetests with the
specified command line options.
* Doesn't run the tests multiple times if you're using vim, Unlike the similar
plugin `nose-watch`.
Installation_
------------
.. code-block:: shell
pip install nose-watcher
Usage
-----
.. code-block:: shell
nosetests --with-watcher
## Instruction:
Switch the badge branches to develop
## Code After:
===============================
Nose-Watcher
===============================
.. image:: https://badge.fury.io/py/nose-watcher.png
:target: http://badge.fury.io/py/nose-watcher
.. image:: https://travis-ci.org/solarnz/nose-watcher.png?branch=develop
:target: https://travis-ci.org/solarnz/nose-watcher
.. image:: https://pypip.in/d/nose-watcher/badge.png
:target: https://pypi.python.org/pypi/nose-watcher
.. image:: https://coveralls.io/repos/solarnz/nose-watcher/badge.png?branch=develop
:target: https://coveralls.io/r/solarnz/nose-watcher?branch=develop
A nose plugin to watch for changes within the local directory.
* Free software: BSD license
* Documentation: http://nose-watcher.readthedocs.org.
Inspired by the `nose-watch <https://github.com/lukaszb/nose-watch>`_ nose
plugin.
Note: nose-watcher will only run on linux, due to the depenency on `python-inotify` and `inotify`.
Features
--------
* Watches for changes in the local directory, then runs nosetests with the
specified command line options.
* Doesn't run the tests multiple times if you're using vim, Unlike the similar
plugin `nose-watch`.
Installation_
------------
.. code-block:: shell
pip install nose-watcher
Usage
-----
.. code-block:: shell
nosetests --with-watcher
| ===============================
Nose-Watcher
===============================
.. image:: https://badge.fury.io/py/nose-watcher.png
:target: http://badge.fury.io/py/nose-watcher
- .. image:: https://travis-ci.org/solarnz/nose-watcher.png?branch=master
? ^^^^ ^
+ .. image:: https://travis-ci.org/solarnz/nose-watcher.png?branch=develop
? ^ ^^^^^
:target: https://travis-ci.org/solarnz/nose-watcher
.. image:: https://pypip.in/d/nose-watcher/badge.png
:target: https://pypi.python.org/pypi/nose-watcher
- .. image:: https://coveralls.io/repos/solarnz/nose-watcher/badge.png?branch=master
? ^^^^ ^
+ .. image:: https://coveralls.io/repos/solarnz/nose-watcher/badge.png?branch=develop
? ^ ^^^^^
- :target: https://coveralls.io/r/solarnz/nose-watcher?branch=master
? ^^^^ ^
+ :target: https://coveralls.io/r/solarnz/nose-watcher?branch=develop
? ^ ^^^^^
A nose plugin to watch for changes within the local directory.
* Free software: BSD license
* Documentation: http://nose-watcher.readthedocs.org.
Inspired by the `nose-watch <https://github.com/lukaszb/nose-watch>`_ nose
plugin.
Note: nose-watcher will only run on linux, due to the depenency on `python-inotify` and `inotify`.
Features
--------
* Watches for changes in the local directory, then runs nosetests with the
specified command line options.
* Doesn't run the tests multiple times if you're using vim, Unlike the similar
plugin `nose-watch`.
Installation_
------------
.. code-block:: shell
pip install nose-watcher
Usage
-----
.. code-block:: shell
nosetests --with-watcher | 6 | 0.12 | 3 | 3 |
0eb5fefaaaedf8664ed36b8903459e3cfdd0e804 | Bluetility/AppDelegate.swift | Bluetility/AppDelegate.swift | //
// AppDelegate.swift
// Bluetility
//
// Created by Joseph Ross on 11/9/15.
// Copyright © 2015 Joseph Ross. All rights reserved.
//
import Cocoa
@NSApplicationMain
class AppDelegate: NSObject, NSApplicationDelegate {
func applicationDidFinishLaunching(aNotification: NSNotification) {
// Insert code here to initialize your application
}
func applicationWillTerminate(aNotification: NSNotification) {
// Insert code here to tear down your application
}
func applicationShouldTerminateAfterLastWindowClosed(sender: NSApplication) -> Bool {
return true
}
}
| //
// AppDelegate.swift
// Bluetility
//
// Created by Joseph Ross on 11/9/15.
// Copyright © 2015 Joseph Ross. All rights reserved.
//
import Cocoa
@NSApplicationMain
class AppDelegate: NSObject, NSApplicationDelegate {
func applicationDidFinishLaunching(aNotification: NSNotification) {
// Insert code here to initialize your application
let defaults = ["NSToolTipAutoWrappingDisabled" : NSNumber(bool: true)]
NSUserDefaults.standardUserDefaults().registerDefaults(defaults)
}
func applicationWillTerminate(aNotification: NSNotification) {
// Insert code here to tear down your application
}
func applicationShouldTerminateAfterLastWindowClosed(sender: NSApplication) -> Bool {
return true
}
}
| Disable tooltip wrapping by default | Disable tooltip wrapping by default
| Swift | mit | jnross/Bluetility,jnross/Bluetility,jnross/Bluetility | swift | ## Code Before:
//
// AppDelegate.swift
// Bluetility
//
// Created by Joseph Ross on 11/9/15.
// Copyright © 2015 Joseph Ross. All rights reserved.
//
import Cocoa
@NSApplicationMain
class AppDelegate: NSObject, NSApplicationDelegate {
func applicationDidFinishLaunching(aNotification: NSNotification) {
// Insert code here to initialize your application
}
func applicationWillTerminate(aNotification: NSNotification) {
// Insert code here to tear down your application
}
func applicationShouldTerminateAfterLastWindowClosed(sender: NSApplication) -> Bool {
return true
}
}
## Instruction:
Disable tooltip wrapping by default
## Code After:
//
// AppDelegate.swift
// Bluetility
//
// Created by Joseph Ross on 11/9/15.
// Copyright © 2015 Joseph Ross. All rights reserved.
//
import Cocoa
@NSApplicationMain
class AppDelegate: NSObject, NSApplicationDelegate {
func applicationDidFinishLaunching(aNotification: NSNotification) {
// Insert code here to initialize your application
let defaults = ["NSToolTipAutoWrappingDisabled" : NSNumber(bool: true)]
NSUserDefaults.standardUserDefaults().registerDefaults(defaults)
}
func applicationWillTerminate(aNotification: NSNotification) {
// Insert code here to tear down your application
}
func applicationShouldTerminateAfterLastWindowClosed(sender: NSApplication) -> Bool {
return true
}
}
| //
// AppDelegate.swift
// Bluetility
//
// Created by Joseph Ross on 11/9/15.
// Copyright © 2015 Joseph Ross. All rights reserved.
//
import Cocoa
@NSApplicationMain
class AppDelegate: NSObject, NSApplicationDelegate {
func applicationDidFinishLaunching(aNotification: NSNotification) {
// Insert code here to initialize your application
+ let defaults = ["NSToolTipAutoWrappingDisabled" : NSNumber(bool: true)]
+ NSUserDefaults.standardUserDefaults().registerDefaults(defaults)
}
func applicationWillTerminate(aNotification: NSNotification) {
// Insert code here to tear down your application
}
func applicationShouldTerminateAfterLastWindowClosed(sender: NSApplication) -> Bool {
return true
}
}
| 2 | 0.068966 | 2 | 0 |
0ec1cc5e5f9986f466942ac4ddcc2f4acfeb100d | .travis.yml | .travis.yml | ---
language: ruby
bundler_args: --without local_only
before_install: rm Gemfile.lock || true
rvm:
- 2.1.10
- 2.2.6
- 2.3.3
- 2.4.0
script: bundle exec rake test
env:
- PUPPET_VERSION="~> 4.6.0"
- PUPPET_VERSION="~> 4.7.0"
- PUPPET_VERSION="~> 4.8.0"
matrix:
exclude:
- rvm: 2.4.0
env: PUPPET_VERSION="~> 4.6.0"
- rvm: 2.4.0
env: PUPPET_VERSION="~> 4.7.0"
notifications:
email:
recipients:
- dfarrell@redhat.com
on_success: change
on_failure: change
| ---
language: ruby
bundler_args: --without local_only
before_install: rm Gemfile.lock || true
rvm:
- 2.1.10
- 2.2.6
- 2.3.3
- 2.4.0
script: bundle exec rake test
env:
- PUPPET_VERSION="~> 4.6.0"
- PUPPET_VERSION="~> 4.7.0"
- PUPPET_VERSION="~> 4.8.0"
- PUPPET_VERSION="~> 4.9.0"
matrix:
exclude:
- rvm: 2.4.0
env: PUPPET_VERSION="~> 4.6.0"
- rvm: 2.4.0
env: PUPPET_VERSION="~> 4.7.0"
notifications:
email:
recipients:
- dfarrell@redhat.com
on_success: change
on_failure: change
| Add Puppet version 4.9.0 to Travis test matrix | Add Puppet version 4.9.0 to Travis test matrix
Signed-off-by: Akshita Jha <21ecdd397585d71ddbb62f3289ec4230a99dde49@gmail.com>
| YAML | bsd-2-clause | umeshksingla/puppet-opendaylight,umeshksingla/puppet-opendaylight | yaml | ## Code Before:
---
language: ruby
bundler_args: --without local_only
before_install: rm Gemfile.lock || true
rvm:
- 2.1.10
- 2.2.6
- 2.3.3
- 2.4.0
script: bundle exec rake test
env:
- PUPPET_VERSION="~> 4.6.0"
- PUPPET_VERSION="~> 4.7.0"
- PUPPET_VERSION="~> 4.8.0"
matrix:
exclude:
- rvm: 2.4.0
env: PUPPET_VERSION="~> 4.6.0"
- rvm: 2.4.0
env: PUPPET_VERSION="~> 4.7.0"
notifications:
email:
recipients:
- dfarrell@redhat.com
on_success: change
on_failure: change
## Instruction:
Add Puppet version 4.9.0 to Travis test matrix
Signed-off-by: Akshita Jha <21ecdd397585d71ddbb62f3289ec4230a99dde49@gmail.com>
## Code After:
---
language: ruby
bundler_args: --without local_only
before_install: rm Gemfile.lock || true
rvm:
- 2.1.10
- 2.2.6
- 2.3.3
- 2.4.0
script: bundle exec rake test
env:
- PUPPET_VERSION="~> 4.6.0"
- PUPPET_VERSION="~> 4.7.0"
- PUPPET_VERSION="~> 4.8.0"
- PUPPET_VERSION="~> 4.9.0"
matrix:
exclude:
- rvm: 2.4.0
env: PUPPET_VERSION="~> 4.6.0"
- rvm: 2.4.0
env: PUPPET_VERSION="~> 4.7.0"
notifications:
email:
recipients:
- dfarrell@redhat.com
on_success: change
on_failure: change
| ---
language: ruby
bundler_args: --without local_only
before_install: rm Gemfile.lock || true
rvm:
- 2.1.10
- 2.2.6
- 2.3.3
- 2.4.0
script: bundle exec rake test
env:
- PUPPET_VERSION="~> 4.6.0"
- PUPPET_VERSION="~> 4.7.0"
- PUPPET_VERSION="~> 4.8.0"
+ - PUPPET_VERSION="~> 4.9.0"
matrix:
exclude:
- rvm: 2.4.0
env: PUPPET_VERSION="~> 4.6.0"
- rvm: 2.4.0
env: PUPPET_VERSION="~> 4.7.0"
notifications:
email:
recipients:
- dfarrell@redhat.com
on_success: change
on_failure: change | 1 | 0.038462 | 1 | 0 |
52004dc05258071dc85dd70b96ad2f7d47e903fd | README.md | README.md |
[](https://travis-ci.org/jonas-schievink/xml-rpc-rs)
[Documentation](https://docs.rs/xmlrpc/)
|
[](https://travis-ci.org/jonas-schievink/xml-rpc-rs)
[Documentation](https://docs.rs/xmlrpc/)
This crate provides a simple implementation of the [XML-RPC spec](http://xmlrpc.scripting.com/spec.html) in pure Rust using `xml-rs` and `hyper`.
See `examples/client.rs` for a small example which connects to a running Python XML-RPC server and calls a method.
| Add some text to the readme | Add some text to the readme
| Markdown | cc0-1.0 | jonas-schievink/xml-rpc-rs | markdown | ## Code Before:
[](https://travis-ci.org/jonas-schievink/xml-rpc-rs)
[Documentation](https://docs.rs/xmlrpc/)
## Instruction:
Add some text to the readme
## Code After:
[](https://travis-ci.org/jonas-schievink/xml-rpc-rs)
[Documentation](https://docs.rs/xmlrpc/)
This crate provides a simple implementation of the [XML-RPC spec](http://xmlrpc.scripting.com/spec.html) in pure Rust using `xml-rs` and `hyper`.
See `examples/client.rs` for a small example which connects to a running Python XML-RPC server and calls a method.
|
[](https://travis-ci.org/jonas-schievink/xml-rpc-rs)
[Documentation](https://docs.rs/xmlrpc/)
+
+ This crate provides a simple implementation of the [XML-RPC spec](http://xmlrpc.scripting.com/spec.html) in pure Rust using `xml-rs` and `hyper`.
+
+ See `examples/client.rs` for a small example which connects to a running Python XML-RPC server and calls a method. | 4 | 1 | 4 | 0 |
23de9ee2c4af848354ecafb73104919ae4137ef4 | src/help_ddis.txt | src/help_ddis.txt | D DISassembler for DCPU-16 v1.7
Usage:
./ddis filename [options]
Options:
-h Show this message
-t or --type <type> Type of file with memory map. lraw -> little endian raw binary ; braw -> big endian raw binary ; ahex -> ascii hexadecimal file ; hexd -> ascii hexadecimal dump file. By default ddis asumes little endian raw binary input file
-c Add comments to all lines of the style [address] xxxx .... where xxxx its the hexadecimal representation of these instruction.
-l Autolabel all jumps (actually only SET PC, ....)
-b<number> Sets the absolute position were begin to disassembly the file. By default it's 0
-e<number> Sets the absolute position were end to disassembly the file. By default it's the end of the file.
| D DISassembler for DCPU-16 v1.7
Usage:
./ddis filename [options]
Options:
-h Show this message
-t or --type <type> Type of file with memory map. lraw -> little endian raw binary ; braw -> big endian raw binary ; ahex -> ascii hexadecimal file ; hexd -> ascii hexadecimal dump file ; dat -> Read DATs from a dasm file. By default ddis asumes little endian raw binary input file.
-c Add comments to all lines of the style [address] xxxx .... where xxxx its the hexadecimal representation of these instruction.
-l Autolabel all jumps (actually only SET PC, ....)
-b<number> Sets the absolute position were begin to disassembly the file. By default it's 0
-e<number> Sets the absolute position were end to disassembly the file. By default it's the end of the file.
| Update help information about ddis | Update help information about ddis
| Text | bsd-3-clause | Zardoz89/DEDCPU-16 | text | ## Code Before:
D DISassembler for DCPU-16 v1.7
Usage:
./ddis filename [options]
Options:
-h Show this message
-t or --type <type> Type of file with memory map. lraw -> little endian raw binary ; braw -> big endian raw binary ; ahex -> ascii hexadecimal file ; hexd -> ascii hexadecimal dump file. By default ddis asumes little endian raw binary input file
-c Add comments to all lines of the style [address] xxxx .... where xxxx its the hexadecimal representation of these instruction.
-l Autolabel all jumps (actually only SET PC, ....)
-b<number> Sets the absolute position were begin to disassembly the file. By default it's 0
-e<number> Sets the absolute position were end to disassembly the file. By default it's the end of the file.
## Instruction:
Update help information about ddis
## Code After:
D DISassembler for DCPU-16 v1.7
Usage:
./ddis filename [options]
Options:
-h Show this message
-t or --type <type> Type of file with memory map. lraw -> little endian raw binary ; braw -> big endian raw binary ; ahex -> ascii hexadecimal file ; hexd -> ascii hexadecimal dump file ; dat -> Read DATs from a dasm file. By default ddis asumes little endian raw binary input file.
-c Add comments to all lines of the style [address] xxxx .... where xxxx its the hexadecimal representation of these instruction.
-l Autolabel all jumps (actually only SET PC, ....)
-b<number> Sets the absolute position were begin to disassembly the file. By default it's 0
-e<number> Sets the absolute position were end to disassembly the file. By default it's the end of the file.
| D DISassembler for DCPU-16 v1.7
Usage:
./ddis filename [options]
Options:
-h Show this message
- -t or --type <type> Type of file with memory map. lraw -> little endian raw binary ; braw -> big endian raw binary ; ahex -> ascii hexadecimal file ; hexd -> ascii hexadecimal dump file. By default ddis asumes little endian raw binary input file
+ -t or --type <type> Type of file with memory map. lraw -> little endian raw binary ; braw -> big endian raw binary ; ahex -> ascii hexadecimal file ; hexd -> ascii hexadecimal dump file ; dat -> Read DATs from a dasm file. By default ddis asumes little endian raw binary input file.
? ++++++++++++++++++++++++++++++++++++ +
-c Add comments to all lines of the style [address] xxxx .... where xxxx its the hexadecimal representation of these instruction.
-l Autolabel all jumps (actually only SET PC, ....)
-b<number> Sets the absolute position were begin to disassembly the file. By default it's 0
-e<number> Sets the absolute position were end to disassembly the file. By default it's the end of the file.
| 2 | 0.166667 | 1 | 1 |
77637d7f1cff1d390b935cab74cbeb6254b703f7 | __template/index.html | __template/index.html | <!DOCTYPE html>
<html>
<head>
<title>My Awesome Screensaver</title>
<script>
// this little javascript snippet will parse any incoming URL
// parameters and place them in the window.urlParams object
window.urlParams = window.location.search.split(/[?&]/).slice(1).map(function(paramPair) {
return paramPair.split(/=(.+)?/).slice(0, 2);
}).reduce(function (obj, pairArray) {
obj[pairArray[0]] = pairArray[1];
return obj;
}, {});
</script>
</head>
<body>
Put the content of your screensaver here!
</body>
</html>
| <!DOCTYPE html>
<html>
<head>
<title>My Awesome Screensaver</title>
<script>
// load any incoming URL parameters. you could just use the
// URLSearchParams object directly to manage these variables, but
// having them in a hash is a little easier sometimes.
var tmpParams = new URLSearchParams(document.location.search);
window.urlParams = {};
for(let k of tmpParams.keys() ) {
window.urlParams[k] = tmpParams.get(k);
}
</script>
</head>
<body>
<h1>Hello! I am a screensaver template!</h1>
<p>I am located in <code><script>document.write(document.location.pathname);</script></code>.</p>
<p>Put the content of your screensaver here!</p>
<h2>Incoming Values</h2>
<p>(These values will be sent to your screensaver when it loads)</p>
<ul>
<script>
for(let k of Object.keys(window.urlParams) ) {
document.write("<li><code>window.urlParams[" + k + "]</code>: " + window.urlParams[k] + "</li>");
}
</script>
</ul>
</body>
</html>
| Clean up the screensaver template | Clean up the screensaver template
| HTML | mit | muffinista/before-dawn-screensavers,muffinista/before-dawn-screensavers,muffinista/before-dawn-screensavers | html | ## Code Before:
<!DOCTYPE html>
<html>
<head>
<title>My Awesome Screensaver</title>
<script>
// this little javascript snippet will parse any incoming URL
// parameters and place them in the window.urlParams object
window.urlParams = window.location.search.split(/[?&]/).slice(1).map(function(paramPair) {
return paramPair.split(/=(.+)?/).slice(0, 2);
}).reduce(function (obj, pairArray) {
obj[pairArray[0]] = pairArray[1];
return obj;
}, {});
</script>
</head>
<body>
Put the content of your screensaver here!
</body>
</html>
## Instruction:
Clean up the screensaver template
## Code After:
<!DOCTYPE html>
<html>
<head>
<title>My Awesome Screensaver</title>
<script>
// load any incoming URL parameters. you could just use the
// URLSearchParams object directly to manage these variables, but
// having them in a hash is a little easier sometimes.
var tmpParams = new URLSearchParams(document.location.search);
window.urlParams = {};
for(let k of tmpParams.keys() ) {
window.urlParams[k] = tmpParams.get(k);
}
</script>
</head>
<body>
<h1>Hello! I am a screensaver template!</h1>
<p>I am located in <code><script>document.write(document.location.pathname);</script></code>.</p>
<p>Put the content of your screensaver here!</p>
<h2>Incoming Values</h2>
<p>(These values will be sent to your screensaver when it loads)</p>
<ul>
<script>
for(let k of Object.keys(window.urlParams) ) {
document.write("<li><code>window.urlParams[" + k + "]</code>: " + window.urlParams[k] + "</li>");
}
</script>
</ul>
</body>
</html>
| <!DOCTYPE html>
<html>
<head>
<title>My Awesome Screensaver</title>
<script>
- // this little javascript snippet will parse any incoming URL
- // parameters and place them in the window.urlParams object
- window.urlParams = window.location.search.split(/[?&]/).slice(1).map(function(paramPair) {
- return paramPair.split(/=(.+)?/).slice(0, 2);
- }).reduce(function (obj, pairArray) {
- obj[pairArray[0]] = pairArray[1];
- return obj;
- }, {});
+ // load any incoming URL parameters. you could just use the
+ // URLSearchParams object directly to manage these variables, but
+ // having them in a hash is a little easier sometimes.
+ var tmpParams = new URLSearchParams(document.location.search);
+ window.urlParams = {};
+
+ for(let k of tmpParams.keys() ) {
+ window.urlParams[k] = tmpParams.get(k);
+ }
</script>
</head>
<body>
+ <h1>Hello! I am a screensaver template!</h1>
+ <p>I am located in <code><script>document.write(document.location.pathname);</script></code>.</p>
- Put the content of your screensaver here!
+ <p>Put the content of your screensaver here!</p>
? +++ ++++
+ <h2>Incoming Values</h2>
+ <p>(These values will be sent to your screensaver when it loads)</p>
+ <ul>
+ <script>
+ for(let k of Object.keys(window.urlParams) ) {
+ document.write("<li><code>window.urlParams[" + k + "]</code>: " + window.urlParams[k] + "</li>");
+ }
+ </script>
+ </ul>
</body>
</html> | 30 | 1.578947 | 21 | 9 |
8207e9b84e9581d286baf5869b9dcbf6351ddd50 | nwn2mdk-lib/cgmath.h | nwn2mdk-lib/cgmath.h |
template <typename T>
class Vector3 {
public:
T x, y, z;
Vector3(T x, T y, T z)
{
this->x = x;
this->y = y;
this->z = z;
}
};
template <typename T>
class Vector4 {
public:
T x, y, z, w;
Vector4() {}
Vector4(T x, T y, T z, T w)
{
this->x = x;
this->y = y;
this->z = z;
this->w = w;
}
T &operator[](unsigned i)
{
return (&x)[i];
}
};
|
template <typename T>
class Vector3 {
public:
T x, y, z;
Vector3(T x, T y, T z)
{
this->x = x;
this->y = y;
this->z = z;
}
T &operator[](unsigned i)
{
return (&x)[i];
}
};
template <typename T>
class Vector4 {
public:
T x, y, z, w;
Vector4() {}
Vector4(T x, T y, T z, T w)
{
this->x = x;
this->y = y;
this->z = z;
this->w = w;
}
T &operator[](unsigned i)
{
return (&x)[i];
}
};
| Add subscript operator to Vector3 | Add subscript operator to Vector3
| C | apache-2.0 | Arbos/nwn2mdk,Arbos/nwn2mdk,Arbos/nwn2mdk | c | ## Code Before:
template <typename T>
class Vector3 {
public:
T x, y, z;
Vector3(T x, T y, T z)
{
this->x = x;
this->y = y;
this->z = z;
}
};
template <typename T>
class Vector4 {
public:
T x, y, z, w;
Vector4() {}
Vector4(T x, T y, T z, T w)
{
this->x = x;
this->y = y;
this->z = z;
this->w = w;
}
T &operator[](unsigned i)
{
return (&x)[i];
}
};
## Instruction:
Add subscript operator to Vector3
## Code After:
template <typename T>
class Vector3 {
public:
T x, y, z;
Vector3(T x, T y, T z)
{
this->x = x;
this->y = y;
this->z = z;
}
T &operator[](unsigned i)
{
return (&x)[i];
}
};
template <typename T>
class Vector4 {
public:
T x, y, z, w;
Vector4() {}
Vector4(T x, T y, T z, T w)
{
this->x = x;
this->y = y;
this->z = z;
this->w = w;
}
T &operator[](unsigned i)
{
return (&x)[i];
}
};
|
template <typename T>
class Vector3 {
public:
T x, y, z;
Vector3(T x, T y, T z)
{
this->x = x;
this->y = y;
this->z = z;
+ }
+
+ T &operator[](unsigned i)
+ {
+ return (&x)[i];
}
};
template <typename T>
class Vector4 {
public:
T x, y, z, w;
Vector4() {}
Vector4(T x, T y, T z, T w)
{
this->x = x;
this->y = y;
this->z = z;
this->w = w;
}
T &operator[](unsigned i)
{
return (&x)[i];
}
}; | 5 | 0.147059 | 5 | 0 |
3b70f211c42d7265a292970e71fc4328d9e4f648 | metadata/io.github.domi04151309.powerapp.yml | metadata/io.github.domi04151309.powerapp.yml | Categories:
- System
License: GPL-3.0-only
AuthorName: Dominik
AuthorWebSite: https://domi04151309.github.io/
WebSite: https://domi04151309.github.io/Android/Powerapp
SourceCode: https://github.com/Domi04151309/Power-App-for-Android
IssueTracker: https://github.com/Domi04151309/Power-App-for-Android/issues
AutoName: Power App
RequiresRoot: 'yes'
RepoType: git
Repo: https://github.com/Domi04151309/Power-App-for-Android
Builds:
- versionName: 1.8.3
versionCode: 183
commit: 1.8.3
subdir: app
gradle:
- yes
- versionName: 1.8.4
versionCode: 184
commit: 1.8.4
subdir: app
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 1.8.4
CurrentVersionCode: 184
| Categories:
- System
License: GPL-3.0-only
AuthorName: Dominik
AuthorWebSite: https://domi04151309.github.io/
WebSite: https://domi04151309.github.io/Android/Powerapp
SourceCode: https://github.com/Domi04151309/Power-App-for-Android
IssueTracker: https://github.com/Domi04151309/Power-App-for-Android/issues
AutoName: Power App
RequiresRoot: 'yes'
RepoType: git
Repo: https://github.com/Domi04151309/Power-App-for-Android
Builds:
- versionName: 1.8.3
versionCode: 183
commit: 1.8.3
subdir: app
gradle:
- yes
- versionName: 1.8.4
versionCode: 184
commit: 1.8.4
subdir: app
gradle:
- yes
- versionName: 1.8.5
versionCode: 185
commit: 1.8.5
subdir: app
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 1.8.5
CurrentVersionCode: 185
| Update Power App to 1.8.5 (185) | Update Power App to 1.8.5 (185)
| YAML | agpl-3.0 | f-droid/fdroiddata,f-droid/fdroiddata | yaml | ## Code Before:
Categories:
- System
License: GPL-3.0-only
AuthorName: Dominik
AuthorWebSite: https://domi04151309.github.io/
WebSite: https://domi04151309.github.io/Android/Powerapp
SourceCode: https://github.com/Domi04151309/Power-App-for-Android
IssueTracker: https://github.com/Domi04151309/Power-App-for-Android/issues
AutoName: Power App
RequiresRoot: 'yes'
RepoType: git
Repo: https://github.com/Domi04151309/Power-App-for-Android
Builds:
- versionName: 1.8.3
versionCode: 183
commit: 1.8.3
subdir: app
gradle:
- yes
- versionName: 1.8.4
versionCode: 184
commit: 1.8.4
subdir: app
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 1.8.4
CurrentVersionCode: 184
## Instruction:
Update Power App to 1.8.5 (185)
## Code After:
Categories:
- System
License: GPL-3.0-only
AuthorName: Dominik
AuthorWebSite: https://domi04151309.github.io/
WebSite: https://domi04151309.github.io/Android/Powerapp
SourceCode: https://github.com/Domi04151309/Power-App-for-Android
IssueTracker: https://github.com/Domi04151309/Power-App-for-Android/issues
AutoName: Power App
RequiresRoot: 'yes'
RepoType: git
Repo: https://github.com/Domi04151309/Power-App-for-Android
Builds:
- versionName: 1.8.3
versionCode: 183
commit: 1.8.3
subdir: app
gradle:
- yes
- versionName: 1.8.4
versionCode: 184
commit: 1.8.4
subdir: app
gradle:
- yes
- versionName: 1.8.5
versionCode: 185
commit: 1.8.5
subdir: app
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 1.8.5
CurrentVersionCode: 185
| Categories:
- System
License: GPL-3.0-only
AuthorName: Dominik
AuthorWebSite: https://domi04151309.github.io/
WebSite: https://domi04151309.github.io/Android/Powerapp
SourceCode: https://github.com/Domi04151309/Power-App-for-Android
IssueTracker: https://github.com/Domi04151309/Power-App-for-Android/issues
AutoName: Power App
RequiresRoot: 'yes'
RepoType: git
Repo: https://github.com/Domi04151309/Power-App-for-Android
Builds:
- versionName: 1.8.3
versionCode: 183
commit: 1.8.3
subdir: app
gradle:
- yes
- versionName: 1.8.4
versionCode: 184
commit: 1.8.4
subdir: app
gradle:
- yes
+ - versionName: 1.8.5
+ versionCode: 185
+ commit: 1.8.5
+ subdir: app
+ gradle:
+ - yes
+
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
- CurrentVersion: 1.8.4
? ^
+ CurrentVersion: 1.8.5
? ^
- CurrentVersionCode: 184
? ^
+ CurrentVersionCode: 185
? ^
| 11 | 0.314286 | 9 | 2 |
917114cbb2d0c367c172bae4bacdfbab5afbdd2f | example/demo_tcp_server.cpp | example/demo_tcp_server.cpp |
int main(int argc, char** argv)
{
//initialize default port number and max connection cout
int port = 12001, max_connection_count= 1000;
// check if there are any passed arguments
if(argc > 1)
{
// initialize string stream from argument
std::istringstream arg_stream(argv[1]);
// bind arguments stream to port int variable if valid
if ( !(arg_stream >> port) )
std::cerr << "Invalid number " << argv[1] << '\n';
}
// create server instance with specified port number
TcpServer server(port);
//start listening to connections
server.Listen(max_connection_count);
return 0;
}
|
int main(int argc, char** argv)
{
//initialize default port number and max connection cout
int port = 12001, max_connection_count= 1000;
// check if there are any passed arguments
if(argc > 1)
{
// initialize string stream from argument
std::istringstream arg_stream(argv[1]);
// bind arguments stream to port int variable if valid
if ( !(arg_stream >> port) )
std::cerr << "Invalid number " << argv[1] << '\n';
}
// create server instance with specified port number
TcpServer server(port);
//start listening to connections
int result = server.Listen(max_connection_count);
if(result != 0)
cout << "Faild to start server !" << std::endl;
return 0;
}
| Update tcp server demo add error msg | Update tcp server demo add error msg
| C++ | mit | samehkamaleldin/socket.cpp | c++ | ## Code Before:
int main(int argc, char** argv)
{
//initialize default port number and max connection cout
int port = 12001, max_connection_count= 1000;
// check if there are any passed arguments
if(argc > 1)
{
// initialize string stream from argument
std::istringstream arg_stream(argv[1]);
// bind arguments stream to port int variable if valid
if ( !(arg_stream >> port) )
std::cerr << "Invalid number " << argv[1] << '\n';
}
// create server instance with specified port number
TcpServer server(port);
//start listening to connections
server.Listen(max_connection_count);
return 0;
}
## Instruction:
Update tcp server demo add error msg
## Code After:
int main(int argc, char** argv)
{
//initialize default port number and max connection cout
int port = 12001, max_connection_count= 1000;
// check if there are any passed arguments
if(argc > 1)
{
// initialize string stream from argument
std::istringstream arg_stream(argv[1]);
// bind arguments stream to port int variable if valid
if ( !(arg_stream >> port) )
std::cerr << "Invalid number " << argv[1] << '\n';
}
// create server instance with specified port number
TcpServer server(port);
//start listening to connections
int result = server.Listen(max_connection_count);
if(result != 0)
cout << "Faild to start server !" << std::endl;
return 0;
}
|
int main(int argc, char** argv)
{
//initialize default port number and max connection cout
int port = 12001, max_connection_count= 1000;
// check if there are any passed arguments
if(argc > 1)
{
// initialize string stream from argument
std::istringstream arg_stream(argv[1]);
// bind arguments stream to port int variable if valid
if ( !(arg_stream >> port) )
std::cerr << "Invalid number " << argv[1] << '\n';
}
// create server instance with specified port number
TcpServer server(port);
//start listening to connections
- server.Listen(max_connection_count);
+ int result = server.Listen(max_connection_count);
? +++++++++++++
+ if(result != 0)
+ cout << "Faild to start server !" << std::endl;
return 0;
} | 4 | 0.16 | 3 | 1 |
5119bb7c571e885790d1e71d7b57fe584f344563 | src/v2/Apps/Partner/Components/__tests__/NavigationTabs.jest.tsx | src/v2/Apps/Partner/Components/__tests__/NavigationTabs.jest.tsx | import { NavigationTabs_Test_PartnerQueryRawResponse } from "v2/__generated__/NavigationTabs_Test_PartnerQuery.graphql"
import { NavigationTabsFragmentContainer as NavigationTabs } from "v2/Apps/Partner/Components/NavigationTabs"
import { renderRelayTree } from "v2/DevTools"
import React from "react"
import { graphql } from "react-relay"
jest.unmock("react-relay")
jest.mock("v2/Components/RouteTabs")
describe("PartnerNavigationTabs", () => {
const getWrapper = async (
response: NavigationTabs_Test_PartnerQueryRawResponse["partner"] = NavigationTabsFixture
) => {
return await renderRelayTree({
Component: ({ partner }: any) => {
return <NavigationTabs partner={partner} />
},
query: graphql`
query NavigationTabs_Test_PartnerQuery @raw_response_type {
partner(id: "white-cube") {
...NavigationTabs_partner
}
}
`,
mockData: {
partner: response,
} as NavigationTabs_Test_PartnerQueryRawResponse,
})
}
it("renders all tabs by default", async () => {
const wrapper = await getWrapper()
const html = wrapper.html()
expect(html).toContain("Overview")
expect(html).toContain("Shows")
expect(html).toContain("Works")
expect(html).toContain("Artists")
expect(html).toContain("Articles")
expect(html).toContain("Contact")
})
})
const NavigationTabsFixture: NavigationTabs_Test_PartnerQueryRawResponse["partner"] = {
id: "white-cube",
slug: "white-cube",
}
| import React from "react"
import { graphql } from "react-relay"
import { setupTestWrapper } from "v2/DevTools/setupTestWrapper"
import { NavigationTabs_Test_PartnerQuery } from "v2/__generated__/NavigationTabs_Test_PartnerQuery.graphql"
import { NavigationTabsFragmentContainer as NavigationTabs } from "v2/Apps/Partner/Components/NavigationTabs"
jest.unmock("react-relay")
jest.mock("v2/Components/RouteTabs")
const { getWrapper } = setupTestWrapper<NavigationTabs_Test_PartnerQuery>({
Component: ({ partner }: any) => {
return <NavigationTabs partner={partner} />
},
query: graphql`
query NavigationTabs_Test_PartnerQuery @raw_response_type {
partner(id: "white-cube") {
...NavigationTabs_partner
}
}
`,
})
describe("PartnerNavigationTabs", () => {
it("renders all tabs by default", async () => {
const wrapper = await getWrapper({
Partner: () => ({ id: "white-cube", slug: "white-cube" }),
})
const html = wrapper.html()
expect(html).toContain("Overview")
expect(html).toContain("Shows")
expect(html).toContain("Works")
expect(html).toContain("Artists")
expect(html).toContain("Articles")
expect(html).toContain("Contact")
})
})
| Update navigation tabs tests to use setupTestWrapper | Update navigation tabs tests to use setupTestWrapper
| TypeScript | mit | joeyAghion/force,artsy/force,joeyAghion/force,artsy/force-public,artsy/force-public,joeyAghion/force,artsy/force,artsy/force,artsy/force,joeyAghion/force | typescript | ## Code Before:
import { NavigationTabs_Test_PartnerQueryRawResponse } from "v2/__generated__/NavigationTabs_Test_PartnerQuery.graphql"
import { NavigationTabsFragmentContainer as NavigationTabs } from "v2/Apps/Partner/Components/NavigationTabs"
import { renderRelayTree } from "v2/DevTools"
import React from "react"
import { graphql } from "react-relay"
jest.unmock("react-relay")
jest.mock("v2/Components/RouteTabs")
describe("PartnerNavigationTabs", () => {
const getWrapper = async (
response: NavigationTabs_Test_PartnerQueryRawResponse["partner"] = NavigationTabsFixture
) => {
return await renderRelayTree({
Component: ({ partner }: any) => {
return <NavigationTabs partner={partner} />
},
query: graphql`
query NavigationTabs_Test_PartnerQuery @raw_response_type {
partner(id: "white-cube") {
...NavigationTabs_partner
}
}
`,
mockData: {
partner: response,
} as NavigationTabs_Test_PartnerQueryRawResponse,
})
}
it("renders all tabs by default", async () => {
const wrapper = await getWrapper()
const html = wrapper.html()
expect(html).toContain("Overview")
expect(html).toContain("Shows")
expect(html).toContain("Works")
expect(html).toContain("Artists")
expect(html).toContain("Articles")
expect(html).toContain("Contact")
})
})
const NavigationTabsFixture: NavigationTabs_Test_PartnerQueryRawResponse["partner"] = {
id: "white-cube",
slug: "white-cube",
}
## Instruction:
Update navigation tabs tests to use setupTestWrapper
## Code After:
import React from "react"
import { graphql } from "react-relay"
import { setupTestWrapper } from "v2/DevTools/setupTestWrapper"
import { NavigationTabs_Test_PartnerQuery } from "v2/__generated__/NavigationTabs_Test_PartnerQuery.graphql"
import { NavigationTabsFragmentContainer as NavigationTabs } from "v2/Apps/Partner/Components/NavigationTabs"
jest.unmock("react-relay")
jest.mock("v2/Components/RouteTabs")
const { getWrapper } = setupTestWrapper<NavigationTabs_Test_PartnerQuery>({
Component: ({ partner }: any) => {
return <NavigationTabs partner={partner} />
},
query: graphql`
query NavigationTabs_Test_PartnerQuery @raw_response_type {
partner(id: "white-cube") {
...NavigationTabs_partner
}
}
`,
})
describe("PartnerNavigationTabs", () => {
it("renders all tabs by default", async () => {
const wrapper = await getWrapper({
Partner: () => ({ id: "white-cube", slug: "white-cube" }),
})
const html = wrapper.html()
expect(html).toContain("Overview")
expect(html).toContain("Shows")
expect(html).toContain("Works")
expect(html).toContain("Artists")
expect(html).toContain("Articles")
expect(html).toContain("Contact")
})
})
| - import { NavigationTabs_Test_PartnerQueryRawResponse } from "v2/__generated__/NavigationTabs_Test_PartnerQuery.graphql"
- import { NavigationTabsFragmentContainer as NavigationTabs } from "v2/Apps/Partner/Components/NavigationTabs"
- import { renderRelayTree } from "v2/DevTools"
import React from "react"
import { graphql } from "react-relay"
+ import { setupTestWrapper } from "v2/DevTools/setupTestWrapper"
+ import { NavigationTabs_Test_PartnerQuery } from "v2/__generated__/NavigationTabs_Test_PartnerQuery.graphql"
+ import { NavigationTabsFragmentContainer as NavigationTabs } from "v2/Apps/Partner/Components/NavigationTabs"
jest.unmock("react-relay")
jest.mock("v2/Components/RouteTabs")
+ const { getWrapper } = setupTestWrapper<NavigationTabs_Test_PartnerQuery>({
+ Component: ({ partner }: any) => {
+ return <NavigationTabs partner={partner} />
+ },
+ query: graphql`
+ query NavigationTabs_Test_PartnerQuery @raw_response_type {
+ partner(id: "white-cube") {
+ ...NavigationTabs_partner
+ }
+ }
+ `,
+ })
+
describe("PartnerNavigationTabs", () => {
+ it("renders all tabs by default", async () => {
+ const wrapper = await getWrapper({
+ Partner: () => ({ id: "white-cube", slug: "white-cube" }),
- const getWrapper = async (
- response: NavigationTabs_Test_PartnerQueryRawResponse["partner"] = NavigationTabsFixture
- ) => {
- return await renderRelayTree({
- Component: ({ partner }: any) => {
- return <NavigationTabs partner={partner} />
- },
- query: graphql`
- query NavigationTabs_Test_PartnerQuery @raw_response_type {
- partner(id: "white-cube") {
- ...NavigationTabs_partner
- }
- }
- `,
- mockData: {
- partner: response,
- } as NavigationTabs_Test_PartnerQueryRawResponse,
})
- }
-
- it("renders all tabs by default", async () => {
- const wrapper = await getWrapper()
const html = wrapper.html()
expect(html).toContain("Overview")
expect(html).toContain("Shows")
expect(html).toContain("Works")
expect(html).toContain("Artists")
expect(html).toContain("Articles")
expect(html).toContain("Contact")
})
})
-
- const NavigationTabsFixture: NavigationTabs_Test_PartnerQueryRawResponse["partner"] = {
- id: "white-cube",
- slug: "white-cube",
- } | 48 | 1.021277 | 19 | 29 |
df3415b87a4c5e528bfeda6d9267c35dbcb69e6f | living-with-django/style/main.css | living-with-django/style/main.css | body {
font-family: sans-serif;
font-size: 12px;
line-height: 1.6em;
color: #000;
margin: 0;
}
.container {
max-width: 1000px;
padding: 30px;
margin: auto;
}
small {
font-size: .7em;
font-weight: normal;
color: #666;
}
pre {
padding-left: 20px;
border-left: solid #888 2px;
}
p code {
padding: 2px;
border-radius: 5px;
background-color: #eee;
}
.entry {
margin-bottom: 80px;
}
h1.big {
color: #888;
margin-bottom: 80px;
font-size: 3em;
}
h1 a {
color: inherit;
text-decoration: inherit;
}
h1 a:hover:after {
content: ' \2190'
}
| body {
font-family: sans-serif;
font-size: 12px;
line-height: 1.6em;
color: #000;
margin: 0;
}
.container {
max-width: 1000px;
padding: 30px;
margin: auto;
}
small {
font-size: .7em;
font-weight: normal;
color: #666;
}
pre {
padding-left: 20px;
border-left: solid #888 2px;
font-size: 1em;
}
p code {
padding: 2px;
border-radius: 5px;
background-color: #eee;
}
.entry {
margin-bottom: 80px;
}
h1.big {
color: #888;
margin-bottom: 80px;
font-size: 3em;
}
h1 a {
color: inherit;
text-decoration: inherit;
}
h1 a:hover:after {
content: ' \2190'
}
| Enforce font size in <pre> blocks. | Enforce font size in <pre> blocks.
| CSS | mit | astex/living-with-django,astex/living-with-django,astex/living-with-django | css | ## Code Before:
body {
font-family: sans-serif;
font-size: 12px;
line-height: 1.6em;
color: #000;
margin: 0;
}
.container {
max-width: 1000px;
padding: 30px;
margin: auto;
}
small {
font-size: .7em;
font-weight: normal;
color: #666;
}
pre {
padding-left: 20px;
border-left: solid #888 2px;
}
p code {
padding: 2px;
border-radius: 5px;
background-color: #eee;
}
.entry {
margin-bottom: 80px;
}
h1.big {
color: #888;
margin-bottom: 80px;
font-size: 3em;
}
h1 a {
color: inherit;
text-decoration: inherit;
}
h1 a:hover:after {
content: ' \2190'
}
## Instruction:
Enforce font size in <pre> blocks.
## Code After:
body {
font-family: sans-serif;
font-size: 12px;
line-height: 1.6em;
color: #000;
margin: 0;
}
.container {
max-width: 1000px;
padding: 30px;
margin: auto;
}
small {
font-size: .7em;
font-weight: normal;
color: #666;
}
pre {
padding-left: 20px;
border-left: solid #888 2px;
font-size: 1em;
}
p code {
padding: 2px;
border-radius: 5px;
background-color: #eee;
}
.entry {
margin-bottom: 80px;
}
h1.big {
color: #888;
margin-bottom: 80px;
font-size: 3em;
}
h1 a {
color: inherit;
text-decoration: inherit;
}
h1 a:hover:after {
content: ' \2190'
}
| body {
font-family: sans-serif;
font-size: 12px;
line-height: 1.6em;
color: #000;
margin: 0;
}
.container {
max-width: 1000px;
padding: 30px;
margin: auto;
}
small {
font-size: .7em;
font-weight: normal;
color: #666;
}
pre {
padding-left: 20px;
border-left: solid #888 2px;
+ font-size: 1em;
}
p code {
padding: 2px;
border-radius: 5px;
background-color: #eee;
}
.entry {
margin-bottom: 80px;
}
h1.big {
color: #888;
margin-bottom: 80px;
font-size: 3em;
}
h1 a {
color: inherit;
text-decoration: inherit;
}
h1 a:hover:after {
content: ' \2190'
} | 1 | 0.020408 | 1 | 0 |
c10afc4ebd4d7ec8571c0685c0d87f76b25b3af9 | scipy/special/_precompute/utils.py | scipy/special/_precompute/utils.py | try:
import mpmath as mp
except ImportError:
pass
try:
from sympy.abc import x # type: ignore[import]
except ImportError:
pass
def lagrange_inversion(a):
"""Given a series
f(x) = a[1]*x + a[2]*x**2 + ... + a[n-1]*x**(n - 1),
use the Lagrange inversion formula to compute a series
g(x) = b[1]*x + b[2]*x**2 + ... + b[n-1]*x**(n - 1)
so that f(g(x)) = g(f(x)) = x mod x**n. We must have a[0] = 0, so
necessarily b[0] = 0 too.
The algorithm is naive and could be improved, but speed isn't an
issue here and it's easy to read.
"""
n = len(a)
f = sum(a[i]*x**i for i in range(len(a)))
h = (x/f).series(x, 0, n).removeO()
hpower = [h**0]
for k in range(n):
hpower.append((hpower[-1]*h).expand())
b = [mp.mpf(0)]
for k in range(1, n):
b.append(hpower[k].coeff(x, k - 1)/k)
b = map(lambda x: mp.mpf(x), b)
return b
| try:
import mpmath as mp
except ImportError:
pass
try:
from sympy.abc import x # type: ignore[import]
except ImportError:
pass
def lagrange_inversion(a):
"""Given a series
f(x) = a[1]*x + a[2]*x**2 + ... + a[n-1]*x**(n - 1),
use the Lagrange inversion formula to compute a series
g(x) = b[1]*x + b[2]*x**2 + ... + b[n-1]*x**(n - 1)
so that f(g(x)) = g(f(x)) = x mod x**n. We must have a[0] = 0, so
necessarily b[0] = 0 too.
The algorithm is naive and could be improved, but speed isn't an
issue here and it's easy to read.
"""
n = len(a)
f = sum(a[i]*x**i for i in range(len(a)))
h = (x/f).series(x, 0, n).removeO()
hpower = [h**0]
for k in range(n):
hpower.append((hpower[-1]*h).expand())
b = [mp.mpf(0)]
for k in range(1, n):
b.append(hpower[k].coeff(x, k - 1)/k)
b = [mp.mpf(x) for x in b]
return b
| Use list comprehension instead of lambda function | Use list comprehension instead of lambda function
| Python | bsd-3-clause | grlee77/scipy,WarrenWeckesser/scipy,vigna/scipy,endolith/scipy,andyfaff/scipy,rgommers/scipy,scipy/scipy,grlee77/scipy,mdhaber/scipy,Stefan-Endres/scipy,zerothi/scipy,rgommers/scipy,andyfaff/scipy,scipy/scipy,zerothi/scipy,tylerjereddy/scipy,endolith/scipy,mdhaber/scipy,endolith/scipy,rgommers/scipy,mdhaber/scipy,endolith/scipy,e-q/scipy,WarrenWeckesser/scipy,ilayn/scipy,matthew-brett/scipy,anntzer/scipy,scipy/scipy,tylerjereddy/scipy,Eric89GXL/scipy,vigna/scipy,WarrenWeckesser/scipy,perimosocordiae/scipy,andyfaff/scipy,perimosocordiae/scipy,e-q/scipy,andyfaff/scipy,WarrenWeckesser/scipy,tylerjereddy/scipy,grlee77/scipy,tylerjereddy/scipy,anntzer/scipy,matthew-brett/scipy,tylerjereddy/scipy,rgommers/scipy,WarrenWeckesser/scipy,vigna/scipy,matthew-brett/scipy,zerothi/scipy,Stefan-Endres/scipy,endolith/scipy,vigna/scipy,ilayn/scipy,ilayn/scipy,zerothi/scipy,mdhaber/scipy,perimosocordiae/scipy,WarrenWeckesser/scipy,anntzer/scipy,zerothi/scipy,perimosocordiae/scipy,zerothi/scipy,andyfaff/scipy,scipy/scipy,vigna/scipy,anntzer/scipy,e-q/scipy,scipy/scipy,perimosocordiae/scipy,matthew-brett/scipy,Eric89GXL/scipy,grlee77/scipy,mdhaber/scipy,anntzer/scipy,ilayn/scipy,ilayn/scipy,matthew-brett/scipy,anntzer/scipy,grlee77/scipy,Stefan-Endres/scipy,e-q/scipy,ilayn/scipy,scipy/scipy,Eric89GXL/scipy,mdhaber/scipy,endolith/scipy,Eric89GXL/scipy,rgommers/scipy,Eric89GXL/scipy,andyfaff/scipy,Eric89GXL/scipy,Stefan-Endres/scipy,Stefan-Endres/scipy,Stefan-Endres/scipy,perimosocordiae/scipy,e-q/scipy | python | ## Code Before:
try:
import mpmath as mp
except ImportError:
pass
try:
from sympy.abc import x # type: ignore[import]
except ImportError:
pass
def lagrange_inversion(a):
"""Given a series
f(x) = a[1]*x + a[2]*x**2 + ... + a[n-1]*x**(n - 1),
use the Lagrange inversion formula to compute a series
g(x) = b[1]*x + b[2]*x**2 + ... + b[n-1]*x**(n - 1)
so that f(g(x)) = g(f(x)) = x mod x**n. We must have a[0] = 0, so
necessarily b[0] = 0 too.
The algorithm is naive and could be improved, but speed isn't an
issue here and it's easy to read.
"""
n = len(a)
f = sum(a[i]*x**i for i in range(len(a)))
h = (x/f).series(x, 0, n).removeO()
hpower = [h**0]
for k in range(n):
hpower.append((hpower[-1]*h).expand())
b = [mp.mpf(0)]
for k in range(1, n):
b.append(hpower[k].coeff(x, k - 1)/k)
b = map(lambda x: mp.mpf(x), b)
return b
## Instruction:
Use list comprehension instead of lambda function
## Code After:
try:
import mpmath as mp
except ImportError:
pass
try:
from sympy.abc import x # type: ignore[import]
except ImportError:
pass
def lagrange_inversion(a):
"""Given a series
f(x) = a[1]*x + a[2]*x**2 + ... + a[n-1]*x**(n - 1),
use the Lagrange inversion formula to compute a series
g(x) = b[1]*x + b[2]*x**2 + ... + b[n-1]*x**(n - 1)
so that f(g(x)) = g(f(x)) = x mod x**n. We must have a[0] = 0, so
necessarily b[0] = 0 too.
The algorithm is naive and could be improved, but speed isn't an
issue here and it's easy to read.
"""
n = len(a)
f = sum(a[i]*x**i for i in range(len(a)))
h = (x/f).series(x, 0, n).removeO()
hpower = [h**0]
for k in range(n):
hpower.append((hpower[-1]*h).expand())
b = [mp.mpf(0)]
for k in range(1, n):
b.append(hpower[k].coeff(x, k - 1)/k)
b = [mp.mpf(x) for x in b]
return b
| try:
import mpmath as mp
except ImportError:
pass
try:
from sympy.abc import x # type: ignore[import]
except ImportError:
pass
def lagrange_inversion(a):
"""Given a series
f(x) = a[1]*x + a[2]*x**2 + ... + a[n-1]*x**(n - 1),
use the Lagrange inversion formula to compute a series
g(x) = b[1]*x + b[2]*x**2 + ... + b[n-1]*x**(n - 1)
so that f(g(x)) = g(f(x)) = x mod x**n. We must have a[0] = 0, so
necessarily b[0] = 0 too.
The algorithm is naive and could be improved, but speed isn't an
issue here and it's easy to read.
"""
n = len(a)
f = sum(a[i]*x**i for i in range(len(a)))
h = (x/f).series(x, 0, n).removeO()
hpower = [h**0]
for k in range(n):
hpower.append((hpower[-1]*h).expand())
b = [mp.mpf(0)]
for k in range(1, n):
b.append(hpower[k].coeff(x, k - 1)/k)
- b = map(lambda x: mp.mpf(x), b)
+ b = [mp.mpf(x) for x in b]
return b | 2 | 0.052632 | 1 | 1 |
45bd83b043027b8b514be1195c403868555a9968 | test/test.js | test/test.js | var mercurius = require('../index.js');
var request = require('supertest');
var assert = require('assert');
describe('mercurius', function() {
before(function() {
return mercurius.ready;
});
it('successfully registers users', function(done) {
request(mercurius.app)
.post('/register')
.send({
endpoint: 'endpoint',
key: 'key'
})
.expect(function(res) {
assert.equal(res.status, 200);
assert.equal(res.text.length, 64);
})
.end(done);
});
});
| var mercurius = require('../index.js');
var request = require('supertest');
var assert = require('assert');
describe('mercurius', function() {
before(function() {
return mercurius.ready;
});
it('successfully registers users', function(done) {
request(mercurius.app)
.post('/register')
.send({
endpoint: 'endpoint',
key: 'key'
})
.expect(function(res) {
assert.equal(res.status, 200);
assert.equal(res.text.length, 64);
})
.end(done);
});
it('sends 404 when a registration doesn\'t exist', function(done) {
request(mercurius.app)
.post('/notify')
.send({
token: 'token_inesistente',
})
.expect(404, done);
});
});
| Test sending a notification with a token that doesn\'t exist | Test sending a notification with a token that doesn\'t exist
| JavaScript | apache-2.0 | marco-c/mercurius,marco-c/mercurius,marco-c/mercurius,zalun/mercurius,zalun/mercurius | javascript | ## Code Before:
var mercurius = require('../index.js');
var request = require('supertest');
var assert = require('assert');
describe('mercurius', function() {
before(function() {
return mercurius.ready;
});
it('successfully registers users', function(done) {
request(mercurius.app)
.post('/register')
.send({
endpoint: 'endpoint',
key: 'key'
})
.expect(function(res) {
assert.equal(res.status, 200);
assert.equal(res.text.length, 64);
})
.end(done);
});
});
## Instruction:
Test sending a notification with a token that doesn\'t exist
## Code After:
var mercurius = require('../index.js');
var request = require('supertest');
var assert = require('assert');
describe('mercurius', function() {
before(function() {
return mercurius.ready;
});
it('successfully registers users', function(done) {
request(mercurius.app)
.post('/register')
.send({
endpoint: 'endpoint',
key: 'key'
})
.expect(function(res) {
assert.equal(res.status, 200);
assert.equal(res.text.length, 64);
})
.end(done);
});
it('sends 404 when a registration doesn\'t exist', function(done) {
request(mercurius.app)
.post('/notify')
.send({
token: 'token_inesistente',
})
.expect(404, done);
});
});
| var mercurius = require('../index.js');
var request = require('supertest');
var assert = require('assert');
describe('mercurius', function() {
before(function() {
return mercurius.ready;
});
it('successfully registers users', function(done) {
request(mercurius.app)
.post('/register')
.send({
endpoint: 'endpoint',
key: 'key'
})
.expect(function(res) {
assert.equal(res.status, 200);
assert.equal(res.text.length, 64);
})
.end(done);
});
+
+ it('sends 404 when a registration doesn\'t exist', function(done) {
+ request(mercurius.app)
+ .post('/notify')
+ .send({
+ token: 'token_inesistente',
+ })
+ .expect(404, done);
+ });
}); | 9 | 0.391304 | 9 | 0 |
644ec4e541f1db7f8b229148e5d19c495a5ff4c5 | ode/public/js/browser.js | ode/public/js/browser.js | $(document).ready(function() {
$('.rule-item').on('mouseenter', function() {
$(this).addClass('highlighted');
var controls = $('<span>').addClass('pull-right controls');
controls.append($.editButton($(this).attr('id')));
controls.append($.removeButton($(this).attr('id')));
$(this).find('h2').append(controls);
$('.edit-button').on('click', function(e) {
var ruleID = $(e.currentTarget).parents('.rule-item').attr('id');
window.location.href =
document.URL + '/' + ruleID + '/input';
});
});
$('.rule-item').on('mouseleave', function() {
$(this).removeClass('highlighted');
$(this).find('.controls').remove();
});
});
| var Rule = Backbone.Model.extend({
initialize: function() {
this.urlRoot = '/rules';
},
});
$(document).ready(function() {
$('.rule-item').on('mouseenter', function() {
$(this).addClass('highlighted');
var controls = $('<span>').addClass('pull-right controls');
controls.append($.editButton($(this).attr('id')));
controls.append($.removeButton($(this).attr('id')));
$(this).find('h2').append(controls);
$('.edit-button').on('click', function(e) {
var ruleID = $(e.currentTarget).parents('.rule-item').attr('id');
window.location.href =
document.URL + '/' + ruleID + '/input';
});
$('.remove-button').on('click', function(e) {
var ruleID = $(e.currentTarget).parents('.rule-item').attr('id')
.substring(1);
var rule = new Rule({ id: ruleID });
rule.destroy();
});
});
$('.rule-item').on('mouseleave', function() {
$(this).removeClass('highlighted');
$(this).find('.controls').remove();
});
});
| Delete feature when clicking its `.remove-button` in Rule Browser. | Delete feature when clicking its `.remove-button` in Rule Browser.
| JavaScript | agpl-3.0 | itsjeyd/ODE,itsjeyd/ODE,itsjeyd/ODE,itsjeyd/ODE | javascript | ## Code Before:
$(document).ready(function() {
$('.rule-item').on('mouseenter', function() {
$(this).addClass('highlighted');
var controls = $('<span>').addClass('pull-right controls');
controls.append($.editButton($(this).attr('id')));
controls.append($.removeButton($(this).attr('id')));
$(this).find('h2').append(controls);
$('.edit-button').on('click', function(e) {
var ruleID = $(e.currentTarget).parents('.rule-item').attr('id');
window.location.href =
document.URL + '/' + ruleID + '/input';
});
});
$('.rule-item').on('mouseleave', function() {
$(this).removeClass('highlighted');
$(this).find('.controls').remove();
});
});
## Instruction:
Delete feature when clicking its `.remove-button` in Rule Browser.
## Code After:
var Rule = Backbone.Model.extend({
initialize: function() {
this.urlRoot = '/rules';
},
});
$(document).ready(function() {
$('.rule-item').on('mouseenter', function() {
$(this).addClass('highlighted');
var controls = $('<span>').addClass('pull-right controls');
controls.append($.editButton($(this).attr('id')));
controls.append($.removeButton($(this).attr('id')));
$(this).find('h2').append(controls);
$('.edit-button').on('click', function(e) {
var ruleID = $(e.currentTarget).parents('.rule-item').attr('id');
window.location.href =
document.URL + '/' + ruleID + '/input';
});
$('.remove-button').on('click', function(e) {
var ruleID = $(e.currentTarget).parents('.rule-item').attr('id')
.substring(1);
var rule = new Rule({ id: ruleID });
rule.destroy();
});
});
$('.rule-item').on('mouseleave', function() {
$(this).removeClass('highlighted');
$(this).find('.controls').remove();
});
});
| + var Rule = Backbone.Model.extend({
+
+ initialize: function() {
+ this.urlRoot = '/rules';
+ },
+
+ });
+
$(document).ready(function() {
$('.rule-item').on('mouseenter', function() {
$(this).addClass('highlighted');
var controls = $('<span>').addClass('pull-right controls');
controls.append($.editButton($(this).attr('id')));
controls.append($.removeButton($(this).attr('id')));
$(this).find('h2').append(controls);
$('.edit-button').on('click', function(e) {
var ruleID = $(e.currentTarget).parents('.rule-item').attr('id');
window.location.href =
document.URL + '/' + ruleID + '/input';
});
-
+ $('.remove-button').on('click', function(e) {
+ var ruleID = $(e.currentTarget).parents('.rule-item').attr('id')
+ .substring(1);
+ var rule = new Rule({ id: ruleID });
+ rule.destroy();
+ });
});
$('.rule-item').on('mouseleave', function() {
$(this).removeClass('highlighted');
$(this).find('.controls').remove();
});
}); | 15 | 0.681818 | 14 | 1 |
1cd61a5ba5abaec39a4e8f28bbb1be4f161d6756 | appveyor.yml | appveyor.yml | version: 1.0.{build}
image: Visual Studio 2017 Preview
build_script:
- cmd: >-
nuget restore
msbuild /m /clp:v=m Codex.sln /bl /logger:"C:\Program Files\AppVeyor\BuildAgent\Appveyor.MSBuildLogger.dll" | version: 1.0.{build}
image: Visual Studio 2017 Preview
build_script:
- cmd: >-
nuget restore
msbuild /m /clp:v=m Codex.sln /bl /logger:"C:\Program Files\AppVeyor\BuildAgent\Appveyor.MSBuildLogger.dll"
artifacts:
- path: msbuild.binlog
name: MSBuild Log | Save msbuild.binlog as an artifact. | Save msbuild.binlog as an artifact.
| YAML | mit | Ref12/Codex,Ref12/Codex,Ref12/Codex,Ref12/Codex,Ref12/Codex,Ref12/Codex | yaml | ## Code Before:
version: 1.0.{build}
image: Visual Studio 2017 Preview
build_script:
- cmd: >-
nuget restore
msbuild /m /clp:v=m Codex.sln /bl /logger:"C:\Program Files\AppVeyor\BuildAgent\Appveyor.MSBuildLogger.dll"
## Instruction:
Save msbuild.binlog as an artifact.
## Code After:
version: 1.0.{build}
image: Visual Studio 2017 Preview
build_script:
- cmd: >-
nuget restore
msbuild /m /clp:v=m Codex.sln /bl /logger:"C:\Program Files\AppVeyor\BuildAgent\Appveyor.MSBuildLogger.dll"
artifacts:
- path: msbuild.binlog
name: MSBuild Log | version: 1.0.{build}
image: Visual Studio 2017 Preview
build_script:
- cmd: >-
nuget restore
msbuild /m /clp:v=m Codex.sln /bl /logger:"C:\Program Files\AppVeyor\BuildAgent\Appveyor.MSBuildLogger.dll"
+ artifacts:
+ - path: msbuild.binlog
+ name: MSBuild Log | 3 | 0.428571 | 3 | 0 |
a5f9c4faa9d94ca598dcf8283a63bb772b7a4848 | packages/front-end/src/index.html | packages/front-end/src/index.html | <!doctype>
<html>
<head>
<script src="scripts/easel/easeljs-0.8.1.combined.js"></script>
<script src="scripts/dc-shell.js"></script>
<link rel="import" href="components/dc-lobby/dc-lobby.html">
<link rel="import" href="components/dc-login.html">
<link rel="import" href="components/dc-game-canvas/dc-game-canvas.html">
<link rel="stylesheet" href="style/base.css">
<script>
(function() {
"use strict";
window.fbAsyncInit = function() {
FB.init({
appId : '409903325878723',
xfbml : false,
version : 'v2.4',
status : true
});
DcShell.afterFbInit();
};
(function(d, s, id){
var js, fjs = d.getElementsByTagName(s)[0];
if (d.getElementById(id)) {return;}
js = d.createElement(s); js.id = id;
js.src = "//connect.facebook.net/en_US/sdk.js";
fjs.parentNode.insertBefore(js, fjs);
}(document, 'script', 'facebook-jssdk'));
}());
</script>
</head>
<body>
<div id="content"></div>
</body>
</html> | <!doctype>
<html>
<head>
<script src="scripts/easel/easeljs-0.8.1.combined.js"></script>
<script src="../../scripts/tween/tweenjs-0.6.1.combined.js"></script>
<script src="../../scripts/preload/preloadjs-0.6.1.combined.js"></script>
<script src="scripts/dc-shell.js"></script>
<link rel="import" href="components/dc-lobby/dc-lobby.html">
<link rel="import" href="components/dc-login.html">
<link rel="import" href="components/dc-game-canvas/dc-game-canvas.html">
<link rel="stylesheet" href="style/base.css">
<script>
(function() {
"use strict";
window.fbAsyncInit = function() {
FB.init({
appId : '409903325878723',
xfbml : false,
version : 'v2.4',
status : true
});
DcShell.afterFbInit();
};
(function(d, s, id){
var js, fjs = d.getElementsByTagName(s)[0];
if (d.getElementById(id)) {return;}
js = d.createElement(s); js.id = id;
js.src = "//connect.facebook.net/en_US/sdk.js";
fjs.parentNode.insertBefore(js, fjs);
}(document, 'script', 'facebook-jssdk'));
}());
</script>
</head>
<body>
<div id="content"></div>
</body>
</html> | Add preload and tween references to main page | Add preload and tween references to main page
| HTML | mit | ryb73/dealers-choice-meta,ryb73/dealers-choice-meta,ryb73/dealers-choice-meta | html | ## Code Before:
<!doctype>
<html>
<head>
<script src="scripts/easel/easeljs-0.8.1.combined.js"></script>
<script src="scripts/dc-shell.js"></script>
<link rel="import" href="components/dc-lobby/dc-lobby.html">
<link rel="import" href="components/dc-login.html">
<link rel="import" href="components/dc-game-canvas/dc-game-canvas.html">
<link rel="stylesheet" href="style/base.css">
<script>
(function() {
"use strict";
window.fbAsyncInit = function() {
FB.init({
appId : '409903325878723',
xfbml : false,
version : 'v2.4',
status : true
});
DcShell.afterFbInit();
};
(function(d, s, id){
var js, fjs = d.getElementsByTagName(s)[0];
if (d.getElementById(id)) {return;}
js = d.createElement(s); js.id = id;
js.src = "//connect.facebook.net/en_US/sdk.js";
fjs.parentNode.insertBefore(js, fjs);
}(document, 'script', 'facebook-jssdk'));
}());
</script>
</head>
<body>
<div id="content"></div>
</body>
</html>
## Instruction:
Add preload and tween references to main page
## Code After:
<!doctype>
<html>
<head>
<script src="scripts/easel/easeljs-0.8.1.combined.js"></script>
<script src="../../scripts/tween/tweenjs-0.6.1.combined.js"></script>
<script src="../../scripts/preload/preloadjs-0.6.1.combined.js"></script>
<script src="scripts/dc-shell.js"></script>
<link rel="import" href="components/dc-lobby/dc-lobby.html">
<link rel="import" href="components/dc-login.html">
<link rel="import" href="components/dc-game-canvas/dc-game-canvas.html">
<link rel="stylesheet" href="style/base.css">
<script>
(function() {
"use strict";
window.fbAsyncInit = function() {
FB.init({
appId : '409903325878723',
xfbml : false,
version : 'v2.4',
status : true
});
DcShell.afterFbInit();
};
(function(d, s, id){
var js, fjs = d.getElementsByTagName(s)[0];
if (d.getElementById(id)) {return;}
js = d.createElement(s); js.id = id;
js.src = "//connect.facebook.net/en_US/sdk.js";
fjs.parentNode.insertBefore(js, fjs);
}(document, 'script', 'facebook-jssdk'));
}());
</script>
</head>
<body>
<div id="content"></div>
</body>
</html> | <!doctype>
<html>
<head>
<script src="scripts/easel/easeljs-0.8.1.combined.js"></script>
+ <script src="../../scripts/tween/tweenjs-0.6.1.combined.js"></script>
+ <script src="../../scripts/preload/preloadjs-0.6.1.combined.js"></script>
<script src="scripts/dc-shell.js"></script>
<link rel="import" href="components/dc-lobby/dc-lobby.html">
<link rel="import" href="components/dc-login.html">
<link rel="import" href="components/dc-game-canvas/dc-game-canvas.html">
<link rel="stylesheet" href="style/base.css">
<script>
(function() {
"use strict";
window.fbAsyncInit = function() {
FB.init({
appId : '409903325878723',
xfbml : false,
version : 'v2.4',
status : true
});
DcShell.afterFbInit();
};
(function(d, s, id){
var js, fjs = d.getElementsByTagName(s)[0];
if (d.getElementById(id)) {return;}
js = d.createElement(s); js.id = id;
js.src = "//connect.facebook.net/en_US/sdk.js";
fjs.parentNode.insertBefore(js, fjs);
}(document, 'script', 'facebook-jssdk'));
}());
</script>
</head>
<body>
<div id="content"></div>
</body>
</html> | 2 | 0.04878 | 2 | 0 |
56a185af5767177d410113a03bd765135e07c9ca | .github/workflows/coq-ci.yml | .github/workflows/coq-ci.yml | name: CI
on: [push, pull_request]
jobs:
build:
runs-on: ubuntu-latest
strategy:
matrix:
image:
- 'coqorg/coq:dev'
- 'coqorg/coq:8.15'
- 'coqorg/coq:8.14'
fail-fast: false
steps:
- uses: actions/checkout@v2
- uses: coq-community/docker-coq-action@v1
with:
opam_file: 'coq-haskell.opam'
custom_image: ${{ matrix.image }}
# See also:
# https://github.com/coq-community/docker-coq-action#readme
# https://github.com/erikmd/docker-coq-github-action-demo
| name: CI
on: [push, pull_request]
jobs:
build:
runs-on: ubuntu-latest
strategy:
matrix:
image:
- 'coqorg/coq:dev'
- 'coqorg/coq:8.16'
- 'coqorg/coq:8.15'
- 'coqorg/coq:8.14'
fail-fast: false
steps:
- uses: actions/checkout@v2
- uses: coq-community/docker-coq-action@v1
with:
opam_file: 'coq-haskell.opam'
custom_image: ${{ matrix.image }}
# See also:
# https://github.com/coq-community/docker-coq-action#readme
# https://github.com/erikmd/docker-coq-github-action-demo
| Add 8.16 to CI file | Add 8.16 to CI file
| YAML | bsd-3-clause | jwiegley/coq-haskell,jwiegley/coq-haskell | yaml | ## Code Before:
name: CI
on: [push, pull_request]
jobs:
build:
runs-on: ubuntu-latest
strategy:
matrix:
image:
- 'coqorg/coq:dev'
- 'coqorg/coq:8.15'
- 'coqorg/coq:8.14'
fail-fast: false
steps:
- uses: actions/checkout@v2
- uses: coq-community/docker-coq-action@v1
with:
opam_file: 'coq-haskell.opam'
custom_image: ${{ matrix.image }}
# See also:
# https://github.com/coq-community/docker-coq-action#readme
# https://github.com/erikmd/docker-coq-github-action-demo
## Instruction:
Add 8.16 to CI file
## Code After:
name: CI
on: [push, pull_request]
jobs:
build:
runs-on: ubuntu-latest
strategy:
matrix:
image:
- 'coqorg/coq:dev'
- 'coqorg/coq:8.16'
- 'coqorg/coq:8.15'
- 'coqorg/coq:8.14'
fail-fast: false
steps:
- uses: actions/checkout@v2
- uses: coq-community/docker-coq-action@v1
with:
opam_file: 'coq-haskell.opam'
custom_image: ${{ matrix.image }}
# See also:
# https://github.com/coq-community/docker-coq-action#readme
# https://github.com/erikmd/docker-coq-github-action-demo
| name: CI
on: [push, pull_request]
jobs:
build:
runs-on: ubuntu-latest
strategy:
matrix:
image:
- 'coqorg/coq:dev'
+ - 'coqorg/coq:8.16'
- 'coqorg/coq:8.15'
- 'coqorg/coq:8.14'
fail-fast: false
steps:
- uses: actions/checkout@v2
- uses: coq-community/docker-coq-action@v1
with:
opam_file: 'coq-haskell.opam'
custom_image: ${{ matrix.image }}
# See also:
# https://github.com/coq-community/docker-coq-action#readme
# https://github.com/erikmd/docker-coq-github-action-demo | 1 | 0.041667 | 1 | 0 |
3a48e7b5debaf566a6a27a59624a3f79b437df25 | ios/ios_src/Helpers/DeviceHelpers.mm | ios/ios_src/Helpers/DeviceHelpers.mm | // Copyright eeGeo Ltd (2012-2015), All Rights Reserved
#include "DeviceHelpers.h"
#include <algorithm>
namespace ExampleApp
{
namespace Helpers
{
int deviceFramerateWhitelist[] = {
GAME_DEVICE_IPHONE8_PLUS,
GAME_DEVICE_IPHONE8,
GAME_DEVICE_IPADPRO_12_9_INCH,
GAME_DEVICE_IPADPRO_9_7_INCH,
GAME_DEVICE_IPHONE_SE,
GAME_DEVICE_IPHONE7_PLUS,
GAME_DEVICE_IPHONE7,
GAME_DEVICE_IPHONE6S_PLUS,
GAME_DEVICE_IPHONE6S,
GAME_DEVICE_IPHONE6PLUS,
GAME_DEVICE_IPHONE6,
GAME_DEVICE_IPHONE5S,
GAME_DEVICE_IPOD6GEN,
GAME_DEVICE_IPADAIR_2,
GAME_DEVICE_IPADMINI_3,
GAME_DEVICE_IPADMINI_4
};
namespace DeviceHelpers
{
NSInteger GetPreferredFramerate(const GameDeviceType& deviceType)
{
bool isWhitelisted = std::find(std::begin(deviceFramerateWhitelist), std::end(deviceFramerateWhitelist), deviceType) != std::end(deviceFramerateWhitelist);
return isWhitelisted ? (NSInteger)60 : (NSInteger)30;
}
}
}
}
| // Copyright eeGeo Ltd (2012-2015), All Rights Reserved
#include "DeviceHelpers.h"
#include <algorithm>
namespace ExampleApp
{
namespace Helpers
{
int deviceFramerateBlacklist[] = {
GAME_DEVICE_IPHONE1,
GAME_DEVICE_IPHONE3G,
GAME_DEVICE_IPHONE4,
GAME_DEVICE_IPHONE4S,
GAME_DEVICE_IPHONE5,
GAME_DEVICE_IPOD1GEN,
GAME_DEVICE_IPOD2GEN,
GAME_DEVICE_IPOD3GEN,
GAME_DEVICE_IPOD4GEN,
GAME_DEVICE_IPOD5GEN,
GAME_DEVICE_IPAD1,
GAME_DEVICE_IPAD2,
GAME_DEVICE_IPAD3,
GAME_DEVICE_IPAD4,
GAME_DEVICE_IPADAIR,
GAME_DEVICE_IPADMINI_1GEN,
GAME_DEVICE_IPADMINI_2
};
namespace DeviceHelpers
{
NSInteger GetPreferredFramerate(const GameDeviceType& deviceType)
{
bool isBlacklisted = std::find(std::begin(deviceFramerateBlacklist), std::end(deviceFramerateBlacklist), deviceType) != std::end(deviceFramerateBlacklist);
return isBlacklisted ? (NSInteger)30 : (NSInteger)60;
}
}
}
}
| Fix for [MPLY-9790]; iOS 60FPS Whitelist is now a 30FPS Blacklist. New devices will no longer need to be added to the list to get the benefit of 60FPS. Buddy: Jonty | Fix for [MPLY-9790]; iOS 60FPS Whitelist is now a 30FPS Blacklist. New devices will no longer need to be added to the list to get the benefit of 60FPS.
Buddy: Jonty
| Objective-C++ | bsd-2-clause | wrld3d/wrld-example-app,eegeo/eegeo-example-app,wrld3d/wrld-example-app,eegeo/eegeo-example-app,wrld3d/wrld-example-app,wrld3d/wrld-example-app,wrld3d/wrld-example-app,eegeo/eegeo-example-app,eegeo/eegeo-example-app,eegeo/eegeo-example-app,wrld3d/wrld-example-app,eegeo/eegeo-example-app,wrld3d/wrld-example-app,eegeo/eegeo-example-app,wrld3d/wrld-example-app | objective-c++ | ## Code Before:
// Copyright eeGeo Ltd (2012-2015), All Rights Reserved
#include "DeviceHelpers.h"
#include <algorithm>
namespace ExampleApp
{
namespace Helpers
{
int deviceFramerateWhitelist[] = {
GAME_DEVICE_IPHONE8_PLUS,
GAME_DEVICE_IPHONE8,
GAME_DEVICE_IPADPRO_12_9_INCH,
GAME_DEVICE_IPADPRO_9_7_INCH,
GAME_DEVICE_IPHONE_SE,
GAME_DEVICE_IPHONE7_PLUS,
GAME_DEVICE_IPHONE7,
GAME_DEVICE_IPHONE6S_PLUS,
GAME_DEVICE_IPHONE6S,
GAME_DEVICE_IPHONE6PLUS,
GAME_DEVICE_IPHONE6,
GAME_DEVICE_IPHONE5S,
GAME_DEVICE_IPOD6GEN,
GAME_DEVICE_IPADAIR_2,
GAME_DEVICE_IPADMINI_3,
GAME_DEVICE_IPADMINI_4
};
namespace DeviceHelpers
{
NSInteger GetPreferredFramerate(const GameDeviceType& deviceType)
{
bool isWhitelisted = std::find(std::begin(deviceFramerateWhitelist), std::end(deviceFramerateWhitelist), deviceType) != std::end(deviceFramerateWhitelist);
return isWhitelisted ? (NSInteger)60 : (NSInteger)30;
}
}
}
}
## Instruction:
Fix for [MPLY-9790]; iOS 60FPS Whitelist is now a 30FPS Blacklist. New devices will no longer need to be added to the list to get the benefit of 60FPS.
Buddy: Jonty
## Code After:
// Copyright eeGeo Ltd (2012-2015), All Rights Reserved
#include "DeviceHelpers.h"
#include <algorithm>
namespace ExampleApp
{
namespace Helpers
{
int deviceFramerateBlacklist[] = {
GAME_DEVICE_IPHONE1,
GAME_DEVICE_IPHONE3G,
GAME_DEVICE_IPHONE4,
GAME_DEVICE_IPHONE4S,
GAME_DEVICE_IPHONE5,
GAME_DEVICE_IPOD1GEN,
GAME_DEVICE_IPOD2GEN,
GAME_DEVICE_IPOD3GEN,
GAME_DEVICE_IPOD4GEN,
GAME_DEVICE_IPOD5GEN,
GAME_DEVICE_IPAD1,
GAME_DEVICE_IPAD2,
GAME_DEVICE_IPAD3,
GAME_DEVICE_IPAD4,
GAME_DEVICE_IPADAIR,
GAME_DEVICE_IPADMINI_1GEN,
GAME_DEVICE_IPADMINI_2
};
namespace DeviceHelpers
{
NSInteger GetPreferredFramerate(const GameDeviceType& deviceType)
{
bool isBlacklisted = std::find(std::begin(deviceFramerateBlacklist), std::end(deviceFramerateBlacklist), deviceType) != std::end(deviceFramerateBlacklist);
return isBlacklisted ? (NSInteger)30 : (NSInteger)60;
}
}
}
}
| // Copyright eeGeo Ltd (2012-2015), All Rights Reserved
#include "DeviceHelpers.h"
#include <algorithm>
namespace ExampleApp
{
namespace Helpers
{
- int deviceFramerateWhitelist[] = {
? ^^^^^
+ int deviceFramerateBlacklist[] = {
? ^^^^^
- GAME_DEVICE_IPHONE8_PLUS,
- GAME_DEVICE_IPHONE8,
? ^
+ GAME_DEVICE_IPHONE1,
? ^
- GAME_DEVICE_IPADPRO_12_9_INCH,
- GAME_DEVICE_IPADPRO_9_7_INCH,
- GAME_DEVICE_IPHONE_SE,
? ^^^
+ GAME_DEVICE_IPHONE3G,
? ^^
- GAME_DEVICE_IPHONE7_PLUS,
- GAME_DEVICE_IPHONE7,
? ^
+ GAME_DEVICE_IPHONE4,
? ^
- GAME_DEVICE_IPHONE6S_PLUS,
- GAME_DEVICE_IPHONE6S,
? ^
+ GAME_DEVICE_IPHONE4S,
? ^
- GAME_DEVICE_IPHONE6PLUS,
- GAME_DEVICE_IPHONE6,
- GAME_DEVICE_IPHONE5S,
? -
+ GAME_DEVICE_IPHONE5,
- GAME_DEVICE_IPOD6GEN,
? ^
+ GAME_DEVICE_IPOD1GEN,
? ^
+ GAME_DEVICE_IPOD2GEN,
+ GAME_DEVICE_IPOD3GEN,
+ GAME_DEVICE_IPOD4GEN,
+ GAME_DEVICE_IPOD5GEN,
+ GAME_DEVICE_IPAD1,
+ GAME_DEVICE_IPAD2,
+ GAME_DEVICE_IPAD3,
+ GAME_DEVICE_IPAD4,
- GAME_DEVICE_IPADAIR_2,
? --
+ GAME_DEVICE_IPADAIR,
- GAME_DEVICE_IPADMINI_3,
? ^
+ GAME_DEVICE_IPADMINI_1GEN,
? ^^^^
- GAME_DEVICE_IPADMINI_4
? ^
+ GAME_DEVICE_IPADMINI_2
? ^
};
namespace DeviceHelpers
{
NSInteger GetPreferredFramerate(const GameDeviceType& deviceType)
{
- bool isWhitelisted = std::find(std::begin(deviceFramerateWhitelist), std::end(deviceFramerateWhitelist), deviceType) != std::end(deviceFramerateWhitelist);
? ^^^^^ ^^^^^ ^^^^^ ^^^^^
+ bool isBlacklisted = std::find(std::begin(deviceFramerateBlacklist), std::end(deviceFramerateBlacklist), deviceType) != std::end(deviceFramerateBlacklist);
? ^^^^^ ^^^^^ ^^^^^ ^^^^^
- return isWhitelisted ? (NSInteger)60 : (NSInteger)30;
? ^^^^^ ^ ^
+ return isBlacklisted ? (NSInteger)30 : (NSInteger)60;
? ^^^^^ ^ ^
}
}
}
} | 39 | 1 | 20 | 19 |
cc6cb859f63472de0a3b8cb21fac33bc6f249d1c | README.md | README.md | pywinauto-64
============
This pywinauto branch (0.5.x) supports 64-bit and 32-bit Python from 2.6 to 3.4.
### Dependencies
* pyWin32 package only ([build 219](http://sourceforge.net/projects/pywin32/files/pywin32/Build%20219/) is recommended).
ActiveState Python 2.x already contains pyWin32 by default.
### Setup
* Install pyWin32 extensions above (no need for Active Python except 3.4 64-bit)
* Download [master branch as ZIP](https://github.com/vasily-v-ryabov/pywinauto-64/archive/master.zip)
* Just unpack and run **python.exe setup.py install**
### Testing status
| Platform: Win7 x64 (1920x1080) | unit tests pass rate |
|-----------------------------|-----------------|
| Python 2.6 32-bit | 97,7% (255/261) |
| Python 2.6 64-bit | 92,7% (242/261) |
| Python 3.4 32-bit | 97,3% (254/261) |
| Python 3.4 64-bit | 92,3% (241/261) |
#### Packages required for running unit tests
* [Pillow](https://pypi.python.org/pypi/Pillow/2.7.0) or PIL
* [coverage](https://pypi.python.org/pypi/coverage) | pywinauto-64
============
This pywinauto branch (0.5.x) supports 64-bit and 32-bit Python from 2.6 to 3.4.
### Dependencies
* pyWin32 package only ([build 219](http://sourceforge.net/projects/pywin32/files/pywin32/Build%20219/) is recommended).
ActiveState Python 2.x already contains pyWin32 by default.
### Setup
* Install pyWin32 extensions above (no need for Active Python except 3.4 64-bit)
* Download [master branch as ZIP](https://github.com/vasily-v-ryabov/pywinauto-64/archive/master.zip)
* Just unpack and run **python.exe setup.py install**
### Testing status
* [Unit tests pass rate for <master> branch](https://github.com/vasily-v-ryabov/pywinauto-64/wiki/Unit-testing-status)
#### Packages required for running unit tests
* [Pillow](https://pypi.python.org/pypi/Pillow/2.7.0) or PIL
* [coverage](https://pypi.python.org/pypi/coverage) | Move unit testing status to Wiki. | Move unit testing status to Wiki.
| Markdown | lgpl-2.1 | mindw/pywinauto,vasily-v-ryabov/pywinauto-64,mindw/pywinauto,airelil/pywinauto,vasily-v-ryabov/pywinauto,cetygamer/pywinauto,pywinauto/pywinauto,vasily-v-ryabov/pywinauto-64,drinkertea/pywinauto,MagazinnikIvan/pywinauto,moden-py/pywinauto,moden-py/pywinauto | markdown | ## Code Before:
pywinauto-64
============
This pywinauto branch (0.5.x) supports 64-bit and 32-bit Python from 2.6 to 3.4.
### Dependencies
* pyWin32 package only ([build 219](http://sourceforge.net/projects/pywin32/files/pywin32/Build%20219/) is recommended).
ActiveState Python 2.x already contains pyWin32 by default.
### Setup
* Install pyWin32 extensions above (no need for Active Python except 3.4 64-bit)
* Download [master branch as ZIP](https://github.com/vasily-v-ryabov/pywinauto-64/archive/master.zip)
* Just unpack and run **python.exe setup.py install**
### Testing status
| Platform: Win7 x64 (1920x1080) | unit tests pass rate |
|-----------------------------|-----------------|
| Python 2.6 32-bit | 97,7% (255/261) |
| Python 2.6 64-bit | 92,7% (242/261) |
| Python 3.4 32-bit | 97,3% (254/261) |
| Python 3.4 64-bit | 92,3% (241/261) |
#### Packages required for running unit tests
* [Pillow](https://pypi.python.org/pypi/Pillow/2.7.0) or PIL
* [coverage](https://pypi.python.org/pypi/coverage)
## Instruction:
Move unit testing status to Wiki.
## Code After:
pywinauto-64
============
This pywinauto branch (0.5.x) supports 64-bit and 32-bit Python from 2.6 to 3.4.
### Dependencies
* pyWin32 package only ([build 219](http://sourceforge.net/projects/pywin32/files/pywin32/Build%20219/) is recommended).
ActiveState Python 2.x already contains pyWin32 by default.
### Setup
* Install pyWin32 extensions above (no need for Active Python except 3.4 64-bit)
* Download [master branch as ZIP](https://github.com/vasily-v-ryabov/pywinauto-64/archive/master.zip)
* Just unpack and run **python.exe setup.py install**
### Testing status
* [Unit tests pass rate for <master> branch](https://github.com/vasily-v-ryabov/pywinauto-64/wiki/Unit-testing-status)
#### Packages required for running unit tests
* [Pillow](https://pypi.python.org/pypi/Pillow/2.7.0) or PIL
* [coverage](https://pypi.python.org/pypi/coverage) | pywinauto-64
============
This pywinauto branch (0.5.x) supports 64-bit and 32-bit Python from 2.6 to 3.4.
### Dependencies
* pyWin32 package only ([build 219](http://sourceforge.net/projects/pywin32/files/pywin32/Build%20219/) is recommended).
ActiveState Python 2.x already contains pyWin32 by default.
### Setup
* Install pyWin32 extensions above (no need for Active Python except 3.4 64-bit)
* Download [master branch as ZIP](https://github.com/vasily-v-ryabov/pywinauto-64/archive/master.zip)
* Just unpack and run **python.exe setup.py install**
### Testing status
+ * [Unit tests pass rate for <master> branch](https://github.com/vasily-v-ryabov/pywinauto-64/wiki/Unit-testing-status)
- | Platform: Win7 x64 (1920x1080) | unit tests pass rate |
- |-----------------------------|-----------------|
- | Python 2.6 32-bit | 97,7% (255/261) |
- | Python 2.6 64-bit | 92,7% (242/261) |
- | Python 3.4 32-bit | 97,3% (254/261) |
- | Python 3.4 64-bit | 92,3% (241/261) |
#### Packages required for running unit tests
* [Pillow](https://pypi.python.org/pypi/Pillow/2.7.0) or PIL
* [coverage](https://pypi.python.org/pypi/coverage) | 7 | 0.25 | 1 | 6 |
a87a5ff51312af66fe2d40075511842f6f6b89a6 | lib/assets/javascripts/cable.coffee.erb | lib/assets/javascripts/cable.coffee.erb |
@Cable =
INTERNAL: <%= ActionCable::INTERNAL.to_json %>
createConsumer: (url = @getConfig("url")) ->
new Cable.Consumer @webSocketURL(url)
getConfig: (name) ->
element = document.head.querySelector("meta[name='action-cable-#{name}']")
element?.getAttribute("content")
webSocketURL: (url) ->
if url and not /^wss?:/i.test(url)
a = document.createElement("a")
a.href = url
# in internet explorer, the next line properly populates the protocol
a.href = a.href
a.protocol = a.protocol.replace("http", "ws")
a.href
else
url
|
@Cable =
INTERNAL: <%= ActionCable::INTERNAL.to_json %>
createConsumer: (url = @getConfig("url")) ->
new Cable.Consumer @createWebSocketURL(url)
getConfig: (name) ->
element = document.head.querySelector("meta[name='action-cable-#{name}']")
element?.getAttribute("content")
createWebSocketURL: (url) ->
if url and not /^wss?:/i.test(url)
a = document.createElement("a")
a.href = url
# Fix populating Location properties in IE. Otherwise, protocol will be blank.
a.href = a.href
a.protocol = a.protocol.replace("http", "ws")
a.href
else
url
| Rename helper and update comment | Rename helper and update comment
| HTML+ERB | mit | valencar/actioncable,valencar/actioncable | html+erb | ## Code Before:
@Cable =
INTERNAL: <%= ActionCable::INTERNAL.to_json %>
createConsumer: (url = @getConfig("url")) ->
new Cable.Consumer @webSocketURL(url)
getConfig: (name) ->
element = document.head.querySelector("meta[name='action-cable-#{name}']")
element?.getAttribute("content")
webSocketURL: (url) ->
if url and not /^wss?:/i.test(url)
a = document.createElement("a")
a.href = url
# in internet explorer, the next line properly populates the protocol
a.href = a.href
a.protocol = a.protocol.replace("http", "ws")
a.href
else
url
## Instruction:
Rename helper and update comment
## Code After:
@Cable =
INTERNAL: <%= ActionCable::INTERNAL.to_json %>
createConsumer: (url = @getConfig("url")) ->
new Cable.Consumer @createWebSocketURL(url)
getConfig: (name) ->
element = document.head.querySelector("meta[name='action-cable-#{name}']")
element?.getAttribute("content")
createWebSocketURL: (url) ->
if url and not /^wss?:/i.test(url)
a = document.createElement("a")
a.href = url
# Fix populating Location properties in IE. Otherwise, protocol will be blank.
a.href = a.href
a.protocol = a.protocol.replace("http", "ws")
a.href
else
url
|
@Cable =
INTERNAL: <%= ActionCable::INTERNAL.to_json %>
createConsumer: (url = @getConfig("url")) ->
- new Cable.Consumer @webSocketURL(url)
? ^
+ new Cable.Consumer @createWebSocketURL(url)
? ^^^^^^^
getConfig: (name) ->
element = document.head.querySelector("meta[name='action-cable-#{name}']")
element?.getAttribute("content")
- webSocketURL: (url) ->
? ^
+ createWebSocketURL: (url) ->
? ^^^^^^^
if url and not /^wss?:/i.test(url)
a = document.createElement("a")
a.href = url
- # in internet explorer, the next line properly populates the protocol
+ # Fix populating Location properties in IE. Otherwise, protocol will be blank.
a.href = a.href
a.protocol = a.protocol.replace("http", "ws")
a.href
else
url | 6 | 0.285714 | 3 | 3 |
e435d5ee1c645759c88ec9a92bd7e61824d49b51 | install.sh | install.sh | DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
PARENT="$( cd $DIR && cd .. && pwd)"
mkdir "$DIR/lib"
echo "#1. compiling the library from source code and dependencies"
mvn clean
mvn install
echo "#2. moving binary to lib folder"
mv "$DIR/target/EventCoreference-v3.0-jar-with-dependencies.jar" "$DIR/lib"
echo "#3. installing the vua-resources"
cd "$PARENT"
git clone https://github.com/cltl/vua-resources.git
echo "#4. cleaning up"
cd "$DIR"
rm -R target
rm -R src
rm EventCoreference.iml
rm EventCoreference.iws
rm EventCoreference.ipr
| set -e
#Requires an installation of maven 2.x and Java 1.6 or higher
# define the location of the install scipt
DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
PARENT="$( cd $DIR && cd .. && pwd)"
mkdir "$DIR/lib"
echo "#1. compiling the library from source code and dependencies"
mvn clean
mvn install
echo "#2. moving binary to lib folder"
mv "$DIR/target/EventCoreference-v3.0-jar-with-dependencies.jar" "$DIR/lib"
echo "#3. installing the vua-resources"
cd "$PARENT"
git clone https://github.com/cltl/vua-resources.git
echo "#4. cleaning up"
cd "$DIR"
rm -R target
rm -R src
rm EventCoreference.iml
rm EventCoreference.iws
rm EventCoreference.ipr
| Add set -e to prevent silent errors | Add set -e to prevent silent errors
| Shell | apache-2.0 | cltl/EventCoreference,cltl/EventCoreference | shell | ## Code Before:
DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
PARENT="$( cd $DIR && cd .. && pwd)"
mkdir "$DIR/lib"
echo "#1. compiling the library from source code and dependencies"
mvn clean
mvn install
echo "#2. moving binary to lib folder"
mv "$DIR/target/EventCoreference-v3.0-jar-with-dependencies.jar" "$DIR/lib"
echo "#3. installing the vua-resources"
cd "$PARENT"
git clone https://github.com/cltl/vua-resources.git
echo "#4. cleaning up"
cd "$DIR"
rm -R target
rm -R src
rm EventCoreference.iml
rm EventCoreference.iws
rm EventCoreference.ipr
## Instruction:
Add set -e to prevent silent errors
## Code After:
set -e
#Requires an installation of maven 2.x and Java 1.6 or higher
# define the location of the install scipt
DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
PARENT="$( cd $DIR && cd .. && pwd)"
mkdir "$DIR/lib"
echo "#1. compiling the library from source code and dependencies"
mvn clean
mvn install
echo "#2. moving binary to lib folder"
mv "$DIR/target/EventCoreference-v3.0-jar-with-dependencies.jar" "$DIR/lib"
echo "#3. installing the vua-resources"
cd "$PARENT"
git clone https://github.com/cltl/vua-resources.git
echo "#4. cleaning up"
cd "$DIR"
rm -R target
rm -R src
rm EventCoreference.iml
rm EventCoreference.iws
rm EventCoreference.ipr
| + set -e
+
+ #Requires an installation of maven 2.x and Java 1.6 or higher
+
+ # define the location of the install scipt
DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
PARENT="$( cd $DIR && cd .. && pwd)"
mkdir "$DIR/lib"
echo "#1. compiling the library from source code and dependencies"
mvn clean
mvn install
echo "#2. moving binary to lib folder"
mv "$DIR/target/EventCoreference-v3.0-jar-with-dependencies.jar" "$DIR/lib"
echo "#3. installing the vua-resources"
cd "$PARENT"
git clone https://github.com/cltl/vua-resources.git
echo "#4. cleaning up"
cd "$DIR"
rm -R target
rm -R src
rm EventCoreference.iml
rm EventCoreference.iws
rm EventCoreference.ipr | 5 | 0.263158 | 5 | 0 |
2e62e364726000c72d31ebbd5311bc01b280029b | wercker.yml | wercker.yml | box: wercker/nodejs
build:
steps:
- npm-install
deploy:
steps:
- add-to-known_hosts:
hostname: server.octohost.io
- add-ssh-key:
keyname: KEY
- script:
name: Push to octohost
code: |
git push $GIT_REMOTE master
after-steps:
- sherzberg/slack-notify:
subdomain: ouisirfudge
token: $SLACK_TOKEN
channel: "#general"
| box: wercker/nodejs
build:
steps:
- npm-install
after-steps:
- wantedly/pretty-slack-notify:
webhook_url: $SLACK_WEBHOOK_URL
deploy:
steps:
- add-to-known_hosts:
hostname: server.octohost.io
- add-ssh-key:
keyname: KEY
- script:
name: Push to octohost
code: |
git push $GIT_REMOTE master
after-steps:
- wantedly/pretty-slack-notify:
webhook_url: $SLACK_WEBHOOK_URL
| Switch to alternate Slack Notify. | Switch to alternate Slack Notify. | YAML | apache-2.0 | octohost/www | yaml | ## Code Before:
box: wercker/nodejs
build:
steps:
- npm-install
deploy:
steps:
- add-to-known_hosts:
hostname: server.octohost.io
- add-ssh-key:
keyname: KEY
- script:
name: Push to octohost
code: |
git push $GIT_REMOTE master
after-steps:
- sherzberg/slack-notify:
subdomain: ouisirfudge
token: $SLACK_TOKEN
channel: "#general"
## Instruction:
Switch to alternate Slack Notify.
## Code After:
box: wercker/nodejs
build:
steps:
- npm-install
after-steps:
- wantedly/pretty-slack-notify:
webhook_url: $SLACK_WEBHOOK_URL
deploy:
steps:
- add-to-known_hosts:
hostname: server.octohost.io
- add-ssh-key:
keyname: KEY
- script:
name: Push to octohost
code: |
git push $GIT_REMOTE master
after-steps:
- wantedly/pretty-slack-notify:
webhook_url: $SLACK_WEBHOOK_URL
| box: wercker/nodejs
build:
steps:
- npm-install
+ after-steps:
+ - wantedly/pretty-slack-notify:
+ webhook_url: $SLACK_WEBHOOK_URL
deploy:
steps:
- add-to-known_hosts:
hostname: server.octohost.io
- add-ssh-key:
keyname: KEY
- script:
name: Push to octohost
code: |
git push $GIT_REMOTE master
after-steps:
+ - wantedly/pretty-slack-notify:
+ webhook_url: $SLACK_WEBHOOK_URL
- - sherzberg/slack-notify:
- subdomain: ouisirfudge
- token: $SLACK_TOKEN
- channel: "#general" | 9 | 0.473684 | 5 | 4 |
0e521e09e392f1584f8d54d5ea657c1f3b19c8dc | test/helper.rb | test/helper.rb |
require 'test/unit'
require 'fluent/test'
require 'fluent/plugin/in_udp_event'
unless defined?(Test::Unit::AssertionFailedError)
class Test::Unit::AssertionFailedError < StandardError
end
end
def unused_port
s = TCPServer.open(0)
port = s.addr[1]
s.close
port
end
def ipv6_enabled?
require 'socket'
begin
TCPServer.open('::1', 0)
true
rescue
false
end
end
|
require 'coveralls'
require 'test/unit'
require 'fluent/test'
require 'fluent/plugin/in_udp_event'
Coveralls.wear!
unless defined?(Test::Unit::AssertionFailedError)
class Test::Unit::AssertionFailedError < StandardError
end
end
def unused_port
s = TCPServer.open(0)
port = s.addr[1]
s.close
port
end
def ipv6_enabled?
require 'socket'
begin
TCPServer.open('::1', 0)
true
rescue
false
end
end
| Add coveralls when running tests | Add coveralls when running tests | Ruby | mit | ablagoev/fluent-plugin-in-udp-event | ruby | ## Code Before:
require 'test/unit'
require 'fluent/test'
require 'fluent/plugin/in_udp_event'
unless defined?(Test::Unit::AssertionFailedError)
class Test::Unit::AssertionFailedError < StandardError
end
end
def unused_port
s = TCPServer.open(0)
port = s.addr[1]
s.close
port
end
def ipv6_enabled?
require 'socket'
begin
TCPServer.open('::1', 0)
true
rescue
false
end
end
## Instruction:
Add coveralls when running tests
## Code After:
require 'coveralls'
require 'test/unit'
require 'fluent/test'
require 'fluent/plugin/in_udp_event'
Coveralls.wear!
unless defined?(Test::Unit::AssertionFailedError)
class Test::Unit::AssertionFailedError < StandardError
end
end
def unused_port
s = TCPServer.open(0)
port = s.addr[1]
s.close
port
end
def ipv6_enabled?
require 'socket'
begin
TCPServer.open('::1', 0)
true
rescue
false
end
end
|
+ require 'coveralls'
require 'test/unit'
require 'fluent/test'
require 'fluent/plugin/in_udp_event'
+
+ Coveralls.wear!
unless defined?(Test::Unit::AssertionFailedError)
class Test::Unit::AssertionFailedError < StandardError
end
end
def unused_port
s = TCPServer.open(0)
port = s.addr[1]
s.close
port
end
def ipv6_enabled?
require 'socket'
begin
TCPServer.open('::1', 0)
true
rescue
false
end
end | 3 | 0.111111 | 3 | 0 |
4bade6d15cf35cc7aca0c2c237c3965723bc3a47 | SpringMVC/src/main/webapp/assets/js/main.js | SpringMVC/src/main/webapp/assets/js/main.js | $(document).ready(function() {
$('#dataTable').DataTable();
} ); | $(document).ready(function() {
$('#dataTable').dataTable( {
language: {
search: "_INPUT_",
searchPlaceholder: "Search"
}
} );
} ); | Add html placeholder option to search input | Add html placeholder option to search input
| JavaScript | apache-2.0 | erdgny/SpringMVCHibernate,erdgny/SpringMVCHibernate | javascript | ## Code Before:
$(document).ready(function() {
$('#dataTable').DataTable();
} );
## Instruction:
Add html placeholder option to search input
## Code After:
$(document).ready(function() {
$('#dataTable').dataTable( {
language: {
search: "_INPUT_",
searchPlaceholder: "Search"
}
} );
} ); | $(document).ready(function() {
- $('#dataTable').DataTable();
? ^ ^^
+ $('#dataTable').dataTable( {
? ^ ^^
+ language: {
+ search: "_INPUT_",
+ searchPlaceholder: "Search"
+ }
+ } );
} ); | 7 | 2.333333 | 6 | 1 |
24377d97c41b042c5bba12e85e23538852b500a8 | content/ourwork.md | content/ourwork.md |
BOH seeks to sustainably empower widows and orphans living in Uganda. In addition, BOH aims to spread compassionate global awareness within our local communities. We hope to unite people from different continents and cultures with the purpose of generating lasting solutions to the problems that arise from poverty.
## Our Committment to Sustainability
As a part of our commitment to sustainable empowerment, BOH works in partnership with a Ugandan organization, WOCAP. Together, through the Beads Project, we generate sustainable income sources for women and their families. By providing widows with a job opportunity, rather than a handout, they are empowered to provide for their families with dignity. In addition, BOH began a student scholarship program (BSSP) to empower students through education. Education is a crucial component in creating opportunities for the future and breaking out of the cycle of poverty.
|
BOH seeks to sustainably empower widows and orphans living in Uganda. In addition, BOH aims to spread compassionate global awareness within our local communities. We hope to unite people from different continents and cultures with the purpose of generating lasting solutions to the problems that arise from poverty.
## Our Committment to Sustainability
We commit to creating lasting, self-sustaining change. To do this, BOH works in partnership with an Ugandan organization, WOCAP, to design, implement, and maintain programs to target specific contributing factors to poverty. Being committed to sustainability means we refuse to give handouts. Instead, BOH and WOCAP use existing resources in new ways to empower the women and children we work with. To learn more about how we do this, take a look at our three programs: the Beads Project, BOH Student Scholarship Program (BSSP), and the Agriculture Program.
| Update Our commitment to Sustainability text on Mission page | Update Our commitment to Sustainability text on Mission page
https://trello.com/c/KhaZU59x/320-update-text-on-mission-page
| Markdown | mit | becauseofhope/because-of-hope,becauseofhope/because-of-hope,becauseofhope/because-of-hope,becauseofhope/because-of-hope | markdown | ## Code Before:
BOH seeks to sustainably empower widows and orphans living in Uganda. In addition, BOH aims to spread compassionate global awareness within our local communities. We hope to unite people from different continents and cultures with the purpose of generating lasting solutions to the problems that arise from poverty.
## Our Committment to Sustainability
As a part of our commitment to sustainable empowerment, BOH works in partnership with a Ugandan organization, WOCAP. Together, through the Beads Project, we generate sustainable income sources for women and their families. By providing widows with a job opportunity, rather than a handout, they are empowered to provide for their families with dignity. In addition, BOH began a student scholarship program (BSSP) to empower students through education. Education is a crucial component in creating opportunities for the future and breaking out of the cycle of poverty.
## Instruction:
Update Our commitment to Sustainability text on Mission page
https://trello.com/c/KhaZU59x/320-update-text-on-mission-page
## Code After:
BOH seeks to sustainably empower widows and orphans living in Uganda. In addition, BOH aims to spread compassionate global awareness within our local communities. We hope to unite people from different continents and cultures with the purpose of generating lasting solutions to the problems that arise from poverty.
## Our Committment to Sustainability
We commit to creating lasting, self-sustaining change. To do this, BOH works in partnership with an Ugandan organization, WOCAP, to design, implement, and maintain programs to target specific contributing factors to poverty. Being committed to sustainability means we refuse to give handouts. Instead, BOH and WOCAP use existing resources in new ways to empower the women and children we work with. To learn more about how we do this, take a look at our three programs: the Beads Project, BOH Student Scholarship Program (BSSP), and the Agriculture Program.
|
BOH seeks to sustainably empower widows and orphans living in Uganda. In addition, BOH aims to spread compassionate global awareness within our local communities. We hope to unite people from different continents and cultures with the purpose of generating lasting solutions to the problems that arise from poverty.
## Our Committment to Sustainability
+ We commit to creating lasting, self-sustaining change. To do this, BOH works in partnership with an Ugandan organization, WOCAP, to design, implement, and maintain programs to target specific contributing factors to poverty. Being committed to sustainability means we refuse to give handouts. Instead, BOH and WOCAP use existing resources in new ways to empower the women and children we work with. To learn more about how we do this, take a look at our three programs: the Beads Project, BOH Student Scholarship Program (BSSP), and the Agriculture Program.
- As a part of our commitment to sustainable empowerment, BOH works in partnership with a Ugandan organization, WOCAP. Together, through the Beads Project, we generate sustainable income sources for women and their families. By providing widows with a job opportunity, rather than a handout, they are empowered to provide for their families with dignity. In addition, BOH began a student scholarship program (BSSP) to empower students through education. Education is a crucial component in creating opportunities for the future and breaking out of the cycle of poverty.
- | 3 | 0.428571 | 1 | 2 |
ea8b3f42dd2cf21aadf333ba9262a864a0fc5267 | autocloud/web/templates/job_output.html | autocloud/web/templates/job_output.html | {% extends "master.html" %}
{% block body %}
<div class="container">
<p>Output for task: {{ job.taskid }}</p>
<pre>{{ job.output }}</pre>
<div>
{% endblock %}
| {% extends "master.html" %}
{% block body %}
<a class="pull-right" href="{{ url_for('job_details') }}">Go to job details</a>
<div class="container">
<p>Output for task: {{ job.taskid }}</p>
<pre>{{ job.output }}</pre>
</div>
{% endblock %}
| Add job details link to output page | Add job details link to output page
| HTML | agpl-3.0 | kushaldas/autocloud,kushaldas/autocloud,maxamillion/autocloud,kushaldas/autocloud,kushaldas/autocloud,maxamillion/autocloud,maxamillion/autocloud,maxamillion/autocloud | html | ## Code Before:
{% extends "master.html" %}
{% block body %}
<div class="container">
<p>Output for task: {{ job.taskid }}</p>
<pre>{{ job.output }}</pre>
<div>
{% endblock %}
## Instruction:
Add job details link to output page
## Code After:
{% extends "master.html" %}
{% block body %}
<a class="pull-right" href="{{ url_for('job_details') }}">Go to job details</a>
<div class="container">
<p>Output for task: {{ job.taskid }}</p>
<pre>{{ job.output }}</pre>
</div>
{% endblock %}
| {% extends "master.html" %}
{% block body %}
+ <a class="pull-right" href="{{ url_for('job_details') }}">Go to job details</a>
<div class="container">
<p>Output for task: {{ job.taskid }}</p>
<pre>{{ job.output }}</pre>
- <div>
+ </div>
? +
{% endblock %} | 3 | 0.375 | 2 | 1 |
e93151490ea9c96b3856ec2e269552d8b52e0355 | lobster/cmssw/__init__.py | lobster/cmssw/__init__.py | from job import JobProvider
from publish import publish
from jobit import JobitStore
from merge import MergeProvider
| from job import JobProvider
from jobit import JobitStore
from merge import MergeProvider
from plotting import plot
from publish import publish
| Add plotting to the default imports of cmssw. | Add plotting to the default imports of cmssw.
| Python | mit | matz-e/lobster,matz-e/lobster,matz-e/lobster | python | ## Code Before:
from job import JobProvider
from publish import publish
from jobit import JobitStore
from merge import MergeProvider
## Instruction:
Add plotting to the default imports of cmssw.
## Code After:
from job import JobProvider
from jobit import JobitStore
from merge import MergeProvider
from plotting import plot
from publish import publish
| from job import JobProvider
- from publish import publish
from jobit import JobitStore
from merge import MergeProvider
+ from plotting import plot
+ from publish import publish | 3 | 0.75 | 2 | 1 |
1a8e07f218c0f467e2b677604888d66e9ec83ab6 | travis/install_ycmd.sh | travis/install_ycmd.sh |
export YCMD_PATH="${HOME}/ycmd"
if [ ! -d "$YCMD_PATH/.git" ]; then
git clone --depth=1 --recursive https://github.com/Valloric/ycmd ${YCMD_PATH}
fi
pushd ${YCMD_PATH}
git pull
git submodule update --init --recursive
export EXTRA_CMAKE_ARGS="-DCMAKE_CXX_COMPILER=/usr/lib/ccache/c++ -DCMAKE_C_COMPILER=/usr/lib/ccache/cc"
python build.py --clang-completer --gocode-completer --tern-completer
valNumResult=$?
if [[ $valNumResult -eq 1 ]]
then
return 1
fi
npm install -g typescript
popd
|
export YCMD_PATH="${HOME}/ycmd"
if [ ! -d "$YCMD_PATH/.git" ]; then
git clone --depth=1 --recursive https://github.com/Valloric/ycmd ${YCMD_PATH}
fi
pushd ${YCMD_PATH}
git pull
git submodule update --init --recursive
export EXTRA_CMAKE_ARGS="-DCMAKE_CXX_COMPILER=/usr/lib/ccache/c++ -DCMAKE_C_COMPILER=/usr/lib/ccache/cc"
python build.py --clang-completer --gocode-completer --tern-completer
npm install -g typescript
popd
| Clean up script for installing ycmd on travis | Clean up script for installing ycmd on travis
| Shell | mit | abingham/emacs-ycmd,emacsmirror/ycmd,ptrv/emacs-ycmd,emacsmirror/ycmd,emacsmirror/ycmd,ptrv/emacs-ycmd,ptrv/emacs-ycmd,emacsmirror/ycmd,emacsmirror/ycmd,abingham/emacs-ycmd,ptrv/emacs-ycmd,abingham/emacs-ycmd,abingham/emacs-ycmd,ptrv/emacs-ycmd,abingham/emacs-ycmd,abingham/emacs-ycmd,emacsmirror/ycmd,ptrv/emacs-ycmd | shell | ## Code Before:
export YCMD_PATH="${HOME}/ycmd"
if [ ! -d "$YCMD_PATH/.git" ]; then
git clone --depth=1 --recursive https://github.com/Valloric/ycmd ${YCMD_PATH}
fi
pushd ${YCMD_PATH}
git pull
git submodule update --init --recursive
export EXTRA_CMAKE_ARGS="-DCMAKE_CXX_COMPILER=/usr/lib/ccache/c++ -DCMAKE_C_COMPILER=/usr/lib/ccache/cc"
python build.py --clang-completer --gocode-completer --tern-completer
valNumResult=$?
if [[ $valNumResult -eq 1 ]]
then
return 1
fi
npm install -g typescript
popd
## Instruction:
Clean up script for installing ycmd on travis
## Code After:
export YCMD_PATH="${HOME}/ycmd"
if [ ! -d "$YCMD_PATH/.git" ]; then
git clone --depth=1 --recursive https://github.com/Valloric/ycmd ${YCMD_PATH}
fi
pushd ${YCMD_PATH}
git pull
git submodule update --init --recursive
export EXTRA_CMAKE_ARGS="-DCMAKE_CXX_COMPILER=/usr/lib/ccache/c++ -DCMAKE_C_COMPILER=/usr/lib/ccache/cc"
python build.py --clang-completer --gocode-completer --tern-completer
npm install -g typescript
popd
|
export YCMD_PATH="${HOME}/ycmd"
if [ ! -d "$YCMD_PATH/.git" ]; then
git clone --depth=1 --recursive https://github.com/Valloric/ycmd ${YCMD_PATH}
fi
pushd ${YCMD_PATH}
git pull
git submodule update --init --recursive
export EXTRA_CMAKE_ARGS="-DCMAKE_CXX_COMPILER=/usr/lib/ccache/c++ -DCMAKE_C_COMPILER=/usr/lib/ccache/cc"
python build.py --clang-completer --gocode-completer --tern-completer
- valNumResult=$?
- if [[ $valNumResult -eq 1 ]]
- then
- return 1
- fi
-
npm install -g typescript
popd | 6 | 0.25 | 0 | 6 |
078091d8761e3f5145df2e9d04c15c1d132b732a | lib/electric_sheep/transports/scp.rb | lib/electric_sheep/transports/scp.rb | module ElectricSheep
module Transports
module SCP
end
end
end
| module ElectricSheep
module Transports
class SCP
include ElectricSheep::Transport
register as: "scp"
def copy
logger.info "Will copy #{resource.basename} " +
"from #{resource.host} " +
"to #{option(:to).to_s}"
end
def move
logger.info "Will move #{resource.basename}" +
"from #{resource.host} " +
"to #{option(:to).to_s}"
end
end
end
end
| Add a skeleton for SCP transport. | Add a skeleton for SCP transport.
| Ruby | mit | benitoDeLaCasita/electric_sheep,servebox/electric_sheep,ehartmann/electric_sheep,ehartmann/electric_sheep,ehartmann/electric_sheep,benitoDeLaCasita/electric_sheep,servebox/electric_sheep,servebox/electric_sheep | ruby | ## Code Before:
module ElectricSheep
module Transports
module SCP
end
end
end
## Instruction:
Add a skeleton for SCP transport.
## Code After:
module ElectricSheep
module Transports
class SCP
include ElectricSheep::Transport
register as: "scp"
def copy
logger.info "Will copy #{resource.basename} " +
"from #{resource.host} " +
"to #{option(:to).to_s}"
end
def move
logger.info "Will move #{resource.basename}" +
"from #{resource.host} " +
"to #{option(:to).to_s}"
end
end
end
end
| module ElectricSheep
module Transports
- module SCP
+ class SCP
+ include ElectricSheep::Transport
+
+ register as: "scp"
+
+ def copy
+ logger.info "Will copy #{resource.basename} " +
+ "from #{resource.host} " +
+ "to #{option(:to).to_s}"
+ end
+
+ def move
+ logger.info "Will move #{resource.basename}" +
+ "from #{resource.host} " +
+ "to #{option(:to).to_s}"
+ end
end
end
end | 17 | 2.428571 | 16 | 1 |
904994432a75a1db0df61ff0782d97dbd2c5a5ad | pytest.ini | pytest.ini | [pytest]
python_files=eve/tests/*.py
addopts = --maxfail=2 -rf
norecursedirs = testsuite .tox
| [pytest]
python_files=eve/tests/*.py
addopts = --maxfail=2 -rf --capture=no
norecursedirs = testsuite .tox
| Disable capture on py.test so ipython works on --pdb | Disable capture on py.test so ipython works on --pdb
| INI | bsd-3-clause | matthieuprat/eve,sebest/eve,EasonYi/eve,superdesk/eve,elpoisterio/eve,kalbasit/eve,hustlzp/eve,bcrochet/eve,mcreenan/eve,yanyanqin/eve,jzorrof/eve,pjs7678/eve,stratosgear/eve,julianhille/eve,amagdas/eve,mugurrus/eve,kidaa/eve,eduardomb/eve | ini | ## Code Before:
[pytest]
python_files=eve/tests/*.py
addopts = --maxfail=2 -rf
norecursedirs = testsuite .tox
## Instruction:
Disable capture on py.test so ipython works on --pdb
## Code After:
[pytest]
python_files=eve/tests/*.py
addopts = --maxfail=2 -rf --capture=no
norecursedirs = testsuite .tox
| [pytest]
python_files=eve/tests/*.py
- addopts = --maxfail=2 -rf
+ addopts = --maxfail=2 -rf --capture=no
? +++++++++++++
norecursedirs = testsuite .tox
- | 3 | 0.6 | 1 | 2 |
f968d0e766c1808621d4e6a045ab2ebe0fd2950b | app/scripts/directives/nobeFooter/nobeFooter.html | app/scripts/directives/nobeFooter/nobeFooter.html |
<footer id="footer" class="midnight-blue">
<div class="container">
<div class="row">
<div class="col-sm-6">
© {{copyrightYear}} NOBE Illinois. All Rights Reserved.
</div>
<div class="col-sm-6">
<ul class="pull-right">
<li><a href="#">Home</a></li>
<li><a href="#">About Us</a></li>
<li><a href="#">Faq</a></li>
<li><a href="#">Contact Us</a></li>
</ul>
</div>
</div>
</div>
</footer><!--/#footer-->
|
<footer id="footer" class="midnight-blue">
<div class="container">
<div class="row">
<div class="col-sm-6">
© {{copyrightYear}} NOBE Illinois. All Rights Reserved.
</div>
</div>
</div>
</footer><!--/#footer-->
| Remove redundant links on bottom right hand corner | Remove redundant links on bottom right hand corner
| HTML | mit | gruan/NOBEIllinois-Website,gruan/NOBEIllinois-Website | html | ## Code Before:
<footer id="footer" class="midnight-blue">
<div class="container">
<div class="row">
<div class="col-sm-6">
© {{copyrightYear}} NOBE Illinois. All Rights Reserved.
</div>
<div class="col-sm-6">
<ul class="pull-right">
<li><a href="#">Home</a></li>
<li><a href="#">About Us</a></li>
<li><a href="#">Faq</a></li>
<li><a href="#">Contact Us</a></li>
</ul>
</div>
</div>
</div>
</footer><!--/#footer-->
## Instruction:
Remove redundant links on bottom right hand corner
## Code After:
<footer id="footer" class="midnight-blue">
<div class="container">
<div class="row">
<div class="col-sm-6">
© {{copyrightYear}} NOBE Illinois. All Rights Reserved.
</div>
</div>
</div>
</footer><!--/#footer-->
|
<footer id="footer" class="midnight-blue">
<div class="container">
<div class="row">
<div class="col-sm-6">
© {{copyrightYear}} NOBE Illinois. All Rights Reserved.
</div>
- <div class="col-sm-6">
- <ul class="pull-right">
- <li><a href="#">Home</a></li>
- <li><a href="#">About Us</a></li>
- <li><a href="#">Faq</a></li>
- <li><a href="#">Contact Us</a></li>
- </ul>
- </div>
</div>
</div>
</footer><!--/#footer--> | 8 | 0.444444 | 0 | 8 |
78ec1cffde6443016bae2c8aefdb67ab26bfab10 | __init__.py | __init__.py | from . import OctoPrintOutputDevicePlugin
from . import DiscoverOctoPrintAction
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "extension",
"plugin": {
"name": "Wifi connection",
"author": "Ultimaker",
"description": catalog.i18nc("Wifi connection", "Wifi connection"),
"api": 3
}
}
def register(app):
return {
"output_device": OctoPrintOutputDevicePlugin.OctoPrintOutputDevicePlugin(),
"machine_action": DiscoverOctoPrintAction.DiscoverOctoPrintAction()
} | from . import OctoPrintOutputDevicePlugin
from . import DiscoverOctoPrintAction
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "extension",
"plugin": {
"name": "OctoPrint connection",
"author": "fieldOfView",
"version": "1.0",
"description": catalog.i18nc("@info:whatsthis", "Allows sending prints to OctoPrint and monitoring the progress"),
"api": 3
}
}
def register(app):
return {
"output_device": OctoPrintOutputDevicePlugin.OctoPrintOutputDevicePlugin(),
"machine_action": DiscoverOctoPrintAction.DiscoverOctoPrintAction()
} | Update plugin information (name, description, version, author) | Update plugin information (name, description, version, author)
| Python | agpl-3.0 | fieldOfView/OctoPrintPlugin | python | ## Code Before:
from . import OctoPrintOutputDevicePlugin
from . import DiscoverOctoPrintAction
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "extension",
"plugin": {
"name": "Wifi connection",
"author": "Ultimaker",
"description": catalog.i18nc("Wifi connection", "Wifi connection"),
"api": 3
}
}
def register(app):
return {
"output_device": OctoPrintOutputDevicePlugin.OctoPrintOutputDevicePlugin(),
"machine_action": DiscoverOctoPrintAction.DiscoverOctoPrintAction()
}
## Instruction:
Update plugin information (name, description, version, author)
## Code After:
from . import OctoPrintOutputDevicePlugin
from . import DiscoverOctoPrintAction
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "extension",
"plugin": {
"name": "OctoPrint connection",
"author": "fieldOfView",
"version": "1.0",
"description": catalog.i18nc("@info:whatsthis", "Allows sending prints to OctoPrint and monitoring the progress"),
"api": 3
}
}
def register(app):
return {
"output_device": OctoPrintOutputDevicePlugin.OctoPrintOutputDevicePlugin(),
"machine_action": DiscoverOctoPrintAction.DiscoverOctoPrintAction()
} | from . import OctoPrintOutputDevicePlugin
from . import DiscoverOctoPrintAction
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "extension",
"plugin": {
- "name": "Wifi connection",
? ^ ^^
+ "name": "OctoPrint connection",
? ^^^^^^ ^^
- "author": "Ultimaker",
? ^ ^ --- ^
+ "author": "fieldOfView",
? ^^^ ^^^^ ^
- "description": catalog.i18nc("Wifi connection", "Wifi connection"),
+ "version": "1.0",
+ "description": catalog.i18nc("@info:whatsthis", "Allows sending prints to OctoPrint and monitoring the progress"),
"api": 3
}
}
def register(app):
return {
"output_device": OctoPrintOutputDevicePlugin.OctoPrintOutputDevicePlugin(),
"machine_action": DiscoverOctoPrintAction.DiscoverOctoPrintAction()
} | 7 | 0.333333 | 4 | 3 |
f4c49daa568f11dfca9a9cb7979fa3261aa44287 | server/package.json | server/package.json | {
"name": "recycling-server",
"description": "recycling API",
"version": "0.0.1",
"private": "true",
"dependencies": {
"express": "3.x",
"lodash": "^2.4.1",
"pg": "3.x",
"sequelize": "^1.7.9",
"sequelize-cli": "^0.2.3"
},
"main": "server.js"
}
| {
"name": "recycling-server",
"description": "recycling API",
"version": "0.0.1",
"private": "true",
"dependencies": {
"express": "3.x",
"lodash": "^2.4.1",
"logfmt": "^1.1.2",
"pg": "3.x",
"sequelize": "^1.7.9",
"sequelize-cli": "^0.2.3"
},
"main": "server.js"
}
| Add logfmt to node modules | Add logfmt to node modules
| JSON | mit | davidjamesknight/recycling,open-city/recycling,open-austin/mybuildingdoesntrecycle,open-city/recycling,davidjamesknight/recycling,open-austin/mybuildingdoesntrecycle,davidjamesknight/recycling,open-austin/mybuildingdoesntrecycle,open-city/recycling | json | ## Code Before:
{
"name": "recycling-server",
"description": "recycling API",
"version": "0.0.1",
"private": "true",
"dependencies": {
"express": "3.x",
"lodash": "^2.4.1",
"pg": "3.x",
"sequelize": "^1.7.9",
"sequelize-cli": "^0.2.3"
},
"main": "server.js"
}
## Instruction:
Add logfmt to node modules
## Code After:
{
"name": "recycling-server",
"description": "recycling API",
"version": "0.0.1",
"private": "true",
"dependencies": {
"express": "3.x",
"lodash": "^2.4.1",
"logfmt": "^1.1.2",
"pg": "3.x",
"sequelize": "^1.7.9",
"sequelize-cli": "^0.2.3"
},
"main": "server.js"
}
| {
"name": "recycling-server",
"description": "recycling API",
"version": "0.0.1",
"private": "true",
"dependencies": {
"express": "3.x",
"lodash": "^2.4.1",
+ "logfmt": "^1.1.2",
"pg": "3.x",
"sequelize": "^1.7.9",
"sequelize-cli": "^0.2.3"
},
"main": "server.js"
} | 1 | 0.071429 | 1 | 0 |
6b1587848a9a8eba96d2df8ba71b24747de4891e | specs/bar-spec.js | specs/bar-spec.js | describe("Process Bar", function () {
var bar;
beforeEach(function () {
jasmine.clock().install();
bar = new EchoesWorks.bar();
});
afterEach(function () {
jasmine.clock().uninstall();
});
it("should return element", function () {
bar.go(30);
jasmine.clock().tick(100);
expect(bar.bars[0].moving).toBe(true);
jasmine.clock().tick(400);
expect(bar.bars[0].here).toBe(30);
expect(bar.bars[0].moving).toBe(false);
});
});
| describe("Process Bar", function () {
var bar;
beforeEach(function () {
jasmine.clock().install();
bar = new EchoesWorks.bar();
});
afterEach(function () {
jasmine.clock().uninstall();
});
it("should return correct process width", function () {
bar.go(30);
jasmine.clock().tick(100);
expect(bar.bars[0].moving).toBe(true);
jasmine.clock().tick(400);
expect(bar.bars[0].here).toBe(30);
expect(bar.bars[0].moving).toBe(false);
});
it("should return 0 when go 100 ", function () {
bar.go(99);
jasmine.clock().tick(1000);
expect(bar.bars[0].here).toBe(99);
bar.go(100);
jasmine.clock().tick(600);
expect(bar.bars[0].here).toBe(0);
});
});
| Add test for 100 width | Add test for 100 width
| JavaScript | mit | Suninus/echoesworks,echoesworks/echoesworks,jaambee/echoesworks,Suninus/echoesworks,phodal/echoesworks,phodal/echoesworks,jaambee/echoesworks,echoesworks/echoesworks | javascript | ## Code Before:
describe("Process Bar", function () {
var bar;
beforeEach(function () {
jasmine.clock().install();
bar = new EchoesWorks.bar();
});
afterEach(function () {
jasmine.clock().uninstall();
});
it("should return element", function () {
bar.go(30);
jasmine.clock().tick(100);
expect(bar.bars[0].moving).toBe(true);
jasmine.clock().tick(400);
expect(bar.bars[0].here).toBe(30);
expect(bar.bars[0].moving).toBe(false);
});
});
## Instruction:
Add test for 100 width
## Code After:
describe("Process Bar", function () {
var bar;
beforeEach(function () {
jasmine.clock().install();
bar = new EchoesWorks.bar();
});
afterEach(function () {
jasmine.clock().uninstall();
});
it("should return correct process width", function () {
bar.go(30);
jasmine.clock().tick(100);
expect(bar.bars[0].moving).toBe(true);
jasmine.clock().tick(400);
expect(bar.bars[0].here).toBe(30);
expect(bar.bars[0].moving).toBe(false);
});
it("should return 0 when go 100 ", function () {
bar.go(99);
jasmine.clock().tick(1000);
expect(bar.bars[0].here).toBe(99);
bar.go(100);
jasmine.clock().tick(600);
expect(bar.bars[0].here).toBe(0);
});
});
| describe("Process Bar", function () {
var bar;
beforeEach(function () {
jasmine.clock().install();
bar = new EchoesWorks.bar();
});
afterEach(function () {
jasmine.clock().uninstall();
});
- it("should return element", function () {
? ^ ^^^
+ it("should return correct process width", function () {
? ++++ ^^^^^^^ ^^^^^^ +
bar.go(30);
jasmine.clock().tick(100);
expect(bar.bars[0].moving).toBe(true);
jasmine.clock().tick(400);
expect(bar.bars[0].here).toBe(30);
expect(bar.bars[0].moving).toBe(false);
});
+ it("should return 0 when go 100 ", function () {
+ bar.go(99);
+ jasmine.clock().tick(1000);
+ expect(bar.bars[0].here).toBe(99);
+ bar.go(100);
+ jasmine.clock().tick(600);
+ expect(bar.bars[0].here).toBe(0);
+ });
+
});
| 11 | 0.478261 | 10 | 1 |
4f05ab7779d3732e3429646416e571ab4b25df70 | zsh/aliases.zsh | zsh/aliases.zsh | update () {
bash "${OSZ_ROOT}/update"
}
alias update=update
# Project folder alias so we can `c [tab]` to it.
setopt AUTO_CD # auto cd by typing the path.
projects="${HOME}/Projects"
alias p='cs ${HOME}/Projects/'
# reload zsh
alias reload='. $HOME/.zshrc'
# Command to compare two directories with sublimerge.
compare () {
subl --command 'sublimerge_compare_paths {"paths": ["$1", "$2"]}'
}
alias compare=compare
# Change default browser.
alias chrome='open -a "/Applications/Google Chrome.app" --args --make-default-browser'
alias canary='open -a "/Applications/Google Chrome Canary.app" --args --make-default-browser'
alias firefox='open -a "Firefox" --args --setDefaultBrowser'
alias firefoxdev='open -a "FirefoxDeveloperEdition" --args --setDefaultBrowser'
alias safari='open -a "Safari" --args --make-default-browser'
| update () {
bash "${OSZ_ROOT}/update"
}
alias update=update
# Project folder alias so we can `c [tab]` to it.
setopt AUTO_CD # auto cd by typing the path.
projects="${HOME}/Projects"
# reload zsh
alias reload='. $HOME/.zshrc'
# Command to compare two directories with sublimerge.
compare () {
subl --command 'sublimerge_compare_paths {"paths": ["$1", "$2"]}'
}
alias compare=compare
# Change default browser.
alias chrome='open -a "/Applications/Google Chrome.app" --args --make-default-browser'
alias canary='open -a "/Applications/Google Chrome Canary.app" --args --make-default-browser'
alias firefox='open -a "Firefox" --args --setDefaultBrowser'
alias firefoxdev='open -a "FirefoxDeveloperEdition" --args --setDefaultBrowser'
alias safari='open -a "Safari" --args --make-default-browser'
| Add `~/Projects` to cd autocomplete path (CDPATH). | feat(zsh): Add `~/Projects` to cd autocomplete path (CDPATH).
| Shell | mit | thezimmee/os-zimmee,thezimmee/os-zimmee,thezimmee/os-zimmee,thezimmee/os-zimmee,thezimmee/os-zimmee,thezimmee/os-zimmee | shell | ## Code Before:
update () {
bash "${OSZ_ROOT}/update"
}
alias update=update
# Project folder alias so we can `c [tab]` to it.
setopt AUTO_CD # auto cd by typing the path.
projects="${HOME}/Projects"
alias p='cs ${HOME}/Projects/'
# reload zsh
alias reload='. $HOME/.zshrc'
# Command to compare two directories with sublimerge.
compare () {
subl --command 'sublimerge_compare_paths {"paths": ["$1", "$2"]}'
}
alias compare=compare
# Change default browser.
alias chrome='open -a "/Applications/Google Chrome.app" --args --make-default-browser'
alias canary='open -a "/Applications/Google Chrome Canary.app" --args --make-default-browser'
alias firefox='open -a "Firefox" --args --setDefaultBrowser'
alias firefoxdev='open -a "FirefoxDeveloperEdition" --args --setDefaultBrowser'
alias safari='open -a "Safari" --args --make-default-browser'
## Instruction:
feat(zsh): Add `~/Projects` to cd autocomplete path (CDPATH).
## Code After:
update () {
bash "${OSZ_ROOT}/update"
}
alias update=update
# Project folder alias so we can `c [tab]` to it.
setopt AUTO_CD # auto cd by typing the path.
projects="${HOME}/Projects"
# reload zsh
alias reload='. $HOME/.zshrc'
# Command to compare two directories with sublimerge.
compare () {
subl --command 'sublimerge_compare_paths {"paths": ["$1", "$2"]}'
}
alias compare=compare
# Change default browser.
alias chrome='open -a "/Applications/Google Chrome.app" --args --make-default-browser'
alias canary='open -a "/Applications/Google Chrome Canary.app" --args --make-default-browser'
alias firefox='open -a "Firefox" --args --setDefaultBrowser'
alias firefoxdev='open -a "FirefoxDeveloperEdition" --args --setDefaultBrowser'
alias safari='open -a "Safari" --args --make-default-browser'
| update () {
bash "${OSZ_ROOT}/update"
}
alias update=update
# Project folder alias so we can `c [tab]` to it.
setopt AUTO_CD # auto cd by typing the path.
- projects="${HOME}/Projects"
+ projects="${HOME}/Projects"
? +
- alias p='cs ${HOME}/Projects/'
# reload zsh
alias reload='. $HOME/.zshrc'
# Command to compare two directories with sublimerge.
compare () {
subl --command 'sublimerge_compare_paths {"paths": ["$1", "$2"]}'
}
alias compare=compare
# Change default browser.
alias chrome='open -a "/Applications/Google Chrome.app" --args --make-default-browser'
alias canary='open -a "/Applications/Google Chrome Canary.app" --args --make-default-browser'
alias firefox='open -a "Firefox" --args --setDefaultBrowser'
alias firefoxdev='open -a "FirefoxDeveloperEdition" --args --setDefaultBrowser'
alias safari='open -a "Safari" --args --make-default-browser' | 3 | 0.12 | 1 | 2 |
54c0b2e2d3589e7ff55f6fb810b01065d5743a44 | src/main/java/com/lichader/alfred/controller/MetroServiceCheckController.java | src/main/java/com/lichader/alfred/controller/MetroServiceCheckController.java | package com.lichader.alfred.controller;
import com.lichader.alfred.logic.TrainlineDisruptionRetrievalLogic;
import com.lichader.alfred.metroapi.v3.model.Disruption;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.web.bind.annotation.GetMapping;
import org.springframework.web.bind.annotation.RequestMapping;
import org.springframework.web.bind.annotation.RestController;
import java.util.List;
@RestController
@RequestMapping("api/metro")
public class MetroServiceCheckController {
@Autowired
private TrainlineDisruptionRetrievalLogic logic;
@RequestMapping("disruptions")
@GetMapping
public List<Disruption> get(){
return logic.findDisruptions();
}
}
| package com.lichader.alfred.controller;
import com.lichader.alfred.logic.TrainlineDisruptionRetrievalLogic;
import com.lichader.alfred.metroapi.v3.model.Disruption;
import com.lichader.alfred.servant.MetroAlfred;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.web.bind.annotation.*;
import java.util.List;
@RestController
@RequestMapping("api/metro")
public class MetroServiceCheckController {
@Autowired
private TrainlineDisruptionRetrievalLogic logic;
@Autowired
private MetroAlfred alfred;
@RequestMapping(value = "disruptions", method = RequestMethod.GET)
public List<Disruption> get(){
return logic.findDisruptions();
}
@RequestMapping(value = "disruptions", method = RequestMethod.POST)
public void sendMessageToSlack(){
alfred.checkDisruption();
}
}
| Add a new POST endpoint to manually trigger sending message functionality | Add a new POST endpoint to manually trigger sending message functionality
| Java | mit | lichader/Alfred,lichader/Alfred | java | ## Code Before:
package com.lichader.alfred.controller;
import com.lichader.alfred.logic.TrainlineDisruptionRetrievalLogic;
import com.lichader.alfred.metroapi.v3.model.Disruption;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.web.bind.annotation.GetMapping;
import org.springframework.web.bind.annotation.RequestMapping;
import org.springframework.web.bind.annotation.RestController;
import java.util.List;
@RestController
@RequestMapping("api/metro")
public class MetroServiceCheckController {
@Autowired
private TrainlineDisruptionRetrievalLogic logic;
@RequestMapping("disruptions")
@GetMapping
public List<Disruption> get(){
return logic.findDisruptions();
}
}
## Instruction:
Add a new POST endpoint to manually trigger sending message functionality
## Code After:
package com.lichader.alfred.controller;
import com.lichader.alfred.logic.TrainlineDisruptionRetrievalLogic;
import com.lichader.alfred.metroapi.v3.model.Disruption;
import com.lichader.alfred.servant.MetroAlfred;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.web.bind.annotation.*;
import java.util.List;
@RestController
@RequestMapping("api/metro")
public class MetroServiceCheckController {
@Autowired
private TrainlineDisruptionRetrievalLogic logic;
@Autowired
private MetroAlfred alfred;
@RequestMapping(value = "disruptions", method = RequestMethod.GET)
public List<Disruption> get(){
return logic.findDisruptions();
}
@RequestMapping(value = "disruptions", method = RequestMethod.POST)
public void sendMessageToSlack(){
alfred.checkDisruption();
}
}
| package com.lichader.alfred.controller;
import com.lichader.alfred.logic.TrainlineDisruptionRetrievalLogic;
import com.lichader.alfred.metroapi.v3.model.Disruption;
+ import com.lichader.alfred.servant.MetroAlfred;
import org.springframework.beans.factory.annotation.Autowired;
- import org.springframework.web.bind.annotation.GetMapping;
? ^^^^^^^^^^
+ import org.springframework.web.bind.annotation.*;
? ^
- import org.springframework.web.bind.annotation.RequestMapping;
- import org.springframework.web.bind.annotation.RestController;
import java.util.List;
@RestController
@RequestMapping("api/metro")
public class MetroServiceCheckController {
@Autowired
private TrainlineDisruptionRetrievalLogic logic;
- @RequestMapping("disruptions")
- @GetMapping
+ @Autowired
+ private MetroAlfred alfred;
+
+ @RequestMapping(value = "disruptions", method = RequestMethod.GET)
public List<Disruption> get(){
return logic.findDisruptions();
}
+ @RequestMapping(value = "disruptions", method = RequestMethod.POST)
+ public void sendMessageToSlack(){
+ alfred.checkDisruption();
+ }
} | 15 | 0.6 | 10 | 5 |
4e454130da7eccd7844ccf624734d429ff4a1444 | app/database/seeds/ComponentTableSeeder.php | app/database/seeds/ComponentTableSeeder.php | <?php
use CachetHQ\Cachet\Models\Component;
use Illuminate\Database\Eloquent\Model;
use Illuminate\Database\Seeder;
class ComponentTableSeeder extends Seeder
{
/**
* Run the database seeding.
*
* @return void
*/
public function run()
{
Model::unguard();
$defaultComponents = [
[
"name" => "API",
"description" => "Used by third-parties to connect to us",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
], [
"name" => "Payments",
"description" => "Backed by Stripe",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
], [
"name" => "Website",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
],
];
Component::truncate();
foreach ($defaultComponents as $component) {
Component::create($component);
}
}
}
| <?php
use CachetHQ\Cachet\Models\Component;
use Illuminate\Database\Eloquent\Model;
use Illuminate\Database\Seeder;
class ComponentTableSeeder extends Seeder
{
/**
* Run the database seeding.
*
* @return void
*/
public function run()
{
Model::unguard();
$defaultComponents = [
[
"name" => "API",
"description" => "Used by third-parties to connect to us",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
"tags" => "",
], [
"name" => "Payments",
"description" => "Backed by Stripe",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
"tags" => "",
], [
"name" => "Website",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
"tags" => "",
],
];
Component::truncate();
foreach ($defaultComponents as $component) {
Component::create($component);
}
}
}
| Add tags to the component seeder | Add tags to the component seeder
| PHP | bsd-3-clause | katzien/Cachet,MicroWorldwide/Cachet,pellaeon/Cachet,billmn/Cachet,g-forks/Cachet,sapk/Cachet,Mebus/Cachet,alobintechnologies/Cachet,karaktaka/Cachet,clbn/Cachet,billmn/Cachet,NossaJureg/Cachet,eduardocruz/Cachet,eduardocruz/Cachet,elektropay/Cachet,wakermahmud/Cachet,sapk/Cachet,karaktaka/Cachet,NossaJureg/Cachet,rogerapras/Cachet,nxtreaming/Cachet,displayn/Cachet,SamuelMoraesF/Cachet,robglas/Cachet,cachethq/Cachet,bthiago/Cachet,ematthews/Cachet,elektropay/Cachet,anujaprasad/Hihat,gm-ah/Cachet,rogerapras/Cachet,coupej/Cachet,MicroWorldwide/Cachet,ephillipe/Cachet,pellaeon/Cachet,coupej/Cachet,CloudA/Cachet,CloudA/Cachet,h3zjp/Cachet,alobintechnologies/Cachet,anujaprasad/Hihat,ZengineChris/Cachet,elektropay/Cachet,gm-ah/Cachet,wngravette/Cachet,ephillipe/Cachet,NossaJureg/Cachet,n0mer/Cachet,whealmedia/Cachet,sapk/Cachet,MatheusRV/Cachet-Sandstorm,ematthews/Cachet,0x73/Cachet,brianjking/openshift-cachet,bthiago/Cachet,brianjking/openshift-cachet,ZengineChris/Cachet,Mebus/Cachet,chaseconey/Cachet,n0mer/Cachet,MatheusRV/Cachet-Sandstorm,MatheusRV/Cachet-Sandstorm,chaseconey/Cachet,0x73/Cachet,ZengineChris/Cachet,withings-sas/Cachet,Surventrix/Cachet,murendie/Cachet,chaseconey/Cachet,billmn/Cachet,SamuelMoraesF/Cachet,Mebus/Cachet,eduardocruz/Cachet,leegeng/Cachet,g-forks/Cachet,h3zjp/Cachet,wakermahmud/Cachet,bthiago/Cachet,n0mer/Cachet,clbn/Cachet,withings-sas/Cachet,murendie/Cachet,coupej/Cachet,coreation/Cachet,coreation/Cachet,MatheusRV/Cachet-Sandstorm,MatheusRV/Cachet-Sandstorm,karaktaka/Cachet,robglas/Cachet,h3zjp/Cachet,ephillipe/Cachet,whealmedia/Cachet,everpay/Cachet,wngravette/Cachet,anujaprasad/Hihat,SamuelMoraesF/Cachet,cachethq/Cachet,withings-sas/Cachet,everpay/Cachet,katzien/Cachet,rogerapras/Cachet,Surventrix/Cachet,pellaeon/Cachet,ematthews/Cachet,murendie/Cachet,leegeng/Cachet,g-forks/Cachet,displayn/Cachet,wakermahmud/Cachet,nxtreaming/Cachet,brianjking/openshift-cachet | php | ## Code Before:
<?php
use CachetHQ\Cachet\Models\Component;
use Illuminate\Database\Eloquent\Model;
use Illuminate\Database\Seeder;
class ComponentTableSeeder extends Seeder
{
/**
* Run the database seeding.
*
* @return void
*/
public function run()
{
Model::unguard();
$defaultComponents = [
[
"name" => "API",
"description" => "Used by third-parties to connect to us",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
], [
"name" => "Payments",
"description" => "Backed by Stripe",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
], [
"name" => "Website",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
],
];
Component::truncate();
foreach ($defaultComponents as $component) {
Component::create($component);
}
}
}
## Instruction:
Add tags to the component seeder
## Code After:
<?php
use CachetHQ\Cachet\Models\Component;
use Illuminate\Database\Eloquent\Model;
use Illuminate\Database\Seeder;
class ComponentTableSeeder extends Seeder
{
/**
* Run the database seeding.
*
* @return void
*/
public function run()
{
Model::unguard();
$defaultComponents = [
[
"name" => "API",
"description" => "Used by third-parties to connect to us",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
"tags" => "",
], [
"name" => "Payments",
"description" => "Backed by Stripe",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
"tags" => "",
], [
"name" => "Website",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
"tags" => "",
],
];
Component::truncate();
foreach ($defaultComponents as $component) {
Component::create($component);
}
}
}
| <?php
use CachetHQ\Cachet\Models\Component;
use Illuminate\Database\Eloquent\Model;
use Illuminate\Database\Seeder;
class ComponentTableSeeder extends Seeder
{
/**
* Run the database seeding.
*
* @return void
*/
public function run()
{
Model::unguard();
$defaultComponents = [
[
"name" => "API",
"description" => "Used by third-parties to connect to us",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
+ "tags" => "",
], [
"name" => "Payments",
"description" => "Backed by Stripe",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
+ "tags" => "",
], [
"name" => "Website",
"status" => 1,
"user_id" => 1,
"order" => 0,
"group_id" => 0,
"link" => "",
+ "tags" => "",
],
];
Component::truncate();
foreach ($defaultComponents as $component) {
Component::create($component);
}
}
} | 3 | 0.058824 | 3 | 0 |
34c126183f75c7d855a58ba98cc53d9f676a25bb | src/app.js | src/app.js | let express = require('express');
let bodyParser = require('body-parser');
let index = require('./routes/index');
let app = express();
app.use(bodyParser.json());
app.use(bodyParser.urlencoded({ extended: false }));
app.use('/', index);
app.use(function(req, res, next) {
let err = new Error('Not Found');
err.status = 404;
next(err);
});
app.use(function(err, req, res) {
res.status(err.status || 500);
res.send(err.message || err);
});
module.exports = app;
| let express = require('express');
let bodyParser = require('body-parser');
let index = require('./routes/index');
let app = express();
app.use(bodyParser.json());
app.use(function (error, req, res, next) {
if (error instanceof SyntaxError) {
res.status(400);
res.send({message: "Malformed JSON document"});
} else {
next();
}
});
app.use(bodyParser.urlencoded({extended: false}));
app.use('/', index);
app.use(function (req, res, next) {
let err = new Error('Not Found');
err.status = 404;
next(err);
});
app.use(function (err, req, res, next) {
res.status(err.status || 500);
res.send(err.message || err);
});
module.exports = app;
| Add handler for non-json requests | Add handler for non-json requests
| JavaScript | mit | jlundan/spacy-nodejs-alpine,jlundan/spacy-nodejs-alpine | javascript | ## Code Before:
let express = require('express');
let bodyParser = require('body-parser');
let index = require('./routes/index');
let app = express();
app.use(bodyParser.json());
app.use(bodyParser.urlencoded({ extended: false }));
app.use('/', index);
app.use(function(req, res, next) {
let err = new Error('Not Found');
err.status = 404;
next(err);
});
app.use(function(err, req, res) {
res.status(err.status || 500);
res.send(err.message || err);
});
module.exports = app;
## Instruction:
Add handler for non-json requests
## Code After:
let express = require('express');
let bodyParser = require('body-parser');
let index = require('./routes/index');
let app = express();
app.use(bodyParser.json());
app.use(function (error, req, res, next) {
if (error instanceof SyntaxError) {
res.status(400);
res.send({message: "Malformed JSON document"});
} else {
next();
}
});
app.use(bodyParser.urlencoded({extended: false}));
app.use('/', index);
app.use(function (req, res, next) {
let err = new Error('Not Found');
err.status = 404;
next(err);
});
app.use(function (err, req, res, next) {
res.status(err.status || 500);
res.send(err.message || err);
});
module.exports = app;
| let express = require('express');
let bodyParser = require('body-parser');
let index = require('./routes/index');
let app = express();
app.use(bodyParser.json());
+ app.use(function (error, req, res, next) {
+ if (error instanceof SyntaxError) {
+ res.status(400);
+ res.send({message: "Malformed JSON document"});
+ } else {
+ next();
+ }
+ });
- app.use(bodyParser.urlencoded({ extended: false }));
? - -
+ app.use(bodyParser.urlencoded({extended: false}));
app.use('/', index);
- app.use(function(req, res, next) {
+ app.use(function (req, res, next) {
? +
- let err = new Error('Not Found');
+ let err = new Error('Not Found');
? ++
- err.status = 404;
+ err.status = 404;
? ++
- next(err);
+ next(err);
? ++
});
- app.use(function(err, req, res) {
+ app.use(function (err, req, res, next) {
? + ++++++
- res.status(err.status || 500);
+ res.status(err.status || 500);
? ++
- res.send(err.message || err);
+ res.send(err.message || err);
? ++
});
module.exports = app; | 24 | 1.043478 | 16 | 8 |
a410bfe7be49c8c2e879685a5dea6d60110d7b73 | package.json | package.json | {
"name": "tempoiq",
"description": "TempoIQ HTTP NodeJS Client",
"version": "1.0.1",
"author": "TempoIQ Tech <software@tempoiq.com>",
"keywords": ["tempoiq", "time-series", "database"],
"main": "lib/tempoiq.js",
"devDependencies": {
"mocha": "1.21.4"
},
"scripts": {
"test": "mocha"
},
"repository": {
"type": "git",
"url": "http://github.com/tempoiq/tempoiq-node-js.git"
}
}
| {
"name": "tempoiq",
"description": "TempoIQ HTTP NodeJS Client",
"version": "1.0.1",
"author": "TempoIQ <support@tempoiq.com>",
"keywords": ["tempoiq", "time-series", "database"],
"main": "lib/tempoiq.js",
"devDependencies": {
"mocha": "1.21.4"
},
"scripts": {
"test": "mocha"
},
"repository": {
"type": "git",
"url": "http://github.com/tempoiq/tempoiq-node-js.git"
}
}
| Update contact email to support@ | Update contact email to support@
| JSON | mit | meshulam/tempoiq-node-js,TempoIQ/tempoiq-node-js | json | ## Code Before:
{
"name": "tempoiq",
"description": "TempoIQ HTTP NodeJS Client",
"version": "1.0.1",
"author": "TempoIQ Tech <software@tempoiq.com>",
"keywords": ["tempoiq", "time-series", "database"],
"main": "lib/tempoiq.js",
"devDependencies": {
"mocha": "1.21.4"
},
"scripts": {
"test": "mocha"
},
"repository": {
"type": "git",
"url": "http://github.com/tempoiq/tempoiq-node-js.git"
}
}
## Instruction:
Update contact email to support@
## Code After:
{
"name": "tempoiq",
"description": "TempoIQ HTTP NodeJS Client",
"version": "1.0.1",
"author": "TempoIQ <support@tempoiq.com>",
"keywords": ["tempoiq", "time-series", "database"],
"main": "lib/tempoiq.js",
"devDependencies": {
"mocha": "1.21.4"
},
"scripts": {
"test": "mocha"
},
"repository": {
"type": "git",
"url": "http://github.com/tempoiq/tempoiq-node-js.git"
}
}
| {
"name": "tempoiq",
"description": "TempoIQ HTTP NodeJS Client",
"version": "1.0.1",
- "author": "TempoIQ Tech <software@tempoiq.com>",
? ----- ^ ----
+ "author": "TempoIQ <support@tempoiq.com>",
? +++ ^
"keywords": ["tempoiq", "time-series", "database"],
"main": "lib/tempoiq.js",
"devDependencies": {
"mocha": "1.21.4"
},
"scripts": {
"test": "mocha"
},
"repository": {
"type": "git",
"url": "http://github.com/tempoiq/tempoiq-node-js.git"
}
} | 2 | 0.111111 | 1 | 1 |
5a84ef8ee5bfaf51ef1eafd7128473aaad69669e | lib/alg/dfs.js | lib/alg/dfs.js | var _ = require("../lodash");
module.exports = dfs;
/*
* A helper that preforms a pre- or post-order traversal on the input graph
* and returns the nodes in the order they were visited. This algorithm treats
* the input as undirected.
*
* Order must be one of "pre" or "post".
*/
function dfs(g, vs, order) {
if (!_.isArray(vs)) {
vs = [vs];
}
var acc = [],
visited = {};
_.each(vs, function(v) {
if (!g.hasNode(v)) {
throw new Error("Graph does not have node: " + v);
}
doDfs(g, v, order === "post", visited, acc);
});
return acc;
}
function doDfs(g, v, postorder, visited, acc) {
if (!_.has(visited, v)) {
visited[v] = true;
if (!postorder) { acc.push(v); }
_.each(g.successors(v), function(w) {
doDfs(g, w, postorder, visited, acc);
});
if (postorder) { acc.push(v); }
}
}
| var _ = require("../lodash");
module.exports = dfs;
/*
* A helper that preforms a pre- or post-order traversal on the input graph
* and returns the nodes in the order they were visited. If the graph is
* undirected then this algorithm will navigate using neighbors. If the graph
* is directed then this algorithm will navigate using successors.
*
* Order must be one of "pre" or "post".
*/
function dfs(g, vs, order) {
if (!_.isArray(vs)) {
vs = [vs];
}
var navigation = (g.isDirected() ? g.successors : g.neighbors).bind(g);
var acc = [],
visited = {};
_.each(vs, function(v) {
if (!g.hasNode(v)) {
throw new Error("Graph does not have node: " + v);
}
doDfs(g, v, order === "post", visited, navigation, acc);
});
return acc;
}
function doDfs(g, v, postorder, visited, navigation, acc) {
if (!_.has(visited, v)) {
visited[v] = true;
if (!postorder) { acc.push(v); }
_.each(navigation(v), function(w) {
doDfs(g, w, postorder, visited, navigation, acc);
});
if (postorder) { acc.push(v); }
}
}
| Use neighbors to perform DFS on undirected graphs | Use neighbors to perform DFS on undirected graphs
| JavaScript | mit | dagrejs/graphlib,cpettitt/graphlib,dagrejs/graphlib,dagrejs/graphlib,cpettitt/graphlib | javascript | ## Code Before:
var _ = require("../lodash");
module.exports = dfs;
/*
* A helper that preforms a pre- or post-order traversal on the input graph
* and returns the nodes in the order they were visited. This algorithm treats
* the input as undirected.
*
* Order must be one of "pre" or "post".
*/
function dfs(g, vs, order) {
if (!_.isArray(vs)) {
vs = [vs];
}
var acc = [],
visited = {};
_.each(vs, function(v) {
if (!g.hasNode(v)) {
throw new Error("Graph does not have node: " + v);
}
doDfs(g, v, order === "post", visited, acc);
});
return acc;
}
function doDfs(g, v, postorder, visited, acc) {
if (!_.has(visited, v)) {
visited[v] = true;
if (!postorder) { acc.push(v); }
_.each(g.successors(v), function(w) {
doDfs(g, w, postorder, visited, acc);
});
if (postorder) { acc.push(v); }
}
}
## Instruction:
Use neighbors to perform DFS on undirected graphs
## Code After:
var _ = require("../lodash");
module.exports = dfs;
/*
* A helper that preforms a pre- or post-order traversal on the input graph
* and returns the nodes in the order they were visited. If the graph is
* undirected then this algorithm will navigate using neighbors. If the graph
* is directed then this algorithm will navigate using successors.
*
* Order must be one of "pre" or "post".
*/
function dfs(g, vs, order) {
if (!_.isArray(vs)) {
vs = [vs];
}
var navigation = (g.isDirected() ? g.successors : g.neighbors).bind(g);
var acc = [],
visited = {};
_.each(vs, function(v) {
if (!g.hasNode(v)) {
throw new Error("Graph does not have node: " + v);
}
doDfs(g, v, order === "post", visited, navigation, acc);
});
return acc;
}
function doDfs(g, v, postorder, visited, navigation, acc) {
if (!_.has(visited, v)) {
visited[v] = true;
if (!postorder) { acc.push(v); }
_.each(navigation(v), function(w) {
doDfs(g, w, postorder, visited, navigation, acc);
});
if (postorder) { acc.push(v); }
}
}
| var _ = require("../lodash");
module.exports = dfs;
/*
* A helper that preforms a pre- or post-order traversal on the input graph
- * and returns the nodes in the order they were visited. This algorithm treats
? ^ -----------------
+ * and returns the nodes in the order they were visited. If the graph is
? ^^^^ ++++++++
- * the input as undirected.
+ * undirected then this algorithm will navigate using neighbors. If the graph
+ * is directed then this algorithm will navigate using successors.
*
* Order must be one of "pre" or "post".
*/
function dfs(g, vs, order) {
if (!_.isArray(vs)) {
vs = [vs];
}
+ var navigation = (g.isDirected() ? g.successors : g.neighbors).bind(g);
+
var acc = [],
visited = {};
_.each(vs, function(v) {
if (!g.hasNode(v)) {
throw new Error("Graph does not have node: " + v);
}
- doDfs(g, v, order === "post", visited, acc);
+ doDfs(g, v, order === "post", visited, navigation, acc);
? ++++++++++++
});
return acc;
}
- function doDfs(g, v, postorder, visited, acc) {
+ function doDfs(g, v, postorder, visited, navigation, acc) {
? ++++++++++++
if (!_.has(visited, v)) {
visited[v] = true;
if (!postorder) { acc.push(v); }
- _.each(g.successors(v), function(w) {
? ^^^^^^^^ ^^
+ _.each(navigation(v), function(w) {
? ++++ ^^^ ^
- doDfs(g, w, postorder, visited, acc);
+ doDfs(g, w, postorder, visited, navigation, acc);
? ++++++++++++
});
if (postorder) { acc.push(v); }
}
} | 15 | 0.384615 | 9 | 6 |
1b5f10b9c975b8a055d830e8ac5e112dd059c7d5 | electron/index.js | electron/index.js | var menubar = require('menubar');
var ipc = require('ipc');
var shell = require('shell')
var mb = menubar({
dir: __dirname,
width: 280,
height: 87
});
mb.on('ready', function ready () {
require('../server');
})
ipc.on('browser', function browser (ev) {
shell.openExternal('http://localhost:3000')
})
ipc.on('terminate', function terminate (ev) {
canQuit = true
mb.app.terminate()
})
ipc.on('homepage', function homepage (ev) {
shell.openExternal('http://lightning-viz.org')
}) | var menubar = require('menubar');
var ipc = require('ipc');
var shell = require('shell')
var mb = menubar({
dir: __dirname,
width: 280,
height: 87,
icon: __dirname + '/Icon.png'
});
mb.on('ready', function ready () {
require('../server');
})
ipc.on('browser', function browser (ev) {
shell.openExternal('http://localhost:3000')
})
ipc.on('terminate', function terminate (ev) {
canQuit = true
mb.app.terminate()
})
ipc.on('homepage', function homepage (ev) {
shell.openExternal('http://lightning-viz.org')
}) | Fix for updated default icon name | Fix for updated default icon name
| JavaScript | mit | origingod/lightning,karissa/lightning,lightning-viz/lightning,cournape/lightning,cournape/lightning,lightning-viz/lightning,origingod/lightning,karissa/lightning | javascript | ## Code Before:
var menubar = require('menubar');
var ipc = require('ipc');
var shell = require('shell')
var mb = menubar({
dir: __dirname,
width: 280,
height: 87
});
mb.on('ready', function ready () {
require('../server');
})
ipc.on('browser', function browser (ev) {
shell.openExternal('http://localhost:3000')
})
ipc.on('terminate', function terminate (ev) {
canQuit = true
mb.app.terminate()
})
ipc.on('homepage', function homepage (ev) {
shell.openExternal('http://lightning-viz.org')
})
## Instruction:
Fix for updated default icon name
## Code After:
var menubar = require('menubar');
var ipc = require('ipc');
var shell = require('shell')
var mb = menubar({
dir: __dirname,
width: 280,
height: 87,
icon: __dirname + '/Icon.png'
});
mb.on('ready', function ready () {
require('../server');
})
ipc.on('browser', function browser (ev) {
shell.openExternal('http://localhost:3000')
})
ipc.on('terminate', function terminate (ev) {
canQuit = true
mb.app.terminate()
})
ipc.on('homepage', function homepage (ev) {
shell.openExternal('http://lightning-viz.org')
}) | var menubar = require('menubar');
var ipc = require('ipc');
var shell = require('shell')
var mb = menubar({
dir: __dirname,
width: 280,
- height: 87
+ height: 87,
? +
+ icon: __dirname + '/Icon.png'
});
mb.on('ready', function ready () {
require('../server');
})
ipc.on('browser', function browser (ev) {
shell.openExternal('http://localhost:3000')
})
ipc.on('terminate', function terminate (ev) {
canQuit = true
mb.app.terminate()
})
ipc.on('homepage', function homepage (ev) {
shell.openExternal('http://lightning-viz.org')
}) | 3 | 0.115385 | 2 | 1 |
11b9a99bf12f16ededdf79af2e46034e96857659 | WebContent/META-INF/context.xml | WebContent/META-INF/context.xml | <?xml version="1.0" encoding="UTF-8"?>
<Context path="/courses">
<Resource
name="jdbc/coursedb"
auth="Container"
type="javax.sql.DataSource"
factory="org.apache.tomcat.jdbc.pool.DataSourceFactory"
driverClassName="com.mysql.jdbc.Driver"
url="jdbc:mysql://devproserv.com:54406/coursedb"
username="javauser"
password="zSxcvb"
initialSize="10"
maxActive="20"
maxIdle="10"
minIdle="5"
maxWait="10000"
testOnBorrow="true"
testOnConnect="true"
testOnReturn="true"
testWhileIdle="true"
validationQuery="SELECT 1"
validationInterval="60000"
minEvictableIdleTimeMillis="60000"
timeBetweenEvictionRunsMillis="60000"
removeAbandoned="true"
removeAbandonedTimeout="60"
logAbandoned="true"
/>
</Context>
| <?xml version="1.0" encoding="UTF-8"?>
<Context path="/courses">
<Resource
name="jdbc/coursedb"
auth="Container"
type="javax.sql.DataSource"
factory="org.apache.tomcat.jdbc.pool.DataSourceFactory"
driverClassName="com.mysql.jdbc.Driver"
url="jdbc:mysql://devproserv.com:54406/coursedb"
connectionProperties="useUnicode=yes;characterEncoding=utf8;"
username="javauser"
password="zSxcvb"
initialSize="10"
maxActive="20"
maxIdle="10"
minIdle="5"
maxWait="10000"
testOnBorrow="true"
testOnConnect="true"
testOnReturn="true"
testWhileIdle="true"
validationQuery="SELECT 1"
validationInterval="60000"
minEvictableIdleTimeMillis="60000"
timeBetweenEvictionRunsMillis="60000"
removeAbandoned="true"
removeAbandonedTimeout="60"
logAbandoned="true"
/>
</Context>
| Set encoding UTF8 in the connection pool | Set encoding UTF8 in the connection pool
| XML | mit | Vovas11/courses,Vovas11/courses,Vovas11/courses | xml | ## Code Before:
<?xml version="1.0" encoding="UTF-8"?>
<Context path="/courses">
<Resource
name="jdbc/coursedb"
auth="Container"
type="javax.sql.DataSource"
factory="org.apache.tomcat.jdbc.pool.DataSourceFactory"
driverClassName="com.mysql.jdbc.Driver"
url="jdbc:mysql://devproserv.com:54406/coursedb"
username="javauser"
password="zSxcvb"
initialSize="10"
maxActive="20"
maxIdle="10"
minIdle="5"
maxWait="10000"
testOnBorrow="true"
testOnConnect="true"
testOnReturn="true"
testWhileIdle="true"
validationQuery="SELECT 1"
validationInterval="60000"
minEvictableIdleTimeMillis="60000"
timeBetweenEvictionRunsMillis="60000"
removeAbandoned="true"
removeAbandonedTimeout="60"
logAbandoned="true"
/>
</Context>
## Instruction:
Set encoding UTF8 in the connection pool
## Code After:
<?xml version="1.0" encoding="UTF-8"?>
<Context path="/courses">
<Resource
name="jdbc/coursedb"
auth="Container"
type="javax.sql.DataSource"
factory="org.apache.tomcat.jdbc.pool.DataSourceFactory"
driverClassName="com.mysql.jdbc.Driver"
url="jdbc:mysql://devproserv.com:54406/coursedb"
connectionProperties="useUnicode=yes;characterEncoding=utf8;"
username="javauser"
password="zSxcvb"
initialSize="10"
maxActive="20"
maxIdle="10"
minIdle="5"
maxWait="10000"
testOnBorrow="true"
testOnConnect="true"
testOnReturn="true"
testWhileIdle="true"
validationQuery="SELECT 1"
validationInterval="60000"
minEvictableIdleTimeMillis="60000"
timeBetweenEvictionRunsMillis="60000"
removeAbandoned="true"
removeAbandonedTimeout="60"
logAbandoned="true"
/>
</Context>
| <?xml version="1.0" encoding="UTF-8"?>
<Context path="/courses">
<Resource
name="jdbc/coursedb"
auth="Container"
type="javax.sql.DataSource"
factory="org.apache.tomcat.jdbc.pool.DataSourceFactory"
driverClassName="com.mysql.jdbc.Driver"
url="jdbc:mysql://devproserv.com:54406/coursedb"
+ connectionProperties="useUnicode=yes;characterEncoding=utf8;"
username="javauser"
password="zSxcvb"
initialSize="10"
maxActive="20"
maxIdle="10"
minIdle="5"
maxWait="10000"
testOnBorrow="true"
testOnConnect="true"
testOnReturn="true"
testWhileIdle="true"
validationQuery="SELECT 1"
validationInterval="60000"
minEvictableIdleTimeMillis="60000"
timeBetweenEvictionRunsMillis="60000"
removeAbandoned="true"
removeAbandonedTimeout="60"
logAbandoned="true"
/>
</Context> | 1 | 0.026316 | 1 | 0 |
97b7681990c3ea7ecedf15a14b181042cc423a28 | README.md | README.md |
Tools to help train the donkey car models.
mktraining: generates a synthetic set of images similar to what a rpi camera might capture.
horizontal FOV 60 degrees; resolution 320x240; random confetti on the floor for each image.
Libraries needed to build: GLFW and GLEW.
blur: blur a PNG image based on a mask given in another PNG image.
All of these tools also want "stb" to be a checkout of github.com/nothings/stb as a peer
to these directories.
|
Tools to help train the donkey car models.
`mktraining`: generates a synthetic set of images similar to what a rpi camera might capture.
horizontal FOV 60 degrees; resolution 320x240; random confetti on the floor for each image.
Libraries needed to build: GLFW and GLEW.
`blur`: blur a PNG image based on a mask given in another PNG image.
`teensy_32_carrier`: KiCad design files for the servo and receiver I/O board for the Teensy 3.2
microcontroller.
All of these tools also want "stb" to be a checkout of github.com/nothings/stb as a peer
to these directories.
| Add information about teensy board files | Add information about teensy board files | Markdown | mit | jwatte/donkey_racing,jwatte/donkey_racing,jwatte/donkey_racing,jwatte/donkey_racing | markdown | ## Code Before:
Tools to help train the donkey car models.
mktraining: generates a synthetic set of images similar to what a rpi camera might capture.
horizontal FOV 60 degrees; resolution 320x240; random confetti on the floor for each image.
Libraries needed to build: GLFW and GLEW.
blur: blur a PNG image based on a mask given in another PNG image.
All of these tools also want "stb" to be a checkout of github.com/nothings/stb as a peer
to these directories.
## Instruction:
Add information about teensy board files
## Code After:
Tools to help train the donkey car models.
`mktraining`: generates a synthetic set of images similar to what a rpi camera might capture.
horizontal FOV 60 degrees; resolution 320x240; random confetti on the floor for each image.
Libraries needed to build: GLFW and GLEW.
`blur`: blur a PNG image based on a mask given in another PNG image.
`teensy_32_carrier`: KiCad design files for the servo and receiver I/O board for the Teensy 3.2
microcontroller.
All of these tools also want "stb" to be a checkout of github.com/nothings/stb as a peer
to these directories.
|
Tools to help train the donkey car models.
- mktraining: generates a synthetic set of images similar to what a rpi camera might capture.
+ `mktraining`: generates a synthetic set of images similar to what a rpi camera might capture.
? + +
horizontal FOV 60 degrees; resolution 320x240; random confetti on the floor for each image.
Libraries needed to build: GLFW and GLEW.
- blur: blur a PNG image based on a mask given in another PNG image.
+ `blur`: blur a PNG image based on a mask given in another PNG image.
? + +
+
+ `teensy_32_carrier`: KiCad design files for the servo and receiver I/O board for the Teensy 3.2
+ microcontroller.
All of these tools also want "stb" to be a checkout of github.com/nothings/stb as a peer
to these directories. | 7 | 0.636364 | 5 | 2 |
bd0b891175fedeb86bc7ba20642bad350cac843b | app/models/user.rb | app/models/user.rb | class User < ActiveRecord::Base
attr_accessor :password
validates_confirmation_of :password
before_save :encrypt_password
belongs_to :region
has_many :sales
has_many :subordinates, class_name: "User", foreign_key: "manager_id"
belongs_to :manager, class_name: "User"
after_initialize :init
def init
self.first_name ||= ""
self.last_name ||= ""
self.username ||= ""
self.password ||= ""
self.role ||= ""
self.manager_id ||= 0.0
self.region_id ||= 0.0
end
def has_role?(role_name)
self.role == role_name
end
def encrypt_password
self.password_salt = BCrypt::Engine.generate_salt
self.password_hash = BCrypt::Engine.hash_secret(password, password_salt)
end
def self.authenticate(email, password)
user = User.where(email: email).first
if user && user.password_hash == BCrypt::Engine.hash_secret(password, user.password_salt)
user
else
nil
end
end
end
| class User < ActiveRecord::Base
attr_accessor :password
validates_confirmation_of :password
before_save :encrypt_password
belongs_to :region
has_many :sales
has_many :subordinates, class_name: "User", foreign_key: "manager_id"
belongs_to :manager, class_name: "User"
after_initialize :init
def init
self.first_name ||= ""
self.last_name ||= ""
self.password ||= ""
self.role ||= ""
self.manager_id ||= 0.0
self.region_id ||= 0.0
end
def has_role?(role_name)
self.role == role_name
end
def encrypt_password
self.password_salt = BCrypt::Engine.generate_salt
self.password_hash = BCrypt::Engine.hash_secret(password, password_salt)
end
def self.authenticate(email, password)
user = User.where(email: email).first
if user && user.password_hash == BCrypt::Engine.hash_secret(password, user.password_salt)
user
else
nil
end
end
end
| Change models to reflect the changes in the db | Change models to reflect the changes in the db
| Ruby | mit | keithrfung/sound-sales,keithrfung/sound-sales,keithrfung/sound-sales | ruby | ## Code Before:
class User < ActiveRecord::Base
attr_accessor :password
validates_confirmation_of :password
before_save :encrypt_password
belongs_to :region
has_many :sales
has_many :subordinates, class_name: "User", foreign_key: "manager_id"
belongs_to :manager, class_name: "User"
after_initialize :init
def init
self.first_name ||= ""
self.last_name ||= ""
self.username ||= ""
self.password ||= ""
self.role ||= ""
self.manager_id ||= 0.0
self.region_id ||= 0.0
end
def has_role?(role_name)
self.role == role_name
end
def encrypt_password
self.password_salt = BCrypt::Engine.generate_salt
self.password_hash = BCrypt::Engine.hash_secret(password, password_salt)
end
def self.authenticate(email, password)
user = User.where(email: email).first
if user && user.password_hash == BCrypt::Engine.hash_secret(password, user.password_salt)
user
else
nil
end
end
end
## Instruction:
Change models to reflect the changes in the db
## Code After:
class User < ActiveRecord::Base
attr_accessor :password
validates_confirmation_of :password
before_save :encrypt_password
belongs_to :region
has_many :sales
has_many :subordinates, class_name: "User", foreign_key: "manager_id"
belongs_to :manager, class_name: "User"
after_initialize :init
def init
self.first_name ||= ""
self.last_name ||= ""
self.password ||= ""
self.role ||= ""
self.manager_id ||= 0.0
self.region_id ||= 0.0
end
def has_role?(role_name)
self.role == role_name
end
def encrypt_password
self.password_salt = BCrypt::Engine.generate_salt
self.password_hash = BCrypt::Engine.hash_secret(password, password_salt)
end
def self.authenticate(email, password)
user = User.where(email: email).first
if user && user.password_hash == BCrypt::Engine.hash_secret(password, user.password_salt)
user
else
nil
end
end
end
| class User < ActiveRecord::Base
attr_accessor :password
validates_confirmation_of :password
before_save :encrypt_password
belongs_to :region
has_many :sales
has_many :subordinates, class_name: "User", foreign_key: "manager_id"
belongs_to :manager, class_name: "User"
after_initialize :init
def init
self.first_name ||= ""
self.last_name ||= ""
- self.username ||= ""
self.password ||= ""
self.role ||= ""
self.manager_id ||= 0.0
self.region_id ||= 0.0
end
def has_role?(role_name)
self.role == role_name
end
def encrypt_password
self.password_salt = BCrypt::Engine.generate_salt
self.password_hash = BCrypt::Engine.hash_secret(password, password_salt)
end
def self.authenticate(email, password)
user = User.where(email: email).first
if user && user.password_hash == BCrypt::Engine.hash_secret(password, user.password_salt)
user
else
nil
end
end
end | 1 | 0.026316 | 0 | 1 |
950c1b327c9a85cd82c1e2b6174ad0b7d4ab78a8 | app/partials/landing_page.html | app/partials/landing_page.html | This is the landing page! | <div class="container">
<div class="starter-template">
<h1>Wait and Eat</h1>
<p class="lead">The smart restaurant waiting list.</p>
</div>
</div><!-- /.container --> | Add title and slogan to landing page | Add title and slogan to landing page
| HTML | mit | himanz/waittime,himanz/waittime,himanz/waittime | html | ## Code Before:
This is the landing page!
## Instruction:
Add title and slogan to landing page
## Code After:
<div class="container">
<div class="starter-template">
<h1>Wait and Eat</h1>
<p class="lead">The smart restaurant waiting list.</p>
</div>
</div><!-- /.container --> | - This is the landing page!
+ <div class="container">
+
+ <div class="starter-template">
+ <h1>Wait and Eat</h1>
+ <p class="lead">The smart restaurant waiting list.</p>
+ </div>
+
+ </div><!-- /.container --> | 9 | 9 | 8 | 1 |
f9cb73670966d7cb3ade12056c92c479404cbb07 | pythran/tests/test_cython.py | pythran/tests/test_cython.py | import os
import unittest
class TestCython(unittest.TestCase):
pass
def add_test(name, runner, target):
setattr(TestCython, "test_" + name, lambda s: runner(s, target))
try:
import Cython
import glob
import sys
targets = glob.glob(os.path.join(os.path.dirname(__file__), "cython", "setup_*.py"))
sys.path.append(os.path.join(os.path.dirname(__file__), "cython"))
for target in targets:
def runner(self, target):
cwd = os.getcwd()
try:
os.chdir(os.path.dirname(target))
exec(open(os.path.basename(target)).read())
except:
raise
finally:
os.chdir(cwd)
name, _ = os.path.splitext(os.path.basename(target))
add_test(name, runner, target)
except ImportError:
pass
| import os
import unittest
class TestCython(unittest.TestCase):
pass
# Needs to wait unil cython supports pythran new builtins naming
#def add_test(name, runner, target):
# setattr(TestCython, "test_" + name, lambda s: runner(s, target))
#
#try:
# import Cython
# import glob
# import sys
# targets = glob.glob(os.path.join(os.path.dirname(__file__), "cython", "setup_*.py"))
# sys.path.append(os.path.join(os.path.dirname(__file__), "cython"))
#
# for target in targets:
# def runner(self, target):
# cwd = os.getcwd()
# try:
# os.chdir(os.path.dirname(target))
# exec(open(os.path.basename(target)).read())
# except:
# raise
# finally:
# os.chdir(cwd)
# name, _ = os.path.splitext(os.path.basename(target))
# add_test(name, runner, target)
#
#
#except ImportError:
# pass
| Disable Cython test until Cython support new pythonic layout | Disable Cython test until Cython support new pythonic layout
| Python | bsd-3-clause | serge-sans-paille/pythran,serge-sans-paille/pythran,pombredanne/pythran,pombredanne/pythran,pombredanne/pythran | python | ## Code Before:
import os
import unittest
class TestCython(unittest.TestCase):
pass
def add_test(name, runner, target):
setattr(TestCython, "test_" + name, lambda s: runner(s, target))
try:
import Cython
import glob
import sys
targets = glob.glob(os.path.join(os.path.dirname(__file__), "cython", "setup_*.py"))
sys.path.append(os.path.join(os.path.dirname(__file__), "cython"))
for target in targets:
def runner(self, target):
cwd = os.getcwd()
try:
os.chdir(os.path.dirname(target))
exec(open(os.path.basename(target)).read())
except:
raise
finally:
os.chdir(cwd)
name, _ = os.path.splitext(os.path.basename(target))
add_test(name, runner, target)
except ImportError:
pass
## Instruction:
Disable Cython test until Cython support new pythonic layout
## Code After:
import os
import unittest
class TestCython(unittest.TestCase):
pass
# Needs to wait unil cython supports pythran new builtins naming
#def add_test(name, runner, target):
# setattr(TestCython, "test_" + name, lambda s: runner(s, target))
#
#try:
# import Cython
# import glob
# import sys
# targets = glob.glob(os.path.join(os.path.dirname(__file__), "cython", "setup_*.py"))
# sys.path.append(os.path.join(os.path.dirname(__file__), "cython"))
#
# for target in targets:
# def runner(self, target):
# cwd = os.getcwd()
# try:
# os.chdir(os.path.dirname(target))
# exec(open(os.path.basename(target)).read())
# except:
# raise
# finally:
# os.chdir(cwd)
# name, _ = os.path.splitext(os.path.basename(target))
# add_test(name, runner, target)
#
#
#except ImportError:
# pass
| import os
import unittest
class TestCython(unittest.TestCase):
pass
+ # Needs to wait unil cython supports pythran new builtins naming
+ #def add_test(name, runner, target):
+ # setattr(TestCython, "test_" + name, lambda s: runner(s, target))
+ #
+ #try:
+ # import Cython
+ # import glob
+ # import sys
+ # targets = glob.glob(os.path.join(os.path.dirname(__file__), "cython", "setup_*.py"))
+ # sys.path.append(os.path.join(os.path.dirname(__file__), "cython"))
+ #
+ # for target in targets:
+ # def runner(self, target):
+ # cwd = os.getcwd()
+ # try:
+ # os.chdir(os.path.dirname(target))
+ # exec(open(os.path.basename(target)).read())
+ # except:
+ # raise
+ # finally:
+ # os.chdir(cwd)
+ # name, _ = os.path.splitext(os.path.basename(target))
+ # add_test(name, runner, target)
+ #
+ #
+ #except ImportError:
+ # pass
- def add_test(name, runner, target):
- setattr(TestCython, "test_" + name, lambda s: runner(s, target))
-
- try:
- import Cython
- import glob
- import sys
- targets = glob.glob(os.path.join(os.path.dirname(__file__), "cython", "setup_*.py"))
- sys.path.append(os.path.join(os.path.dirname(__file__), "cython"))
-
- for target in targets:
- def runner(self, target):
- cwd = os.getcwd()
- try:
- os.chdir(os.path.dirname(target))
- exec(open(os.path.basename(target)).read())
- except:
- raise
- finally:
- os.chdir(cwd)
- name, _ = os.path.splitext(os.path.basename(target))
- add_test(name, runner, target)
-
-
- except ImportError:
- pass
- | 54 | 1.636364 | 27 | 27 |
d0f4ac416d496670f043af0edfaae9548d92f13c | CHANGELOG.md | CHANGELOG.md |
All notable changes to this project will be documented in this file.
This project adheres to [Semantic Versioning](http://semver.org/).
**1.0.0**
- Initial release
|
All notable changes to this project will be documented in this file.
This project adheres to [Semantic Versioning](http://semver.org/).
**1.1.0**
- Make `toArray` operate recursively to flatten nested structures. PR#1
**1.0.0**
- Initial release
| Update changelog for v1.1.0 release | Update changelog for v1.1.0 release
| Markdown | mit | equip/structure | markdown | ## Code Before:
All notable changes to this project will be documented in this file.
This project adheres to [Semantic Versioning](http://semver.org/).
**1.0.0**
- Initial release
## Instruction:
Update changelog for v1.1.0 release
## Code After:
All notable changes to this project will be documented in this file.
This project adheres to [Semantic Versioning](http://semver.org/).
**1.1.0**
- Make `toArray` operate recursively to flatten nested structures. PR#1
**1.0.0**
- Initial release
|
All notable changes to this project will be documented in this file.
This project adheres to [Semantic Versioning](http://semver.org/).
+ **1.1.0**
+
+ - Make `toArray` operate recursively to flatten nested structures. PR#1
+
**1.0.0**
- Initial release | 4 | 0.571429 | 4 | 0 |
c74e3fa443e193afb5274b672eccd58d2d7cff28 | test/test_rtp.c | test/test_rtp.c |
static KmsEndpoint*
create_endpoint() {
KmsEndpoint *ep;
gchar *name;
name = g_strdup_printf(LOCALNAME);
ep = g_object_new(KMS_TYPE_RTP_ENDPOINT, "localname", name, NULL);
g_free(name);
return ep;
}
static void
check_endpoint(KmsEndpoint *ep) {
gchar *name;
g_object_get(ep, "localname", &name, NULL);
g_assert(g_strcmp0(name, LOCALNAME) == 0);
g_free(name);
}
gint
main(gint argc, gchar **argv) {
KmsEndpoint *ep;
g_type_init();
ep = create_endpoint();
check_endpoint(ep);
g_object_unref(ep);
return 0;
}
|
static KmsEndpoint*
create_endpoint() {
KmsEndpoint *ep;
gchar *name;
name = g_strdup_printf(LOCALNAME);
ep = g_object_new(KMS_TYPE_RTP_ENDPOINT, "localname", name, NULL);
g_free(name);
return ep;
}
static void
check_endpoint(KmsEndpoint *ep) {
gchar *name;
g_object_get(ep, "localname", &name, NULL);
g_assert(g_strcmp0(name, LOCALNAME) == 0);
g_free(name);
}
gint
main(gint argc, gchar **argv) {
KmsEndpoint *ep;
KmsConnection *conn;
g_type_init();
ep = create_endpoint();
check_endpoint(ep);
conn = kms_endpoint_create_connection(ep, KMS_CONNECTION_TYPE_RTP,
NULL);
kms_endpoint_delete_connection(ep, conn, NULL);
g_object_unref(conn);
check_endpoint(ep);
g_object_unref(ep);
return 0;
}
| Create and delete rtp connections | Create and delete rtp connections
| C | lgpl-2.1 | mparis/kurento-media-server,mparis/kurento-media-server,lulufei/kurento-media-server,shelsonjava/kurento-media-server,Kurento/kurento-media-server,todotobe1/kurento-media-server,Kurento/kurento-media-server,lulufei/kurento-media-server,TribeMedia/kurento-media-server,todotobe1/kurento-media-server,TribeMedia/kurento-media-server,shelsonjava/kurento-media-server | c | ## Code Before:
static KmsEndpoint*
create_endpoint() {
KmsEndpoint *ep;
gchar *name;
name = g_strdup_printf(LOCALNAME);
ep = g_object_new(KMS_TYPE_RTP_ENDPOINT, "localname", name, NULL);
g_free(name);
return ep;
}
static void
check_endpoint(KmsEndpoint *ep) {
gchar *name;
g_object_get(ep, "localname", &name, NULL);
g_assert(g_strcmp0(name, LOCALNAME) == 0);
g_free(name);
}
gint
main(gint argc, gchar **argv) {
KmsEndpoint *ep;
g_type_init();
ep = create_endpoint();
check_endpoint(ep);
g_object_unref(ep);
return 0;
}
## Instruction:
Create and delete rtp connections
## Code After:
static KmsEndpoint*
create_endpoint() {
KmsEndpoint *ep;
gchar *name;
name = g_strdup_printf(LOCALNAME);
ep = g_object_new(KMS_TYPE_RTP_ENDPOINT, "localname", name, NULL);
g_free(name);
return ep;
}
static void
check_endpoint(KmsEndpoint *ep) {
gchar *name;
g_object_get(ep, "localname", &name, NULL);
g_assert(g_strcmp0(name, LOCALNAME) == 0);
g_free(name);
}
gint
main(gint argc, gchar **argv) {
KmsEndpoint *ep;
KmsConnection *conn;
g_type_init();
ep = create_endpoint();
check_endpoint(ep);
conn = kms_endpoint_create_connection(ep, KMS_CONNECTION_TYPE_RTP,
NULL);
kms_endpoint_delete_connection(ep, conn, NULL);
g_object_unref(conn);
check_endpoint(ep);
g_object_unref(ep);
return 0;
}
|
static KmsEndpoint*
create_endpoint() {
KmsEndpoint *ep;
gchar *name;
name = g_strdup_printf(LOCALNAME);
ep = g_object_new(KMS_TYPE_RTP_ENDPOINT, "localname", name, NULL);
g_free(name);
return ep;
}
static void
check_endpoint(KmsEndpoint *ep) {
gchar *name;
g_object_get(ep, "localname", &name, NULL);
g_assert(g_strcmp0(name, LOCALNAME) == 0);
g_free(name);
}
gint
main(gint argc, gchar **argv) {
KmsEndpoint *ep;
+ KmsConnection *conn;
g_type_init();
ep = create_endpoint();
check_endpoint(ep);
+ conn = kms_endpoint_create_connection(ep, KMS_CONNECTION_TYPE_RTP,
+ NULL);
+ kms_endpoint_delete_connection(ep, conn, NULL);
+ g_object_unref(conn);
+
+ check_endpoint(ep);
+
g_object_unref(ep);
return 0;
} | 8 | 0.210526 | 8 | 0 |
3d84e8e871b1049102815136ef23e3e630461918 | connman_dispatcher/utils.py | connman_dispatcher/utils.py | import os
import subprocess
import logbook
logger = logbook.Logger('connman-dispatcher')
def execute_scripts_in_dirs(paths, state):
for path in sorted(paths):
if os.path.exists(path) and os.path.isdir(path):
execute_scripts_in_dir(path, state)
def execute_scripts_in_dir(path, state):
for script in sorted(os.listdir(path)):
full_scirpt_path = os.path.join(path, script)
if os.path.exists(full_scirpt_path):
logger.info('executing: %s %s' % (full_scirpt_path, state))
subprocess.Popen([full_scirpt_path, state])
| import os
import subprocess
import logbook
logger = logbook.Logger('connman-dispatcher')
def is_executable(path):
return all([os.path.isfile(path), os.access(path, os.X_OK)])
def execute_scripts_in_dirs(paths, state):
for path in sorted(paths):
if os.path.exists(path) and os.path.isdir(path):
execute_scripts_in_dir(path, state)
def execute_scripts_in_dir(path, state):
for script in sorted(os.listdir(path)):
full_scirpt_path = os.path.join(path, script)
if os.path.exists(full_scirpt_path):
if is_executable(full_scirpt_path):
logger.info('executing: %s %s' % (full_scirpt_path, state))
subprocess.Popen([full_scirpt_path, state])
else:
logger.error('%s is not executable file' % full_scirpt_path)
| Check if file is executable, before executing it | Check if file is executable, before executing it
| Python | isc | a-sk/connman-dispatcher | python | ## Code Before:
import os
import subprocess
import logbook
logger = logbook.Logger('connman-dispatcher')
def execute_scripts_in_dirs(paths, state):
for path in sorted(paths):
if os.path.exists(path) and os.path.isdir(path):
execute_scripts_in_dir(path, state)
def execute_scripts_in_dir(path, state):
for script in sorted(os.listdir(path)):
full_scirpt_path = os.path.join(path, script)
if os.path.exists(full_scirpt_path):
logger.info('executing: %s %s' % (full_scirpt_path, state))
subprocess.Popen([full_scirpt_path, state])
## Instruction:
Check if file is executable, before executing it
## Code After:
import os
import subprocess
import logbook
logger = logbook.Logger('connman-dispatcher')
def is_executable(path):
return all([os.path.isfile(path), os.access(path, os.X_OK)])
def execute_scripts_in_dirs(paths, state):
for path in sorted(paths):
if os.path.exists(path) and os.path.isdir(path):
execute_scripts_in_dir(path, state)
def execute_scripts_in_dir(path, state):
for script in sorted(os.listdir(path)):
full_scirpt_path = os.path.join(path, script)
if os.path.exists(full_scirpt_path):
if is_executable(full_scirpt_path):
logger.info('executing: %s %s' % (full_scirpt_path, state))
subprocess.Popen([full_scirpt_path, state])
else:
logger.error('%s is not executable file' % full_scirpt_path)
| import os
import subprocess
import logbook
logger = logbook.Logger('connman-dispatcher')
+
+ def is_executable(path):
+ return all([os.path.isfile(path), os.access(path, os.X_OK)])
def execute_scripts_in_dirs(paths, state):
for path in sorted(paths):
if os.path.exists(path) and os.path.isdir(path):
execute_scripts_in_dir(path, state)
def execute_scripts_in_dir(path, state):
for script in sorted(os.listdir(path)):
full_scirpt_path = os.path.join(path, script)
if os.path.exists(full_scirpt_path):
+ if is_executable(full_scirpt_path):
- logger.info('executing: %s %s' % (full_scirpt_path, state))
+ logger.info('executing: %s %s' % (full_scirpt_path, state))
? ++++
- subprocess.Popen([full_scirpt_path, state])
+ subprocess.Popen([full_scirpt_path, state])
? ++++
+ else:
+ logger.error('%s is not executable file' % full_scirpt_path)
| 10 | 0.555556 | 8 | 2 |
1b4dccd27d66efc7f66bf458bd37a09de913dd1d | _sass/_misc.scss | _sass/_misc.scss | hr {
border: 0;
border-bottom: 1px dashed #cfcfcf;
margin: 30px 0;
}
table {
width: 100%;
max-width: 100%;
background-color: transparent;
margin-bottom: 20px;
border: 2px solid #ddd;
border-spacing: 0;
-webkit-border-radius: 4px;
-moz-border-radius: 4px;
border-radius: 4px;
}
table tr>th:not(:first-child), table tr>td:not(:first-child) {
border-left: 2px solid #ddd;
}
table tr>td {
border-top: 2px solid #ddd;
}
table th, table td {
padding: 8px;
vertical-align: top;
}
table tr:nth-child(2n)>td {
background-color : rgba(102,128,153,.05);
}
table tr:last-child>td:first-child {
border-radius: 0 0 0 2px;
}
table tr:last-child>td:last-child {
border-radius: 0 0 2px 0;
} | hr {
border: 0;
border-bottom: 1px dashed #cfcfcf;
margin: 30px 0;
}
table {
width: 100%;
max-width: 100%;
table-layout: fixed;
background-color: transparent;
margin-bottom: 20px;
border: 2px solid #ddd;
border-spacing: 0;
-webkit-border-radius: 4px;
-moz-border-radius: 4px;
border-radius: 4px;
}
table tr>th:not(:first-child), table tr>td:not(:first-child) {
border-left: 2px solid #ddd;
}
table tr>td {
border-top: 2px solid #ddd;
word-break: break-all;
word-wrap: break-word;
}
table th, table td {
padding: 8px;
vertical-align: top;
}
table tr:nth-child(2n)>td {
background-color : rgba(102,128,153,.05);
}
table tr:last-child>td:first-child {
border-radius: 0 0 0 2px;
}
table tr:last-child>td:last-child {
border-radius: 0 0 2px 0;
} | Add rules to control table layout. | Add rules to control table layout.
| SCSS | mit | wangbangkun/wangbangkun.github.io,wangbangkun/wangbangkun.github.io,wangbangkun/wangbangkun.github.io | scss | ## Code Before:
hr {
border: 0;
border-bottom: 1px dashed #cfcfcf;
margin: 30px 0;
}
table {
width: 100%;
max-width: 100%;
background-color: transparent;
margin-bottom: 20px;
border: 2px solid #ddd;
border-spacing: 0;
-webkit-border-radius: 4px;
-moz-border-radius: 4px;
border-radius: 4px;
}
table tr>th:not(:first-child), table tr>td:not(:first-child) {
border-left: 2px solid #ddd;
}
table tr>td {
border-top: 2px solid #ddd;
}
table th, table td {
padding: 8px;
vertical-align: top;
}
table tr:nth-child(2n)>td {
background-color : rgba(102,128,153,.05);
}
table tr:last-child>td:first-child {
border-radius: 0 0 0 2px;
}
table tr:last-child>td:last-child {
border-radius: 0 0 2px 0;
}
## Instruction:
Add rules to control table layout.
## Code After:
hr {
border: 0;
border-bottom: 1px dashed #cfcfcf;
margin: 30px 0;
}
table {
width: 100%;
max-width: 100%;
table-layout: fixed;
background-color: transparent;
margin-bottom: 20px;
border: 2px solid #ddd;
border-spacing: 0;
-webkit-border-radius: 4px;
-moz-border-radius: 4px;
border-radius: 4px;
}
table tr>th:not(:first-child), table tr>td:not(:first-child) {
border-left: 2px solid #ddd;
}
table tr>td {
border-top: 2px solid #ddd;
word-break: break-all;
word-wrap: break-word;
}
table th, table td {
padding: 8px;
vertical-align: top;
}
table tr:nth-child(2n)>td {
background-color : rgba(102,128,153,.05);
}
table tr:last-child>td:first-child {
border-radius: 0 0 0 2px;
}
table tr:last-child>td:last-child {
border-radius: 0 0 2px 0;
} | hr {
border: 0;
border-bottom: 1px dashed #cfcfcf;
margin: 30px 0;
}
table {
width: 100%;
max-width: 100%;
+ table-layout: fixed;
background-color: transparent;
margin-bottom: 20px;
border: 2px solid #ddd;
border-spacing: 0;
-webkit-border-radius: 4px;
-moz-border-radius: 4px;
border-radius: 4px;
}
table tr>th:not(:first-child), table tr>td:not(:first-child) {
border-left: 2px solid #ddd;
}
table tr>td {
border-top: 2px solid #ddd;
+ word-break: break-all;
+ word-wrap: break-word;
}
table th, table td {
padding: 8px;
vertical-align: top;
}
table tr:nth-child(2n)>td {
background-color : rgba(102,128,153,.05);
}
table tr:last-child>td:first-child {
border-radius: 0 0 0 2px;
}
table tr:last-child>td:last-child {
border-radius: 0 0 2px 0;
} | 3 | 0.071429 | 3 | 0 |
c1629102be68260b1d9d8d8032aa9575173f54d1 | metadata/org.poirsouille.tinc_gui.txt | metadata/org.poirsouille.tinc_gui.txt | Categories:Internet
License:GPLv3+
Web Site:http://tinc_gui.poirsouille.org
Source Code:https://github.com/Vilbrekin/tinc_gui
Issue Tracker:https://github.com/Vilbrekin/tinc_gui/issues
Auto Name:Tinc
Summary:Port of Tinc VPN
Description:
Tinc GUI for Android is a (slightly modified) cross-compiled version of tincd,
associated with a basic GUI for daemon management.
Root is not needed, even if highly recommended for correct tinc daemon usage.
.
Repo Type:git
Repo:https://github.com/Vilbrekin/tinc_gui.git
Build:0.9.7-arm,8
commit=ad35972299ad697a0770e7b8d9cb5b3778415a7f
forceversion=yes
submodules=yes
rm=res/raw/tincd
build=sed -i 1i'SHELL := /bin/bash' src_tinc/Makefile && \
make -C src_tinc/ install
Maintainer Notes:
Uses armeabi binary in res/raw/, so no native-code appears and fdroid can't
filter by architecture. This is why -arm is appended, to at least let people
running x86 or MIPS know that it won't work for them yet.
.
Auto Update Mode:None
Update Check Mode:Tags
Current Version:0.9.7
Current Version Code:8
| Categories:Internet
License:GPLv3+
Web Site:http://tinc_gui.poirsouille.org
Source Code:https://github.com/Vilbrekin/tinc_gui
Issue Tracker:https://github.com/Vilbrekin/tinc_gui/issues
Auto Name:Tinc
Summary:Port of Tinc VPN
Description:
Tinc GUI for Android is a (slightly modified) cross-compiled version of tincd,
associated with a basic GUI for daemon management.
Root is not needed, even if highly recommended for correct tinc daemon usage.
.
Repo Type:git
Repo:https://github.com/Vilbrekin/tinc_gui.git
Build:0.9.7-arm,8
commit=ad35972299ad697a0770e7b8d9cb5b3778415a7f
forceversion=yes
submodules=yes
rm=res/raw/tincd
build=make -C src_tinc/ install
Maintainer Notes:
Uses armeabi binary in res/raw/, so no native-code appears and fdroid can't
filter by architecture. This is why -arm is appended, to at least let people
running x86 or MIPS know that it won't work for them yet.
.
Auto Update Mode:None
Update Check Mode:Tags
Current Version:0.9.7
Current Version Code:8
| Revert "Tinc: fix build" - no longer needed? | Revert "Tinc: fix build" - no longer needed?
This reverts commit 618975403a28b8574e7418248ed60997b8dd571d.
| Text | agpl-3.0 | f-droid/fdroiddata,f-droid/fdroid-data,f-droid/fdroiddata | text | ## Code Before:
Categories:Internet
License:GPLv3+
Web Site:http://tinc_gui.poirsouille.org
Source Code:https://github.com/Vilbrekin/tinc_gui
Issue Tracker:https://github.com/Vilbrekin/tinc_gui/issues
Auto Name:Tinc
Summary:Port of Tinc VPN
Description:
Tinc GUI for Android is a (slightly modified) cross-compiled version of tincd,
associated with a basic GUI for daemon management.
Root is not needed, even if highly recommended for correct tinc daemon usage.
.
Repo Type:git
Repo:https://github.com/Vilbrekin/tinc_gui.git
Build:0.9.7-arm,8
commit=ad35972299ad697a0770e7b8d9cb5b3778415a7f
forceversion=yes
submodules=yes
rm=res/raw/tincd
build=sed -i 1i'SHELL := /bin/bash' src_tinc/Makefile && \
make -C src_tinc/ install
Maintainer Notes:
Uses armeabi binary in res/raw/, so no native-code appears and fdroid can't
filter by architecture. This is why -arm is appended, to at least let people
running x86 or MIPS know that it won't work for them yet.
.
Auto Update Mode:None
Update Check Mode:Tags
Current Version:0.9.7
Current Version Code:8
## Instruction:
Revert "Tinc: fix build" - no longer needed?
This reverts commit 618975403a28b8574e7418248ed60997b8dd571d.
## Code After:
Categories:Internet
License:GPLv3+
Web Site:http://tinc_gui.poirsouille.org
Source Code:https://github.com/Vilbrekin/tinc_gui
Issue Tracker:https://github.com/Vilbrekin/tinc_gui/issues
Auto Name:Tinc
Summary:Port of Tinc VPN
Description:
Tinc GUI for Android is a (slightly modified) cross-compiled version of tincd,
associated with a basic GUI for daemon management.
Root is not needed, even if highly recommended for correct tinc daemon usage.
.
Repo Type:git
Repo:https://github.com/Vilbrekin/tinc_gui.git
Build:0.9.7-arm,8
commit=ad35972299ad697a0770e7b8d9cb5b3778415a7f
forceversion=yes
submodules=yes
rm=res/raw/tincd
build=make -C src_tinc/ install
Maintainer Notes:
Uses armeabi binary in res/raw/, so no native-code appears and fdroid can't
filter by architecture. This is why -arm is appended, to at least let people
running x86 or MIPS know that it won't work for them yet.
.
Auto Update Mode:None
Update Check Mode:Tags
Current Version:0.9.7
Current Version Code:8
| Categories:Internet
License:GPLv3+
Web Site:http://tinc_gui.poirsouille.org
Source Code:https://github.com/Vilbrekin/tinc_gui
Issue Tracker:https://github.com/Vilbrekin/tinc_gui/issues
Auto Name:Tinc
Summary:Port of Tinc VPN
Description:
Tinc GUI for Android is a (slightly modified) cross-compiled version of tincd,
associated with a basic GUI for daemon management.
Root is not needed, even if highly recommended for correct tinc daemon usage.
.
Repo Type:git
Repo:https://github.com/Vilbrekin/tinc_gui.git
Build:0.9.7-arm,8
commit=ad35972299ad697a0770e7b8d9cb5b3778415a7f
forceversion=yes
submodules=yes
rm=res/raw/tincd
- build=sed -i 1i'SHELL := /bin/bash' src_tinc/Makefile && \
- make -C src_tinc/ install
? ^^^^^^
+ build=make -C src_tinc/ install
? ^^^^^^
Maintainer Notes:
Uses armeabi binary in res/raw/, so no native-code appears and fdroid can't
filter by architecture. This is why -arm is appended, to at least let people
running x86 or MIPS know that it won't work for them yet.
.
Auto Update Mode:None
Update Check Mode:Tags
Current Version:0.9.7
Current Version Code:8
| 3 | 0.081081 | 1 | 2 |
78c1f4f9266b4954b11819747218a69b2c3b2eee | lib/spreedly/payment_methods/credit_card.rb | lib/spreedly/payment_methods/credit_card.rb | module Spreedly
class CreditCard < PaymentMethod
field :first_name, :last_name, :number, :last_four_digits, :card_type, :verification_value
field :month, :year, :address1, :address2, :city, :state, :country, :phone_number
end
end
| module Spreedly
class CreditCard < PaymentMethod
field :first_name, :last_name, :number, :last_four_digits, :card_type, :verification_value
field :full_name, :month, :year, :address1, :address2, :city, :state, :country, :phone_number
end
end
| Support full_name in credit card | Support full_name in credit card
| Ruby | mit | SturdyCo/spreedly-gem,BitPerformanceLabs/spreedly-gem,kieranklaassen/spreedly-gem,spreedly/spreedly-gem | ruby | ## Code Before:
module Spreedly
class CreditCard < PaymentMethod
field :first_name, :last_name, :number, :last_four_digits, :card_type, :verification_value
field :month, :year, :address1, :address2, :city, :state, :country, :phone_number
end
end
## Instruction:
Support full_name in credit card
## Code After:
module Spreedly
class CreditCard < PaymentMethod
field :first_name, :last_name, :number, :last_four_digits, :card_type, :verification_value
field :full_name, :month, :year, :address1, :address2, :city, :state, :country, :phone_number
end
end
| module Spreedly
class CreditCard < PaymentMethod
field :first_name, :last_name, :number, :last_four_digits, :card_type, :verification_value
- field :month, :year, :address1, :address2, :city, :state, :country, :phone_number
+ field :full_name, :month, :year, :address1, :address2, :city, :state, :country, :phone_number
? ++++++++++++
end
end | 2 | 0.222222 | 1 | 1 |
f4476f14ebcd224a46305fac0562639e43e7dc2f | README.md | README.md |
[](https://travis-ci.org/refile/refile-fog)
A backend for [Refile](https://github.com/elabs/refile) which provides storage
in multiple cloud storage services via the [Fog](https://github.com/fog/fog)
cloud services gem.
If you're looking for a backend for Amazon S3, please use Refile's build in S3
backend. It is far superior to using S3 via Fog.
## Installation
Add this line to your application's Gemfile:
```ruby
gem 'refile-fog'
```
Set up Refile to use the fog backend:
``` ruby
# config/initializers/refile.rb
require "refile/fog"
credentials = {
provider: "Google",
aws_access_key_id: "zyx",
aws_secret_access_key: "12345",
directory: "my-app"
}
Refile.configure do |config|
config.cache = Refile::Fog::Backend.new(prefix: "cache", **credentials)
config.store = Refile::Fog::Backend.new(prefix: "store", **credentials)
end
```
## License
[MIT](License.txt)
|
[](https://travis-ci.org/refile/refile-fog)
A backend for [Refile](https://github.com/elabs/refile) which provides storage
in multiple cloud storage services via the [Fog](https://github.com/fog/fog)
cloud services gem.
If you're looking for a backend for Amazon S3, please use Refile's build in S3
backend. It is far superior to using S3 via Fog.
## Installation
Add this line to your application's Gemfile:
```ruby
gem 'refile-fog'
```
You also need to add the storage provider you use.
Example for google storage:
```ruby
gem 'fog-google'
gem 'google-api-client', '~> 0.8.6'
```
Set up Refile to use the fog backend:
``` ruby
# config/initializers/refile.rb
require "fog/google"
require "refile/fog"
credentials = {
provider: "Google",
aws_access_key_id: "zyx",
aws_secret_access_key: "12345",
directory: "my-app"
}
Refile.configure do |config|
config.cache = Refile::Fog::Backend.new(prefix: "cache", **credentials)
config.store = Refile::Fog::Backend.new(prefix: "store", **credentials)
end
```
## License
[MIT](License.txt)
| Add information about adding providers | Add information about adding providers
| Markdown | mit | refile/refile-fog | markdown | ## Code Before:
[](https://travis-ci.org/refile/refile-fog)
A backend for [Refile](https://github.com/elabs/refile) which provides storage
in multiple cloud storage services via the [Fog](https://github.com/fog/fog)
cloud services gem.
If you're looking for a backend for Amazon S3, please use Refile's build in S3
backend. It is far superior to using S3 via Fog.
## Installation
Add this line to your application's Gemfile:
```ruby
gem 'refile-fog'
```
Set up Refile to use the fog backend:
``` ruby
# config/initializers/refile.rb
require "refile/fog"
credentials = {
provider: "Google",
aws_access_key_id: "zyx",
aws_secret_access_key: "12345",
directory: "my-app"
}
Refile.configure do |config|
config.cache = Refile::Fog::Backend.new(prefix: "cache", **credentials)
config.store = Refile::Fog::Backend.new(prefix: "store", **credentials)
end
```
## License
[MIT](License.txt)
## Instruction:
Add information about adding providers
## Code After:
[](https://travis-ci.org/refile/refile-fog)
A backend for [Refile](https://github.com/elabs/refile) which provides storage
in multiple cloud storage services via the [Fog](https://github.com/fog/fog)
cloud services gem.
If you're looking for a backend for Amazon S3, please use Refile's build in S3
backend. It is far superior to using S3 via Fog.
## Installation
Add this line to your application's Gemfile:
```ruby
gem 'refile-fog'
```
You also need to add the storage provider you use.
Example for google storage:
```ruby
gem 'fog-google'
gem 'google-api-client', '~> 0.8.6'
```
Set up Refile to use the fog backend:
``` ruby
# config/initializers/refile.rb
require "fog/google"
require "refile/fog"
credentials = {
provider: "Google",
aws_access_key_id: "zyx",
aws_secret_access_key: "12345",
directory: "my-app"
}
Refile.configure do |config|
config.cache = Refile::Fog::Backend.new(prefix: "cache", **credentials)
config.store = Refile::Fog::Backend.new(prefix: "store", **credentials)
end
```
## License
[MIT](License.txt)
|
[](https://travis-ci.org/refile/refile-fog)
A backend for [Refile](https://github.com/elabs/refile) which provides storage
in multiple cloud storage services via the [Fog](https://github.com/fog/fog)
cloud services gem.
If you're looking for a backend for Amazon S3, please use Refile's build in S3
backend. It is far superior to using S3 via Fog.
## Installation
Add this line to your application's Gemfile:
```ruby
gem 'refile-fog'
```
+ You also need to add the storage provider you use.
+
+ Example for google storage:
+
+ ```ruby
+ gem 'fog-google'
+ gem 'google-api-client', '~> 0.8.6'
+
+ ```
+
Set up Refile to use the fog backend:
``` ruby
# config/initializers/refile.rb
+ require "fog/google"
require "refile/fog"
credentials = {
provider: "Google",
aws_access_key_id: "zyx",
aws_secret_access_key: "12345",
directory: "my-app"
}
Refile.configure do |config|
config.cache = Refile::Fog::Backend.new(prefix: "cache", **credentials)
config.store = Refile::Fog::Backend.new(prefix: "store", **credentials)
end
```
## License
[MIT](License.txt) | 11 | 0.268293 | 11 | 0 |
012415a1dc8f15c0918874bd001d8f19cd602560 | templates/zerver/help/restrict-editing-of-old-messages-and-topics.md | templates/zerver/help/restrict-editing-of-old-messages-and-topics.md |
If you are an administrator of a Zulip organization, you can easily change the
time limit that your organization's users have to change their messages after sending
them. Alternatively, you can choose to disable message editing for your organization
users.
{!go-to-the.md!} [Organization Settings](/#administration/organization-settings)
{!admin.md!}
4. Locate the **Users can edit old messages**
checkbox and **Message edit limit in minutes (0 for no limit)** input field
underneath it.
By default, user message editing is enabled for 10 minutes after sending.
* **Users can edit old messages** - Uncheck this option if you wish to
disable message editing. Upon doing so, the **Message edit limit in minutes (0 for no limit)**
input field will be grayed out.
* **Message edit limit in minutes (0 for no limit)** - If you enable message
editing in your organization, you can restrict the time that organization
users have to edit their messages. Simply input the time limit in minutes
that you would like to set; for example, if you want to set a message edit
time limit of 5 minutes, enter **5** in the field.
!!! tip ""
If you would like to disable the message editing time limit for your
organization, enter **0** in the field. This enables users to edit
their messages whenever they want.
5. To save any changes you have made to your organization settings, click the
**Save changes** button at the bottom of the **Organizations settings**
section.
|
If you are an administrator of a Zulip organization, you can easily change the
time limit that your organization's users have to change their messages after sending
them. Alternatively, you can choose to disable message editing for your organization
users.
{!go-to-the.md!} [Organization Settings](/#administration/organization-settings)
{!admin.md!}
4. Locate the **Users can edit old messages**
checkbox and **Message edit limit in minutes (0 for no limit)** input field
underneath it.
By default, user message editing is enabled for 10 minutes after sending.
* **Users can edit old messages** - Uncheck this option if you wish to
disable message editing. Upon doing so, the **Message edit limit in minutes (0 for no limit)**
input field will be grayed out.
* **Message edit limit in minutes (0 for no limit)** - If you enable message
editing in your organization, you can restrict the time that organization
users have to edit their messages. Simply input the time limit in minutes
that you would like to set; for example, if you want to set a message edit
time limit of 5 minutes, enter **5** in the field.
!!! tip ""
If you would like to disable the message editing time limit for your
organization, enter **0** in the field. This enables users to edit
their messages whenever they want.
{!save-changes.md!} organization settings.
| Add *Save changes* macro to *Restrict editing of old messages and topics* | docs: Add *Save changes* macro to *Restrict editing of old messages and topics*
| Markdown | apache-2.0 | timabbott/zulip,brainwane/zulip,verma-varsha/zulip,verma-varsha/zulip,isht3/zulip,brockwhittaker/zulip,andersk/zulip,rht/zulip,rht/zulip,rishig/zulip,j831/zulip,SmartPeople/zulip,isht3/zulip,zulip/zulip,tommyip/zulip,amyliu345/zulip,jphilipsen05/zulip,punchagan/zulip,andersk/zulip,amyliu345/zulip,showell/zulip,vabs22/zulip,dawran6/zulip,isht3/zulip,rishig/zulip,amyliu345/zulip,susansls/zulip,PhilSk/zulip,jrowan/zulip,jackrzhang/zulip,jphilipsen05/zulip,hackerkid/zulip,jrowan/zulip,verma-varsha/zulip,Galexrt/zulip,rishig/zulip,brainwane/zulip,jainayush975/zulip,Galexrt/zulip,sonali0901/zulip,Galexrt/zulip,brainwane/zulip,jrowan/zulip,christi3k/zulip,blaze225/zulip,PhilSk/zulip,zulip/zulip,amyliu345/zulip,hackerkid/zulip,dhcrzf/zulip,brainwane/zulip,PhilSk/zulip,showell/zulip,christi3k/zulip,eeshangarg/zulip,hackerkid/zulip,jackrzhang/zulip,blaze225/zulip,eeshangarg/zulip,susansls/zulip,rishig/zulip,kou/zulip,samatdav/zulip,synicalsyntax/zulip,JPJPJPOPOP/zulip,ryanbackman/zulip,jphilipsen05/zulip,tommyip/zulip,vabs22/zulip,blaze225/zulip,tommyip/zulip,Galexrt/zulip,jainayush975/zulip,zulip/zulip,rht/zulip,dawran6/zulip,JPJPJPOPOP/zulip,punchagan/zulip,eeshangarg/zulip,synicalsyntax/zulip,amanharitsh123/zulip,ryanbackman/zulip,synicalsyntax/zulip,showell/zulip,blaze225/zulip,mahim97/zulip,tommyip/zulip,brockwhittaker/zulip,Galexrt/zulip,rht/zulip,j831/zulip,jainayush975/zulip,sonali0901/zulip,amanharitsh123/zulip,susansls/zulip,amyliu345/zulip,brockwhittaker/zulip,JPJPJPOPOP/zulip,Galexrt/zulip,blaze225/zulip,samatdav/zulip,sonali0901/zulip,shubhamdhama/zulip,SmartPeople/zulip,amanharitsh123/zulip,andersk/zulip,amanharitsh123/zulip,christi3k/zulip,jackrzhang/zulip,kou/zulip,synicalsyntax/zulip,dhcrzf/zulip,dattatreya303/zulip,synicalsyntax/zulip,JPJPJPOPOP/zulip,zulip/zulip,susansls/zulip,sonali0901/zulip,sharmaeklavya2/zulip,Galexrt/zulip,timabbott/zulip,brainwane/zulip,JPJPJPOPOP/zulip,kou/zulip,vaidap/zulip,vaidap/zulip,jackrzhang/zulip,jrowan/zulip,j831/zulip,brockwhittaker/zulip,dawran6/zulip,sonali0901/zulip,souravbadami/zulip,timabbott/zulip,tommyip/zulip,sharmaeklavya2/zulip,showell/zulip,brockwhittaker/zulip,shubhamdhama/zulip,dattatreya303/zulip,showell/zulip,vaidap/zulip,kou/zulip,dattatreya303/zulip,aakash-cr7/zulip,brainwane/zulip,jphilipsen05/zulip,isht3/zulip,vabs22/zulip,aakash-cr7/zulip,punchagan/zulip,shubhamdhama/zulip,aakash-cr7/zulip,punchagan/zulip,SmartPeople/zulip,andersk/zulip,rht/zulip,hackerkid/zulip,showell/zulip,kou/zulip,zulip/zulip,shubhamdhama/zulip,hackerkid/zulip,eeshangarg/zulip,dhcrzf/zulip,jainayush975/zulip,verma-varsha/zulip,souravbadami/zulip,vabs22/zulip,mahim97/zulip,samatdav/zulip,JPJPJPOPOP/zulip,dattatreya303/zulip,shubhamdhama/zulip,christi3k/zulip,mahim97/zulip,hackerkid/zulip,jphilipsen05/zulip,dhcrzf/zulip,punchagan/zulip,andersk/zulip,dhcrzf/zulip,kou/zulip,vabs22/zulip,mahim97/zulip,samatdav/zulip,SmartPeople/zulip,souravbadami/zulip,blaze225/zulip,sonali0901/zulip,mahim97/zulip,jainayush975/zulip,kou/zulip,mahim97/zulip,souravbadami/zulip,brockwhittaker/zulip,susansls/zulip,amanharitsh123/zulip,PhilSk/zulip,punchagan/zulip,vaidap/zulip,shubhamdhama/zulip,sharmaeklavya2/zulip,SmartPeople/zulip,tommyip/zulip,eeshangarg/zulip,jackrzhang/zulip,verma-varsha/zulip,brainwane/zulip,isht3/zulip,ryanbackman/zulip,j831/zulip,rht/zulip,timabbott/zulip,PhilSk/zulip,sharmaeklavya2/zulip,samatdav/zulip,j831/zulip,eeshangarg/zulip,zulip/zulip,rishig/zulip,jrowan/zulip,ryanbackman/zulip,ryanbackman/zulip,tommyip/zulip,susansls/zulip,christi3k/zulip,aakash-cr7/zulip,vabs22/zulip,vaidap/zulip,timabbott/zulip,ryanbackman/zulip,andersk/zulip,shubhamdhama/zulip,synicalsyntax/zulip,dhcrzf/zulip,synicalsyntax/zulip,PhilSk/zulip,vaidap/zulip,amyliu345/zulip,rht/zulip,showell/zulip,aakash-cr7/zulip,hackerkid/zulip,jackrzhang/zulip,dhcrzf/zulip,andersk/zulip,samatdav/zulip,timabbott/zulip,eeshangarg/zulip,dawran6/zulip,amanharitsh123/zulip,christi3k/zulip,aakash-cr7/zulip,jrowan/zulip,dawran6/zulip,sharmaeklavya2/zulip,jphilipsen05/zulip,timabbott/zulip,jainayush975/zulip,souravbadami/zulip,j831/zulip,dattatreya303/zulip,sharmaeklavya2/zulip,zulip/zulip,isht3/zulip,verma-varsha/zulip,jackrzhang/zulip,rishig/zulip,SmartPeople/zulip,dattatreya303/zulip,rishig/zulip,dawran6/zulip,souravbadami/zulip,punchagan/zulip | markdown | ## Code Before:
If you are an administrator of a Zulip organization, you can easily change the
time limit that your organization's users have to change their messages after sending
them. Alternatively, you can choose to disable message editing for your organization
users.
{!go-to-the.md!} [Organization Settings](/#administration/organization-settings)
{!admin.md!}
4. Locate the **Users can edit old messages**
checkbox and **Message edit limit in minutes (0 for no limit)** input field
underneath it.
By default, user message editing is enabled for 10 minutes after sending.
* **Users can edit old messages** - Uncheck this option if you wish to
disable message editing. Upon doing so, the **Message edit limit in minutes (0 for no limit)**
input field will be grayed out.
* **Message edit limit in minutes (0 for no limit)** - If you enable message
editing in your organization, you can restrict the time that organization
users have to edit their messages. Simply input the time limit in minutes
that you would like to set; for example, if you want to set a message edit
time limit of 5 minutes, enter **5** in the field.
!!! tip ""
If you would like to disable the message editing time limit for your
organization, enter **0** in the field. This enables users to edit
their messages whenever they want.
5. To save any changes you have made to your organization settings, click the
**Save changes** button at the bottom of the **Organizations settings**
section.
## Instruction:
docs: Add *Save changes* macro to *Restrict editing of old messages and topics*
## Code After:
If you are an administrator of a Zulip organization, you can easily change the
time limit that your organization's users have to change their messages after sending
them. Alternatively, you can choose to disable message editing for your organization
users.
{!go-to-the.md!} [Organization Settings](/#administration/organization-settings)
{!admin.md!}
4. Locate the **Users can edit old messages**
checkbox and **Message edit limit in minutes (0 for no limit)** input field
underneath it.
By default, user message editing is enabled for 10 minutes after sending.
* **Users can edit old messages** - Uncheck this option if you wish to
disable message editing. Upon doing so, the **Message edit limit in minutes (0 for no limit)**
input field will be grayed out.
* **Message edit limit in minutes (0 for no limit)** - If you enable message
editing in your organization, you can restrict the time that organization
users have to edit their messages. Simply input the time limit in minutes
that you would like to set; for example, if you want to set a message edit
time limit of 5 minutes, enter **5** in the field.
!!! tip ""
If you would like to disable the message editing time limit for your
organization, enter **0** in the field. This enables users to edit
their messages whenever they want.
{!save-changes.md!} organization settings.
|
If you are an administrator of a Zulip organization, you can easily change the
time limit that your organization's users have to change their messages after sending
them. Alternatively, you can choose to disable message editing for your organization
users.
{!go-to-the.md!} [Organization Settings](/#administration/organization-settings)
{!admin.md!}
4. Locate the **Users can edit old messages**
checkbox and **Message edit limit in minutes (0 for no limit)** input field
underneath it.
By default, user message editing is enabled for 10 minutes after sending.
* **Users can edit old messages** - Uncheck this option if you wish to
disable message editing. Upon doing so, the **Message edit limit in minutes (0 for no limit)**
input field will be grayed out.
* **Message edit limit in minutes (0 for no limit)** - If you enable message
editing in your organization, you can restrict the time that organization
users have to edit their messages. Simply input the time limit in minutes
that you would like to set; for example, if you want to set a message edit
time limit of 5 minutes, enter **5** in the field.
!!! tip ""
If you would like to disable the message editing time limit for your
organization, enter **0** in the field. This enables users to edit
their messages whenever they want.
+ {!save-changes.md!} organization settings.
- 5. To save any changes you have made to your organization settings, click the
- **Save changes** button at the bottom of the **Organizations settings**
- section. | 4 | 0.121212 | 1 | 3 |
1695df5386808815fac919486be2fb378a9634aa | android/src/com/eegeo/surveys/SurveyView.java | android/src/com/eegeo/surveys/SurveyView.java | // Copyright eeGeo Ltd (2012-2016), All Rights Reserved
package com.eegeo.surveys;
import com.eegeo.entrypointinfrastructure.MainActivity;
import android.app.AlertDialog;
import android.content.DialogInterface;
import android.content.Intent;
import android.net.Uri;
public class SurveyView {
private MainActivity m_activity = null;
private static AlertDialog m_options = null;
public SurveyView(MainActivity activity) {
m_activity = activity;
}
protected void startUxCallback(String timerSurveyUrl)
{
AlertDialog.Builder builder = new AlertDialog.Builder(m_activity);
builder.setTitle("Want to help us improve our map?");
builder.setMessage("Just a few quick questions - honest!");
builder.setPositiveButton("Yes", showSurvey(timerSurveyUrl));
builder.setNegativeButton("No", null);
builder.setCancelable(false);
m_options = builder.show();
}
private DialogInterface.OnClickListener showSurvey(final String timerSurveyUrl)
{
return new DialogInterface.OnClickListener()
{
@Override
public void onClick(DialogInterface dialog, int which)
{
Intent browserIntent = new Intent(Intent.ACTION_VIEW, Uri.parse(timerSurveyUrl));
m_activity.startActivity(browserIntent);
m_options = null;
}
};
}
}
| // Copyright eeGeo Ltd (2012-2016), All Rights Reserved
package com.eegeo.surveys;
import com.eegeo.entrypointinfrastructure.MainActivity;
import android.app.AlertDialog;
import android.content.DialogInterface;
import android.content.Intent;
import android.net.Uri;
public class SurveyView {
private MainActivity m_activity = null;
private static AlertDialog m_options = null;
public SurveyView(MainActivity activity) {
m_activity = activity;
}
protected void startUxCallback(String timerSurveyUrl)
{
AlertDialog.Builder builder = new AlertDialog.Builder(m_activity);
builder.setTitle("Want to help us improve our map?");
builder.setMessage("Just a few quick questions - honest!");
builder.setPositiveButton("Yes", showSurvey(timerSurveyUrl));
builder.setNegativeButton("No", null);
builder.setCancelable(false);
m_options = builder.show();
}
private DialogInterface.OnClickListener showSurvey(final String timerSurveyUrl)
{
return new DialogInterface.OnClickListener() {
@Override
public void onClick(DialogInterface dialog, int which) {
String prefixedUrl = "";
if (!timerSurveyUrl.isEmpty())
{
if (!timerSurveyUrl.startsWith("https://") && !timerSurveyUrl.startsWith("http://")) {
prefixedUrl = "http://" + timerSurveyUrl;
}
Intent browserIntent = new Intent(Intent.ACTION_VIEW, Uri.parse(prefixedUrl));
m_activity.startActivity(browserIntent);
}
m_options = null;
}
};
}
}
| Fix for MPLY-8221. Survey now handle a null survey url and urls that don't start with http or https. Buddy: Vimmy | Fix for MPLY-8221. Survey now handle a null survey url and urls that don't start with http or https. Buddy: Vimmy
| Java | bsd-2-clause | eegeo/eegeo-example-app,wrld3d/wrld-example-app,wrld3d/wrld-example-app,wrld3d/wrld-example-app,wrld3d/wrld-example-app,eegeo/eegeo-example-app,wrld3d/wrld-example-app,eegeo/eegeo-example-app,wrld3d/wrld-example-app,eegeo/eegeo-example-app,eegeo/eegeo-example-app,wrld3d/wrld-example-app,eegeo/eegeo-example-app,eegeo/eegeo-example-app,wrld3d/wrld-example-app | java | ## Code Before:
// Copyright eeGeo Ltd (2012-2016), All Rights Reserved
package com.eegeo.surveys;
import com.eegeo.entrypointinfrastructure.MainActivity;
import android.app.AlertDialog;
import android.content.DialogInterface;
import android.content.Intent;
import android.net.Uri;
public class SurveyView {
private MainActivity m_activity = null;
private static AlertDialog m_options = null;
public SurveyView(MainActivity activity) {
m_activity = activity;
}
protected void startUxCallback(String timerSurveyUrl)
{
AlertDialog.Builder builder = new AlertDialog.Builder(m_activity);
builder.setTitle("Want to help us improve our map?");
builder.setMessage("Just a few quick questions - honest!");
builder.setPositiveButton("Yes", showSurvey(timerSurveyUrl));
builder.setNegativeButton("No", null);
builder.setCancelable(false);
m_options = builder.show();
}
private DialogInterface.OnClickListener showSurvey(final String timerSurveyUrl)
{
return new DialogInterface.OnClickListener()
{
@Override
public void onClick(DialogInterface dialog, int which)
{
Intent browserIntent = new Intent(Intent.ACTION_VIEW, Uri.parse(timerSurveyUrl));
m_activity.startActivity(browserIntent);
m_options = null;
}
};
}
}
## Instruction:
Fix for MPLY-8221. Survey now handle a null survey url and urls that don't start with http or https. Buddy: Vimmy
## Code After:
// Copyright eeGeo Ltd (2012-2016), All Rights Reserved
package com.eegeo.surveys;
import com.eegeo.entrypointinfrastructure.MainActivity;
import android.app.AlertDialog;
import android.content.DialogInterface;
import android.content.Intent;
import android.net.Uri;
public class SurveyView {
private MainActivity m_activity = null;
private static AlertDialog m_options = null;
public SurveyView(MainActivity activity) {
m_activity = activity;
}
protected void startUxCallback(String timerSurveyUrl)
{
AlertDialog.Builder builder = new AlertDialog.Builder(m_activity);
builder.setTitle("Want to help us improve our map?");
builder.setMessage("Just a few quick questions - honest!");
builder.setPositiveButton("Yes", showSurvey(timerSurveyUrl));
builder.setNegativeButton("No", null);
builder.setCancelable(false);
m_options = builder.show();
}
private DialogInterface.OnClickListener showSurvey(final String timerSurveyUrl)
{
return new DialogInterface.OnClickListener() {
@Override
public void onClick(DialogInterface dialog, int which) {
String prefixedUrl = "";
if (!timerSurveyUrl.isEmpty())
{
if (!timerSurveyUrl.startsWith("https://") && !timerSurveyUrl.startsWith("http://")) {
prefixedUrl = "http://" + timerSurveyUrl;
}
Intent browserIntent = new Intent(Intent.ACTION_VIEW, Uri.parse(prefixedUrl));
m_activity.startActivity(browserIntent);
}
m_options = null;
}
};
}
}
| // Copyright eeGeo Ltd (2012-2016), All Rights Reserved
package com.eegeo.surveys;
import com.eegeo.entrypointinfrastructure.MainActivity;
import android.app.AlertDialog;
import android.content.DialogInterface;
import android.content.Intent;
import android.net.Uri;
public class SurveyView {
private MainActivity m_activity = null;
private static AlertDialog m_options = null;
public SurveyView(MainActivity activity) {
m_activity = activity;
}
protected void startUxCallback(String timerSurveyUrl)
{
AlertDialog.Builder builder = new AlertDialog.Builder(m_activity);
builder.setTitle("Want to help us improve our map?");
builder.setMessage("Just a few quick questions - honest!");
builder.setPositiveButton("Yes", showSurvey(timerSurveyUrl));
builder.setNegativeButton("No", null);
builder.setCancelable(false);
m_options = builder.show();
}
private DialogInterface.OnClickListener showSurvey(final String timerSurveyUrl)
{
- return new DialogInterface.OnClickListener()
+ return new DialogInterface.OnClickListener() {
? ++
- {
@Override
- public void onClick(DialogInterface dialog, int which)
+ public void onClick(DialogInterface dialog, int which) {
? ++
+ String prefixedUrl = "";
+ if (!timerSurveyUrl.isEmpty())
- {
+ {
? ++++
+ if (!timerSurveyUrl.startsWith("https://") && !timerSurveyUrl.startsWith("http://")) {
+ prefixedUrl = "http://" + timerSurveyUrl;
+ }
- Intent browserIntent = new Intent(Intent.ACTION_VIEW, Uri.parse(timerSurveyUrl));
? ^ ^ ^ ^^^^^^^
+ Intent browserIntent = new Intent(Intent.ACTION_VIEW, Uri.parse(prefixedUrl));
? ^^^^^^^^ ^^^^ ^ ^
- m_activity.startActivity(browserIntent);
? ^
+ m_activity.startActivity(browserIntent);
? ^^^^^^^^
+ }
- m_options = null;
? ^
+ m_options = null;
? ^^^^
}
};
}
} | 19 | 0.404255 | 12 | 7 |
e229349a1a12100d89802f6402129f8ed93e0857 | appveyor.yml | appveyor.yml | environment:
matrix:
- PYTHON: "C:\\Python34"
- PYTHON: "C:\\Python27"
cache:
- C:\container -> appveyor.yml
install:
- ps: $env:PATH="${env:PYTHON};${env:PYTHON}\\Scripts;${env:PATH};C:\MinGW\bin;C:\Python34"
- ps: python ciscripts/bootstrap.py -d C:/container -s container-setup.py -r $(pwd) -e powershell -p test-env.ps1 --no-mdl
- ps: . ./test-env
build: false
test_script:
- ps: polysquare_run check/python/check.py --coverage-exclude "*/__init__.py" "*/_scripts/*" "tmp*" --lint-exclude="*/sample/*" --no-mdl
after_test:
- ps: polysquare_cleanup
| environment:
matrix:
- PYTHON: "C:\\Python34"
- PYTHON: "C:\\Python27"
cache:
- C:\container -> appveyor.yml
install:
- ps: $env:PATH="${env:PYTHON};${env:PYTHON}\\Scripts;${env:PATH};C:\\MinGW\bin;C:\\Python34;C:\\Python34\\Scripts;C:\\Ruby21\bin"
- ps: python ciscripts/bootstrap.py -d C:/container -s container-setup.py -r $(pwd) -e powershell -p test-env.ps1 --no-mdl
- ps: . ./test-env
build: false
test_script:
- ps: polysquare_run check/python/check.py --coverage-exclude "*/__init__.py" "*/_scripts/*" "tmp*" --lint-exclude="*/sample/*" --no-mdl
after_test:
- ps: polysquare_cleanup
| Include Python34 and Python34/Scripts in PATH | Include Python34 and Python34/Scripts in PATH
| YAML | mit | polysquare/polysquare-ci-scripts,polysquare/polysquare-ci-scripts | yaml | ## Code Before:
environment:
matrix:
- PYTHON: "C:\\Python34"
- PYTHON: "C:\\Python27"
cache:
- C:\container -> appveyor.yml
install:
- ps: $env:PATH="${env:PYTHON};${env:PYTHON}\\Scripts;${env:PATH};C:\MinGW\bin;C:\Python34"
- ps: python ciscripts/bootstrap.py -d C:/container -s container-setup.py -r $(pwd) -e powershell -p test-env.ps1 --no-mdl
- ps: . ./test-env
build: false
test_script:
- ps: polysquare_run check/python/check.py --coverage-exclude "*/__init__.py" "*/_scripts/*" "tmp*" --lint-exclude="*/sample/*" --no-mdl
after_test:
- ps: polysquare_cleanup
## Instruction:
Include Python34 and Python34/Scripts in PATH
## Code After:
environment:
matrix:
- PYTHON: "C:\\Python34"
- PYTHON: "C:\\Python27"
cache:
- C:\container -> appveyor.yml
install:
- ps: $env:PATH="${env:PYTHON};${env:PYTHON}\\Scripts;${env:PATH};C:\\MinGW\bin;C:\\Python34;C:\\Python34\\Scripts;C:\\Ruby21\bin"
- ps: python ciscripts/bootstrap.py -d C:/container -s container-setup.py -r $(pwd) -e powershell -p test-env.ps1 --no-mdl
- ps: . ./test-env
build: false
test_script:
- ps: polysquare_run check/python/check.py --coverage-exclude "*/__init__.py" "*/_scripts/*" "tmp*" --lint-exclude="*/sample/*" --no-mdl
after_test:
- ps: polysquare_cleanup
| environment:
matrix:
- PYTHON: "C:\\Python34"
- PYTHON: "C:\\Python27"
cache:
- C:\container -> appveyor.yml
install:
- - ps: $env:PATH="${env:PYTHON};${env:PYTHON}\\Scripts;${env:PATH};C:\MinGW\bin;C:\Python34"
+ - ps: $env:PATH="${env:PYTHON};${env:PYTHON}\\Scripts;${env:PATH};C:\\MinGW\bin;C:\\Python34;C:\\Python34\\Scripts;C:\\Ruby21\bin"
? + + +++++++++++++++++++++++++++++++++++++
- ps: python ciscripts/bootstrap.py -d C:/container -s container-setup.py -r $(pwd) -e powershell -p test-env.ps1 --no-mdl
- ps: . ./test-env
build: false
test_script:
- ps: polysquare_run check/python/check.py --coverage-exclude "*/__init__.py" "*/_scripts/*" "tmp*" --lint-exclude="*/sample/*" --no-mdl
after_test:
- ps: polysquare_cleanup | 2 | 0.1 | 1 | 1 |
e83e8d197a04ae6d2ae7f0bd4cbe5e8ccf9d303e | test/functional/no-drm-player.html | test/functional/no-drm-player.html | <!DOCTYPE html>
<head>
<meta charset="utf-8">
<meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1">
<title>Test Player</title>
<meta name="description" content="">
<meta name="viewport" content="width=device-width">
<link rel="stylesheet" href="../../node_modules/video.js/dist/video-js/video-js.css">
<link rel="stylesheet" href="../../src/css/videojs-dash.css">
</head>
<body>
<video id="example-video" width=600 height=300 class="video-js vjs-default-skin" controls>
</video>
<script src="../../node_modules/video.js/dist/video-js/video.novtt.dev.js"></script>
<!-- Dash.js -->
<script src="../../node_modules/dashjs/dist/dash.all.js"></script>
<!-- videojs-contrib-dash -->
<script src="../../src/js/videojs-dash.js"></script>
<script>
var player = videojs('example-video');
player.ready(function(){
player.src({
src: "http://dash.edgesuite.net/dash264/TestCases/2c/qualcomm/2/MultiRes.mpd",
type: "application/dash+xml"
});
});
</script>
</body>
</html>
| <!DOCTYPE html>
<head>
<meta charset="utf-8">
<meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1">
<title>Test Player</title>
<meta name="description" content="">
<meta name="viewport" content="width=device-width">
<link rel="stylesheet" href="../../node_modules/video.js/dist/video-js/video-js.css">
<link rel="stylesheet" href="../../src/css/videojs-dash.css">
</head>
<body>
<video id="example-video" width=600 height=300 class="video-js vjs-default-skin" controls>
</video>
<script src="../../node_modules/video.js/dist/video-js/video.novtt.dev.js"></script>
<!-- Dash.js -->
<script src="../../node_modules/dashjs/dist/dash.all.js"></script>
<!-- videojs-contrib-dash -->
<script src="../../src/js/videojs-dash.js"></script>
<script>
var player = videojs('example-video');
player.ready(function(){
player.src({
src: "http://wams.edgesuite.net/media/SintelTrailer_MP4_from_WAME/sintel_trailer-1080p.ism/manifest(format=mpd-time-csf)",
type: "application/dash+xml"
});
});
</script>
</body>
</html>
| Use a different test stream for non-protected playback | Use a different test stream for non-protected playback
| HTML | apache-2.0 | Afrostream/videojs-contrib-dash,videojs/videojs-contrib-dash,streamroot/videojs5-dashjs-p2p-source-handler,Afrostream/videojs-contrib-dash,radut/videojs-contrib-dash,cmwd/videojs-contrib-dash,radut/videojs-contrib-dash,videojs/videojs-contrib-dash,streamroot/videojs5-dashjs-p2p-source-handler,cmwd/videojs-contrib-dash | html | ## Code Before:
<!DOCTYPE html>
<head>
<meta charset="utf-8">
<meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1">
<title>Test Player</title>
<meta name="description" content="">
<meta name="viewport" content="width=device-width">
<link rel="stylesheet" href="../../node_modules/video.js/dist/video-js/video-js.css">
<link rel="stylesheet" href="../../src/css/videojs-dash.css">
</head>
<body>
<video id="example-video" width=600 height=300 class="video-js vjs-default-skin" controls>
</video>
<script src="../../node_modules/video.js/dist/video-js/video.novtt.dev.js"></script>
<!-- Dash.js -->
<script src="../../node_modules/dashjs/dist/dash.all.js"></script>
<!-- videojs-contrib-dash -->
<script src="../../src/js/videojs-dash.js"></script>
<script>
var player = videojs('example-video');
player.ready(function(){
player.src({
src: "http://dash.edgesuite.net/dash264/TestCases/2c/qualcomm/2/MultiRes.mpd",
type: "application/dash+xml"
});
});
</script>
</body>
</html>
## Instruction:
Use a different test stream for non-protected playback
## Code After:
<!DOCTYPE html>
<head>
<meta charset="utf-8">
<meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1">
<title>Test Player</title>
<meta name="description" content="">
<meta name="viewport" content="width=device-width">
<link rel="stylesheet" href="../../node_modules/video.js/dist/video-js/video-js.css">
<link rel="stylesheet" href="../../src/css/videojs-dash.css">
</head>
<body>
<video id="example-video" width=600 height=300 class="video-js vjs-default-skin" controls>
</video>
<script src="../../node_modules/video.js/dist/video-js/video.novtt.dev.js"></script>
<!-- Dash.js -->
<script src="../../node_modules/dashjs/dist/dash.all.js"></script>
<!-- videojs-contrib-dash -->
<script src="../../src/js/videojs-dash.js"></script>
<script>
var player = videojs('example-video');
player.ready(function(){
player.src({
src: "http://wams.edgesuite.net/media/SintelTrailer_MP4_from_WAME/sintel_trailer-1080p.ism/manifest(format=mpd-time-csf)",
type: "application/dash+xml"
});
});
</script>
</body>
</html>
| <!DOCTYPE html>
<head>
<meta charset="utf-8">
<meta http-equiv="X-UA-Compatible" content="IE=edge,chrome=1">
<title>Test Player</title>
<meta name="description" content="">
<meta name="viewport" content="width=device-width">
<link rel="stylesheet" href="../../node_modules/video.js/dist/video-js/video-js.css">
<link rel="stylesheet" href="../../src/css/videojs-dash.css">
</head>
<body>
<video id="example-video" width=600 height=300 class="video-js vjs-default-skin" controls>
</video>
<script src="../../node_modules/video.js/dist/video-js/video.novtt.dev.js"></script>
<!-- Dash.js -->
<script src="../../node_modules/dashjs/dist/dash.all.js"></script>
<!-- videojs-contrib-dash -->
<script src="../../src/js/videojs-dash.js"></script>
<script>
var player = videojs('example-video');
player.ready(function(){
player.src({
- src: "http://dash.edgesuite.net/dash264/TestCases/2c/qualcomm/2/MultiRes.mpd",
+ src: "http://wams.edgesuite.net/media/SintelTrailer_MP4_from_WAME/sintel_trailer-1080p.ism/manifest(format=mpd-time-csf)",
type: "application/dash+xml"
});
});
</script>
</body>
</html> | 2 | 0.060606 | 1 | 1 |
81c763ca2c9d1e120a4ef516e135ea6ef2046953 | src/Oro/Bundle/ImportExportBundle/Resources/translations/messages.en.yml | src/Oro/Bundle/ImportExportBundle/Resources/translations/messages.en.yml | oro_importexport.export.exported_entities_count %count%: "Exported entities: %count%"
oro.importexport.import_error %number%: "Error in row #%number%."
oro_importexport.import.import_success: "File was successful imported."
oro_importexport.import.import_error: "Import error."
| oro_importexport.export.exported_entities_count %count%: "Exported entities: %count%"
oro.importexport.import_error %number%: "Error in row #%number%."
oro_importexport.import.import_success: "File was successful imported."
oro_importexport.import.import_error: "Errors occurred during file import."
| Create universal import/export controller - ddd | CRM-483: Create universal import/export controller
- ddd
| YAML | mit | morontt/platform,2ndkauboy/platform,hugeval/platform,geoffroycochard/platform,Djamy/platform,akeneo/platform,umpirsky/platform,mszajner/platform,trustify/oroplatform,ramunasd/platform,northdakota/platform,mszajner/platform,orocrm/platform,ramunasd/platform,orocrm/platform,morontt/platform,orocrm/platform,2ndkauboy/platform,geoffroycochard/platform,umpirsky/platform,northdakota/platform,Djamy/platform,akeneo/platform,Djamy/platform,2ndkauboy/platform,hugeval/platform,trustify/oroplatform,hugeval/platform,morontt/platform,trustify/oroplatform,ramunasd/platform,northdakota/platform,akeneo/platform,geoffroycochard/platform,mszajner/platform | yaml | ## Code Before:
oro_importexport.export.exported_entities_count %count%: "Exported entities: %count%"
oro.importexport.import_error %number%: "Error in row #%number%."
oro_importexport.import.import_success: "File was successful imported."
oro_importexport.import.import_error: "Import error."
## Instruction:
CRM-483: Create universal import/export controller
- ddd
## Code After:
oro_importexport.export.exported_entities_count %count%: "Exported entities: %count%"
oro.importexport.import_error %number%: "Error in row #%number%."
oro_importexport.import.import_success: "File was successful imported."
oro_importexport.import.import_error: "Errors occurred during file import."
| oro_importexport.export.exported_entities_count %count%: "Exported entities: %count%"
oro.importexport.import_error %number%: "Error in row #%number%."
oro_importexport.import.import_success: "File was successful imported."
- oro_importexport.import.import_error: "Import error."
+ oro_importexport.import.import_error: "Errors occurred during file import." | 2 | 0.5 | 1 | 1 |
c455f0a48666f726fa8f522c969cd6b01ac380cd | modules/mod_ginger_base/support/ginger_config.erl | modules/mod_ginger_base/support/ginger_config.erl | -module(ginger_config).
-author("Driebit <tech@driebit.nl").
-export([
install_config/1
]).
-include_lib("zotonic.hrl").
%% @doc Set preferred default config (but don't overwrite if already set)
-spec install_config(#context{}) -> ok.
install_config(Context) ->
lists:foreach(
fun({Mod, Key, Value}) ->
m_config:get(Mod, Key, Context),
case m_config:get(Mod, Key, Context) of
undefined ->
m_config:set_value(Mod, Key, Value, Context);
_ ->
noop
end
end,
get_config()
),
ok.
%% @doc Get default Ginger config parameters
-spec get_config() -> list().
get_config() ->
[
{mod_l10n, timezone, <<"Europe/Berlin">>},
{mod_l10n, timezone_is_fixed, true}
].
| -module(ginger_config).
-author("Driebit <tech@driebit.nl").
-export([
install_config/1
]).
-include_lib("zotonic.hrl").
%% @doc Set preferred default config (but don't overwrite if already set)
-spec install_config(#context{}) -> ok.
install_config(Context) ->
lists:foreach(
fun({Mod, Key, Value}) ->
m_config:get(Mod, Key, Context),
case m_config:get(Mod, Key, Context) of
undefined ->
m_config:set_value(Mod, Key, Value, Context);
_ ->
noop
end
end,
get_config()
),
ok.
%% @doc Get default Ginger config parameters
-spec get_config() -> list().
get_config() ->
[
{i18n, language, nl},
{mod_l10n, timezone, <<"Europe/Berlin">>},
{mod_l10n, timezone_is_fixed, true}
].
| Set Dutch as default site language | [base] Set Dutch as default site language
| Erlang | apache-2.0 | driebit/ginger,driebit/ginger,driebit/ginger | erlang | ## Code Before:
-module(ginger_config).
-author("Driebit <tech@driebit.nl").
-export([
install_config/1
]).
-include_lib("zotonic.hrl").
%% @doc Set preferred default config (but don't overwrite if already set)
-spec install_config(#context{}) -> ok.
install_config(Context) ->
lists:foreach(
fun({Mod, Key, Value}) ->
m_config:get(Mod, Key, Context),
case m_config:get(Mod, Key, Context) of
undefined ->
m_config:set_value(Mod, Key, Value, Context);
_ ->
noop
end
end,
get_config()
),
ok.
%% @doc Get default Ginger config parameters
-spec get_config() -> list().
get_config() ->
[
{mod_l10n, timezone, <<"Europe/Berlin">>},
{mod_l10n, timezone_is_fixed, true}
].
## Instruction:
[base] Set Dutch as default site language
## Code After:
-module(ginger_config).
-author("Driebit <tech@driebit.nl").
-export([
install_config/1
]).
-include_lib("zotonic.hrl").
%% @doc Set preferred default config (but don't overwrite if already set)
-spec install_config(#context{}) -> ok.
install_config(Context) ->
lists:foreach(
fun({Mod, Key, Value}) ->
m_config:get(Mod, Key, Context),
case m_config:get(Mod, Key, Context) of
undefined ->
m_config:set_value(Mod, Key, Value, Context);
_ ->
noop
end
end,
get_config()
),
ok.
%% @doc Get default Ginger config parameters
-spec get_config() -> list().
get_config() ->
[
{i18n, language, nl},
{mod_l10n, timezone, <<"Europe/Berlin">>},
{mod_l10n, timezone_is_fixed, true}
].
| -module(ginger_config).
-author("Driebit <tech@driebit.nl").
-export([
install_config/1
]).
-include_lib("zotonic.hrl").
%% @doc Set preferred default config (but don't overwrite if already set)
-spec install_config(#context{}) -> ok.
install_config(Context) ->
lists:foreach(
fun({Mod, Key, Value}) ->
m_config:get(Mod, Key, Context),
case m_config:get(Mod, Key, Context) of
undefined ->
m_config:set_value(Mod, Key, Value, Context);
_ ->
noop
end
end,
get_config()
),
ok.
%% @doc Get default Ginger config parameters
-spec get_config() -> list().
get_config() ->
[
+ {i18n, language, nl},
{mod_l10n, timezone, <<"Europe/Berlin">>},
{mod_l10n, timezone_is_fixed, true}
]. | 1 | 0.030303 | 1 | 0 |
7b4dac2b9cbcf5d3fd75b031f47b44dbe417d6c8 | app/views/admin/users/_datasets_table_row.html.erb | app/views/admin/users/_datasets_table_row.html.erb | <tr>
<td><%= permission.datasetID %></td>
<td colspan="2"><%= permission.access_level.datasetname %></td>
<td>
<% if table_name == :accessible && category == :a && permission.user_has_access %>
<!-- Revoke button -->
<%= link_to_delete_permission(@user.user, permission.datasetID, 'revoke', category) do %>
<%= image_tag "button_drop.png", :alt => "revoke" %>
<% end %>
<% elsif table_name == :pending %>
<!-- Checkbox to grant access to pending dataset -->
<%= check_box_tag "user[datasets_cat_#{category}_pending_to_grant][]", permission.datasetID, false, :id => "pending_#{ddi_to_id(permission.datasetID)}" %>
<% elsif category == :b %>
<!-- Permissions for category B dataset -->
<%= UserPermissionB::PERMISSION_VALUE_S[permission.permissionvalue] %>
<% end %>
</td>
<td>
<!-- Delete button -->
<%= link_to_delete_permission(@user.user, permission.datasetID, 'delete', category) do %>
<%= image_tag "button_drop.png", :alt => "delete" %>
<% end %>
</td>
</tr>
| <tr>
<td><%= permission.datasetID %></td>
<td colspan="2"><%= permission.access_level ? permission.access_level.datasetname : '-' %></td>
<td>
<% if table_name == :accessible && category == :a && permission.user_has_access %>
<!-- Revoke button -->
<%= link_to_delete_permission(@user.user, permission.datasetID, 'revoke', category) do %>
<%= image_tag "button_drop.png", :alt => "revoke" %>
<% end %>
<% elsif table_name == :pending %>
<!-- Checkbox to grant access to pending dataset -->
<%= check_box_tag "user[datasets_cat_#{category}_pending_to_grant][]", permission.datasetID, false, :id => "pending_#{ddi_to_id(permission.datasetID)}" %>
<% elsif category == :b %>
<!-- Permissions for category B dataset -->
<%= UserPermissionB::PERMISSION_VALUE_S[permission.permissionvalue] %>
<% end %>
</td>
<td>
<!-- Delete button -->
<%= link_to_delete_permission(@user.user, permission.datasetID, 'delete', category) do %>
<%= image_tag "button_drop.png", :alt => "delete" %>
<% end %>
</td>
</tr>
| Handle users with permission to deleted datasets | Handle users with permission to deleted datasets
| HTML+ERB | mit | ANUSF/ADAUsers,ANUSF/ADAUsers,ANUSF/ADAUsers | html+erb | ## Code Before:
<tr>
<td><%= permission.datasetID %></td>
<td colspan="2"><%= permission.access_level.datasetname %></td>
<td>
<% if table_name == :accessible && category == :a && permission.user_has_access %>
<!-- Revoke button -->
<%= link_to_delete_permission(@user.user, permission.datasetID, 'revoke', category) do %>
<%= image_tag "button_drop.png", :alt => "revoke" %>
<% end %>
<% elsif table_name == :pending %>
<!-- Checkbox to grant access to pending dataset -->
<%= check_box_tag "user[datasets_cat_#{category}_pending_to_grant][]", permission.datasetID, false, :id => "pending_#{ddi_to_id(permission.datasetID)}" %>
<% elsif category == :b %>
<!-- Permissions for category B dataset -->
<%= UserPermissionB::PERMISSION_VALUE_S[permission.permissionvalue] %>
<% end %>
</td>
<td>
<!-- Delete button -->
<%= link_to_delete_permission(@user.user, permission.datasetID, 'delete', category) do %>
<%= image_tag "button_drop.png", :alt => "delete" %>
<% end %>
</td>
</tr>
## Instruction:
Handle users with permission to deleted datasets
## Code After:
<tr>
<td><%= permission.datasetID %></td>
<td colspan="2"><%= permission.access_level ? permission.access_level.datasetname : '-' %></td>
<td>
<% if table_name == :accessible && category == :a && permission.user_has_access %>
<!-- Revoke button -->
<%= link_to_delete_permission(@user.user, permission.datasetID, 'revoke', category) do %>
<%= image_tag "button_drop.png", :alt => "revoke" %>
<% end %>
<% elsif table_name == :pending %>
<!-- Checkbox to grant access to pending dataset -->
<%= check_box_tag "user[datasets_cat_#{category}_pending_to_grant][]", permission.datasetID, false, :id => "pending_#{ddi_to_id(permission.datasetID)}" %>
<% elsif category == :b %>
<!-- Permissions for category B dataset -->
<%= UserPermissionB::PERMISSION_VALUE_S[permission.permissionvalue] %>
<% end %>
</td>
<td>
<!-- Delete button -->
<%= link_to_delete_permission(@user.user, permission.datasetID, 'delete', category) do %>
<%= image_tag "button_drop.png", :alt => "delete" %>
<% end %>
</td>
</tr>
| <tr>
<td><%= permission.datasetID %></td>
- <td colspan="2"><%= permission.access_level.datasetname %></td>
+ <td colspan="2"><%= permission.access_level ? permission.access_level.datasetname : '-' %></td>
? ++++++++++++++++++++++++++ ++++++
<td>
<% if table_name == :accessible && category == :a && permission.user_has_access %>
<!-- Revoke button -->
<%= link_to_delete_permission(@user.user, permission.datasetID, 'revoke', category) do %>
<%= image_tag "button_drop.png", :alt => "revoke" %>
<% end %>
<% elsif table_name == :pending %>
<!-- Checkbox to grant access to pending dataset -->
<%= check_box_tag "user[datasets_cat_#{category}_pending_to_grant][]", permission.datasetID, false, :id => "pending_#{ddi_to_id(permission.datasetID)}" %>
<% elsif category == :b %>
<!-- Permissions for category B dataset -->
<%= UserPermissionB::PERMISSION_VALUE_S[permission.permissionvalue] %>
<% end %>
</td>
<td>
<!-- Delete button -->
<%= link_to_delete_permission(@user.user, permission.datasetID, 'delete', category) do %>
<%= image_tag "button_drop.png", :alt => "delete" %>
<% end %>
</td>
</tr> | 2 | 0.083333 | 1 | 1 |
a37282b5bb4ef054703e5d2a6c3a893b1e2148c1 | src/main/frontend/css/venus.css | src/main/frontend/css/venus.css | body, html {
overflow: hidden;
}
#simulator-tab-view {
display: none;
}
#editor-tab-view {
overflow: hidden;
height: 100%;
width: 100%;
}
#asm-editor {
width: 100%;
height: 88vh;
max-height: 88vh;
font-family: Courier New, Courier, monospace;
}
.section {
padding: 0rem 1.5rem;
}
#mc-column {
width: 20%;
}
#oc-column {
width: 80%;
}
#program-listing {
font-family: Courier New, Courier, monospace;
}
#program-listing-body {
overflow: auto;
height: 100%;
}
#program-listing-container {
min-height: 60vh;
max-height: 60vh;
overflow-y: auto;
}
#register-listing-container {
min-height: 80vh;
max-height: 80vh;
overflow-y: auto;
}
#register-listing {
font-family: Courier New, Courier, monospace;
}
#console-output {
height: 100%;
}
| body {
height: 100vh;
}
#simulator-tab-view {
display: none;
}
#editor-tab-view {
overflow: hidden;
height: 100%;
width: 100%;
}
#asm-editor {
width: 100%;
height: 88vh;
max-height: 88vh;
font-family: Courier New, Courier, monospace;
}
.section {
padding: 0rem 1.5rem;
}
#mc-column {
width: 20%;
}
#oc-column {
width: 80%;
}
#program-listing {
font-family: Courier New, Courier, monospace;
}
#program-listing-body {
overflow: auto;
height: 100%;
}
#program-listing-container {
min-height: 60vh;
max-height: 60vh;
overflow-y: auto;
}
#register-listing-container {
min-height: 80vh;
max-height: 80vh;
overflow-y: auto;
}
#register-listing {
font-family: Courier New, Courier, monospace;
}
#console-output {
height: 100%;
}
| Change to allow scrolling on small screens | Change to allow scrolling on small screens
| CSS | mit | kvakil/venus,kvakil/venus,kvakil/venus | css | ## Code Before:
body, html {
overflow: hidden;
}
#simulator-tab-view {
display: none;
}
#editor-tab-view {
overflow: hidden;
height: 100%;
width: 100%;
}
#asm-editor {
width: 100%;
height: 88vh;
max-height: 88vh;
font-family: Courier New, Courier, monospace;
}
.section {
padding: 0rem 1.5rem;
}
#mc-column {
width: 20%;
}
#oc-column {
width: 80%;
}
#program-listing {
font-family: Courier New, Courier, monospace;
}
#program-listing-body {
overflow: auto;
height: 100%;
}
#program-listing-container {
min-height: 60vh;
max-height: 60vh;
overflow-y: auto;
}
#register-listing-container {
min-height: 80vh;
max-height: 80vh;
overflow-y: auto;
}
#register-listing {
font-family: Courier New, Courier, monospace;
}
#console-output {
height: 100%;
}
## Instruction:
Change to allow scrolling on small screens
## Code After:
body {
height: 100vh;
}
#simulator-tab-view {
display: none;
}
#editor-tab-view {
overflow: hidden;
height: 100%;
width: 100%;
}
#asm-editor {
width: 100%;
height: 88vh;
max-height: 88vh;
font-family: Courier New, Courier, monospace;
}
.section {
padding: 0rem 1.5rem;
}
#mc-column {
width: 20%;
}
#oc-column {
width: 80%;
}
#program-listing {
font-family: Courier New, Courier, monospace;
}
#program-listing-body {
overflow: auto;
height: 100%;
}
#program-listing-container {
min-height: 60vh;
max-height: 60vh;
overflow-y: auto;
}
#register-listing-container {
min-height: 80vh;
max-height: 80vh;
overflow-y: auto;
}
#register-listing {
font-family: Courier New, Courier, monospace;
}
#console-output {
height: 100%;
}
| - body, html {
- overflow: hidden;
+ body {
+ height: 100vh;
}
#simulator-tab-view {
display: none;
}
#editor-tab-view {
overflow: hidden;
height: 100%;
width: 100%;
}
#asm-editor {
width: 100%;
height: 88vh;
max-height: 88vh;
font-family: Courier New, Courier, monospace;
}
.section {
padding: 0rem 1.5rem;
}
#mc-column {
width: 20%;
}
#oc-column {
width: 80%;
}
#program-listing {
font-family: Courier New, Courier, monospace;
}
#program-listing-body {
overflow: auto;
height: 100%;
}
#program-listing-container {
min-height: 60vh;
max-height: 60vh;
overflow-y: auto;
}
#register-listing-container {
min-height: 80vh;
max-height: 80vh;
overflow-y: auto;
}
#register-listing {
font-family: Courier New, Courier, monospace;
}
#console-output {
height: 100%;
} | 4 | 0.065574 | 2 | 2 |
380c5445bc53f0060c7bccb85fd04cf6d87782ea | Tofu/Password.swift | Tofu/Password.swift | import Foundation
class Password {
var algorithm: Algorithm = .sha1
var counter = 0
var digits = 6
var period = 30
var secret = Data()
var timeBased = false
func valueForDate(_ date: Date) -> String {
let counter = timeBased ? UInt64(date.timeIntervalSince1970) / UInt64(period) : UInt64(self.counter)
var input = counter.bigEndian
let digest = UnsafeMutablePointer<UInt8>.allocate(capacity: algorithm.digestLength)
defer { digest.deallocate() }
secret.withUnsafeBytes { secretBytes in CCHmac(algorithm.hmacAlgorithm, secretBytes, secret.count, &input, MemoryLayout.size(ofValue: input), digest) }
let offset = digest[algorithm.digestLength - 1] & 0x0f
let number = (digest + Int(offset)).withMemoryRebound(to: UInt32.self, capacity: 1) { UInt32(bigEndian: $0.pointee) } & 0x7fffffff
return String(format: "%0\(digits)d", number % UInt32(pow(10, Float(digits))))
}
func progressForDate(_ date: Date) -> Double {
return timeIntervalRemainingForDate(date) / Double(period)
}
func timeIntervalRemainingForDate(_ date: Date) -> Double {
let period = Double(self.period)
return period - (date.timeIntervalSince1970.truncatingRemainder(dividingBy: period))
}
}
| import Foundation
class Password {
var algorithm: Algorithm = .sha1
var counter = 0
var digits = 6
var period = 30
var secret = Data()
var timeBased = false
func valueForDate(_ date: Date) -> String {
let counter = timeBased ? UInt64(date.timeIntervalSince1970) / UInt64(period) : UInt64(self.counter)
var input = counter.bigEndian
let digest = UnsafeMutablePointer<UInt8>.allocate(capacity: algorithm.digestLength)
defer { digest.deallocate() }
secret.withUnsafeBytes { secretBytes in CCHmac(algorithm.hmacAlgorithm, secretBytes.baseAddress, secret.count, &input, MemoryLayout.size(ofValue: input), digest) }
let offset = digest[algorithm.digestLength - 1] & 0x0f
let number = (digest + Int(offset)).withMemoryRebound(to: UInt32.self, capacity: 1) { UInt32(bigEndian: $0.pointee) } & 0x7fffffff
return String(format: "%0\(digits)d", number % UInt32(pow(10, Float(digits))))
}
func progressForDate(_ date: Date) -> Double {
return timeIntervalRemainingForDate(date) / Double(period)
}
func timeIntervalRemainingForDate(_ date: Date) -> Double {
let period = Double(self.period)
return period - (date.timeIntervalSince1970.truncatingRemainder(dividingBy: period))
}
}
| Use the non-deprecated version of withUnsafeBytes | Use the non-deprecated version of withUnsafeBytes
The one that provides an UnsafeRawBufferPointer instead of an
UnsafePointer that is.
| Swift | isc | calleerlandsson/Tofu | swift | ## Code Before:
import Foundation
class Password {
var algorithm: Algorithm = .sha1
var counter = 0
var digits = 6
var period = 30
var secret = Data()
var timeBased = false
func valueForDate(_ date: Date) -> String {
let counter = timeBased ? UInt64(date.timeIntervalSince1970) / UInt64(period) : UInt64(self.counter)
var input = counter.bigEndian
let digest = UnsafeMutablePointer<UInt8>.allocate(capacity: algorithm.digestLength)
defer { digest.deallocate() }
secret.withUnsafeBytes { secretBytes in CCHmac(algorithm.hmacAlgorithm, secretBytes, secret.count, &input, MemoryLayout.size(ofValue: input), digest) }
let offset = digest[algorithm.digestLength - 1] & 0x0f
let number = (digest + Int(offset)).withMemoryRebound(to: UInt32.self, capacity: 1) { UInt32(bigEndian: $0.pointee) } & 0x7fffffff
return String(format: "%0\(digits)d", number % UInt32(pow(10, Float(digits))))
}
func progressForDate(_ date: Date) -> Double {
return timeIntervalRemainingForDate(date) / Double(period)
}
func timeIntervalRemainingForDate(_ date: Date) -> Double {
let period = Double(self.period)
return period - (date.timeIntervalSince1970.truncatingRemainder(dividingBy: period))
}
}
## Instruction:
Use the non-deprecated version of withUnsafeBytes
The one that provides an UnsafeRawBufferPointer instead of an
UnsafePointer that is.
## Code After:
import Foundation
class Password {
var algorithm: Algorithm = .sha1
var counter = 0
var digits = 6
var period = 30
var secret = Data()
var timeBased = false
func valueForDate(_ date: Date) -> String {
let counter = timeBased ? UInt64(date.timeIntervalSince1970) / UInt64(period) : UInt64(self.counter)
var input = counter.bigEndian
let digest = UnsafeMutablePointer<UInt8>.allocate(capacity: algorithm.digestLength)
defer { digest.deallocate() }
secret.withUnsafeBytes { secretBytes in CCHmac(algorithm.hmacAlgorithm, secretBytes.baseAddress, secret.count, &input, MemoryLayout.size(ofValue: input), digest) }
let offset = digest[algorithm.digestLength - 1] & 0x0f
let number = (digest + Int(offset)).withMemoryRebound(to: UInt32.self, capacity: 1) { UInt32(bigEndian: $0.pointee) } & 0x7fffffff
return String(format: "%0\(digits)d", number % UInt32(pow(10, Float(digits))))
}
func progressForDate(_ date: Date) -> Double {
return timeIntervalRemainingForDate(date) / Double(period)
}
func timeIntervalRemainingForDate(_ date: Date) -> Double {
let period = Double(self.period)
return period - (date.timeIntervalSince1970.truncatingRemainder(dividingBy: period))
}
}
| import Foundation
class Password {
var algorithm: Algorithm = .sha1
var counter = 0
var digits = 6
var period = 30
var secret = Data()
var timeBased = false
func valueForDate(_ date: Date) -> String {
let counter = timeBased ? UInt64(date.timeIntervalSince1970) / UInt64(period) : UInt64(self.counter)
var input = counter.bigEndian
let digest = UnsafeMutablePointer<UInt8>.allocate(capacity: algorithm.digestLength)
defer { digest.deallocate() }
- secret.withUnsafeBytes { secretBytes in CCHmac(algorithm.hmacAlgorithm, secretBytes, secret.count, &input, MemoryLayout.size(ofValue: input), digest) }
+ secret.withUnsafeBytes { secretBytes in CCHmac(algorithm.hmacAlgorithm, secretBytes.baseAddress, secret.count, &input, MemoryLayout.size(ofValue: input), digest) }
? ++++++++++++
let offset = digest[algorithm.digestLength - 1] & 0x0f
let number = (digest + Int(offset)).withMemoryRebound(to: UInt32.self, capacity: 1) { UInt32(bigEndian: $0.pointee) } & 0x7fffffff
return String(format: "%0\(digits)d", number % UInt32(pow(10, Float(digits))))
}
func progressForDate(_ date: Date) -> Double {
return timeIntervalRemainingForDate(date) / Double(period)
}
func timeIntervalRemainingForDate(_ date: Date) -> Double {
let period = Double(self.period)
return period - (date.timeIntervalSince1970.truncatingRemainder(dividingBy: period))
}
} | 2 | 0.066667 | 1 | 1 |
7f9f691449d58aff3b47c3c3ad1cc1408b223ffa | lib/wakame/actions/propagate_instances.rb | lib/wakame/actions/propagate_instances.rb | module Wakame
module Actions
class PropagateInstances < Action
def initialize(resource, propagate_num=0)
raise ArgumentError unless resource.is_a?(Wakame::Service::Resource)
@resource = resource
@propagate_num = propagate_num
end
def run
svc_to_start = []
StatusDB.barrier {
@propagate_num.times {
service_cluster.propagate(@resource)
}
# First, look for the service instances which are already created in the cluster. Then they will be scheduled to start the services later.
online_svc = []
service_cluster.each_instance(@resource.class) { |svc|
if svc.status == Service::STATUS_ONLINE || svc.status == Service::STATUS_STARTING
online_svc << svc
else
svc_to_start << svc
end
}
# The list is empty means that this action is called to propagate a new service instance instead of just starting scheduled instances.
svc_count = service_cluster.instance_count(@resource)
if svc_count > online_svc.size + svc_to_start.size
Wakame.log.debug("#{self.class}: @resource.instance_count - online_svc.size=#{svc_count - online_svc.size}")
(svc_count - (online_svc.size + svc_to_start.size)).times {
svc_to_start << service_cluster.propagate(@resource.class)
}
end
}
acquire_lock { |ary|
svc_to_start.each { |svc|
ary << svc.resource.id
}
}
svc_to_start.each { |svc|
target_agent = nil
if svc.resource.require_agent && !svc.host.mapped?
# Try to arrange agent from existing agent pool.
StatusDB.barrier {
AgentPool.instance.group_active.each { |agent_id|
agent = Agent.find(agent_id)
if agent.has_resource_type?(svc.resource) # && svc.resource.vm_spec.current.satisfy?(agent.vm_attrs)
target_agent = agent
break
end
}
}
unless target_agent.nil?
Wakame.log.debug("#{self.class}: arranged agent for #{svc.resource.class}: #{target_agent.id}")
end
end
trigger_action(StartService.new(svc, target_agent))
}
flush_subactions
end
end
end
end
| module Wakame
module Actions
class PropagateInstances < Action
def initialize(svc, host_id=nil)
raise ArgumentError unless svc.is_a?(Wakame::Service::ServiceInstance)
@svc = svc
@host_id = host_id
end
def run
acquire_lock { |ary|
ary << @svc.resource.class.to_s
}
newsvc = nil
StatusDB.barrier {
newsvc = cluster.propagate_service(@svc.id, @host_id)
}
trigger_action(StartService.new(newsvc))
flush_subactions
end
end
end
end
| Update PropagateInstance action to handle new model structure. | Update PropagateInstance action to handle new model structure.
| Ruby | apache-2.0 | axsh/wakame-fuel,axsh/wakame-fuel,axsh/wakame-fuel | ruby | ## Code Before:
module Wakame
module Actions
class PropagateInstances < Action
def initialize(resource, propagate_num=0)
raise ArgumentError unless resource.is_a?(Wakame::Service::Resource)
@resource = resource
@propagate_num = propagate_num
end
def run
svc_to_start = []
StatusDB.barrier {
@propagate_num.times {
service_cluster.propagate(@resource)
}
# First, look for the service instances which are already created in the cluster. Then they will be scheduled to start the services later.
online_svc = []
service_cluster.each_instance(@resource.class) { |svc|
if svc.status == Service::STATUS_ONLINE || svc.status == Service::STATUS_STARTING
online_svc << svc
else
svc_to_start << svc
end
}
# The list is empty means that this action is called to propagate a new service instance instead of just starting scheduled instances.
svc_count = service_cluster.instance_count(@resource)
if svc_count > online_svc.size + svc_to_start.size
Wakame.log.debug("#{self.class}: @resource.instance_count - online_svc.size=#{svc_count - online_svc.size}")
(svc_count - (online_svc.size + svc_to_start.size)).times {
svc_to_start << service_cluster.propagate(@resource.class)
}
end
}
acquire_lock { |ary|
svc_to_start.each { |svc|
ary << svc.resource.id
}
}
svc_to_start.each { |svc|
target_agent = nil
if svc.resource.require_agent && !svc.host.mapped?
# Try to arrange agent from existing agent pool.
StatusDB.barrier {
AgentPool.instance.group_active.each { |agent_id|
agent = Agent.find(agent_id)
if agent.has_resource_type?(svc.resource) # && svc.resource.vm_spec.current.satisfy?(agent.vm_attrs)
target_agent = agent
break
end
}
}
unless target_agent.nil?
Wakame.log.debug("#{self.class}: arranged agent for #{svc.resource.class}: #{target_agent.id}")
end
end
trigger_action(StartService.new(svc, target_agent))
}
flush_subactions
end
end
end
end
## Instruction:
Update PropagateInstance action to handle new model structure.
## Code After:
module Wakame
module Actions
class PropagateInstances < Action
def initialize(svc, host_id=nil)
raise ArgumentError unless svc.is_a?(Wakame::Service::ServiceInstance)
@svc = svc
@host_id = host_id
end
def run
acquire_lock { |ary|
ary << @svc.resource.class.to_s
}
newsvc = nil
StatusDB.barrier {
newsvc = cluster.propagate_service(@svc.id, @host_id)
}
trigger_action(StartService.new(newsvc))
flush_subactions
end
end
end
end
| module Wakame
module Actions
class PropagateInstances < Action
- def initialize(resource, propagate_num=0)
+ def initialize(svc, host_id=nil)
- raise ArgumentError unless resource.is_a?(Wakame::Service::Resource)
? -- ^^^ - ^ ^^^
+ raise ArgumentError unless svc.is_a?(Wakame::Service::ServiceInstance)
? ^ ^ +++++++ ^^^
- @resource = resource
- @propagate_num = propagate_num
+ @svc = svc
+ @host_id = host_id
end
+
def run
+ acquire_lock { |ary|
+ ary << @svc.resource.class.to_s
- svc_to_start = []
-
- StatusDB.barrier {
- @propagate_num.times {
- service_cluster.propagate(@resource)
- }
-
- # First, look for the service instances which are already created in the cluster. Then they will be scheduled to start the services later.
- online_svc = []
- service_cluster.each_instance(@resource.class) { |svc|
- if svc.status == Service::STATUS_ONLINE || svc.status == Service::STATUS_STARTING
- online_svc << svc
- else
- svc_to_start << svc
- end
- }
-
- # The list is empty means that this action is called to propagate a new service instance instead of just starting scheduled instances.
- svc_count = service_cluster.instance_count(@resource)
- if svc_count > online_svc.size + svc_to_start.size
- Wakame.log.debug("#{self.class}: @resource.instance_count - online_svc.size=#{svc_count - online_svc.size}")
- (svc_count - (online_svc.size + svc_to_start.size)).times {
- svc_to_start << service_cluster.propagate(@resource.class)
- }
- end
}
+ newsvc = nil
+ StatusDB.barrier {
+ newsvc = cluster.propagate_service(@svc.id, @host_id)
- acquire_lock { |ary|
- svc_to_start.each { |svc|
- ary << svc.resource.id
- }
}
-
- svc_to_start.each { |svc|
- target_agent = nil
- if svc.resource.require_agent && !svc.host.mapped?
- # Try to arrange agent from existing agent pool.
- StatusDB.barrier {
- AgentPool.instance.group_active.each { |agent_id|
- agent = Agent.find(agent_id)
- if agent.has_resource_type?(svc.resource) # && svc.resource.vm_spec.current.satisfy?(agent.vm_attrs)
- target_agent = agent
- break
- end
- }
- }
-
- unless target_agent.nil?
- Wakame.log.debug("#{self.class}: arranged agent for #{svc.resource.class}: #{target_agent.id}")
- end
- end
-
- trigger_action(StartService.new(svc, target_agent))
? -- --------------
+ trigger_action(StartService.new(newsvc))
? +++
- }
flush_subactions
end
-
+
end
end
end | 68 | 0.957746 | 12 | 56 |
76959b42af944723a51330a260689902ec681a2e | README.md | README.md | [](https://travis-ci.org/TF2Stadium/TF2RconWrapper)
[](https://coveralls.io/github/TF2Stadium/TF2RconWrapper?branch=master)
Docs are at https://godoc.org/github.com/TF2Stadium/TF2RconWrapper
|
[](https://travis-ci.org/TF2Stadium/TF2RconWrapper)
[](https://coveralls.io/github/TF2Stadium/TF2RconWrapper?branch=master)
[](https://godoc.org/github.com/TF2Stadium/TF2RconWrapper)
## License
Released under the [MIT License](https://github.com/TF2Stadium/TF2RconWrapper/blob/master/LICENSE)
| Add License and GoDoc badge | Add License and GoDoc badge
| Markdown | mit | TF2Stadium/TF2RconWrapper | markdown | ## Code Before:
[](https://travis-ci.org/TF2Stadium/TF2RconWrapper)
[](https://coveralls.io/github/TF2Stadium/TF2RconWrapper?branch=master)
Docs are at https://godoc.org/github.com/TF2Stadium/TF2RconWrapper
## Instruction:
Add License and GoDoc badge
## Code After:
[](https://travis-ci.org/TF2Stadium/TF2RconWrapper)
[](https://coveralls.io/github/TF2Stadium/TF2RconWrapper?branch=master)
[](https://godoc.org/github.com/TF2Stadium/TF2RconWrapper)
## License
Released under the [MIT License](https://github.com/TF2Stadium/TF2RconWrapper/blob/master/LICENSE)
| +
[](https://travis-ci.org/TF2Stadium/TF2RconWrapper)
[](https://coveralls.io/github/TF2Stadium/TF2RconWrapper?branch=master)
+ [](https://godoc.org/github.com/TF2Stadium/TF2RconWrapper)
- Docs are at https://godoc.org/github.com/TF2Stadium/TF2RconWrapper
+ ## License
+ Released under the [MIT License](https://github.com/TF2Stadium/TF2RconWrapper/blob/master/LICENSE) | 5 | 1.25 | 4 | 1 |
6a2cd92ca25805200c50fa229a550dbf8035d240 | src/lib/nameOldEigenQueries.ts | src/lib/nameOldEigenQueries.ts | import { error } from "./loggers"
import { RequestHandler } from "express"
export const nameOldEigenQueries: RequestHandler = (req, _res, next) => {
const agent = req.headers["user-agent"]
if (agent && agent.includes("Eigen") && req.body.query) {
const { query } = req.body as { query: string }
if (!query.startsWith("query ")) {
if (query.includes("saved_artworks")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/favorites.graphql
req.body.query = `query FavoritesQuery ${query}`
} else if (query.includes("sale_artworks")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/artworks_in_sale.graphql
req.body.query = `query ArtworksInSaleQuery ${query}`
} else if (query.includes("totalUnreadCount")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/conversations.graphql
req.body.query = `query TotalUnreadCountQuery ${query}`
} else {
error(`Unexpected Eigen query: ${query}`)
}
}
}
next()
}
| import { error } from "./loggers"
import { RequestHandler } from "express"
export const nameOldEigenQueries: RequestHandler = (req, _res, next) => {
const agent = req.headers["user-agent"]
if (req.body.query && agent && agent.includes("Eigen")) {
const { query } = req.body as { query: string }
if (!query.startsWith("query ")) {
if (query.includes("saved_artworks")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/favorites.graphql
req.body.query = `query FavoritesQuery ${query}`
} else if (query.includes("sale_artworks")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/artworks_in_sale.graphql
req.body.query = `query ArtworksInSaleQuery ${query}`
} else if (query.includes("totalUnreadCount")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/conversations.graphql
req.body.query = `query TotalUnreadCountQuery ${query}`
} else {
error(`Unexpected Eigen query: ${query}`)
}
}
}
next()
}
| Refactor the query check in the eigen queries | Refactor the query check in the eigen queries
| TypeScript | mit | mzikherman/metaphysics-1,mzikherman/metaphysics-1,artsy/metaphysics,mzikherman/metaphysics-1,artsy/metaphysics,artsy/metaphysics,mzikherman/metaphysics-1 | typescript | ## Code Before:
import { error } from "./loggers"
import { RequestHandler } from "express"
export const nameOldEigenQueries: RequestHandler = (req, _res, next) => {
const agent = req.headers["user-agent"]
if (agent && agent.includes("Eigen") && req.body.query) {
const { query } = req.body as { query: string }
if (!query.startsWith("query ")) {
if (query.includes("saved_artworks")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/favorites.graphql
req.body.query = `query FavoritesQuery ${query}`
} else if (query.includes("sale_artworks")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/artworks_in_sale.graphql
req.body.query = `query ArtworksInSaleQuery ${query}`
} else if (query.includes("totalUnreadCount")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/conversations.graphql
req.body.query = `query TotalUnreadCountQuery ${query}`
} else {
error(`Unexpected Eigen query: ${query}`)
}
}
}
next()
}
## Instruction:
Refactor the query check in the eigen queries
## Code After:
import { error } from "./loggers"
import { RequestHandler } from "express"
export const nameOldEigenQueries: RequestHandler = (req, _res, next) => {
const agent = req.headers["user-agent"]
if (req.body.query && agent && agent.includes("Eigen")) {
const { query } = req.body as { query: string }
if (!query.startsWith("query ")) {
if (query.includes("saved_artworks")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/favorites.graphql
req.body.query = `query FavoritesQuery ${query}`
} else if (query.includes("sale_artworks")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/artworks_in_sale.graphql
req.body.query = `query ArtworksInSaleQuery ${query}`
} else if (query.includes("totalUnreadCount")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/conversations.graphql
req.body.query = `query TotalUnreadCountQuery ${query}`
} else {
error(`Unexpected Eigen query: ${query}`)
}
}
}
next()
}
| import { error } from "./loggers"
import { RequestHandler } from "express"
export const nameOldEigenQueries: RequestHandler = (req, _res, next) => {
const agent = req.headers["user-agent"]
- if (agent && agent.includes("Eigen") && req.body.query) {
+ if (req.body.query && agent && agent.includes("Eigen")) {
const { query } = req.body as { query: string }
if (!query.startsWith("query ")) {
if (query.includes("saved_artworks")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/favorites.graphql
req.body.query = `query FavoritesQuery ${query}`
} else if (query.includes("sale_artworks")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/artworks_in_sale.graphql
req.body.query = `query ArtworksInSaleQuery ${query}`
} else if (query.includes("totalUnreadCount")) {
// https://github.com/artsy/eigen/blob/master/Artsy/Networking/conversations.graphql
req.body.query = `query TotalUnreadCountQuery ${query}`
} else {
error(`Unexpected Eigen query: ${query}`)
}
}
}
next()
} | 2 | 0.083333 | 1 | 1 |
77f876b35ee0d60c2bae42ac2a5c4870c805fbde | requirements.txt | requirements.txt | requests>=2.6.0
singledispatch>=3.4.0
pyyaml>=3.1.1
| requests>=2.6.0
singledispatch>=3.4.0
pyyaml>=3.1.1
ordereddict>=1.1
| Add missing requirement for python 2.6 | Add missing requirement for python 2.6
| Text | mit | travispavek/testrail,travispavek/testrail-python | text | ## Code Before:
requests>=2.6.0
singledispatch>=3.4.0
pyyaml>=3.1.1
## Instruction:
Add missing requirement for python 2.6
## Code After:
requests>=2.6.0
singledispatch>=3.4.0
pyyaml>=3.1.1
ordereddict>=1.1
| requests>=2.6.0
singledispatch>=3.4.0
pyyaml>=3.1.1
+ ordereddict>=1.1 | 1 | 0.333333 | 1 | 0 |
c986a269797a21d56e1d6bcbd045634dd6e21b94 | recipes-debian/php-auth/php-auth_debian.bb | recipes-debian/php-auth/php-auth_debian.bb | DESCRIPTION = "PHP PEAR modules for creating an authentication system"
PR = "r0"
inherit debian-package
LICENSE = "PHP-3.1"
LIC_FILES_CHKSUM = "file://debian/copyright;md5=02f173e28791fe9054ea54631abf4dc3"
do_install() {
install -d ${D}${datadir}/php/Auth/Container
install -d ${D}${datadir}/php/Auth/Frontend
install -m 644 ${S}/Auth-*/Auth.php ${D}${datadir}/php/
install -m 644 ${S}/Auth-*/Auth/*.php ${D}${datadir}/php/Auth
install -m 644 ${S}/Auth-*/Auth/Container/* ${D}${datadir}/php/Auth/Container
install -m 644 ${S}/Auth-*/Auth/Frontend/* ${D}${datadir}/php/Auth/Frontend
}
FILES_${PN}="${datadir}"
| DESCRIPTION = "PHP PEAR modules for creating an authentication system"
PR = "r0"
inherit debian-package
LICENSE = "PHP-3.0 & BSD-3-Clause"
LIC_FILES_CHKSUM = " \
file://Auth-1.6.4/Auth.php;beginline=9;endline=13;md5=6863d257d87ab97c7e821ca3fe49a316 \
file://Auth-1.6.4/Auth/Frontend/md5.js;endline=8;md5=915fa4679c019c47741472ce7ce6afd1 \
"
# source format is 3.0 (quilt) but there is no debian patch
DEBIAN_QUILT_PATCHES = ""
do_install() {
install -d ${D}${datadir}/php/Auth/Container
install -d ${D}${datadir}/php/Auth/Frontend
install -m 644 ${S}/Auth-*/Auth.php ${D}${datadir}/php/
install -m 644 ${S}/Auth-*/Auth/*.php ${D}${datadir}/php/Auth
install -m 644 ${S}/Auth-*/Auth/Container/* ${D}${datadir}/php/Auth/Container
install -m 644 ${S}/Auth-*/Auth/Frontend/* ${D}${datadir}/php/Auth/Frontend
}
FILES_${PN}="${datadir}"
| Correct license and fix error "debian/patches not found" | php-auth: Correct license and fix error "debian/patches not found"
- Correct license: PHP-3.0 & BSD-3-Clause
- Source format is 3.0 (quilt) but there is no debian patch,
empty DEBIAN_QUILT_PATCHES to avoid error "debian/patches not found".
Signed-off-by: Trung Do <64081e6b7eb2b24289b8a4e51cef9836ddf28af2@toshiba-tsdv.com>
| BitBake | mit | tienlee/meta-debian,meta-debian/meta-debian,meta-debian/meta-debian,meta-debian/meta-debian,nghiaht-tsdv/meta-debian,meta-debian/meta-debian,tienlee/meta-debian,nghiaht-tsdv/meta-debian,tienlee/meta-debian,meta-debian/meta-debian,tienlee/meta-debian,meta-debian/meta-debian,tienlee/meta-debian,nghiaht-tsdv/meta-debian,nghiaht-tsdv/meta-debian,nghiaht-tsdv/meta-debian,nghiaht-tsdv/meta-debian,tienlee/meta-debian | bitbake | ## Code Before:
DESCRIPTION = "PHP PEAR modules for creating an authentication system"
PR = "r0"
inherit debian-package
LICENSE = "PHP-3.1"
LIC_FILES_CHKSUM = "file://debian/copyright;md5=02f173e28791fe9054ea54631abf4dc3"
do_install() {
install -d ${D}${datadir}/php/Auth/Container
install -d ${D}${datadir}/php/Auth/Frontend
install -m 644 ${S}/Auth-*/Auth.php ${D}${datadir}/php/
install -m 644 ${S}/Auth-*/Auth/*.php ${D}${datadir}/php/Auth
install -m 644 ${S}/Auth-*/Auth/Container/* ${D}${datadir}/php/Auth/Container
install -m 644 ${S}/Auth-*/Auth/Frontend/* ${D}${datadir}/php/Auth/Frontend
}
FILES_${PN}="${datadir}"
## Instruction:
php-auth: Correct license and fix error "debian/patches not found"
- Correct license: PHP-3.0 & BSD-3-Clause
- Source format is 3.0 (quilt) but there is no debian patch,
empty DEBIAN_QUILT_PATCHES to avoid error "debian/patches not found".
Signed-off-by: Trung Do <64081e6b7eb2b24289b8a4e51cef9836ddf28af2@toshiba-tsdv.com>
## Code After:
DESCRIPTION = "PHP PEAR modules for creating an authentication system"
PR = "r0"
inherit debian-package
LICENSE = "PHP-3.0 & BSD-3-Clause"
LIC_FILES_CHKSUM = " \
file://Auth-1.6.4/Auth.php;beginline=9;endline=13;md5=6863d257d87ab97c7e821ca3fe49a316 \
file://Auth-1.6.4/Auth/Frontend/md5.js;endline=8;md5=915fa4679c019c47741472ce7ce6afd1 \
"
# source format is 3.0 (quilt) but there is no debian patch
DEBIAN_QUILT_PATCHES = ""
do_install() {
install -d ${D}${datadir}/php/Auth/Container
install -d ${D}${datadir}/php/Auth/Frontend
install -m 644 ${S}/Auth-*/Auth.php ${D}${datadir}/php/
install -m 644 ${S}/Auth-*/Auth/*.php ${D}${datadir}/php/Auth
install -m 644 ${S}/Auth-*/Auth/Container/* ${D}${datadir}/php/Auth/Container
install -m 644 ${S}/Auth-*/Auth/Frontend/* ${D}${datadir}/php/Auth/Frontend
}
FILES_${PN}="${datadir}"
| DESCRIPTION = "PHP PEAR modules for creating an authentication system"
PR = "r0"
inherit debian-package
- LICENSE = "PHP-3.1"
- LIC_FILES_CHKSUM = "file://debian/copyright;md5=02f173e28791fe9054ea54631abf4dc3"
+ LICENSE = "PHP-3.0 & BSD-3-Clause"
+ LIC_FILES_CHKSUM = " \
+ file://Auth-1.6.4/Auth.php;beginline=9;endline=13;md5=6863d257d87ab97c7e821ca3fe49a316 \
+ file://Auth-1.6.4/Auth/Frontend/md5.js;endline=8;md5=915fa4679c019c47741472ce7ce6afd1 \
+ "
+
+ # source format is 3.0 (quilt) but there is no debian patch
+ DEBIAN_QUILT_PATCHES = ""
do_install() {
install -d ${D}${datadir}/php/Auth/Container
install -d ${D}${datadir}/php/Auth/Frontend
install -m 644 ${S}/Auth-*/Auth.php ${D}${datadir}/php/
install -m 644 ${S}/Auth-*/Auth/*.php ${D}${datadir}/php/Auth
install -m 644 ${S}/Auth-*/Auth/Container/* ${D}${datadir}/php/Auth/Container
install -m 644 ${S}/Auth-*/Auth/Frontend/* ${D}${datadir}/php/Auth/Frontend
}
FILES_${PN}="${datadir}"
| 10 | 0.5 | 8 | 2 |
e6bfa9f7036d4ad51af84a9205616b4782a5db91 | geoportal/geoportailv3_geoportal/static-ngeo/js/backgroundselector/BackgroundselectorDirective.js | geoportal/geoportailv3_geoportal/static-ngeo/js/backgroundselector/BackgroundselectorDirective.js | import appModule from '../module.js';
/**
* @param {string} appBackgroundselectorTemplateUrl Url to backgroundselector template.
* @return {angular.IDirective}
* @ngInject
*/
const exports = function(appBackgroundselectorTemplateUrl) {
return {
restrict: 'E',
scope: {
'map': '=appBackgroundselectorMap'
},
controller: 'AppBackgroundselectorController',
controllerAs: 'ctrl',
bindToController: true,
templateUrl: appBackgroundselectorTemplateUrl
}
}
appModule.directive('appBackgroundselector', exports);
export default exports;
| import appModule from '../module.js';
/**
* @param {string} appBackgroundselectorTemplateUrl Url to backgroundselector template.
* @return {angular.IDirective}
* @ngInject
*/
const exports = function(appBackgroundselectorTemplateUrl) {
return {
restrict: 'E',
scope: {
'map': '=appBackgroundselectorMap'
},
controller: 'AppBackgroundselectorController',
controllerAs: 'ctrl',
bindToController: true,
templateUrl: appBackgroundselectorTemplateUrl
}
}
// Custom directive for the "vector tiles style" upload button
appModule.directive('customOnChange', function() {
return {
restrict: 'A',
link: function (scope, element, attrs) {
var onChangeHandler = scope.$eval(attrs.customOnChange);
element.on('change', onChangeHandler);
element.on('$destroy', function() {
element.off();
});
}
};
});
appModule.directive('appBackgroundselector', exports);
export default exports;
| Add missing customOnChange directive for the new background selector component | Add missing customOnChange directive for the new background selector component
| JavaScript | mit | Geoportail-Luxembourg/geoportailv3,Geoportail-Luxembourg/geoportailv3,Geoportail-Luxembourg/geoportailv3,Geoportail-Luxembourg/geoportailv3,Geoportail-Luxembourg/geoportailv3 | javascript | ## Code Before:
import appModule from '../module.js';
/**
* @param {string} appBackgroundselectorTemplateUrl Url to backgroundselector template.
* @return {angular.IDirective}
* @ngInject
*/
const exports = function(appBackgroundselectorTemplateUrl) {
return {
restrict: 'E',
scope: {
'map': '=appBackgroundselectorMap'
},
controller: 'AppBackgroundselectorController',
controllerAs: 'ctrl',
bindToController: true,
templateUrl: appBackgroundselectorTemplateUrl
}
}
appModule.directive('appBackgroundselector', exports);
export default exports;
## Instruction:
Add missing customOnChange directive for the new background selector component
## Code After:
import appModule from '../module.js';
/**
* @param {string} appBackgroundselectorTemplateUrl Url to backgroundselector template.
* @return {angular.IDirective}
* @ngInject
*/
const exports = function(appBackgroundselectorTemplateUrl) {
return {
restrict: 'E',
scope: {
'map': '=appBackgroundselectorMap'
},
controller: 'AppBackgroundselectorController',
controllerAs: 'ctrl',
bindToController: true,
templateUrl: appBackgroundselectorTemplateUrl
}
}
// Custom directive for the "vector tiles style" upload button
appModule.directive('customOnChange', function() {
return {
restrict: 'A',
link: function (scope, element, attrs) {
var onChangeHandler = scope.$eval(attrs.customOnChange);
element.on('change', onChangeHandler);
element.on('$destroy', function() {
element.off();
});
}
};
});
appModule.directive('appBackgroundselector', exports);
export default exports;
| import appModule from '../module.js';
/**
* @param {string} appBackgroundselectorTemplateUrl Url to backgroundselector template.
* @return {angular.IDirective}
* @ngInject
*/
const exports = function(appBackgroundselectorTemplateUrl) {
return {
restrict: 'E',
scope: {
'map': '=appBackgroundselectorMap'
},
controller: 'AppBackgroundselectorController',
controllerAs: 'ctrl',
bindToController: true,
templateUrl: appBackgroundselectorTemplateUrl
}
}
+ // Custom directive for the "vector tiles style" upload button
+ appModule.directive('customOnChange', function() {
+ return {
+ restrict: 'A',
+ link: function (scope, element, attrs) {
+ var onChangeHandler = scope.$eval(attrs.customOnChange);
+ element.on('change', onChangeHandler);
+ element.on('$destroy', function() {
+ element.off();
+ });
+ }
+ };
+ });
+
appModule.directive('appBackgroundselector', exports);
export default exports; | 14 | 0.608696 | 14 | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.