prompt stringlengths 77 6.9k | completion stringlengths 1 3.38k | text stringlengths 163 9.03k | commit stringlengths 40 40 | old_file stringlengths 4 264 | new_file stringlengths 4 264 | lang stringclasses 277
values |
|---|---|---|---|---|---|---|
<|file_sep|>package.json.diff
original:
updated:
"karma-cli": "0.0.4",
<|file_sep|>original/package.json
{
"name": "ng-simplePagination",
"version": "0.0.9",
"private": true,
"author": "Svilen Gospodinov <svilen.gospodinov@gmail.com>",
"license": "MIT",
"devDependencies": {
"karma": "^0.12.19",
"karma-jasmine": "^0.1.5",
"karma-phantomjs-launcher": "^0.1.4"
}
}
<|file_sep|>current/package.json
{
"name": "ng-simplePagination",
"version": "0.0.9",
"private": true,
"author": "Svilen Gospodinov <svilen.gospodinov@gmail.com>",
"license": "MIT",
"devDependencies": {
"karma": "^0.12.19",
"karma-cli": "0.0.4",
"karma-jasmine": "^0.1.5",
"karma-phantomjs-launcher": "^0.1.4"
}
}
<|file_sep|>updated/package.json | {
"name": "ng-simplePagination",
"version": "0.0.9",
"private": true,
"author": "Svilen Gospodinov <svilen.gospodinov@gmail.com>",
"license": "MIT",
"devDependencies": {
"karma": "^0.12.19",
"karma-cli": "0.0.4",
"karma-jasmine": "^0.1.5",
"karma-phantomjs-launcher": "^0.1.4"
},
"scripts": {
"test": "karma start"
}
} | <|file_sep|>package.json.diff
original:
updated:
"karma-cli": "0.0.4",
<|file_sep|>original/package.json
{
"name": "ng-simplePagination",
"version": "0.0.9",
"private": true,
"author": "Svilen Gospodinov <svilen.gospodinov@gmail.com>",
"license": "MIT",
"devDependencies": {
"karma": "^0.12.19",
"karma-jasmine": "^0.1.5",
"karma-phantomjs-launcher": "^0.1.4"
}
}
<|file_sep|>current/package.json
{
"name": "ng-simplePagination",
"version": "0.0.9",
"private": true,
"author": "Svilen Gospodinov <svilen.gospodinov@gmail.com>",
"license": "MIT",
"devDependencies": {
"karma": "^0.12.19",
"karma-cli": "0.0.4",
"karma-jasmine": "^0.1.5",
"karma-phantomjs-launcher": "^0.1.4"
}
}
<|file_sep|>updated/package.json
{
"name": "ng-simplePagination",
"version": "0.0.9",
"private": true,
"author": "Svilen Gospodinov <svilen.gospodinov@gmail.com>",
"license": "MIT",
"devDependencies": {
"karma": "^0.12.19",
"karma-cli": "0.0.4",
"karma-jasmine": "^0.1.5",
"karma-phantomjs-launcher": "^0.1.4"
},
"scripts": {
"test": "karma start"
}
} | 89210657ca9f876acee97d8a6732c500c8e69704 | package.json | package.json | JSON |
<|file_sep|>subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp.diff
original:
updated:
#include <bitset>
<|file_sep|>subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp.diff
original:
#define DATA_SIZE (1<<8)
updated:
<|file_sep|>subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp.diff
original:
updated:
enum{
read,
write
};
<|file_sep|>subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp.diff
original:
updated:
// Memory simulated
unsigned char rawMemory[MEM_SIZE];
int cacheDirection[CACHE_SIZE];
unsigned char cacheData[CACHE_SIZE];
// To simulate the circular queue
int front = 0, tail = 0;
// Initialize the cached Directions
memset(cacheDirection, -1, sizeof cacheDirection);
<|file_sep|>subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp.diff
original:
// 19...0 bits for dir,
updated:
cmd = rand();
// 19...0 bits for dir
dir = ((cmd)&(0xFFFFF));
<|file_sep|>original/subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp
int dir, cmd, op;
unsigned char dat, wait;
// Plant the seed
srand(time(0));
// Infinit Loop
while(true){
// Each command is 32 bits,
// 19...0 bits for dir,
// 27...20 bits for data
// 28 bit indicates operation type 1:read 0:write
cmd = rand();
dir = cmd & 0xFFFFF;
dat = (cmd & 0xF00000)>>20;
op = (cmd>>28)&1;
printf("op=%c on dir(0x%x) data:%d\n","wr"[op],dir,dat);
wait = getchar();
}
return 0;
}
<|file_sep|>current/subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp
// Plant the seed
srand(time(0));
// Infinit Loop
while(true){
// Each command is 32 bits,
cmd = rand();
// 19...0 bits for dir
dir = ((cmd)&(0xFFFFF));
// 27...20 bits for data
// 28 bit indicates operation type 1:read 0:write
cmd = rand();
dir = cmd & 0xFFFFF;
dat = (cmd & 0xF00000)>>20;
op = (cmd>>28)&1;
printf("op=%c on dir(0x%x) data:%d\n","wr"[op],dir,dat);
wait = getchar();
}
return 0;
}
<|file_sep|>updated/subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp |
// Infinit Loop
while(true){
// Each command is 32 bits,
cmd = rand();
// 19...0 bits for dir
dir = ((cmd)&(0xFFFFF));
// 27...20 bits for data
dat = (((cmd)&(0xF00000))>>20);
// 28 bit indicates operation type 0:read 1:write
op = ((cmd>>28)&(1));
//Perform operation according
if(op == read) {
} else { // write
}
wait = getchar();
} | <|file_sep|>subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp.diff
original:
updated:
#include <bitset>
<|file_sep|>subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp.diff
original:
#define DATA_SIZE (1<<8)
updated:
<|file_sep|>subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp.diff
original:
updated:
enum{
read,
write
};
<|file_sep|>subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp.diff
original:
updated:
// Memory simulated
unsigned char rawMemory[MEM_SIZE];
int cacheDirection[CACHE_SIZE];
unsigned char cacheData[CACHE_SIZE];
// To simulate the circular queue
int front = 0, tail = 0;
// Initialize the cached Directions
memset(cacheDirection, -1, sizeof cacheDirection);
<|file_sep|>subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp.diff
original:
// 19...0 bits for dir,
updated:
cmd = rand();
// 19...0 bits for dir
dir = ((cmd)&(0xFFFFF));
<|file_sep|>original/subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp
int dir, cmd, op;
unsigned char dat, wait;
// Plant the seed
srand(time(0));
// Infinit Loop
while(true){
// Each command is 32 bits,
// 19...0 bits for dir,
// 27...20 bits for data
// 28 bit indicates operation type 1:read 0:write
cmd = rand();
dir = cmd & 0xFFFFF;
dat = (cmd & 0xF00000)>>20;
op = (cmd>>28)&1;
printf("op=%c on dir(0x%x) data:%d\n","wr"[op],dir,dat);
wait = getchar();
}
return 0;
}
<|file_sep|>current/subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp
// Plant the seed
srand(time(0));
// Infinit Loop
while(true){
// Each command is 32 bits,
cmd = rand();
// 19...0 bits for dir
dir = ((cmd)&(0xFFFFF));
// 27...20 bits for data
// 28 bit indicates operation type 1:read 0:write
cmd = rand();
dir = cmd & 0xFFFFF;
dat = (cmd & 0xF00000)>>20;
op = (cmd>>28)&1;
printf("op=%c on dir(0x%x) data:%d\n","wr"[op],dir,dat);
wait = getchar();
}
return 0;
}
<|file_sep|>updated/subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp
// Infinit Loop
while(true){
// Each command is 32 bits,
cmd = rand();
// 19...0 bits for dir
dir = ((cmd)&(0xFFFFF));
// 27...20 bits for data
dat = (((cmd)&(0xF00000))>>20);
// 28 bit indicates operation type 0:read 1:write
op = ((cmd>>28)&(1));
//Perform operation according
if(op == read) {
} else { // write
}
wait = getchar();
} | 09c4779cb1473cc885b4175548ad6df32207dead | subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp | subjects/Microcomputer-Architecture/cache-simulation/full-back.cpp | C++ |
<|file_sep|>actionmailer/lib/action_mailer/log_subscriber.rb.diff
original:
"\nSent mail to #{recipients} (#{event.duration.round(1)}ms)"
updated:
"Sent mail to #{recipients} (#{event.duration.round(1)}ms)"
<|file_sep|>actionmailer/lib/action_mailer/log_subscriber.rb.diff
original:
info { "\nReceived mail (#{event.duration.round(1)}ms)" }
updated:
info { "Received mail (#{event.duration.round(1)}ms)" }
<|file_sep|>original/actionmailer/lib/action_mailer/log_subscriber.rb
def receive(event)
info { "\nReceived mail (#{event.duration.round(1)}ms)" }
debug { event.payload[:mail] }
end
# An email was generated.
def process(event)
debug do
mailer = event.payload[:mailer]
action = event.payload[:action]
"\n#{mailer}##{action}: processed outbound mail in #{event.duration.round(1)}ms"
end
end
# Use the logger configured for ActionMailer::Base.
def logger
ActionMailer::Base.logger
end
end
end
<|file_sep|>current/actionmailer/lib/action_mailer/log_subscriber.rb
def receive(event)
info { "Received mail (#{event.duration.round(1)}ms)" }
debug { event.payload[:mail] }
end
# An email was generated.
def process(event)
debug do
mailer = event.payload[:mailer]
action = event.payload[:action]
"\n#{mailer}##{action}: processed outbound mail in #{event.duration.round(1)}ms"
end
end
# Use the logger configured for ActionMailer::Base.
def logger
ActionMailer::Base.logger
end
end
end
<|file_sep|>updated/actionmailer/lib/action_mailer/log_subscriber.rb | def receive(event)
info { "Received mail (#{event.duration.round(1)}ms)" }
debug { event.payload[:mail] }
end
# An email was generated.
def process(event)
debug do
mailer = event.payload[:mailer]
action = event.payload[:action]
"#{mailer}##{action}: processed outbound mail in #{event.duration.round(1)}ms"
end
end
# Use the logger configured for ActionMailer::Base.
def logger
ActionMailer::Base.logger
end
end
end
| <|file_sep|>actionmailer/lib/action_mailer/log_subscriber.rb.diff
original:
"\nSent mail to #{recipients} (#{event.duration.round(1)}ms)"
updated:
"Sent mail to #{recipients} (#{event.duration.round(1)}ms)"
<|file_sep|>actionmailer/lib/action_mailer/log_subscriber.rb.diff
original:
info { "\nReceived mail (#{event.duration.round(1)}ms)" }
updated:
info { "Received mail (#{event.duration.round(1)}ms)" }
<|file_sep|>original/actionmailer/lib/action_mailer/log_subscriber.rb
def receive(event)
info { "\nReceived mail (#{event.duration.round(1)}ms)" }
debug { event.payload[:mail] }
end
# An email was generated.
def process(event)
debug do
mailer = event.payload[:mailer]
action = event.payload[:action]
"\n#{mailer}##{action}: processed outbound mail in #{event.duration.round(1)}ms"
end
end
# Use the logger configured for ActionMailer::Base.
def logger
ActionMailer::Base.logger
end
end
end
<|file_sep|>current/actionmailer/lib/action_mailer/log_subscriber.rb
def receive(event)
info { "Received mail (#{event.duration.round(1)}ms)" }
debug { event.payload[:mail] }
end
# An email was generated.
def process(event)
debug do
mailer = event.payload[:mailer]
action = event.payload[:action]
"\n#{mailer}##{action}: processed outbound mail in #{event.duration.round(1)}ms"
end
end
# Use the logger configured for ActionMailer::Base.
def logger
ActionMailer::Base.logger
end
end
end
<|file_sep|>updated/actionmailer/lib/action_mailer/log_subscriber.rb
def receive(event)
info { "Received mail (#{event.duration.round(1)}ms)" }
debug { event.payload[:mail] }
end
# An email was generated.
def process(event)
debug do
mailer = event.payload[:mailer]
action = event.payload[:action]
"#{mailer}##{action}: processed outbound mail in #{event.duration.round(1)}ms"
end
end
# Use the logger configured for ActionMailer::Base.
def logger
ActionMailer::Base.logger
end
end
end
| 5e4f82a5f39054b27232fbf0840496ab5ddb06a7 | actionmailer/lib/action_mailer/log_subscriber.rb | actionmailer/lib/action_mailer/log_subscriber.rb | Ruby |
<|file_sep|>src/applications/files/transform/__tests__/PhabricatorFileTransformTestCase.php.diff
original:
updated:
protected function getPhabricatorTestCaseConfiguration() {
return array(
self::PHABRICATOR_TESTCONFIG_BUILD_STORAGE_FIXTURES => true,
);
}
<|file_sep|>original/src/applications/files/transform/__tests__/PhabricatorFileTransformTestCase.php
<?php
final class PhabricatorFileTransformTestCase extends PhabricatorTestCase {
public function testGetAllTransforms() {
PhabricatorFileTransform::getAllTransforms();
$this->assertTrue(true);
}
}
<|file_sep|>current/src/applications/files/transform/__tests__/PhabricatorFileTransformTestCase.php
<?php
final class PhabricatorFileTransformTestCase extends PhabricatorTestCase {
protected function getPhabricatorTestCaseConfiguration() {
return array(
self::PHABRICATOR_TESTCONFIG_BUILD_STORAGE_FIXTURES => true,
);
}
public function testGetAllTransforms() {
PhabricatorFileTransform::getAllTransforms();
$this->assertTrue(true);
}
}
<|file_sep|>updated/src/applications/files/transform/__tests__/PhabricatorFileTransformTestCase.php | return array(
self::PHABRICATOR_TESTCONFIG_BUILD_STORAGE_FIXTURES => true,
);
}
public function testGetAllTransforms() {
PhabricatorFileTransform::getAllTransforms();
$this->assertTrue(true);
}
public function testThumbTransformDefaults() {
$xforms = PhabricatorFileTransform::getAllTransforms();
$file = new PhabricatorFile();
foreach ($xforms as $xform) {
if (!($xform instanceof PhabricatorFileThumbnailTransform)) {
continue;
}
// For thumbnails, generate the default thumbnail. This should be able
// to generate something rather than throwing an exception because we | <|file_sep|>src/applications/files/transform/__tests__/PhabricatorFileTransformTestCase.php.diff
original:
updated:
protected function getPhabricatorTestCaseConfiguration() {
return array(
self::PHABRICATOR_TESTCONFIG_BUILD_STORAGE_FIXTURES => true,
);
}
<|file_sep|>original/src/applications/files/transform/__tests__/PhabricatorFileTransformTestCase.php
<?php
final class PhabricatorFileTransformTestCase extends PhabricatorTestCase {
public function testGetAllTransforms() {
PhabricatorFileTransform::getAllTransforms();
$this->assertTrue(true);
}
}
<|file_sep|>current/src/applications/files/transform/__tests__/PhabricatorFileTransformTestCase.php
<?php
final class PhabricatorFileTransformTestCase extends PhabricatorTestCase {
protected function getPhabricatorTestCaseConfiguration() {
return array(
self::PHABRICATOR_TESTCONFIG_BUILD_STORAGE_FIXTURES => true,
);
}
public function testGetAllTransforms() {
PhabricatorFileTransform::getAllTransforms();
$this->assertTrue(true);
}
}
<|file_sep|>updated/src/applications/files/transform/__tests__/PhabricatorFileTransformTestCase.php
return array(
self::PHABRICATOR_TESTCONFIG_BUILD_STORAGE_FIXTURES => true,
);
}
public function testGetAllTransforms() {
PhabricatorFileTransform::getAllTransforms();
$this->assertTrue(true);
}
public function testThumbTransformDefaults() {
$xforms = PhabricatorFileTransform::getAllTransforms();
$file = new PhabricatorFile();
foreach ($xforms as $xform) {
if (!($xform instanceof PhabricatorFileThumbnailTransform)) {
continue;
}
// For thumbnails, generate the default thumbnail. This should be able
// to generate something rather than throwing an exception because we | bb2b91d28e7e9c81722949cbed6624f46f694d1c | src/applications/files/transform/__tests__/PhabricatorFileTransformTestCase.php | src/applications/files/transform/__tests__/PhabricatorFileTransformTestCase.php | PHP |
<|file_sep|>original/manifests/cf-manifest/operations.d/530-ipsec.yml
<|file_sep|>current/manifests/cf-manifest/operations.d/530-ipsec.yml
<|file_sep|>updated/manifests/cf-manifest/operations.d/530-ipsec.yml | ---
- type: replace
path: /releases/-
value:
name: ipsec
sha1: 6a9e252162519e50f6c511c489ba304a286d496a
url: https://s3-eu-west-1.amazonaws.com/gds-paas-build-releases/ipsec-0.1.3.tgz
version: 0.1.3
- type: replace
path: /instance_groups/name=router/jobs/-
value:
name: racoon
release: ipsec
properties:
racoon:
ports:
- name: router
targets: ((terraform_outputs_cell_subnet_cidr_blocks))
certificate_authority_private_key: "((ipsec_ca.private_key))" | <|file_sep|>original/manifests/cf-manifest/operations.d/530-ipsec.yml
<|file_sep|>current/manifests/cf-manifest/operations.d/530-ipsec.yml
<|file_sep|>updated/manifests/cf-manifest/operations.d/530-ipsec.yml
---
- type: replace
path: /releases/-
value:
name: ipsec
sha1: 6a9e252162519e50f6c511c489ba304a286d496a
url: https://s3-eu-west-1.amazonaws.com/gds-paas-build-releases/ipsec-0.1.3.tgz
version: 0.1.3
- type: replace
path: /instance_groups/name=router/jobs/-
value:
name: racoon
release: ipsec
properties:
racoon:
ports:
- name: router
targets: ((terraform_outputs_cell_subnet_cidr_blocks))
certificate_authority_private_key: "((ipsec_ca.private_key))" | b3876aeb793c476c146dd9fbc30781886b1024b7 | manifests/cf-manifest/operations.d/530-ipsec.yml | manifests/cf-manifest/operations.d/530-ipsec.yml | YAML |
<|file_sep|>original/rci.bat
:: Revision: 01/29/2011 10:20:16 AM
@echo off
setlocal
:: echo initial:
:: echo %%0 = %0
:: echo %%* = %*
:: fix ruby.exe invocation when explicitly disabling RubyGems
if "x%1" == "x--disable-gems" (
set NOGEM=%1
for /F "tokens=1*" %%i in ("%*") do set RB_ARGS=%%j
) else (
set NOGEM=
set RB_ARGS=%*
)
:: echo pre call ruby:
:: echo NOGEM = %NOGEM%
:: echo RB_ARGS = %RB_ARGS%
ruby.exe %NOGEM% -x %~f0 %RB_ARGS%
<|file_sep|>current/rci.bat
:: Revision: 01/29/2011 10:20:16 AM
@echo off
setlocal
:: echo initial:
:: echo %%0 = %0
:: echo %%* = %*
:: fix ruby.exe invocation when explicitly disabling RubyGems
if "x%1" == "x--disable-gems" (
set NOGEM=%1
for /F "tokens=1*" %%i in ("%*") do set RB_ARGS=%%j
) else (
set NOGEM=
set RB_ARGS=%*
)
:: echo pre call ruby:
:: echo NOGEM = %NOGEM%
:: echo RB_ARGS = %RB_ARGS%
ruby.exe %NOGEM% -x %~f0 %RB_ARGS%
<|file_sep|>updated/rci.bat | :: Revision: 01/29/2011 10:20:16 AM
@echo off
setlocal
:: echo initial:
:: echo %%0 = %0
:: echo %%* = %*
:: fix ruby.exe invocation when explicitly disabling RubyGems
if "x%1" == "x--disable-gems" (
set NOGEM=%1
for /F "tokens=1*" %%i in ("%*") do (
set RB_ARGS=%%j
)
) else (
set NOGEM=
set RB_ARGS=%*
)
:: echo pre call ruby:
:: echo NOGEM = %NOGEM%
:: echo RB_ARGS = %RB_ARGS% | <|file_sep|>original/rci.bat
:: Revision: 01/29/2011 10:20:16 AM
@echo off
setlocal
:: echo initial:
:: echo %%0 = %0
:: echo %%* = %*
:: fix ruby.exe invocation when explicitly disabling RubyGems
if "x%1" == "x--disable-gems" (
set NOGEM=%1
for /F "tokens=1*" %%i in ("%*") do set RB_ARGS=%%j
) else (
set NOGEM=
set RB_ARGS=%*
)
:: echo pre call ruby:
:: echo NOGEM = %NOGEM%
:: echo RB_ARGS = %RB_ARGS%
ruby.exe %NOGEM% -x %~f0 %RB_ARGS%
<|file_sep|>current/rci.bat
:: Revision: 01/29/2011 10:20:16 AM
@echo off
setlocal
:: echo initial:
:: echo %%0 = %0
:: echo %%* = %*
:: fix ruby.exe invocation when explicitly disabling RubyGems
if "x%1" == "x--disable-gems" (
set NOGEM=%1
for /F "tokens=1*" %%i in ("%*") do set RB_ARGS=%%j
) else (
set NOGEM=
set RB_ARGS=%*
)
:: echo pre call ruby:
:: echo NOGEM = %NOGEM%
:: echo RB_ARGS = %RB_ARGS%
ruby.exe %NOGEM% -x %~f0 %RB_ARGS%
<|file_sep|>updated/rci.bat
:: Revision: 01/29/2011 10:20:16 AM
@echo off
setlocal
:: echo initial:
:: echo %%0 = %0
:: echo %%* = %*
:: fix ruby.exe invocation when explicitly disabling RubyGems
if "x%1" == "x--disable-gems" (
set NOGEM=%1
for /F "tokens=1*" %%i in ("%*") do (
set RB_ARGS=%%j
)
) else (
set NOGEM=
set RB_ARGS=%*
)
:: echo pre call ruby:
:: echo NOGEM = %NOGEM%
:: echo RB_ARGS = %RB_ARGS% | 1adef8ad02df39ff478ab7af21d42925ed011e3c | rci.bat | rci.bat | Batchfile |
<|file_sep|>original/MKLog.podspec
# Be sure to run `pod lib lint MKLog.podspec' to ensure this is a
# valid spec and remove all comments before submitting the spec.
#
# Any lines starting with a # are optional, but encouraged
#
# To learn more about a Podspec see http://guides.cocoapods.org/syntax/podspec.html
#
Pod::Spec.new do |s|
s.name = "MKLog"
s.version = "0.1.1"
s.summary = "Lightweight log with multiple log levels"
s.description = <<-DESC
A lightweight logger implementation for Objective-C with multiple log levels.
DESC
s.homepage = "https://github.com/mikumi/MKLog"
s.license = 'MIT'
s.author = { "Michael Kuck" => "me@michael-kuck.com" }
s.source = { :git => "https://github.com/mikumi/MKLog.git", :tag => s.version.to_s }
s.platform = :ios, '7.0'
<|file_sep|>current/MKLog.podspec
# Be sure to run `pod lib lint MKLog.podspec' to ensure this is a
# valid spec and remove all comments before submitting the spec.
#
# Any lines starting with a # are optional, but encouraged
#
# To learn more about a Podspec see http://guides.cocoapods.org/syntax/podspec.html
#
Pod::Spec.new do |s|
s.name = "MKLog"
s.version = "0.1.1"
s.summary = "Lightweight log with multiple log levels"
s.description = <<-DESC
A lightweight logger implementation for Objective-C with multiple log levels.
DESC
s.homepage = "https://github.com/mikumi/MKLog"
s.license = 'MIT'
s.author = { "Michael Kuck" => "me@michael-kuck.com" }
s.source = { :git => "https://github.com/mikumi/MKLog.git", :tag => s.version.to_s }
s.platform = :ios, '7.0'
<|file_sep|>updated/MKLog.podspec | # Be sure to run `pod lib lint MKLog.podspec' to ensure this is a
# valid spec and remove all comments before submitting the spec.
#
# Any lines starting with a # are optional, but encouraged
#
# To learn more about a Podspec see http://guides.cocoapods.org/syntax/podspec.html
#
Pod::Spec.new do |s|
s.name = "MKLog"
s.version = "0.1.2"
s.summary = "Lightweight log with multiple log levels"
s.description = <<-DESC
A lightweight logger implementation for Objective-C with multiple log levels.
DESC
s.homepage = "https://github.com/mikumi/MKLog"
s.license = 'MIT'
s.author = { "Michael Kuck" => "me@michael-kuck.com" }
s.source = { :git => "https://github.com/mikumi/MKLog.git", :tag => s.version.to_s }
s.platform = :ios, '7.0' | <|file_sep|>original/MKLog.podspec
# Be sure to run `pod lib lint MKLog.podspec' to ensure this is a
# valid spec and remove all comments before submitting the spec.
#
# Any lines starting with a # are optional, but encouraged
#
# To learn more about a Podspec see http://guides.cocoapods.org/syntax/podspec.html
#
Pod::Spec.new do |s|
s.name = "MKLog"
s.version = "0.1.1"
s.summary = "Lightweight log with multiple log levels"
s.description = <<-DESC
A lightweight logger implementation for Objective-C with multiple log levels.
DESC
s.homepage = "https://github.com/mikumi/MKLog"
s.license = 'MIT'
s.author = { "Michael Kuck" => "me@michael-kuck.com" }
s.source = { :git => "https://github.com/mikumi/MKLog.git", :tag => s.version.to_s }
s.platform = :ios, '7.0'
<|file_sep|>current/MKLog.podspec
# Be sure to run `pod lib lint MKLog.podspec' to ensure this is a
# valid spec and remove all comments before submitting the spec.
#
# Any lines starting with a # are optional, but encouraged
#
# To learn more about a Podspec see http://guides.cocoapods.org/syntax/podspec.html
#
Pod::Spec.new do |s|
s.name = "MKLog"
s.version = "0.1.1"
s.summary = "Lightweight log with multiple log levels"
s.description = <<-DESC
A lightweight logger implementation for Objective-C with multiple log levels.
DESC
s.homepage = "https://github.com/mikumi/MKLog"
s.license = 'MIT'
s.author = { "Michael Kuck" => "me@michael-kuck.com" }
s.source = { :git => "https://github.com/mikumi/MKLog.git", :tag => s.version.to_s }
s.platform = :ios, '7.0'
<|file_sep|>updated/MKLog.podspec
# Be sure to run `pod lib lint MKLog.podspec' to ensure this is a
# valid spec and remove all comments before submitting the spec.
#
# Any lines starting with a # are optional, but encouraged
#
# To learn more about a Podspec see http://guides.cocoapods.org/syntax/podspec.html
#
Pod::Spec.new do |s|
s.name = "MKLog"
s.version = "0.1.2"
s.summary = "Lightweight log with multiple log levels"
s.description = <<-DESC
A lightweight logger implementation for Objective-C with multiple log levels.
DESC
s.homepage = "https://github.com/mikumi/MKLog"
s.license = 'MIT'
s.author = { "Michael Kuck" => "me@michael-kuck.com" }
s.source = { :git => "https://github.com/mikumi/MKLog.git", :tag => s.version.to_s }
s.platform = :ios, '7.0' | 525e6fd1e426efea83527d116a42cc5efab632f0 | MKLog.podspec | MKLog.podspec | Ruby |
<|file_sep|>original/examples/SetPictures.cs
if(args.Length < 2) {
Console.Error.WriteLine("USAGE: mono SetPictures.exe AUDIO_PATH IMAGE_PATH_1[...IMAGE_PATH_N]");
return;
}
TagLib.File file = TagLib.File.Create(args[0]);
Console.WriteLine("Current picture count: " + file.Tag.Pictures.Length);
Picture [] pictures = new Picture[args.Length - 1];
for(int i = 1; i < args.Length; i++) {
Picture picture = Picture.CreateFromPath(args[i]);
pictures[i - 1] = picture;
}
file.Tag.Pictures = pictures;
file.Save();
Console.WriteLine("New picture count: " + file.Tag.Pictures.Length);
}
}
<|file_sep|>current/examples/SetPictures.cs
if(args.Length < 2) {
Console.Error.WriteLine("USAGE: mono SetPictures.exe AUDIO_PATH IMAGE_PATH_1[...IMAGE_PATH_N]");
return;
}
TagLib.File file = TagLib.File.Create(args[0]);
Console.WriteLine("Current picture count: " + file.Tag.Pictures.Length);
Picture [] pictures = new Picture[args.Length - 1];
for(int i = 1; i < args.Length; i++) {
Picture picture = Picture.CreateFromPath(args[i]);
pictures[i - 1] = picture;
}
file.Tag.Pictures = pictures;
file.Save();
Console.WriteLine("New picture count: " + file.Tag.Pictures.Length);
}
}
<|file_sep|>updated/examples/SetPictures.cs | if(args.Length < 2) {
Console.Error.WriteLine("USAGE: mono SetPictures.exe AUDIO_PATH IMAGE_PATH_1[...IMAGE_PATH_N]");
return;
}
TagLib.File file = TagLib.File.Create(args[0]);
Console.WriteLine("Current picture count: " + file.Tag.Pictures.Length);
Picture [] pictures = new Picture[args.Length - 1];
for(int i = 1; i < args.Length; i++) {
Picture picture = new Picture(args[i]);
pictures[i - 1] = picture;
}
file.Tag.Pictures = pictures;
file.Save();
Console.WriteLine("New picture count: " + file.Tag.Pictures.Length);
}
} | <|file_sep|>original/examples/SetPictures.cs
if(args.Length < 2) {
Console.Error.WriteLine("USAGE: mono SetPictures.exe AUDIO_PATH IMAGE_PATH_1[...IMAGE_PATH_N]");
return;
}
TagLib.File file = TagLib.File.Create(args[0]);
Console.WriteLine("Current picture count: " + file.Tag.Pictures.Length);
Picture [] pictures = new Picture[args.Length - 1];
for(int i = 1; i < args.Length; i++) {
Picture picture = Picture.CreateFromPath(args[i]);
pictures[i - 1] = picture;
}
file.Tag.Pictures = pictures;
file.Save();
Console.WriteLine("New picture count: " + file.Tag.Pictures.Length);
}
}
<|file_sep|>current/examples/SetPictures.cs
if(args.Length < 2) {
Console.Error.WriteLine("USAGE: mono SetPictures.exe AUDIO_PATH IMAGE_PATH_1[...IMAGE_PATH_N]");
return;
}
TagLib.File file = TagLib.File.Create(args[0]);
Console.WriteLine("Current picture count: " + file.Tag.Pictures.Length);
Picture [] pictures = new Picture[args.Length - 1];
for(int i = 1; i < args.Length; i++) {
Picture picture = Picture.CreateFromPath(args[i]);
pictures[i - 1] = picture;
}
file.Tag.Pictures = pictures;
file.Save();
Console.WriteLine("New picture count: " + file.Tag.Pictures.Length);
}
}
<|file_sep|>updated/examples/SetPictures.cs
if(args.Length < 2) {
Console.Error.WriteLine("USAGE: mono SetPictures.exe AUDIO_PATH IMAGE_PATH_1[...IMAGE_PATH_N]");
return;
}
TagLib.File file = TagLib.File.Create(args[0]);
Console.WriteLine("Current picture count: " + file.Tag.Pictures.Length);
Picture [] pictures = new Picture[args.Length - 1];
for(int i = 1; i < args.Length; i++) {
Picture picture = new Picture(args[i]);
pictures[i - 1] = picture;
}
file.Tag.Pictures = pictures;
file.Save();
Console.WriteLine("New picture count: " + file.Tag.Pictures.Length);
}
} | 9288a468d586e88533ebd380ba0eabe71e90a254 | examples/SetPictures.cs | examples/SetPictures.cs | C# |
<|file_sep|>original/.travis.yml
language: php
php:
- 5.5
- 5.6
addons:
code_climate:
repo_token: 26489f9e854fcdf7e7660ba29c1455694685465b1f90329a79f7d2bf448acb61
install:
- rm composer.lock
- composer install
- php artisan env
- mv -v .env.testing .env
script:
- phpunit
after_script:
- php vendor/bin/coveralls
<|file_sep|>current/.travis.yml
language: php
php:
- 5.5
- 5.6
addons:
code_climate:
repo_token: 26489f9e854fcdf7e7660ba29c1455694685465b1f90329a79f7d2bf448acb61
install:
- rm composer.lock
- composer install
- php artisan env
- mv -v .env.testing .env
script:
- phpunit
after_script:
- php vendor/bin/coveralls
<|file_sep|>updated/.travis.yml | language: php
php:
- 5.5
- 5.6
addons:
code_climate:
repo_token: 26489f9e854fcdf7e7660ba29c1455694685465b1f90329a79f7d2bf448acb61
install:
- rm composer.lock
- composer install
- php artisan env
- mv -v .env.testing .env
script:
- phpunit --debug
after_script:
- php vendor/bin/coveralls | <|file_sep|>original/.travis.yml
language: php
php:
- 5.5
- 5.6
addons:
code_climate:
repo_token: 26489f9e854fcdf7e7660ba29c1455694685465b1f90329a79f7d2bf448acb61
install:
- rm composer.lock
- composer install
- php artisan env
- mv -v .env.testing .env
script:
- phpunit
after_script:
- php vendor/bin/coveralls
<|file_sep|>current/.travis.yml
language: php
php:
- 5.5
- 5.6
addons:
code_climate:
repo_token: 26489f9e854fcdf7e7660ba29c1455694685465b1f90329a79f7d2bf448acb61
install:
- rm composer.lock
- composer install
- php artisan env
- mv -v .env.testing .env
script:
- phpunit
after_script:
- php vendor/bin/coveralls
<|file_sep|>updated/.travis.yml
language: php
php:
- 5.5
- 5.6
addons:
code_climate:
repo_token: 26489f9e854fcdf7e7660ba29c1455694685465b1f90329a79f7d2bf448acb61
install:
- rm composer.lock
- composer install
- php artisan env
- mv -v .env.testing .env
script:
- phpunit --debug
after_script:
- php vendor/bin/coveralls | 20c2ff344344be18168ef66398eb6c894cb4a69f | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/mbed_lib.json
{
"name": "core",
"config": {
"stdio-convert-newlines": {
"help": "Enable conversion to standard newlines on stdin/stdout",
"value": false
},
"stdio-baud-rate": {
"help": "Baud rate for stdio",
"value": 9600
}
}
}
<|file_sep|>current/mbed_lib.json
{
"name": "core",
"config": {
"stdio-convert-newlines": {
"help": "Enable conversion to standard newlines on stdin/stdout",
"value": false
},
"stdio-baud-rate": {
"help": "Baud rate for stdio",
"value": 9600
}
}
}
<|file_sep|>updated/mbed_lib.json | {
"name": "core",
"config": {
"stdio-convert-newlines": {
"help": "Enable conversion to standard newlines on stdin/stdout",
"value": false
},
"stdio-baud-rate": {
"help": "Baud rate for stdio",
"value": 9600
}
},
"target_overrides": {
"EFM32": {
"stdio-baud-rate": 115200
}
}
} | <|file_sep|>original/mbed_lib.json
{
"name": "core",
"config": {
"stdio-convert-newlines": {
"help": "Enable conversion to standard newlines on stdin/stdout",
"value": false
},
"stdio-baud-rate": {
"help": "Baud rate for stdio",
"value": 9600
}
}
}
<|file_sep|>current/mbed_lib.json
{
"name": "core",
"config": {
"stdio-convert-newlines": {
"help": "Enable conversion to standard newlines on stdin/stdout",
"value": false
},
"stdio-baud-rate": {
"help": "Baud rate for stdio",
"value": 9600
}
}
}
<|file_sep|>updated/mbed_lib.json
{
"name": "core",
"config": {
"stdio-convert-newlines": {
"help": "Enable conversion to standard newlines on stdin/stdout",
"value": false
},
"stdio-baud-rate": {
"help": "Baud rate for stdio",
"value": 9600
}
},
"target_overrides": {
"EFM32": {
"stdio-baud-rate": 115200
}
}
} | 3206d96487ea6f2419b0bbd250f655d6c508f7a3 | mbed_lib.json | mbed_lib.json | JSON |
<|file_sep|>original/setup.py
long_description=long_description,
long_description_content_type='text/markdown',
author='Valentin Berlier',
author_email='berlier.v@gmail.com',
url='https://github.com/vberlier/nbtlib',
platforms=['any'],
python_requires='>=3.6',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Software Development :: Libraries :: Python Modules',
],
keywords='nbt schema minecraft package library parser reader module',
packages=find_packages(),
<|file_sep|>current/setup.py
long_description=long_description,
long_description_content_type='text/markdown',
author='Valentin Berlier',
author_email='berlier.v@gmail.com',
url='https://github.com/vberlier/nbtlib',
platforms=['any'],
python_requires='>=3.6',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Software Development :: Libraries :: Python Modules',
],
keywords='nbt schema minecraft package library parser reader module',
packages=find_packages(),
<|file_sep|>updated/setup.py | long_description=long_description,
long_description_content_type='text/markdown',
author='Valentin Berlier',
author_email='berlier.v@gmail.com',
url='https://github.com/vberlier/nbtlib',
platforms=['any'],
python_requires='>=3.6',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Software Development :: Libraries :: Python Modules',
],
keywords='nbt schema minecraft package library parser reader module',
packages=find_packages(), | <|file_sep|>original/setup.py
long_description=long_description,
long_description_content_type='text/markdown',
author='Valentin Berlier',
author_email='berlier.v@gmail.com',
url='https://github.com/vberlier/nbtlib',
platforms=['any'],
python_requires='>=3.6',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Software Development :: Libraries :: Python Modules',
],
keywords='nbt schema minecraft package library parser reader module',
packages=find_packages(),
<|file_sep|>current/setup.py
long_description=long_description,
long_description_content_type='text/markdown',
author='Valentin Berlier',
author_email='berlier.v@gmail.com',
url='https://github.com/vberlier/nbtlib',
platforms=['any'],
python_requires='>=3.6',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Software Development :: Libraries :: Python Modules',
],
keywords='nbt schema minecraft package library parser reader module',
packages=find_packages(),
<|file_sep|>updated/setup.py
long_description=long_description,
long_description_content_type='text/markdown',
author='Valentin Berlier',
author_email='berlier.v@gmail.com',
url='https://github.com/vberlier/nbtlib',
platforms=['any'],
python_requires='>=3.6',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Software Development :: Libraries :: Python Modules',
],
keywords='nbt schema minecraft package library parser reader module',
packages=find_packages(), | c2297672afff206bdae91aa479cb946b932a35e9 | setup.py | setup.py | Python |
<|file_sep|>original/WindowsStore/Service/CameraService.cs
using MyDocs.Common.Model;
using MyDocs.WindowsStore.Storage;
using System;
using System.Threading.Tasks;
using Windows.Media.Capture;
namespace MyDocs.WindowsStore.Service
{
public class CameraService : ICameraService
{
public async Task<Photo> CapturePhotoAsync()
{
CameraCaptureUI camera = new CameraCaptureUI();
var file = await camera.CaptureFileAsync(CameraCaptureUIMode.Photo);
if (file == null) {
return null;
}
return new Photo(DateTime.Now.ToString("G"), new WindowsStoreFile(file));
}
}
}
<|file_sep|>current/WindowsStore/Service/CameraService.cs
using MyDocs.Common.Model;
using MyDocs.WindowsStore.Storage;
using System;
using System.Threading.Tasks;
using Windows.Media.Capture;
namespace MyDocs.WindowsStore.Service
{
public class CameraService : ICameraService
{
public async Task<Photo> CapturePhotoAsync()
{
CameraCaptureUI camera = new CameraCaptureUI();
var file = await camera.CaptureFileAsync(CameraCaptureUIMode.Photo);
if (file == null) {
return null;
}
return new Photo(DateTime.Now.ToString("G"), new WindowsStoreFile(file));
}
}
}
<|file_sep|>updated/WindowsStore/Service/CameraService.cs | using MyDocs.Common.Model;
using MyDocs.WindowsStore.Storage;
using System;
using System.Threading.Tasks;
using Windows.Media.Capture;
namespace MyDocs.WindowsStore.Service
{
public class CameraService : ICameraService
{
public async Task<Photo> CapturePhotoAsync()
{
var camera = new CameraCaptureUI();
var file = await camera.CaptureFileAsync(CameraCaptureUIMode.Photo);
if (file == null) {
return null;
}
return new Photo(DateTime.Now.ToString("G"), new WindowsStoreFile(file));
}
}
} | <|file_sep|>original/WindowsStore/Service/CameraService.cs
using MyDocs.Common.Model;
using MyDocs.WindowsStore.Storage;
using System;
using System.Threading.Tasks;
using Windows.Media.Capture;
namespace MyDocs.WindowsStore.Service
{
public class CameraService : ICameraService
{
public async Task<Photo> CapturePhotoAsync()
{
CameraCaptureUI camera = new CameraCaptureUI();
var file = await camera.CaptureFileAsync(CameraCaptureUIMode.Photo);
if (file == null) {
return null;
}
return new Photo(DateTime.Now.ToString("G"), new WindowsStoreFile(file));
}
}
}
<|file_sep|>current/WindowsStore/Service/CameraService.cs
using MyDocs.Common.Model;
using MyDocs.WindowsStore.Storage;
using System;
using System.Threading.Tasks;
using Windows.Media.Capture;
namespace MyDocs.WindowsStore.Service
{
public class CameraService : ICameraService
{
public async Task<Photo> CapturePhotoAsync()
{
CameraCaptureUI camera = new CameraCaptureUI();
var file = await camera.CaptureFileAsync(CameraCaptureUIMode.Photo);
if (file == null) {
return null;
}
return new Photo(DateTime.Now.ToString("G"), new WindowsStoreFile(file));
}
}
}
<|file_sep|>updated/WindowsStore/Service/CameraService.cs
using MyDocs.Common.Model;
using MyDocs.WindowsStore.Storage;
using System;
using System.Threading.Tasks;
using Windows.Media.Capture;
namespace MyDocs.WindowsStore.Service
{
public class CameraService : ICameraService
{
public async Task<Photo> CapturePhotoAsync()
{
var camera = new CameraCaptureUI();
var file = await camera.CaptureFileAsync(CameraCaptureUIMode.Photo);
if (file == null) {
return null;
}
return new Photo(DateTime.Now.ToString("G"), new WindowsStoreFile(file));
}
}
} | 36bb2086420e20a1eae477b958d9e58915ec8047 | WindowsStore/Service/CameraService.cs | WindowsStore/Service/CameraService.cs | C# |
<|file_sep|>original/README.md
# zend-modulemanager
[](https://secure.travis-ci.org/zendframework/zend-modulemanager)
[](https://coveralls.io/r/zendframework/zend-modulemanager)
Zend Framework 2.0 introduces a new and powerful approach to modules. This new
module system is designed with flexibility, simplicity, and re-usability in mind.
A module may contain just about anything: PHP code, including MVC functionality;
library code; view scripts; and/or public assets such as images, CSS, and
JavaScript. The possibilities are endless.
`Zend\ModuleManager` is the component that enables the design of a module
architecture for PHP applcations.
- File issues at https://github.com/zendframework/zend-modulemanager/issues
- Documentation is at http://framework.zend.com/manual/current/en/index.html#zend-modulemanager
<|file_sep|>current/README.md
# zend-modulemanager
[](https://secure.travis-ci.org/zendframework/zend-modulemanager)
[](https://coveralls.io/r/zendframework/zend-modulemanager)
Zend Framework 2.0 introduces a new and powerful approach to modules. This new
module system is designed with flexibility, simplicity, and re-usability in mind.
A module may contain just about anything: PHP code, including MVC functionality;
library code; view scripts; and/or public assets such as images, CSS, and
JavaScript. The possibilities are endless.
`Zend\ModuleManager` is the component that enables the design of a module
architecture for PHP applcations.
- File issues at https://github.com/zendframework/zend-modulemanager/issues
- Documentation is at http://framework.zend.com/manual/current/en/index.html#zend-modulemanager
<|file_sep|>updated/README.md | # zend-modulemanager
[](https://secure.travis-ci.org/zendframework/zend-modulemanager)
[](https://coveralls.io/r/zendframework/zend-modulemanager?branch=master)
Zend Framework 2.0 introduces a new and powerful approach to modules. This new
module system is designed with flexibility, simplicity, and re-usability in mind.
A module may contain just about anything: PHP code, including MVC functionality;
library code; view scripts; and/or public assets such as images, CSS, and
JavaScript. The possibilities are endless.
`Zend\ModuleManager` is the component that enables the design of a module
architecture for PHP applcations.
- File issues at https://github.com/zendframework/zend-modulemanager/issues
- Documentation is at http://framework.zend.com/manual/current/en/index.html#zend-modulemanager | <|file_sep|>original/README.md
# zend-modulemanager
[](https://secure.travis-ci.org/zendframework/zend-modulemanager)
[](https://coveralls.io/r/zendframework/zend-modulemanager)
Zend Framework 2.0 introduces a new and powerful approach to modules. This new
module system is designed with flexibility, simplicity, and re-usability in mind.
A module may contain just about anything: PHP code, including MVC functionality;
library code; view scripts; and/or public assets such as images, CSS, and
JavaScript. The possibilities are endless.
`Zend\ModuleManager` is the component that enables the design of a module
architecture for PHP applcations.
- File issues at https://github.com/zendframework/zend-modulemanager/issues
- Documentation is at http://framework.zend.com/manual/current/en/index.html#zend-modulemanager
<|file_sep|>current/README.md
# zend-modulemanager
[](https://secure.travis-ci.org/zendframework/zend-modulemanager)
[](https://coveralls.io/r/zendframework/zend-modulemanager)
Zend Framework 2.0 introduces a new and powerful approach to modules. This new
module system is designed with flexibility, simplicity, and re-usability in mind.
A module may contain just about anything: PHP code, including MVC functionality;
library code; view scripts; and/or public assets such as images, CSS, and
JavaScript. The possibilities are endless.
`Zend\ModuleManager` is the component that enables the design of a module
architecture for PHP applcations.
- File issues at https://github.com/zendframework/zend-modulemanager/issues
- Documentation is at http://framework.zend.com/manual/current/en/index.html#zend-modulemanager
<|file_sep|>updated/README.md
# zend-modulemanager
[](https://secure.travis-ci.org/zendframework/zend-modulemanager)
[](https://coveralls.io/r/zendframework/zend-modulemanager?branch=master)
Zend Framework 2.0 introduces a new and powerful approach to modules. This new
module system is designed with flexibility, simplicity, and re-usability in mind.
A module may contain just about anything: PHP code, including MVC functionality;
library code; view scripts; and/or public assets such as images, CSS, and
JavaScript. The possibilities are endless.
`Zend\ModuleManager` is the component that enables the design of a module
architecture for PHP applcations.
- File issues at https://github.com/zendframework/zend-modulemanager/issues
- Documentation is at http://framework.zend.com/manual/current/en/index.html#zend-modulemanager | be4fef94edd4d89eb0f468e0fbcaf2dd46176ecf | README.md | README.md | Markdown |
<|file_sep|>original/db/migrate/20190212152221_change_dataset_id_to_bigint.rb
<|file_sep|>current/db/migrate/20190212152221_change_dataset_id_to_bigint.rb
<|file_sep|>updated/db/migrate/20190212152221_change_dataset_id_to_bigint.rb | class ChangeDatasetIdToBigint < ActiveRecord::Migration[6.0]
def up
change_column :datasets, :id, :bigint
change_column :agreement_event_datasets, :dataset_id, :bigint
change_column :dataset_file_audits, :dataset_id, :bigint
change_column :dataset_files, :dataset_id, :bigint
change_column :dataset_page_audits, :dataset_id, :bigint
change_column :dataset_reviews, :dataset_id, :bigint
change_column :dataset_users, :dataset_id, :bigint
change_column :dataset_versions, :dataset_id, :bigint
change_column :domains, :dataset_id, :bigint
change_column :forms, :dataset_id, :bigint
change_column :legal_document_datasets, :dataset_id, :bigint
change_column :notifications, :dataset_id, :bigint
change_column :requests, :dataset_id, :bigint
change_column :variable_forms, :dataset_id, :bigint
change_column :variables, :dataset_id, :bigint
end
def down | <|file_sep|>original/db/migrate/20190212152221_change_dataset_id_to_bigint.rb
<|file_sep|>current/db/migrate/20190212152221_change_dataset_id_to_bigint.rb
<|file_sep|>updated/db/migrate/20190212152221_change_dataset_id_to_bigint.rb
class ChangeDatasetIdToBigint < ActiveRecord::Migration[6.0]
def up
change_column :datasets, :id, :bigint
change_column :agreement_event_datasets, :dataset_id, :bigint
change_column :dataset_file_audits, :dataset_id, :bigint
change_column :dataset_files, :dataset_id, :bigint
change_column :dataset_page_audits, :dataset_id, :bigint
change_column :dataset_reviews, :dataset_id, :bigint
change_column :dataset_users, :dataset_id, :bigint
change_column :dataset_versions, :dataset_id, :bigint
change_column :domains, :dataset_id, :bigint
change_column :forms, :dataset_id, :bigint
change_column :legal_document_datasets, :dataset_id, :bigint
change_column :notifications, :dataset_id, :bigint
change_column :requests, :dataset_id, :bigint
change_column :variable_forms, :dataset_id, :bigint
change_column :variables, :dataset_id, :bigint
end
def down | 39b5d899e5646b936e3d479c43fbee321f6c9b1a | db/migrate/20190212152221_change_dataset_id_to_bigint.rb | db/migrate/20190212152221_change_dataset_id_to_bigint.rb | Ruby |
<|file_sep|>tests/CrawlerProcess/asyncio_deferred_signal.py.diff
original:
import scrapy
updated:
from scrapy import Spider
<|file_sep|>tests/CrawlerProcess/asyncio_deferred_signal.py.diff
original:
updated:
from scrapy.utils.defer import deferred_from_coro
<|file_sep|>tests/CrawlerProcess/asyncio_deferred_signal.py.diff
original:
return Deferred.fromFuture(loop.create_task(self._open_spider(spider)))
updated:
return deferred_from_coro(self._open_spider(spider))
<|file_sep|>original/tests/CrawlerProcess/asyncio_deferred_signal.py
await asyncio.sleep(0.1)
def open_spider(self, spider):
loop = asyncio.get_event_loop()
return Deferred.fromFuture(loop.create_task(self._open_spider(spider)))
def process_item(self, item, spider):
return {"url": item["url"].upper()}
class UrlSpider(scrapy.Spider):
name = "url_spider"
start_urls = ["data:,"]
custom_settings = {
"ITEM_PIPELINES": {UppercasePipeline: 100},
}
def parse(self, response):
yield {"url": response.url}
<|file_sep|>current/tests/CrawlerProcess/asyncio_deferred_signal.py
await asyncio.sleep(0.1)
def open_spider(self, spider):
loop = asyncio.get_event_loop()
return deferred_from_coro(self._open_spider(spider))
def process_item(self, item, spider):
return {"url": item["url"].upper()}
class UrlSpider(scrapy.Spider):
name = "url_spider"
start_urls = ["data:,"]
custom_settings = {
"ITEM_PIPELINES": {UppercasePipeline: 100},
}
def parse(self, response):
yield {"url": response.url}
<|file_sep|>updated/tests/CrawlerProcess/asyncio_deferred_signal.py | await asyncio.sleep(0.1)
def open_spider(self, spider):
loop = asyncio.get_event_loop()
return deferred_from_coro(self._open_spider(spider))
def process_item(self, item, spider):
return {"url": item["url"].upper()}
class UrlSpider(Spider):
name = "url_spider"
start_urls = ["data:,"]
custom_settings = {
"ITEM_PIPELINES": {UppercasePipeline: 100},
}
def parse(self, response):
yield {"url": response.url}
| <|file_sep|>tests/CrawlerProcess/asyncio_deferred_signal.py.diff
original:
import scrapy
updated:
from scrapy import Spider
<|file_sep|>tests/CrawlerProcess/asyncio_deferred_signal.py.diff
original:
updated:
from scrapy.utils.defer import deferred_from_coro
<|file_sep|>tests/CrawlerProcess/asyncio_deferred_signal.py.diff
original:
return Deferred.fromFuture(loop.create_task(self._open_spider(spider)))
updated:
return deferred_from_coro(self._open_spider(spider))
<|file_sep|>original/tests/CrawlerProcess/asyncio_deferred_signal.py
await asyncio.sleep(0.1)
def open_spider(self, spider):
loop = asyncio.get_event_loop()
return Deferred.fromFuture(loop.create_task(self._open_spider(spider)))
def process_item(self, item, spider):
return {"url": item["url"].upper()}
class UrlSpider(scrapy.Spider):
name = "url_spider"
start_urls = ["data:,"]
custom_settings = {
"ITEM_PIPELINES": {UppercasePipeline: 100},
}
def parse(self, response):
yield {"url": response.url}
<|file_sep|>current/tests/CrawlerProcess/asyncio_deferred_signal.py
await asyncio.sleep(0.1)
def open_spider(self, spider):
loop = asyncio.get_event_loop()
return deferred_from_coro(self._open_spider(spider))
def process_item(self, item, spider):
return {"url": item["url"].upper()}
class UrlSpider(scrapy.Spider):
name = "url_spider"
start_urls = ["data:,"]
custom_settings = {
"ITEM_PIPELINES": {UppercasePipeline: 100},
}
def parse(self, response):
yield {"url": response.url}
<|file_sep|>updated/tests/CrawlerProcess/asyncio_deferred_signal.py
await asyncio.sleep(0.1)
def open_spider(self, spider):
loop = asyncio.get_event_loop()
return deferred_from_coro(self._open_spider(spider))
def process_item(self, item, spider):
return {"url": item["url"].upper()}
class UrlSpider(Spider):
name = "url_spider"
start_urls = ["data:,"]
custom_settings = {
"ITEM_PIPELINES": {UppercasePipeline: 100},
}
def parse(self, response):
yield {"url": response.url}
| 7e98a76ac455a8c69950104766719cde313bbb74 | tests/CrawlerProcess/asyncio_deferred_signal.py | tests/CrawlerProcess/asyncio_deferred_signal.py | Python |
<|file_sep|>original/.travis.yml
language: php
php:
- '5.6'
- '7.0'
- hhvm
- nightly
script:
- composer install --no-dev
- npm install
- bower install
# - php -S localhost:7000
<|file_sep|>current/.travis.yml
language: php
php:
- '5.6'
- '7.0'
- hhvm
- nightly
script:
- composer install --no-dev
- npm install
- bower install
# - php -S localhost:7000
<|file_sep|>updated/.travis.yml | language: php
php:
- '5.6'
- '7.0'
# - hhvm
# - nightly
script:
- composer install --no-dev
- npm install
- bower install
# - php -S localhost:7000 | <|file_sep|>original/.travis.yml
language: php
php:
- '5.6'
- '7.0'
- hhvm
- nightly
script:
- composer install --no-dev
- npm install
- bower install
# - php -S localhost:7000
<|file_sep|>current/.travis.yml
language: php
php:
- '5.6'
- '7.0'
- hhvm
- nightly
script:
- composer install --no-dev
- npm install
- bower install
# - php -S localhost:7000
<|file_sep|>updated/.travis.yml
language: php
php:
- '5.6'
- '7.0'
# - hhvm
# - nightly
script:
- composer install --no-dev
- npm install
- bower install
# - php -S localhost:7000 | b8310eea4052d4bc04e08b30a73a4ab58bf75565 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/lib/tasks/deployment/20191004110122_update_the_brain.rake
<|file_sep|>current/lib/tasks/deployment/20191004110122_update_the_brain.rake
<|file_sep|>updated/lib/tasks/deployment/20191004110122_update_the_brain.rake | namespace :after_party do
desc 'Deployment task: update_the_brain'
task update_the_brain: :environment do
puts "Running deploy task 'update_the_brain'"
# Put your task implementation HERE.
InterventionTypeGroup.find_by(icon: 'head-side-brain').update(icon: 'users')
# Update task as completed. If you remove the line below, the task will
# run with every deploy (or every time you call after_party:run).
AfterParty::TaskRecord.create version: AfterParty::TaskRecorder.new(__FILE__).timestamp
end
end | <|file_sep|>original/lib/tasks/deployment/20191004110122_update_the_brain.rake
<|file_sep|>current/lib/tasks/deployment/20191004110122_update_the_brain.rake
<|file_sep|>updated/lib/tasks/deployment/20191004110122_update_the_brain.rake
namespace :after_party do
desc 'Deployment task: update_the_brain'
task update_the_brain: :environment do
puts "Running deploy task 'update_the_brain'"
# Put your task implementation HERE.
InterventionTypeGroup.find_by(icon: 'head-side-brain').update(icon: 'users')
# Update task as completed. If you remove the line below, the task will
# run with every deploy (or every time you call after_party:run).
AfterParty::TaskRecord.create version: AfterParty::TaskRecorder.new(__FILE__).timestamp
end
end | f3bca6910732c92fa080e6dae7b1fa471d3a213d | lib/tasks/deployment/20191004110122_update_the_brain.rake | lib/tasks/deployment/20191004110122_update_the_brain.rake | Ruby |
<|file_sep|>test/web_test.rb.diff
original:
def test_ifttt_unauthorized
updated:
def test_when_ifttt_unauthorized_returns_401
<|file_sep|>test/web_test.rb.diff
original:
updated:
Http::Actions::PostIfttt.any_instance.stubs(:call).raises(UnathorizedRequest)
<|file_sep|>test/web_test.rb.diff
original:
skip "this is a glorified testcase with too many deps"
subject.stub :slack_client, Minitest::Mock.new do
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: { token: "abc" }.to_json)
updated:
Http::Actions::PostIfttt.any_instance.stubs(:call).returns("OK")
<|file_sep|>original/test/web_test.rb
assert_equal "OK", response.body
end
def test_ifttt_unauthorized
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: "{}")
assert_equal 401, response.status
end
def test_ifttt_returns_ok
skip "this is a glorified testcase with too many deps"
subject.stub :slack_client, Minitest::Mock.new do
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: { token: "abc" }.to_json)
assert_equal 200, response.status
assert_equal "OK", response.body
end
<|file_sep|>current/test/web_test.rb
assert_equal "OK", response.body
end
def test_when_ifttt_unauthorized_returns_401
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: "{}")
Http::Actions::PostIfttt.any_instance.stubs(:call).raises(UnathorizedRequest)
assert_equal 401, response.status
end
def test_ifttt_returns_ok
Http::Actions::PostIfttt.any_instance.stubs(:call).returns("OK")
assert_equal 200, response.status
assert_equal "OK", response.body
end
end
end
<|file_sep|>updated/test/web_test.rb | def test_when_ifttt_unauthorized_returns_401
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: "{}")
Http::Actions::PostIfttt.any_instance.stubs(:call).raises(UnathorizedRequest)
assert_equal 401, response.status
end
def test_ifttt_returns_ok
Http::Actions::PostIfttt.any_instance.stubs(:call).returns("OK")
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: "{}")
assert_equal 200, response.status
assert_equal "OK", response.body
end
end | <|file_sep|>test/web_test.rb.diff
original:
def test_ifttt_unauthorized
updated:
def test_when_ifttt_unauthorized_returns_401
<|file_sep|>test/web_test.rb.diff
original:
updated:
Http::Actions::PostIfttt.any_instance.stubs(:call).raises(UnathorizedRequest)
<|file_sep|>test/web_test.rb.diff
original:
skip "this is a glorified testcase with too many deps"
subject.stub :slack_client, Minitest::Mock.new do
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: { token: "abc" }.to_json)
updated:
Http::Actions::PostIfttt.any_instance.stubs(:call).returns("OK")
<|file_sep|>original/test/web_test.rb
assert_equal "OK", response.body
end
def test_ifttt_unauthorized
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: "{}")
assert_equal 401, response.status
end
def test_ifttt_returns_ok
skip "this is a glorified testcase with too many deps"
subject.stub :slack_client, Minitest::Mock.new do
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: { token: "abc" }.to_json)
assert_equal 200, response.status
assert_equal "OK", response.body
end
<|file_sep|>current/test/web_test.rb
assert_equal "OK", response.body
end
def test_when_ifttt_unauthorized_returns_401
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: "{}")
Http::Actions::PostIfttt.any_instance.stubs(:call).raises(UnathorizedRequest)
assert_equal 401, response.status
end
def test_ifttt_returns_ok
Http::Actions::PostIfttt.any_instance.stubs(:call).returns("OK")
assert_equal 200, response.status
assert_equal "OK", response.body
end
end
end
<|file_sep|>updated/test/web_test.rb
def test_when_ifttt_unauthorized_returns_401
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: "{}")
Http::Actions::PostIfttt.any_instance.stubs(:call).raises(UnathorizedRequest)
assert_equal 401, response.status
end
def test_ifttt_returns_ok
Http::Actions::PostIfttt.any_instance.stubs(:call).returns("OK")
response = post(url: "/ifttt",
headers: { "Content-Type" => "application/json" },
payload: "{}")
assert_equal 200, response.status
assert_equal "OK", response.body
end
end | b53f9fc45dd3bd81a1bd313ed59db92a23218cc0 | test/web_test.rb | test/web_test.rb | Ruby |
<|file_sep|>original/requirements.txt
alembic==0.6.4
argparse==1.2.1
coverage==3.7.1
cx_Freeze==4.3.2
esky==0.9.8
faulthandler==2.3
mako==0.9.1
nose==1.3.1
poster==0.8.1
psutil==2.0.0
pycrypto==2.6.1
SQLAlchemy==0.9.4
<|file_sep|>current/requirements.txt
alembic==0.6.4
argparse==1.2.1
coverage==3.7.1
cx_Freeze==4.3.2
esky==0.9.8
faulthandler==2.3
mako==0.9.1
nose==1.3.1
poster==0.8.1
psutil==2.0.0
pycrypto==2.6.1
SQLAlchemy==0.9.4
<|file_sep|>updated/requirements.txt | alembic==0.6.4
argparse==1.2.1
coverage==3.7.1
cx_Freeze==4.3.2
esky==0.9.8
faulthandler==2.3
mako==0.9.1
nose==1.3.1
poster==0.8.1
psutil==2.1.1
pycrypto==2.6.1
SQLAlchemy==0.9.4 | <|file_sep|>original/requirements.txt
alembic==0.6.4
argparse==1.2.1
coverage==3.7.1
cx_Freeze==4.3.2
esky==0.9.8
faulthandler==2.3
mako==0.9.1
nose==1.3.1
poster==0.8.1
psutil==2.0.0
pycrypto==2.6.1
SQLAlchemy==0.9.4
<|file_sep|>current/requirements.txt
alembic==0.6.4
argparse==1.2.1
coverage==3.7.1
cx_Freeze==4.3.2
esky==0.9.8
faulthandler==2.3
mako==0.9.1
nose==1.3.1
poster==0.8.1
psutil==2.0.0
pycrypto==2.6.1
SQLAlchemy==0.9.4
<|file_sep|>updated/requirements.txt
alembic==0.6.4
argparse==1.2.1
coverage==3.7.1
cx_Freeze==4.3.2
esky==0.9.8
faulthandler==2.3
mako==0.9.1
nose==1.3.1
poster==0.8.1
psutil==2.1.1
pycrypto==2.6.1
SQLAlchemy==0.9.4 | 4e43e1228b41d42e40e654e2f22266ab412b95c7 | requirements.txt | requirements.txt | Text |
<|file_sep|>original/scripts/airdate.sh
if [ -z "$2" ]; then
cat <<USAGE
$caller: usage: ~script $SCRIPTNAME "<SHOW NAME>"
e.g. '~script $SCRIPTNAME "rick and morty"'
Uses 'http://epguides.com' to locate the next air date for
a given show.
USAGE
exit 0
fi
url="http://epguides.com/$show/"
episodesList=$(curl "$url" 2> /dev/null | grep -E "$EPISODE_REGEX +$DATE_REGEX.*")
lastEpisode=$(echo "$episodesList" | tail -n 1)
epNum="S"$(echo "$lastEpisode" | grep -Eo "$EPISODE_REGEX" | sed 's/-/-E/')
date=$(echo "$lastEpisode" | grep -Eo "$DATE_REGEX")
echo "$caller: The next episode of $2 ($epNum) airs on $date"
<|file_sep|>current/scripts/airdate.sh
if [ -z "$2" ]; then
cat <<USAGE
$caller: usage: ~script $SCRIPTNAME "<SHOW NAME>"
e.g. '~script $SCRIPTNAME "rick and morty"'
Uses 'http://epguides.com' to locate the next air date for
a given show.
USAGE
exit 0
fi
url="http://epguides.com/$show/"
episodesList=$(curl "$url" 2> /dev/null | grep -E "$EPISODE_REGEX +$DATE_REGEX.*")
lastEpisode=$(echo "$episodesList" | tail -n 1)
epNum="S"$(echo "$lastEpisode" | grep -Eo "$EPISODE_REGEX" | sed 's/-/-E/')
date=$(echo "$lastEpisode" | grep -Eo "$DATE_REGEX")
echo "$caller: The next episode of $2 ($epNum) airs on $date"
<|file_sep|>updated/scripts/airdate.sh | Uses 'http://epguides.com' to locate the next air date for
a given show.
USAGE
exit 0
fi
url="http://epguides.com/$show/"
episodesList=$(curl "$url" 2> /dev/null | grep -E "$EPISODE_REGEX +$DATE_REGEX.*")
if [ -z "$episodesList" ]; then
echo "$caller: Could not find episode guide for '$2'"
exit 0
fi
lastEpisode=$(echo "$episodesList" | tail -n 1)
epNum="S"$(echo "$lastEpisode" | grep -Eo "$EPISODE_REGEX" | sed 's/-/-E/')
date=$(echo "$lastEpisode" | grep -Eo "$DATE_REGEX")
echo "$caller: The next episode of $2 ($epNum) airs on $date" | <|file_sep|>original/scripts/airdate.sh
if [ -z "$2" ]; then
cat <<USAGE
$caller: usage: ~script $SCRIPTNAME "<SHOW NAME>"
e.g. '~script $SCRIPTNAME "rick and morty"'
Uses 'http://epguides.com' to locate the next air date for
a given show.
USAGE
exit 0
fi
url="http://epguides.com/$show/"
episodesList=$(curl "$url" 2> /dev/null | grep -E "$EPISODE_REGEX +$DATE_REGEX.*")
lastEpisode=$(echo "$episodesList" | tail -n 1)
epNum="S"$(echo "$lastEpisode" | grep -Eo "$EPISODE_REGEX" | sed 's/-/-E/')
date=$(echo "$lastEpisode" | grep -Eo "$DATE_REGEX")
echo "$caller: The next episode of $2 ($epNum) airs on $date"
<|file_sep|>current/scripts/airdate.sh
if [ -z "$2" ]; then
cat <<USAGE
$caller: usage: ~script $SCRIPTNAME "<SHOW NAME>"
e.g. '~script $SCRIPTNAME "rick and morty"'
Uses 'http://epguides.com' to locate the next air date for
a given show.
USAGE
exit 0
fi
url="http://epguides.com/$show/"
episodesList=$(curl "$url" 2> /dev/null | grep -E "$EPISODE_REGEX +$DATE_REGEX.*")
lastEpisode=$(echo "$episodesList" | tail -n 1)
epNum="S"$(echo "$lastEpisode" | grep -Eo "$EPISODE_REGEX" | sed 's/-/-E/')
date=$(echo "$lastEpisode" | grep -Eo "$DATE_REGEX")
echo "$caller: The next episode of $2 ($epNum) airs on $date"
<|file_sep|>updated/scripts/airdate.sh
Uses 'http://epguides.com' to locate the next air date for
a given show.
USAGE
exit 0
fi
url="http://epguides.com/$show/"
episodesList=$(curl "$url" 2> /dev/null | grep -E "$EPISODE_REGEX +$DATE_REGEX.*")
if [ -z "$episodesList" ]; then
echo "$caller: Could not find episode guide for '$2'"
exit 0
fi
lastEpisode=$(echo "$episodesList" | tail -n 1)
epNum="S"$(echo "$lastEpisode" | grep -Eo "$EPISODE_REGEX" | sed 's/-/-E/')
date=$(echo "$lastEpisode" | grep -Eo "$DATE_REGEX")
echo "$caller: The next episode of $2 ($epNum) airs on $date" | acac9b10662399646883ecc7a3758e0de18e0632 | scripts/airdate.sh | scripts/airdate.sh | Shell |
<|file_sep|>Code/CLOS/class-initialization-defmethods.lisp.diff
original:
(defmethod initialize-instance :after
((class regular-class) &rest initargs &key &allow-other-keys)
(apply #'initialize-instance-after-regular-class-default
class initargs))
updated:
<|file_sep|>original/Code/CLOS/class-initialization-defmethods.lisp
class
slot-names
initargs))
;;; According to the AMOP, calling initialize-instance on a built-in
;;; class (i.e., on an instance of a subclass of the class
;;; BUILT-IN-CLASS) signals an error, because built-in classes can not
;;; be created by the user. But during the bootstrapping phase we
;;; need to create built-in classes. We solve this problem by
;;; removing this method once the bootstrapping phase is finished.
;;;
;;; We do not add readers and writers here, because we do it in
;;; ENSURE-BUILT-IN-CLASS after we have finalized inheritance. The
;;; reason for that is that we then know the slot location.
(defmethod initialize-instance :after
((class built-in-class) &rest initargs &key &allow-other-keys)
(apply #'initialize-instance-after-built-in-class-default
class initargs))
;;; I don't know why this definition makes SBCL go into an infinite
;;; recursion.
<|file_sep|>current/Code/CLOS/class-initialization-defmethods.lisp
;;; class (i.e., on an instance of a subclass of the class
;;; BUILT-IN-CLASS) signals an error, because built-in classes can not
;;; be created by the user. But during the bootstrapping phase we
;;; need to create built-in classes. We solve this problem by
;;; removing this method once the bootstrapping phase is finished.
;;;
;;; We do not add readers and writers here, because we do it in
;;; ENSURE-BUILT-IN-CLASS after we have finalized inheritance. The
;;; reason for that is that we then know the slot location.
(defmethod initialize-instance :after
((class built-in-class) &rest initargs &key &allow-other-keys)
(apply #'initialize-instance-after-built-in-class-default
class initargs))
;;; I don't know why this definition makes SBCL go into an infinite
;;; recursion.
;; (defmethod reinitialize-instance :after
;; ((class standard-class)
;; &rest args
;; &key
;; &allow-other-keys)
<|file_sep|>updated/Code/CLOS/class-initialization-defmethods.lisp | (cl:in-package #:sicl-clos)
(defmethod shared-initialize :around
((class real-class)
slot-names
&rest initargs
&key
&allow-other-keys)
(apply #'shared-initialize-around-real-class-default
#'call-next-method
class
slot-names
initargs)) | <|file_sep|>Code/CLOS/class-initialization-defmethods.lisp.diff
original:
(defmethod initialize-instance :after
((class regular-class) &rest initargs &key &allow-other-keys)
(apply #'initialize-instance-after-regular-class-default
class initargs))
updated:
<|file_sep|>original/Code/CLOS/class-initialization-defmethods.lisp
class
slot-names
initargs))
;;; According to the AMOP, calling initialize-instance on a built-in
;;; class (i.e., on an instance of a subclass of the class
;;; BUILT-IN-CLASS) signals an error, because built-in classes can not
;;; be created by the user. But during the bootstrapping phase we
;;; need to create built-in classes. We solve this problem by
;;; removing this method once the bootstrapping phase is finished.
;;;
;;; We do not add readers and writers here, because we do it in
;;; ENSURE-BUILT-IN-CLASS after we have finalized inheritance. The
;;; reason for that is that we then know the slot location.
(defmethod initialize-instance :after
((class built-in-class) &rest initargs &key &allow-other-keys)
(apply #'initialize-instance-after-built-in-class-default
class initargs))
;;; I don't know why this definition makes SBCL go into an infinite
;;; recursion.
<|file_sep|>current/Code/CLOS/class-initialization-defmethods.lisp
;;; class (i.e., on an instance of a subclass of the class
;;; BUILT-IN-CLASS) signals an error, because built-in classes can not
;;; be created by the user. But during the bootstrapping phase we
;;; need to create built-in classes. We solve this problem by
;;; removing this method once the bootstrapping phase is finished.
;;;
;;; We do not add readers and writers here, because we do it in
;;; ENSURE-BUILT-IN-CLASS after we have finalized inheritance. The
;;; reason for that is that we then know the slot location.
(defmethod initialize-instance :after
((class built-in-class) &rest initargs &key &allow-other-keys)
(apply #'initialize-instance-after-built-in-class-default
class initargs))
;;; I don't know why this definition makes SBCL go into an infinite
;;; recursion.
;; (defmethod reinitialize-instance :after
;; ((class standard-class)
;; &rest args
;; &key
;; &allow-other-keys)
<|file_sep|>updated/Code/CLOS/class-initialization-defmethods.lisp
(cl:in-package #:sicl-clos)
(defmethod shared-initialize :around
((class real-class)
slot-names
&rest initargs
&key
&allow-other-keys)
(apply #'shared-initialize-around-real-class-default
#'call-next-method
class
slot-names
initargs)) | 21835b0a11fb8bac53f5113408ea455ea66657f6 | Code/CLOS/class-initialization-defmethods.lisp | Code/CLOS/class-initialization-defmethods.lisp | Common Lisp |
<|file_sep|>original/kresources/birthdays/CMakeLists.txt
include_directories( ${CMAKE_SOURCE_DIR}/kaddressbook/common ${CMAKE_SOURCE_DIR}/kaddressbook/ ${CMAKE_BINARY_DIR}/libkdepim ${CMAKE_SOURCE_DIR}/libkdepim ${CMAKE_BINARY_DIR}/kaddressbook/common )
add_definitions (-DQT3_SUPPORT -DQT3_SUPPORT_WARNINGS)
########### next target ###############
set(kcal_kabc_PART_SRCS resourcekabc.cpp resourcekabcconfig.cpp )
kde4_add_plugin(kcal_kabc ${kcal_kabc_PART_SRCS})
target_link_libraries(kcal_kabc ${KDE4_KDECORE_LIBS} ${KDEPIMLIBS_KCAL_LIBS} ${KDEPIMLIBS_KPIMUTILS_LIBS} ${QT_QT3SUPPORT_LIBRARY} kdepim)
install(TARGETS kcal_kabc DESTINATION ${PLUGIN_INSTALL_DIR})
########### install files ###############
install( FILES kabc.desktop DESTINATION ${SERVICES_INSTALL_DIR}/kresources/kcal)
<|file_sep|>current/kresources/birthdays/CMakeLists.txt
include_directories( ${CMAKE_SOURCE_DIR}/kaddressbook/common ${CMAKE_SOURCE_DIR}/kaddressbook/ ${CMAKE_BINARY_DIR}/libkdepim ${CMAKE_SOURCE_DIR}/libkdepim ${CMAKE_BINARY_DIR}/kaddressbook/common )
add_definitions (-DQT3_SUPPORT -DQT3_SUPPORT_WARNINGS)
########### next target ###############
set(kcal_kabc_PART_SRCS resourcekabc.cpp resourcekabcconfig.cpp )
kde4_add_plugin(kcal_kabc ${kcal_kabc_PART_SRCS})
target_link_libraries(kcal_kabc ${KDE4_KDECORE_LIBS} ${KDEPIMLIBS_KCAL_LIBS} ${KDEPIMLIBS_KPIMUTILS_LIBS} ${QT_QT3SUPPORT_LIBRARY} kdepim)
install(TARGETS kcal_kabc DESTINATION ${PLUGIN_INSTALL_DIR})
########### install files ###############
install( FILES kabc.desktop DESTINATION ${SERVICES_INSTALL_DIR}/kresources/kcal)
<|file_sep|>updated/kresources/birthdays/CMakeLists.txt |
include_directories( ${CMAKE_SOURCE_DIR}/kaddressbook/common ${CMAKE_SOURCE_DIR}/kaddressbook/ ${CMAKE_BINARY_DIR}/libkdepim ${CMAKE_SOURCE_DIR}/libkdepim ${CMAKE_BINARY_DIR}/kaddressbook/common )
add_definitions (-DQT3_SUPPORT -DQT3_SUPPORT_WARNINGS)
########### next target ###############
set(kcal_kabc_PART_SRCS resourcekabc.cpp resourcekabcconfig.cpp )
kde4_add_plugin(kcal_kabc ${kcal_kabc_PART_SRCS})
target_link_libraries(kcal_kabc ${KDE4_KDECORE_LIBS} ${KDEPIMLIBS_KCAL_LIBS} ${KDEPIMLIBS_KPIMUTILS_LIBS} ${QT_QT3SUPPORT_LIBRARY} kdepim kabcommon)
install(TARGETS kcal_kabc DESTINATION ${PLUGIN_INSTALL_DIR})
########### install files ###############
install( FILES kabc.desktop DESTINATION ${SERVICES_INSTALL_DIR}/kresources/kcal) | <|file_sep|>original/kresources/birthdays/CMakeLists.txt
include_directories( ${CMAKE_SOURCE_DIR}/kaddressbook/common ${CMAKE_SOURCE_DIR}/kaddressbook/ ${CMAKE_BINARY_DIR}/libkdepim ${CMAKE_SOURCE_DIR}/libkdepim ${CMAKE_BINARY_DIR}/kaddressbook/common )
add_definitions (-DQT3_SUPPORT -DQT3_SUPPORT_WARNINGS)
########### next target ###############
set(kcal_kabc_PART_SRCS resourcekabc.cpp resourcekabcconfig.cpp )
kde4_add_plugin(kcal_kabc ${kcal_kabc_PART_SRCS})
target_link_libraries(kcal_kabc ${KDE4_KDECORE_LIBS} ${KDEPIMLIBS_KCAL_LIBS} ${KDEPIMLIBS_KPIMUTILS_LIBS} ${QT_QT3SUPPORT_LIBRARY} kdepim)
install(TARGETS kcal_kabc DESTINATION ${PLUGIN_INSTALL_DIR})
########### install files ###############
install( FILES kabc.desktop DESTINATION ${SERVICES_INSTALL_DIR}/kresources/kcal)
<|file_sep|>current/kresources/birthdays/CMakeLists.txt
include_directories( ${CMAKE_SOURCE_DIR}/kaddressbook/common ${CMAKE_SOURCE_DIR}/kaddressbook/ ${CMAKE_BINARY_DIR}/libkdepim ${CMAKE_SOURCE_DIR}/libkdepim ${CMAKE_BINARY_DIR}/kaddressbook/common )
add_definitions (-DQT3_SUPPORT -DQT3_SUPPORT_WARNINGS)
########### next target ###############
set(kcal_kabc_PART_SRCS resourcekabc.cpp resourcekabcconfig.cpp )
kde4_add_plugin(kcal_kabc ${kcal_kabc_PART_SRCS})
target_link_libraries(kcal_kabc ${KDE4_KDECORE_LIBS} ${KDEPIMLIBS_KCAL_LIBS} ${KDEPIMLIBS_KPIMUTILS_LIBS} ${QT_QT3SUPPORT_LIBRARY} kdepim)
install(TARGETS kcal_kabc DESTINATION ${PLUGIN_INSTALL_DIR})
########### install files ###############
install( FILES kabc.desktop DESTINATION ${SERVICES_INSTALL_DIR}/kresources/kcal)
<|file_sep|>updated/kresources/birthdays/CMakeLists.txt
include_directories( ${CMAKE_SOURCE_DIR}/kaddressbook/common ${CMAKE_SOURCE_DIR}/kaddressbook/ ${CMAKE_BINARY_DIR}/libkdepim ${CMAKE_SOURCE_DIR}/libkdepim ${CMAKE_BINARY_DIR}/kaddressbook/common )
add_definitions (-DQT3_SUPPORT -DQT3_SUPPORT_WARNINGS)
########### next target ###############
set(kcal_kabc_PART_SRCS resourcekabc.cpp resourcekabcconfig.cpp )
kde4_add_plugin(kcal_kabc ${kcal_kabc_PART_SRCS})
target_link_libraries(kcal_kabc ${KDE4_KDECORE_LIBS} ${KDEPIMLIBS_KCAL_LIBS} ${KDEPIMLIBS_KPIMUTILS_LIBS} ${QT_QT3SUPPORT_LIBRARY} kdepim kabcommon)
install(TARGETS kcal_kabc DESTINATION ${PLUGIN_INSTALL_DIR})
########### install files ###############
install( FILES kabc.desktop DESTINATION ${SERVICES_INSTALL_DIR}/kresources/kcal) | 9e720af5ff5325f8ac3af86052d796f0bf09838b | kresources/birthdays/CMakeLists.txt | kresources/birthdays/CMakeLists.txt | Text |
<|file_sep|>original/recipes/pytrends/meta.yaml
- pandas >=0.25
- python >=3.3
- requests
test:
imports:
- pytrends
commands:
- pip check
requires:
- pip
about:
home: https://github.com/dreyco676/pytrends
summary: Pseudo API for Google Trends
license: Apache-2.0
license_file: LICENSE
extra:
recipe-maintainers:
- IlanaRadinsky
<|file_sep|>current/recipes/pytrends/meta.yaml
- pandas >=0.25
- python >=3.3
- requests
test:
imports:
- pytrends
commands:
- pip check
requires:
- pip
about:
home: https://github.com/dreyco676/pytrends
summary: Pseudo API for Google Trends
license: Apache-2.0
license_file: LICENSE
extra:
recipe-maintainers:
- IlanaRadinsky
<|file_sep|>updated/recipes/pytrends/meta.yaml | - python >=3.3
- requests
test:
imports:
- pytrends
commands:
- pip check
requires:
- pip
about:
home: https://github.com/dreyco676/pytrends
summary: Pseudo API for Google Trends
license: Apache-2.0
# license file manually packaged. See https://github.com/GeneralMills/pytrends/pull/443
license_file: LICENSE
extra:
recipe-maintainers:
- IlanaRadinsky | <|file_sep|>original/recipes/pytrends/meta.yaml
- pandas >=0.25
- python >=3.3
- requests
test:
imports:
- pytrends
commands:
- pip check
requires:
- pip
about:
home: https://github.com/dreyco676/pytrends
summary: Pseudo API for Google Trends
license: Apache-2.0
license_file: LICENSE
extra:
recipe-maintainers:
- IlanaRadinsky
<|file_sep|>current/recipes/pytrends/meta.yaml
- pandas >=0.25
- python >=3.3
- requests
test:
imports:
- pytrends
commands:
- pip check
requires:
- pip
about:
home: https://github.com/dreyco676/pytrends
summary: Pseudo API for Google Trends
license: Apache-2.0
license_file: LICENSE
extra:
recipe-maintainers:
- IlanaRadinsky
<|file_sep|>updated/recipes/pytrends/meta.yaml
- python >=3.3
- requests
test:
imports:
- pytrends
commands:
- pip check
requires:
- pip
about:
home: https://github.com/dreyco676/pytrends
summary: Pseudo API for Google Trends
license: Apache-2.0
# license file manually packaged. See https://github.com/GeneralMills/pytrends/pull/443
license_file: LICENSE
extra:
recipe-maintainers:
- IlanaRadinsky | ea3859d06a3771d1b0f710ef39c6e32c7612b88d | recipes/pytrends/meta.yaml | recipes/pytrends/meta.yaml | YAML |
<|file_sep|>README.md.diff
original:
updated:
Building the soundboard requires having [Node.js](https://nodejs.org/), [PHP](https://secure.php.net/) and [Composer](https://getcomposer.org/) installed on your system.
<|file_sep|>original/README.md
# Soundboard
> More like knäckebröd, har jag rätt?
## How to _deal with it?_
1. Add some sound samples to `public/samples/`.
2. Run `npm install` to install dependencies _and_ build the assets.
3. __Optional:__ Run `gulp php-server` to start a PHP server.
4. __Optional:__ Run `gulp watch` to start watchers that will rebuild assets on change.
## Features
- Every wav/mp3/ogg you add to `public/samples/` will be glorified instantaneously
- Responsive search
- Direct, obfuscated, linking to soundbytes
- Random button
- Hold shift to play soundbytes simultaneously
Samples not included!
Design by [Mesoptier](https://github.com/mesoptier)!
<|file_sep|>current/README.md
# Soundboard
> More like knäckebröd, har jag rätt?
## How to _deal with it?_
1. Add some sound samples to `public/samples/`.
2. Run `npm install` to install dependencies _and_ build the assets.
3. __Optional:__ Run `gulp php-server` to start a PHP server.
4. __Optional:__ Run `gulp watch` to start watchers that will rebuild assets on change.
Building the soundboard requires having [Node.js](https://nodejs.org/), [PHP](https://secure.php.net/) and [Composer](https://getcomposer.org/) installed on your system.
## Features
- Every wav/mp3/ogg you add to `public/samples/` will be glorified instantaneously
- Responsive search
- Direct, obfuscated, linking to soundbytes
- Random button
- Hold shift to play soundbytes simultaneously
Samples not included!
Design by [Mesoptier](https://github.com/mesoptier)!
<|file_sep|>updated/README.md | > More like knäckebröd, har jag rätt?
## How to _deal with it?_
1. Add some sound samples to `public/samples/`.
2. Run `npm install` to install dependencies _and_ build the assets.
3. __Optional:__ Run `gulp php-server` to start a PHP server.
4. __Optional:__ Run `gulp watch` to start watchers that will rebuild assets on change.
Building the soundboard requires having [Node.js](https://nodejs.org/), [PHP](https://secure.php.net/) and [Composer](https://getcomposer.org/) installed on your system.
## Features
- Every wav/mp3/ogg you add to `public/samples/` will be glorified instantaneously
- Responsive search
- Direct, obfuscated, linking to soundbytes
- Random button
- Hold shift to play soundbytes simultaneously
- Hold control to play soundbytes on a loop
Samples not included!
Design by [Mesoptier](https://github.com/mesoptier)! | <|file_sep|>README.md.diff
original:
updated:
Building the soundboard requires having [Node.js](https://nodejs.org/), [PHP](https://secure.php.net/) and [Composer](https://getcomposer.org/) installed on your system.
<|file_sep|>original/README.md
# Soundboard
> More like knäckebröd, har jag rätt?
## How to _deal with it?_
1. Add some sound samples to `public/samples/`.
2. Run `npm install` to install dependencies _and_ build the assets.
3. __Optional:__ Run `gulp php-server` to start a PHP server.
4. __Optional:__ Run `gulp watch` to start watchers that will rebuild assets on change.
## Features
- Every wav/mp3/ogg you add to `public/samples/` will be glorified instantaneously
- Responsive search
- Direct, obfuscated, linking to soundbytes
- Random button
- Hold shift to play soundbytes simultaneously
Samples not included!
Design by [Mesoptier](https://github.com/mesoptier)!
<|file_sep|>current/README.md
# Soundboard
> More like knäckebröd, har jag rätt?
## How to _deal with it?_
1. Add some sound samples to `public/samples/`.
2. Run `npm install` to install dependencies _and_ build the assets.
3. __Optional:__ Run `gulp php-server` to start a PHP server.
4. __Optional:__ Run `gulp watch` to start watchers that will rebuild assets on change.
Building the soundboard requires having [Node.js](https://nodejs.org/), [PHP](https://secure.php.net/) and [Composer](https://getcomposer.org/) installed on your system.
## Features
- Every wav/mp3/ogg you add to `public/samples/` will be glorified instantaneously
- Responsive search
- Direct, obfuscated, linking to soundbytes
- Random button
- Hold shift to play soundbytes simultaneously
Samples not included!
Design by [Mesoptier](https://github.com/mesoptier)!
<|file_sep|>updated/README.md
> More like knäckebröd, har jag rätt?
## How to _deal with it?_
1. Add some sound samples to `public/samples/`.
2. Run `npm install` to install dependencies _and_ build the assets.
3. __Optional:__ Run `gulp php-server` to start a PHP server.
4. __Optional:__ Run `gulp watch` to start watchers that will rebuild assets on change.
Building the soundboard requires having [Node.js](https://nodejs.org/), [PHP](https://secure.php.net/) and [Composer](https://getcomposer.org/) installed on your system.
## Features
- Every wav/mp3/ogg you add to `public/samples/` will be glorified instantaneously
- Responsive search
- Direct, obfuscated, linking to soundbytes
- Random button
- Hold shift to play soundbytes simultaneously
- Hold control to play soundbytes on a loop
Samples not included!
Design by [Mesoptier](https://github.com/mesoptier)! | 4f63e8039ae419000123b2642214c67c789ca8bd | README.md | README.md | Markdown |
<|file_sep|>original/src/webservice/WebService.php
<|file_sep|>current/src/webservice/WebService.php
<|file_sep|>updated/src/webservice/WebService.php | <?php
require_once 'webservice/db.php';
/**
* Abstract Base class for all web services.
*/
abstract class WebService {
/**
* Executes the web service.
* @param Array $data User-specified parameters. Usually, url parameters are included as $data[query1], $data[query2],... and $_GET and _POST are merged in
*/
abstract public function execute($data);
/**
* ouputs $dataArray as (on php>=5.4 pretty) JSON
* @param Array $dataArray
*/
public static function output($dataArray) {
echo json_encode($dataArray, defined('JSON_PRETTY_PRINT')?JSON_PRETTY_PRINT:0); | <|file_sep|>original/src/webservice/WebService.php
<|file_sep|>current/src/webservice/WebService.php
<|file_sep|>updated/src/webservice/WebService.php
<?php
require_once 'webservice/db.php';
/**
* Abstract Base class for all web services.
*/
abstract class WebService {
/**
* Executes the web service.
* @param Array $data User-specified parameters. Usually, url parameters are included as $data[query1], $data[query2],... and $_GET and _POST are merged in
*/
abstract public function execute($data);
/**
* ouputs $dataArray as (on php>=5.4 pretty) JSON
* @param Array $dataArray
*/
public static function output($dataArray) {
echo json_encode($dataArray, defined('JSON_PRETTY_PRINT')?JSON_PRETTY_PRINT:0); | 025078711677667c976d99ae08a523e2584f907b | src/webservice/WebService.php | src/webservice/WebService.php | PHP |
<|file_sep|>original/spec/01-algorithms/find.spec.js
import { find } from '../../src/c/find'
describe('chapter1.find', function() {
it('should return -1 if no such element', function() {
expect(find([2, 1, 4, 2, 3], 5)).toBe(-1);
});
it('should return first index at which element is found', function() {
expect(find([5, 1, 2, 4, 2, 3, 8, 9], 2)).toBe(2);
});
});
<|file_sep|>current/spec/01-algorithms/find.spec.js
import { find } from '../../src/c/find'
describe('chapter1.find', function() {
it('should return -1 if no such element', function() {
expect(find([2, 1, 4, 2, 3], 5)).toBe(-1);
});
it('should return first index at which element is found', function() {
expect(find([5, 1, 2, 4, 2, 3, 8, 9], 2)).toBe(2);
});
});
<|file_sep|>updated/spec/01-algorithms/find.spec.js | import { find } from '../../src/01-algorithms/find'
describe('chapter1.find', function() {
it('should return -1 if no such element', function() {
expect(find([2, 1, 4, 2, 3], 5)).toBe(-1);
});
it('should return first index at which element is found', function() {
expect(find([5, 1, 2, 4, 2, 3, 8, 9], 2)).toBe(2);
});
}); | <|file_sep|>original/spec/01-algorithms/find.spec.js
import { find } from '../../src/c/find'
describe('chapter1.find', function() {
it('should return -1 if no such element', function() {
expect(find([2, 1, 4, 2, 3], 5)).toBe(-1);
});
it('should return first index at which element is found', function() {
expect(find([5, 1, 2, 4, 2, 3, 8, 9], 2)).toBe(2);
});
});
<|file_sep|>current/spec/01-algorithms/find.spec.js
import { find } from '../../src/c/find'
describe('chapter1.find', function() {
it('should return -1 if no such element', function() {
expect(find([2, 1, 4, 2, 3], 5)).toBe(-1);
});
it('should return first index at which element is found', function() {
expect(find([5, 1, 2, 4, 2, 3, 8, 9], 2)).toBe(2);
});
});
<|file_sep|>updated/spec/01-algorithms/find.spec.js
import { find } from '../../src/01-algorithms/find'
describe('chapter1.find', function() {
it('should return -1 if no such element', function() {
expect(find([2, 1, 4, 2, 3], 5)).toBe(-1);
});
it('should return first index at which element is found', function() {
expect(find([5, 1, 2, 4, 2, 3, 8, 9], 2)).toBe(2);
});
}); | c3ae1bbc51bb480ab1b5393df52b16f57e047445 | spec/01-algorithms/find.spec.js | spec/01-algorithms/find.spec.js | JavaScript |
<|file_sep|>original/environment-3.6.yml
name: cmac_env
channels:
- conda-forge
- defaults
dependencies:
- python=3.6
- arm_pyart
- cartopy
- coincbc
- cython=0.27
- distributed
- scikit-fuzzy
- pip:
- git+https://github.com/CSU-Radarmet/CSU_RadarTools.git
- git+https://github.com/EVS-ATMOS/cmac2.0.git
<|file_sep|>current/environment-3.6.yml
name: cmac_env
channels:
- conda-forge
- defaults
dependencies:
- python=3.6
- arm_pyart
- cartopy
- coincbc
- cython=0.27
- distributed
- scikit-fuzzy
- pip:
- git+https://github.com/CSU-Radarmet/CSU_RadarTools.git
- git+https://github.com/EVS-ATMOS/cmac2.0.git
<|file_sep|>updated/environment-3.6.yml | name: cmac_env
channels:
- conda-forge
- defaults
dependencies:
- python=3.6
- arm_pyart
- cartopy
- coincbc
- cython=0.27
- distributed
- scikit-fuzzy
- pip
- pip:
- git+https://github.com/CSU-Radarmet/CSU_RadarTools.git
- git+https://github.com/EVS-ATMOS/cmac2.0.git | <|file_sep|>original/environment-3.6.yml
name: cmac_env
channels:
- conda-forge
- defaults
dependencies:
- python=3.6
- arm_pyart
- cartopy
- coincbc
- cython=0.27
- distributed
- scikit-fuzzy
- pip:
- git+https://github.com/CSU-Radarmet/CSU_RadarTools.git
- git+https://github.com/EVS-ATMOS/cmac2.0.git
<|file_sep|>current/environment-3.6.yml
name: cmac_env
channels:
- conda-forge
- defaults
dependencies:
- python=3.6
- arm_pyart
- cartopy
- coincbc
- cython=0.27
- distributed
- scikit-fuzzy
- pip:
- git+https://github.com/CSU-Radarmet/CSU_RadarTools.git
- git+https://github.com/EVS-ATMOS/cmac2.0.git
<|file_sep|>updated/environment-3.6.yml
name: cmac_env
channels:
- conda-forge
- defaults
dependencies:
- python=3.6
- arm_pyart
- cartopy
- coincbc
- cython=0.27
- distributed
- scikit-fuzzy
- pip
- pip:
- git+https://github.com/CSU-Radarmet/CSU_RadarTools.git
- git+https://github.com/EVS-ATMOS/cmac2.0.git | c1daac20b765792af6e3010d542212c98974bf1f | environment-3.6.yml | environment-3.6.yml | YAML |
<|file_sep|>original/source/keepassx20.sh
<|file_sep|>current/source/keepassx20.sh
<|file_sep|>updated/source/keepassx20.sh | #!/bin/bash
wget https://www.keepassx.org/dev/attachments/download/115/keepassx-2.0-beta2.tar.gz
tar xvzf keepassx-2.0-beta2.tar.gz
cd keepassx-2.0-beta2
mkdir build
cd build
cmake -DCMAKE_INSTALL_PREFIX=$HOME/dotfiles/local -DCMAKE_VERBOSE_MAKEFILE=ON ..
make
make install
cd ../..
rm -rf keepassx-2.0-beta2
rm -f tar xvzf keepassx-2.0-beta2.tar.gz | <|file_sep|>original/source/keepassx20.sh
<|file_sep|>current/source/keepassx20.sh
<|file_sep|>updated/source/keepassx20.sh
#!/bin/bash
wget https://www.keepassx.org/dev/attachments/download/115/keepassx-2.0-beta2.tar.gz
tar xvzf keepassx-2.0-beta2.tar.gz
cd keepassx-2.0-beta2
mkdir build
cd build
cmake -DCMAKE_INSTALL_PREFIX=$HOME/dotfiles/local -DCMAKE_VERBOSE_MAKEFILE=ON ..
make
make install
cd ../..
rm -rf keepassx-2.0-beta2
rm -f tar xvzf keepassx-2.0-beta2.tar.gz | 6aef450585bff853bfa73062fb08cd2e4fe0d55e | source/keepassx20.sh | source/keepassx20.sh | Shell |
<|file_sep|>original/guacamole-common/src/main/java/org/apache/guacamole/GuacamoleResourceClosedException.java
<|file_sep|>current/guacamole-common/src/main/java/org/apache/guacamole/GuacamoleResourceClosedException.java
<|file_sep|>updated/guacamole-common/src/main/java/org/apache/guacamole/GuacamoleResourceClosedException.java | /*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.guacamole;
| <|file_sep|>original/guacamole-common/src/main/java/org/apache/guacamole/GuacamoleResourceClosedException.java
<|file_sep|>current/guacamole-common/src/main/java/org/apache/guacamole/GuacamoleResourceClosedException.java
<|file_sep|>updated/guacamole-common/src/main/java/org/apache/guacamole/GuacamoleResourceClosedException.java
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.guacamole;
| 32610e9af225a9e7eeb63707048434c42ad5c702 | guacamole-common/src/main/java/org/apache/guacamole/GuacamoleResourceClosedException.java | guacamole-common/src/main/java/org/apache/guacamole/GuacamoleResourceClosedException.java | Java |
<|file_sep|>original/projects/view/client/viewProjects/viewProjects.html
<template name="viewProjects">
<div class="page-header">
<a class="btn btn-primary pull-right" href="/projects/add" role="button">Add Project</a>
<h1>View all projects</h1>
</div>
<div class="row">
{{> categoryDropdown }}
{{> locationDropdown }}
{{> searchResults }}
</div>
</template>
<|file_sep|>current/projects/view/client/viewProjects/viewProjects.html
<template name="viewProjects">
<div class="page-header">
<a class="btn btn-primary pull-right" href="/projects/add" role="button">Add Project</a>
<h1>View all projects</h1>
</div>
<div class="row">
{{> categoryDropdown }}
{{> locationDropdown }}
{{> searchResults }}
</div>
</template>
<|file_sep|>updated/projects/view/client/viewProjects/viewProjects.html | <template name="viewProjects">
<div class="page-header">
<a class="btn btn-primary pull-right" href="/projects/add" role="button">Add Project</a>
<h1>View all projects</h1>
</div>
<div class="row">
{{> filterProjects }}
</div>
<div class="row">
{{> searchResults }}
</div>
</template> | <|file_sep|>original/projects/view/client/viewProjects/viewProjects.html
<template name="viewProjects">
<div class="page-header">
<a class="btn btn-primary pull-right" href="/projects/add" role="button">Add Project</a>
<h1>View all projects</h1>
</div>
<div class="row">
{{> categoryDropdown }}
{{> locationDropdown }}
{{> searchResults }}
</div>
</template>
<|file_sep|>current/projects/view/client/viewProjects/viewProjects.html
<template name="viewProjects">
<div class="page-header">
<a class="btn btn-primary pull-right" href="/projects/add" role="button">Add Project</a>
<h1>View all projects</h1>
</div>
<div class="row">
{{> categoryDropdown }}
{{> locationDropdown }}
{{> searchResults }}
</div>
</template>
<|file_sep|>updated/projects/view/client/viewProjects/viewProjects.html
<template name="viewProjects">
<div class="page-header">
<a class="btn btn-primary pull-right" href="/projects/add" role="button">Add Project</a>
<h1>View all projects</h1>
</div>
<div class="row">
{{> filterProjects }}
</div>
<div class="row">
{{> searchResults }}
</div>
</template> | 69c25d7d04968c4815eb6c9acf27b2c5b8417a96 | projects/view/client/viewProjects/viewProjects.html | projects/view/client/viewProjects/viewProjects.html | HTML |
<|file_sep|>original/api/build.sbt
name := "detective-api"
organization := "me.frmr.kafka"
scalaVersion := "2.12.4"
version := Version.version
libraryDependencies += "com.typesafe.scala-logging" %% "scala-logging" % "3.5.0"
libraryDependencies += "org.scalatest" %% "scalatest" % "3.0.1" % "test"
<|file_sep|>current/api/build.sbt
name := "detective-api"
organization := "me.frmr.kafka"
scalaVersion := "2.12.4"
version := Version.version
libraryDependencies += "com.typesafe.scala-logging" %% "scala-logging" % "3.5.0"
libraryDependencies += "org.scalatest" %% "scalatest" % "3.0.1" % "test"
<|file_sep|>updated/api/build.sbt | name := "detective-api"
organization := "me.frmr.kafka"
scalaVersion := "2.12.4"
version := Version.version
libraryDependencies += "com.typesafe.scala-logging" %% "scala-logging" % "3.5.0"
libraryDependencies += "org.scalatest" %% "scalatest" % "3.0.1" % "test"
pomExtra :=
<url>https://github.com/farmdawgnation/kafka-detective</url>
<licenses>
<license>
<name>Apache 2</name>
<url>http://www.apache.org/licenses/LICENSE-2.0.txt</url>
<distribution>repo</distribution>
</license>
</licenses>
<scm> | <|file_sep|>original/api/build.sbt
name := "detective-api"
organization := "me.frmr.kafka"
scalaVersion := "2.12.4"
version := Version.version
libraryDependencies += "com.typesafe.scala-logging" %% "scala-logging" % "3.5.0"
libraryDependencies += "org.scalatest" %% "scalatest" % "3.0.1" % "test"
<|file_sep|>current/api/build.sbt
name := "detective-api"
organization := "me.frmr.kafka"
scalaVersion := "2.12.4"
version := Version.version
libraryDependencies += "com.typesafe.scala-logging" %% "scala-logging" % "3.5.0"
libraryDependencies += "org.scalatest" %% "scalatest" % "3.0.1" % "test"
<|file_sep|>updated/api/build.sbt
name := "detective-api"
organization := "me.frmr.kafka"
scalaVersion := "2.12.4"
version := Version.version
libraryDependencies += "com.typesafe.scala-logging" %% "scala-logging" % "3.5.0"
libraryDependencies += "org.scalatest" %% "scalatest" % "3.0.1" % "test"
pomExtra :=
<url>https://github.com/farmdawgnation/kafka-detective</url>
<licenses>
<license>
<name>Apache 2</name>
<url>http://www.apache.org/licenses/LICENSE-2.0.txt</url>
<distribution>repo</distribution>
</license>
</licenses>
<scm> | 1570c4e69988f366841a9be3ae777239aa312de0 | api/build.sbt | api/build.sbt | Scala |
<|file_sep|>original/configs/mac.conf.yaml
- defaults:
link:
relink: true
shell:
stdout: true
stderr: true
- clean: ['~']
- link:
~/.Brewfile: mac/Brewfile
~/.chunkwmrc: mac/chunkwmrc
~/.skhdrc: mac/skhdrc
- shell:
- [brew services start crisidev/chunkwm/chunkwm, Launchd start chunkwm]
- [brew services start koekeishiya/formulae/skhd, Launchd start skhd]
<|file_sep|>current/configs/mac.conf.yaml
- defaults:
link:
relink: true
shell:
stdout: true
stderr: true
- clean: ['~']
- link:
~/.Brewfile: mac/Brewfile
~/.chunkwmrc: mac/chunkwmrc
~/.skhdrc: mac/skhdrc
- shell:
- [brew services start crisidev/chunkwm/chunkwm, Launchd start chunkwm]
- [brew services start koekeishiya/formulae/skhd, Launchd start skhd]
<|file_sep|>updated/configs/mac.conf.yaml | - defaults:
link:
relink: true
shell:
stdout: true
stderr: true
- clean: ['~']
- link:
~/.Brewfile: mac/Brewfile
~/.chunkwmrc: mac/chunkwmrc
~/.skhdrc: mac/skhdrc
- shell:
- [brew bundle cleanup --file=$HOME/.Brewfile --force, Uninstall all deps not in Brewfile]
- [brew bundle install --file=$HOME/.Brewfile, Install or upgrade all deps not in Brewfile]
- [brew services start crisidev/chunkwm/chunkwm, Launchd start chunkwm]
- [brew services start koekeishiya/formulae/skhd, Launchd start skhd] | <|file_sep|>original/configs/mac.conf.yaml
- defaults:
link:
relink: true
shell:
stdout: true
stderr: true
- clean: ['~']
- link:
~/.Brewfile: mac/Brewfile
~/.chunkwmrc: mac/chunkwmrc
~/.skhdrc: mac/skhdrc
- shell:
- [brew services start crisidev/chunkwm/chunkwm, Launchd start chunkwm]
- [brew services start koekeishiya/formulae/skhd, Launchd start skhd]
<|file_sep|>current/configs/mac.conf.yaml
- defaults:
link:
relink: true
shell:
stdout: true
stderr: true
- clean: ['~']
- link:
~/.Brewfile: mac/Brewfile
~/.chunkwmrc: mac/chunkwmrc
~/.skhdrc: mac/skhdrc
- shell:
- [brew services start crisidev/chunkwm/chunkwm, Launchd start chunkwm]
- [brew services start koekeishiya/formulae/skhd, Launchd start skhd]
<|file_sep|>updated/configs/mac.conf.yaml
- defaults:
link:
relink: true
shell:
stdout: true
stderr: true
- clean: ['~']
- link:
~/.Brewfile: mac/Brewfile
~/.chunkwmrc: mac/chunkwmrc
~/.skhdrc: mac/skhdrc
- shell:
- [brew bundle cleanup --file=$HOME/.Brewfile --force, Uninstall all deps not in Brewfile]
- [brew bundle install --file=$HOME/.Brewfile, Install or upgrade all deps not in Brewfile]
- [brew services start crisidev/chunkwm/chunkwm, Launchd start chunkwm]
- [brew services start koekeishiya/formulae/skhd, Launchd start skhd] | c1d0fb9f0f314e558226f068793207e5866afd06 | configs/mac.conf.yaml | configs/mac.conf.yaml | YAML |
<|file_sep|>spec/controllers/users_controller_spec.rb.diff
original:
end
end
end
describe "GET #new" do
it "devrait réussir" do
get :new
expect(response).to have_http_status(:success)
updated:
end
<|file_sep|>spec/controllers/users_controller_spec.rb.diff
original:
it "devrait avoir le titre adéquat" do
get :new
expect(response).should have_selector("head title", :content => "Sign up")
end
end
updated:
describe "GET #new" do
it "devrait réussir" do
get :new
expect(response).to have_http_status(:success)
end
<|file_sep|>original/spec/controllers/users_controller_spec.rb
response.should have_selector("title", :content => @user.nom)
end
it "devrait inclure le nom de l'utilisateur" do
get :show, :id => @user
response.should have_selector("h1", :content => @user.nom)
end
it "devrait avoir une image de profil" do
get :show, :id => @user
response.should have_selector("h1>img", :class => "gravatar")
end
end
end
describe "GET #new" do
it "devrait réussir" do
get :new
expect(response).to have_http_status(:success)
end
<|file_sep|>current/spec/controllers/users_controller_spec.rb
response.should have_selector("title", :content => @user.nom)
end
it "devrait inclure le nom de l'utilisateur" do
get :show, :id => @user
response.should have_selector("h1", :content => @user.nom)
end
it "devrait avoir une image de profil" do
get :show, :id => @user
response.should have_selector("h1>img", :class => "gravatar")
end
end
describe "GET #new" do
it "devrait réussir" do
get :new
expect(response).to have_http_status(:success)
end
end
<|file_sep|>updated/spec/controllers/users_controller_spec.rb | response.should have_selector("h1", :content => @user.nom)
end
it "devrait avoir une image de profil" do
get :show, :id => @user
response.should have_selector("h1>img", :class => "gravatar")
end
end
describe "GET #new" do
it "devrait réussir" do
get :new
expect(response).to have_http_status(:success)
end
it "devrait avoir le titre adéquat" do
get :new
expect(response).should have_selector("head title", :content => "Sign up")
end
end
end | <|file_sep|>spec/controllers/users_controller_spec.rb.diff
original:
end
end
end
describe "GET #new" do
it "devrait réussir" do
get :new
expect(response).to have_http_status(:success)
updated:
end
<|file_sep|>spec/controllers/users_controller_spec.rb.diff
original:
it "devrait avoir le titre adéquat" do
get :new
expect(response).should have_selector("head title", :content => "Sign up")
end
end
updated:
describe "GET #new" do
it "devrait réussir" do
get :new
expect(response).to have_http_status(:success)
end
<|file_sep|>original/spec/controllers/users_controller_spec.rb
response.should have_selector("title", :content => @user.nom)
end
it "devrait inclure le nom de l'utilisateur" do
get :show, :id => @user
response.should have_selector("h1", :content => @user.nom)
end
it "devrait avoir une image de profil" do
get :show, :id => @user
response.should have_selector("h1>img", :class => "gravatar")
end
end
end
describe "GET #new" do
it "devrait réussir" do
get :new
expect(response).to have_http_status(:success)
end
<|file_sep|>current/spec/controllers/users_controller_spec.rb
response.should have_selector("title", :content => @user.nom)
end
it "devrait inclure le nom de l'utilisateur" do
get :show, :id => @user
response.should have_selector("h1", :content => @user.nom)
end
it "devrait avoir une image de profil" do
get :show, :id => @user
response.should have_selector("h1>img", :class => "gravatar")
end
end
describe "GET #new" do
it "devrait réussir" do
get :new
expect(response).to have_http_status(:success)
end
end
<|file_sep|>updated/spec/controllers/users_controller_spec.rb
response.should have_selector("h1", :content => @user.nom)
end
it "devrait avoir une image de profil" do
get :show, :id => @user
response.should have_selector("h1>img", :class => "gravatar")
end
end
describe "GET #new" do
it "devrait réussir" do
get :new
expect(response).to have_http_status(:success)
end
it "devrait avoir le titre adéquat" do
get :new
expect(response).should have_selector("head title", :content => "Sign up")
end
end
end | ddae5b4015e0b827a1d770189fe28075367e6a81 | spec/controllers/users_controller_spec.rb | spec/controllers/users_controller_spec.rb | Ruby |
<|file_sep|>original/package.json
"url": "git://github.com/veged/coa.git"
},
"directories": {
"lib": "./lib"
},
"dependencies": {
"q": "~0.8.8"
},
"devDependencies": {
"coffee-script": "~1.4.0",
"vows": "~0.6.4"
},
"engines": {
"node": ">= 0.6.0"
},
"licenses": [
{
"type": "MIT"
}
],
"optionalDependencies": {}
<|file_sep|>current/package.json
"url": "git://github.com/veged/coa.git"
},
"directories": {
"lib": "./lib"
},
"dependencies": {
"q": "~0.8.8"
},
"devDependencies": {
"coffee-script": "~1.4.0",
"vows": "~0.6.4"
},
"engines": {
"node": ">= 0.6.0"
},
"licenses": [
{
"type": "MIT"
}
],
"optionalDependencies": {}
<|file_sep|>updated/package.json | "url": "git://github.com/veged/coa.git"
},
"directories": {
"lib": "./lib"
},
"dependencies": {
"q": "~0.8.8"
},
"devDependencies": {
"coffee-script": "~1.4.0",
"istanbul": "~0.1.11",
"vows": "~0.6.4"
},
"engines": {
"node": ">= 0.6.0"
},
"licenses": [
{
"type": "MIT"
}
], | <|file_sep|>original/package.json
"url": "git://github.com/veged/coa.git"
},
"directories": {
"lib": "./lib"
},
"dependencies": {
"q": "~0.8.8"
},
"devDependencies": {
"coffee-script": "~1.4.0",
"vows": "~0.6.4"
},
"engines": {
"node": ">= 0.6.0"
},
"licenses": [
{
"type": "MIT"
}
],
"optionalDependencies": {}
<|file_sep|>current/package.json
"url": "git://github.com/veged/coa.git"
},
"directories": {
"lib": "./lib"
},
"dependencies": {
"q": "~0.8.8"
},
"devDependencies": {
"coffee-script": "~1.4.0",
"vows": "~0.6.4"
},
"engines": {
"node": ">= 0.6.0"
},
"licenses": [
{
"type": "MIT"
}
],
"optionalDependencies": {}
<|file_sep|>updated/package.json
"url": "git://github.com/veged/coa.git"
},
"directories": {
"lib": "./lib"
},
"dependencies": {
"q": "~0.8.8"
},
"devDependencies": {
"coffee-script": "~1.4.0",
"istanbul": "~0.1.11",
"vows": "~0.6.4"
},
"engines": {
"node": ">= 0.6.0"
},
"licenses": [
{
"type": "MIT"
}
], | 2a0c0f67bcb1b4079b22ccf6ddf7a7d79f640bf4 | package.json | package.json | JSON |
<|file_sep|>src/Data.js.diff
original:
updated:
const db = new minimongo();
db.debug = false;
<|file_sep|>original/src/Data.js
process.nextTick = setImmediate;
export default {
_endpoint: null,
_options: null,
ddp: null,
subscriptions: {},
db: new minimongo(),
calls: [],
hasBeenConnected: false,
getUrl() {
return this._endpoint.substring(0, this._endpoint.indexOf('/websocket'));
},
waitDdpReady(cb) {
if(this.ddp) {
cb();
} else {
setTimeout(()=>{
this.waitDdpReady(cb);
<|file_sep|>current/src/Data.js
process.nextTick = setImmediate;
const db = new minimongo();
db.debug = false;
export default {
_endpoint: null,
_options: null,
ddp: null,
subscriptions: {},
db: new minimongo(),
calls: [],
hasBeenConnected: false,
getUrl() {
return this._endpoint.substring(0, this._endpoint.indexOf('/websocket'));
},
waitDdpReady(cb) {
if(this.ddp) {
cb();
<|file_sep|>updated/src/Data.js | process.nextTick = setImmediate;
const db = new minimongo();
db.debug = false;
export default {
_endpoint: null,
_options: null,
ddp: null,
subscriptions: {},
db: db,
calls: [],
hasBeenConnected: false,
getUrl() {
return this._endpoint.substring(0, this._endpoint.indexOf('/websocket'));
},
waitDdpReady(cb) {
if(this.ddp) {
cb(); | <|file_sep|>src/Data.js.diff
original:
updated:
const db = new minimongo();
db.debug = false;
<|file_sep|>original/src/Data.js
process.nextTick = setImmediate;
export default {
_endpoint: null,
_options: null,
ddp: null,
subscriptions: {},
db: new minimongo(),
calls: [],
hasBeenConnected: false,
getUrl() {
return this._endpoint.substring(0, this._endpoint.indexOf('/websocket'));
},
waitDdpReady(cb) {
if(this.ddp) {
cb();
} else {
setTimeout(()=>{
this.waitDdpReady(cb);
<|file_sep|>current/src/Data.js
process.nextTick = setImmediate;
const db = new minimongo();
db.debug = false;
export default {
_endpoint: null,
_options: null,
ddp: null,
subscriptions: {},
db: new minimongo(),
calls: [],
hasBeenConnected: false,
getUrl() {
return this._endpoint.substring(0, this._endpoint.indexOf('/websocket'));
},
waitDdpReady(cb) {
if(this.ddp) {
cb();
<|file_sep|>updated/src/Data.js
process.nextTick = setImmediate;
const db = new minimongo();
db.debug = false;
export default {
_endpoint: null,
_options: null,
ddp: null,
subscriptions: {},
db: db,
calls: [],
hasBeenConnected: false,
getUrl() {
return this._endpoint.substring(0, this._endpoint.indexOf('/websocket'));
},
waitDdpReady(cb) {
if(this.ddp) {
cb(); | 369dde81fc50f11c0fba432305b5188e0a245d26 | src/Data.js | src/Data.js | JavaScript |
<|file_sep|>original/bounce.py
<|file_sep|>current/bounce.py
<|file_sep|>updated/bounce.py | import curses
import random
def bounce(stdscr):
curses.curs_set(0)
height, width = stdscr.getmaxyx()
ypos, xpos = random.randrange(0, height - 1), random.randrange(0, width - 1)
ydir, xdir = 1, 1
stdscr.addstr(height//2, width//2, "Type a character")
stdscr.nodelay(1)
c = None
while True:
c = get_chr(stdscr) or c
if not c:
continue
stdscr.clear()
if ypos in {0, height - 1}:
ydir *= -1
if xpos in {0, width - 1}:
xdir *= -1
ypos += ydir | <|file_sep|>original/bounce.py
<|file_sep|>current/bounce.py
<|file_sep|>updated/bounce.py
import curses
import random
def bounce(stdscr):
curses.curs_set(0)
height, width = stdscr.getmaxyx()
ypos, xpos = random.randrange(0, height - 1), random.randrange(0, width - 1)
ydir, xdir = 1, 1
stdscr.addstr(height//2, width//2, "Type a character")
stdscr.nodelay(1)
c = None
while True:
c = get_chr(stdscr) or c
if not c:
continue
stdscr.clear()
if ypos in {0, height - 1}:
ydir *= -1
if xpos in {0, width - 1}:
xdir *= -1
ypos += ydir | 836b483e06f1221f35d543e534ddfab6020b06ed | bounce.py | bounce.py | Python |
<|file_sep|>original/pkg_server_ports.sh
<|file_sep|>current/pkg_server_ports.sh
<|file_sep|>updated/pkg_server_ports.sh | #!/usr/bin/env bash
PKGSERVER_SERVICES=$(svcs -H pkg/server | grep -v disabled | awk '{print $3}')
for service in ${PKGSERVER_SERVICES}; do
instance=$(echo $service | awk -F ':' '{print $3}')
echo -n "Repository service instance '$instance' is on port "
svccfg -s $instance listprop | grep port | awk '{print $3}'
done | sort -nk8 | <|file_sep|>original/pkg_server_ports.sh
<|file_sep|>current/pkg_server_ports.sh
<|file_sep|>updated/pkg_server_ports.sh
#!/usr/bin/env bash
PKGSERVER_SERVICES=$(svcs -H pkg/server | grep -v disabled | awk '{print $3}')
for service in ${PKGSERVER_SERVICES}; do
instance=$(echo $service | awk -F ':' '{print $3}')
echo -n "Repository service instance '$instance' is on port "
svccfg -s $instance listprop | grep port | awk '{print $3}'
done | sort -nk8 | b61d7738b5b586b96eb4711ccce4a0bcd18ce310 | pkg_server_ports.sh | pkg_server_ports.sh | Shell |
<|file_sep|>original/time_tracker_backend/lib/time_tracker_backend/endpoint.ex
plug Plug.Parsers,
parsers: [:urlencoded, :multipart, :json],
pass: ["*/*"],
json_decoder: Poison
plug Plug.MethodOverride
plug Plug.Head
# The session will be stored in the cookie and signed,
# this means its contents can be read but not tampered with.
# Set :encryption_salt if you would also like to encrypt it.
plug Plug.Session,
store: :cookie,
key: "_time_tracker_backend_key",
signing_salt: "oT8ZqGe8"
plug CORSPlug, expose: ["link", "total", "per-page"]
plug TimeTrackerBackend.Router
end
<|file_sep|>current/time_tracker_backend/lib/time_tracker_backend/endpoint.ex
plug Plug.Parsers,
parsers: [:urlencoded, :multipart, :json],
pass: ["*/*"],
json_decoder: Poison
plug Plug.MethodOverride
plug Plug.Head
# The session will be stored in the cookie and signed,
# this means its contents can be read but not tampered with.
# Set :encryption_salt if you would also like to encrypt it.
plug Plug.Session,
store: :cookie,
key: "_time_tracker_backend_key",
signing_salt: "oT8ZqGe8"
plug CORSPlug, expose: ["link", "total", "per-page"]
plug TimeTrackerBackend.Router
end
<|file_sep|>updated/time_tracker_backend/lib/time_tracker_backend/endpoint.ex |
plug Plug.Parsers,
parsers: [:urlencoded, :multipart, :json],
pass: ["*/*"],
json_decoder: Poison
plug Plug.MethodOverride
plug Plug.Head
# The session will be stored in the cookie and signed,
# this means its contents can be read but not tampered with.
# Set :encryption_salt if you would also like to encrypt it.
plug Plug.Session,
store: :cookie,
key: "_time_tracker_backend_key",
signing_salt: "oT8ZqGe8"
plug CORSPlug, expose: ["link", "total", "per-page", "total-pages", "page-number"]
plug TimeTrackerBackend.Router
end | <|file_sep|>original/time_tracker_backend/lib/time_tracker_backend/endpoint.ex
plug Plug.Parsers,
parsers: [:urlencoded, :multipart, :json],
pass: ["*/*"],
json_decoder: Poison
plug Plug.MethodOverride
plug Plug.Head
# The session will be stored in the cookie and signed,
# this means its contents can be read but not tampered with.
# Set :encryption_salt if you would also like to encrypt it.
plug Plug.Session,
store: :cookie,
key: "_time_tracker_backend_key",
signing_salt: "oT8ZqGe8"
plug CORSPlug, expose: ["link", "total", "per-page"]
plug TimeTrackerBackend.Router
end
<|file_sep|>current/time_tracker_backend/lib/time_tracker_backend/endpoint.ex
plug Plug.Parsers,
parsers: [:urlencoded, :multipart, :json],
pass: ["*/*"],
json_decoder: Poison
plug Plug.MethodOverride
plug Plug.Head
# The session will be stored in the cookie and signed,
# this means its contents can be read but not tampered with.
# Set :encryption_salt if you would also like to encrypt it.
plug Plug.Session,
store: :cookie,
key: "_time_tracker_backend_key",
signing_salt: "oT8ZqGe8"
plug CORSPlug, expose: ["link", "total", "per-page"]
plug TimeTrackerBackend.Router
end
<|file_sep|>updated/time_tracker_backend/lib/time_tracker_backend/endpoint.ex
plug Plug.Parsers,
parsers: [:urlencoded, :multipart, :json],
pass: ["*/*"],
json_decoder: Poison
plug Plug.MethodOverride
plug Plug.Head
# The session will be stored in the cookie and signed,
# this means its contents can be read but not tampered with.
# Set :encryption_salt if you would also like to encrypt it.
plug Plug.Session,
store: :cookie,
key: "_time_tracker_backend_key",
signing_salt: "oT8ZqGe8"
plug CORSPlug, expose: ["link", "total", "per-page", "total-pages", "page-number"]
plug TimeTrackerBackend.Router
end | ebab7baff9882537254e7c73bf4ccb3c8be761fb | time_tracker_backend/lib/time_tracker_backend/endpoint.ex | time_tracker_backend/lib/time_tracker_backend/endpoint.ex | Elixir |
<|file_sep|>opps/contrib/logging/models.py.diff
original:
updated:
verbose_name=_(u'User')
<|file_sep|>original/opps/contrib/logging/models.py
settings.AUTH_USER_MODEL,
null=True, blank=True,
)
application = models.CharField(
_(u"Application"),
max_length=75,
null=True, blank=True,
db_index=True)
action = models.CharField(
_(u"Action"),
max_length=50,
null=True, blank=True,
db_index=True)
text = models.TextField(
_(u"Text"),
null=True, blank=True,
db_index=True)
def save(self, *args, **kwargs):
self.published = True
super(Logging, self).save(*args, **kwargs)
<|file_sep|>current/opps/contrib/logging/models.py
null=True, blank=True,
verbose_name=_(u'User')
)
application = models.CharField(
_(u"Application"),
max_length=75,
null=True, blank=True,
db_index=True)
action = models.CharField(
_(u"Action"),
max_length=50,
null=True, blank=True,
db_index=True)
text = models.TextField(
_(u"Text"),
null=True, blank=True,
db_index=True)
def save(self, *args, **kwargs):
self.published = True
super(Logging, self).save(*args, **kwargs)
<|file_sep|>updated/opps/contrib/logging/models.py | _(u"Application"),
max_length=75,
null=True, blank=True,
db_index=True)
action = models.CharField(
_(u"Action"),
max_length=50,
null=True, blank=True,
db_index=True)
text = models.TextField(
_(u"Text"),
null=True, blank=True,
db_index=True)
def save(self, *args, **kwargs):
self.published = True
super(Logging, self).save(*args, **kwargs)
class Meta:
verbose_name = _(u'Logging')
verbose_name_plural = _(u'Loggings') | <|file_sep|>opps/contrib/logging/models.py.diff
original:
updated:
verbose_name=_(u'User')
<|file_sep|>original/opps/contrib/logging/models.py
settings.AUTH_USER_MODEL,
null=True, blank=True,
)
application = models.CharField(
_(u"Application"),
max_length=75,
null=True, blank=True,
db_index=True)
action = models.CharField(
_(u"Action"),
max_length=50,
null=True, blank=True,
db_index=True)
text = models.TextField(
_(u"Text"),
null=True, blank=True,
db_index=True)
def save(self, *args, **kwargs):
self.published = True
super(Logging, self).save(*args, **kwargs)
<|file_sep|>current/opps/contrib/logging/models.py
null=True, blank=True,
verbose_name=_(u'User')
)
application = models.CharField(
_(u"Application"),
max_length=75,
null=True, blank=True,
db_index=True)
action = models.CharField(
_(u"Action"),
max_length=50,
null=True, blank=True,
db_index=True)
text = models.TextField(
_(u"Text"),
null=True, blank=True,
db_index=True)
def save(self, *args, **kwargs):
self.published = True
super(Logging, self).save(*args, **kwargs)
<|file_sep|>updated/opps/contrib/logging/models.py
_(u"Application"),
max_length=75,
null=True, blank=True,
db_index=True)
action = models.CharField(
_(u"Action"),
max_length=50,
null=True, blank=True,
db_index=True)
text = models.TextField(
_(u"Text"),
null=True, blank=True,
db_index=True)
def save(self, *args, **kwargs):
self.published = True
super(Logging, self).save(*args, **kwargs)
class Meta:
verbose_name = _(u'Logging')
verbose_name_plural = _(u'Loggings') | 528edba420089249bd58c0621e06225db84e223f | opps/contrib/logging/models.py | opps/contrib/logging/models.py | Python |
<|file_sep|>.storybook/webpack.config.babel.js.diff
original:
export default merge(
initConf({
context: path.resolve(__dirname, '..'),
mode: 'development',
injectStyles: true,
}),
{
plugins: [
// Set up mock for WebExt APIs. Idea from:
// https://github.com/aeksco/react-typescript-web-extension-starter/blob/f62486ec8518eb5cf78687a2e658505fd528dc8a/.storybook/webpack.config.js
new NormalModuleReplacementPlugin(
/webextension-polyfill-ts/,
resource => {
const absMockPath = path.resolve(
__dirname,
'mocks/webextension-polyfill-ts.js',
)
updated:
export default async () =>
merge(
await initConf({
context: path.resolve(__dirname, '..'),
mode: 'development',
injectStyles: true,
}),
{
plugins: [
// Set up mock for WebExt APIs. Idea from:
// https://github.com/aeksco/react-typescript-web-extension-starter/blob/f62486ec8518eb5cf78687a2e658505fd528dc8a/.storybook/webpack.config.js
new NormalModuleReplacementPlugin(
/webextension-polyfill-ts/,
(resource) => {
const absMockPath = path.resolve(
__dirname,
'mocks/webextension-polyfill-ts.js',
)
<|file_sep|>original/.storybook/webpack.config.babel.js
{
plugins: [
// Set up mock for WebExt APIs. Idea from:
// https://github.com/aeksco/react-typescript-web-extension-starter/blob/f62486ec8518eb5cf78687a2e658505fd528dc8a/.storybook/webpack.config.js
new NormalModuleReplacementPlugin(
/webextension-polyfill-ts/,
resource => {
const absMockPath = path.resolve(
__dirname,
'mocks/webextension-polyfill-ts.js',
)
resource.request = path.relative(
resource.context,
absMockPath,
)
},
),
],
},
)
<|file_sep|>current/.storybook/webpack.config.babel.js
{
plugins: [
// Set up mock for WebExt APIs. Idea from:
// https://github.com/aeksco/react-typescript-web-extension-starter/blob/f62486ec8518eb5cf78687a2e658505fd528dc8a/.storybook/webpack.config.js
new NormalModuleReplacementPlugin(
/webextension-polyfill-ts/,
(resource) => {
const absMockPath = path.resolve(
__dirname,
'mocks/webextension-polyfill-ts.js',
)
resource.request = path.relative(
resource.context,
absMockPath,
)
},
),
],
},
)
<|file_sep|>updated/.storybook/webpack.config.babel.js | {
plugins: [
// Set up mock for WebExt APIs. Idea from:
// https://github.com/aeksco/react-typescript-web-extension-starter/blob/f62486ec8518eb5cf78687a2e658505fd528dc8a/.storybook/webpack.config.js
new NormalModuleReplacementPlugin(
/webextension-polyfill-ts/,
(resource) => {
const absMockPath = path.resolve(
__dirname,
'mocks/webextension-polyfill-ts.js',
)
resource.request = path.relative(
resource.context,
absMockPath,
)
},
),
],
},
) | <|file_sep|>.storybook/webpack.config.babel.js.diff
original:
export default merge(
initConf({
context: path.resolve(__dirname, '..'),
mode: 'development',
injectStyles: true,
}),
{
plugins: [
// Set up mock for WebExt APIs. Idea from:
// https://github.com/aeksco/react-typescript-web-extension-starter/blob/f62486ec8518eb5cf78687a2e658505fd528dc8a/.storybook/webpack.config.js
new NormalModuleReplacementPlugin(
/webextension-polyfill-ts/,
resource => {
const absMockPath = path.resolve(
__dirname,
'mocks/webextension-polyfill-ts.js',
)
updated:
export default async () =>
merge(
await initConf({
context: path.resolve(__dirname, '..'),
mode: 'development',
injectStyles: true,
}),
{
plugins: [
// Set up mock for WebExt APIs. Idea from:
// https://github.com/aeksco/react-typescript-web-extension-starter/blob/f62486ec8518eb5cf78687a2e658505fd528dc8a/.storybook/webpack.config.js
new NormalModuleReplacementPlugin(
/webextension-polyfill-ts/,
(resource) => {
const absMockPath = path.resolve(
__dirname,
'mocks/webextension-polyfill-ts.js',
)
<|file_sep|>original/.storybook/webpack.config.babel.js
{
plugins: [
// Set up mock for WebExt APIs. Idea from:
// https://github.com/aeksco/react-typescript-web-extension-starter/blob/f62486ec8518eb5cf78687a2e658505fd528dc8a/.storybook/webpack.config.js
new NormalModuleReplacementPlugin(
/webextension-polyfill-ts/,
resource => {
const absMockPath = path.resolve(
__dirname,
'mocks/webextension-polyfill-ts.js',
)
resource.request = path.relative(
resource.context,
absMockPath,
)
},
),
],
},
)
<|file_sep|>current/.storybook/webpack.config.babel.js
{
plugins: [
// Set up mock for WebExt APIs. Idea from:
// https://github.com/aeksco/react-typescript-web-extension-starter/blob/f62486ec8518eb5cf78687a2e658505fd528dc8a/.storybook/webpack.config.js
new NormalModuleReplacementPlugin(
/webextension-polyfill-ts/,
(resource) => {
const absMockPath = path.resolve(
__dirname,
'mocks/webextension-polyfill-ts.js',
)
resource.request = path.relative(
resource.context,
absMockPath,
)
},
),
],
},
)
<|file_sep|>updated/.storybook/webpack.config.babel.js
{
plugins: [
// Set up mock for WebExt APIs. Idea from:
// https://github.com/aeksco/react-typescript-web-extension-starter/blob/f62486ec8518eb5cf78687a2e658505fd528dc8a/.storybook/webpack.config.js
new NormalModuleReplacementPlugin(
/webextension-polyfill-ts/,
(resource) => {
const absMockPath = path.resolve(
__dirname,
'mocks/webextension-polyfill-ts.js',
)
resource.request = path.relative(
resource.context,
absMockPath,
)
},
),
],
},
) | 2a9e9e8640780e5faa0d47f5d2da1028f8104f69 | .storybook/webpack.config.babel.js | .storybook/webpack.config.babel.js | JavaScript |
<|file_sep|>original/layouts/partials/showcase/display-horizontal.html
<div class="px-6 py-4 w-full md:w-1/3 z-0 md:order-1">
<div class="md:bg-white md:border py-2 pl-6 pr-2 -ml-8 rounded-r-lg ">
<p>Showcase</p>
<div class="font-bold text-base mb-2">
<a href="{{ .Permalink }}" class="dim text-xl text-black no-underline title">
{{ .Title }}
</a>
</div>
</div>
</div>
{{ with $image }}
<a href="{{ $.URL }}" class="no-underline w-full md:w-2/3 md:order-0 z-50">
{{ if in (print $.Site.BaseURL) "localhost" }}
<img data-src="{{ . }}" src="/images/placeholder.svg" class="lazyload mw-100 shadow ">
{{ else }}
{{ $img := replace . "/uploads" ($.Param "dev.image_url") }}
<img src="/images/placeholder.svg" data-src="{{ $img }}?{{ (querify "w" "460") | safeURL }}" class="lazyload shadow mw-100">
{{ end }}
</a>
{{ end }}
</div>
<|file_sep|>current/layouts/partials/showcase/display-horizontal.html
<div class="px-6 py-4 w-full md:w-1/3 z-0 md:order-1">
<div class="md:bg-white md:border py-2 pl-6 pr-2 -ml-8 rounded-r-lg ">
<p>Showcase</p>
<div class="font-bold text-base mb-2">
<a href="{{ .Permalink }}" class="dim text-xl text-black no-underline title">
{{ .Title }}
</a>
</div>
</div>
</div>
{{ with $image }}
<a href="{{ $.URL }}" class="no-underline w-full md:w-2/3 md:order-0 z-50">
{{ if in (print $.Site.BaseURL) "localhost" }}
<img data-src="{{ . }}" src="/images/placeholder.svg" class="lazyload mw-100 shadow ">
{{ else }}
{{ $img := replace . "/uploads" ($.Param "dev.image_url") }}
<img src="/images/placeholder.svg" data-src="{{ $img }}?{{ (querify "w" "460") | safeURL }}" class="lazyload shadow mw-100">
{{ end }}
</a>
{{ end }}
</div>
<|file_sep|>updated/layouts/partials/showcase/display-horizontal.html | <div class="px-6 py-4 w-full md:w-1/3 z-0 md:order-1">
<div class="md:bg-white md:border py-2 pl-6 pr-2 -ml-8 rounded-r-lg ">
<p>Showcase</p>
<div class="font-bold text-base mb-2">
<a href="{{ .Permalink }}" class="dim text-xl text-black no-underline title">
{{ .Title }}
</a>
</div>
</div>
</div>
{{ with $image }}
<a href="{{ $.URL }}" class="no-underline w-full md:w-2/3 md:order-0 z-50">
{{ if in (print $.Site.BaseURL) "localhost" }}
<img data-src="{{ . }}" src="/images/placeholder.svg" class="lazyload mw-100 shadow ">
{{ else }}
{{ $img := replace . "/uploads" ($.Param "dev.image_url") }}
<img src="/images/placeholder.svg" data-src="{{ $img }}?{{ (querify "w" "460" "fit" "crop") | safeURL }}" class="lazyload shadow mw-100">
{{ end }}
</a>
{{ end }}
</div> | <|file_sep|>original/layouts/partials/showcase/display-horizontal.html
<div class="px-6 py-4 w-full md:w-1/3 z-0 md:order-1">
<div class="md:bg-white md:border py-2 pl-6 pr-2 -ml-8 rounded-r-lg ">
<p>Showcase</p>
<div class="font-bold text-base mb-2">
<a href="{{ .Permalink }}" class="dim text-xl text-black no-underline title">
{{ .Title }}
</a>
</div>
</div>
</div>
{{ with $image }}
<a href="{{ $.URL }}" class="no-underline w-full md:w-2/3 md:order-0 z-50">
{{ if in (print $.Site.BaseURL) "localhost" }}
<img data-src="{{ . }}" src="/images/placeholder.svg" class="lazyload mw-100 shadow ">
{{ else }}
{{ $img := replace . "/uploads" ($.Param "dev.image_url") }}
<img src="/images/placeholder.svg" data-src="{{ $img }}?{{ (querify "w" "460") | safeURL }}" class="lazyload shadow mw-100">
{{ end }}
</a>
{{ end }}
</div>
<|file_sep|>current/layouts/partials/showcase/display-horizontal.html
<div class="px-6 py-4 w-full md:w-1/3 z-0 md:order-1">
<div class="md:bg-white md:border py-2 pl-6 pr-2 -ml-8 rounded-r-lg ">
<p>Showcase</p>
<div class="font-bold text-base mb-2">
<a href="{{ .Permalink }}" class="dim text-xl text-black no-underline title">
{{ .Title }}
</a>
</div>
</div>
</div>
{{ with $image }}
<a href="{{ $.URL }}" class="no-underline w-full md:w-2/3 md:order-0 z-50">
{{ if in (print $.Site.BaseURL) "localhost" }}
<img data-src="{{ . }}" src="/images/placeholder.svg" class="lazyload mw-100 shadow ">
{{ else }}
{{ $img := replace . "/uploads" ($.Param "dev.image_url") }}
<img src="/images/placeholder.svg" data-src="{{ $img }}?{{ (querify "w" "460") | safeURL }}" class="lazyload shadow mw-100">
{{ end }}
</a>
{{ end }}
</div>
<|file_sep|>updated/layouts/partials/showcase/display-horizontal.html
<div class="px-6 py-4 w-full md:w-1/3 z-0 md:order-1">
<div class="md:bg-white md:border py-2 pl-6 pr-2 -ml-8 rounded-r-lg ">
<p>Showcase</p>
<div class="font-bold text-base mb-2">
<a href="{{ .Permalink }}" class="dim text-xl text-black no-underline title">
{{ .Title }}
</a>
</div>
</div>
</div>
{{ with $image }}
<a href="{{ $.URL }}" class="no-underline w-full md:w-2/3 md:order-0 z-50">
{{ if in (print $.Site.BaseURL) "localhost" }}
<img data-src="{{ . }}" src="/images/placeholder.svg" class="lazyload mw-100 shadow ">
{{ else }}
{{ $img := replace . "/uploads" ($.Param "dev.image_url") }}
<img src="/images/placeholder.svg" data-src="{{ $img }}?{{ (querify "w" "460" "fit" "crop") | safeURL }}" class="lazyload shadow mw-100">
{{ end }}
</a>
{{ end }}
</div> | a441154ec5a2fea3674afc5f4c9e114393d2aaed | layouts/partials/showcase/display-horizontal.html | layouts/partials/showcase/display-horizontal.html | HTML |
<|file_sep|>original/tests/stop_job.yml
<|file_sep|>current/tests/stop_job.yml
<|file_sep|>updated/tests/stop_job.yml | ---
- hosts: all
tasks:
- name: Check whether example job is submitted.
command: 'nomad status'
register: job_status_nomad
changed_when: no
- name: Stop example job
command: 'nomad stop example'
when: "'example' in job_status_nomad.stdout" | <|file_sep|>original/tests/stop_job.yml
<|file_sep|>current/tests/stop_job.yml
<|file_sep|>updated/tests/stop_job.yml
---
- hosts: all
tasks:
- name: Check whether example job is submitted.
command: 'nomad status'
register: job_status_nomad
changed_when: no
- name: Stop example job
command: 'nomad stop example'
when: "'example' in job_status_nomad.stdout" | bb6c0618070f88da1e8868cb217819e7d6e61bfa | tests/stop_job.yml | tests/stop_job.yml | YAML |
<|file_sep|>original/.github/workflows/gradle-java14.yml
<|file_sep|>current/.github/workflows/gradle-java14.yml
<|file_sep|>updated/.github/workflows/gradle-java14.yml | # This workflow will build a Java project with Gradle
# For more information see: https://help.github.com/actions/language-and-framework-guides/building-and-testing-java-with-gradle
name: Java 14
on:
push:
branches:
- master
jobs:
build:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- name: Set up JDK 14
uses: actions/setup-java@v1
with:
java-version: 14 | <|file_sep|>original/.github/workflows/gradle-java14.yml
<|file_sep|>current/.github/workflows/gradle-java14.yml
<|file_sep|>updated/.github/workflows/gradle-java14.yml
# This workflow will build a Java project with Gradle
# For more information see: https://help.github.com/actions/language-and-framework-guides/building-and-testing-java-with-gradle
name: Java 14
on:
push:
branches:
- master
jobs:
build:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- name: Set up JDK 14
uses: actions/setup-java@v1
with:
java-version: 14 | 4e56506951f44884c56bd0aac3a9e9411baf11f3 | .github/workflows/gradle-java14.yml | .github/workflows/gradle-java14.yml | YAML |
<|file_sep|>package.json.diff
original:
"version": "1.0.1",
updated:
"version": "1.0.2",
<|file_sep|>package.json.diff
original:
"test": "echo \"Error: no test specified\" && exit 1",
updated:
"test": "standard index.jsx",
<|file_sep|>original/package.json
},
"keywords": [
"react",
"react-component",
"keyboard",
"shortcuts"
],
"author": "Florian Hartmann",
"license": "MIT",
"bugs": {
"url": "https://github.com/florian/react-shortcut-chooser/issues"
},
"homepage": "https://github.com/florian/react-shortcut-chooser#readme",
"dependencies": {
"key-event-to-string": "^1.1.0",
"react": "^0.14.7"
},
"devDependencies": {
"react-tools": "^0.13.3"
}
}
<|file_sep|>current/package.json
},
"keywords": [
"react",
"react-component",
"keyboard",
"shortcuts"
],
"author": "Florian Hartmann",
"license": "MIT",
"bugs": {
"url": "https://github.com/florian/react-shortcut-chooser/issues"
},
"homepage": "https://github.com/florian/react-shortcut-chooser#readme",
"dependencies": {
"key-event-to-string": "^1.1.0",
"react": "^0.14.7"
},
"devDependencies": {
"react-tools": "^0.13.3"
}
}
<|file_sep|>updated/package.json | "react",
"react-component",
"keyboard",
"shortcuts"
],
"author": "Florian Hartmann",
"license": "MIT",
"bugs": {
"url": "https://github.com/florian/react-shortcut-chooser/issues"
},
"homepage": "https://github.com/florian/react-shortcut-chooser#readme",
"dependencies": {
"key-event-to-string": "^1.1.0"
},
"peerDependencies": {
"react": "^0.14.7"
},
"devDependencies": {
"react-tools": "^0.13.3"
}
} | <|file_sep|>package.json.diff
original:
"version": "1.0.1",
updated:
"version": "1.0.2",
<|file_sep|>package.json.diff
original:
"test": "echo \"Error: no test specified\" && exit 1",
updated:
"test": "standard index.jsx",
<|file_sep|>original/package.json
},
"keywords": [
"react",
"react-component",
"keyboard",
"shortcuts"
],
"author": "Florian Hartmann",
"license": "MIT",
"bugs": {
"url": "https://github.com/florian/react-shortcut-chooser/issues"
},
"homepage": "https://github.com/florian/react-shortcut-chooser#readme",
"dependencies": {
"key-event-to-string": "^1.1.0",
"react": "^0.14.7"
},
"devDependencies": {
"react-tools": "^0.13.3"
}
}
<|file_sep|>current/package.json
},
"keywords": [
"react",
"react-component",
"keyboard",
"shortcuts"
],
"author": "Florian Hartmann",
"license": "MIT",
"bugs": {
"url": "https://github.com/florian/react-shortcut-chooser/issues"
},
"homepage": "https://github.com/florian/react-shortcut-chooser#readme",
"dependencies": {
"key-event-to-string": "^1.1.0",
"react": "^0.14.7"
},
"devDependencies": {
"react-tools": "^0.13.3"
}
}
<|file_sep|>updated/package.json
"react",
"react-component",
"keyboard",
"shortcuts"
],
"author": "Florian Hartmann",
"license": "MIT",
"bugs": {
"url": "https://github.com/florian/react-shortcut-chooser/issues"
},
"homepage": "https://github.com/florian/react-shortcut-chooser#readme",
"dependencies": {
"key-event-to-string": "^1.1.0"
},
"peerDependencies": {
"react": "^0.14.7"
},
"devDependencies": {
"react-tools": "^0.13.3"
}
} | 9e46355a2e98333239dbd8cf874a7347873903f7 | package.json | package.json | JSON |
<|file_sep|>original/lib/rrj/janus/response_error.rb
<|file_sep|>current/lib/rrj/janus/response_error.rb
<|file_sep|>updated/lib/rrj/janus/response_error.rb | # frozen_string_literal: true
require 'bunny'
module RubyRabbitmqJanus
# @author VAILLANT Jeremy <jeremy.vaillant@dazzl.tv>
# Class for test if response return an janus error
class ResponseError
# Return an Hash to request
def initialize(request)
@request = Has.new(request)
end
# Test if response is an error
def test_request_return
@request['janus'] == 'error' ? true : false
end
# Return a reaon to error returning by janus
def reason
@request['error']['reason'] | <|file_sep|>original/lib/rrj/janus/response_error.rb
<|file_sep|>current/lib/rrj/janus/response_error.rb
<|file_sep|>updated/lib/rrj/janus/response_error.rb
# frozen_string_literal: true
require 'bunny'
module RubyRabbitmqJanus
# @author VAILLANT Jeremy <jeremy.vaillant@dazzl.tv>
# Class for test if response return an janus error
class ResponseError
# Return an Hash to request
def initialize(request)
@request = Has.new(request)
end
# Test if response is an error
def test_request_return
@request['janus'] == 'error' ? true : false
end
# Return a reaon to error returning by janus
def reason
@request['error']['reason'] | a45c28134a4d0ed74dfb2504896980e10eea2f96 | lib/rrj/janus/response_error.rb | lib/rrj/janus/response_error.rb | Ruby |
<|file_sep|>README.rdoc.diff
original:
updated:
In some applications it's also necessary to add routes in #{Rails.roo}/config/routes.rb for the ActionCost dashboard pages:
namespace :action_cost do
get '/' => 'dashboards#index'
match '/ca/:ca' => 'dashboards#ca', :as => 'ca', :constraints => { :ca => /.*/ }
end
<|file_sep|>README.rdoc.diff
original:
2. After each controller action completes, an action_cost summary is printed in the application logfile at the debug log_level.
updated:
2. After each controller action completes, an action_cost summary is #{Rails.root}/log/action_cost.log
<|file_sep|>original/README.rdoc
= ActionCost
ActionCost is a Rails 3 engine implemented as a gem.
It hooks into ActiveRecord (and RecordCache if used) and counts the number of SQL queries
per controller action and per table.
== Setup
Simply include action_cost into your Rails 3 application Gemfile.
gem 'action_cost', '0.0.1'
== Where is the data stored?
In the Rails process memory. The data goes away when the process dies.
== Where is the data visible?
1. Go to http://yourapp/action_cost, it will display a page with a list of controller action and the average number of queries made in each one of them. Click on a controller action to find which tables were queried.
2. After each controller action completes, an action_cost summary is printed in the application logfile at the debug log_level.
<|file_sep|>current/README.rdoc
== Setup
Simply include action_cost into your Rails 3 application Gemfile.
gem 'action_cost', '0.0.1'
In some applications it's also necessary to add routes in #{Rails.roo}/config/routes.rb for the ActionCost dashboard pages:
namespace :action_cost do
get '/' => 'dashboards#index'
match '/ca/:ca' => 'dashboards#ca', :as => 'ca', :constraints => { :ca => /.*/ }
end
== Where is the data stored?
In the Rails process memory. The data goes away when the process dies.
== Where is the data visible?
1. Go to http://yourapp/action_cost, it will display a page with a list of controller action and the average number of queries made in each one of them. Click on a controller action to find which tables were queried.
2. After each controller action completes, an action_cost summary is #{Rails.root}/log/action_cost.log
<|file_sep|>updated/README.rdoc | == Setup
Simply include action_cost into your Rails 3 application Gemfile.
gem 'action_cost', '0.0.1'
In some applications it's also necessary to add routes in #{Rails.roo}/config/routes.rb for the ActionCost dashboard pages:
namespace :action_cost do
get '/' => 'dashboards#index'
match '/ca/:ca' => 'dashboards#ca', :as => 'ca', :constraints => { :ca => /.*/ }
end
== Where is the data stored?
In the Rails process memory. The data goes away when the process dies.
== Where is the data visible?
1. Go to http://yourapp/action_cost, it will display a page with a list of controller action and the average number of queries made in each one of them. Click on a controller action to find which tables were queried.
2. After each controller action completes, an action_cost summary is #{Rails.root}/log/action_cost.log
| <|file_sep|>README.rdoc.diff
original:
updated:
In some applications it's also necessary to add routes in #{Rails.roo}/config/routes.rb for the ActionCost dashboard pages:
namespace :action_cost do
get '/' => 'dashboards#index'
match '/ca/:ca' => 'dashboards#ca', :as => 'ca', :constraints => { :ca => /.*/ }
end
<|file_sep|>README.rdoc.diff
original:
2. After each controller action completes, an action_cost summary is printed in the application logfile at the debug log_level.
updated:
2. After each controller action completes, an action_cost summary is #{Rails.root}/log/action_cost.log
<|file_sep|>original/README.rdoc
= ActionCost
ActionCost is a Rails 3 engine implemented as a gem.
It hooks into ActiveRecord (and RecordCache if used) and counts the number of SQL queries
per controller action and per table.
== Setup
Simply include action_cost into your Rails 3 application Gemfile.
gem 'action_cost', '0.0.1'
== Where is the data stored?
In the Rails process memory. The data goes away when the process dies.
== Where is the data visible?
1. Go to http://yourapp/action_cost, it will display a page with a list of controller action and the average number of queries made in each one of them. Click on a controller action to find which tables were queried.
2. After each controller action completes, an action_cost summary is printed in the application logfile at the debug log_level.
<|file_sep|>current/README.rdoc
== Setup
Simply include action_cost into your Rails 3 application Gemfile.
gem 'action_cost', '0.0.1'
In some applications it's also necessary to add routes in #{Rails.roo}/config/routes.rb for the ActionCost dashboard pages:
namespace :action_cost do
get '/' => 'dashboards#index'
match '/ca/:ca' => 'dashboards#ca', :as => 'ca', :constraints => { :ca => /.*/ }
end
== Where is the data stored?
In the Rails process memory. The data goes away when the process dies.
== Where is the data visible?
1. Go to http://yourapp/action_cost, it will display a page with a list of controller action and the average number of queries made in each one of them. Click on a controller action to find which tables were queried.
2. After each controller action completes, an action_cost summary is #{Rails.root}/log/action_cost.log
<|file_sep|>updated/README.rdoc
== Setup
Simply include action_cost into your Rails 3 application Gemfile.
gem 'action_cost', '0.0.1'
In some applications it's also necessary to add routes in #{Rails.roo}/config/routes.rb for the ActionCost dashboard pages:
namespace :action_cost do
get '/' => 'dashboards#index'
match '/ca/:ca' => 'dashboards#ca', :as => 'ca', :constraints => { :ca => /.*/ }
end
== Where is the data stored?
In the Rails process memory. The data goes away when the process dies.
== Where is the data visible?
1. Go to http://yourapp/action_cost, it will display a page with a list of controller action and the average number of queries made in each one of them. Click on a controller action to find which tables were queried.
2. After each controller action completes, an action_cost summary is #{Rails.root}/log/action_cost.log
| dc5b06df6c93c6235941621191de18b0576dd3c9 | README.rdoc | README.rdoc | RDoc |
<|file_sep|>original/pubspec.yaml
name: win32
description: A Dart library for accessing common Win32 APIs using FFI. No C required!
version: 2.3.6
homepage: https://win32.pub
repository: https://github.com/timsneath/win32
issue_tracker: https://github.com/timsneath/win32/issues
environment:
sdk: '>=2.14.0 <3.0.0'
dependencies:
ffi: ^1.0.0
dev_dependencies:
args: ^2.3.0
lints: ^1.0.1
path: ^1.8.1
test: ^1.20.1
winmd: 1.0.23 # Pinned deliberately
# dependency_overrides:
<|file_sep|>current/pubspec.yaml
name: win32
description: A Dart library for accessing common Win32 APIs using FFI. No C required!
version: 2.3.6
homepage: https://win32.pub
repository: https://github.com/timsneath/win32
issue_tracker: https://github.com/timsneath/win32/issues
environment:
sdk: '>=2.14.0 <3.0.0'
dependencies:
ffi: ^1.0.0
dev_dependencies:
args: ^2.3.0
lints: ^1.0.1
path: ^1.8.1
test: ^1.20.1
winmd: 1.0.23 # Pinned deliberately
# dependency_overrides:
<|file_sep|>updated/pubspec.yaml | name: win32
description: A Dart library for accessing common Win32 APIs using FFI. No C required!
version: 2.3.6
homepage: https://win32.pub
repository: https://github.com/timsneath/win32
issue_tracker: https://github.com/timsneath/win32/issues
environment:
sdk: '>=2.14.0 <3.0.0'
# Declare that this package only works on Windows.
platforms:
windows:
dependencies:
ffi: ^1.0.0
dev_dependencies:
args: ^2.3.0
lints: ^1.0.1
path: ^1.8.1 | <|file_sep|>original/pubspec.yaml
name: win32
description: A Dart library for accessing common Win32 APIs using FFI. No C required!
version: 2.3.6
homepage: https://win32.pub
repository: https://github.com/timsneath/win32
issue_tracker: https://github.com/timsneath/win32/issues
environment:
sdk: '>=2.14.0 <3.0.0'
dependencies:
ffi: ^1.0.0
dev_dependencies:
args: ^2.3.0
lints: ^1.0.1
path: ^1.8.1
test: ^1.20.1
winmd: 1.0.23 # Pinned deliberately
# dependency_overrides:
<|file_sep|>current/pubspec.yaml
name: win32
description: A Dart library for accessing common Win32 APIs using FFI. No C required!
version: 2.3.6
homepage: https://win32.pub
repository: https://github.com/timsneath/win32
issue_tracker: https://github.com/timsneath/win32/issues
environment:
sdk: '>=2.14.0 <3.0.0'
dependencies:
ffi: ^1.0.0
dev_dependencies:
args: ^2.3.0
lints: ^1.0.1
path: ^1.8.1
test: ^1.20.1
winmd: 1.0.23 # Pinned deliberately
# dependency_overrides:
<|file_sep|>updated/pubspec.yaml
name: win32
description: A Dart library for accessing common Win32 APIs using FFI. No C required!
version: 2.3.6
homepage: https://win32.pub
repository: https://github.com/timsneath/win32
issue_tracker: https://github.com/timsneath/win32/issues
environment:
sdk: '>=2.14.0 <3.0.0'
# Declare that this package only works on Windows.
platforms:
windows:
dependencies:
ffi: ^1.0.0
dev_dependencies:
args: ^2.3.0
lints: ^1.0.1
path: ^1.8.1 | 8bae327c74033d1afe16fe6fc804f1cc474c5452 | pubspec.yaml | pubspec.yaml | YAML |
<|file_sep|>src/tools/StackScrollMouseWheelTool.js.diff
original:
updated:
invert: false,
<|file_sep|>src/tools/StackScrollMouseWheelTool.js.diff
original:
const { loop, allowSkipping } = this.configuration;
updated:
const { loop, allowSkipping, invert } = this.configuration;
const direction = invert ? -images : images;
<|file_sep|>original/src/tools/StackScrollMouseWheelTool.js
name: 'StackScrollMouseWheel',
supportedInteractionTypes: ['MouseWheel'],
configuration: {
loop: false,
allowSkipping: true,
},
};
const initialConfiguration = Object.assign(defaultConfig, configuration);
super(initialConfiguration);
this.initialConfiguration = initialConfiguration;
}
mouseWheelCallback(evt) {
const { direction: images, element } = evt.detail;
const { loop, allowSkipping } = this.configuration;
scroll(element, -images, loop, allowSkipping);
}
}
<|file_sep|>current/src/tools/StackScrollMouseWheelTool.js
configuration: {
loop: false,
allowSkipping: true,
invert: false,
},
};
const initialConfiguration = Object.assign(defaultConfig, configuration);
super(initialConfiguration);
this.initialConfiguration = initialConfiguration;
}
mouseWheelCallback(evt) {
const { direction: images, element } = evt.detail;
const { loop, allowSkipping, invert } = this.configuration;
const direction = invert ? -images : images;
scroll(element, -images, loop, allowSkipping);
}
}
<|file_sep|>updated/src/tools/StackScrollMouseWheelTool.js | configuration: {
loop: false,
allowSkipping: true,
invert: false,
},
};
const initialConfiguration = Object.assign(defaultConfig, configuration);
super(initialConfiguration);
this.initialConfiguration = initialConfiguration;
}
mouseWheelCallback(evt) {
const { direction: images, element } = evt.detail;
const { loop, allowSkipping, invert } = this.configuration;
const direction = invert ? -images : images;
scroll(element, direction, loop, allowSkipping);
}
} | <|file_sep|>src/tools/StackScrollMouseWheelTool.js.diff
original:
updated:
invert: false,
<|file_sep|>src/tools/StackScrollMouseWheelTool.js.diff
original:
const { loop, allowSkipping } = this.configuration;
updated:
const { loop, allowSkipping, invert } = this.configuration;
const direction = invert ? -images : images;
<|file_sep|>original/src/tools/StackScrollMouseWheelTool.js
name: 'StackScrollMouseWheel',
supportedInteractionTypes: ['MouseWheel'],
configuration: {
loop: false,
allowSkipping: true,
},
};
const initialConfiguration = Object.assign(defaultConfig, configuration);
super(initialConfiguration);
this.initialConfiguration = initialConfiguration;
}
mouseWheelCallback(evt) {
const { direction: images, element } = evt.detail;
const { loop, allowSkipping } = this.configuration;
scroll(element, -images, loop, allowSkipping);
}
}
<|file_sep|>current/src/tools/StackScrollMouseWheelTool.js
configuration: {
loop: false,
allowSkipping: true,
invert: false,
},
};
const initialConfiguration = Object.assign(defaultConfig, configuration);
super(initialConfiguration);
this.initialConfiguration = initialConfiguration;
}
mouseWheelCallback(evt) {
const { direction: images, element } = evt.detail;
const { loop, allowSkipping, invert } = this.configuration;
const direction = invert ? -images : images;
scroll(element, -images, loop, allowSkipping);
}
}
<|file_sep|>updated/src/tools/StackScrollMouseWheelTool.js
configuration: {
loop: false,
allowSkipping: true,
invert: false,
},
};
const initialConfiguration = Object.assign(defaultConfig, configuration);
super(initialConfiguration);
this.initialConfiguration = initialConfiguration;
}
mouseWheelCallback(evt) {
const { direction: images, element } = evt.detail;
const { loop, allowSkipping, invert } = this.configuration;
const direction = invert ? -images : images;
scroll(element, direction, loop, allowSkipping);
}
} | 6b4d4f20be70cfb905d84d5af846714461f34736 | src/tools/StackScrollMouseWheelTool.js | src/tools/StackScrollMouseWheelTool.js | JavaScript |
<|file_sep|>original/test/test_helper.rb
$:.unshift lib_dir unless $:.include?(lib_dir)
require 'haml'
require 'sass'
Sass::RAILS_LOADED = true unless defined?(Sass::RAILS_LOADED)
class Test::Unit::TestCase
def munge_filename(opts)
return if opts[:filename]
test_name = Haml::Util.caller_info(caller[1])[2]
opts[:filename] = "#{test_name}_inline.sass"
end
def clean_up_sassc
path = File.dirname(__FILE__) + "/../.sass-cache"
FileUtils.rm_r(path) if File.exist?(path)
end
def assert_warning(message)
the_real_stderr, $stderr = $stderr, StringIO.new
yield
<|file_sep|>current/test/test_helper.rb
$:.unshift lib_dir unless $:.include?(lib_dir)
require 'haml'
require 'sass'
Sass::RAILS_LOADED = true unless defined?(Sass::RAILS_LOADED)
class Test::Unit::TestCase
def munge_filename(opts)
return if opts[:filename]
test_name = Haml::Util.caller_info(caller[1])[2]
opts[:filename] = "#{test_name}_inline.sass"
end
def clean_up_sassc
path = File.dirname(__FILE__) + "/../.sass-cache"
FileUtils.rm_r(path) if File.exist?(path)
end
def assert_warning(message)
the_real_stderr, $stderr = $stderr, StringIO.new
yield
<|file_sep|>updated/test/test_helper.rb | $:.unshift lib_dir unless $:.include?(lib_dir)
require 'haml'
require 'sass'
Sass::RAILS_LOADED = true unless defined?(Sass::RAILS_LOADED)
class Test::Unit::TestCase
def munge_filename(opts)
return if opts[:filename]
test_name = Haml::Util.caller_info(caller[1])[2]
test_name.sub!(/^block in /, '')
opts[:filename] = "#{test_name}_inline.sass"
end
def clean_up_sassc
path = File.dirname(__FILE__) + "/../.sass-cache"
FileUtils.rm_r(path) if File.exist?(path)
end
def assert_warning(message)
the_real_stderr, $stderr = $stderr, StringIO.new | <|file_sep|>original/test/test_helper.rb
$:.unshift lib_dir unless $:.include?(lib_dir)
require 'haml'
require 'sass'
Sass::RAILS_LOADED = true unless defined?(Sass::RAILS_LOADED)
class Test::Unit::TestCase
def munge_filename(opts)
return if opts[:filename]
test_name = Haml::Util.caller_info(caller[1])[2]
opts[:filename] = "#{test_name}_inline.sass"
end
def clean_up_sassc
path = File.dirname(__FILE__) + "/../.sass-cache"
FileUtils.rm_r(path) if File.exist?(path)
end
def assert_warning(message)
the_real_stderr, $stderr = $stderr, StringIO.new
yield
<|file_sep|>current/test/test_helper.rb
$:.unshift lib_dir unless $:.include?(lib_dir)
require 'haml'
require 'sass'
Sass::RAILS_LOADED = true unless defined?(Sass::RAILS_LOADED)
class Test::Unit::TestCase
def munge_filename(opts)
return if opts[:filename]
test_name = Haml::Util.caller_info(caller[1])[2]
opts[:filename] = "#{test_name}_inline.sass"
end
def clean_up_sassc
path = File.dirname(__FILE__) + "/../.sass-cache"
FileUtils.rm_r(path) if File.exist?(path)
end
def assert_warning(message)
the_real_stderr, $stderr = $stderr, StringIO.new
yield
<|file_sep|>updated/test/test_helper.rb
$:.unshift lib_dir unless $:.include?(lib_dir)
require 'haml'
require 'sass'
Sass::RAILS_LOADED = true unless defined?(Sass::RAILS_LOADED)
class Test::Unit::TestCase
def munge_filename(opts)
return if opts[:filename]
test_name = Haml::Util.caller_info(caller[1])[2]
test_name.sub!(/^block in /, '')
opts[:filename] = "#{test_name}_inline.sass"
end
def clean_up_sassc
path = File.dirname(__FILE__) + "/../.sass-cache"
FileUtils.rm_r(path) if File.exist?(path)
end
def assert_warning(message)
the_real_stderr, $stderr = $stderr, StringIO.new | 2ebda4eba8927239823b9b1b1b01cca71912dacb | test/test_helper.rb | test/test_helper.rb | Ruby |
<|file_sep|>original/deploy_website.sh
# Copy website files from real repo
cp -R ../website/* .
# Download the latest javadoc
curl -L "http://repository.sonatype.org/service/local/artifact/maven/redirect?r=central-proxy&g=$GROUP_ID&a=$ARTIFACT_ID&v=LATEST&c=javadoc" > javadoc.zip
mkdir javadoc
unzip javadoc.zip -d javadoc
rm javadoc.zip
exit 0
# Stage all files in git and create a commit
git add .
git add -u
git commit -m "Website at $(date)"
# Push the new files up to GitHub
git push origin gh-pages
# Delete our temp folder
<|file_sep|>current/deploy_website.sh
# Copy website files from real repo
cp -R ../website/* .
# Download the latest javadoc
curl -L "http://repository.sonatype.org/service/local/artifact/maven/redirect?r=central-proxy&g=$GROUP_ID&a=$ARTIFACT_ID&v=LATEST&c=javadoc" > javadoc.zip
mkdir javadoc
unzip javadoc.zip -d javadoc
rm javadoc.zip
exit 0
# Stage all files in git and create a commit
git add .
git add -u
git commit -m "Website at $(date)"
# Push the new files up to GitHub
git push origin gh-pages
# Delete our temp folder
<|file_sep|>updated/deploy_website.sh |
# Copy website files from real repo
cp -R ../website/* .
# Download the latest javadoc
curl -L "http://repository.sonatype.org/service/local/artifact/maven/redirect?r=central-proxy&g=$GROUP_ID&a=$ARTIFACT_ID&v=LATEST&c=javadoc" > javadoc.zip
mkdir javadoc
unzip javadoc.zip -d javadoc
rm javadoc.zip
# Stage all files in git and create a commit
git add .
git add -u
git commit -m "Website at $(date)"
# Push the new files up to GitHub
git push origin gh-pages
# Delete our temp folder
cd ..
rm -rf $DIR | <|file_sep|>original/deploy_website.sh
# Copy website files from real repo
cp -R ../website/* .
# Download the latest javadoc
curl -L "http://repository.sonatype.org/service/local/artifact/maven/redirect?r=central-proxy&g=$GROUP_ID&a=$ARTIFACT_ID&v=LATEST&c=javadoc" > javadoc.zip
mkdir javadoc
unzip javadoc.zip -d javadoc
rm javadoc.zip
exit 0
# Stage all files in git and create a commit
git add .
git add -u
git commit -m "Website at $(date)"
# Push the new files up to GitHub
git push origin gh-pages
# Delete our temp folder
<|file_sep|>current/deploy_website.sh
# Copy website files from real repo
cp -R ../website/* .
# Download the latest javadoc
curl -L "http://repository.sonatype.org/service/local/artifact/maven/redirect?r=central-proxy&g=$GROUP_ID&a=$ARTIFACT_ID&v=LATEST&c=javadoc" > javadoc.zip
mkdir javadoc
unzip javadoc.zip -d javadoc
rm javadoc.zip
exit 0
# Stage all files in git and create a commit
git add .
git add -u
git commit -m "Website at $(date)"
# Push the new files up to GitHub
git push origin gh-pages
# Delete our temp folder
<|file_sep|>updated/deploy_website.sh
# Copy website files from real repo
cp -R ../website/* .
# Download the latest javadoc
curl -L "http://repository.sonatype.org/service/local/artifact/maven/redirect?r=central-proxy&g=$GROUP_ID&a=$ARTIFACT_ID&v=LATEST&c=javadoc" > javadoc.zip
mkdir javadoc
unzip javadoc.zip -d javadoc
rm javadoc.zip
# Stage all files in git and create a commit
git add .
git add -u
git commit -m "Website at $(date)"
# Push the new files up to GitHub
git push origin gh-pages
# Delete our temp folder
cd ..
rm -rf $DIR | ece261be29a18d979ccc3cafb4703e4505f8e66a | deploy_website.sh | deploy_website.sh | Shell |
<|file_sep|>original/package.json
"dependencies": {
"react": "^15.4.2",
"react-dom": "^15.4.2",
"react-sound": "^1.1.0"
},
"devDependencies": {
"enzyme": "^2.7.1",
"gh-pages": "^0.12.0",
"react-addons-test-utils": "^15.4.2",
"react-scripts": "0.9.5",
"react-test-utils": "0.0.1"
},
"scripts": {
"start": "react-scripts start",
"build": "react-scripts build",
"test": "react-scripts test --env=jsdom",
"eject": "react-scripts eject",
"deploy": "npm run build&&gh-pages -d build"
},
"homepage": "http://crhain.github.io/fcc-simon"
}
<|file_sep|>current/package.json
"dependencies": {
"react": "^15.4.2",
"react-dom": "^15.4.2",
"react-sound": "^1.1.0"
},
"devDependencies": {
"enzyme": "^2.7.1",
"gh-pages": "^0.12.0",
"react-addons-test-utils": "^15.4.2",
"react-scripts": "0.9.5",
"react-test-utils": "0.0.1"
},
"scripts": {
"start": "react-scripts start",
"build": "react-scripts build",
"test": "react-scripts test --env=jsdom",
"eject": "react-scripts eject",
"deploy": "npm run build&&gh-pages -d build"
},
"homepage": "http://crhain.github.io/fcc-simon"
}
<|file_sep|>updated/package.json | "dependencies": {
"react": "^15.4.2",
"react-dom": "^15.4.2",
"react-sound": "^1.1.0"
},
"devDependencies": {
"enzyme": "^2.7.1",
"gh-pages": "^0.12.0",
"react-addons-test-utils": "^15.4.2",
"react-scripts": "0.9.5",
"react-test-utils": "0.0.1"
},
"scripts": {
"start": "react-scripts start",
"build": "react-scripts build",
"test": "react-scripts test --env=jsdom",
"eject": "react-scripts eject",
"deploy": "npm run build&&gh-pages -d build"
},
"homepage": "http://crhwebdev.github.io/fcc-simon"
} | <|file_sep|>original/package.json
"dependencies": {
"react": "^15.4.2",
"react-dom": "^15.4.2",
"react-sound": "^1.1.0"
},
"devDependencies": {
"enzyme": "^2.7.1",
"gh-pages": "^0.12.0",
"react-addons-test-utils": "^15.4.2",
"react-scripts": "0.9.5",
"react-test-utils": "0.0.1"
},
"scripts": {
"start": "react-scripts start",
"build": "react-scripts build",
"test": "react-scripts test --env=jsdom",
"eject": "react-scripts eject",
"deploy": "npm run build&&gh-pages -d build"
},
"homepage": "http://crhain.github.io/fcc-simon"
}
<|file_sep|>current/package.json
"dependencies": {
"react": "^15.4.2",
"react-dom": "^15.4.2",
"react-sound": "^1.1.0"
},
"devDependencies": {
"enzyme": "^2.7.1",
"gh-pages": "^0.12.0",
"react-addons-test-utils": "^15.4.2",
"react-scripts": "0.9.5",
"react-test-utils": "0.0.1"
},
"scripts": {
"start": "react-scripts start",
"build": "react-scripts build",
"test": "react-scripts test --env=jsdom",
"eject": "react-scripts eject",
"deploy": "npm run build&&gh-pages -d build"
},
"homepage": "http://crhain.github.io/fcc-simon"
}
<|file_sep|>updated/package.json
"dependencies": {
"react": "^15.4.2",
"react-dom": "^15.4.2",
"react-sound": "^1.1.0"
},
"devDependencies": {
"enzyme": "^2.7.1",
"gh-pages": "^0.12.0",
"react-addons-test-utils": "^15.4.2",
"react-scripts": "0.9.5",
"react-test-utils": "0.0.1"
},
"scripts": {
"start": "react-scripts start",
"build": "react-scripts build",
"test": "react-scripts test --env=jsdom",
"eject": "react-scripts eject",
"deploy": "npm run build&&gh-pages -d build"
},
"homepage": "http://crhwebdev.github.io/fcc-simon"
} | 1deb098a931bc5faa3d9ea9cac1adf8fff121903 | package.json | package.json | JSON |
<|file_sep|>original/README.md
## Mermaid plugin for GitBook
[](https://travis-ci.org/JozoVilcek/gitbook-mermaid-plugin)
[](http://badge.fury.io/js/gitbook-mermaid)
Plugin for [GitBook](https://github.com/GitbookIO/gitbook) which renders [Mermaid](https://github.com/knsv/mermaid) diagrams and flow charts detected in the book markdown.
### How to install it?
You can use install via **NPM**:
```
$ npm install gitbook-mermaid-plugin
```
And use it for your book with in the book.json:
```
{
"plugins": ["gitbook-mermaid"]
}
```
<|file_sep|>current/README.md
## Mermaid plugin for GitBook
[](https://travis-ci.org/JozoVilcek/gitbook-mermaid-plugin)
[](http://badge.fury.io/js/gitbook-mermaid)
Plugin for [GitBook](https://github.com/GitbookIO/gitbook) which renders [Mermaid](https://github.com/knsv/mermaid) diagrams and flow charts detected in the book markdown.
### How to install it?
You can use install via **NPM**:
```
$ npm install gitbook-mermaid-plugin
```
And use it for your book with in the book.json:
```
{
"plugins": ["gitbook-mermaid"]
}
```
<|file_sep|>updated/README.md | ## Mermaid plugin for GitBook
[](https://travis-ci.org/JozoVilcek/gitbook-plugin-mermaid)
[](http://badge.fury.io/js/gitbook-plugin-mermaid)
Plugin for [GitBook](https://github.com/GitbookIO/gitbook) which renders [Mermaid](https://github.com/knsv/mermaid) diagrams and flow charts detected in the book markdown.
### How to install it?
You can use install via **NPM**:
```
$ npm install gitbook-mermaid-plugin
```
And use it for your book with in the book.json:
```
{
"plugins": ["gitbook-mermaid"]
}
``` | <|file_sep|>original/README.md
## Mermaid plugin for GitBook
[](https://travis-ci.org/JozoVilcek/gitbook-mermaid-plugin)
[](http://badge.fury.io/js/gitbook-mermaid)
Plugin for [GitBook](https://github.com/GitbookIO/gitbook) which renders [Mermaid](https://github.com/knsv/mermaid) diagrams and flow charts detected in the book markdown.
### How to install it?
You can use install via **NPM**:
```
$ npm install gitbook-mermaid-plugin
```
And use it for your book with in the book.json:
```
{
"plugins": ["gitbook-mermaid"]
}
```
<|file_sep|>current/README.md
## Mermaid plugin for GitBook
[](https://travis-ci.org/JozoVilcek/gitbook-mermaid-plugin)
[](http://badge.fury.io/js/gitbook-mermaid)
Plugin for [GitBook](https://github.com/GitbookIO/gitbook) which renders [Mermaid](https://github.com/knsv/mermaid) diagrams and flow charts detected in the book markdown.
### How to install it?
You can use install via **NPM**:
```
$ npm install gitbook-mermaid-plugin
```
And use it for your book with in the book.json:
```
{
"plugins": ["gitbook-mermaid"]
}
```
<|file_sep|>updated/README.md
## Mermaid plugin for GitBook
[](https://travis-ci.org/JozoVilcek/gitbook-plugin-mermaid)
[](http://badge.fury.io/js/gitbook-plugin-mermaid)
Plugin for [GitBook](https://github.com/GitbookIO/gitbook) which renders [Mermaid](https://github.com/knsv/mermaid) diagrams and flow charts detected in the book markdown.
### How to install it?
You can use install via **NPM**:
```
$ npm install gitbook-mermaid-plugin
```
And use it for your book with in the book.json:
```
{
"plugins": ["gitbook-mermaid"]
}
``` | 8224190364d1c82382239c0d243a2c0902ad1cf8 | README.md | README.md | Markdown |
<|file_sep|>original/.travis.yml
language: haskell
before_install:
- sudo add-apt-repository -y ppa:h-rayflood/llvm
- sudo apt-get update -qq
- sudo apt-get -f install llvm-3.3 llvm-3.3-dev libclang-common-dev libclang-dev
notifications:
email: false
<|file_sep|>current/.travis.yml
language: haskell
before_install:
- sudo add-apt-repository -y ppa:h-rayflood/llvm
- sudo apt-get update -qq
- sudo apt-get -f install llvm-3.3 llvm-3.3-dev libclang-common-dev libclang-dev
notifications:
email: false
<|file_sep|>updated/.travis.yml | language: haskell
before_install:
- wget -q -O - http://llvm.org/apt/llvm-snapshot.gpg.key | sudo apt-key add -
- sudo apt-add-repository -y 'deb http://llvm.org/apt/precise/ llvm-toolchain-precise main'
- sudo apt-get update -qq
- sudo apt-get -f install llvm-3.3 llvm-3.3-dev libclang-common-dev libclang-dev
notifications:
email: false | <|file_sep|>original/.travis.yml
language: haskell
before_install:
- sudo add-apt-repository -y ppa:h-rayflood/llvm
- sudo apt-get update -qq
- sudo apt-get -f install llvm-3.3 llvm-3.3-dev libclang-common-dev libclang-dev
notifications:
email: false
<|file_sep|>current/.travis.yml
language: haskell
before_install:
- sudo add-apt-repository -y ppa:h-rayflood/llvm
- sudo apt-get update -qq
- sudo apt-get -f install llvm-3.3 llvm-3.3-dev libclang-common-dev libclang-dev
notifications:
email: false
<|file_sep|>updated/.travis.yml
language: haskell
before_install:
- wget -q -O - http://llvm.org/apt/llvm-snapshot.gpg.key | sudo apt-key add -
- sudo apt-add-repository -y 'deb http://llvm.org/apt/precise/ llvm-toolchain-precise main'
- sudo apt-get update -qq
- sudo apt-get -f install llvm-3.3 llvm-3.3-dev libclang-common-dev libclang-dev
notifications:
email: false | 6b5b16fffa04a62593b6cc508598bec8e185f63c | .travis.yml | .travis.yml | YAML |
<|file_sep|>package.json.diff
original:
"homepage": "https://github.com/shama/isndarray",
updated:
"homepage": "https://github.com/scijs/isndarray",
<|file_sep|>package.json.diff
original:
"url": "git://github.com/shama/isndarray.git"
updated:
"url": "git://github.com/scijs/isndarray.git"
<|file_sep|>package.json.diff
original:
"url": "https://github.com/shama/isndarray/issues"
updated:
"url": "https://github.com/scijs/isndarray/issues"
<|file_sep|>original/package.json
},
"repository": {
"type": "git",
"url": "git://github.com/shama/isndarray.git"
},
"bugs": {
"url": "https://github.com/shama/isndarray/issues"
},
"licenses": [
{
"type": "MIT",
"url": "https://github.com/shama/isndarray/blob/master/LICENSE-MIT"
}
],
"engines": {
"node": ">= 0.8.0"
},
"scripts": {
"test": "node test.js"
},
"dependencies": {},
<|file_sep|>current/package.json
},
"repository": {
"type": "git",
"url": "git://github.com/scijs/isndarray.git"
},
"bugs": {
"url": "https://github.com/scijs/isndarray/issues"
},
"licenses": [
{
"type": "MIT",
"url": "https://github.com/shama/isndarray/blob/master/LICENSE-MIT"
}
],
"engines": {
"node": ">= 0.8.0"
},
"scripts": {
"test": "node test.js"
},
"dependencies": {},
<|file_sep|>updated/package.json | },
"repository": {
"type": "git",
"url": "git://github.com/scijs/isndarray.git"
},
"bugs": {
"url": "https://github.com/scijs/isndarray/issues"
},
"license": "MIT",
"engines": {
"node": ">= 0.8.0"
},
"scripts": {
"test": "node test.js"
},
"dependencies": {},
"devDependencies": {
"tape": "~1.0.4",
"ndarray": "~1.0.3"
},
"keywords": [ | <|file_sep|>package.json.diff
original:
"homepage": "https://github.com/shama/isndarray",
updated:
"homepage": "https://github.com/scijs/isndarray",
<|file_sep|>package.json.diff
original:
"url": "git://github.com/shama/isndarray.git"
updated:
"url": "git://github.com/scijs/isndarray.git"
<|file_sep|>package.json.diff
original:
"url": "https://github.com/shama/isndarray/issues"
updated:
"url": "https://github.com/scijs/isndarray/issues"
<|file_sep|>original/package.json
},
"repository": {
"type": "git",
"url": "git://github.com/shama/isndarray.git"
},
"bugs": {
"url": "https://github.com/shama/isndarray/issues"
},
"licenses": [
{
"type": "MIT",
"url": "https://github.com/shama/isndarray/blob/master/LICENSE-MIT"
}
],
"engines": {
"node": ">= 0.8.0"
},
"scripts": {
"test": "node test.js"
},
"dependencies": {},
<|file_sep|>current/package.json
},
"repository": {
"type": "git",
"url": "git://github.com/scijs/isndarray.git"
},
"bugs": {
"url": "https://github.com/scijs/isndarray/issues"
},
"licenses": [
{
"type": "MIT",
"url": "https://github.com/shama/isndarray/blob/master/LICENSE-MIT"
}
],
"engines": {
"node": ">= 0.8.0"
},
"scripts": {
"test": "node test.js"
},
"dependencies": {},
<|file_sep|>updated/package.json
},
"repository": {
"type": "git",
"url": "git://github.com/scijs/isndarray.git"
},
"bugs": {
"url": "https://github.com/scijs/isndarray/issues"
},
"license": "MIT",
"engines": {
"node": ">= 0.8.0"
},
"scripts": {
"test": "node test.js"
},
"dependencies": {},
"devDependencies": {
"tape": "~1.0.4",
"ndarray": "~1.0.3"
},
"keywords": [ | 9792af5924be2b371d5225dd5c82852d6db3d39e | package.json | package.json | JSON |
<|file_sep|>original/recipes/default.rb
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
include_recipe 'build-essential'
case node['platform_family']
when 'debian'
package 'zlib1g-dev' do
action :install
end.run_action(:install)
end
chef_gem 'fog' do
version node['dnsimple']['fog_version']
action :install
end
require 'fog'
<|file_sep|>current/recipes/default.rb
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
include_recipe 'build-essential'
case node['platform_family']
when 'debian'
package 'zlib1g-dev' do
action :install
end.run_action(:install)
end
chef_gem 'fog' do
version node['dnsimple']['fog_version']
action :install
end
require 'fog'
<|file_sep|>updated/recipes/default.rb | # See the License for the specific language governing permissions and
# limitations under the License.
#
include_recipe 'build-essential'
case node['platform_family']
when 'debian'
package 'zlib1g-dev' do
action :install
end.run_action(:install)
end
chef_gem 'fog' do
version node['dnsimple']['fog_version']
compile_time false if Chef::Resource::ChefGem.method_defined?(:compile_time)
action :install
end
require 'fog' | <|file_sep|>original/recipes/default.rb
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
include_recipe 'build-essential'
case node['platform_family']
when 'debian'
package 'zlib1g-dev' do
action :install
end.run_action(:install)
end
chef_gem 'fog' do
version node['dnsimple']['fog_version']
action :install
end
require 'fog'
<|file_sep|>current/recipes/default.rb
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
include_recipe 'build-essential'
case node['platform_family']
when 'debian'
package 'zlib1g-dev' do
action :install
end.run_action(:install)
end
chef_gem 'fog' do
version node['dnsimple']['fog_version']
action :install
end
require 'fog'
<|file_sep|>updated/recipes/default.rb
# See the License for the specific language governing permissions and
# limitations under the License.
#
include_recipe 'build-essential'
case node['platform_family']
when 'debian'
package 'zlib1g-dev' do
action :install
end.run_action(:install)
end
chef_gem 'fog' do
version node['dnsimple']['fog_version']
compile_time false if Chef::Resource::ChefGem.method_defined?(:compile_time)
action :install
end
require 'fog' | 06c69c37e470158236ff23bc6f9a420ef6cd0654 | recipes/default.rb | recipes/default.rb | Ruby |
<|file_sep|>original/static/domotica.js
function lightswitch()
{
id = this.id.replace("light_", "");
$.post("/lightswitch", { id: id });
};
$(document).ready(function () {
$(".toggle").each(function(index) {
obj = $(".toggle")[index];
document.querySelector('#' + obj.id).addEventListener('toggle', lightswitch, obj.id);
});
});
<|file_sep|>current/static/domotica.js
function lightswitch()
{
id = this.id.replace("light_", "");
$.post("/lightswitch", { id: id });
};
$(document).ready(function () {
$(".toggle").each(function(index) {
obj = $(".toggle")[index];
document.querySelector('#' + obj.id).addEventListener('toggle', lightswitch, obj.id);
});
});
<|file_sep|>updated/static/domotica.js | function lightswitch()
{
id = this.id.replace("light_", "");
$.post("/lightswitch", { id: id })
.fail(location.reload());
};
$(document).ready(function () {
$(".toggle").each(function(index) {
obj = $(".toggle")[index];
document.querySelector('#' + obj.id).addEventListener('toggle', lightswitch, obj.id);
});
}); | <|file_sep|>original/static/domotica.js
function lightswitch()
{
id = this.id.replace("light_", "");
$.post("/lightswitch", { id: id });
};
$(document).ready(function () {
$(".toggle").each(function(index) {
obj = $(".toggle")[index];
document.querySelector('#' + obj.id).addEventListener('toggle', lightswitch, obj.id);
});
});
<|file_sep|>current/static/domotica.js
function lightswitch()
{
id = this.id.replace("light_", "");
$.post("/lightswitch", { id: id });
};
$(document).ready(function () {
$(".toggle").each(function(index) {
obj = $(".toggle")[index];
document.querySelector('#' + obj.id).addEventListener('toggle', lightswitch, obj.id);
});
});
<|file_sep|>updated/static/domotica.js
function lightswitch()
{
id = this.id.replace("light_", "");
$.post("/lightswitch", { id: id })
.fail(location.reload());
};
$(document).ready(function () {
$(".toggle").each(function(index) {
obj = $(".toggle")[index];
document.querySelector('#' + obj.id).addEventListener('toggle', lightswitch, obj.id);
});
}); | 1bce8ff8d0914c861e7b0e3d2db1b94a38b6ee06 | static/domotica.js | static/domotica.js | JavaScript |
<|file_sep|>original/ninjapack/root/opt/ninjablocks/factory-reset/bin/check-nand.sh
<|file_sep|>current/ninjapack/root/opt/ninjablocks/factory-reset/bin/check-nand.sh
<|file_sep|>updated/ninjapack/root/opt/ninjablocks/factory-reset/bin/check-nand.sh | #!/usr/bin/env bash
#
# Checks that the first X bytes of /dev/mtdN have an md5sum that matches md5sum /etc/firmware-versions/mtdN-X.md5
#
# Excludes mtd6, mtd7 and mtd9 since these partitions are known to be different from the checksum.
#
# Used during factory flashing process to detect bad blocks outside of the UBI managed partition (mtd9).
#
die() {
echo "$*" 1>&2
exit 1
}
md5sum() {
openssl md5 | cut -f2 -d' '
}
main() {
rc=0
find /etc/firmware-versions -maxdepth 1 -mindepth 1 -type f -name 'mtd*.md5' | egrep -v "mtd9|mtd6|mtd7" | while read f; do | <|file_sep|>original/ninjapack/root/opt/ninjablocks/factory-reset/bin/check-nand.sh
<|file_sep|>current/ninjapack/root/opt/ninjablocks/factory-reset/bin/check-nand.sh
<|file_sep|>updated/ninjapack/root/opt/ninjablocks/factory-reset/bin/check-nand.sh
#!/usr/bin/env bash
#
# Checks that the first X bytes of /dev/mtdN have an md5sum that matches md5sum /etc/firmware-versions/mtdN-X.md5
#
# Excludes mtd6, mtd7 and mtd9 since these partitions are known to be different from the checksum.
#
# Used during factory flashing process to detect bad blocks outside of the UBI managed partition (mtd9).
#
die() {
echo "$*" 1>&2
exit 1
}
md5sum() {
openssl md5 | cut -f2 -d' '
}
main() {
rc=0
find /etc/firmware-versions -maxdepth 1 -mindepth 1 -type f -name 'mtd*.md5' | egrep -v "mtd9|mtd6|mtd7" | while read f; do | 5ab688e26821277d9f46b4955bf5b535359713d6 | ninjapack/root/opt/ninjablocks/factory-reset/bin/check-nand.sh | ninjapack/root/opt/ninjablocks/factory-reset/bin/check-nand.sh | Shell |
<|file_sep|>original/README.md
# negroni-cache
[](https://travis-ci.org/trumanw/negroni-cache)
A standard compatible([RFC 7234](http://www.rfc-base.org/rfc-7234.html)) HTTP Cache middleware for negroni.
# Usage
Take a peek at the basic example or the custom example, the latter of which includes setting a custom Before and After function on the cache middleware.
<|file_sep|>current/README.md
# negroni-cache
[](https://travis-ci.org/trumanw/negroni-cache)
A standard compatible([RFC 7234](http://www.rfc-base.org/rfc-7234.html)) HTTP Cache middleware for negroni.
# Usage
Take a peek at the basic example or the custom example, the latter of which includes setting a custom Before and After function on the cache middleware.
<|file_sep|>updated/README.md | # negroni-cache
[](https://travis-ci.org/trumanw/negroni-cache)
A standard compatible([RFC 7234](http://www.rfc-base.org/rfc-7234.html)) HTTP Cache middleware for negroni.
# Usage
Take a peek at the basic example or the custom example, the latter of which includes setting a custom Before and After function on the cache middleware. | <|file_sep|>original/README.md
# negroni-cache
[](https://travis-ci.org/trumanw/negroni-cache)
A standard compatible([RFC 7234](http://www.rfc-base.org/rfc-7234.html)) HTTP Cache middleware for negroni.
# Usage
Take a peek at the basic example or the custom example, the latter of which includes setting a custom Before and After function on the cache middleware.
<|file_sep|>current/README.md
# negroni-cache
[](https://travis-ci.org/trumanw/negroni-cache)
A standard compatible([RFC 7234](http://www.rfc-base.org/rfc-7234.html)) HTTP Cache middleware for negroni.
# Usage
Take a peek at the basic example or the custom example, the latter of which includes setting a custom Before and After function on the cache middleware.
<|file_sep|>updated/README.md
# negroni-cache
[](https://travis-ci.org/trumanw/negroni-cache)
A standard compatible([RFC 7234](http://www.rfc-base.org/rfc-7234.html)) HTTP Cache middleware for negroni.
# Usage
Take a peek at the basic example or the custom example, the latter of which includes setting a custom Before and After function on the cache middleware. | 8299e33fa274fa91e22112a00a69d7222d14b9c2 | README.md | README.md | Markdown |
<|file_sep|>.travis.yml.diff
original:
updated:
- blead-thr
<|file_sep|>.travis.yml.diff
original:
- '5.22'
- '5.20'
- '5.18'
- '5.16'
- '5.14'
- '5.12'
- '5.10'
- '5.8'
updated:
- dev-thr
- 5.22.0
- 5.22.0-thr
- 5.20.2
- 5.20.2-thr
- 5.18.3
- 5.18.3-thr
- 5.16.3
- 5.16.3-thr
- 5.14.4
- 5.14.4-thr
- 5.12.5
- 5.12.5-thr
- 5.10.1
- 5.10.1-thr
- 5.8.8
- 5.8.8-thr
<|file_sep|>original/.travis.yml
- dev
- '5.22'
- '5.20'
- '5.18'
- '5.16'
- '5.14'
- '5.12'
- '5.10'
- '5.8'
matrix:
allow_failures:
- perl: blead
include:
- env: COVERAGE=1
perl: '5.22'
env:
global:
- RELEASE_TESTING=1
- AUTHOR_TESTING=1
before_install:
- eval $(curl https://travis-perl.github.io/init) --auto
<|file_sep|>current/.travis.yml
- 5.16.3-thr
- 5.14.4
- 5.14.4-thr
- 5.12.5
- 5.12.5-thr
- 5.10.1
- 5.10.1-thr
- 5.8.8
- 5.8.8-thr
matrix:
allow_failures:
- perl: blead
include:
- env: COVERAGE=1
perl: '5.22'
env:
global:
- RELEASE_TESTING=1
- AUTHOR_TESTING=1
before_install:
- eval $(curl https://travis-perl.github.io/init) --auto
<|file_sep|>updated/.travis.yml | - 5.14.4
- 5.14.4-thr
- 5.12.5
- 5.12.5-thr
- 5.10.1
- 5.10.1-thr
- 5.8.8
- 5.8.8-thr
matrix:
allow_failures:
- perl: blead
- perl: blead-thr
include:
- env: COVERAGE=1
perl: '5.22'
env:
global:
- RELEASE_TESTING=1
- AUTHOR_TESTING=1
before_install:
- eval $(curl https://travis-perl.github.io/init) --auto | <|file_sep|>.travis.yml.diff
original:
updated:
- blead-thr
<|file_sep|>.travis.yml.diff
original:
- '5.22'
- '5.20'
- '5.18'
- '5.16'
- '5.14'
- '5.12'
- '5.10'
- '5.8'
updated:
- dev-thr
- 5.22.0
- 5.22.0-thr
- 5.20.2
- 5.20.2-thr
- 5.18.3
- 5.18.3-thr
- 5.16.3
- 5.16.3-thr
- 5.14.4
- 5.14.4-thr
- 5.12.5
- 5.12.5-thr
- 5.10.1
- 5.10.1-thr
- 5.8.8
- 5.8.8-thr
<|file_sep|>original/.travis.yml
- dev
- '5.22'
- '5.20'
- '5.18'
- '5.16'
- '5.14'
- '5.12'
- '5.10'
- '5.8'
matrix:
allow_failures:
- perl: blead
include:
- env: COVERAGE=1
perl: '5.22'
env:
global:
- RELEASE_TESTING=1
- AUTHOR_TESTING=1
before_install:
- eval $(curl https://travis-perl.github.io/init) --auto
<|file_sep|>current/.travis.yml
- 5.16.3-thr
- 5.14.4
- 5.14.4-thr
- 5.12.5
- 5.12.5-thr
- 5.10.1
- 5.10.1-thr
- 5.8.8
- 5.8.8-thr
matrix:
allow_failures:
- perl: blead
include:
- env: COVERAGE=1
perl: '5.22'
env:
global:
- RELEASE_TESTING=1
- AUTHOR_TESTING=1
before_install:
- eval $(curl https://travis-perl.github.io/init) --auto
<|file_sep|>updated/.travis.yml
- 5.14.4
- 5.14.4-thr
- 5.12.5
- 5.12.5-thr
- 5.10.1
- 5.10.1-thr
- 5.8.8
- 5.8.8-thr
matrix:
allow_failures:
- perl: blead
- perl: blead-thr
include:
- env: COVERAGE=1
perl: '5.22'
env:
global:
- RELEASE_TESTING=1
- AUTHOR_TESTING=1
before_install:
- eval $(curl https://travis-perl.github.io/init) --auto | 1439df0e0ad416ab5368ba5b7f77b7405645b74f | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/docs/deploy.md
# Building and Deploying the cAdvisor Docker Container
## Building
Building the cAdvisor Docker container is simple, just run:
```
$ ./deploy/build.sh
```
Which will statically build the cAdvisor binary and then build the Docker image. The resulting Docker image will be called `google/cadvisor:beta`. This image is very bare, containing the cAdvisor binary and nothing else.
## Deploying
All cAdvisor releases are tagged and correspond to a Docker image. The latest supported release uses the `latest` tag. We have a `beta` and `canary` tag for pre-release versions with newer features. You can see more details about this in the cAdvisor Docker [registry](https://registry.hub.docker.com/u/google/cadvisor/) page.
<|file_sep|>current/docs/deploy.md
# Building and Deploying the cAdvisor Docker Container
## Building
Building the cAdvisor Docker container is simple, just run:
```
$ ./deploy/build.sh
```
Which will statically build the cAdvisor binary and then build the Docker image. The resulting Docker image will be called `google/cadvisor:beta`. This image is very bare, containing the cAdvisor binary and nothing else.
## Deploying
All cAdvisor releases are tagged and correspond to a Docker image. The latest supported release uses the `latest` tag. We have a `beta` and `canary` tag for pre-release versions with newer features. You can see more details about this in the cAdvisor Docker [registry](https://registry.hub.docker.com/u/google/cadvisor/) page.
<|file_sep|>updated/docs/deploy.md | # Building and Deploying the cAdvisor Docker Container
## Building
Building the cAdvisor Docker container is simple, just run:
```
$ ./deploy/build.sh
```
Which will statically build the cAdvisor binary and then build the Docker image. The resulting Docker image will be called `google/cadvisor:beta`. This image is very bare, containing the cAdvisor binary and nothing else.
## Deploying
All cAdvisor releases are tagged and correspond to a Docker image. The latest supported release uses the `latest` tag. We have a `beta` and `canary` tag for pre-release versions with newer features. You can see more details about this in the cAdvisor [Google Container Registry](https://gcr.io/cadvisor/cadvisor) page. | <|file_sep|>original/docs/deploy.md
# Building and Deploying the cAdvisor Docker Container
## Building
Building the cAdvisor Docker container is simple, just run:
```
$ ./deploy/build.sh
```
Which will statically build the cAdvisor binary and then build the Docker image. The resulting Docker image will be called `google/cadvisor:beta`. This image is very bare, containing the cAdvisor binary and nothing else.
## Deploying
All cAdvisor releases are tagged and correspond to a Docker image. The latest supported release uses the `latest` tag. We have a `beta` and `canary` tag for pre-release versions with newer features. You can see more details about this in the cAdvisor Docker [registry](https://registry.hub.docker.com/u/google/cadvisor/) page.
<|file_sep|>current/docs/deploy.md
# Building and Deploying the cAdvisor Docker Container
## Building
Building the cAdvisor Docker container is simple, just run:
```
$ ./deploy/build.sh
```
Which will statically build the cAdvisor binary and then build the Docker image. The resulting Docker image will be called `google/cadvisor:beta`. This image is very bare, containing the cAdvisor binary and nothing else.
## Deploying
All cAdvisor releases are tagged and correspond to a Docker image. The latest supported release uses the `latest` tag. We have a `beta` and `canary` tag for pre-release versions with newer features. You can see more details about this in the cAdvisor Docker [registry](https://registry.hub.docker.com/u/google/cadvisor/) page.
<|file_sep|>updated/docs/deploy.md
# Building and Deploying the cAdvisor Docker Container
## Building
Building the cAdvisor Docker container is simple, just run:
```
$ ./deploy/build.sh
```
Which will statically build the cAdvisor binary and then build the Docker image. The resulting Docker image will be called `google/cadvisor:beta`. This image is very bare, containing the cAdvisor binary and nothing else.
## Deploying
All cAdvisor releases are tagged and correspond to a Docker image. The latest supported release uses the `latest` tag. We have a `beta` and `canary` tag for pre-release versions with newer features. You can see more details about this in the cAdvisor [Google Container Registry](https://gcr.io/cadvisor/cadvisor) page. | f05750a6e386203726be3bfffd895bea703a6acf | docs/deploy.md | docs/deploy.md | Markdown |
<|file_sep|>original/docs/using.rst
For their main and mobile website:
* http://www.croisedanslemetro.com
* http://m.croisedanslemetro.com
The Molly Project
-----------------
Molly is a framework for the rapid development of information and service
portals targeted at mobile internet devices: http://mollyproject.org
It powers the University of Oxford's mobile portal: http://m.ox.ac.uk/
Mozilla
-------
Powering mozilla.org:
* https://www.mozilla.org
* https://github.com/mozilla/bedrock
<|file_sep|>current/docs/using.rst
For their main and mobile website:
* http://www.croisedanslemetro.com
* http://m.croisedanslemetro.com
The Molly Project
-----------------
Molly is a framework for the rapid development of information and service
portals targeted at mobile internet devices: http://mollyproject.org
It powers the University of Oxford's mobile portal: http://m.ox.ac.uk/
Mozilla
-------
Powering mozilla.org:
* https://www.mozilla.org
* https://github.com/mozilla/bedrock
<|file_sep|>updated/docs/using.rst |
For their main and mobile website:
* http://www.croisedanslemetro.com
* http://m.croisedanslemetro.com
The Molly Project
-----------------
Molly is a framework for the rapid development of information and service
portals targeted at mobile internet devices: https://github.com/mollyproject/mollyproject
It powers the University of Oxford's mobile portal: http://m.ox.ac.uk/
Mozilla
-------
Powering mozilla.org:
* https://www.mozilla.org
* https://github.com/mozilla/bedrock | <|file_sep|>original/docs/using.rst
For their main and mobile website:
* http://www.croisedanslemetro.com
* http://m.croisedanslemetro.com
The Molly Project
-----------------
Molly is a framework for the rapid development of information and service
portals targeted at mobile internet devices: http://mollyproject.org
It powers the University of Oxford's mobile portal: http://m.ox.ac.uk/
Mozilla
-------
Powering mozilla.org:
* https://www.mozilla.org
* https://github.com/mozilla/bedrock
<|file_sep|>current/docs/using.rst
For their main and mobile website:
* http://www.croisedanslemetro.com
* http://m.croisedanslemetro.com
The Molly Project
-----------------
Molly is a framework for the rapid development of information and service
portals targeted at mobile internet devices: http://mollyproject.org
It powers the University of Oxford's mobile portal: http://m.ox.ac.uk/
Mozilla
-------
Powering mozilla.org:
* https://www.mozilla.org
* https://github.com/mozilla/bedrock
<|file_sep|>updated/docs/using.rst
For their main and mobile website:
* http://www.croisedanslemetro.com
* http://m.croisedanslemetro.com
The Molly Project
-----------------
Molly is a framework for the rapid development of information and service
portals targeted at mobile internet devices: https://github.com/mollyproject/mollyproject
It powers the University of Oxford's mobile portal: http://m.ox.ac.uk/
Mozilla
-------
Powering mozilla.org:
* https://www.mozilla.org
* https://github.com/mozilla/bedrock | f9852471e64c5c99f7e6ab225183ac59216b25b3 | docs/using.rst | docs/using.rst | reStructuredText |
<|file_sep|>original/test/Transforms/InstCombine/cast.ll
; Tests to make sure elimination of casts is working correctly
; RUN: if as < %s | opt -instcombine | grep '%c' | grep cast
; RUN: then exit 1
; RUN: else exit 0
; RUN: fi
implementation
int %test1(int %A) {
%c1 = cast int %A to uint
%c2 = cast uint %c1 to int
ret int %c2
}
ulong %test2(ubyte %A) {
%c1 = cast ubyte %A to ushort
%c2 = cast ushort %c1 to uint
%Ret = cast uint %c2 to ulong
ret ulong %Ret
}
<|file_sep|>current/test/Transforms/InstCombine/cast.ll
; Tests to make sure elimination of casts is working correctly
; RUN: if as < %s | opt -instcombine | grep '%c' | grep cast
; RUN: then exit 1
; RUN: else exit 0
; RUN: fi
implementation
int %test1(int %A) {
%c1 = cast int %A to uint
%c2 = cast uint %c1 to int
ret int %c2
}
ulong %test2(ubyte %A) {
%c1 = cast ubyte %A to ushort
%c2 = cast ushort %c1 to uint
%Ret = cast uint %c2 to ulong
ret ulong %Ret
}
<|file_sep|>updated/test/Transforms/InstCombine/cast.ll | ; Tests to make sure elimination of casts is working correctly
; RUN: if as < %s | opt -instcombine -die | dis | grep '%c' | grep cast
; RUN: then exit 1
; RUN: else exit 0
; RUN: fi
implementation
int %test1(int %A) {
%c1 = cast int %A to uint
%c2 = cast uint %c1 to int
ret int %c2
}
ulong %test2(ubyte %A) {
%c1 = cast ubyte %A to ushort
%c2 = cast ushort %c1 to uint
%Ret = cast uint %c2 to ulong
ret ulong %Ret
} | <|file_sep|>original/test/Transforms/InstCombine/cast.ll
; Tests to make sure elimination of casts is working correctly
; RUN: if as < %s | opt -instcombine | grep '%c' | grep cast
; RUN: then exit 1
; RUN: else exit 0
; RUN: fi
implementation
int %test1(int %A) {
%c1 = cast int %A to uint
%c2 = cast uint %c1 to int
ret int %c2
}
ulong %test2(ubyte %A) {
%c1 = cast ubyte %A to ushort
%c2 = cast ushort %c1 to uint
%Ret = cast uint %c2 to ulong
ret ulong %Ret
}
<|file_sep|>current/test/Transforms/InstCombine/cast.ll
; Tests to make sure elimination of casts is working correctly
; RUN: if as < %s | opt -instcombine | grep '%c' | grep cast
; RUN: then exit 1
; RUN: else exit 0
; RUN: fi
implementation
int %test1(int %A) {
%c1 = cast int %A to uint
%c2 = cast uint %c1 to int
ret int %c2
}
ulong %test2(ubyte %A) {
%c1 = cast ubyte %A to ushort
%c2 = cast ushort %c1 to uint
%Ret = cast uint %c2 to ulong
ret ulong %Ret
}
<|file_sep|>updated/test/Transforms/InstCombine/cast.ll
; Tests to make sure elimination of casts is working correctly
; RUN: if as < %s | opt -instcombine -die | dis | grep '%c' | grep cast
; RUN: then exit 1
; RUN: else exit 0
; RUN: fi
implementation
int %test1(int %A) {
%c1 = cast int %A to uint
%c2 = cast uint %c1 to int
ret int %c2
}
ulong %test2(ubyte %A) {
%c1 = cast ubyte %A to ushort
%c2 = cast ushort %c1 to uint
%Ret = cast uint %c2 to ulong
ret ulong %Ret
} | 8360d0922ab683a3b57ecefe1890e75150b95294 | test/Transforms/InstCombine/cast.ll | test/Transforms/InstCombine/cast.ll | LLVM |
<|file_sep|>original/.travis.yml
language: node_js
node_js:
- "7"
os: osx
script:
- yarn run publish
cache:
yarn: true
<|file_sep|>current/.travis.yml
language: node_js
node_js:
- "7"
os: osx
script:
- yarn run publish
cache:
yarn: true
<|file_sep|>updated/.travis.yml | language: node_js
node_js:
- "10"
os: osx
script:
- yarn run publish
cache:
yarn: true | <|file_sep|>original/.travis.yml
language: node_js
node_js:
- "7"
os: osx
script:
- yarn run publish
cache:
yarn: true
<|file_sep|>current/.travis.yml
language: node_js
node_js:
- "7"
os: osx
script:
- yarn run publish
cache:
yarn: true
<|file_sep|>updated/.travis.yml
language: node_js
node_js:
- "10"
os: osx
script:
- yarn run publish
cache:
yarn: true | 2dfd51c442f842171ff198a3293e7bf44087d0a2 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/app/templates/components/show-more.html
{% macro show_more(url=None, label=None, with_border=True) %}
{% if url and label %}
<a
href="{{ url }}"
class="show-more{% if not with_border %}-no-border{% endif %}"
><span>{{ label }}</span></a>
{% else %}
<span class="show-more-empty"></span>
{% endif %}
{% endmacro %}
<|file_sep|>current/app/templates/components/show-more.html
{% macro show_more(url=None, label=None, with_border=True) %}
{% if url and label %}
<a
href="{{ url }}"
class="show-more{% if not with_border %}-no-border{% endif %}"
><span>{{ label }}</span></a>
{% else %}
<span class="show-more-empty"></span>
{% endif %}
{% endmacro %}
<|file_sep|>updated/app/templates/components/show-more.html | {% macro show_more(url, label, with_border=True) %}
<a
href="{{ url }}"
class="show-more{% if not with_border %}-no-border{% endif %}"
><span>{{ label }}</span></a>
{% endmacro %} | <|file_sep|>original/app/templates/components/show-more.html
{% macro show_more(url=None, label=None, with_border=True) %}
{% if url and label %}
<a
href="{{ url }}"
class="show-more{% if not with_border %}-no-border{% endif %}"
><span>{{ label }}</span></a>
{% else %}
<span class="show-more-empty"></span>
{% endif %}
{% endmacro %}
<|file_sep|>current/app/templates/components/show-more.html
{% macro show_more(url=None, label=None, with_border=True) %}
{% if url and label %}
<a
href="{{ url }}"
class="show-more{% if not with_border %}-no-border{% endif %}"
><span>{{ label }}</span></a>
{% else %}
<span class="show-more-empty"></span>
{% endif %}
{% endmacro %}
<|file_sep|>updated/app/templates/components/show-more.html
{% macro show_more(url, label, with_border=True) %}
<a
href="{{ url }}"
class="show-more{% if not with_border %}-no-border{% endif %}"
><span>{{ label }}</span></a>
{% endmacro %} | 3eb550d297e694dbfe9a202b043845c924ccdf3b | app/templates/components/show-more.html | app/templates/components/show-more.html | HTML |
<|file_sep|>original/src/bookshelf_fs.erl
<|file_sep|>current/src/bookshelf_fs.erl
<|file_sep|>updated/src/bookshelf_fs.erl | %% @copyright 2012 Opscode, Inc. All Rights Reserved
%% @author Tim Dysinger <timd@opscode.com>
%%
%% Licensed to the Apache Software Foundation (ASF) under one or more
%% contributor license agreements. See the NOTICE file distributed
%% with this work for additional information regarding copyright
%% ownership. The ASF licenses this file to you under the Apache
%% License, Version 2.0 (the "License"); you may not use this file
%% except in compliance with the License. You may obtain a copy of
%% the License at http://www.apache.org/licenses/LICENSE-2.0
%%
%% Unless required by applicable law or agreed to in writing, software
%% distributed under the License is distributed on an "AS IS" BASIS,
%% WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
%% implied. See the License for the specific language governing
%% permissions and limitations under the License.
-module(bookshelf_fs).
-include("bookshelf.hrl").
-export([
bucket_list/1, | <|file_sep|>original/src/bookshelf_fs.erl
<|file_sep|>current/src/bookshelf_fs.erl
<|file_sep|>updated/src/bookshelf_fs.erl
%% @copyright 2012 Opscode, Inc. All Rights Reserved
%% @author Tim Dysinger <timd@opscode.com>
%%
%% Licensed to the Apache Software Foundation (ASF) under one or more
%% contributor license agreements. See the NOTICE file distributed
%% with this work for additional information regarding copyright
%% ownership. The ASF licenses this file to you under the Apache
%% License, Version 2.0 (the "License"); you may not use this file
%% except in compliance with the License. You may obtain a copy of
%% the License at http://www.apache.org/licenses/LICENSE-2.0
%%
%% Unless required by applicable law or agreed to in writing, software
%% distributed under the License is distributed on an "AS IS" BASIS,
%% WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
%% implied. See the License for the specific language governing
%% permissions and limitations under the License.
-module(bookshelf_fs).
-include("bookshelf.hrl").
-export([
bucket_list/1, | 12325306679c695ddcecd3ce2e5843fd0958ad8e | src/bookshelf_fs.erl | src/bookshelf_fs.erl | Erlang |
<|file_sep|>circle.yml.diff
original:
updated:
PROJECT: "${HOME}/.go_project/src/github.com/jamesnetherton/homehub-cli"
<|file_sep|>circle.yml.diff
original:
- cd ${GOPATH} && make build
updated:
- cd ${PROJECT} && make build
<|file_sep|>original/circle.yml
PATH: "/usr/local/go/bin:/usr/local/go_workspace/bin:~/.go_workspace/bin:${PATH}"
GOPATH: "${HOME}/.go_workspace:/usr/local/go_workspace:${HOME}/.go_project"
post:
- mkdir -p download
- mkdir -p ${HOME}/.go_project/src/github.com/jamesnetherton/
- cd ${HOME}/.go_project/src/github.com/jamesnetherton/ && ln -s ${HOME}/homehub-cli homehub-cli
- test -e download/$GODIST || curl -o download/$GODIST https://storage.googleapis.com/golang/$GODIST
- sudo rm -rf /usr/local/go
- sudo tar -C /usr/local -xzf download/$GODIST
dependencies:
cache_directories:
- ~/download
override:
- cd ${GOPATH} && make build
deployment:
release:
branch: release
commands:
- cd ${GOPATH} && make release
<|file_sep|>current/circle.yml
GOPATH: "${HOME}/.go_workspace:/usr/local/go_workspace:${HOME}/.go_project"
PROJECT: "${HOME}/.go_project/src/github.com/jamesnetherton/homehub-cli"
post:
- mkdir -p download
- mkdir -p ${HOME}/.go_project/src/github.com/jamesnetherton/
- cd ${HOME}/.go_project/src/github.com/jamesnetherton/ && ln -s ${HOME}/homehub-cli homehub-cli
- test -e download/$GODIST || curl -o download/$GODIST https://storage.googleapis.com/golang/$GODIST
- sudo rm -rf /usr/local/go
- sudo tar -C /usr/local -xzf download/$GODIST
dependencies:
cache_directories:
- ~/download
override:
- cd ${PROJECT} && make build
deployment:
release:
branch: release
commands:
- cd ${GOPATH} && make release
<|file_sep|>updated/circle.yml | GOPATH: "${HOME}/.go_workspace:/usr/local/go_workspace:${HOME}/.go_project"
PROJECT: "${HOME}/.go_project/src/github.com/jamesnetherton/homehub-cli"
post:
- mkdir -p download
- mkdir -p ${HOME}/.go_project/src/github.com/jamesnetherton/
- cd ${HOME}/.go_project/src/github.com/jamesnetherton/ && ln -s ${HOME}/homehub-cli homehub-cli
- test -e download/$GODIST || curl -o download/$GODIST https://storage.googleapis.com/golang/$GODIST
- sudo rm -rf /usr/local/go
- sudo tar -C /usr/local -xzf download/$GODIST
dependencies:
cache_directories:
- ~/download
override:
- cd ${PROJECT} && make build
deployment:
release:
branch: release
commands:
- cd ${PROJECT} && make release | <|file_sep|>circle.yml.diff
original:
updated:
PROJECT: "${HOME}/.go_project/src/github.com/jamesnetherton/homehub-cli"
<|file_sep|>circle.yml.diff
original:
- cd ${GOPATH} && make build
updated:
- cd ${PROJECT} && make build
<|file_sep|>original/circle.yml
PATH: "/usr/local/go/bin:/usr/local/go_workspace/bin:~/.go_workspace/bin:${PATH}"
GOPATH: "${HOME}/.go_workspace:/usr/local/go_workspace:${HOME}/.go_project"
post:
- mkdir -p download
- mkdir -p ${HOME}/.go_project/src/github.com/jamesnetherton/
- cd ${HOME}/.go_project/src/github.com/jamesnetherton/ && ln -s ${HOME}/homehub-cli homehub-cli
- test -e download/$GODIST || curl -o download/$GODIST https://storage.googleapis.com/golang/$GODIST
- sudo rm -rf /usr/local/go
- sudo tar -C /usr/local -xzf download/$GODIST
dependencies:
cache_directories:
- ~/download
override:
- cd ${GOPATH} && make build
deployment:
release:
branch: release
commands:
- cd ${GOPATH} && make release
<|file_sep|>current/circle.yml
GOPATH: "${HOME}/.go_workspace:/usr/local/go_workspace:${HOME}/.go_project"
PROJECT: "${HOME}/.go_project/src/github.com/jamesnetherton/homehub-cli"
post:
- mkdir -p download
- mkdir -p ${HOME}/.go_project/src/github.com/jamesnetherton/
- cd ${HOME}/.go_project/src/github.com/jamesnetherton/ && ln -s ${HOME}/homehub-cli homehub-cli
- test -e download/$GODIST || curl -o download/$GODIST https://storage.googleapis.com/golang/$GODIST
- sudo rm -rf /usr/local/go
- sudo tar -C /usr/local -xzf download/$GODIST
dependencies:
cache_directories:
- ~/download
override:
- cd ${PROJECT} && make build
deployment:
release:
branch: release
commands:
- cd ${GOPATH} && make release
<|file_sep|>updated/circle.yml
GOPATH: "${HOME}/.go_workspace:/usr/local/go_workspace:${HOME}/.go_project"
PROJECT: "${HOME}/.go_project/src/github.com/jamesnetherton/homehub-cli"
post:
- mkdir -p download
- mkdir -p ${HOME}/.go_project/src/github.com/jamesnetherton/
- cd ${HOME}/.go_project/src/github.com/jamesnetherton/ && ln -s ${HOME}/homehub-cli homehub-cli
- test -e download/$GODIST || curl -o download/$GODIST https://storage.googleapis.com/golang/$GODIST
- sudo rm -rf /usr/local/go
- sudo tar -C /usr/local -xzf download/$GODIST
dependencies:
cache_directories:
- ~/download
override:
- cd ${PROJECT} && make build
deployment:
release:
branch: release
commands:
- cd ${PROJECT} && make release | 2198fbdc97c17bdf4e08112ab33a08276bf874d1 | circle.yml | circle.yml | YAML |
<|file_sep|>original/src/elliptical_slice.jl
<|file_sep|>current/src/elliptical_slice.jl
<|file_sep|>updated/src/elliptical_slice.jl |
# http://homepages.inf.ed.ac.uk/imurray2/pub/10ess/elliptical_slice.m
# https://github.com/jobovy/bovy_mcmc/blob/master/bovy_mcmc/elliptical_slice.py
function elliptical_slice(xx, prior::Array{Float64,}, log_like_fn, cur_log_like)
D = length(xx);
cur_log_like = log_like_fn(xx);
angle_range = 0;
hh = log(rand()) + cur_log_like;
# Set up a bracket of angles and pick a first proposal.
# "phi = (theta'-theta)" is a change in angle.
if angle_range <= 0
# Bracket whole ellipse with both edges at first proposed point
phi = rand() * 2*pi;
phi_min = phi - 2*pi;
phi_max = phi;
else
# Randomly center bracket on current point
phi_min = -angle_range*rand(); | <|file_sep|>original/src/elliptical_slice.jl
<|file_sep|>current/src/elliptical_slice.jl
<|file_sep|>updated/src/elliptical_slice.jl
# http://homepages.inf.ed.ac.uk/imurray2/pub/10ess/elliptical_slice.m
# https://github.com/jobovy/bovy_mcmc/blob/master/bovy_mcmc/elliptical_slice.py
function elliptical_slice(xx, prior::Array{Float64,}, log_like_fn, cur_log_like)
D = length(xx);
cur_log_like = log_like_fn(xx);
angle_range = 0;
hh = log(rand()) + cur_log_like;
# Set up a bracket of angles and pick a first proposal.
# "phi = (theta'-theta)" is a change in angle.
if angle_range <= 0
# Bracket whole ellipse with both edges at first proposed point
phi = rand() * 2*pi;
phi_min = phi - 2*pi;
phi_max = phi;
else
# Randomly center bracket on current point
phi_min = -angle_range*rand(); | f2e28a96b1b208faa612b11247293957bbb93af3 | src/elliptical_slice.jl | src/elliptical_slice.jl | Julia |
<|file_sep|>original/_posts/2018-01-03-season-1-episode-1.markdown
---
layout: post
title: "001: Introduction to the FLOSS for Science podcast"
date: 2018-01-03 13:59:59
author: Admin
categories: podcast
#- blog
img: post01.jpg
thumb: thumb01.jpg
podcast_link: https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3
podcast_file_size: 5.0 MB
podcast_duration: "10:48"
podcast_length: 5220352
podcast_guid: 454393ac623ec4173b3a93c0bb2d943b2f81074979185bf8c4350cd9812c2677
---
In this epsiode the two hosts David Brassard and Patrick Diehl of FLOSS for science introduce themself and the explain the aims of this podcast. Listen this episode here or add our rss feed to your favourite podcast app:
<iframe src="https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3" width="500" height="30" frameborder="0" webkitallowfullscreen="true" mozallowfullscreen="true" allowfullscreen></iframe>
<|file_sep|>current/_posts/2018-01-03-season-1-episode-1.markdown
---
layout: post
title: "001: Introduction to the FLOSS for Science podcast"
date: 2018-01-03 13:59:59
author: Admin
categories: podcast
#- blog
img: post01.jpg
thumb: thumb01.jpg
podcast_link: https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3
podcast_file_size: 5.0 MB
podcast_duration: "10:48"
podcast_length: 5220352
podcast_guid: 454393ac623ec4173b3a93c0bb2d943b2f81074979185bf8c4350cd9812c2677
---
In this epsiode the two hosts David Brassard and Patrick Diehl of FLOSS for science introduce themself and the explain the aims of this podcast. Listen this episode here or add our rss feed to your favourite podcast app:
<iframe src="https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3" width="500" height="30" frameborder="0" webkitallowfullscreen="true" mozallowfullscreen="true" allowfullscreen></iframe>
<|file_sep|>updated/_posts/2018-01-03-season-1-episode-1.markdown | layout: post
title: "001: Introduction to the FLOSS for Science podcast"
date: 2018-01-03 13:59:59
author: Admin
categories: podcast
#- blog
img: post01.jpg
thumb: thumb01.jpg
podcast_link: https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3
podcast_file_size: 5.0 MB
podcast_duration: "10:48"
podcast_length: 5220352
podcast_guid: 454393ac623ec4173b3a93c0bb2d943b2f81074979185bf8c4350cd9812c2677
---
In this epsiode the two hosts David Brassard and Patrick Diehl of FLOSS for science introduce themself and the explain the aims of this podcast. Listen this episode here or add our rss feed to your favourite podcast app:
<audio controls>
<source src="https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3" type="audio/mpeg">
Your browser does not support the audio element.
</audio>
| <|file_sep|>original/_posts/2018-01-03-season-1-episode-1.markdown
---
layout: post
title: "001: Introduction to the FLOSS for Science podcast"
date: 2018-01-03 13:59:59
author: Admin
categories: podcast
#- blog
img: post01.jpg
thumb: thumb01.jpg
podcast_link: https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3
podcast_file_size: 5.0 MB
podcast_duration: "10:48"
podcast_length: 5220352
podcast_guid: 454393ac623ec4173b3a93c0bb2d943b2f81074979185bf8c4350cd9812c2677
---
In this epsiode the two hosts David Brassard and Patrick Diehl of FLOSS for science introduce themself and the explain the aims of this podcast. Listen this episode here or add our rss feed to your favourite podcast app:
<iframe src="https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3" width="500" height="30" frameborder="0" webkitallowfullscreen="true" mozallowfullscreen="true" allowfullscreen></iframe>
<|file_sep|>current/_posts/2018-01-03-season-1-episode-1.markdown
---
layout: post
title: "001: Introduction to the FLOSS for Science podcast"
date: 2018-01-03 13:59:59
author: Admin
categories: podcast
#- blog
img: post01.jpg
thumb: thumb01.jpg
podcast_link: https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3
podcast_file_size: 5.0 MB
podcast_duration: "10:48"
podcast_length: 5220352
podcast_guid: 454393ac623ec4173b3a93c0bb2d943b2f81074979185bf8c4350cd9812c2677
---
In this epsiode the two hosts David Brassard and Patrick Diehl of FLOSS for science introduce themself and the explain the aims of this podcast. Listen this episode here or add our rss feed to your favourite podcast app:
<iframe src="https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3" width="500" height="30" frameborder="0" webkitallowfullscreen="true" mozallowfullscreen="true" allowfullscreen></iframe>
<|file_sep|>updated/_posts/2018-01-03-season-1-episode-1.markdown
layout: post
title: "001: Introduction to the FLOSS for Science podcast"
date: 2018-01-03 13:59:59
author: Admin
categories: podcast
#- blog
img: post01.jpg
thumb: thumb01.jpg
podcast_link: https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3
podcast_file_size: 5.0 MB
podcast_duration: "10:48"
podcast_length: 5220352
podcast_guid: 454393ac623ec4173b3a93c0bb2d943b2f81074979185bf8c4350cd9812c2677
---
In this epsiode the two hosts David Brassard and Patrick Diehl of FLOSS for science introduce themself and the explain the aims of this podcast. Listen this episode here or add our rss feed to your favourite podcast app:
<audio controls>
<source src="https://archive.org/download/FlossforscienceEp001-Introduction/FlossforscienceEp001-Introduction.mp3" type="audio/mpeg">
Your browser does not support the audio element.
</audio>
| 7fdd24a9745917da6b39c81ebd19450d6c789f5c | _posts/2018-01-03-season-1-episode-1.markdown | _posts/2018-01-03-season-1-episode-1.markdown | Markdown |
<|file_sep|>original/README.md
n3554
=====
A proposal for a standard parallel algorithms library for ISO C++.
Building the PDF:
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.pdf
Just generating LaTex (you can edit the `.tex` and then `pdflatex` the output):
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.tex
<|file_sep|>current/README.md
n3554
=====
A proposal for a standard parallel algorithms library for ISO C++.
Building the PDF:
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.pdf
Just generating LaTex (you can edit the `.tex` and then `pdflatex` the output):
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.tex
<|file_sep|>updated/README.md | n3554
=====
A proposal for a standard parallel algorithms library for ISO C++.
Building the PDF:
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.pdf
Just generating LaTex (you can edit the `.tex` and then `pdflatex` the output):
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.tex
[Get `pandoc` here.](http://johnmacfarlane.net/pandoc/installing.html)
| <|file_sep|>original/README.md
n3554
=====
A proposal for a standard parallel algorithms library for ISO C++.
Building the PDF:
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.pdf
Just generating LaTex (you can edit the `.tex` and then `pdflatex` the output):
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.tex
<|file_sep|>current/README.md
n3554
=====
A proposal for a standard parallel algorithms library for ISO C++.
Building the PDF:
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.pdf
Just generating LaTex (you can edit the `.tex` and then `pdflatex` the output):
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.tex
<|file_sep|>updated/README.md
n3554
=====
A proposal for a standard parallel algorithms library for ISO C++.
Building the PDF:
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.pdf
Just generating LaTex (you can edit the `.tex` and then `pdflatex` the output):
$ pandoc --number-sections -H n3554_header.tex n3554_a_parallel_algorithms_library.*.md -o n3554_a_parallel_algorithms_library.tex
[Get `pandoc` here.](http://johnmacfarlane.net/pandoc/installing.html)
| 83e5a7cebb6082bc46773f483c68176564cccc19 | README.md | README.md | Markdown |
<|file_sep|>original/static/js/sh_init.js
if (elPre.tagName != "PRE") continue;
// Prepare for highlighting
// elPre.className = elPre.className.replace("sourceCode ","");
if (elPre.className.match("sourceCode"))
switch (elPre.className.replace("sourceCode ","")) {
case 'css':
case 'haskell':
case 'html':
case 'php':
case 'python':
elPre.className = "sourceCode sh_" + elPre.className.replace("sourceCode ", "");
break;
case 'js':
elPre.className += " sh_javascript";
case 'literate haskell':
elPre.className += " sh_haskell";
break;
}
}
<|file_sep|>current/static/js/sh_init.js
if (elPre.tagName != "PRE") continue;
// Prepare for highlighting
// elPre.className = elPre.className.replace("sourceCode ","");
if (elPre.className.match("sourceCode"))
switch (elPre.className.replace("sourceCode ","")) {
case 'css':
case 'haskell':
case 'html':
case 'php':
case 'python':
elPre.className = "sourceCode sh_" + elPre.className.replace("sourceCode ", "");
break;
case 'js':
elPre.className += " sh_javascript";
case 'literate haskell':
elPre.className += " sh_haskell";
break;
}
}
<|file_sep|>updated/static/js/sh_init.js | if (elPre.tagName != "PRE") continue;
// Prepare for highlighting
// elPre.className = elPre.className.replace("sourceCode ","");
if (elPre.className.match("sourceCode"))
switch (elPre.className.replace("sourceCode ","")) {
case 'css':
case 'haskell':
case 'html':
case 'php':
case 'java':
case 'python':
elPre.className = "sourceCode sh_" + elPre.className.replace("sourceCode ", "");
break;
case 'js':
elPre.className += " sh_javascript";
case 'literate haskell':
elPre.className += " sh_haskell";
break;
}
} | <|file_sep|>original/static/js/sh_init.js
if (elPre.tagName != "PRE") continue;
// Prepare for highlighting
// elPre.className = elPre.className.replace("sourceCode ","");
if (elPre.className.match("sourceCode"))
switch (elPre.className.replace("sourceCode ","")) {
case 'css':
case 'haskell':
case 'html':
case 'php':
case 'python':
elPre.className = "sourceCode sh_" + elPre.className.replace("sourceCode ", "");
break;
case 'js':
elPre.className += " sh_javascript";
case 'literate haskell':
elPre.className += " sh_haskell";
break;
}
}
<|file_sep|>current/static/js/sh_init.js
if (elPre.tagName != "PRE") continue;
// Prepare for highlighting
// elPre.className = elPre.className.replace("sourceCode ","");
if (elPre.className.match("sourceCode"))
switch (elPre.className.replace("sourceCode ","")) {
case 'css':
case 'haskell':
case 'html':
case 'php':
case 'python':
elPre.className = "sourceCode sh_" + elPre.className.replace("sourceCode ", "");
break;
case 'js':
elPre.className += " sh_javascript";
case 'literate haskell':
elPre.className += " sh_haskell";
break;
}
}
<|file_sep|>updated/static/js/sh_init.js
if (elPre.tagName != "PRE") continue;
// Prepare for highlighting
// elPre.className = elPre.className.replace("sourceCode ","");
if (elPre.className.match("sourceCode"))
switch (elPre.className.replace("sourceCode ","")) {
case 'css':
case 'haskell':
case 'html':
case 'php':
case 'java':
case 'python':
elPre.className = "sourceCode sh_" + elPre.className.replace("sourceCode ", "");
break;
case 'js':
elPre.className += " sh_javascript";
case 'literate haskell':
elPre.className += " sh_haskell";
break;
}
} | a439b552acbd586615b15b86d79ed003979aeaed | static/js/sh_init.js | static/js/sh_init.js | JavaScript |
<|file_sep|>.travis.yml.diff
original:
updated:
<|file_sep|>original/.travis.yml
language: rust
rust:
- 1.0.0-beta
- nightly
<|file_sep|>current/.travis.yml
language: rust
rust:
- 1.0.0-beta
- nightly
<|file_sep|>updated/.travis.yml | language: rust
rust:
- 1.0.0-beta
- nightly
after_success: |
[ $TRAVIS_BRANCH = master ] &&
[ $TRAVIS_PULL_REQUEST = false ] &&
cargo doc &&
echo "<meta http-equiv=refresh content=0;url=`echo $TRAVIS_REPO_SLUG | cut -d '/' -f 2`/index.html>" > target/doc/index.html &&
sudo pip install ghp-import &&
ghp-import -n target/doc &&
git push -fq https://${GH_TOKEN}@github.com/${TRAVIS_REPO_SLUG}.git gh-pages
env:
global:
secure: ueWZoYHNmficGPMDmcERVtC3dcEjdJxPA4RPSknYGXMV9LJgvRRMAqa/9lG4uTb/mgIppALZTRF+F+6Tl8cxrySANSmClT11p9dZHI7fmlIE/eRqwviDHnYNOP7FCsWSQsFkjjzK4Q78h3cqsqjLuIaf7UjgsjPKNOrLmY2PIlg= | <|file_sep|>.travis.yml.diff
original:
updated:
<|file_sep|>original/.travis.yml
language: rust
rust:
- 1.0.0-beta
- nightly
<|file_sep|>current/.travis.yml
language: rust
rust:
- 1.0.0-beta
- nightly
<|file_sep|>updated/.travis.yml
language: rust
rust:
- 1.0.0-beta
- nightly
after_success: |
[ $TRAVIS_BRANCH = master ] &&
[ $TRAVIS_PULL_REQUEST = false ] &&
cargo doc &&
echo "<meta http-equiv=refresh content=0;url=`echo $TRAVIS_REPO_SLUG | cut -d '/' -f 2`/index.html>" > target/doc/index.html &&
sudo pip install ghp-import &&
ghp-import -n target/doc &&
git push -fq https://${GH_TOKEN}@github.com/${TRAVIS_REPO_SLUG}.git gh-pages
env:
global:
secure: ueWZoYHNmficGPMDmcERVtC3dcEjdJxPA4RPSknYGXMV9LJgvRRMAqa/9lG4uTb/mgIppALZTRF+F+6Tl8cxrySANSmClT11p9dZHI7fmlIE/eRqwviDHnYNOP7FCsWSQsFkjjzK4Q78h3cqsqjLuIaf7UjgsjPKNOrLmY2PIlg= | 45110420e832d5392f3c86e861ca3ba268fb6be4 | .travis.yml | .travis.yml | YAML |
<|file_sep|>html.js.diff
original:
const head = Helmet.rewind()
updated:
const {title} = Helmet.rewind()
<|file_sep|>original/html.js
css = <style dangerouslySetInnerHTML={ { __html: require('!raw!./public/styles.css')} } />
}
return (
<html lang="en">
<head>
<meta charSet="utf-8" />
<meta httpEquiv="X-UA-Compatible" content="IE=edge" />
<meta name="viewport" content="width=device-width, initial-scale=1.0 maximum-scale=5.0" />
<title>
{ head.title }
</title>
{ font }
{ css }
</head>
<body>
<div id="react-mount" dangerouslySetInnerHTML={ { __html: this.props.body} } />
<script src={ prefixLink(`/bundle.js?t=${BUILD_TIME}`) } />
</body>
</html>
)
<|file_sep|>current/html.js
css = <style dangerouslySetInnerHTML={ { __html: require('!raw!./public/styles.css')} } />
}
return (
<html lang="en">
<head>
<meta charSet="utf-8" />
<meta httpEquiv="X-UA-Compatible" content="IE=edge" />
<meta name="viewport" content="width=device-width, initial-scale=1.0 maximum-scale=5.0" />
<title>
{ head.title }
</title>
{ font }
{ css }
</head>
<body>
<div id="react-mount" dangerouslySetInnerHTML={ { __html: this.props.body} } />
<script src={ prefixLink(`/bundle.js?t=${BUILD_TIME}`) } />
</body>
</html>
)
<|file_sep|>updated/html.js | css = <style dangerouslySetInnerHTML={ { __html: require('!raw!./public/styles.css')} } />
}
return (
<html lang="en">
<head>
<meta charSet="utf-8" />
<meta httpEquiv="X-UA-Compatible" content="IE=edge" />
<meta name="viewport" content="width=device-width, initial-scale=1.0 maximum-scale=5.0" />
{ title.toComponent() }
{ font }
{ css }
</head>
<body>
<div id="react-mount" dangerouslySetInnerHTML={ { __html: this.props.body} } />
<script src={ prefixLink(`/bundle.js?t=${BUILD_TIME}`) } />
</body>
</html>
)
},
}) | <|file_sep|>html.js.diff
original:
const head = Helmet.rewind()
updated:
const {title} = Helmet.rewind()
<|file_sep|>original/html.js
css = <style dangerouslySetInnerHTML={ { __html: require('!raw!./public/styles.css')} } />
}
return (
<html lang="en">
<head>
<meta charSet="utf-8" />
<meta httpEquiv="X-UA-Compatible" content="IE=edge" />
<meta name="viewport" content="width=device-width, initial-scale=1.0 maximum-scale=5.0" />
<title>
{ head.title }
</title>
{ font }
{ css }
</head>
<body>
<div id="react-mount" dangerouslySetInnerHTML={ { __html: this.props.body} } />
<script src={ prefixLink(`/bundle.js?t=${BUILD_TIME}`) } />
</body>
</html>
)
<|file_sep|>current/html.js
css = <style dangerouslySetInnerHTML={ { __html: require('!raw!./public/styles.css')} } />
}
return (
<html lang="en">
<head>
<meta charSet="utf-8" />
<meta httpEquiv="X-UA-Compatible" content="IE=edge" />
<meta name="viewport" content="width=device-width, initial-scale=1.0 maximum-scale=5.0" />
<title>
{ head.title }
</title>
{ font }
{ css }
</head>
<body>
<div id="react-mount" dangerouslySetInnerHTML={ { __html: this.props.body} } />
<script src={ prefixLink(`/bundle.js?t=${BUILD_TIME}`) } />
</body>
</html>
)
<|file_sep|>updated/html.js
css = <style dangerouslySetInnerHTML={ { __html: require('!raw!./public/styles.css')} } />
}
return (
<html lang="en">
<head>
<meta charSet="utf-8" />
<meta httpEquiv="X-UA-Compatible" content="IE=edge" />
<meta name="viewport" content="width=device-width, initial-scale=1.0 maximum-scale=5.0" />
{ title.toComponent() }
{ font }
{ css }
</head>
<body>
<div id="react-mount" dangerouslySetInnerHTML={ { __html: this.props.body} } />
<script src={ prefixLink(`/bundle.js?t=${BUILD_TIME}`) } />
</body>
</html>
)
},
}) | 67696ac08bf39662aa796ce44e556b73e96594a5 | html.js | html.js | JavaScript |
<|file_sep|>original/qlf/dashboard/templates/dashboard/embed_bokeh.html
{% block content %}
<div class="container">
<div class="page-header">
<!-- TODO: improve this -->
{% if "exposures" in bokeh_app %}
<h1>Exposures</h1>
{% else %}
<a class="btn btn-default"
href="/dashboard/exposures/"
role="button">Back</a>
{% endif %}
</div>
<div class="container">
{{ bokeh_script|safe }}
</div>
<|file_sep|>current/qlf/dashboard/templates/dashboard/embed_bokeh.html
{% block content %}
<div class="container">
<div class="page-header">
<!-- TODO: improve this -->
{% if "exposures" in bokeh_app %}
<h1>Exposures</h1>
{% else %}
<a class="btn btn-default"
href="/dashboard/exposures/"
role="button">Back</a>
{% endif %}
</div>
<div class="container">
{{ bokeh_script|safe }}
</div>
<|file_sep|>updated/qlf/dashboard/templates/dashboard/embed_bokeh.html |
{% block content %}
<div class="container">
<div class="page-header">
<!-- TODO: improve this -->
{% if "exposures" in bokeh_app %}
<h1>Exposures</h1>
{% elif "monitor" in bokeh_app %}
<h1>Monitor</h1>
{% else %}
<a class="btn btn-default"
href="/dashboard/exposures/"
role="button">Back</a>
{% endif %}
</div>
<div class="container">
{{ bokeh_script|safe }} | <|file_sep|>original/qlf/dashboard/templates/dashboard/embed_bokeh.html
{% block content %}
<div class="container">
<div class="page-header">
<!-- TODO: improve this -->
{% if "exposures" in bokeh_app %}
<h1>Exposures</h1>
{% else %}
<a class="btn btn-default"
href="/dashboard/exposures/"
role="button">Back</a>
{% endif %}
</div>
<div class="container">
{{ bokeh_script|safe }}
</div>
<|file_sep|>current/qlf/dashboard/templates/dashboard/embed_bokeh.html
{% block content %}
<div class="container">
<div class="page-header">
<!-- TODO: improve this -->
{% if "exposures" in bokeh_app %}
<h1>Exposures</h1>
{% else %}
<a class="btn btn-default"
href="/dashboard/exposures/"
role="button">Back</a>
{% endif %}
</div>
<div class="container">
{{ bokeh_script|safe }}
</div>
<|file_sep|>updated/qlf/dashboard/templates/dashboard/embed_bokeh.html
{% block content %}
<div class="container">
<div class="page-header">
<!-- TODO: improve this -->
{% if "exposures" in bokeh_app %}
<h1>Exposures</h1>
{% elif "monitor" in bokeh_app %}
<h1>Monitor</h1>
{% else %}
<a class="btn btn-default"
href="/dashboard/exposures/"
role="button">Back</a>
{% endif %}
</div>
<div class="container">
{{ bokeh_script|safe }} | 642fa098eedc2ef20ec9720ae0cc851acda68479 | qlf/dashboard/templates/dashboard/embed_bokeh.html | qlf/dashboard/templates/dashboard/embed_bokeh.html | HTML |
<|file_sep|>appveyor.yml.diff
original:
updated:
# Output useful info for debugging
- node --version
- npm --version
<|file_sep|>original/appveyor.yml
- nodejs_version: "4" # LTS - Recommended for most users
- nodejs_version: "6" # Current - Latest features
platform:
- x86
- x64
install:
- ps: Install-Product node $env:nodejs_version $env:platform
- cd ./src
- npm install
test_script:
# Output useful info for debugging.
- node --version
- npm --version
# run tests
- npm test
# Don't actually build.
build: off
<|file_sep|>current/appveyor.yml
platform:
- x86
- x64
install:
- ps: Install-Product node $env:nodejs_version $env:platform
- cd ./src
- npm install
# Output useful info for debugging
- node --version
- npm --version
test_script:
# Output useful info for debugging.
- node --version
- npm --version
# run tests
- npm test
# Don't actually build.
build: off
<|file_sep|>updated/appveyor.yml | matrix:
- nodejs_version: "4" # LTS - Recommended for most users
- nodejs_version: "6" # Current - Latest features
platform:
- x86
- x64
install:
- ps: Install-Product node $env:nodejs_version $env:platform
- cd ./src
- npm install
# Output useful info for debugging
- node --version
- npm --version
test_script:
- npm test
# Don't actually build.
build: off | <|file_sep|>appveyor.yml.diff
original:
updated:
# Output useful info for debugging
- node --version
- npm --version
<|file_sep|>original/appveyor.yml
- nodejs_version: "4" # LTS - Recommended for most users
- nodejs_version: "6" # Current - Latest features
platform:
- x86
- x64
install:
- ps: Install-Product node $env:nodejs_version $env:platform
- cd ./src
- npm install
test_script:
# Output useful info for debugging.
- node --version
- npm --version
# run tests
- npm test
# Don't actually build.
build: off
<|file_sep|>current/appveyor.yml
platform:
- x86
- x64
install:
- ps: Install-Product node $env:nodejs_version $env:platform
- cd ./src
- npm install
# Output useful info for debugging
- node --version
- npm --version
test_script:
# Output useful info for debugging.
- node --version
- npm --version
# run tests
- npm test
# Don't actually build.
build: off
<|file_sep|>updated/appveyor.yml
matrix:
- nodejs_version: "4" # LTS - Recommended for most users
- nodejs_version: "6" # Current - Latest features
platform:
- x86
- x64
install:
- ps: Install-Product node $env:nodejs_version $env:platform
- cd ./src
- npm install
# Output useful info for debugging
- node --version
- npm --version
test_script:
- npm test
# Don't actually build.
build: off | 69a9e6070d2d3ae38c68a98e49824594fa5d908c | appveyor.yml | appveyor.yml | YAML |
<|file_sep|>spec/features/work_generator_spec.rb.diff
original:
Rails::Generators.invoke('hyrax:work', ['Catapult'], destination_root: Rails.root)
updated:
Rails::Generators.invoke('hyrax:work', ['Catapult', '--quiet'], destination_root: Rails.root)
<|file_sep|>original/spec/features/work_generator_spec.rb
Rails::Generators.invoke('hyrax:work', ['Catapult'], destination_root: Rails.root)
load "#{EngineCart.destination}/app/models/catapult.rb"
load "#{EngineCart.destination}/app/controllers/hyrax/catapults_controller.rb"
load "#{EngineCart.destination}/app/actors/hyrax/actors/catapult_actor.rb"
load "#{EngineCart.destination}/app/forms/hyrax/catapult_form.rb"
load "#{EngineCart.destination}/config/initializers/hyrax.rb"
load "#{EngineCart.destination}/config/routes.rb"
load "app/helpers/hyrax/url_helper.rb"
end
after do
Rails::Generators.invoke('hyrax:work', ['Catapult'], behavior: :revoke, destination_root: Rails.root)
end
it 'catapults should behave like generic works' do
expect(Hyrax.config.curation_concerns).to include Catapult
expect(defined? Hyrax::Actors::CatapultActor).to eq 'constant'
expect(defined? Hyrax::CatapultsController).to eq 'constant'
expect(defined? Hyrax::CatapultForm).to eq 'constant'
end
end
<|file_sep|>current/spec/features/work_generator_spec.rb
Rails::Generators.invoke('hyrax:work', ['Catapult', '--quiet'], destination_root: Rails.root)
load "#{EngineCart.destination}/app/models/catapult.rb"
load "#{EngineCart.destination}/app/controllers/hyrax/catapults_controller.rb"
load "#{EngineCart.destination}/app/actors/hyrax/actors/catapult_actor.rb"
load "#{EngineCart.destination}/app/forms/hyrax/catapult_form.rb"
load "#{EngineCart.destination}/config/initializers/hyrax.rb"
load "#{EngineCart.destination}/config/routes.rb"
load "app/helpers/hyrax/url_helper.rb"
end
after do
Rails::Generators.invoke('hyrax:work', ['Catapult'], behavior: :revoke, destination_root: Rails.root)
end
it 'catapults should behave like generic works' do
expect(Hyrax.config.curation_concerns).to include Catapult
expect(defined? Hyrax::Actors::CatapultActor).to eq 'constant'
expect(defined? Hyrax::CatapultsController).to eq 'constant'
expect(defined? Hyrax::CatapultForm).to eq 'constant'
end
end
<|file_sep|>updated/spec/features/work_generator_spec.rb | Rails::Generators.invoke('hyrax:work', ['Catapult', '--quiet'], destination_root: Rails.root)
load "#{EngineCart.destination}/app/models/catapult.rb"
load "#{EngineCart.destination}/app/controllers/hyrax/catapults_controller.rb"
load "#{EngineCart.destination}/app/actors/hyrax/actors/catapult_actor.rb"
load "#{EngineCart.destination}/app/forms/hyrax/catapult_form.rb"
load "#{EngineCart.destination}/config/initializers/hyrax.rb"
load "#{EngineCart.destination}/config/routes.rb"
load "app/helpers/hyrax/url_helper.rb"
end
after do
Rails::Generators.invoke('hyrax:work', ['Catapult', '--quiet'], behavior: :revoke, destination_root: Rails.root)
end
it 'catapults should behave like generic works' do
expect(Hyrax.config.curation_concerns).to include Catapult
expect(defined? Hyrax::Actors::CatapultActor).to eq 'constant'
expect(defined? Hyrax::CatapultsController).to eq 'constant'
expect(defined? Hyrax::CatapultForm).to eq 'constant'
end
end | <|file_sep|>spec/features/work_generator_spec.rb.diff
original:
Rails::Generators.invoke('hyrax:work', ['Catapult'], destination_root: Rails.root)
updated:
Rails::Generators.invoke('hyrax:work', ['Catapult', '--quiet'], destination_root: Rails.root)
<|file_sep|>original/spec/features/work_generator_spec.rb
Rails::Generators.invoke('hyrax:work', ['Catapult'], destination_root: Rails.root)
load "#{EngineCart.destination}/app/models/catapult.rb"
load "#{EngineCart.destination}/app/controllers/hyrax/catapults_controller.rb"
load "#{EngineCart.destination}/app/actors/hyrax/actors/catapult_actor.rb"
load "#{EngineCart.destination}/app/forms/hyrax/catapult_form.rb"
load "#{EngineCart.destination}/config/initializers/hyrax.rb"
load "#{EngineCart.destination}/config/routes.rb"
load "app/helpers/hyrax/url_helper.rb"
end
after do
Rails::Generators.invoke('hyrax:work', ['Catapult'], behavior: :revoke, destination_root: Rails.root)
end
it 'catapults should behave like generic works' do
expect(Hyrax.config.curation_concerns).to include Catapult
expect(defined? Hyrax::Actors::CatapultActor).to eq 'constant'
expect(defined? Hyrax::CatapultsController).to eq 'constant'
expect(defined? Hyrax::CatapultForm).to eq 'constant'
end
end
<|file_sep|>current/spec/features/work_generator_spec.rb
Rails::Generators.invoke('hyrax:work', ['Catapult', '--quiet'], destination_root: Rails.root)
load "#{EngineCart.destination}/app/models/catapult.rb"
load "#{EngineCart.destination}/app/controllers/hyrax/catapults_controller.rb"
load "#{EngineCart.destination}/app/actors/hyrax/actors/catapult_actor.rb"
load "#{EngineCart.destination}/app/forms/hyrax/catapult_form.rb"
load "#{EngineCart.destination}/config/initializers/hyrax.rb"
load "#{EngineCart.destination}/config/routes.rb"
load "app/helpers/hyrax/url_helper.rb"
end
after do
Rails::Generators.invoke('hyrax:work', ['Catapult'], behavior: :revoke, destination_root: Rails.root)
end
it 'catapults should behave like generic works' do
expect(Hyrax.config.curation_concerns).to include Catapult
expect(defined? Hyrax::Actors::CatapultActor).to eq 'constant'
expect(defined? Hyrax::CatapultsController).to eq 'constant'
expect(defined? Hyrax::CatapultForm).to eq 'constant'
end
end
<|file_sep|>updated/spec/features/work_generator_spec.rb
Rails::Generators.invoke('hyrax:work', ['Catapult', '--quiet'], destination_root: Rails.root)
load "#{EngineCart.destination}/app/models/catapult.rb"
load "#{EngineCart.destination}/app/controllers/hyrax/catapults_controller.rb"
load "#{EngineCart.destination}/app/actors/hyrax/actors/catapult_actor.rb"
load "#{EngineCart.destination}/app/forms/hyrax/catapult_form.rb"
load "#{EngineCart.destination}/config/initializers/hyrax.rb"
load "#{EngineCart.destination}/config/routes.rb"
load "app/helpers/hyrax/url_helper.rb"
end
after do
Rails::Generators.invoke('hyrax:work', ['Catapult', '--quiet'], behavior: :revoke, destination_root: Rails.root)
end
it 'catapults should behave like generic works' do
expect(Hyrax.config.curation_concerns).to include Catapult
expect(defined? Hyrax::Actors::CatapultActor).to eq 'constant'
expect(defined? Hyrax::CatapultsController).to eq 'constant'
expect(defined? Hyrax::CatapultForm).to eq 'constant'
end
end | bd0ef6a613cf5f82b394553e5cb30c6dfe575558 | spec/features/work_generator_spec.rb | spec/features/work_generator_spec.rb | Ruby |
<|file_sep|>original/ext/gtest/CMakeLists.txt
cmake_minimum_required(VERSION 2.8.8)
project(gtest_builder C CXX)
include(ExternalProject)
#set(GTEST_FORCE_SHARED_CRT ON)
#set(GTEST_DISABLE_PTHREADS OFF)
ExternalProject_Add(googletest
GIT_REPOSITORY https://github.com/google/googletest.git
CMAKE_ARGS
# -Dgtest_force_shared_crt=${GTEST_FORCE_SHARED_CRT}
# -Dgtest_disable_pthreads=${GTEST_DISABLE_PTHREADS}
-DBUILD_GTEST=ON
PREFIX "${CMAKE_CURRENT_BINARY_DIR}"
INSTALL_COMMAND ""
)
# Specify include dir
ExternalProject_Get_Property(googletest source_dir)
set(GTEST_INCLUDE_DIRS ${source_dir}/googletest/include PARENT_SCOPE)
<|file_sep|>current/ext/gtest/CMakeLists.txt
cmake_minimum_required(VERSION 2.8.8)
project(gtest_builder C CXX)
include(ExternalProject)
#set(GTEST_FORCE_SHARED_CRT ON)
#set(GTEST_DISABLE_PTHREADS OFF)
ExternalProject_Add(googletest
GIT_REPOSITORY https://github.com/google/googletest.git
CMAKE_ARGS
# -Dgtest_force_shared_crt=${GTEST_FORCE_SHARED_CRT}
# -Dgtest_disable_pthreads=${GTEST_DISABLE_PTHREADS}
-DBUILD_GTEST=ON
PREFIX "${CMAKE_CURRENT_BINARY_DIR}"
INSTALL_COMMAND ""
)
# Specify include dir
ExternalProject_Get_Property(googletest source_dir)
set(GTEST_INCLUDE_DIRS ${source_dir}/googletest/include PARENT_SCOPE)
<|file_sep|>updated/ext/gtest/CMakeLists.txt | cmake_minimum_required(VERSION 2.8.8)
project(gtest_builder C CXX)
include(ExternalProject)
#set(GTEST_FORCE_SHARED_CRT ON)
#set(GTEST_DISABLE_PTHREADS OFF)
ExternalProject_Add(googletest
DEPENDS iproute2 # TODO - not really the right place for this.
GIT_REPOSITORY https://github.com/google/googletest.git
CMAKE_ARGS
# -Dgtest_force_shared_crt=${GTEST_FORCE_SHARED_CRT}
# -Dgtest_disable_pthreads=${GTEST_DISABLE_PTHREADS}
-DBUILD_GTEST=ON
PREFIX "${CMAKE_CURRENT_BINARY_DIR}"
INSTALL_COMMAND ""
)
# Specify include dir
ExternalProject_Get_Property(googletest source_dir) | <|file_sep|>original/ext/gtest/CMakeLists.txt
cmake_minimum_required(VERSION 2.8.8)
project(gtest_builder C CXX)
include(ExternalProject)
#set(GTEST_FORCE_SHARED_CRT ON)
#set(GTEST_DISABLE_PTHREADS OFF)
ExternalProject_Add(googletest
GIT_REPOSITORY https://github.com/google/googletest.git
CMAKE_ARGS
# -Dgtest_force_shared_crt=${GTEST_FORCE_SHARED_CRT}
# -Dgtest_disable_pthreads=${GTEST_DISABLE_PTHREADS}
-DBUILD_GTEST=ON
PREFIX "${CMAKE_CURRENT_BINARY_DIR}"
INSTALL_COMMAND ""
)
# Specify include dir
ExternalProject_Get_Property(googletest source_dir)
set(GTEST_INCLUDE_DIRS ${source_dir}/googletest/include PARENT_SCOPE)
<|file_sep|>current/ext/gtest/CMakeLists.txt
cmake_minimum_required(VERSION 2.8.8)
project(gtest_builder C CXX)
include(ExternalProject)
#set(GTEST_FORCE_SHARED_CRT ON)
#set(GTEST_DISABLE_PTHREADS OFF)
ExternalProject_Add(googletest
GIT_REPOSITORY https://github.com/google/googletest.git
CMAKE_ARGS
# -Dgtest_force_shared_crt=${GTEST_FORCE_SHARED_CRT}
# -Dgtest_disable_pthreads=${GTEST_DISABLE_PTHREADS}
-DBUILD_GTEST=ON
PREFIX "${CMAKE_CURRENT_BINARY_DIR}"
INSTALL_COMMAND ""
)
# Specify include dir
ExternalProject_Get_Property(googletest source_dir)
set(GTEST_INCLUDE_DIRS ${source_dir}/googletest/include PARENT_SCOPE)
<|file_sep|>updated/ext/gtest/CMakeLists.txt
cmake_minimum_required(VERSION 2.8.8)
project(gtest_builder C CXX)
include(ExternalProject)
#set(GTEST_FORCE_SHARED_CRT ON)
#set(GTEST_DISABLE_PTHREADS OFF)
ExternalProject_Add(googletest
DEPENDS iproute2 # TODO - not really the right place for this.
GIT_REPOSITORY https://github.com/google/googletest.git
CMAKE_ARGS
# -Dgtest_force_shared_crt=${GTEST_FORCE_SHARED_CRT}
# -Dgtest_disable_pthreads=${GTEST_DISABLE_PTHREADS}
-DBUILD_GTEST=ON
PREFIX "${CMAKE_CURRENT_BINARY_DIR}"
INSTALL_COMMAND ""
)
# Specify include dir
ExternalProject_Get_Property(googletest source_dir) | be0df090828820a3a17eed92f57baf2881576ada | ext/gtest/CMakeLists.txt | ext/gtest/CMakeLists.txt | Text |
<|file_sep|>code/get_parser_outputs.sh.diff
original:
echo "$0: ERROR: can not traverse input article_filename" >&2
updated:
echo "$0: ERROR: can not traverse input directory" >&2
<|file_sep|>original/code/get_parser_outputs.sh
if [ $# -ne 2 ]; then
echo "$0: ERROR: wrong number of arguments" >&2
echo "$0: USAGE: $0 INPUT_DIR OUTPUT_DIR" >&2
exit 1
fi
if [ \( ! -r $1 \) -o \( ! -x $1 \) ]; then
echo "$0: ERROR: can not traverse input article_filename" >&2
exit 1
fi
mkdir -p $2
for article_filename in `find $1 -maxdepth 1 -type d`; do
if [ -r ${article_filename}/input.text ]; then
if [ ! -r $2/`basename ${article_filename}` ]; then
ln -s ${article_filename}/input.text $2/`basename ${article_filename}`
fi
fi
done
<|file_sep|>current/code/get_parser_outputs.sh
if [ $# -ne 2 ]; then
echo "$0: ERROR: wrong number of arguments" >&2
echo "$0: USAGE: $0 INPUT_DIR OUTPUT_DIR" >&2
exit 1
fi
if [ \( ! -r $1 \) -o \( ! -x $1 \) ]; then
echo "$0: ERROR: can not traverse input directory" >&2
exit 1
fi
mkdir -p $2
for article_filename in `find $1 -maxdepth 1 -type d`; do
if [ -r ${article_filename}/input.text ]; then
if [ ! -r $2/`basename ${article_filename}` ]; then
ln -s ${article_filename}/input.text $2/`basename ${article_filename}`
fi
fi
done
<|file_sep|>updated/code/get_parser_outputs.sh | echo "$0: ERROR: wrong number of arguments" >&2
echo "$0: USAGE: $0 INPUT_DIR OUTPUT_DIR" >&2
exit 1
fi
if [ \( ! -r $1 \) -o \( ! -x $1 \) ]; then
echo "$0: ERROR: can not traverse input directory" >&2
exit 1
fi
mkdir -p $2
for article_filename in `find $1 -maxdepth 1 -type d`; do
if [ -r ${article_filename}/input.text ]; then
NAME=`basename ${article_filename}`
if [ ! -r $2/${NAME} ]; then
ln -s `readlink -f ${article_filename}/input.text` $2/${NAME}
fi
fi
done
| <|file_sep|>code/get_parser_outputs.sh.diff
original:
echo "$0: ERROR: can not traverse input article_filename" >&2
updated:
echo "$0: ERROR: can not traverse input directory" >&2
<|file_sep|>original/code/get_parser_outputs.sh
if [ $# -ne 2 ]; then
echo "$0: ERROR: wrong number of arguments" >&2
echo "$0: USAGE: $0 INPUT_DIR OUTPUT_DIR" >&2
exit 1
fi
if [ \( ! -r $1 \) -o \( ! -x $1 \) ]; then
echo "$0: ERROR: can not traverse input article_filename" >&2
exit 1
fi
mkdir -p $2
for article_filename in `find $1 -maxdepth 1 -type d`; do
if [ -r ${article_filename}/input.text ]; then
if [ ! -r $2/`basename ${article_filename}` ]; then
ln -s ${article_filename}/input.text $2/`basename ${article_filename}`
fi
fi
done
<|file_sep|>current/code/get_parser_outputs.sh
if [ $# -ne 2 ]; then
echo "$0: ERROR: wrong number of arguments" >&2
echo "$0: USAGE: $0 INPUT_DIR OUTPUT_DIR" >&2
exit 1
fi
if [ \( ! -r $1 \) -o \( ! -x $1 \) ]; then
echo "$0: ERROR: can not traverse input directory" >&2
exit 1
fi
mkdir -p $2
for article_filename in `find $1 -maxdepth 1 -type d`; do
if [ -r ${article_filename}/input.text ]; then
if [ ! -r $2/`basename ${article_filename}` ]; then
ln -s ${article_filename}/input.text $2/`basename ${article_filename}`
fi
fi
done
<|file_sep|>updated/code/get_parser_outputs.sh
echo "$0: ERROR: wrong number of arguments" >&2
echo "$0: USAGE: $0 INPUT_DIR OUTPUT_DIR" >&2
exit 1
fi
if [ \( ! -r $1 \) -o \( ! -x $1 \) ]; then
echo "$0: ERROR: can not traverse input directory" >&2
exit 1
fi
mkdir -p $2
for article_filename in `find $1 -maxdepth 1 -type d`; do
if [ -r ${article_filename}/input.text ]; then
NAME=`basename ${article_filename}`
if [ ! -r $2/${NAME} ]; then
ln -s `readlink -f ${article_filename}/input.text` $2/${NAME}
fi
fi
done
| d084dd1e67dc4e41f283de6e0bc1122b1f96bf4d | code/get_parser_outputs.sh | code/get_parser_outputs.sh | Shell |
<|file_sep|>original/app/assets/javascripts/components/overview/my_exercises/components/Header.jsx
export const Header = ({ completed = false, course, remaining = [], text }) => (
<header className="header">
<h3 className={completed ? 'completed' : ''}>
{ text }
{
remaining.length
? <small>{remaining.length} additional exercises remaining.</small>
: null
}
</h3>
<NavLink to={`../../${course.slug}/resources`} className="resources-link">
View all exercises
</NavLink>
</header>
);
Header.propTypes = {
course: PropTypes.shape({
slug: PropTypes.string.isRequired
}).isRequired,
remaining: PropTypes.array,
<|file_sep|>current/app/assets/javascripts/components/overview/my_exercises/components/Header.jsx
export const Header = ({ completed = false, course, remaining = [], text }) => (
<header className="header">
<h3 className={completed ? 'completed' : ''}>
{ text }
{
remaining.length
? <small>{remaining.length} additional exercises remaining.</small>
: null
}
</h3>
<NavLink to={`../../${course.slug}/resources`} className="resources-link">
View all exercises
</NavLink>
</header>
);
Header.propTypes = {
course: PropTypes.shape({
slug: PropTypes.string.isRequired
}).isRequired,
remaining: PropTypes.array,
<|file_sep|>updated/app/assets/javascripts/components/overview/my_exercises/components/Header.jsx | export const Header = ({ completed = false, course, remaining = [], text }) => (
<header className="header">
<h3 className={completed ? 'completed' : ''}>
{ text }
{
remaining.length
? <small>{remaining.length} additional exercises remaining.</small>
: null
}
</h3>
<NavLink exact to={`/courses/${course.slug}/resources`} className="resources-link">
View all exercises
</NavLink>
</header>
);
Header.propTypes = {
course: PropTypes.shape({
slug: PropTypes.string.isRequired
}).isRequired,
remaining: PropTypes.array, | <|file_sep|>original/app/assets/javascripts/components/overview/my_exercises/components/Header.jsx
export const Header = ({ completed = false, course, remaining = [], text }) => (
<header className="header">
<h3 className={completed ? 'completed' : ''}>
{ text }
{
remaining.length
? <small>{remaining.length} additional exercises remaining.</small>
: null
}
</h3>
<NavLink to={`../../${course.slug}/resources`} className="resources-link">
View all exercises
</NavLink>
</header>
);
Header.propTypes = {
course: PropTypes.shape({
slug: PropTypes.string.isRequired
}).isRequired,
remaining: PropTypes.array,
<|file_sep|>current/app/assets/javascripts/components/overview/my_exercises/components/Header.jsx
export const Header = ({ completed = false, course, remaining = [], text }) => (
<header className="header">
<h3 className={completed ? 'completed' : ''}>
{ text }
{
remaining.length
? <small>{remaining.length} additional exercises remaining.</small>
: null
}
</h3>
<NavLink to={`../../${course.slug}/resources`} className="resources-link">
View all exercises
</NavLink>
</header>
);
Header.propTypes = {
course: PropTypes.shape({
slug: PropTypes.string.isRequired
}).isRequired,
remaining: PropTypes.array,
<|file_sep|>updated/app/assets/javascripts/components/overview/my_exercises/components/Header.jsx
export const Header = ({ completed = false, course, remaining = [], text }) => (
<header className="header">
<h3 className={completed ? 'completed' : ''}>
{ text }
{
remaining.length
? <small>{remaining.length} additional exercises remaining.</small>
: null
}
</h3>
<NavLink exact to={`/courses/${course.slug}/resources`} className="resources-link">
View all exercises
</NavLink>
</header>
);
Header.propTypes = {
course: PropTypes.shape({
slug: PropTypes.string.isRequired
}).isRequired,
remaining: PropTypes.array, | 219e06070c1c85cc893e1fbfb889d7e6abaaeee6 | app/assets/javascripts/components/overview/my_exercises/components/Header.jsx | app/assets/javascripts/components/overview/my_exercises/components/Header.jsx | JSX |
<|file_sep|>chapters/pereira.json.diff
original:
"cronopio",
updated:
"alsuga",
"aerendon",
"caal-15",
"h3ct0rjs",
"jdmorales92",
<|file_sep|>original/chapters/pereira.json
{
"name": "NodeSchool Pereira",
"location": "Pereira",
"country": "CO",
"region": "Latin America",
"organizers": [
"cronopio",
"pin3da",
"jhonber"
],
"repo": "http://github.com/nodeschool/pereira",
"website": "http://nodeschool.io/pereira",
"twitter": "PereiraJS_"
}
<|file_sep|>current/chapters/pereira.json
{
"name": "NodeSchool Pereira",
"location": "Pereira",
"country": "CO",
"region": "Latin America",
"organizers": [
"alsuga",
"aerendon",
"caal-15",
"h3ct0rjs",
"jdmorales92",
"pin3da",
"jhonber"
],
"repo": "http://github.com/nodeschool/pereira",
"website": "http://nodeschool.io/pereira",
"twitter": "PereiraJS_"
}
<|file_sep|>updated/chapters/pereira.json | {
"name": "NodeSchool Pereira",
"location": "Pereira",
"country": "CO",
"region": "Latin America",
"organizers": [
"alsuga",
"aerendon",
"caal-15",
"h3ct0rjs",
"jdmorales92",
"pin3da",
"xergioalex"
],
"repo": "http://github.com/nodeschool/pereira",
"website": "http://nodeschool.io/pereira",
"twitter": "PereiraJS_"
} | <|file_sep|>chapters/pereira.json.diff
original:
"cronopio",
updated:
"alsuga",
"aerendon",
"caal-15",
"h3ct0rjs",
"jdmorales92",
<|file_sep|>original/chapters/pereira.json
{
"name": "NodeSchool Pereira",
"location": "Pereira",
"country": "CO",
"region": "Latin America",
"organizers": [
"cronopio",
"pin3da",
"jhonber"
],
"repo": "http://github.com/nodeschool/pereira",
"website": "http://nodeschool.io/pereira",
"twitter": "PereiraJS_"
}
<|file_sep|>current/chapters/pereira.json
{
"name": "NodeSchool Pereira",
"location": "Pereira",
"country": "CO",
"region": "Latin America",
"organizers": [
"alsuga",
"aerendon",
"caal-15",
"h3ct0rjs",
"jdmorales92",
"pin3da",
"jhonber"
],
"repo": "http://github.com/nodeschool/pereira",
"website": "http://nodeschool.io/pereira",
"twitter": "PereiraJS_"
}
<|file_sep|>updated/chapters/pereira.json
{
"name": "NodeSchool Pereira",
"location": "Pereira",
"country": "CO",
"region": "Latin America",
"organizers": [
"alsuga",
"aerendon",
"caal-15",
"h3ct0rjs",
"jdmorales92",
"pin3da",
"xergioalex"
],
"repo": "http://github.com/nodeschool/pereira",
"website": "http://nodeschool.io/pereira",
"twitter": "PereiraJS_"
} | 25e9ac106be35ec9d916b29a240689d5508bbcc4 | chapters/pereira.json | chapters/pereira.json | JSON |
<|file_sep|>original/app/assets/stylesheets/votes.css
width: 100px;
height: 100px;
}
#winner-photo {
border: solid #9FE695 5px;
}
#loser-photo {
border: solid #E69595 5px;
}
.last-photo {
margin-top: 5px;
width: 90px;
height: 90px;
}
.prev-winner, .prev-loser {
text-align: center;
}
<|file_sep|>current/app/assets/stylesheets/votes.css
width: 100px;
height: 100px;
}
#winner-photo {
border: solid #9FE695 5px;
}
#loser-photo {
border: solid #E69595 5px;
}
.last-photo {
margin-top: 5px;
width: 90px;
height: 90px;
}
.prev-winner, .prev-loser {
text-align: center;
}
<|file_sep|>updated/app/assets/stylesheets/votes.css | width: 100px;
height: 100px;
}
#winner-photo {
border: solid #9FE695 5px;
}
#loser-photo {
border: solid #E69595 5px;
opacity: 0.5;
}
.last-photo {
margin-top: 5px;
width: 90px;
height: 90px;
}
.prev-winner, .prev-loser {
text-align: center; | <|file_sep|>original/app/assets/stylesheets/votes.css
width: 100px;
height: 100px;
}
#winner-photo {
border: solid #9FE695 5px;
}
#loser-photo {
border: solid #E69595 5px;
}
.last-photo {
margin-top: 5px;
width: 90px;
height: 90px;
}
.prev-winner, .prev-loser {
text-align: center;
}
<|file_sep|>current/app/assets/stylesheets/votes.css
width: 100px;
height: 100px;
}
#winner-photo {
border: solid #9FE695 5px;
}
#loser-photo {
border: solid #E69595 5px;
}
.last-photo {
margin-top: 5px;
width: 90px;
height: 90px;
}
.prev-winner, .prev-loser {
text-align: center;
}
<|file_sep|>updated/app/assets/stylesheets/votes.css
width: 100px;
height: 100px;
}
#winner-photo {
border: solid #9FE695 5px;
}
#loser-photo {
border: solid #E69595 5px;
opacity: 0.5;
}
.last-photo {
margin-top: 5px;
width: 90px;
height: 90px;
}
.prev-winner, .prev-loser {
text-align: center; | aba065171eda23b4125caf3676eb0f6c165c7906 | app/assets/stylesheets/votes.css | app/assets/stylesheets/votes.css | CSS |
<|file_sep|>original/maven-plugins/pom.xml
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
<modelVersion>4.0.0</modelVersion>
<parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>parent</artifactId>
<version>7-SNAPSHOT</version>
<relativePath>../parent/pom.xml</relativePath>
</parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>maven-plugins</artifactId>
<name>maven-plugins</name>
<packaging>pom</packaging>
<version>7-SNAPSHOT</version>
<description>Parent artifact for Vespa maven plugins.</description>
<url>http://yahoo.github.io/vespa</url>
<modules>
<module>../annotations</module>
<module>../bundle-plugin</module>
<module>../configgen</module>
<module>../config-class-plugin</module>
<|file_sep|>current/maven-plugins/pom.xml
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
<modelVersion>4.0.0</modelVersion>
<parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>parent</artifactId>
<version>7-SNAPSHOT</version>
<relativePath>../parent/pom.xml</relativePath>
</parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>maven-plugins</artifactId>
<name>maven-plugins</name>
<packaging>pom</packaging>
<version>7-SNAPSHOT</version>
<description>Parent artifact for Vespa maven plugins.</description>
<url>http://yahoo.github.io/vespa</url>
<modules>
<module>../annotations</module>
<module>../bundle-plugin</module>
<module>../configgen</module>
<module>../config-class-plugin</module>
<|file_sep|>updated/maven-plugins/pom.xml | <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
<modelVersion>4.0.0</modelVersion>
<parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>parent</artifactId>
<version>7-SNAPSHOT</version>
<relativePath>../parent/pom.xml</relativePath>
</parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>maven-plugins</artifactId>
<packaging>pom</packaging>
<version>7-SNAPSHOT</version>
<description>Parent artifact for Vespa maven plugins.</description>
<url>http://yahoo.github.io/vespa</url>
<modules>
<module>../annotations</module>
<module>../bundle-plugin</module>
<module>../configgen</module>
<module>../config-class-plugin</module>
<module>../vespa-application-maven-plugin</module> | <|file_sep|>original/maven-plugins/pom.xml
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
<modelVersion>4.0.0</modelVersion>
<parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>parent</artifactId>
<version>7-SNAPSHOT</version>
<relativePath>../parent/pom.xml</relativePath>
</parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>maven-plugins</artifactId>
<name>maven-plugins</name>
<packaging>pom</packaging>
<version>7-SNAPSHOT</version>
<description>Parent artifact for Vespa maven plugins.</description>
<url>http://yahoo.github.io/vespa</url>
<modules>
<module>../annotations</module>
<module>../bundle-plugin</module>
<module>../configgen</module>
<module>../config-class-plugin</module>
<|file_sep|>current/maven-plugins/pom.xml
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
<modelVersion>4.0.0</modelVersion>
<parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>parent</artifactId>
<version>7-SNAPSHOT</version>
<relativePath>../parent/pom.xml</relativePath>
</parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>maven-plugins</artifactId>
<name>maven-plugins</name>
<packaging>pom</packaging>
<version>7-SNAPSHOT</version>
<description>Parent artifact for Vespa maven plugins.</description>
<url>http://yahoo.github.io/vespa</url>
<modules>
<module>../annotations</module>
<module>../bundle-plugin</module>
<module>../configgen</module>
<module>../config-class-plugin</module>
<|file_sep|>updated/maven-plugins/pom.xml
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
<modelVersion>4.0.0</modelVersion>
<parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>parent</artifactId>
<version>7-SNAPSHOT</version>
<relativePath>../parent/pom.xml</relativePath>
</parent>
<groupId>com.yahoo.vespa</groupId>
<artifactId>maven-plugins</artifactId>
<packaging>pom</packaging>
<version>7-SNAPSHOT</version>
<description>Parent artifact for Vespa maven plugins.</description>
<url>http://yahoo.github.io/vespa</url>
<modules>
<module>../annotations</module>
<module>../bundle-plugin</module>
<module>../configgen</module>
<module>../config-class-plugin</module>
<module>../vespa-application-maven-plugin</module> | 6eb73a86086dcc40f76d0dc64b899af4a1fe5a58 | maven-plugins/pom.xml | maven-plugins/pom.xml | XML |
<|file_sep|>original/.circleci/config.yml
filters:
branches:
only:
- master
- maint
- ci/all
- ci/circle
matrix:
parameters:
os:
- linux
py:
- "2.7"
- "3.5"
- "3.6"
- "3.7"
- "3.8"
- "3.9"
mpi:
- "mpich"
- "openmpi"
<|file_sep|>current/.circleci/config.yml
filters:
branches:
only:
- master
- maint
- ci/all
- ci/circle
matrix:
parameters:
os:
- linux
py:
- "2.7"
- "3.5"
- "3.6"
- "3.7"
- "3.8"
- "3.9"
mpi:
- "mpich"
- "openmpi"
<|file_sep|>updated/.circleci/config.yml | - test:
filters:
branches:
only:
- master
- maint
- ci/all
- ci/circle
matrix:
parameters:
os:
- linux
py:
- "3.5"
- "3.6"
- "3.7"
- "3.8"
- "3.9"
mpi:
- "mpich"
- "openmpi" | <|file_sep|>original/.circleci/config.yml
filters:
branches:
only:
- master
- maint
- ci/all
- ci/circle
matrix:
parameters:
os:
- linux
py:
- "2.7"
- "3.5"
- "3.6"
- "3.7"
- "3.8"
- "3.9"
mpi:
- "mpich"
- "openmpi"
<|file_sep|>current/.circleci/config.yml
filters:
branches:
only:
- master
- maint
- ci/all
- ci/circle
matrix:
parameters:
os:
- linux
py:
- "2.7"
- "3.5"
- "3.6"
- "3.7"
- "3.8"
- "3.9"
mpi:
- "mpich"
- "openmpi"
<|file_sep|>updated/.circleci/config.yml
- test:
filters:
branches:
only:
- master
- maint
- ci/all
- ci/circle
matrix:
parameters:
os:
- linux
py:
- "3.5"
- "3.6"
- "3.7"
- "3.8"
- "3.9"
mpi:
- "mpich"
- "openmpi" | 0b04e77d77381192a8fd560f14e1020034a57b86 | .circleci/config.yml | .circleci/config.yml | YAML |
<|file_sep|>original/.travis.yml
language: objective-c
xcode_project: JenkinsiOS.xcodeproj
xcode_scheme: JenkinsiOSTests
osx_image: xcode8.3
xcode_sdk: iphonesimulator10.0
script:
- xcodebuild clean build test -project JenkinsiOS.xcodeproj -scheme JenkinsiOSTests -destination "platform=iOS Simulator,id=34FA4749-C467-4D45-9F8E-E31AEDDC39C2" -json
<|file_sep|>current/.travis.yml
language: objective-c
xcode_project: JenkinsiOS.xcodeproj
xcode_scheme: JenkinsiOSTests
osx_image: xcode8.3
xcode_sdk: iphonesimulator10.0
script:
- xcodebuild clean build test -project JenkinsiOS.xcodeproj -scheme JenkinsiOSTests -destination "platform=iOS Simulator,id=34FA4749-C467-4D45-9F8E-E31AEDDC39C2" -json
<|file_sep|>updated/.travis.yml | language: objective-c
xcode_project: JenkinsiOS.xcodeproj
xcode_scheme: JenkinsiOSTests
osx_image: xcode8.3
xcode_sdk: iphonesimulator10.0
script:
- xcodebuild clean build test -project JenkinsiOS.xcodeproj -scheme JenkinsiOSTests -destination "platform=iOS Simulator,id=85AF1F51-F2EB-4A4D-930B-0E5761511A45" -json | <|file_sep|>original/.travis.yml
language: objective-c
xcode_project: JenkinsiOS.xcodeproj
xcode_scheme: JenkinsiOSTests
osx_image: xcode8.3
xcode_sdk: iphonesimulator10.0
script:
- xcodebuild clean build test -project JenkinsiOS.xcodeproj -scheme JenkinsiOSTests -destination "platform=iOS Simulator,id=34FA4749-C467-4D45-9F8E-E31AEDDC39C2" -json
<|file_sep|>current/.travis.yml
language: objective-c
xcode_project: JenkinsiOS.xcodeproj
xcode_scheme: JenkinsiOSTests
osx_image: xcode8.3
xcode_sdk: iphonesimulator10.0
script:
- xcodebuild clean build test -project JenkinsiOS.xcodeproj -scheme JenkinsiOSTests -destination "platform=iOS Simulator,id=34FA4749-C467-4D45-9F8E-E31AEDDC39C2" -json
<|file_sep|>updated/.travis.yml
language: objective-c
xcode_project: JenkinsiOS.xcodeproj
xcode_scheme: JenkinsiOSTests
osx_image: xcode8.3
xcode_sdk: iphonesimulator10.0
script:
- xcodebuild clean build test -project JenkinsiOS.xcodeproj -scheme JenkinsiOSTests -destination "platform=iOS Simulator,id=85AF1F51-F2EB-4A4D-930B-0E5761511A45" -json | 0099d7d3f8999c1a4ca1e7803a9df607db5ad1ca | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/metadata/net.chilon.matt.teacup.txt
Build:1.0,1
commit=a544fdbe5c043f37952cc5624d937cd6212cc1ef
Build:1.1,2
commit=a288fbc41c
Build:1.2,3
commit=v1.2
Build:1.3,4
commit=v1.3
Build:1.5,6
commit=v1.5
Auto Update Mode:None
Update Check Mode:Tags
Current Version:1.6
Current Version Code:7
<|file_sep|>current/metadata/net.chilon.matt.teacup.txt
Build:1.0,1
commit=a544fdbe5c043f37952cc5624d937cd6212cc1ef
Build:1.1,2
commit=a288fbc41c
Build:1.2,3
commit=v1.2
Build:1.3,4
commit=v1.3
Build:1.5,6
commit=v1.5
Auto Update Mode:None
Update Check Mode:Tags
Current Version:1.6
Current Version Code:7
<|file_sep|>updated/metadata/net.chilon.matt.teacup.txt |
Build:1.1,2
commit=a288fbc41c
Build:1.2,3
commit=v1.2
Build:1.3,4
commit=v1.3
Build:1.5,6
commit=v1.5
Build:1.6,7
commit=v1.6
Auto Update Mode:None
Update Check Mode:Tags
Current Version:1.6
Current Version Code:7
| <|file_sep|>original/metadata/net.chilon.matt.teacup.txt
Build:1.0,1
commit=a544fdbe5c043f37952cc5624d937cd6212cc1ef
Build:1.1,2
commit=a288fbc41c
Build:1.2,3
commit=v1.2
Build:1.3,4
commit=v1.3
Build:1.5,6
commit=v1.5
Auto Update Mode:None
Update Check Mode:Tags
Current Version:1.6
Current Version Code:7
<|file_sep|>current/metadata/net.chilon.matt.teacup.txt
Build:1.0,1
commit=a544fdbe5c043f37952cc5624d937cd6212cc1ef
Build:1.1,2
commit=a288fbc41c
Build:1.2,3
commit=v1.2
Build:1.3,4
commit=v1.3
Build:1.5,6
commit=v1.5
Auto Update Mode:None
Update Check Mode:Tags
Current Version:1.6
Current Version Code:7
<|file_sep|>updated/metadata/net.chilon.matt.teacup.txt
Build:1.1,2
commit=a288fbc41c
Build:1.2,3
commit=v1.2
Build:1.3,4
commit=v1.3
Build:1.5,6
commit=v1.5
Build:1.6,7
commit=v1.6
Auto Update Mode:None
Update Check Mode:Tags
Current Version:1.6
Current Version Code:7
| 468f922d9cc4184213faa205c11062863518b8f3 | metadata/net.chilon.matt.teacup.txt | metadata/net.chilon.matt.teacup.txt | Text |
<|file_sep|>original/.travis.yml
sudo: false
language: python
python:
# https://docs.python.org/devguide/#status-of-python-branches
- "2.7"
- "3.3"
- "3.4"
- "3.5"
- "3.6"
- "pypy"
- "pypy3"
before_install: pip install codecov
install: pip install --upgrade -r dev_requirements.txt
script: PYTHONHASHSEED=random pytest
after_success: codecov
<|file_sep|>current/.travis.yml
sudo: false
language: python
python:
# https://docs.python.org/devguide/#status-of-python-branches
- "2.7"
- "3.3"
- "3.4"
- "3.5"
- "3.6"
- "pypy"
- "pypy3"
before_install: pip install codecov
install: pip install --upgrade -r dev_requirements.txt
script: PYTHONHASHSEED=random pytest
after_success: codecov
<|file_sep|>updated/.travis.yml | sudo: false
language: python
python:
# https://docs.python.org/devguide/#status-of-python-branches
- "2.7"
- "3.4"
- "3.5"
- "3.6"
- "pypy"
- "pypy3"
before_install: pip install codecov
install: pip install --upgrade -r dev_requirements.txt
script: PYTHONHASHSEED=random pytest
after_success: codecov | <|file_sep|>original/.travis.yml
sudo: false
language: python
python:
# https://docs.python.org/devguide/#status-of-python-branches
- "2.7"
- "3.3"
- "3.4"
- "3.5"
- "3.6"
- "pypy"
- "pypy3"
before_install: pip install codecov
install: pip install --upgrade -r dev_requirements.txt
script: PYTHONHASHSEED=random pytest
after_success: codecov
<|file_sep|>current/.travis.yml
sudo: false
language: python
python:
# https://docs.python.org/devguide/#status-of-python-branches
- "2.7"
- "3.3"
- "3.4"
- "3.5"
- "3.6"
- "pypy"
- "pypy3"
before_install: pip install codecov
install: pip install --upgrade -r dev_requirements.txt
script: PYTHONHASHSEED=random pytest
after_success: codecov
<|file_sep|>updated/.travis.yml
sudo: false
language: python
python:
# https://docs.python.org/devguide/#status-of-python-branches
- "2.7"
- "3.4"
- "3.5"
- "3.6"
- "pypy"
- "pypy3"
before_install: pip install codecov
install: pip install --upgrade -r dev_requirements.txt
script: PYTHONHASHSEED=random pytest
after_success: codecov | 2e3da5971845ab1999491d397f2b5eef8b5d3bc5 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/README.md
# moe-sounds
Moe sounds, moe problems. Check out the live [site.](http://www.moesounds.com)
## Documentation
For information on how to setup your own moe sounds, read the uncompleted [wiki.](https://github.com/NYPD/moe-sounds/wiki)
<|file_sep|>current/README.md
# moe-sounds
Moe sounds, moe problems. Check out the live [site.](http://www.moesounds.com)
## Documentation
For information on how to setup your own moe sounds, read the uncompleted [wiki.](https://github.com/NYPD/moe-sounds/wiki)
<|file_sep|>updated/README.md | <br>
<sup>(Logo by [Louis Wabnitz @ studio peep](http://studiopeep.com/))</sup>
# moe-sounds
Moe sounds, moe problems. Check out the live [site.](http://www.moesounds.com)
## Documentation
For information on how to setup your own moe sounds, read the uncompleted [wiki.](https://github.com/NYPD/moe-sounds/wiki) | <|file_sep|>original/README.md
# moe-sounds
Moe sounds, moe problems. Check out the live [site.](http://www.moesounds.com)
## Documentation
For information on how to setup your own moe sounds, read the uncompleted [wiki.](https://github.com/NYPD/moe-sounds/wiki)
<|file_sep|>current/README.md
# moe-sounds
Moe sounds, moe problems. Check out the live [site.](http://www.moesounds.com)
## Documentation
For information on how to setup your own moe sounds, read the uncompleted [wiki.](https://github.com/NYPD/moe-sounds/wiki)
<|file_sep|>updated/README.md
<br>
<sup>(Logo by [Louis Wabnitz @ studio peep](http://studiopeep.com/))</sup>
# moe-sounds
Moe sounds, moe problems. Check out the live [site.](http://www.moesounds.com)
## Documentation
For information on how to setup your own moe sounds, read the uncompleted [wiki.](https://github.com/NYPD/moe-sounds/wiki) | a1f87221db127550a7165d4dbc2edcb6b5a16716 | README.md | README.md | Markdown |
<|file_sep|>src/modules/client/client.native.js.diff
original:
updated:
import 'core-js/es6/object';
<|file_sep|>original/src/modules/client/client.native.js
/* @flow */
/* eslint-disable import/no-commonjs */
import 'core-js/es6/symbol';
import 'core-js/es6/array';
global.navigator.userAgent = 'react-native';
require('./client-base');
<|file_sep|>current/src/modules/client/client.native.js
/* @flow */
/* eslint-disable import/no-commonjs */
import 'core-js/es6/symbol';
import 'core-js/es6/object';
import 'core-js/es6/array';
global.navigator.userAgent = 'react-native';
require('./client-base');
<|file_sep|>updated/src/modules/client/client.native.js | /* @flow */
/* eslint-disable import/no-commonjs */
import 'core-js/es6/symbol';
import 'core-js/es6/object';
import 'core-js/es6/array';
import 'core-js/es6/function';
global.navigator.userAgent = 'react-native';
require('./client-base'); | <|file_sep|>src/modules/client/client.native.js.diff
original:
updated:
import 'core-js/es6/object';
<|file_sep|>original/src/modules/client/client.native.js
/* @flow */
/* eslint-disable import/no-commonjs */
import 'core-js/es6/symbol';
import 'core-js/es6/array';
global.navigator.userAgent = 'react-native';
require('./client-base');
<|file_sep|>current/src/modules/client/client.native.js
/* @flow */
/* eslint-disable import/no-commonjs */
import 'core-js/es6/symbol';
import 'core-js/es6/object';
import 'core-js/es6/array';
global.navigator.userAgent = 'react-native';
require('./client-base');
<|file_sep|>updated/src/modules/client/client.native.js
/* @flow */
/* eslint-disable import/no-commonjs */
import 'core-js/es6/symbol';
import 'core-js/es6/object';
import 'core-js/es6/array';
import 'core-js/es6/function';
global.navigator.userAgent = 'react-native';
require('./client-base'); | 2fa8ee8bbc29caf13e454868a620f2c2dc35f0f6 | src/modules/client/client.native.js | src/modules/client/client.native.js | JavaScript |
<|file_sep|>original/spec/javascripts/fixtures/emojis.rb
require 'spec_helper'
describe 'Emojis (JavaScript fixtures)' do
include JavaScriptFixturesHelpers
before(:all) do
clean_frontend_fixtures('emojis/')
end
it 'emojis/emojis.json' do |example|
# Copying the emojis.json from the public folder
fixture_file_name = File.expand_path('emojis/emojis.json', JavaScriptFixturesHelpers::FIXTURE_PATH)
FileUtils.mkdir_p(File.dirname(fixture_file_name))
FileUtils.cp(Rails.root.join('public/-/emojis/1/emojis.json'), fixture_file_name)
end
end
<|file_sep|>current/spec/javascripts/fixtures/emojis.rb
require 'spec_helper'
describe 'Emojis (JavaScript fixtures)' do
include JavaScriptFixturesHelpers
before(:all) do
clean_frontend_fixtures('emojis/')
end
it 'emojis/emojis.json' do |example|
# Copying the emojis.json from the public folder
fixture_file_name = File.expand_path('emojis/emojis.json', JavaScriptFixturesHelpers::FIXTURE_PATH)
FileUtils.mkdir_p(File.dirname(fixture_file_name))
FileUtils.cp(Rails.root.join('public/-/emojis/1/emojis.json'), fixture_file_name)
end
end
<|file_sep|>updated/spec/javascripts/fixtures/emojis.rb | require 'spec_helper'
describe 'Emojis (JavaScript fixtures)' do
include JavaScriptFixturesHelpers
before(:all) do
clean_frontend_fixtures('emojis/')
end
it 'emojis/emojis.json' do |example|
JavaScriptFixturesHelpers::FIXTURE_PATHS.each do |fixture_path|
# Copying the emojis.json from the public folder
fixture_file_name = File.expand_path('emojis/emojis.json', fixture_path)
FileUtils.mkdir_p(File.dirname(fixture_file_name))
FileUtils.cp(Rails.root.join('public/-/emojis/1/emojis.json'), fixture_file_name)
end
end
end | <|file_sep|>original/spec/javascripts/fixtures/emojis.rb
require 'spec_helper'
describe 'Emojis (JavaScript fixtures)' do
include JavaScriptFixturesHelpers
before(:all) do
clean_frontend_fixtures('emojis/')
end
it 'emojis/emojis.json' do |example|
# Copying the emojis.json from the public folder
fixture_file_name = File.expand_path('emojis/emojis.json', JavaScriptFixturesHelpers::FIXTURE_PATH)
FileUtils.mkdir_p(File.dirname(fixture_file_name))
FileUtils.cp(Rails.root.join('public/-/emojis/1/emojis.json'), fixture_file_name)
end
end
<|file_sep|>current/spec/javascripts/fixtures/emojis.rb
require 'spec_helper'
describe 'Emojis (JavaScript fixtures)' do
include JavaScriptFixturesHelpers
before(:all) do
clean_frontend_fixtures('emojis/')
end
it 'emojis/emojis.json' do |example|
# Copying the emojis.json from the public folder
fixture_file_name = File.expand_path('emojis/emojis.json', JavaScriptFixturesHelpers::FIXTURE_PATH)
FileUtils.mkdir_p(File.dirname(fixture_file_name))
FileUtils.cp(Rails.root.join('public/-/emojis/1/emojis.json'), fixture_file_name)
end
end
<|file_sep|>updated/spec/javascripts/fixtures/emojis.rb
require 'spec_helper'
describe 'Emojis (JavaScript fixtures)' do
include JavaScriptFixturesHelpers
before(:all) do
clean_frontend_fixtures('emojis/')
end
it 'emojis/emojis.json' do |example|
JavaScriptFixturesHelpers::FIXTURE_PATHS.each do |fixture_path|
# Copying the emojis.json from the public folder
fixture_file_name = File.expand_path('emojis/emojis.json', fixture_path)
FileUtils.mkdir_p(File.dirname(fixture_file_name))
FileUtils.cp(Rails.root.join('public/-/emojis/1/emojis.json'), fixture_file_name)
end
end
end | 42df6c854dbf14eb0eb3fad839b93d3b3d9649f7 | spec/javascripts/fixtures/emojis.rb | spec/javascripts/fixtures/emojis.rb | Ruby |
<|file_sep|>original/requirements-dev.txt
coverage==4.2
docker-py==1.10.6
flake8==3.0.4
ipdb==0.10.1
ipython==5.1.0
pyodbc==3.0.10
pytest==3.0.4
pytest-cov==2.4.0
pytest-sugar==0.7.1
sphinx==1.4.8
sphinxcontrib-asyncio==0.2.0
uvloop==0.6.5
<|file_sep|>current/requirements-dev.txt
coverage==4.2
docker-py==1.10.6
flake8==3.0.4
ipdb==0.10.1
ipython==5.1.0
pyodbc==3.0.10
pytest==3.0.4
pytest-cov==2.4.0
pytest-sugar==0.7.1
sphinx==1.4.8
sphinxcontrib-asyncio==0.2.0
uvloop==0.6.5
<|file_sep|>updated/requirements-dev.txt | coverage==4.2
docker-py==1.10.6
flake8==3.1.0
ipdb==0.10.1
ipython==5.1.0
pyodbc==3.0.10
pytest==3.0.4
pytest-cov==2.4.0
pytest-sugar==0.7.1
sphinx==1.4.8
sphinxcontrib-asyncio==0.2.0
uvloop==0.6.5 | <|file_sep|>original/requirements-dev.txt
coverage==4.2
docker-py==1.10.6
flake8==3.0.4
ipdb==0.10.1
ipython==5.1.0
pyodbc==3.0.10
pytest==3.0.4
pytest-cov==2.4.0
pytest-sugar==0.7.1
sphinx==1.4.8
sphinxcontrib-asyncio==0.2.0
uvloop==0.6.5
<|file_sep|>current/requirements-dev.txt
coverage==4.2
docker-py==1.10.6
flake8==3.0.4
ipdb==0.10.1
ipython==5.1.0
pyodbc==3.0.10
pytest==3.0.4
pytest-cov==2.4.0
pytest-sugar==0.7.1
sphinx==1.4.8
sphinxcontrib-asyncio==0.2.0
uvloop==0.6.5
<|file_sep|>updated/requirements-dev.txt
coverage==4.2
docker-py==1.10.6
flake8==3.1.0
ipdb==0.10.1
ipython==5.1.0
pyodbc==3.0.10
pytest==3.0.4
pytest-cov==2.4.0
pytest-sugar==0.7.1
sphinx==1.4.8
sphinxcontrib-asyncio==0.2.0
uvloop==0.6.5 | 98c16bcea58853ecf83bb9247fc4ac93ca9ec0b4 | requirements-dev.txt | requirements-dev.txt | Text |
<|file_sep|>original/_includes/blog-masthead.html
<div class="blog-masthead">
<div class="container">
<nav class="blog-nav">
<a class="blog-nav-item {{ active.home }}" href="/">Home</a>
<a class="blog-nav-item" href="http://wiki.hacksburg.org">Wiki</a>
<a class="blog-nav-item {{ active.about }}" href="/about.html">About</a>
</nav>
</div>
</div>
<|file_sep|>current/_includes/blog-masthead.html
<div class="blog-masthead">
<div class="container">
<nav class="blog-nav">
<a class="blog-nav-item {{ active.home }}" href="/">Home</a>
<a class="blog-nav-item" href="http://wiki.hacksburg.org">Wiki</a>
<a class="blog-nav-item {{ active.about }}" href="/about.html">About</a>
</nav>
</div>
</div>
<|file_sep|>updated/_includes/blog-masthead.html | <div class="blog-masthead">
<div class="container">
<nav class="blog-nav">
<a class="blog-nav-item {{ active.home }}" href="/">Home</a>
<a class="blog-nav-item" href="http://wiki.hacksburg.org">Wiki</a>
<a class="blog-nav-item {{ active.about }}" href="/about.html">About</a>
<a class="blog-nav-item" href="mailto:board@hacksburg.org">Email</a>
</nav>
</div>
</div> | <|file_sep|>original/_includes/blog-masthead.html
<div class="blog-masthead">
<div class="container">
<nav class="blog-nav">
<a class="blog-nav-item {{ active.home }}" href="/">Home</a>
<a class="blog-nav-item" href="http://wiki.hacksburg.org">Wiki</a>
<a class="blog-nav-item {{ active.about }}" href="/about.html">About</a>
</nav>
</div>
</div>
<|file_sep|>current/_includes/blog-masthead.html
<div class="blog-masthead">
<div class="container">
<nav class="blog-nav">
<a class="blog-nav-item {{ active.home }}" href="/">Home</a>
<a class="blog-nav-item" href="http://wiki.hacksburg.org">Wiki</a>
<a class="blog-nav-item {{ active.about }}" href="/about.html">About</a>
</nav>
</div>
</div>
<|file_sep|>updated/_includes/blog-masthead.html
<div class="blog-masthead">
<div class="container">
<nav class="blog-nav">
<a class="blog-nav-item {{ active.home }}" href="/">Home</a>
<a class="blog-nav-item" href="http://wiki.hacksburg.org">Wiki</a>
<a class="blog-nav-item {{ active.about }}" href="/about.html">About</a>
<a class="blog-nav-item" href="mailto:board@hacksburg.org">Email</a>
</nav>
</div>
</div> | 8a7f24cc0c5dcd5f3dc2d526f40ab3550715007c | _includes/blog-masthead.html | _includes/blog-masthead.html | HTML |
<|file_sep|>app/views/root/_transaction_cross_domain_analytics.html.erb.diff
original:
(function(i,s,o,g,r,a,m){i['GoogleAnalyticsObject']=r;i[r]=i[r]||function(){
(i[r].q=i[r].q||[]).push(arguments)},i[r].l=1*new Date();a=s.createElement(o),
m=s.getElementsByTagName(o)[0];a.async=1;a.src=g;m.parentNode.insertBefore(a,m)
})(window,document,'script','//www.google-analytics.com/analytics.js','ga');
updated:
(function(i,s,o,g,r,a,m){
i['GoogleAnalyticsObject']=r;i[r]=i[r]||function(){
(i[r].q=i[r].q||[]).push(arguments)},i[r].l=1*new Date();
})(window,document,'script','//www.google-analytics.com/analytics.js','ga');
<|file_sep|>app/views/root/_transaction_cross_domain_analytics.html.erb.diff
original:
ga('create', '<%= ga_account_code %>');
updated:
ga('create', '<%= ga_account_code %>', 'auto', {'name': 'transactionTracker'});
<|file_sep|>original/app/views/root/_transaction_cross_domain_analytics.html.erb
<script id="transaction_cross_domain_analytics">
(function(i,s,o,g,r,a,m){i['GoogleAnalyticsObject']=r;i[r]=i[r]||function(){
(i[r].q=i[r].q||[]).push(arguments)},i[r].l=1*new Date();a=s.createElement(o),
m=s.getElementsByTagName(o)[0];a.async=1;a.src=g;m.parentNode.insertBefore(a,m)
})(window,document,'script','//www.google-analytics.com/analytics.js','ga');
ga('create', '<%= ga_account_code %>');
// Load the plugin.
ga('require', 'linker');
// Define which domains to autoLink.
ga('linker:autoLink', ['service.gov.uk']);
ga('send', 'pageview');
</script>
<|file_sep|>current/app/views/root/_transaction_cross_domain_analytics.html.erb
<script id="transaction_cross_domain_analytics">
(function(i,s,o,g,r,a,m){
i['GoogleAnalyticsObject']=r;i[r]=i[r]||function(){
(i[r].q=i[r].q||[]).push(arguments)},i[r].l=1*new Date();
})(window,document,'script','//www.google-analytics.com/analytics.js','ga');
ga('create', '<%= ga_account_code %>', 'auto', {'name': 'transactionTracker'});
// Load the plugin.
ga('require', 'linker');
// Define which domains to autoLink.
ga('linker:autoLink', ['service.gov.uk']);
ga('send', 'pageview');
</script>
<|file_sep|>updated/app/views/root/_transaction_cross_domain_analytics.html.erb | <script id="transaction_cross_domain_analytics">
(function(i,s,o,g,r,a,m){
i['GoogleAnalyticsObject']=r;i[r]=i[r]||function(){
(i[r].q=i[r].q||[]).push(arguments)},i[r].l=1*new Date();
})(window,document,'script','//www.google-analytics.com/analytics.js','ga');
ga('create', '<%= ga_account_code %>', 'auto', {'name': 'transactionTracker'});
// Load the plugin.
ga('require', 'linker');
// Define which domains to autoLink.
ga('linker:autoLink', ['service.gov.uk']);
ga('transactionTracker.send', 'pageview');
</script> | <|file_sep|>app/views/root/_transaction_cross_domain_analytics.html.erb.diff
original:
(function(i,s,o,g,r,a,m){i['GoogleAnalyticsObject']=r;i[r]=i[r]||function(){
(i[r].q=i[r].q||[]).push(arguments)},i[r].l=1*new Date();a=s.createElement(o),
m=s.getElementsByTagName(o)[0];a.async=1;a.src=g;m.parentNode.insertBefore(a,m)
})(window,document,'script','//www.google-analytics.com/analytics.js','ga');
updated:
(function(i,s,o,g,r,a,m){
i['GoogleAnalyticsObject']=r;i[r]=i[r]||function(){
(i[r].q=i[r].q||[]).push(arguments)},i[r].l=1*new Date();
})(window,document,'script','//www.google-analytics.com/analytics.js','ga');
<|file_sep|>app/views/root/_transaction_cross_domain_analytics.html.erb.diff
original:
ga('create', '<%= ga_account_code %>');
updated:
ga('create', '<%= ga_account_code %>', 'auto', {'name': 'transactionTracker'});
<|file_sep|>original/app/views/root/_transaction_cross_domain_analytics.html.erb
<script id="transaction_cross_domain_analytics">
(function(i,s,o,g,r,a,m){i['GoogleAnalyticsObject']=r;i[r]=i[r]||function(){
(i[r].q=i[r].q||[]).push(arguments)},i[r].l=1*new Date();a=s.createElement(o),
m=s.getElementsByTagName(o)[0];a.async=1;a.src=g;m.parentNode.insertBefore(a,m)
})(window,document,'script','//www.google-analytics.com/analytics.js','ga');
ga('create', '<%= ga_account_code %>');
// Load the plugin.
ga('require', 'linker');
// Define which domains to autoLink.
ga('linker:autoLink', ['service.gov.uk']);
ga('send', 'pageview');
</script>
<|file_sep|>current/app/views/root/_transaction_cross_domain_analytics.html.erb
<script id="transaction_cross_domain_analytics">
(function(i,s,o,g,r,a,m){
i['GoogleAnalyticsObject']=r;i[r]=i[r]||function(){
(i[r].q=i[r].q||[]).push(arguments)},i[r].l=1*new Date();
})(window,document,'script','//www.google-analytics.com/analytics.js','ga');
ga('create', '<%= ga_account_code %>', 'auto', {'name': 'transactionTracker'});
// Load the plugin.
ga('require', 'linker');
// Define which domains to autoLink.
ga('linker:autoLink', ['service.gov.uk']);
ga('send', 'pageview');
</script>
<|file_sep|>updated/app/views/root/_transaction_cross_domain_analytics.html.erb
<script id="transaction_cross_domain_analytics">
(function(i,s,o,g,r,a,m){
i['GoogleAnalyticsObject']=r;i[r]=i[r]||function(){
(i[r].q=i[r].q||[]).push(arguments)},i[r].l=1*new Date();
})(window,document,'script','//www.google-analytics.com/analytics.js','ga');
ga('create', '<%= ga_account_code %>', 'auto', {'name': 'transactionTracker'});
// Load the plugin.
ga('require', 'linker');
// Define which domains to autoLink.
ga('linker:autoLink', ['service.gov.uk']);
ga('transactionTracker.send', 'pageview');
</script> | 37ae3261062bccf164fed1b88f2bdff9edecc1fb | app/views/root/_transaction_cross_domain_analytics.html.erb | app/views/root/_transaction_cross_domain_analytics.html.erb | HTML+ERB |
<|file_sep|>original/requirements.txt
PyYAML
websockets
arrow
# Stability of this package is backed by piqueserver
git+git://github.com/godwhoa/pytimeparse
<|file_sep|>current/requirements.txt
PyYAML
websockets
arrow
# Stability of this package is backed by piqueserver
git+git://github.com/godwhoa/pytimeparse
<|file_sep|>updated/requirements.txt | PyYAML
websockets
arrow
pytimeparse-plus | <|file_sep|>original/requirements.txt
PyYAML
websockets
arrow
# Stability of this package is backed by piqueserver
git+git://github.com/godwhoa/pytimeparse
<|file_sep|>current/requirements.txt
PyYAML
websockets
arrow
# Stability of this package is backed by piqueserver
git+git://github.com/godwhoa/pytimeparse
<|file_sep|>updated/requirements.txt
PyYAML
websockets
arrow
pytimeparse-plus | 9796f0b9a8eb422dbf16394023b580b6d960f8bd | requirements.txt | requirements.txt | Text |
<|file_sep|>original/infra/gcp/terraform/k8s-infra-oci-proxy-prod/monitoring.tf
display_name = "k8s-infra-alerts@kubernetes.io"
project = google_project.project.project_id
type = "email"
labels = {
email_address = "k8s-infra-alerts@kubernetes.io"
}
}
module "alerts" {
project_id = google_project.project.project_id
source = "../modules/monitoring/uptime-alert"
documentation_text = "${local.domain} is down"
domain = local.domain
notification_channels = [
# Manually created. Monitoring channels can't be created with Terraform.
# See: https://github.com/hashicorp/terraform-provider-google/issues/1134
"${google_project.project.id}/notificationChannels/15334306215710275143",
google_monitoring_notification_channel.emails.name,
]
}
<|file_sep|>current/infra/gcp/terraform/k8s-infra-oci-proxy-prod/monitoring.tf
display_name = "k8s-infra-alerts@kubernetes.io"
project = google_project.project.project_id
type = "email"
labels = {
email_address = "k8s-infra-alerts@kubernetes.io"
}
}
module "alerts" {
project_id = google_project.project.project_id
source = "../modules/monitoring/uptime-alert"
documentation_text = "${local.domain} is down"
domain = local.domain
notification_channels = [
# Manually created. Monitoring channels can't be created with Terraform.
# See: https://github.com/hashicorp/terraform-provider-google/issues/1134
"${google_project.project.id}/notificationChannels/15334306215710275143",
google_monitoring_notification_channel.emails.name,
]
}
<|file_sep|>updated/infra/gcp/terraform/k8s-infra-oci-proxy-prod/monitoring.tf | display_name = "k8s-infra-alerts@kubernetes.io"
project = google_project.project.project_id
type = "email"
labels = {
email_address = "k8s-infra-alerts@kubernetes.io"
}
}
module "alerts" {
project_id = google_project.project.project_id
source = "../modules/monitoring/uptime-alert"
documentation_text = "${var.domain} is down"
domain = var.domain
notification_channels = [
# Manually created. Monitoring channels can't be created with Terraform.
# See: https://github.com/hashicorp/terraform-provider-google/issues/1134
"${google_project.project.id}/notificationChannels/15334306215710275143",
google_monitoring_notification_channel.emails.name,
]
} | <|file_sep|>original/infra/gcp/terraform/k8s-infra-oci-proxy-prod/monitoring.tf
display_name = "k8s-infra-alerts@kubernetes.io"
project = google_project.project.project_id
type = "email"
labels = {
email_address = "k8s-infra-alerts@kubernetes.io"
}
}
module "alerts" {
project_id = google_project.project.project_id
source = "../modules/monitoring/uptime-alert"
documentation_text = "${local.domain} is down"
domain = local.domain
notification_channels = [
# Manually created. Monitoring channels can't be created with Terraform.
# See: https://github.com/hashicorp/terraform-provider-google/issues/1134
"${google_project.project.id}/notificationChannels/15334306215710275143",
google_monitoring_notification_channel.emails.name,
]
}
<|file_sep|>current/infra/gcp/terraform/k8s-infra-oci-proxy-prod/monitoring.tf
display_name = "k8s-infra-alerts@kubernetes.io"
project = google_project.project.project_id
type = "email"
labels = {
email_address = "k8s-infra-alerts@kubernetes.io"
}
}
module "alerts" {
project_id = google_project.project.project_id
source = "../modules/monitoring/uptime-alert"
documentation_text = "${local.domain} is down"
domain = local.domain
notification_channels = [
# Manually created. Monitoring channels can't be created with Terraform.
# See: https://github.com/hashicorp/terraform-provider-google/issues/1134
"${google_project.project.id}/notificationChannels/15334306215710275143",
google_monitoring_notification_channel.emails.name,
]
}
<|file_sep|>updated/infra/gcp/terraform/k8s-infra-oci-proxy-prod/monitoring.tf
display_name = "k8s-infra-alerts@kubernetes.io"
project = google_project.project.project_id
type = "email"
labels = {
email_address = "k8s-infra-alerts@kubernetes.io"
}
}
module "alerts" {
project_id = google_project.project.project_id
source = "../modules/monitoring/uptime-alert"
documentation_text = "${var.domain} is down"
domain = var.domain
notification_channels = [
# Manually created. Monitoring channels can't be created with Terraform.
# See: https://github.com/hashicorp/terraform-provider-google/issues/1134
"${google_project.project.id}/notificationChannels/15334306215710275143",
google_monitoring_notification_channel.emails.name,
]
} | 126414daeb1422987c56b5cf2ef1154baed5f4bb | infra/gcp/terraform/k8s-infra-oci-proxy-prod/monitoring.tf | infra/gcp/terraform/k8s-infra-oci-proxy-prod/monitoring.tf | HCL |
<|file_sep|>original/app/css/_select.scss
.select-prefix {
display: inline-block;
width: 2em;
&.selected:before {
margin-left: 0.75em;
content: "\2713 ";
}
}
.selectable button {
text-align: left;
display: block;
width: 100%;
padding: 0.5em;
padding-left: 0;
font-size: 16px;
border: 0;
background: $selectBackground;
}
<|file_sep|>current/app/css/_select.scss
.select-prefix {
display: inline-block;
width: 2em;
&.selected:before {
margin-left: 0.75em;
content: "\2713 ";
}
}
.selectable button {
text-align: left;
display: block;
width: 100%;
padding: 0.5em;
padding-left: 0;
font-size: 16px;
border: 0;
background: $selectBackground;
}
<|file_sep|>updated/app/css/_select.scss | .select-prefix {
display: inline-block;
width: 2em;
&.selected:before {
margin-left: 0.75em;
content: "\2713 ";
}
}
.selectable button {
text-align: left;
display: block;
width: 100%;
padding: 0.5em;
padding-left: 0;
padding-right: 1.25em;
font-size: 16px;
border: 0;
background: $selectBackground;
} | <|file_sep|>original/app/css/_select.scss
.select-prefix {
display: inline-block;
width: 2em;
&.selected:before {
margin-left: 0.75em;
content: "\2713 ";
}
}
.selectable button {
text-align: left;
display: block;
width: 100%;
padding: 0.5em;
padding-left: 0;
font-size: 16px;
border: 0;
background: $selectBackground;
}
<|file_sep|>current/app/css/_select.scss
.select-prefix {
display: inline-block;
width: 2em;
&.selected:before {
margin-left: 0.75em;
content: "\2713 ";
}
}
.selectable button {
text-align: left;
display: block;
width: 100%;
padding: 0.5em;
padding-left: 0;
font-size: 16px;
border: 0;
background: $selectBackground;
}
<|file_sep|>updated/app/css/_select.scss
.select-prefix {
display: inline-block;
width: 2em;
&.selected:before {
margin-left: 0.75em;
content: "\2713 ";
}
}
.selectable button {
text-align: left;
display: block;
width: 100%;
padding: 0.5em;
padding-left: 0;
padding-right: 1.25em;
font-size: 16px;
border: 0;
background: $selectBackground;
} | 696c8f91973bfded8787680c12808a07d1827f07 | app/css/_select.scss | app/css/_select.scss | SCSS |
<|file_sep|>original/src/util/libbb/CMakeLists.txt
SET(_SRC
bb_utils bb_alias
bb_core bb_simple
bb_sha1 bb_tools)
INCLUDE_DIRECTORIES(${CMAKE_SOURCE_DIR}/src/util/libbb)
ADD_LIBRARY(bb ${_SRC})
TARGET_LINK_LIBRARIES(bb pthread)
ADD_SUBDIRECTORY(bbtools)
IF(BUILD_SCRIPT_BINDING)
ADD_SUBDIRECTORY(scripting)
ENDIF(BUILD_SCRIPT_BINDING)
<|file_sep|>current/src/util/libbb/CMakeLists.txt
SET(_SRC
bb_utils bb_alias
bb_core bb_simple
bb_sha1 bb_tools)
INCLUDE_DIRECTORIES(${CMAKE_SOURCE_DIR}/src/util/libbb)
ADD_LIBRARY(bb ${_SRC})
TARGET_LINK_LIBRARIES(bb pthread)
ADD_SUBDIRECTORY(bbtools)
IF(BUILD_SCRIPT_BINDING)
ADD_SUBDIRECTORY(scripting)
ENDIF(BUILD_SCRIPT_BINDING)
<|file_sep|>updated/src/util/libbb/CMakeLists.txt | SET(_SRC
bb_utils bb_alias
bb_core bb_simple
bb_sha1 bb_tools bb_core_sysv.c)
INCLUDE_DIRECTORIES(${CMAKE_SOURCE_DIR}/src/util/libbb)
ADD_LIBRARY(bb ${_SRC})
TARGET_LINK_LIBRARIES(bb pthread)
ADD_SUBDIRECTORY(bbtools)
IF(BUILD_SCRIPT_BINDING)
ADD_SUBDIRECTORY(scripting)
ENDIF(BUILD_SCRIPT_BINDING)
| <|file_sep|>original/src/util/libbb/CMakeLists.txt
SET(_SRC
bb_utils bb_alias
bb_core bb_simple
bb_sha1 bb_tools)
INCLUDE_DIRECTORIES(${CMAKE_SOURCE_DIR}/src/util/libbb)
ADD_LIBRARY(bb ${_SRC})
TARGET_LINK_LIBRARIES(bb pthread)
ADD_SUBDIRECTORY(bbtools)
IF(BUILD_SCRIPT_BINDING)
ADD_SUBDIRECTORY(scripting)
ENDIF(BUILD_SCRIPT_BINDING)
<|file_sep|>current/src/util/libbb/CMakeLists.txt
SET(_SRC
bb_utils bb_alias
bb_core bb_simple
bb_sha1 bb_tools)
INCLUDE_DIRECTORIES(${CMAKE_SOURCE_DIR}/src/util/libbb)
ADD_LIBRARY(bb ${_SRC})
TARGET_LINK_LIBRARIES(bb pthread)
ADD_SUBDIRECTORY(bbtools)
IF(BUILD_SCRIPT_BINDING)
ADD_SUBDIRECTORY(scripting)
ENDIF(BUILD_SCRIPT_BINDING)
<|file_sep|>updated/src/util/libbb/CMakeLists.txt
SET(_SRC
bb_utils bb_alias
bb_core bb_simple
bb_sha1 bb_tools bb_core_sysv.c)
INCLUDE_DIRECTORIES(${CMAKE_SOURCE_DIR}/src/util/libbb)
ADD_LIBRARY(bb ${_SRC})
TARGET_LINK_LIBRARIES(bb pthread)
ADD_SUBDIRECTORY(bbtools)
IF(BUILD_SCRIPT_BINDING)
ADD_SUBDIRECTORY(scripting)
ENDIF(BUILD_SCRIPT_BINDING)
| 60adeb7a0d19eb4730abdd344d7e48c89155a20e | src/util/libbb/CMakeLists.txt | src/util/libbb/CMakeLists.txt | Text |
<|file_sep|>original/mungegithub/submit-queue/deployment/community/configmap.yaml
# basic config options.
http-cache-dir: /cache/httpcache
organization: kubernetes
project: community
# Make sure approval-handler and blunderbuss run before submit-queue.
# Otherwise it's going to take an extra-cycle to detect the label change.
# Run blunderbuss before approval-handler, so that we can suggest approvers
# based on assigned reviewer.
pr-mungers: blunderbuss, submit-queue, needs-rebase, sig-mention-handler, lgtm-after-commit, comment-deleter
state: open
token-file: /etc/secret-volume/token
period: 5m
repo-dir: /gitrepos
github-key-file: /etc/hook-secret-volume/secret
# status contexts options.
protected-branches-extra-contexts: "cla/linuxfoundation"
required-retest-contexts: ""
# submit-queue options.
protected-branches: "master"
<|file_sep|>current/mungegithub/submit-queue/deployment/community/configmap.yaml
# basic config options.
http-cache-dir: /cache/httpcache
organization: kubernetes
project: community
# Make sure approval-handler and blunderbuss run before submit-queue.
# Otherwise it's going to take an extra-cycle to detect the label change.
# Run blunderbuss before approval-handler, so that we can suggest approvers
# based on assigned reviewer.
pr-mungers: blunderbuss, submit-queue, needs-rebase, sig-mention-handler, lgtm-after-commit, comment-deleter
state: open
token-file: /etc/secret-volume/token
period: 5m
repo-dir: /gitrepos
github-key-file: /etc/hook-secret-volume/secret
# status contexts options.
protected-branches-extra-contexts: "cla/linuxfoundation"
required-retest-contexts: ""
# submit-queue options.
protected-branches: "master"
<|file_sep|>updated/mungegithub/submit-queue/deployment/community/configmap.yaml | # basic config options.
http-cache-dir: /cache/httpcache
organization: kubernetes
project: community
# Make sure approval-handler and blunderbuss run before submit-queue.
# Otherwise it's going to take an extra-cycle to detect the label change.
# Run blunderbuss before approval-handler, so that we can suggest approvers
# based on assigned reviewer.
pr-mungers: blunderbuss, submit-queue, needs-rebase, sig-mention-handler, lgtm-after-commit, comment-deleter, close-stale
state: open
token-file: /etc/secret-volume/token
period: 5m
repo-dir: /gitrepos
github-key-file: /etc/hook-secret-volume/secret
# status contexts options.
protected-branches-extra-contexts: "cla/linuxfoundation"
required-retest-contexts: ""
# submit-queue options.
protected-branches: "master" | <|file_sep|>original/mungegithub/submit-queue/deployment/community/configmap.yaml
# basic config options.
http-cache-dir: /cache/httpcache
organization: kubernetes
project: community
# Make sure approval-handler and blunderbuss run before submit-queue.
# Otherwise it's going to take an extra-cycle to detect the label change.
# Run blunderbuss before approval-handler, so that we can suggest approvers
# based on assigned reviewer.
pr-mungers: blunderbuss, submit-queue, needs-rebase, sig-mention-handler, lgtm-after-commit, comment-deleter
state: open
token-file: /etc/secret-volume/token
period: 5m
repo-dir: /gitrepos
github-key-file: /etc/hook-secret-volume/secret
# status contexts options.
protected-branches-extra-contexts: "cla/linuxfoundation"
required-retest-contexts: ""
# submit-queue options.
protected-branches: "master"
<|file_sep|>current/mungegithub/submit-queue/deployment/community/configmap.yaml
# basic config options.
http-cache-dir: /cache/httpcache
organization: kubernetes
project: community
# Make sure approval-handler and blunderbuss run before submit-queue.
# Otherwise it's going to take an extra-cycle to detect the label change.
# Run blunderbuss before approval-handler, so that we can suggest approvers
# based on assigned reviewer.
pr-mungers: blunderbuss, submit-queue, needs-rebase, sig-mention-handler, lgtm-after-commit, comment-deleter
state: open
token-file: /etc/secret-volume/token
period: 5m
repo-dir: /gitrepos
github-key-file: /etc/hook-secret-volume/secret
# status contexts options.
protected-branches-extra-contexts: "cla/linuxfoundation"
required-retest-contexts: ""
# submit-queue options.
protected-branches: "master"
<|file_sep|>updated/mungegithub/submit-queue/deployment/community/configmap.yaml
# basic config options.
http-cache-dir: /cache/httpcache
organization: kubernetes
project: community
# Make sure approval-handler and blunderbuss run before submit-queue.
# Otherwise it's going to take an extra-cycle to detect the label change.
# Run blunderbuss before approval-handler, so that we can suggest approvers
# based on assigned reviewer.
pr-mungers: blunderbuss, submit-queue, needs-rebase, sig-mention-handler, lgtm-after-commit, comment-deleter, close-stale
state: open
token-file: /etc/secret-volume/token
period: 5m
repo-dir: /gitrepos
github-key-file: /etc/hook-secret-volume/secret
# status contexts options.
protected-branches-extra-contexts: "cla/linuxfoundation"
required-retest-contexts: ""
# submit-queue options.
protected-branches: "master" | 6ca06d79156264123dc0dc962924326df32e7152 | mungegithub/submit-queue/deployment/community/configmap.yaml | mungegithub/submit-queue/deployment/community/configmap.yaml | YAML |
<|file_sep|>src/config_app.erl.diff
original:
updated:
DefaultExists = lists:filter(fun filelib:is_file/1, Default),
<|file_sep|>src/config_app.erl.diff
original:
Default;
updated:
DefaultExists;
<|file_sep|>original/src/config_app.erl
%% ===================================================================
%% Application callbacks
%% ===================================================================
start(_StartType, _StartArgs) ->
config_sup:start_link(get_ini_files()).
stop(_State) ->
ok.
get_ini_files() ->
Etc = filename:join(code:root_dir(), "etc"),
Default = [filename:join(Etc,"default.ini"), filename:join(Etc,"local.ini")],
case init:get_argument(couch_ini) of
error ->
Default;
{ok, [[]]} ->
Default;
{ok, [Values]} ->
Values
end.
<|file_sep|>current/src/config_app.erl
%% Application callbacks
%% ===================================================================
start(_StartType, _StartArgs) ->
config_sup:start_link(get_ini_files()).
stop(_State) ->
ok.
get_ini_files() ->
Etc = filename:join(code:root_dir(), "etc"),
Default = [filename:join(Etc,"default.ini"), filename:join(Etc,"local.ini")],
DefaultExists = lists:filter(fun filelib:is_file/1, Default),
case init:get_argument(couch_ini) of
error ->
DefaultExists;
{ok, [[]]} ->
Default;
{ok, [Values]} ->
Values
end.
<|file_sep|>updated/src/config_app.erl | %% Application callbacks
%% ===================================================================
start(_StartType, _StartArgs) ->
config_sup:start_link(get_ini_files()).
stop(_State) ->
ok.
get_ini_files() ->
Etc = filename:join(code:root_dir(), "etc"),
Default = [filename:join(Etc,"default.ini"), filename:join(Etc,"local.ini")],
DefaultExists = lists:filter(fun filelib:is_file/1, Default),
case init:get_argument(couch_ini) of
error ->
DefaultExists;
{ok, [[]]} ->
DefaultExists;
{ok, [Values]} ->
Values
end. | <|file_sep|>src/config_app.erl.diff
original:
updated:
DefaultExists = lists:filter(fun filelib:is_file/1, Default),
<|file_sep|>src/config_app.erl.diff
original:
Default;
updated:
DefaultExists;
<|file_sep|>original/src/config_app.erl
%% ===================================================================
%% Application callbacks
%% ===================================================================
start(_StartType, _StartArgs) ->
config_sup:start_link(get_ini_files()).
stop(_State) ->
ok.
get_ini_files() ->
Etc = filename:join(code:root_dir(), "etc"),
Default = [filename:join(Etc,"default.ini"), filename:join(Etc,"local.ini")],
case init:get_argument(couch_ini) of
error ->
Default;
{ok, [[]]} ->
Default;
{ok, [Values]} ->
Values
end.
<|file_sep|>current/src/config_app.erl
%% Application callbacks
%% ===================================================================
start(_StartType, _StartArgs) ->
config_sup:start_link(get_ini_files()).
stop(_State) ->
ok.
get_ini_files() ->
Etc = filename:join(code:root_dir(), "etc"),
Default = [filename:join(Etc,"default.ini"), filename:join(Etc,"local.ini")],
DefaultExists = lists:filter(fun filelib:is_file/1, Default),
case init:get_argument(couch_ini) of
error ->
DefaultExists;
{ok, [[]]} ->
Default;
{ok, [Values]} ->
Values
end.
<|file_sep|>updated/src/config_app.erl
%% Application callbacks
%% ===================================================================
start(_StartType, _StartArgs) ->
config_sup:start_link(get_ini_files()).
stop(_State) ->
ok.
get_ini_files() ->
Etc = filename:join(code:root_dir(), "etc"),
Default = [filename:join(Etc,"default.ini"), filename:join(Etc,"local.ini")],
DefaultExists = lists:filter(fun filelib:is_file/1, Default),
case init:get_argument(couch_ini) of
error ->
DefaultExists;
{ok, [[]]} ->
DefaultExists;
{ok, [Values]} ->
Values
end. | 21dd1938ea3365dbce81f363678e27fa2ae35828 | src/config_app.erl | src/config_app.erl | Erlang |
<|file_sep|>static/atom.css.diff
original:
.row > .pane + .pane {
updated:
.row > * + * {
<|file_sep|>original/static/atom.css
position: absolute;
width: 100%;
height: 100%;
overflow-x: hidden;
}
#root-view #panes .pane {
position: absolute;
width: 100%;
height: 100%;
box-sizing: border-box;
-webkit-box-shadow: inset 0px 0px 2px 2px rgba(0, 0, 0, .1);
}
.row > .pane + .pane {
border-left: 5px solid #515151;
}
.column > .pane + .pane {
border-top: 5px solid #515151;
}
<|file_sep|>current/static/atom.css
position: absolute;
width: 100%;
height: 100%;
overflow-x: hidden;
}
#root-view #panes .pane {
position: absolute;
width: 100%;
height: 100%;
box-sizing: border-box;
-webkit-box-shadow: inset 0px 0px 2px 2px rgba(0, 0, 0, .1);
}
.row > * + * {
border-left: 5px solid #515151;
}
.column > .pane + .pane {
border-top: 5px solid #515151;
}
<|file_sep|>updated/static/atom.css | position: absolute;
width: 100%;
height: 100%;
overflow-x: hidden;
}
#root-view #panes .pane {
position: absolute;
width: 100%;
height: 100%;
box-sizing: border-box;
-webkit-box-shadow: inset 0px 0px 2px 2px rgba(0, 0, 0, .1);
}
.row > * + * {
border-left: 5px solid #515151;
}
.column > * + * {
border-top: 5px solid #515151;
} | <|file_sep|>static/atom.css.diff
original:
.row > .pane + .pane {
updated:
.row > * + * {
<|file_sep|>original/static/atom.css
position: absolute;
width: 100%;
height: 100%;
overflow-x: hidden;
}
#root-view #panes .pane {
position: absolute;
width: 100%;
height: 100%;
box-sizing: border-box;
-webkit-box-shadow: inset 0px 0px 2px 2px rgba(0, 0, 0, .1);
}
.row > .pane + .pane {
border-left: 5px solid #515151;
}
.column > .pane + .pane {
border-top: 5px solid #515151;
}
<|file_sep|>current/static/atom.css
position: absolute;
width: 100%;
height: 100%;
overflow-x: hidden;
}
#root-view #panes .pane {
position: absolute;
width: 100%;
height: 100%;
box-sizing: border-box;
-webkit-box-shadow: inset 0px 0px 2px 2px rgba(0, 0, 0, .1);
}
.row > * + * {
border-left: 5px solid #515151;
}
.column > .pane + .pane {
border-top: 5px solid #515151;
}
<|file_sep|>updated/static/atom.css
position: absolute;
width: 100%;
height: 100%;
overflow-x: hidden;
}
#root-view #panes .pane {
position: absolute;
width: 100%;
height: 100%;
box-sizing: border-box;
-webkit-box-shadow: inset 0px 0px 2px 2px rgba(0, 0, 0, .1);
}
.row > * + * {
border-left: 5px solid #515151;
}
.column > * + * {
border-top: 5px solid #515151;
} | 08b661e3bf7ac1af744a8b24ae3e95edc1bb1cfc | static/atom.css | static/atom.css | CSS |
<|file_sep|>metadata/com.craigd.lmsmaterial.app.yml.diff
original:
updated:
- versionName: 0.1.3
versionCode: 103
commit: 0.1.3
subdir: lms-material
gradle:
- yes
<|file_sep|>original/metadata/com.craigd.lmsmaterial.app.yml
Repo: https://github.com/CDrummond/lms-material-app
Builds:
- versionName: 0.0.2
versionCode: 2
commit: 0.0.2
subdir: lms-material
gradle:
- yes
- versionName: 0.1.2
versionCode: 102
commit: 0.1.2
subdir: lms-material
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 0.1.2
CurrentVersionCode: 102
<|file_sep|>current/metadata/com.craigd.lmsmaterial.app.yml
gradle:
- yes
- versionName: 0.1.2
versionCode: 102
commit: 0.1.2
subdir: lms-material
gradle:
- yes
- versionName: 0.1.3
versionCode: 103
commit: 0.1.3
subdir: lms-material
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 0.1.2
CurrentVersionCode: 102
<|file_sep|>updated/metadata/com.craigd.lmsmaterial.app.yml | gradle:
- yes
- versionName: 0.1.2
versionCode: 102
commit: 0.1.2
subdir: lms-material
gradle:
- yes
- versionName: 0.1.3
versionCode: 103
commit: 0.1.3
subdir: lms-material
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 0.1.3
CurrentVersionCode: 103 | <|file_sep|>metadata/com.craigd.lmsmaterial.app.yml.diff
original:
updated:
- versionName: 0.1.3
versionCode: 103
commit: 0.1.3
subdir: lms-material
gradle:
- yes
<|file_sep|>original/metadata/com.craigd.lmsmaterial.app.yml
Repo: https://github.com/CDrummond/lms-material-app
Builds:
- versionName: 0.0.2
versionCode: 2
commit: 0.0.2
subdir: lms-material
gradle:
- yes
- versionName: 0.1.2
versionCode: 102
commit: 0.1.2
subdir: lms-material
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 0.1.2
CurrentVersionCode: 102
<|file_sep|>current/metadata/com.craigd.lmsmaterial.app.yml
gradle:
- yes
- versionName: 0.1.2
versionCode: 102
commit: 0.1.2
subdir: lms-material
gradle:
- yes
- versionName: 0.1.3
versionCode: 103
commit: 0.1.3
subdir: lms-material
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 0.1.2
CurrentVersionCode: 102
<|file_sep|>updated/metadata/com.craigd.lmsmaterial.app.yml
gradle:
- yes
- versionName: 0.1.2
versionCode: 102
commit: 0.1.2
subdir: lms-material
gradle:
- yes
- versionName: 0.1.3
versionCode: 103
commit: 0.1.3
subdir: lms-material
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 0.1.3
CurrentVersionCode: 103 | 40cee4f2d16643d9baa96e93b55ce2b07b109b52 | metadata/com.craigd.lmsmaterial.app.yml | metadata/com.craigd.lmsmaterial.app.yml | YAML |
<|file_sep|>README.md.diff
original:
# First, create a chest - a container for your drops
updated:
First, create a chest - a container for your drops
<|file_sep|>original/README.md
# Archloot
Quick and simple loot table creation.
## Examples:
# First, create a chest - a container for your drops
chest = Chest.new
# Initlialize your item
# It can be anything - Archloot class will accept any object
your_item = YourItemClass.new(.....)
your_item2 = YourItemClass.new(.....)
first_drop = Drop.new(your_item)
second_drop = Drop.new(your_item2)
chest.add(first_drop)
chest.add(second_drop)
chest.get_drops
<|file_sep|>current/README.md
Quick and simple loot table creation.
## Examples:
First, create a chest - a container for your drops
chest = Chest.new
# Initlialize your item
# It can be anything - Archloot class will accept any object
your_item = YourItemClass.new(.....)
your_item2 = YourItemClass.new(.....)
first_drop = Drop.new(your_item)
second_drop = Drop.new(your_item2)
chest.add(first_drop)
chest.add(second_drop)
chest.get_drops
<|file_sep|>updated/README.md |
Quick and simple loot table creation.
## Examples:
First, create a chest - a container for your drops
chest = Chest.new
Initlialize your item
It can be anything - Archloot class will accept any object
your_item = YourItemClass.new({name: "Random Ring of Spec"})
your_item2 = YourItemClass.new({name: "Talisman of Testing"})
first_drop = Drop.new(your_item)
second_drop = Drop.new(your_item2)
chest.add(first_drop)
chest.add(second_drop)
| <|file_sep|>README.md.diff
original:
# First, create a chest - a container for your drops
updated:
First, create a chest - a container for your drops
<|file_sep|>original/README.md
# Archloot
Quick and simple loot table creation.
## Examples:
# First, create a chest - a container for your drops
chest = Chest.new
# Initlialize your item
# It can be anything - Archloot class will accept any object
your_item = YourItemClass.new(.....)
your_item2 = YourItemClass.new(.....)
first_drop = Drop.new(your_item)
second_drop = Drop.new(your_item2)
chest.add(first_drop)
chest.add(second_drop)
chest.get_drops
<|file_sep|>current/README.md
Quick and simple loot table creation.
## Examples:
First, create a chest - a container for your drops
chest = Chest.new
# Initlialize your item
# It can be anything - Archloot class will accept any object
your_item = YourItemClass.new(.....)
your_item2 = YourItemClass.new(.....)
first_drop = Drop.new(your_item)
second_drop = Drop.new(your_item2)
chest.add(first_drop)
chest.add(second_drop)
chest.get_drops
<|file_sep|>updated/README.md
Quick and simple loot table creation.
## Examples:
First, create a chest - a container for your drops
chest = Chest.new
Initlialize your item
It can be anything - Archloot class will accept any object
your_item = YourItemClass.new({name: "Random Ring of Spec"})
your_item2 = YourItemClass.new({name: "Talisman of Testing"})
first_drop = Drop.new(your_item)
second_drop = Drop.new(your_item2)
chest.add(first_drop)
chest.add(second_drop)
| 2462c1aefa1ebc7c54114252a0ce2c72a22d7d6e | README.md | README.md | Markdown |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.