prompt stringlengths 77 6.9k | completion stringlengths 1 3.38k | text stringlengths 163 9.03k | commit stringlengths 40 40 | old_file stringlengths 4 264 | new_file stringlengths 4 264 | lang stringclasses 277
values |
|---|---|---|---|---|---|---|
<|file_sep|>src/qgr-cntrl-pills.js.diff
original:
}
updated:
},
<|file_sep|>original/src/qgr-cntrl-pills.js
]
};
}
});
var PillChoiceView = Backbone.View.extend({
// Represent an individual checkbox with a view.
tagName: 'label',
class: 'checkbox',
tmpl: Handlebars.compile(pills_tmpl),
events: {
'click li': 'set_choice',
'click': 'clicked',
},
initialize: function(options) {
<|file_sep|>current/src/qgr-cntrl-pills.js
]
};
}
});
var PillChoiceView = Backbone.View.extend({
// Represent an individual checkbox with a view.
tagName: 'label',
class: 'checkbox',
tmpl: Handlebars.compile(pills_tmpl),
events: {
'click li': 'set_choice',
'click': 'clicked',
},
initialize: function(options) {
<|file_sep|>updated/src/qgr-cntrl-pills.js | ]
};
}
});
var PillChoiceView = Backbone.View.extend({
// Represent an individual checkbox with a view.
tmpl: Handlebars.compile(pills_tmpl),
events: {
'click li': 'set_choice',
'click': 'clicked',
},
initialize: function(options) {
// Initialize with a choice model in the options hash.
_.bindAll(this, 'render', 'set_choice')
this.choices = this.options.choices;
}, | <|file_sep|>src/qgr-cntrl-pills.js.diff
original:
}
updated:
},
<|file_sep|>original/src/qgr-cntrl-pills.js
]
};
}
});
var PillChoiceView = Backbone.View.extend({
// Represent an individual checkbox with a view.
tagName: 'label',
class: 'checkbox',
tmpl: Handlebars.compile(pills_tmpl),
events: {
'click li': 'set_choice',
'click': 'clicked',
},
initialize: function(options) {
<|file_sep|>current/src/qgr-cntrl-pills.js
]
};
}
});
var PillChoiceView = Backbone.View.extend({
// Represent an individual checkbox with a view.
tagName: 'label',
class: 'checkbox',
tmpl: Handlebars.compile(pills_tmpl),
events: {
'click li': 'set_choice',
'click': 'clicked',
},
initialize: function(options) {
<|file_sep|>updated/src/qgr-cntrl-pills.js
]
};
}
});
var PillChoiceView = Backbone.View.extend({
// Represent an individual checkbox with a view.
tmpl: Handlebars.compile(pills_tmpl),
events: {
'click li': 'set_choice',
'click': 'clicked',
},
initialize: function(options) {
// Initialize with a choice model in the options hash.
_.bindAll(this, 'render', 'set_choice')
this.choices = this.options.choices;
}, | d89d9b2bc6d759b630ec5607ea67d3733b3cae26 | src/qgr-cntrl-pills.js | src/qgr-cntrl-pills.js | JavaScript |
<|file_sep|>original/data/script.js
if (!$(this).attr('data-url-expanded')) {
data_expanded_url = $(this).attr('data-expanded-url');
if (data_expanded_url) {
$(this).attr('href', data_expanded_url);
$(this).attr('data-url-expanded', '1');
$('> .js-display-url', this).html(data_expanded_url); }}
});
}
$(document).ready(function(){
replace_t_co_links();
console.error("Voila");
var observer = new MutationObserver(function(mutations){
replace_t_co_links();
});
var config = { attributes: true, childList: true, characterData: true, subtree: true };
observer.observe(document.body, config);
});
<|file_sep|>current/data/script.js
if (!$(this).attr('data-url-expanded')) {
data_expanded_url = $(this).attr('data-expanded-url');
if (data_expanded_url) {
$(this).attr('href', data_expanded_url);
$(this).attr('data-url-expanded', '1');
$('> .js-display-url', this).html(data_expanded_url); }}
});
}
$(document).ready(function(){
replace_t_co_links();
console.error("Voila");
var observer = new MutationObserver(function(mutations){
replace_t_co_links();
});
var config = { attributes: true, childList: true, characterData: true, subtree: true };
observer.observe(document.body, config);
});
<|file_sep|>updated/data/script.js | if (!$(this).attr('data-url-expanded')) {
data_expanded_url = $(this).attr('data-expanded-url');
if (data_expanded_url) {
$(this).attr('href', data_expanded_url);
$(this).attr('data-url-expanded', '1');
$('> .js-display-url', this).html(data_expanded_url); }}
});
}
$(document).ready(function(){
replace_t_co_links();
var observer = new MutationObserver(function(mutations){
replace_t_co_links();
});
var config = { attributes: true, childList: true, characterData: true, subtree: true };
observer.observe(document.body, config);
}); | <|file_sep|>original/data/script.js
if (!$(this).attr('data-url-expanded')) {
data_expanded_url = $(this).attr('data-expanded-url');
if (data_expanded_url) {
$(this).attr('href', data_expanded_url);
$(this).attr('data-url-expanded', '1');
$('> .js-display-url', this).html(data_expanded_url); }}
});
}
$(document).ready(function(){
replace_t_co_links();
console.error("Voila");
var observer = new MutationObserver(function(mutations){
replace_t_co_links();
});
var config = { attributes: true, childList: true, characterData: true, subtree: true };
observer.observe(document.body, config);
});
<|file_sep|>current/data/script.js
if (!$(this).attr('data-url-expanded')) {
data_expanded_url = $(this).attr('data-expanded-url');
if (data_expanded_url) {
$(this).attr('href', data_expanded_url);
$(this).attr('data-url-expanded', '1');
$('> .js-display-url', this).html(data_expanded_url); }}
});
}
$(document).ready(function(){
replace_t_co_links();
console.error("Voila");
var observer = new MutationObserver(function(mutations){
replace_t_co_links();
});
var config = { attributes: true, childList: true, characterData: true, subtree: true };
observer.observe(document.body, config);
});
<|file_sep|>updated/data/script.js
if (!$(this).attr('data-url-expanded')) {
data_expanded_url = $(this).attr('data-expanded-url');
if (data_expanded_url) {
$(this).attr('href', data_expanded_url);
$(this).attr('data-url-expanded', '1');
$('> .js-display-url', this).html(data_expanded_url); }}
});
}
$(document).ready(function(){
replace_t_co_links();
var observer = new MutationObserver(function(mutations){
replace_t_co_links();
});
var config = { attributes: true, childList: true, characterData: true, subtree: true };
observer.observe(document.body, config);
}); | c0ee940a1de0ce38d504a55abca5050618ae945d | data/script.js | data/script.js | JavaScript |
<|file_sep|>demo/node/Readme.md.diff
original:
var jsroot = require("jsroot");
updated:
let jsroot = require("jsroot");
<|file_sep|>demo/node/Readme.md.diff
original:
Example <makesvg.js> you will find in this directory. Just call it:
updated:
Examples <makesvg.js> or <geomsvg.js> you will find in this directory. Just call them:
<|file_sep|>original/demo/node/Readme.md
var jsroot = require("jsroot");
Provided package.json file allows to use demos directly with local jsroot installation:
npm install
Main motivation to use JSROOT from Node.js is creation of SVG files.
Example <makesvg.js> you will find in this directory. Just call it:
node makesvg.js
JSROOT also provides possibility to read arbitrary TTree data without involving
any peace of native ROOT code. <tree_dump.js> demonstrate how simple dump of TTree
data can be done:
node tree_dump.js
JSROOT also implements extensive tree-draw functionality, shown in <tree_draw.js> example:
node tree_draw.js
<|file_sep|>current/demo/node/Readme.md
let jsroot = require("jsroot");
Provided package.json file allows to use demos directly with local jsroot installation:
npm install
Main motivation to use JSROOT from Node.js is creation of SVG files.
Examples <makesvg.js> or <geomsvg.js> you will find in this directory. Just call them:
node makesvg.js
JSROOT also provides possibility to read arbitrary TTree data without involving
any peace of native ROOT code. <tree_dump.js> demonstrate how simple dump of TTree
data can be done:
node tree_dump.js
JSROOT also implements extensive tree-draw functionality, shown in <tree_draw.js> example:
node tree_draw.js
<|file_sep|>updated/demo/node/Readme.md | let jsroot = require("jsroot");
Provided package.json file allows to use demos directly with local jsroot installation:
npm install
Main motivation to use JSROOT from Node.js is creation of SVG files.
Examples <makesvg.js> or <geomsvg.js> you will find in this directory. Just call them:
node makesvg.js
node geomsvg.js
JSROOT also provides possibility to read arbitrary TTree data without involving
any peace of native ROOT code. <tree_dump.js> demonstrate how simple dump of TTree
data can be done:
node tree_dump.js
JSROOT also implements extensive tree-draw functionality, shown in <tree_draw.js> example:
node tree_draw.js | <|file_sep|>demo/node/Readme.md.diff
original:
var jsroot = require("jsroot");
updated:
let jsroot = require("jsroot");
<|file_sep|>demo/node/Readme.md.diff
original:
Example <makesvg.js> you will find in this directory. Just call it:
updated:
Examples <makesvg.js> or <geomsvg.js> you will find in this directory. Just call them:
<|file_sep|>original/demo/node/Readme.md
var jsroot = require("jsroot");
Provided package.json file allows to use demos directly with local jsroot installation:
npm install
Main motivation to use JSROOT from Node.js is creation of SVG files.
Example <makesvg.js> you will find in this directory. Just call it:
node makesvg.js
JSROOT also provides possibility to read arbitrary TTree data without involving
any peace of native ROOT code. <tree_dump.js> demonstrate how simple dump of TTree
data can be done:
node tree_dump.js
JSROOT also implements extensive tree-draw functionality, shown in <tree_draw.js> example:
node tree_draw.js
<|file_sep|>current/demo/node/Readme.md
let jsroot = require("jsroot");
Provided package.json file allows to use demos directly with local jsroot installation:
npm install
Main motivation to use JSROOT from Node.js is creation of SVG files.
Examples <makesvg.js> or <geomsvg.js> you will find in this directory. Just call them:
node makesvg.js
JSROOT also provides possibility to read arbitrary TTree data without involving
any peace of native ROOT code. <tree_dump.js> demonstrate how simple dump of TTree
data can be done:
node tree_dump.js
JSROOT also implements extensive tree-draw functionality, shown in <tree_draw.js> example:
node tree_draw.js
<|file_sep|>updated/demo/node/Readme.md
let jsroot = require("jsroot");
Provided package.json file allows to use demos directly with local jsroot installation:
npm install
Main motivation to use JSROOT from Node.js is creation of SVG files.
Examples <makesvg.js> or <geomsvg.js> you will find in this directory. Just call them:
node makesvg.js
node geomsvg.js
JSROOT also provides possibility to read arbitrary TTree data without involving
any peace of native ROOT code. <tree_dump.js> demonstrate how simple dump of TTree
data can be done:
node tree_dump.js
JSROOT also implements extensive tree-draw functionality, shown in <tree_draw.js> example:
node tree_draw.js | 80d4095509d98e35a227360e657f15f1c46a6ef1 | demo/node/Readme.md | demo/node/Readme.md | Markdown |
<|file_sep|>original/examples/built/index.html
<head>
<style>
body {
font-family: sans-serif;
font-size: 14px;
line-height: 18px;
}
</style>
<script type="text/javascript" src="//platform.twitter.com/widgets.js"></script>
<script src="http://cdn.livefyre.com/libs/livefyre-auth/0.1.0/livefyre-auth.min.js"></script>
<script src="http://cdn.livefyre.com/libs/sdk/v2.12.9/streamhub-sdk.min.js"></script>
<script src="../../dist/streamhub-wall.min.js"></script>
</head>
<body>
<div id="wall"></div>
<script>
(function () {
var SDK = Livefyre.require('streamhub-sdk');
var LiveMediaWall = Livefyre.require('streamhub-wall');
<|file_sep|>current/examples/built/index.html
<head>
<style>
body {
font-family: sans-serif;
font-size: 14px;
line-height: 18px;
}
</style>
<script type="text/javascript" src="//platform.twitter.com/widgets.js"></script>
<script src="http://cdn.livefyre.com/libs/livefyre-auth/0.1.0/livefyre-auth.min.js"></script>
<script src="http://cdn.livefyre.com/libs/sdk/v2.12.9/streamhub-sdk.min.js"></script>
<script src="../../dist/streamhub-wall.min.js"></script>
</head>
<body>
<div id="wall"></div>
<script>
(function () {
var SDK = Livefyre.require('streamhub-sdk');
var LiveMediaWall = Livefyre.require('streamhub-wall');
<|file_sep|>updated/examples/built/index.html | <head>
<style>
body {
font-family: sans-serif;
font-size: 14px;
line-height: 18px;
}
</style>
<script type="text/javascript" src="//platform.twitter.com/widgets.js"></script>
<script src="http://cdn.livefyre.com/libs/livefyre-auth/0.1.0/livefyre-auth.min.js"></script>
<script src="http://cdn.livefyre.com/libs/sdk/v2.12.10/streamhub-sdk.min.js"></script>
<script src="../../dist/streamhub-wall.min.js"></script>
</head>
<body>
<div id="wall"></div>
<script>
(function () {
var SDK = Livefyre.require('streamhub-sdk');
var LiveMediaWall = Livefyre.require('streamhub-wall'); | <|file_sep|>original/examples/built/index.html
<head>
<style>
body {
font-family: sans-serif;
font-size: 14px;
line-height: 18px;
}
</style>
<script type="text/javascript" src="//platform.twitter.com/widgets.js"></script>
<script src="http://cdn.livefyre.com/libs/livefyre-auth/0.1.0/livefyre-auth.min.js"></script>
<script src="http://cdn.livefyre.com/libs/sdk/v2.12.9/streamhub-sdk.min.js"></script>
<script src="../../dist/streamhub-wall.min.js"></script>
</head>
<body>
<div id="wall"></div>
<script>
(function () {
var SDK = Livefyre.require('streamhub-sdk');
var LiveMediaWall = Livefyre.require('streamhub-wall');
<|file_sep|>current/examples/built/index.html
<head>
<style>
body {
font-family: sans-serif;
font-size: 14px;
line-height: 18px;
}
</style>
<script type="text/javascript" src="//platform.twitter.com/widgets.js"></script>
<script src="http://cdn.livefyre.com/libs/livefyre-auth/0.1.0/livefyre-auth.min.js"></script>
<script src="http://cdn.livefyre.com/libs/sdk/v2.12.9/streamhub-sdk.min.js"></script>
<script src="../../dist/streamhub-wall.min.js"></script>
</head>
<body>
<div id="wall"></div>
<script>
(function () {
var SDK = Livefyre.require('streamhub-sdk');
var LiveMediaWall = Livefyre.require('streamhub-wall');
<|file_sep|>updated/examples/built/index.html
<head>
<style>
body {
font-family: sans-serif;
font-size: 14px;
line-height: 18px;
}
</style>
<script type="text/javascript" src="//platform.twitter.com/widgets.js"></script>
<script src="http://cdn.livefyre.com/libs/livefyre-auth/0.1.0/livefyre-auth.min.js"></script>
<script src="http://cdn.livefyre.com/libs/sdk/v2.12.10/streamhub-sdk.min.js"></script>
<script src="../../dist/streamhub-wall.min.js"></script>
</head>
<body>
<div id="wall"></div>
<script>
(function () {
var SDK = Livefyre.require('streamhub-sdk');
var LiveMediaWall = Livefyre.require('streamhub-wall'); | c76851b647b8853eb48bd3ba2100e9a8a76c5074 | examples/built/index.html | examples/built/index.html | HTML |
<|file_sep|>original/readux/annotations/migrations/0003_annotation_group_and_permissions.py
operations = [
migrations.CreateModel(
name='AnnotationGroup',
fields=[
('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')),
('notes', models.TextField(blank=True)),
('created', models.DateTimeField(auto_now_add=True)),
('updated', models.DateTimeField(auto_now=True)),
],
bases=('auth.group',),
),
migrations.AlterModelOptions(
name='annotation',
options={'permissions': (('view_annotation', 'View annotation'),)},
),
migrations.AlterModelOptions(
name='annotation',
options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))},
),
]
<|file_sep|>current/readux/annotations/migrations/0003_annotation_group_and_permissions.py
operations = [
migrations.CreateModel(
name='AnnotationGroup',
fields=[
('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')),
('notes', models.TextField(blank=True)),
('created', models.DateTimeField(auto_now_add=True)),
('updated', models.DateTimeField(auto_now=True)),
],
bases=('auth.group',),
),
migrations.AlterModelOptions(
name='annotation',
options={'permissions': (('view_annotation', 'View annotation'),)},
),
migrations.AlterModelOptions(
name='annotation',
options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))},
),
]
<|file_sep|>updated/readux/annotations/migrations/0003_annotation_group_and_permissions.py | dependencies = [
('auth', '0006_require_contenttypes_0002'),
('annotations', '0002_add_volume_uri'),
]
operations = [
migrations.CreateModel(
name='AnnotationGroup',
fields=[
('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')),
('notes', models.TextField(blank=True)),
('created', models.DateTimeField(auto_now_add=True)),
('updated', models.DateTimeField(auto_now=True)),
],
bases=('auth.group',),
),
migrations.AlterModelOptions(
name='annotation',
options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))},
),
] | <|file_sep|>original/readux/annotations/migrations/0003_annotation_group_and_permissions.py
operations = [
migrations.CreateModel(
name='AnnotationGroup',
fields=[
('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')),
('notes', models.TextField(blank=True)),
('created', models.DateTimeField(auto_now_add=True)),
('updated', models.DateTimeField(auto_now=True)),
],
bases=('auth.group',),
),
migrations.AlterModelOptions(
name='annotation',
options={'permissions': (('view_annotation', 'View annotation'),)},
),
migrations.AlterModelOptions(
name='annotation',
options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))},
),
]
<|file_sep|>current/readux/annotations/migrations/0003_annotation_group_and_permissions.py
operations = [
migrations.CreateModel(
name='AnnotationGroup',
fields=[
('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')),
('notes', models.TextField(blank=True)),
('created', models.DateTimeField(auto_now_add=True)),
('updated', models.DateTimeField(auto_now=True)),
],
bases=('auth.group',),
),
migrations.AlterModelOptions(
name='annotation',
options={'permissions': (('view_annotation', 'View annotation'),)},
),
migrations.AlterModelOptions(
name='annotation',
options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))},
),
]
<|file_sep|>updated/readux/annotations/migrations/0003_annotation_group_and_permissions.py
dependencies = [
('auth', '0006_require_contenttypes_0002'),
('annotations', '0002_add_volume_uri'),
]
operations = [
migrations.CreateModel(
name='AnnotationGroup',
fields=[
('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')),
('notes', models.TextField(blank=True)),
('created', models.DateTimeField(auto_now_add=True)),
('updated', models.DateTimeField(auto_now=True)),
],
bases=('auth.group',),
),
migrations.AlterModelOptions(
name='annotation',
options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))},
),
] | a9354124f4905f4befe9ff2ca8274406fbbb0dad | readux/annotations/migrations/0003_annotation_group_and_permissions.py | readux/annotations/migrations/0003_annotation_group_and_permissions.py | Python |
<|file_sep|>original/.travis.yml
language: python
python:
- 2.7
- 3.2
- 3.6
install:
- pip install -r requirements.txt
- sudo rm -rf /dev/shm && sudo ln -s /run/shm /dev/shm
script: coverage run --source=breeze setup.py test
notifications:
email:
recipients:
- aortiz32@gmail.com
on_success: [always|never|change] # default: change
on_failure: [always|never|change] # default: always
after_success: coveralls
<|file_sep|>current/.travis.yml
language: python
python:
- 2.7
- 3.2
- 3.6
install:
- pip install -r requirements.txt
- sudo rm -rf /dev/shm && sudo ln -s /run/shm /dev/shm
script: coverage run --source=breeze setup.py test
notifications:
email:
recipients:
- aortiz32@gmail.com
on_success: [always|never|change] # default: change
on_failure: [always|never|change] # default: always
after_success: coveralls
<|file_sep|>updated/.travis.yml | language: python
python:
- 2.7
install:
- pip install -r requirements.txt
- sudo rm -rf /dev/shm && sudo ln -s /run/shm /dev/shm
script: coverage run --source=breeze setup.py test
notifications:
email:
recipients:
- aortiz32@gmail.com
on_success: [always|never|change] # default: change
on_failure: [always|never|change] # default: always
after_success: coveralls
| <|file_sep|>original/.travis.yml
language: python
python:
- 2.7
- 3.2
- 3.6
install:
- pip install -r requirements.txt
- sudo rm -rf /dev/shm && sudo ln -s /run/shm /dev/shm
script: coverage run --source=breeze setup.py test
notifications:
email:
recipients:
- aortiz32@gmail.com
on_success: [always|never|change] # default: change
on_failure: [always|never|change] # default: always
after_success: coveralls
<|file_sep|>current/.travis.yml
language: python
python:
- 2.7
- 3.2
- 3.6
install:
- pip install -r requirements.txt
- sudo rm -rf /dev/shm && sudo ln -s /run/shm /dev/shm
script: coverage run --source=breeze setup.py test
notifications:
email:
recipients:
- aortiz32@gmail.com
on_success: [always|never|change] # default: change
on_failure: [always|never|change] # default: always
after_success: coveralls
<|file_sep|>updated/.travis.yml
language: python
python:
- 2.7
install:
- pip install -r requirements.txt
- sudo rm -rf /dev/shm && sudo ln -s /run/shm /dev/shm
script: coverage run --source=breeze setup.py test
notifications:
email:
recipients:
- aortiz32@gmail.com
on_success: [always|never|change] # default: change
on_failure: [always|never|change] # default: always
after_success: coveralls
| acfb307e878d3073cf7845f6a91412bbb9083664 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/.github/workflows/php.yml
paths:
- 'composer.*'
- phpcs.xml
- phpstan.neon
- '**.php'
jobs:
build:
strategy:
matrix:
php-versions: ['8.0', '7.4', '7.3', '7.2']
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v1
- name: Set up PHP versions
uses: shivammathur/setup-php@v1
with:
php-version: ${{ matrix.php-versions }}
<|file_sep|>current/.github/workflows/php.yml
paths:
- 'composer.*'
- phpcs.xml
- phpstan.neon
- '**.php'
jobs:
build:
strategy:
matrix:
php-versions: ['8.0', '7.4', '7.3', '7.2']
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v1
- name: Set up PHP versions
uses: shivammathur/setup-php@v1
with:
php-version: ${{ matrix.php-versions }}
<|file_sep|>updated/.github/workflows/php.yml | paths:
- 'composer.*'
- phpcs.xml
- phpstan.neon
- '**.php'
jobs:
build:
strategy:
matrix:
php-versions: ['7.4', '7.3', '7.2']
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v1
- name: Set up PHP versions
uses: shivammathur/setup-php@v1
with:
php-version: ${{ matrix.php-versions }} | <|file_sep|>original/.github/workflows/php.yml
paths:
- 'composer.*'
- phpcs.xml
- phpstan.neon
- '**.php'
jobs:
build:
strategy:
matrix:
php-versions: ['8.0', '7.4', '7.3', '7.2']
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v1
- name: Set up PHP versions
uses: shivammathur/setup-php@v1
with:
php-version: ${{ matrix.php-versions }}
<|file_sep|>current/.github/workflows/php.yml
paths:
- 'composer.*'
- phpcs.xml
- phpstan.neon
- '**.php'
jobs:
build:
strategy:
matrix:
php-versions: ['8.0', '7.4', '7.3', '7.2']
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v1
- name: Set up PHP versions
uses: shivammathur/setup-php@v1
with:
php-version: ${{ matrix.php-versions }}
<|file_sep|>updated/.github/workflows/php.yml
paths:
- 'composer.*'
- phpcs.xml
- phpstan.neon
- '**.php'
jobs:
build:
strategy:
matrix:
php-versions: ['7.4', '7.3', '7.2']
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v1
- name: Set up PHP versions
uses: shivammathur/setup-php@v1
with:
php-version: ${{ matrix.php-versions }} | 9c1649930a3b1bb87422b7ca8aae93409b7b4e4a | .github/workflows/php.yml | .github/workflows/php.yml | YAML |
<|file_sep|>original/reviewboard/static/rb/js/models/reviewReplyModel.js
listKey: 'replies',
toJSON: function() {
return {
'public': this.get('public'),
'body_top': this.get('bodyTop'),
'body_bottom': this.get('bodyBottom')
};
},
parse: function(rsp) {
var result = RB.BaseResource.prototype.parse.call(this, rsp),
rspData = rsp[this.rspNamespace];
result.bodyTop = rspData.body_top;
result.bodyBottom = rspData.body_bottom;
result.public = rspData.public;
return result;
},
});
<|file_sep|>current/reviewboard/static/rb/js/models/reviewReplyModel.js
listKey: 'replies',
toJSON: function() {
return {
'public': this.get('public'),
'body_top': this.get('bodyTop'),
'body_bottom': this.get('bodyBottom')
};
},
parse: function(rsp) {
var result = RB.BaseResource.prototype.parse.call(this, rsp),
rspData = rsp[this.rspNamespace];
result.bodyTop = rspData.body_top;
result.bodyBottom = rspData.body_bottom;
result.public = rspData.public;
return result;
},
});
<|file_sep|>updated/reviewboard/static/rb/js/models/reviewReplyModel.js | listKey: 'replies',
toJSON: function() {
return {
'public': this.get('public'),
'body_top': this.get('bodyTop'),
'body_bottom': this.get('bodyBottom')
};
},
parse: function(rsp) {
var result = RB.BaseResource.prototype.parse.call(this, rsp),
rspData = rsp[this.rspNamespace];
result.bodyTop = rspData.body_top;
result.bodyBottom = rspData.body_bottom;
result.public = rspData.public;
return result;
}
}); | <|file_sep|>original/reviewboard/static/rb/js/models/reviewReplyModel.js
listKey: 'replies',
toJSON: function() {
return {
'public': this.get('public'),
'body_top': this.get('bodyTop'),
'body_bottom': this.get('bodyBottom')
};
},
parse: function(rsp) {
var result = RB.BaseResource.prototype.parse.call(this, rsp),
rspData = rsp[this.rspNamespace];
result.bodyTop = rspData.body_top;
result.bodyBottom = rspData.body_bottom;
result.public = rspData.public;
return result;
},
});
<|file_sep|>current/reviewboard/static/rb/js/models/reviewReplyModel.js
listKey: 'replies',
toJSON: function() {
return {
'public': this.get('public'),
'body_top': this.get('bodyTop'),
'body_bottom': this.get('bodyBottom')
};
},
parse: function(rsp) {
var result = RB.BaseResource.prototype.parse.call(this, rsp),
rspData = rsp[this.rspNamespace];
result.bodyTop = rspData.body_top;
result.bodyBottom = rspData.body_bottom;
result.public = rspData.public;
return result;
},
});
<|file_sep|>updated/reviewboard/static/rb/js/models/reviewReplyModel.js
listKey: 'replies',
toJSON: function() {
return {
'public': this.get('public'),
'body_top': this.get('bodyTop'),
'body_bottom': this.get('bodyBottom')
};
},
parse: function(rsp) {
var result = RB.BaseResource.prototype.parse.call(this, rsp),
rspData = rsp[this.rspNamespace];
result.bodyTop = rspData.body_top;
result.bodyBottom = rspData.body_bottom;
result.public = rspData.public;
return result;
}
}); | f0eac5b36cba5263c61b95f5d13c97bed4aab652 | reviewboard/static/rb/js/models/reviewReplyModel.js | reviewboard/static/rb/js/models/reviewReplyModel.js | JavaScript |
<|file_sep|>src/java/test/org/jaxen/javabean/DocumentNavigatorTest.java.diff
original:
updated:
import org.jaxen.saxpath.helpers.XPathReaderFactory;
<|file_sep|>original/src/java/test/org/jaxen/javabean/DocumentNavigatorTest.java
package org.jaxen.javabean;
import junit.framework.TestCase;
public class DocumentNavigatorTest
extends TestCase
{
public void testNothing()
throws Exception
{
JavaBeanXPath xpath = new JavaBeanXPath( "brother[position()<4]/name" );
Person bob = new Person( "bob", 30 );
bob.addBrother( new Person( "billy", 34 ) );
bob.addBrother( new Person( "seth", 29 ) );
bob.addBrother( new Person( "dave", 32 ) );
bob.addBrother( new Person( "jim", 29 ) );
bob.addBrother( new Person( "larry", 42 ) );
bob.addBrother( new Person( "ted", 22 ) );
<|file_sep|>current/src/java/test/org/jaxen/javabean/DocumentNavigatorTest.java
package org.jaxen.javabean;
import junit.framework.TestCase;
import org.jaxen.saxpath.helpers.XPathReaderFactory;
public class DocumentNavigatorTest
extends TestCase
{
public void testNothing()
throws Exception
{
JavaBeanXPath xpath = new JavaBeanXPath( "brother[position()<4]/name" );
Person bob = new Person( "bob", 30 );
bob.addBrother( new Person( "billy", 34 ) );
bob.addBrother( new Person( "seth", 29 ) );
bob.addBrother( new Person( "dave", 32 ) );
bob.addBrother( new Person( "jim", 29 ) );
bob.addBrother( new Person( "larry", 42 ) );
<|file_sep|>updated/src/java/test/org/jaxen/javabean/DocumentNavigatorTest.java | package org.jaxen.javabean;
import junit.framework.TestCase;
import org.jaxen.saxpath.helpers.XPathReaderFactory;
public class DocumentNavigatorTest
extends TestCase
{
protected void setUp() throws Exception
{
System.setProperty( XPathReaderFactory.DRIVER_PROPERTY,
"" );
}
public void testNothing()
throws Exception
{
JavaBeanXPath xpath = new JavaBeanXPath( "brother[position()<4]/name" );
Person bob = new Person( "bob", 30 ); | <|file_sep|>src/java/test/org/jaxen/javabean/DocumentNavigatorTest.java.diff
original:
updated:
import org.jaxen.saxpath.helpers.XPathReaderFactory;
<|file_sep|>original/src/java/test/org/jaxen/javabean/DocumentNavigatorTest.java
package org.jaxen.javabean;
import junit.framework.TestCase;
public class DocumentNavigatorTest
extends TestCase
{
public void testNothing()
throws Exception
{
JavaBeanXPath xpath = new JavaBeanXPath( "brother[position()<4]/name" );
Person bob = new Person( "bob", 30 );
bob.addBrother( new Person( "billy", 34 ) );
bob.addBrother( new Person( "seth", 29 ) );
bob.addBrother( new Person( "dave", 32 ) );
bob.addBrother( new Person( "jim", 29 ) );
bob.addBrother( new Person( "larry", 42 ) );
bob.addBrother( new Person( "ted", 22 ) );
<|file_sep|>current/src/java/test/org/jaxen/javabean/DocumentNavigatorTest.java
package org.jaxen.javabean;
import junit.framework.TestCase;
import org.jaxen.saxpath.helpers.XPathReaderFactory;
public class DocumentNavigatorTest
extends TestCase
{
public void testNothing()
throws Exception
{
JavaBeanXPath xpath = new JavaBeanXPath( "brother[position()<4]/name" );
Person bob = new Person( "bob", 30 );
bob.addBrother( new Person( "billy", 34 ) );
bob.addBrother( new Person( "seth", 29 ) );
bob.addBrother( new Person( "dave", 32 ) );
bob.addBrother( new Person( "jim", 29 ) );
bob.addBrother( new Person( "larry", 42 ) );
<|file_sep|>updated/src/java/test/org/jaxen/javabean/DocumentNavigatorTest.java
package org.jaxen.javabean;
import junit.framework.TestCase;
import org.jaxen.saxpath.helpers.XPathReaderFactory;
public class DocumentNavigatorTest
extends TestCase
{
protected void setUp() throws Exception
{
System.setProperty( XPathReaderFactory.DRIVER_PROPERTY,
"" );
}
public void testNothing()
throws Exception
{
JavaBeanXPath xpath = new JavaBeanXPath( "brother[position()<4]/name" );
Person bob = new Person( "bob", 30 ); | f04e482a92c8faca43c3d4f098488cc4e54236d2 | src/java/test/org/jaxen/javabean/DocumentNavigatorTest.java | src/java/test/org/jaxen/javabean/DocumentNavigatorTest.java | Java |
<|file_sep|>original/json-stream.gemspec
require './lib/json/stream/version'
Gem::Specification.new do |s|
s.name = 'json-stream'
s.version = JSON::Stream::VERSION
s.summary = %q[A streaming JSON parser that generates SAX-like events.]
s.description = %q[A parser best suited for huge JSON documents that don't fit in memory.]
s.authors = ['David Graham']
s.email = %w[david.malcom.graham@gmail.com]
s.homepage = 'http://dgraham.github.io/json-stream/'
s.license = 'MIT'
s.files = Dir['[A-Z]*', 'json-stream.gemspec', '{lib}/**/*']
s.test_files = Dir['spec/**/*']
s.require_path = 'lib'
s.add_development_dependency 'rake'
s.required_ruby_version = '>= 1.9.2'
end
<|file_sep|>current/json-stream.gemspec
require './lib/json/stream/version'
Gem::Specification.new do |s|
s.name = 'json-stream'
s.version = JSON::Stream::VERSION
s.summary = %q[A streaming JSON parser that generates SAX-like events.]
s.description = %q[A parser best suited for huge JSON documents that don't fit in memory.]
s.authors = ['David Graham']
s.email = %w[david.malcom.graham@gmail.com]
s.homepage = 'http://dgraham.github.io/json-stream/'
s.license = 'MIT'
s.files = Dir['[A-Z]*', 'json-stream.gemspec', '{lib}/**/*']
s.test_files = Dir['spec/**/*']
s.require_path = 'lib'
s.add_development_dependency 'rake'
s.required_ruby_version = '>= 1.9.2'
end
<|file_sep|>updated/json-stream.gemspec | require './lib/json/stream/version'
Gem::Specification.new do |s|
s.name = 'json-stream'
s.version = JSON::Stream::VERSION
s.summary = %q[A streaming JSON parser that generates SAX-like events.]
s.description = %q[A parser best suited for huge JSON documents that don't fit in memory.]
s.authors = ['David Graham']
s.email = %w[david.malcom.graham@gmail.com]
s.homepage = 'http://dgraham.github.io/json-stream/'
s.license = 'MIT'
s.files = Dir['[A-Z]*', 'json-stream.gemspec', '{lib}/**/*']
s.test_files = Dir['spec/**/*']
s.require_path = 'lib'
s.add_development_dependency 'rake', '~> 10.3'
s.required_ruby_version = '>= 1.9.2'
end | <|file_sep|>original/json-stream.gemspec
require './lib/json/stream/version'
Gem::Specification.new do |s|
s.name = 'json-stream'
s.version = JSON::Stream::VERSION
s.summary = %q[A streaming JSON parser that generates SAX-like events.]
s.description = %q[A parser best suited for huge JSON documents that don't fit in memory.]
s.authors = ['David Graham']
s.email = %w[david.malcom.graham@gmail.com]
s.homepage = 'http://dgraham.github.io/json-stream/'
s.license = 'MIT'
s.files = Dir['[A-Z]*', 'json-stream.gemspec', '{lib}/**/*']
s.test_files = Dir['spec/**/*']
s.require_path = 'lib'
s.add_development_dependency 'rake'
s.required_ruby_version = '>= 1.9.2'
end
<|file_sep|>current/json-stream.gemspec
require './lib/json/stream/version'
Gem::Specification.new do |s|
s.name = 'json-stream'
s.version = JSON::Stream::VERSION
s.summary = %q[A streaming JSON parser that generates SAX-like events.]
s.description = %q[A parser best suited for huge JSON documents that don't fit in memory.]
s.authors = ['David Graham']
s.email = %w[david.malcom.graham@gmail.com]
s.homepage = 'http://dgraham.github.io/json-stream/'
s.license = 'MIT'
s.files = Dir['[A-Z]*', 'json-stream.gemspec', '{lib}/**/*']
s.test_files = Dir['spec/**/*']
s.require_path = 'lib'
s.add_development_dependency 'rake'
s.required_ruby_version = '>= 1.9.2'
end
<|file_sep|>updated/json-stream.gemspec
require './lib/json/stream/version'
Gem::Specification.new do |s|
s.name = 'json-stream'
s.version = JSON::Stream::VERSION
s.summary = %q[A streaming JSON parser that generates SAX-like events.]
s.description = %q[A parser best suited for huge JSON documents that don't fit in memory.]
s.authors = ['David Graham']
s.email = %w[david.malcom.graham@gmail.com]
s.homepage = 'http://dgraham.github.io/json-stream/'
s.license = 'MIT'
s.files = Dir['[A-Z]*', 'json-stream.gemspec', '{lib}/**/*']
s.test_files = Dir['spec/**/*']
s.require_path = 'lib'
s.add_development_dependency 'rake', '~> 10.3'
s.required_ruby_version = '>= 1.9.2'
end | d200ce0f63fc55f3b144c75afb2b1364c8098060 | json-stream.gemspec | json-stream.gemspec | Ruby |
<|file_sep|>original/Handler/HistoryBG.hs
module Handler.HistoryBG where
import Import
------------------------------------------------------------------------
getHistoryBGR :: Handler Html
getHistoryBGR = do
Entity uid _ <- requireAuth
-- TODO: pagination
sugars <- fmap (map entityVal) $
runDB $ selectList [BloodGlucoseHistoryUid ==. uid] [LimitTo 10]
let bloodGlucoseHistoryDate _ = "TODO" :: Text -- TODO
bloodGlucoseHistoryTime _ = "TODO" :: Text -- TODO
defaultLayout $ do
setTitle "Betty : Blood Sugar Logs"
$(widgetFile "bg.history")
<|file_sep|>current/Handler/HistoryBG.hs
module Handler.HistoryBG where
import Import
------------------------------------------------------------------------
getHistoryBGR :: Handler Html
getHistoryBGR = do
Entity uid _ <- requireAuth
-- TODO: pagination
sugars <- fmap (map entityVal) $
runDB $ selectList [BloodGlucoseHistoryUid ==. uid] [LimitTo 10]
let bloodGlucoseHistoryDate _ = "TODO" :: Text -- TODO
bloodGlucoseHistoryTime _ = "TODO" :: Text -- TODO
defaultLayout $ do
setTitle "Betty : Blood Sugar Logs"
$(widgetFile "bg.history")
<|file_sep|>updated/Handler/HistoryBG.hs | {-# LANGUAGE NoDisambiguateRecordFields, NoRecordWildCards #-}
module Handler.HistoryBG where
import Import
------------------------------------------------------------------------
getHistoryBGR :: Handler Html
getHistoryBGR = do
Entity uid _ <- requireAuth
-- TODO: pagination
sugars <- fmap (map entityVal) $
runDB $ selectList [BloodGlucoseHistoryUid ==. uid] [LimitTo 10]
let bloodGlucoseHistoryDate _ = "TODO" :: Text -- TODO
bloodGlucoseHistoryTime _ = "TODO" :: Text -- TODO
defaultLayout $ do
setTitle "Betty : Blood Sugar Logs" | <|file_sep|>original/Handler/HistoryBG.hs
module Handler.HistoryBG where
import Import
------------------------------------------------------------------------
getHistoryBGR :: Handler Html
getHistoryBGR = do
Entity uid _ <- requireAuth
-- TODO: pagination
sugars <- fmap (map entityVal) $
runDB $ selectList [BloodGlucoseHistoryUid ==. uid] [LimitTo 10]
let bloodGlucoseHistoryDate _ = "TODO" :: Text -- TODO
bloodGlucoseHistoryTime _ = "TODO" :: Text -- TODO
defaultLayout $ do
setTitle "Betty : Blood Sugar Logs"
$(widgetFile "bg.history")
<|file_sep|>current/Handler/HistoryBG.hs
module Handler.HistoryBG where
import Import
------------------------------------------------------------------------
getHistoryBGR :: Handler Html
getHistoryBGR = do
Entity uid _ <- requireAuth
-- TODO: pagination
sugars <- fmap (map entityVal) $
runDB $ selectList [BloodGlucoseHistoryUid ==. uid] [LimitTo 10]
let bloodGlucoseHistoryDate _ = "TODO" :: Text -- TODO
bloodGlucoseHistoryTime _ = "TODO" :: Text -- TODO
defaultLayout $ do
setTitle "Betty : Blood Sugar Logs"
$(widgetFile "bg.history")
<|file_sep|>updated/Handler/HistoryBG.hs
{-# LANGUAGE NoDisambiguateRecordFields, NoRecordWildCards #-}
module Handler.HistoryBG where
import Import
------------------------------------------------------------------------
getHistoryBGR :: Handler Html
getHistoryBGR = do
Entity uid _ <- requireAuth
-- TODO: pagination
sugars <- fmap (map entityVal) $
runDB $ selectList [BloodGlucoseHistoryUid ==. uid] [LimitTo 10]
let bloodGlucoseHistoryDate _ = "TODO" :: Text -- TODO
bloodGlucoseHistoryTime _ = "TODO" :: Text -- TODO
defaultLayout $ do
setTitle "Betty : Blood Sugar Logs" | 8fc2a76c654cb8f38e422dcc41237696c50b279c | Handler/HistoryBG.hs | Handler/HistoryBG.hs | Haskell |
<|file_sep|>original/src/main/java/com/github/aureliano/damihilogs/annotation/NotNull.java
<|file_sep|>current/src/main/java/com/github/aureliano/damihilogs/annotation/NotNull.java
<|file_sep|>updated/src/main/java/com/github/aureliano/damihilogs/annotation/NotNull.java | package com.github.aureliano.damihilogs.annotation;
import java.lang.annotation.ElementType;
import java.lang.annotation.Retention;
import java.lang.annotation.RetentionPolicy;
import java.lang.annotation.Target;
@Retention(RetentionPolicy.RUNTIME)
@Target(ElementType.METHOD)
public @interface NotNull {
public abstract String message() default "Expected to find a not null value for field ?.";
} | <|file_sep|>original/src/main/java/com/github/aureliano/damihilogs/annotation/NotNull.java
<|file_sep|>current/src/main/java/com/github/aureliano/damihilogs/annotation/NotNull.java
<|file_sep|>updated/src/main/java/com/github/aureliano/damihilogs/annotation/NotNull.java
package com.github.aureliano.damihilogs.annotation;
import java.lang.annotation.ElementType;
import java.lang.annotation.Retention;
import java.lang.annotation.RetentionPolicy;
import java.lang.annotation.Target;
@Retention(RetentionPolicy.RUNTIME)
@Target(ElementType.METHOD)
public @interface NotNull {
public abstract String message() default "Expected to find a not null value for field ?.";
} | 8b4b49f749faab98bc6b79c5e3f3f2f997484a6a | src/main/java/com/github/aureliano/damihilogs/annotation/NotNull.java | src/main/java/com/github/aureliano/damihilogs/annotation/NotNull.java | Java |
<|file_sep|>original/README.md
# security-fairy
This is a tool to that helps give developers a STARTING POINT to build IAM policies that are as minimally permissive as possible.
<|file_sep|>current/README.md
# security-fairy
This is a tool to that helps give developers a STARTING POINT to build IAM policies that are as minimally permissive as possible.
<|file_sep|>updated/README.md | ## Scoping Permissions
While IAM is a fantastic way to protect your AWS Resources/Accounts, properly scoping IAM Policies can be very difficult. There are two general approaches to correctly scoping permissions:
1. Granting overly broad permissions to "get things working" and then reining in the permissions later.
1. Running the application and adding permissions one by one (white-listing) as denies crop up.
The first approach is very common and can greatly increase the speed at which development can move. The catch, however, is that "later" often never comes and the user/system retains its overly permissive state.
The second approach is somewhat less common because it can be *extremely* pain-staking. If an application has even a little complexity, this can take a very long time, and that says nothing about the system breaking when a rarely used feature runs into an access denied that wasn't encountered during the initial permission scoping.
To help streamline the process on the AWS, we built a serverless tool that can bring together the benefits of both.
## Security-Fairy
The intention of this tool is to help give developers a **STARTING POINT** to build properly scoped IAM policies that are as minimally permissive as possible.
This tool is only to be used in a sandbox/dev environment and should **never** be used in Production or other secure environments.
What this tool will do:
- Create a managed policy for a given AWS Role which contains only the "necessary" set of AWS actions the application needs to function. This is determined by what AWS API actions a given role has taken based on the past X days of CloudTrail Logs.
- The resulting policy will serve as a STARTING POINT for an IAM Policy that can be used in later environments | <|file_sep|>original/README.md
# security-fairy
This is a tool to that helps give developers a STARTING POINT to build IAM policies that are as minimally permissive as possible.
<|file_sep|>current/README.md
# security-fairy
This is a tool to that helps give developers a STARTING POINT to build IAM policies that are as minimally permissive as possible.
<|file_sep|>updated/README.md
## Scoping Permissions
While IAM is a fantastic way to protect your AWS Resources/Accounts, properly scoping IAM Policies can be very difficult. There are two general approaches to correctly scoping permissions:
1. Granting overly broad permissions to "get things working" and then reining in the permissions later.
1. Running the application and adding permissions one by one (white-listing) as denies crop up.
The first approach is very common and can greatly increase the speed at which development can move. The catch, however, is that "later" often never comes and the user/system retains its overly permissive state.
The second approach is somewhat less common because it can be *extremely* pain-staking. If an application has even a little complexity, this can take a very long time, and that says nothing about the system breaking when a rarely used feature runs into an access denied that wasn't encountered during the initial permission scoping.
To help streamline the process on the AWS, we built a serverless tool that can bring together the benefits of both.
## Security-Fairy
The intention of this tool is to help give developers a **STARTING POINT** to build properly scoped IAM policies that are as minimally permissive as possible.
This tool is only to be used in a sandbox/dev environment and should **never** be used in Production or other secure environments.
What this tool will do:
- Create a managed policy for a given AWS Role which contains only the "necessary" set of AWS actions the application needs to function. This is determined by what AWS API actions a given role has taken based on the past X days of CloudTrail Logs.
- The resulting policy will serve as a STARTING POINT for an IAM Policy that can be used in later environments | 5dddb90157f8971a6752d531eeaaa1a7b9bb9453 | README.md | README.md | Markdown |
<|file_sep|>lib/ggake/heroku.rb.diff
original:
mkdir_p "package/#{@name}"
cp "heroku/Procfile.#{@name}", "package/#{@name}/Procfile"
cp artefact, "package/#{@name}"
updated:
tap "package/#{@name}" do |pkg_dir|
mkdir_p pkg_dir
cp "heroku/Procfile.#{@name}", "#{pkg_dir}/Procfile"
cp artefact, pkg_dir
end
<|file_sep|>original/lib/ggake/heroku.rb
end
end
class HerokuApp
include Rake::DSL
def initialize(name, heroku_name = nil)
@name, @heroku_name = name, heroku_name
@heroku_name ||= @name
end
def package(artefact)
mkdir_p "package/#{@name}"
cp "heroku/Procfile.#{@name}", "package/#{@name}/Procfile"
cp artefact, "package/#{@name}"
end
def push
heroku "push package/#{@name} --app #{@heroku_name}"
end
end
<|file_sep|>current/lib/ggake/heroku.rb
class HerokuApp
include Rake::DSL
def initialize(name, heroku_name = nil)
@name, @heroku_name = name, heroku_name
@heroku_name ||= @name
end
def package(artefact)
tap "package/#{@name}" do |pkg_dir|
mkdir_p pkg_dir
cp "heroku/Procfile.#{@name}", "#{pkg_dir}/Procfile"
cp artefact, pkg_dir
end
end
def push
heroku "push package/#{@name} --app #{@heroku_name}"
end
end
<|file_sep|>updated/lib/ggake/heroku.rb | def initialize(name, heroku_name = nil)
@name, @heroku_name = name, heroku_name
@heroku_name ||= @name
end
def package(artefact)
tap "package/#{@name}" do |pkg_dir|
mkdir_p pkg_dir
cp "heroku/Procfile.#{@name}", "#{pkg_dir}/Procfile"
cp artefact, pkg_dir
end
end
def push
heroku "push package/#{@name} --app #{@heroku_name}"
end
def config(var)
heroku_config_get(var, @heroku_name)
end
end | <|file_sep|>lib/ggake/heroku.rb.diff
original:
mkdir_p "package/#{@name}"
cp "heroku/Procfile.#{@name}", "package/#{@name}/Procfile"
cp artefact, "package/#{@name}"
updated:
tap "package/#{@name}" do |pkg_dir|
mkdir_p pkg_dir
cp "heroku/Procfile.#{@name}", "#{pkg_dir}/Procfile"
cp artefact, pkg_dir
end
<|file_sep|>original/lib/ggake/heroku.rb
end
end
class HerokuApp
include Rake::DSL
def initialize(name, heroku_name = nil)
@name, @heroku_name = name, heroku_name
@heroku_name ||= @name
end
def package(artefact)
mkdir_p "package/#{@name}"
cp "heroku/Procfile.#{@name}", "package/#{@name}/Procfile"
cp artefact, "package/#{@name}"
end
def push
heroku "push package/#{@name} --app #{@heroku_name}"
end
end
<|file_sep|>current/lib/ggake/heroku.rb
class HerokuApp
include Rake::DSL
def initialize(name, heroku_name = nil)
@name, @heroku_name = name, heroku_name
@heroku_name ||= @name
end
def package(artefact)
tap "package/#{@name}" do |pkg_dir|
mkdir_p pkg_dir
cp "heroku/Procfile.#{@name}", "#{pkg_dir}/Procfile"
cp artefact, pkg_dir
end
end
def push
heroku "push package/#{@name} --app #{@heroku_name}"
end
end
<|file_sep|>updated/lib/ggake/heroku.rb
def initialize(name, heroku_name = nil)
@name, @heroku_name = name, heroku_name
@heroku_name ||= @name
end
def package(artefact)
tap "package/#{@name}" do |pkg_dir|
mkdir_p pkg_dir
cp "heroku/Procfile.#{@name}", "#{pkg_dir}/Procfile"
cp artefact, pkg_dir
end
end
def push
heroku "push package/#{@name} --app #{@heroku_name}"
end
def config(var)
heroku_config_get(var, @heroku_name)
end
end | 8fa4e9f0c179f6c9df05c47291ea51fa40370fd2 | lib/ggake/heroku.rb | lib/ggake/heroku.rb | Ruby |
<|file_sep|>original/saylua/templates/banned.html
{% extends "layout.html" %}
{% block body %}
<h1>You are banned.</h1>
<p class="center">
Sorry, you've been banned from Saylua.
</p>
<p class="center">
Reason: <strong>{{ g.user.ban.reason }}</strong>
</p>
{% endblock %}
<|file_sep|>current/saylua/templates/banned.html
{% extends "layout.html" %}
{% block body %}
<h1>You are banned.</h1>
<p class="center">
Sorry, you've been banned from Saylua.
</p>
<p class="center">
Reason: <strong>{{ g.user.ban.reason }}</strong>
</p>
{% endblock %}
<|file_sep|>updated/saylua/templates/banned.html | {% extends "layout.html" %}
{% block body %}
<h1>You are banned.</h1>
<p class="center">
You have been banned from Saylua.
</p>
<p class="center">
Reason: <strong>{{ g.user.ban.reason }}</strong>
</p>
{% endblock %} | <|file_sep|>original/saylua/templates/banned.html
{% extends "layout.html" %}
{% block body %}
<h1>You are banned.</h1>
<p class="center">
Sorry, you've been banned from Saylua.
</p>
<p class="center">
Reason: <strong>{{ g.user.ban.reason }}</strong>
</p>
{% endblock %}
<|file_sep|>current/saylua/templates/banned.html
{% extends "layout.html" %}
{% block body %}
<h1>You are banned.</h1>
<p class="center">
Sorry, you've been banned from Saylua.
</p>
<p class="center">
Reason: <strong>{{ g.user.ban.reason }}</strong>
</p>
{% endblock %}
<|file_sep|>updated/saylua/templates/banned.html
{% extends "layout.html" %}
{% block body %}
<h1>You are banned.</h1>
<p class="center">
You have been banned from Saylua.
</p>
<p class="center">
Reason: <strong>{{ g.user.ban.reason }}</strong>
</p>
{% endblock %} | cf65b555f73cb3e29cfcbcf004cc2872643cf2c6 | saylua/templates/banned.html | saylua/templates/banned.html | HTML |
<|file_sep|>original/download_test_data.sh
wget -P tests/data/ http://www.segelflug.de/vereine/welt2000/download/WELT2000.TXT
<|file_sep|>current/download_test_data.sh
wget -P tests/data/ http://www.segelflug.de/vereine/welt2000/download/WELT2000.TXT
<|file_sep|>updated/download_test_data.sh | wget -NP tests/data/ http://www.segelflug.de/vereine/welt2000/download/WELT2000.TXT | <|file_sep|>original/download_test_data.sh
wget -P tests/data/ http://www.segelflug.de/vereine/welt2000/download/WELT2000.TXT
<|file_sep|>current/download_test_data.sh
wget -P tests/data/ http://www.segelflug.de/vereine/welt2000/download/WELT2000.TXT
<|file_sep|>updated/download_test_data.sh
wget -NP tests/data/ http://www.segelflug.de/vereine/welt2000/download/WELT2000.TXT | 01e74d61fe933c3ab6e39fe40e13b4b61f4d9b0b | download_test_data.sh | download_test_data.sh | Shell |
<|file_sep|>original/src/Application/Sonata/ProductBundle/Resources/views/Training/view.html.twig
{% set variations_properties = ['level', 'duration'] %}
{% block product_title %}{{ product.name }}{% endblock %}
{% block product_delivery %}{% endblock %}
{% block product_properties_before_price %}
{% if not product.isMaster %}
<dt style="width: auto;">{{ 'training.level_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.level|trans([], 'SonataProductBundle') }}</dd>
<dt style="width: auto;">{{ 'training.instructor_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.instructorName|trans([], 'SonataProductBundle') }}</dd>
{% if product.startDate %}
<dt style="width: auto;">{{ 'training.start_date_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.startDate|date() }}</dd>
{% endif %}
<dt style="width: auto;">{{ 'training.duration_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.duration|trans([], 'SonataProductBundle') }}</dd>
{% endif %}
{% endblock %}
<|file_sep|>current/src/Application/Sonata/ProductBundle/Resources/views/Training/view.html.twig
{% set variations_properties = ['level', 'duration'] %}
{% block product_title %}{{ product.name }}{% endblock %}
{% block product_delivery %}{% endblock %}
{% block product_properties_before_price %}
{% if not product.isMaster %}
<dt style="width: auto;">{{ 'training.level_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.level|trans([], 'SonataProductBundle') }}</dd>
<dt style="width: auto;">{{ 'training.instructor_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.instructorName|trans([], 'SonataProductBundle') }}</dd>
{% if product.startDate %}
<dt style="width: auto;">{{ 'training.start_date_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.startDate|date() }}</dd>
{% endif %}
<dt style="width: auto;">{{ 'training.duration_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.duration|trans([], 'SonataProductBundle') }}</dd>
{% endif %}
{% endblock %}
<|file_sep|>updated/src/Application/Sonata/ProductBundle/Resources/views/Training/view.html.twig | (c) Thomas Rabaix <thomas.rabaix@sonata-project.org>
For the full copyright and license information, please view the LICENSE
file that was distributed with this source code.
#}
{#
Default product template file, you can
=> customize the global template : Product:Product:view.twig
=> rewrite your own template
#}
{% extends 'SonataProductBundle:Product:view.html.twig' %}
{% set variations_properties = ['level', 'duration'] %}
{% block product_title %}{{ product.name }}{% endblock %}
{% block product_delivery %}{% endblock %} | <|file_sep|>original/src/Application/Sonata/ProductBundle/Resources/views/Training/view.html.twig
{% set variations_properties = ['level', 'duration'] %}
{% block product_title %}{{ product.name }}{% endblock %}
{% block product_delivery %}{% endblock %}
{% block product_properties_before_price %}
{% if not product.isMaster %}
<dt style="width: auto;">{{ 'training.level_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.level|trans([], 'SonataProductBundle') }}</dd>
<dt style="width: auto;">{{ 'training.instructor_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.instructorName|trans([], 'SonataProductBundle') }}</dd>
{% if product.startDate %}
<dt style="width: auto;">{{ 'training.start_date_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.startDate|date() }}</dd>
{% endif %}
<dt style="width: auto;">{{ 'training.duration_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.duration|trans([], 'SonataProductBundle') }}</dd>
{% endif %}
{% endblock %}
<|file_sep|>current/src/Application/Sonata/ProductBundle/Resources/views/Training/view.html.twig
{% set variations_properties = ['level', 'duration'] %}
{% block product_title %}{{ product.name }}{% endblock %}
{% block product_delivery %}{% endblock %}
{% block product_properties_before_price %}
{% if not product.isMaster %}
<dt style="width: auto;">{{ 'training.level_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.level|trans([], 'SonataProductBundle') }}</dd>
<dt style="width: auto;">{{ 'training.instructor_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.instructorName|trans([], 'SonataProductBundle') }}</dd>
{% if product.startDate %}
<dt style="width: auto;">{{ 'training.start_date_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.startDate|date() }}</dd>
{% endif %}
<dt style="width: auto;">{{ 'training.duration_title'|trans([], 'SonataProductBundle') }}</dt>
<dd style="margin-left: 110px;">{{ product.duration|trans([], 'SonataProductBundle') }}</dd>
{% endif %}
{% endblock %}
<|file_sep|>updated/src/Application/Sonata/ProductBundle/Resources/views/Training/view.html.twig
(c) Thomas Rabaix <thomas.rabaix@sonata-project.org>
For the full copyright and license information, please view the LICENSE
file that was distributed with this source code.
#}
{#
Default product template file, you can
=> customize the global template : Product:Product:view.twig
=> rewrite your own template
#}
{% extends 'SonataProductBundle:Product:view.html.twig' %}
{% set variations_properties = ['level', 'duration'] %}
{% block product_title %}{{ product.name }}{% endblock %}
{% block product_delivery %}{% endblock %} | c9958fe3f0ea330abfe7ff4fb5f926801eb7b2b4 | src/Application/Sonata/ProductBundle/Resources/views/Training/view.html.twig | src/Application/Sonata/ProductBundle/Resources/views/Training/view.html.twig | Twig |
<|file_sep|>original/test/unit/lookups/maxmind_local_test.rb
def setup
Geocoder.configure(ip_lookup: :maxmind_local)
end
def test_result_attributes
result = Geocoder.search('8.8.8.8').first
assert_equal result.address, 'Mountain View, CA 94043, United States'
assert_equal result.city, 'Mountain View'
assert_equal result.state, 'CA'
assert_equal result.country, 'United States'
assert_equal result.country_code, 'USA'
assert_equal result.postal_code, '94043'
assert_equal result.latitude, 37.41919999999999
assert_equal result.longitude, -122.0574
end
def test_loopback
results = Geocoder.search('127.0.0.1')
assert_equal [], results
end
<|file_sep|>current/test/unit/lookups/maxmind_local_test.rb
def setup
Geocoder.configure(ip_lookup: :maxmind_local)
end
def test_result_attributes
result = Geocoder.search('8.8.8.8').first
assert_equal result.address, 'Mountain View, CA 94043, United States'
assert_equal result.city, 'Mountain View'
assert_equal result.state, 'CA'
assert_equal result.country, 'United States'
assert_equal result.country_code, 'USA'
assert_equal result.postal_code, '94043'
assert_equal result.latitude, 37.41919999999999
assert_equal result.longitude, -122.0574
end
def test_loopback
results = Geocoder.search('127.0.0.1')
assert_equal [], results
end
<|file_sep|>updated/test/unit/lookups/maxmind_local_test.rb |
def setup
Geocoder.configure(ip_lookup: :maxmind_local)
end
def test_result_attributes
result = Geocoder.search('8.8.8.8').first
assert_equal 'Mountain View, CA 94043, United States', result.address
assert_equal 'Mountain View', result.city
assert_equal 'CA', result.state
assert_equal 'United States', result.country
assert_equal 'USA', result.country_code
assert_equal '94043', result.postal_code
assert_equal 37.41919999999999, result.latitude
assert_equal -122.0574, result.longitude
end
def test_loopback
results = Geocoder.search('127.0.0.1')
assert_equal [], results
end | <|file_sep|>original/test/unit/lookups/maxmind_local_test.rb
def setup
Geocoder.configure(ip_lookup: :maxmind_local)
end
def test_result_attributes
result = Geocoder.search('8.8.8.8').first
assert_equal result.address, 'Mountain View, CA 94043, United States'
assert_equal result.city, 'Mountain View'
assert_equal result.state, 'CA'
assert_equal result.country, 'United States'
assert_equal result.country_code, 'USA'
assert_equal result.postal_code, '94043'
assert_equal result.latitude, 37.41919999999999
assert_equal result.longitude, -122.0574
end
def test_loopback
results = Geocoder.search('127.0.0.1')
assert_equal [], results
end
<|file_sep|>current/test/unit/lookups/maxmind_local_test.rb
def setup
Geocoder.configure(ip_lookup: :maxmind_local)
end
def test_result_attributes
result = Geocoder.search('8.8.8.8').first
assert_equal result.address, 'Mountain View, CA 94043, United States'
assert_equal result.city, 'Mountain View'
assert_equal result.state, 'CA'
assert_equal result.country, 'United States'
assert_equal result.country_code, 'USA'
assert_equal result.postal_code, '94043'
assert_equal result.latitude, 37.41919999999999
assert_equal result.longitude, -122.0574
end
def test_loopback
results = Geocoder.search('127.0.0.1')
assert_equal [], results
end
<|file_sep|>updated/test/unit/lookups/maxmind_local_test.rb
def setup
Geocoder.configure(ip_lookup: :maxmind_local)
end
def test_result_attributes
result = Geocoder.search('8.8.8.8').first
assert_equal 'Mountain View, CA 94043, United States', result.address
assert_equal 'Mountain View', result.city
assert_equal 'CA', result.state
assert_equal 'United States', result.country
assert_equal 'USA', result.country_code
assert_equal '94043', result.postal_code
assert_equal 37.41919999999999, result.latitude
assert_equal -122.0574, result.longitude
end
def test_loopback
results = Geocoder.search('127.0.0.1')
assert_equal [], results
end | 076521a2d9a56437cb419063bf1bc5f70d20cf56 | test/unit/lookups/maxmind_local_test.rb | test/unit/lookups/maxmind_local_test.rb | Ruby |
<|file_sep|>original/requirements/prod.txt
# General
flask==1.1.2
requests==2.25.1
beautifulsoup4==4.9.3
eventlet==0.30.1
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
jsmin==2.2.2
# Sockets
flask_socketio==5.0.1
python_engineio==4.0.0
python_socketio==5.0.4
# Celery
celery==5.0.5
redis==3.5.3
# Octave interaction
<|file_sep|>current/requirements/prod.txt
# General
flask==1.1.2
requests==2.25.1
beautifulsoup4==4.9.3
eventlet==0.30.1
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
jsmin==2.2.2
# Sockets
flask_socketio==5.0.1
python_engineio==4.0.0
python_socketio==5.0.4
# Celery
celery==5.0.5
redis==3.5.3
# Octave interaction
<|file_sep|>updated/requirements/prod.txt | # General
flask==1.1.2
requests==2.25.1
beautifulsoup4==4.9.3
eventlet==0.30.2
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
jsmin==2.2.2
# Sockets
flask_socketio==5.0.1
python_engineio==4.0.0
python_socketio==5.0.4
# Celery
celery==5.0.5
redis==3.5.3
# Octave interaction | <|file_sep|>original/requirements/prod.txt
# General
flask==1.1.2
requests==2.25.1
beautifulsoup4==4.9.3
eventlet==0.30.1
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
jsmin==2.2.2
# Sockets
flask_socketio==5.0.1
python_engineio==4.0.0
python_socketio==5.0.4
# Celery
celery==5.0.5
redis==3.5.3
# Octave interaction
<|file_sep|>current/requirements/prod.txt
# General
flask==1.1.2
requests==2.25.1
beautifulsoup4==4.9.3
eventlet==0.30.1
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
jsmin==2.2.2
# Sockets
flask_socketio==5.0.1
python_engineio==4.0.0
python_socketio==5.0.4
# Celery
celery==5.0.5
redis==3.5.3
# Octave interaction
<|file_sep|>updated/requirements/prod.txt
# General
flask==1.1.2
requests==2.25.1
beautifulsoup4==4.9.3
eventlet==0.30.2
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
jsmin==2.2.2
# Sockets
flask_socketio==5.0.1
python_engineio==4.0.0
python_socketio==5.0.4
# Celery
celery==5.0.5
redis==3.5.3
# Octave interaction | 7e7e405996da766d8f598eb1765a1920b7323da7 | requirements/prod.txt | requirements/prod.txt | Text |
<|file_sep|>original/.travis.yml
language: objective-c
osx_image: xcode7.3
before_install:
- brew update >/dev/null
- brew list autoconf || brew install autoconf
- brew list automake || brew install automake
- brew list doxygen || brew install doxygen
- brew uninstall libtool
- brew list libtool || brew install libtool
- cd libmpdclient
- sed -i .bak 's/libtoolize/glibtoolize/g' autogen.sh
- ./autogen.sh
- make
- cd ..
install:
- version_number=`git describe --tags --abbrev=0 2>/dev/null || echo 0.0`
- if [[ $? != 0 ]]; then version_number="0.0"; fi
- agvtool new-version -all $TRAVIS_BUILD_NUMBER
- agvtool new-marketing-version $version_number
- xcodebuild -target kmbmpdc -configuration Release
before_script:
<|file_sep|>current/.travis.yml
language: objective-c
osx_image: xcode7.3
before_install:
- brew update >/dev/null
- brew list autoconf || brew install autoconf
- brew list automake || brew install automake
- brew list doxygen || brew install doxygen
- brew uninstall libtool
- brew list libtool || brew install libtool
- cd libmpdclient
- sed -i .bak 's/libtoolize/glibtoolize/g' autogen.sh
- ./autogen.sh
- make
- cd ..
install:
- version_number=`git describe --tags --abbrev=0 2>/dev/null || echo 0.0`
- if [[ $? != 0 ]]; then version_number="0.0"; fi
- agvtool new-version -all $TRAVIS_BUILD_NUMBER
- agvtool new-marketing-version $version_number
- xcodebuild -target kmbmpdc -configuration Release
before_script:
<|file_sep|>updated/.travis.yml | language: objective-c
osx_image: xcode8
before_install:
- brew update >/dev/null
- brew list autoconf || brew install autoconf
- brew list automake || brew install automake
- brew list doxygen || brew install doxygen
- brew uninstall libtool
- brew list libtool || brew install libtool
- cd libmpdclient
- sed -i .bak 's/libtoolize/glibtoolize/g' autogen.sh
- ./autogen.sh
- make
- cd ..
install:
- version_number=`git describe --tags --abbrev=0 2>/dev/null || echo 0.0`
- if [[ $? != 0 ]]; then version_number="0.0"; fi
- agvtool new-version -all $TRAVIS_BUILD_NUMBER
- agvtool new-marketing-version $version_number
- xcodebuild -target kmbmpdc -configuration Release
before_script: | <|file_sep|>original/.travis.yml
language: objective-c
osx_image: xcode7.3
before_install:
- brew update >/dev/null
- brew list autoconf || brew install autoconf
- brew list automake || brew install automake
- brew list doxygen || brew install doxygen
- brew uninstall libtool
- brew list libtool || brew install libtool
- cd libmpdclient
- sed -i .bak 's/libtoolize/glibtoolize/g' autogen.sh
- ./autogen.sh
- make
- cd ..
install:
- version_number=`git describe --tags --abbrev=0 2>/dev/null || echo 0.0`
- if [[ $? != 0 ]]; then version_number="0.0"; fi
- agvtool new-version -all $TRAVIS_BUILD_NUMBER
- agvtool new-marketing-version $version_number
- xcodebuild -target kmbmpdc -configuration Release
before_script:
<|file_sep|>current/.travis.yml
language: objective-c
osx_image: xcode7.3
before_install:
- brew update >/dev/null
- brew list autoconf || brew install autoconf
- brew list automake || brew install automake
- brew list doxygen || brew install doxygen
- brew uninstall libtool
- brew list libtool || brew install libtool
- cd libmpdclient
- sed -i .bak 's/libtoolize/glibtoolize/g' autogen.sh
- ./autogen.sh
- make
- cd ..
install:
- version_number=`git describe --tags --abbrev=0 2>/dev/null || echo 0.0`
- if [[ $? != 0 ]]; then version_number="0.0"; fi
- agvtool new-version -all $TRAVIS_BUILD_NUMBER
- agvtool new-marketing-version $version_number
- xcodebuild -target kmbmpdc -configuration Release
before_script:
<|file_sep|>updated/.travis.yml
language: objective-c
osx_image: xcode8
before_install:
- brew update >/dev/null
- brew list autoconf || brew install autoconf
- brew list automake || brew install automake
- brew list doxygen || brew install doxygen
- brew uninstall libtool
- brew list libtool || brew install libtool
- cd libmpdclient
- sed -i .bak 's/libtoolize/glibtoolize/g' autogen.sh
- ./autogen.sh
- make
- cd ..
install:
- version_number=`git describe --tags --abbrev=0 2>/dev/null || echo 0.0`
- if [[ $? != 0 ]]; then version_number="0.0"; fi
- agvtool new-version -all $TRAVIS_BUILD_NUMBER
- agvtool new-marketing-version $version_number
- xcodebuild -target kmbmpdc -configuration Release
before_script: | 49daf63786e450319039c3daacb38c6fc9f9f64d | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/.bob.json
{
"build": "clean lint complexity test coverage doc",
"lint": {
"type": "jshint"
},
"complexity": {
"type": "plato"
},
"test": {
"type": "buster"
},
"coverage": {
"type": "buster-istanbul"
},
"doc": {
"type": "dox-foundation"
}
}
<|file_sep|>current/.bob.json
{
"build": "clean lint complexity test coverage doc",
"lint": {
"type": "jshint"
},
"complexity": {
"type": "plato"
},
"test": {
"type": "buster"
},
"coverage": {
"type": "buster-istanbul"
},
"doc": {
"type": "dox-foundation"
}
}
<|file_sep|>updated/.bob.json | {
"build": "clean lint complexity test coverage doc",
"lint": {
"type": "jshint"
},
"complexity": {
"type": "plato"
},
"test": {
"type": "buster"
},
"coverage": {
"type": "buster-istanbul"
},
"doc": {
"type": "dox-foundation"
},
"site": {
"type": "gh-pages-cli"
}
} | <|file_sep|>original/.bob.json
{
"build": "clean lint complexity test coverage doc",
"lint": {
"type": "jshint"
},
"complexity": {
"type": "plato"
},
"test": {
"type": "buster"
},
"coverage": {
"type": "buster-istanbul"
},
"doc": {
"type": "dox-foundation"
}
}
<|file_sep|>current/.bob.json
{
"build": "clean lint complexity test coverage doc",
"lint": {
"type": "jshint"
},
"complexity": {
"type": "plato"
},
"test": {
"type": "buster"
},
"coverage": {
"type": "buster-istanbul"
},
"doc": {
"type": "dox-foundation"
}
}
<|file_sep|>updated/.bob.json
{
"build": "clean lint complexity test coverage doc",
"lint": {
"type": "jshint"
},
"complexity": {
"type": "plato"
},
"test": {
"type": "buster"
},
"coverage": {
"type": "buster-istanbul"
},
"doc": {
"type": "dox-foundation"
},
"site": {
"type": "gh-pages-cli"
}
} | c3f770413fd2989ce0b872100d770e528a2a5adf | .bob.json | .bob.json | JSON |
<|file_sep|>original/src/Ilios/WebBundle/Controller/IndexController.php
class IndexController extends Controller
{
public function indexAction()
{
$fs = $this->get('ilioscore.symfonyfilesystem');
$path = $this->getParameter('kernel.cache_dir') . '/' . UpdateFrontendCommand::CACHE_FILE_NAME;
if (!$fs->exists($path)) {
throw new \Exception(
'Unable to load the index file. Run ilios:maintenance:update-frontend to create it.'
);
}
$contents = $fs->readFile($path);
$response = new Response($contents);
$response->headers->set('Content-Type', 'text/html');
$response->setPublic();
$response->setMaxAge(60);
<|file_sep|>current/src/Ilios/WebBundle/Controller/IndexController.php
class IndexController extends Controller
{
public function indexAction()
{
$fs = $this->get('ilioscore.symfonyfilesystem');
$path = $this->getParameter('kernel.cache_dir') . '/' . UpdateFrontendCommand::CACHE_FILE_NAME;
if (!$fs->exists($path)) {
throw new \Exception(
'Unable to load the index file. Run ilios:maintenance:update-frontend to create it.'
);
}
$contents = $fs->readFile($path);
$response = new Response($contents);
$response->headers->set('Content-Type', 'text/html');
$response->setPublic();
$response->setMaxAge(60);
<|file_sep|>updated/src/Ilios/WebBundle/Controller/IndexController.php |
class IndexController extends Controller
{
public function indexAction()
{
$fs = $this->get('ilioscore.symfonyfilesystem');
$path = $this->getParameter('kernel.cache_dir') . '/' . UpdateFrontendCommand::CACHE_FILE_NAME;
if (!$fs->exists($path)) {
throw new \Exception(
"Unable to load the index file at {$path}. Run ilios:maintenance:update-frontend to create it."
);
}
$contents = $fs->readFile($path);
$response = new Response($contents);
$response->headers->set('Content-Type', 'text/html');
$response->setPublic();
$response->setMaxAge(60);
| <|file_sep|>original/src/Ilios/WebBundle/Controller/IndexController.php
class IndexController extends Controller
{
public function indexAction()
{
$fs = $this->get('ilioscore.symfonyfilesystem');
$path = $this->getParameter('kernel.cache_dir') . '/' . UpdateFrontendCommand::CACHE_FILE_NAME;
if (!$fs->exists($path)) {
throw new \Exception(
'Unable to load the index file. Run ilios:maintenance:update-frontend to create it.'
);
}
$contents = $fs->readFile($path);
$response = new Response($contents);
$response->headers->set('Content-Type', 'text/html');
$response->setPublic();
$response->setMaxAge(60);
<|file_sep|>current/src/Ilios/WebBundle/Controller/IndexController.php
class IndexController extends Controller
{
public function indexAction()
{
$fs = $this->get('ilioscore.symfonyfilesystem');
$path = $this->getParameter('kernel.cache_dir') . '/' . UpdateFrontendCommand::CACHE_FILE_NAME;
if (!$fs->exists($path)) {
throw new \Exception(
'Unable to load the index file. Run ilios:maintenance:update-frontend to create it.'
);
}
$contents = $fs->readFile($path);
$response = new Response($contents);
$response->headers->set('Content-Type', 'text/html');
$response->setPublic();
$response->setMaxAge(60);
<|file_sep|>updated/src/Ilios/WebBundle/Controller/IndexController.php
class IndexController extends Controller
{
public function indexAction()
{
$fs = $this->get('ilioscore.symfonyfilesystem');
$path = $this->getParameter('kernel.cache_dir') . '/' . UpdateFrontendCommand::CACHE_FILE_NAME;
if (!$fs->exists($path)) {
throw new \Exception(
"Unable to load the index file at {$path}. Run ilios:maintenance:update-frontend to create it."
);
}
$contents = $fs->readFile($path);
$response = new Response($contents);
$response->headers->set('Content-Type', 'text/html');
$response->setPublic();
$response->setMaxAge(60);
| 418dd74f2d89129e03fdbde232e1a846b607c190 | src/Ilios/WebBundle/Controller/IndexController.php | src/Ilios/WebBundle/Controller/IndexController.php | PHP |
<|file_sep|>original/submit.sh
echo "Program usage: $0 <directory> <output>" >&2
exit 1
fi
}
submit() {
echo "$@" | qsub -S /bin/bash -cwd -sync n -pe parallel 24 -N build-images
}
build_images() {
local directory="$1"
local output_name="$2"
echo /home/sw/anaconda/bin/python ./plot.py -o ${output_name} "${directory}/*.fits"
}
main() {
verify_args "$@"
submit $(build_images "$1" "$2")
}
main "$@"
<|file_sep|>current/submit.sh
echo "Program usage: $0 <directory> <output>" >&2
exit 1
fi
}
submit() {
echo "$@" | qsub -S /bin/bash -cwd -sync n -pe parallel 24 -N build-images
}
build_images() {
local directory="$1"
local output_name="$2"
echo /home/sw/anaconda/bin/python ./plot.py -o ${output_name} "${directory}/*.fits"
}
main() {
verify_args "$@"
submit $(build_images "$1" "$2")
}
main "$@"
<|file_sep|>updated/submit.sh | fi
}
submit() {
echo "$@" | qsub -S /bin/bash -cwd -sync n -pe parallel 24 -N build-images
}
build_images() {
local directory="$1"
local output_name="$2"
echo /home/sw/anaconda/bin/python ./plot.py -o ${output_name} "$(getfiles ${directory})"
}
getfiles() {
local directory="$1"
ls ${directory}/proc*.fits | grep -v skybkgmap
}
main() {
verify_args "$@"
submit $(build_images "$1" "$2") | <|file_sep|>original/submit.sh
echo "Program usage: $0 <directory> <output>" >&2
exit 1
fi
}
submit() {
echo "$@" | qsub -S /bin/bash -cwd -sync n -pe parallel 24 -N build-images
}
build_images() {
local directory="$1"
local output_name="$2"
echo /home/sw/anaconda/bin/python ./plot.py -o ${output_name} "${directory}/*.fits"
}
main() {
verify_args "$@"
submit $(build_images "$1" "$2")
}
main "$@"
<|file_sep|>current/submit.sh
echo "Program usage: $0 <directory> <output>" >&2
exit 1
fi
}
submit() {
echo "$@" | qsub -S /bin/bash -cwd -sync n -pe parallel 24 -N build-images
}
build_images() {
local directory="$1"
local output_name="$2"
echo /home/sw/anaconda/bin/python ./plot.py -o ${output_name} "${directory}/*.fits"
}
main() {
verify_args "$@"
submit $(build_images "$1" "$2")
}
main "$@"
<|file_sep|>updated/submit.sh
fi
}
submit() {
echo "$@" | qsub -S /bin/bash -cwd -sync n -pe parallel 24 -N build-images
}
build_images() {
local directory="$1"
local output_name="$2"
echo /home/sw/anaconda/bin/python ./plot.py -o ${output_name} "$(getfiles ${directory})"
}
getfiles() {
local directory="$1"
ls ${directory}/proc*.fits | grep -v skybkgmap
}
main() {
verify_args "$@"
submit $(build_images "$1" "$2") | da7e3c5a5e9b5bd392f271bc1c1e871f9ae4fa4c | submit.sh | submit.sh | Shell |
<|file_sep|>original/tests/testloop.sh
#!/bin/bash
while [ true ]; do
phpunit -c phpunit.xml
sleep 3
done
<|file_sep|>current/tests/testloop.sh
#!/bin/bash
while [ true ]; do
phpunit -c phpunit.xml
sleep 3
done
<|file_sep|>updated/tests/testloop.sh | #!/bin/bash
while [ true ]; do
../vendor/bin/phpunit -c phpunit.xml
sleep 3
done | <|file_sep|>original/tests/testloop.sh
#!/bin/bash
while [ true ]; do
phpunit -c phpunit.xml
sleep 3
done
<|file_sep|>current/tests/testloop.sh
#!/bin/bash
while [ true ]; do
phpunit -c phpunit.xml
sleep 3
done
<|file_sep|>updated/tests/testloop.sh
#!/bin/bash
while [ true ]; do
../vendor/bin/phpunit -c phpunit.xml
sleep 3
done | 64e4eac7ef6d8f67e56e18cd8b7ac459e680059c | tests/testloop.sh | tests/testloop.sh | Shell |
<|file_sep|>original/slave/functions/network.php
<?php
function collect_networkUsage($debug, $log) {
$usage = shell_exec('ifstat -S 0.1 1');
$usage = explode(" ", $usage);
if ($debug == TRUE) {
debug("[DEBUG_COLLECT] Network Usage collected: IN:" . $usage[0] . " | OUT: " . $usage[1] . "\n", $log);
}
$return = $usage[0] . ',' . $usage[1];
return $return;
}
?>
<|file_sep|>current/slave/functions/network.php
<?php
function collect_networkUsage($debug, $log) {
$usage = shell_exec('ifstat -S 0.1 1');
$usage = explode(" ", $usage);
if ($debug == TRUE) {
debug("[DEBUG_COLLECT] Network Usage collected: IN:" . $usage[0] . " | OUT: " . $usage[1] . "\n", $log);
}
$return = $usage[0] . ',' . $usage[1];
return $return;
}
?>
<|file_sep|>updated/slave/functions/network.php | <?php
function collect_networkUsage($debug, $log) {
$usage = trim(exec('ifstat -S 0.1 1'));
$usage = explode(" ", $usage);
if ($debug == TRUE) {
debug("[DEBUG_COLLECT] Network Usage collected: IN:" . $usage[0] . " | OUT: " . $usage[1] . "\n", $log);
}
$return = $usage[0] . ',' . $usage[1];
return $return;
}
?> | <|file_sep|>original/slave/functions/network.php
<?php
function collect_networkUsage($debug, $log) {
$usage = shell_exec('ifstat -S 0.1 1');
$usage = explode(" ", $usage);
if ($debug == TRUE) {
debug("[DEBUG_COLLECT] Network Usage collected: IN:" . $usage[0] . " | OUT: " . $usage[1] . "\n", $log);
}
$return = $usage[0] . ',' . $usage[1];
return $return;
}
?>
<|file_sep|>current/slave/functions/network.php
<?php
function collect_networkUsage($debug, $log) {
$usage = shell_exec('ifstat -S 0.1 1');
$usage = explode(" ", $usage);
if ($debug == TRUE) {
debug("[DEBUG_COLLECT] Network Usage collected: IN:" . $usage[0] . " | OUT: " . $usage[1] . "\n", $log);
}
$return = $usage[0] . ',' . $usage[1];
return $return;
}
?>
<|file_sep|>updated/slave/functions/network.php
<?php
function collect_networkUsage($debug, $log) {
$usage = trim(exec('ifstat -S 0.1 1'));
$usage = explode(" ", $usage);
if ($debug == TRUE) {
debug("[DEBUG_COLLECT] Network Usage collected: IN:" . $usage[0] . " | OUT: " . $usage[1] . "\n", $log);
}
$return = $usage[0] . ',' . $usage[1];
return $return;
}
?> | ec9fcd770211fe8882f50db0a5958f94458aa584 | slave/functions/network.php | slave/functions/network.php | PHP |
<|file_sep|>original/pom.xml
</dependencies>
<profiles>
<profile>
<id>excluding_mac</id>
<activation>
<os>
<family>!mac</family>
</os>
</activation>
<dependencies>
<dependency>
<groupId>com.sun</groupId>
<artifactId>tools</artifactId>
<version>1.5.0</version>
<scope>system</scope>
<systemPath>${java.home}/../lib/tools.jar</systemPath>
</dependency>
</dependencies>
</profile>
</profiles>
</project>
<|file_sep|>current/pom.xml
</dependencies>
<profiles>
<profile>
<id>excluding_mac</id>
<activation>
<os>
<family>!mac</family>
</os>
</activation>
<dependencies>
<dependency>
<groupId>com.sun</groupId>
<artifactId>tools</artifactId>
<version>1.5.0</version>
<scope>system</scope>
<systemPath>${java.home}/../lib/tools.jar</systemPath>
</dependency>
</dependencies>
</profile>
</profiles>
</project>
<|file_sep|>updated/pom.xml | <artifactId>maven-compiler-plugin</artifactId>
<configuration>
<target>1.8</target>
</configuration>
</plugin>
</plugins>
</build>
<repositories>
<repository>
<id>spigot</id>
<url>https://hub.spigotmc.org/nexus/content/groups/public/</url>
</repository>
</repositories>
<dependencies>
<dependency>
<groupId>org.spigotmc</groupId>
<artifactId>spigot-api</artifactId>
<version>1.8.8-R0.1-SNAPSHOT</version>
</dependency>
</dependencies>
</project> | <|file_sep|>original/pom.xml
</dependencies>
<profiles>
<profile>
<id>excluding_mac</id>
<activation>
<os>
<family>!mac</family>
</os>
</activation>
<dependencies>
<dependency>
<groupId>com.sun</groupId>
<artifactId>tools</artifactId>
<version>1.5.0</version>
<scope>system</scope>
<systemPath>${java.home}/../lib/tools.jar</systemPath>
</dependency>
</dependencies>
</profile>
</profiles>
</project>
<|file_sep|>current/pom.xml
</dependencies>
<profiles>
<profile>
<id>excluding_mac</id>
<activation>
<os>
<family>!mac</family>
</os>
</activation>
<dependencies>
<dependency>
<groupId>com.sun</groupId>
<artifactId>tools</artifactId>
<version>1.5.0</version>
<scope>system</scope>
<systemPath>${java.home}/../lib/tools.jar</systemPath>
</dependency>
</dependencies>
</profile>
</profiles>
</project>
<|file_sep|>updated/pom.xml
<artifactId>maven-compiler-plugin</artifactId>
<configuration>
<target>1.8</target>
</configuration>
</plugin>
</plugins>
</build>
<repositories>
<repository>
<id>spigot</id>
<url>https://hub.spigotmc.org/nexus/content/groups/public/</url>
</repository>
</repositories>
<dependencies>
<dependency>
<groupId>org.spigotmc</groupId>
<artifactId>spigot-api</artifactId>
<version>1.8.8-R0.1-SNAPSHOT</version>
</dependency>
</dependencies>
</project> | 8b01e59a83b14f4308b05aad855a3918a0cb4a48 | pom.xml | pom.xml | XML |
<|file_sep|>original/lib/letter_opener.rb
require "fileutils"
require "digest/sha1"
require "cgi"
require "letter_opener/message"
require "letter_opener/delivery_method"
require "letter_opener/railtie" if defined? Rails
<|file_sep|>current/lib/letter_opener.rb
require "fileutils"
require "digest/sha1"
require "cgi"
require "letter_opener/message"
require "letter_opener/delivery_method"
require "letter_opener/railtie" if defined? Rails
<|file_sep|>updated/lib/letter_opener.rb | require "fileutils"
require "digest/sha1"
require "cgi"
require "launchy"
require "letter_opener/message"
require "letter_opener/delivery_method"
require "letter_opener/railtie" if defined? Rails | <|file_sep|>original/lib/letter_opener.rb
require "fileutils"
require "digest/sha1"
require "cgi"
require "letter_opener/message"
require "letter_opener/delivery_method"
require "letter_opener/railtie" if defined? Rails
<|file_sep|>current/lib/letter_opener.rb
require "fileutils"
require "digest/sha1"
require "cgi"
require "letter_opener/message"
require "letter_opener/delivery_method"
require "letter_opener/railtie" if defined? Rails
<|file_sep|>updated/lib/letter_opener.rb
require "fileutils"
require "digest/sha1"
require "cgi"
require "launchy"
require "letter_opener/message"
require "letter_opener/delivery_method"
require "letter_opener/railtie" if defined? Rails | 4c3ed3ca379257ce8bacfffa69978669cb80b1e7 | lib/letter_opener.rb | lib/letter_opener.rb | Ruby |
<|file_sep|>src/Views/tickets/partials/modal_attachment_edit.blade.php.diff
original:
@include('panichd::tickets.partials.attachment_form_fields')
updated:
<!-- Div edit attachment -->
<fieldset id="edit_ticket_attachment">
@include('panichd::tickets.partials.attachment_form_fields')
<|file_sep|>src/Views/tickets/partials/modal_attachment_edit.blade.php.diff
original:
{!! CollectiveForm::hidden(null, 'modal-attachment-edit', ['id'=>'hide_modal_id']) !!}
updated:
{!! CollectiveForm::hidden(null, 'modal-attachment-edit', ['id'=>'hide_modal_id']) !!}
<|file_sep|>original/src/Views/tickets/partials/modal_attachment_edit.blade.php
<div class="modal-header">
<h4 class="modal-title">{{ trans('panichd::lang.attachment-edit') }}</h4>
<button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">{{ trans('panichd::lang.flash-x') }}</span></button>
</div>
<div class="modal-body">
@include('panichd::tickets.partials.attachment_form_fields')
{!! CollectiveForm::hidden(null, 'modal-attachment-edit', ['id'=>'hide_modal_id']) !!}
<div class="text-right col-md-12">
{!! CollectiveForm::button( trans('panichd::lang.update'), [
'type' => 'button',
'class' => 'btn btn-primary attachment_form_submit'
]) !!}
</div>
</div>
</div>
</div>
</div>
@append
<|file_sep|>current/src/Views/tickets/partials/modal_attachment_edit.blade.php
<button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">{{ trans('panichd::lang.flash-x') }}</span></button>
</div>
<div class="modal-body">
<!-- Div edit attachment -->
<fieldset id="edit_ticket_attachment">
@include('panichd::tickets.partials.attachment_form_fields')
{!! CollectiveForm::hidden(null, 'modal-attachment-edit', ['id'=>'hide_modal_id']) !!}
<div class="text-right col-md-12">
{!! CollectiveForm::button( trans('panichd::lang.update'), [
'type' => 'button',
'class' => 'btn btn-primary attachment_form_submit'
]) !!}
</div>
</div>
</div>
</div>
</div>
@append
<|file_sep|>updated/src/Views/tickets/partials/modal_attachment_edit.blade.php | <div class="modal-body">
<!-- Div edit attachment -->
<fieldset id="edit_ticket_attachment">
@include('panichd::tickets.partials.attachment_form_fields')
{!! CollectiveForm::hidden(null, 'modal-attachment-edit', ['id'=>'hide_modal_id']) !!}
<div class="text-right col-md-12">
{!! CollectiveForm::button( trans('panichd::lang.update'), [
'type' => 'button',
'class' => 'btn btn-primary attachment_form_submit',
'data-back-div' => 'ticket_attached'
]) !!}
</div>
</fieldset>
</div>
</div>
</div>
</div>
@append | <|file_sep|>src/Views/tickets/partials/modal_attachment_edit.blade.php.diff
original:
@include('panichd::tickets.partials.attachment_form_fields')
updated:
<!-- Div edit attachment -->
<fieldset id="edit_ticket_attachment">
@include('panichd::tickets.partials.attachment_form_fields')
<|file_sep|>src/Views/tickets/partials/modal_attachment_edit.blade.php.diff
original:
{!! CollectiveForm::hidden(null, 'modal-attachment-edit', ['id'=>'hide_modal_id']) !!}
updated:
{!! CollectiveForm::hidden(null, 'modal-attachment-edit', ['id'=>'hide_modal_id']) !!}
<|file_sep|>original/src/Views/tickets/partials/modal_attachment_edit.blade.php
<div class="modal-header">
<h4 class="modal-title">{{ trans('panichd::lang.attachment-edit') }}</h4>
<button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">{{ trans('panichd::lang.flash-x') }}</span></button>
</div>
<div class="modal-body">
@include('panichd::tickets.partials.attachment_form_fields')
{!! CollectiveForm::hidden(null, 'modal-attachment-edit', ['id'=>'hide_modal_id']) !!}
<div class="text-right col-md-12">
{!! CollectiveForm::button( trans('panichd::lang.update'), [
'type' => 'button',
'class' => 'btn btn-primary attachment_form_submit'
]) !!}
</div>
</div>
</div>
</div>
</div>
@append
<|file_sep|>current/src/Views/tickets/partials/modal_attachment_edit.blade.php
<button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">{{ trans('panichd::lang.flash-x') }}</span></button>
</div>
<div class="modal-body">
<!-- Div edit attachment -->
<fieldset id="edit_ticket_attachment">
@include('panichd::tickets.partials.attachment_form_fields')
{!! CollectiveForm::hidden(null, 'modal-attachment-edit', ['id'=>'hide_modal_id']) !!}
<div class="text-right col-md-12">
{!! CollectiveForm::button( trans('panichd::lang.update'), [
'type' => 'button',
'class' => 'btn btn-primary attachment_form_submit'
]) !!}
</div>
</div>
</div>
</div>
</div>
@append
<|file_sep|>updated/src/Views/tickets/partials/modal_attachment_edit.blade.php
<div class="modal-body">
<!-- Div edit attachment -->
<fieldset id="edit_ticket_attachment">
@include('panichd::tickets.partials.attachment_form_fields')
{!! CollectiveForm::hidden(null, 'modal-attachment-edit', ['id'=>'hide_modal_id']) !!}
<div class="text-right col-md-12">
{!! CollectiveForm::button( trans('panichd::lang.update'), [
'type' => 'button',
'class' => 'btn btn-primary attachment_form_submit',
'data-back-div' => 'ticket_attached'
]) !!}
</div>
</fieldset>
</div>
</div>
</div>
</div>
@append | e6b1c6fc355585ebd2f9b4a284ae151303cf0488 | src/Views/tickets/partials/modal_attachment_edit.blade.php | src/Views/tickets/partials/modal_attachment_edit.blade.php | PHP |
<|file_sep|>original/landscape/trees/_pakmak.tab
OPTIONS
compile yes
merge trees.all.pak
<|file_sep|>current/landscape/trees/_pakmak.tab
OPTIONS
compile yes
merge trees.all.pak
<|file_sep|>updated/landscape/trees/_pakmak.tab | OPTIONS
compile yes
merge trees.all.pak
# There's relatively little sense in allowing users to manipulate single trees, usually they are only all deleted for network games. | <|file_sep|>original/landscape/trees/_pakmak.tab
OPTIONS
compile yes
merge trees.all.pak
<|file_sep|>current/landscape/trees/_pakmak.tab
OPTIONS
compile yes
merge trees.all.pak
<|file_sep|>updated/landscape/trees/_pakmak.tab
OPTIONS
compile yes
merge trees.all.pak
# There's relatively little sense in allowing users to manipulate single trees, usually they are only all deleted for network games. | 31727063360030407eaf691e11fbbfc73f60c88f | landscape/trees/_pakmak.tab | landscape/trees/_pakmak.tab | SQL |
<|file_sep|>plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl.diff
original:
# Libc.flush_cstdio()
updated:
Libc.flush_cstdio()
<|file_sep|>plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl.diff
original:
# flush(STDERR)
updated:
flush(STDERR)
<|file_sep|>original/plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl
println("<<<<eox>>>>")
end
end
function flush_all()
# Libc.flush_cstdio()
flush(STDOUT)
# flush(STDERR)
end
#import Base.flush
#const StdioPipe = Base.PipeEndpoint
#function flush(io::StdioPipe)
# invoke(flush, (supertype(StdioPipe),), io)
# if io == STDOUT
# oslibuv_flush()
# send_stream("stdout")
# elseif io == STDERR
# oslibuv_flush()
# send_stream("stderr")
# end
#end
end;
<|file_sep|>current/plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl
println("<<<<eox>>>>")
end
end
function flush_all()
Libc.flush_cstdio()
flush(STDOUT)
flush(STDERR)
end
#import Base.flush
#const StdioPipe = Base.PipeEndpoint
#function flush(io::StdioPipe)
# invoke(flush, (supertype(StdioPipe),), io)
# if io == STDOUT
# oslibuv_flush()
# send_stream("stdout")
# elseif io == STDERR
# oslibuv_flush()
# send_stream("stderr")
# end
#end
end;
<|file_sep|>updated/plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl | println(result)
end
end
end
catch e
showerror(STDOUT, e); println()
status = "error"
println()
finally
flush_all()
println("<<<<$status>>>>")
println("<<<<$mimeType>>>>")
println("<<<<eox>>>>")
end
end
function flush_all()
Libc.flush_cstdio()
flush(STDOUT)
flush(STDERR)
end
end; | <|file_sep|>plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl.diff
original:
# Libc.flush_cstdio()
updated:
Libc.flush_cstdio()
<|file_sep|>plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl.diff
original:
# flush(STDERR)
updated:
flush(STDERR)
<|file_sep|>original/plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl
println("<<<<eox>>>>")
end
end
function flush_all()
# Libc.flush_cstdio()
flush(STDOUT)
# flush(STDERR)
end
#import Base.flush
#const StdioPipe = Base.PipeEndpoint
#function flush(io::StdioPipe)
# invoke(flush, (supertype(StdioPipe),), io)
# if io == STDOUT
# oslibuv_flush()
# send_stream("stdout")
# elseif io == STDERR
# oslibuv_flush()
# send_stream("stderr")
# end
#end
end;
<|file_sep|>current/plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl
println("<<<<eox>>>>")
end
end
function flush_all()
Libc.flush_cstdio()
flush(STDOUT)
flush(STDERR)
end
#import Base.flush
#const StdioPipe = Base.PipeEndpoint
#function flush(io::StdioPipe)
# invoke(flush, (supertype(StdioPipe),), io)
# if io == STDOUT
# oslibuv_flush()
# send_stream("stdout")
# elseif io == STDERR
# oslibuv_flush()
# send_stream("stderr")
# end
#end
end;
<|file_sep|>updated/plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl
println(result)
end
end
end
catch e
showerror(STDOUT, e); println()
status = "error"
println()
finally
flush_all()
println("<<<<$status>>>>")
println("<<<<$mimeType>>>>")
println("<<<<eox>>>>")
end
end
function flush_all()
Libc.flush_cstdio()
flush(STDOUT)
flush(STDERR)
end
end; | 198ccbb51237bdea2dbc41726f16a1aa2f2f9c1a | plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl | plugins/com.juliacomputing.jldt.eclipse.ui.console/script/repl-wrapper.jl | Julia |
<|file_sep|>original/.travis.yml
sudo: false
language: python
python:
- 2.7
git:
submodules: false
depth: 500
addons:
postgresql: 9.6
install:
- pip install -U pip setuptools
- pip install -r requirements.dev.txt
- pip install -e .
- npm install
script:
- FORCE_COLOR=1 npx react-jsx-i18n extract --ext jsx indico/web/client/ indico/modules/ > /dev/null
- isort -rc --diff --check-only indico/
- python bin/maintenance/update_backrefs.py --ci
- npx eslint --ext .js --ext .jsx indico/modules/rb_new/ indico/modules/events/logs/ indico/web/client/js/react/
- pytest
- python bin/maintenance/build-assets.py indico --dev
<|file_sep|>current/.travis.yml
sudo: false
language: python
python:
- 2.7
git:
submodules: false
depth: 500
addons:
postgresql: 9.6
install:
- pip install -U pip setuptools
- pip install -r requirements.dev.txt
- pip install -e .
- npm install
script:
- FORCE_COLOR=1 npx react-jsx-i18n extract --ext jsx indico/web/client/ indico/modules/ > /dev/null
- isort -rc --diff --check-only indico/
- python bin/maintenance/update_backrefs.py --ci
- npx eslint --ext .js --ext .jsx indico/modules/rb_new/ indico/modules/events/logs/ indico/web/client/js/react/
- pytest
- python bin/maintenance/build-assets.py indico --dev
<|file_sep|>updated/.travis.yml | language: python
python:
- 2.7
git:
submodules: false
depth: 500
addons:
postgresql: 9.6
install:
- pip install -U pip setuptools
- pip install -r requirements.dev.txt
- pip install -e .
- npm install
script:
- FORCE_COLOR=1 npx react-jsx-i18n extract --ext jsx indico/web/client/ indico/modules/ > /dev/null
- isort -rc --diff --check-only indico/
- python bin/maintenance/update_backrefs.py --ci
- npx eslint --ext .js --ext .jsx indico/modules/rb_new/ indico/modules/events/logs/ indico/web/client/js/react/
- pytest
- python bin/maintenance/build-assets.py indico --dev
notifications: | <|file_sep|>original/.travis.yml
sudo: false
language: python
python:
- 2.7
git:
submodules: false
depth: 500
addons:
postgresql: 9.6
install:
- pip install -U pip setuptools
- pip install -r requirements.dev.txt
- pip install -e .
- npm install
script:
- FORCE_COLOR=1 npx react-jsx-i18n extract --ext jsx indico/web/client/ indico/modules/ > /dev/null
- isort -rc --diff --check-only indico/
- python bin/maintenance/update_backrefs.py --ci
- npx eslint --ext .js --ext .jsx indico/modules/rb_new/ indico/modules/events/logs/ indico/web/client/js/react/
- pytest
- python bin/maintenance/build-assets.py indico --dev
<|file_sep|>current/.travis.yml
sudo: false
language: python
python:
- 2.7
git:
submodules: false
depth: 500
addons:
postgresql: 9.6
install:
- pip install -U pip setuptools
- pip install -r requirements.dev.txt
- pip install -e .
- npm install
script:
- FORCE_COLOR=1 npx react-jsx-i18n extract --ext jsx indico/web/client/ indico/modules/ > /dev/null
- isort -rc --diff --check-only indico/
- python bin/maintenance/update_backrefs.py --ci
- npx eslint --ext .js --ext .jsx indico/modules/rb_new/ indico/modules/events/logs/ indico/web/client/js/react/
- pytest
- python bin/maintenance/build-assets.py indico --dev
<|file_sep|>updated/.travis.yml
language: python
python:
- 2.7
git:
submodules: false
depth: 500
addons:
postgresql: 9.6
install:
- pip install -U pip setuptools
- pip install -r requirements.dev.txt
- pip install -e .
- npm install
script:
- FORCE_COLOR=1 npx react-jsx-i18n extract --ext jsx indico/web/client/ indico/modules/ > /dev/null
- isort -rc --diff --check-only indico/
- python bin/maintenance/update_backrefs.py --ci
- npx eslint --ext .js --ext .jsx indico/modules/rb_new/ indico/modules/events/logs/ indico/web/client/js/react/
- pytest
- python bin/maintenance/build-assets.py indico --dev
notifications: | f84adaee75ee229024d217ed68977fb6c3a126dc | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/Cargo.toml
[workspace]
members = [
"webrender",
"webrender_api",
"wrench",
]
[profile.release]
debug = true
<|file_sep|>current/Cargo.toml
[workspace]
members = [
"webrender",
"webrender_api",
"wrench",
]
[profile.release]
debug = true
<|file_sep|>updated/Cargo.toml | [workspace]
members = [
"webrender",
"webrender_api",
"wrench",
]
[profile.release]
debug = true
panic = "abort" | <|file_sep|>original/Cargo.toml
[workspace]
members = [
"webrender",
"webrender_api",
"wrench",
]
[profile.release]
debug = true
<|file_sep|>current/Cargo.toml
[workspace]
members = [
"webrender",
"webrender_api",
"wrench",
]
[profile.release]
debug = true
<|file_sep|>updated/Cargo.toml
[workspace]
members = [
"webrender",
"webrender_api",
"wrench",
]
[profile.release]
debug = true
panic = "abort" | 6af8c836e1dac195eff8e52232f5f7466012f5bf | Cargo.toml | Cargo.toml | TOML |
<|file_sep|>original/app/views/main.html
<textarea class="form-control" rows="4" placeholder="Cheep cheep cheep?"></textarea>
</div>
<button type="submit" class="btn btn-success">Cheep!</button>
</form>
</div>
<div class="col-md-9 col-lg-8">
<ul class="media-list">
<li class="media" ng-repeat="cheep in stream">
<a class="pull-left" href="#">
<img class="media-object" width="64px" height="64px" src="{{ cheep.author.thumbnail_url }}" alt="Generic placeholder image">
</a>
<div class="media-body">
<h4 class="media-heading">{{ cheep.author.name }} <small>@{{ cheep.author.handle }}</small></h4>
<p>{{ cheep.content }}</p>
</div>
</li>
</ul>
</div>
</div>
<|file_sep|>current/app/views/main.html
<textarea class="form-control" rows="4" placeholder="Cheep cheep cheep?"></textarea>
</div>
<button type="submit" class="btn btn-success">Cheep!</button>
</form>
</div>
<div class="col-md-9 col-lg-8">
<ul class="media-list">
<li class="media" ng-repeat="cheep in stream">
<a class="pull-left" href="#">
<img class="media-object" width="64px" height="64px" src="{{ cheep.author.thumbnail_url }}" alt="Generic placeholder image">
</a>
<div class="media-body">
<h4 class="media-heading">{{ cheep.author.name }} <small>@{{ cheep.author.handle }}</small></h4>
<p>{{ cheep.content }}</p>
</div>
</li>
</ul>
</div>
</div>
<|file_sep|>updated/app/views/main.html | <textarea class="form-control" rows="4" placeholder="Cheep cheep cheep?"></textarea>
</div>
<button type="submit" class="btn btn-success">Cheep!</button>
</form>
</div>
<div class="col-md-9 col-lg-8">
<ul class="media-list">
<li class="media" ng-repeat="cheep in stream">
<a class="pull-left" href="#">
<img class="media-object" width="64px" height="64px" ng-src="{{ cheep.author.thumbnail_url }}" alt="Generic placeholder image">
</a>
<div class="media-body">
<h4 class="media-heading">{{ cheep.author.name }} <small>@{{ cheep.author.handle }}</small></h4>
<p>{{ cheep.content }}</p>
</div>
</li>
</ul>
</div>
</div> | <|file_sep|>original/app/views/main.html
<textarea class="form-control" rows="4" placeholder="Cheep cheep cheep?"></textarea>
</div>
<button type="submit" class="btn btn-success">Cheep!</button>
</form>
</div>
<div class="col-md-9 col-lg-8">
<ul class="media-list">
<li class="media" ng-repeat="cheep in stream">
<a class="pull-left" href="#">
<img class="media-object" width="64px" height="64px" src="{{ cheep.author.thumbnail_url }}" alt="Generic placeholder image">
</a>
<div class="media-body">
<h4 class="media-heading">{{ cheep.author.name }} <small>@{{ cheep.author.handle }}</small></h4>
<p>{{ cheep.content }}</p>
</div>
</li>
</ul>
</div>
</div>
<|file_sep|>current/app/views/main.html
<textarea class="form-control" rows="4" placeholder="Cheep cheep cheep?"></textarea>
</div>
<button type="submit" class="btn btn-success">Cheep!</button>
</form>
</div>
<div class="col-md-9 col-lg-8">
<ul class="media-list">
<li class="media" ng-repeat="cheep in stream">
<a class="pull-left" href="#">
<img class="media-object" width="64px" height="64px" src="{{ cheep.author.thumbnail_url }}" alt="Generic placeholder image">
</a>
<div class="media-body">
<h4 class="media-heading">{{ cheep.author.name }} <small>@{{ cheep.author.handle }}</small></h4>
<p>{{ cheep.content }}</p>
</div>
</li>
</ul>
</div>
</div>
<|file_sep|>updated/app/views/main.html
<textarea class="form-control" rows="4" placeholder="Cheep cheep cheep?"></textarea>
</div>
<button type="submit" class="btn btn-success">Cheep!</button>
</form>
</div>
<div class="col-md-9 col-lg-8">
<ul class="media-list">
<li class="media" ng-repeat="cheep in stream">
<a class="pull-left" href="#">
<img class="media-object" width="64px" height="64px" ng-src="{{ cheep.author.thumbnail_url }}" alt="Generic placeholder image">
</a>
<div class="media-body">
<h4 class="media-heading">{{ cheep.author.name }} <small>@{{ cheep.author.handle }}</small></h4>
<p>{{ cheep.content }}</p>
</div>
</li>
</ul>
</div>
</div> | c1603a84541526ff2fbe1a43f0193fe7caaed18a | app/views/main.html | app/views/main.html | HTML |
<|file_sep|>original/README.md
To have an always up-to-date, bleeding edge, possibly broken
version available, you can use the folowing snippet in your `.bashrc`.
```shell
alias shoreman='sh -c "$(curl -fsSL https://raw.github.com/hecticjeff/shoreman/master/shoreman.sh)"'
```
## Usage
```
cd project-with-procfile
shoreman
```
## Credits
Inspired by the original [Foreman](https://github.com/ddollar/foreman)
by David Dollar (@ddollar) and [Norman](https://github.com/josh/norman) (foreman for
Node) by Josh Peek (@josh).
<|file_sep|>current/README.md
To have an always up-to-date, bleeding edge, possibly broken
version available, you can use the folowing snippet in your `.bashrc`.
```shell
alias shoreman='sh -c "$(curl -fsSL https://raw.github.com/hecticjeff/shoreman/master/shoreman.sh)"'
```
## Usage
```
cd project-with-procfile
shoreman
```
## Credits
Inspired by the original [Foreman](https://github.com/ddollar/foreman)
by David Dollar (@ddollar) and [Norman](https://github.com/josh/norman) (foreman for
Node) by Josh Peek (@josh).
<|file_sep|>updated/README.md |
To have an always up-to-date, bleeding edge, possibly broken
version available, you can use the folowing snippet in your `.bashrc`.
```shell
alias shoreman='sh -c "$(curl -fsSL https://raw.github.com/hecticjeff/shoreman/master/shoreman.sh)"'
```
## Usage
Head into a project that has a `Procfile` in it, then simply run the
`shoreman` command, it will read your Procfile, and start up the
processes it finds.
```
cd project-with-procfile
shoreman
```
## Credits
| <|file_sep|>original/README.md
To have an always up-to-date, bleeding edge, possibly broken
version available, you can use the folowing snippet in your `.bashrc`.
```shell
alias shoreman='sh -c "$(curl -fsSL https://raw.github.com/hecticjeff/shoreman/master/shoreman.sh)"'
```
## Usage
```
cd project-with-procfile
shoreman
```
## Credits
Inspired by the original [Foreman](https://github.com/ddollar/foreman)
by David Dollar (@ddollar) and [Norman](https://github.com/josh/norman) (foreman for
Node) by Josh Peek (@josh).
<|file_sep|>current/README.md
To have an always up-to-date, bleeding edge, possibly broken
version available, you can use the folowing snippet in your `.bashrc`.
```shell
alias shoreman='sh -c "$(curl -fsSL https://raw.github.com/hecticjeff/shoreman/master/shoreman.sh)"'
```
## Usage
```
cd project-with-procfile
shoreman
```
## Credits
Inspired by the original [Foreman](https://github.com/ddollar/foreman)
by David Dollar (@ddollar) and [Norman](https://github.com/josh/norman) (foreman for
Node) by Josh Peek (@josh).
<|file_sep|>updated/README.md
To have an always up-to-date, bleeding edge, possibly broken
version available, you can use the folowing snippet in your `.bashrc`.
```shell
alias shoreman='sh -c "$(curl -fsSL https://raw.github.com/hecticjeff/shoreman/master/shoreman.sh)"'
```
## Usage
Head into a project that has a `Procfile` in it, then simply run the
`shoreman` command, it will read your Procfile, and start up the
processes it finds.
```
cd project-with-procfile
shoreman
```
## Credits
| b417a10231b602bbec890ccc1325f59c7fb6ba9e | README.md | README.md | Markdown |
<|file_sep|>README.md.diff
original:
Since it's not in WebStore at the moment, install it as unpacked extension:
updated:
Install it from Chrome WebStore:
https://chrome.google.com/webstore/detail/lingualeo-dictionary-expo/mpaohidlipnfnkbogpmanchjfjpdgcml
Development
===========
Install it as unpacked extension:
<|file_sep|>README.md.diff
original:
- Publish in Chrome WebStore
updated:
<|file_sep|>original/README.md
Motivation
==========
LinguaLeo word trainings lacking interval repetition support. So I decided to export my LinguaLeo dictionary to Anki via CSV.
Installation
============
Since it's not in WebStore at the moment, install it as unpacked extension:
1. Clone the repo
2. Go to Chrome -> Tools -> Extensions
3. Check 'Developer mode'
4. Click 'Load unpacked extension' and point to the 'src' folder of the repo
Todo
====
- Publish in Chrome WebStore
- Go Bower!
<|file_sep|>current/README.md
============
Install it from Chrome WebStore:
https://chrome.google.com/webstore/detail/lingualeo-dictionary-expo/mpaohidlipnfnkbogpmanchjfjpdgcml
Development
===========
Install it as unpacked extension:
1. Clone the repo
2. Go to Chrome -> Tools -> Extensions
3. Check 'Developer mode'
4. Click 'Load unpacked extension' and point to the 'src' folder of the repo
Todo
====
- Go Bower!
<|file_sep|>updated/README.md |
Install it from Chrome WebStore:
https://chrome.google.com/webstore/detail/lingualeo-dictionary-expo/mpaohidlipnfnkbogpmanchjfjpdgcml
Development
===========
Install it as unpacked extension:
1. Clone the repo
2. Go to Chrome -> Tools -> Extensions
3. Check 'Developer mode'
4. Click 'Load unpacked extension' and point to the 'src' folder of the repo
Todo
====
- Go Bower!
- Decide what to do with several translations. Now they are spat in CSV separated by <br/> | <|file_sep|>README.md.diff
original:
Since it's not in WebStore at the moment, install it as unpacked extension:
updated:
Install it from Chrome WebStore:
https://chrome.google.com/webstore/detail/lingualeo-dictionary-expo/mpaohidlipnfnkbogpmanchjfjpdgcml
Development
===========
Install it as unpacked extension:
<|file_sep|>README.md.diff
original:
- Publish in Chrome WebStore
updated:
<|file_sep|>original/README.md
Motivation
==========
LinguaLeo word trainings lacking interval repetition support. So I decided to export my LinguaLeo dictionary to Anki via CSV.
Installation
============
Since it's not in WebStore at the moment, install it as unpacked extension:
1. Clone the repo
2. Go to Chrome -> Tools -> Extensions
3. Check 'Developer mode'
4. Click 'Load unpacked extension' and point to the 'src' folder of the repo
Todo
====
- Publish in Chrome WebStore
- Go Bower!
<|file_sep|>current/README.md
============
Install it from Chrome WebStore:
https://chrome.google.com/webstore/detail/lingualeo-dictionary-expo/mpaohidlipnfnkbogpmanchjfjpdgcml
Development
===========
Install it as unpacked extension:
1. Clone the repo
2. Go to Chrome -> Tools -> Extensions
3. Check 'Developer mode'
4. Click 'Load unpacked extension' and point to the 'src' folder of the repo
Todo
====
- Go Bower!
<|file_sep|>updated/README.md
Install it from Chrome WebStore:
https://chrome.google.com/webstore/detail/lingualeo-dictionary-expo/mpaohidlipnfnkbogpmanchjfjpdgcml
Development
===========
Install it as unpacked extension:
1. Clone the repo
2. Go to Chrome -> Tools -> Extensions
3. Check 'Developer mode'
4. Click 'Load unpacked extension' and point to the 'src' folder of the repo
Todo
====
- Go Bower!
- Decide what to do with several translations. Now they are spat in CSV separated by <br/> | 2e03b8c100447da4788a802546e4d846e45d07c6 | README.md | README.md | Markdown |
<|file_sep|>pkgs/applications/misc/hello/default.nix.diff
original:
updated:
, runCommand
<|file_sep|>pkgs/applications/misc/hello/default.nix.diff
original:
stdenv.mkDerivation rec {
updated:
stdenv.mkDerivation (self: {
<|file_sep|>pkgs/applications/misc/hello/default.nix.diff
original:
url = "mirror://gnu/hello/${pname}-${version}.tar.gz";
updated:
url = "mirror://gnu/hello/${self.pname}-${self.version}.tar.gz";
<|file_sep|>pkgs/applications/misc/hello/default.nix.diff
original:
updated:
passthru.tests.run = runCommand "hello-test-run" {
nativeBuildInputs = [ self ];
} ''
diff -U3 --color=auto <(hello) <(echo 'Hello, world!')
touch $out
'';
<|file_sep|>pkgs/applications/misc/hello/default.nix.diff
original:
changelog = "https://git.savannah.gnu.org/cgit/hello.git/plain/NEWS?h=v${version}";
updated:
changelog = "https://git.savannah.gnu.org/cgit/hello.git/plain/NEWS?h=v${self.version}";
<|file_sep|>original/pkgs/applications/misc/hello/default.nix
invariant-under-noXlibs =
testers.testEqualDerivation
"hello must not be rebuilt when environment.noXlibs is set."
hello
(nixos { environment.noXlibs = true; }).pkgs.hello;
};
meta = with lib; {
description = "A program that produces a familiar, friendly greeting";
longDescription = ''
GNU Hello is a program that prints "Hello, world!" when you run it.
It is fully customizable.
'';
homepage = "https://www.gnu.org/software/hello/manual/";
changelog = "https://git.savannah.gnu.org/cgit/hello.git/plain/NEWS?h=v${version}";
license = licenses.gpl3Plus;
maintainers = [ maintainers.eelco ];
platforms = platforms.all;
};
}
<|file_sep|>current/pkgs/applications/misc/hello/default.nix
passthru.tests.run = runCommand "hello-test-run" {
nativeBuildInputs = [ self ];
} ''
diff -U3 --color=auto <(hello) <(echo 'Hello, world!')
touch $out
'';
meta = with lib; {
description = "A program that produces a familiar, friendly greeting";
longDescription = ''
GNU Hello is a program that prints "Hello, world!" when you run it.
It is fully customizable.
'';
homepage = "https://www.gnu.org/software/hello/manual/";
changelog = "https://git.savannah.gnu.org/cgit/hello.git/plain/NEWS?h=v${self.version}";
license = licenses.gpl3Plus;
maintainers = [ maintainers.eelco ];
platforms = platforms.all;
};
}
<|file_sep|>updated/pkgs/applications/misc/hello/default.nix |
passthru.tests.run = runCommand "hello-test-run" {
nativeBuildInputs = [ self ];
} ''
diff -U3 --color=auto <(hello) <(echo 'Hello, world!')
touch $out
'';
meta = with lib; {
description = "A program that produces a familiar, friendly greeting";
longDescription = ''
GNU Hello is a program that prints "Hello, world!" when you run it.
It is fully customizable.
'';
homepage = "https://www.gnu.org/software/hello/manual/";
changelog = "https://git.savannah.gnu.org/cgit/hello.git/plain/NEWS?h=v${self.version}";
license = licenses.gpl3Plus;
maintainers = [ maintainers.eelco ];
platforms = platforms.all;
};
}) | <|file_sep|>pkgs/applications/misc/hello/default.nix.diff
original:
updated:
, runCommand
<|file_sep|>pkgs/applications/misc/hello/default.nix.diff
original:
stdenv.mkDerivation rec {
updated:
stdenv.mkDerivation (self: {
<|file_sep|>pkgs/applications/misc/hello/default.nix.diff
original:
url = "mirror://gnu/hello/${pname}-${version}.tar.gz";
updated:
url = "mirror://gnu/hello/${self.pname}-${self.version}.tar.gz";
<|file_sep|>pkgs/applications/misc/hello/default.nix.diff
original:
updated:
passthru.tests.run = runCommand "hello-test-run" {
nativeBuildInputs = [ self ];
} ''
diff -U3 --color=auto <(hello) <(echo 'Hello, world!')
touch $out
'';
<|file_sep|>pkgs/applications/misc/hello/default.nix.diff
original:
changelog = "https://git.savannah.gnu.org/cgit/hello.git/plain/NEWS?h=v${version}";
updated:
changelog = "https://git.savannah.gnu.org/cgit/hello.git/plain/NEWS?h=v${self.version}";
<|file_sep|>original/pkgs/applications/misc/hello/default.nix
invariant-under-noXlibs =
testers.testEqualDerivation
"hello must not be rebuilt when environment.noXlibs is set."
hello
(nixos { environment.noXlibs = true; }).pkgs.hello;
};
meta = with lib; {
description = "A program that produces a familiar, friendly greeting";
longDescription = ''
GNU Hello is a program that prints "Hello, world!" when you run it.
It is fully customizable.
'';
homepage = "https://www.gnu.org/software/hello/manual/";
changelog = "https://git.savannah.gnu.org/cgit/hello.git/plain/NEWS?h=v${version}";
license = licenses.gpl3Plus;
maintainers = [ maintainers.eelco ];
platforms = platforms.all;
};
}
<|file_sep|>current/pkgs/applications/misc/hello/default.nix
passthru.tests.run = runCommand "hello-test-run" {
nativeBuildInputs = [ self ];
} ''
diff -U3 --color=auto <(hello) <(echo 'Hello, world!')
touch $out
'';
meta = with lib; {
description = "A program that produces a familiar, friendly greeting";
longDescription = ''
GNU Hello is a program that prints "Hello, world!" when you run it.
It is fully customizable.
'';
homepage = "https://www.gnu.org/software/hello/manual/";
changelog = "https://git.savannah.gnu.org/cgit/hello.git/plain/NEWS?h=v${self.version}";
license = licenses.gpl3Plus;
maintainers = [ maintainers.eelco ];
platforms = platforms.all;
};
}
<|file_sep|>updated/pkgs/applications/misc/hello/default.nix
passthru.tests.run = runCommand "hello-test-run" {
nativeBuildInputs = [ self ];
} ''
diff -U3 --color=auto <(hello) <(echo 'Hello, world!')
touch $out
'';
meta = with lib; {
description = "A program that produces a familiar, friendly greeting";
longDescription = ''
GNU Hello is a program that prints "Hello, world!" when you run it.
It is fully customizable.
'';
homepage = "https://www.gnu.org/software/hello/manual/";
changelog = "https://git.savannah.gnu.org/cgit/hello.git/plain/NEWS?h=v${self.version}";
license = licenses.gpl3Plus;
maintainers = [ maintainers.eelco ];
platforms = platforms.all;
};
}) | 2afc03a084cf736cc76f629a45a720c3c44eb27d | pkgs/applications/misc/hello/default.nix | pkgs/applications/misc/hello/default.nix | Nix |
<|file_sep|>test/correctness/thread_safety.cpp.diff
original:
#include <thread>
updated:
#include <future>
<|file_sep|>test/correctness/thread_safety.cpp.diff
original:
updated:
static std::atomic<int> foo;
<|file_sep|>test/correctness/thread_safety.cpp.diff
original:
std::vector<std::thread> threads;
updated:
std::vector<std::future<void>> futures;
<|file_sep|>test/correctness/thread_safety.cpp.diff
original:
threads.emplace_back([]{
updated:
futures.emplace_back(std::async(std::launch::async, []{
<|file_sep|>test/correctness/thread_safety.cpp.diff
original:
});
updated:
}));
<|file_sep|>original/test/correctness/thread_safety.cpp
int main(int argc, char **argv) {
// Test if the compiler itself is thread-safe. This test is
// intended to be run in a thread-sanitizer.
std::vector<std::thread> threads;
for (int i = 0; i < 1000; i++) {
threads.emplace_back([]{
Func f;
Var x;
f(x) = x;
f.realize(100);
});
}
for (auto &t : threads) {
t.join();
}
printf("Success!\n");
return 0;
}
<|file_sep|>current/test/correctness/thread_safety.cpp
int main(int argc, char **argv) {
// Test if the compiler itself is thread-safe. This test is
// intended to be run in a thread-sanitizer.
std::vector<std::future<void>> futures;
for (int i = 0; i < 1000; i++) {
futures.emplace_back(std::async(std::launch::async, []{
Func f;
Var x;
f(x) = x;
f.realize(100);
}));
}
for (auto &t : threads) {
t.join();
}
printf("Success!\n");
return 0;
}
<|file_sep|>updated/test/correctness/thread_safety.cpp | int main(int argc, char **argv) {
// Test if the compiler itself is thread-safe. This test is
// intended to be run in a thread-sanitizer.
std::vector<std::future<void>> futures;
for (int i = 0; i < 1000; i++) {
futures.emplace_back(std::async(std::launch::async, []{
Func f;
Var x;
f(x) = x;
f.realize(100);
}));
}
for (auto &f : futures) {
f.wait();
}
printf("Success!\n");
return 0;
} | <|file_sep|>test/correctness/thread_safety.cpp.diff
original:
#include <thread>
updated:
#include <future>
<|file_sep|>test/correctness/thread_safety.cpp.diff
original:
updated:
static std::atomic<int> foo;
<|file_sep|>test/correctness/thread_safety.cpp.diff
original:
std::vector<std::thread> threads;
updated:
std::vector<std::future<void>> futures;
<|file_sep|>test/correctness/thread_safety.cpp.diff
original:
threads.emplace_back([]{
updated:
futures.emplace_back(std::async(std::launch::async, []{
<|file_sep|>test/correctness/thread_safety.cpp.diff
original:
});
updated:
}));
<|file_sep|>original/test/correctness/thread_safety.cpp
int main(int argc, char **argv) {
// Test if the compiler itself is thread-safe. This test is
// intended to be run in a thread-sanitizer.
std::vector<std::thread> threads;
for (int i = 0; i < 1000; i++) {
threads.emplace_back([]{
Func f;
Var x;
f(x) = x;
f.realize(100);
});
}
for (auto &t : threads) {
t.join();
}
printf("Success!\n");
return 0;
}
<|file_sep|>current/test/correctness/thread_safety.cpp
int main(int argc, char **argv) {
// Test if the compiler itself is thread-safe. This test is
// intended to be run in a thread-sanitizer.
std::vector<std::future<void>> futures;
for (int i = 0; i < 1000; i++) {
futures.emplace_back(std::async(std::launch::async, []{
Func f;
Var x;
f(x) = x;
f.realize(100);
}));
}
for (auto &t : threads) {
t.join();
}
printf("Success!\n");
return 0;
}
<|file_sep|>updated/test/correctness/thread_safety.cpp
int main(int argc, char **argv) {
// Test if the compiler itself is thread-safe. This test is
// intended to be run in a thread-sanitizer.
std::vector<std::future<void>> futures;
for (int i = 0; i < 1000; i++) {
futures.emplace_back(std::async(std::launch::async, []{
Func f;
Var x;
f(x) = x;
f.realize(100);
}));
}
for (auto &f : futures) {
f.wait();
}
printf("Success!\n");
return 0;
} | 9ff6ca8d4dee0a014e5da47866734567b09944e0 | test/correctness/thread_safety.cpp | test/correctness/thread_safety.cpp | C++ |
<|file_sep|>original/.zuul.yaml
- project:
templates:
- check-requirements
- lib-forward-testing-python3
- openstack-python3-xena-jobs
- periodic-stable-jobs
- publish-openstack-docs-pti
- release-notes-jobs-python3
<|file_sep|>current/.zuul.yaml
- project:
templates:
- check-requirements
- lib-forward-testing-python3
- openstack-python3-xena-jobs
- periodic-stable-jobs
- publish-openstack-docs-pti
- release-notes-jobs-python3
<|file_sep|>updated/.zuul.yaml | - project:
templates:
- check-requirements
- lib-forward-testing-python3
- openstack-python3-yoga-jobs
- periodic-stable-jobs
- publish-openstack-docs-pti
- release-notes-jobs-python3 | <|file_sep|>original/.zuul.yaml
- project:
templates:
- check-requirements
- lib-forward-testing-python3
- openstack-python3-xena-jobs
- periodic-stable-jobs
- publish-openstack-docs-pti
- release-notes-jobs-python3
<|file_sep|>current/.zuul.yaml
- project:
templates:
- check-requirements
- lib-forward-testing-python3
- openstack-python3-xena-jobs
- periodic-stable-jobs
- publish-openstack-docs-pti
- release-notes-jobs-python3
<|file_sep|>updated/.zuul.yaml
- project:
templates:
- check-requirements
- lib-forward-testing-python3
- openstack-python3-yoga-jobs
- periodic-stable-jobs
- publish-openstack-docs-pti
- release-notes-jobs-python3 | fa138406f76964949708932683d93967975f88db | .zuul.yaml | .zuul.yaml | YAML |
<|file_sep|>original/README.md
# company-cli
create a company from the command line interface
<|file_sep|>current/README.md
# company-cli
create a company from the command line interface
<|file_sep|>updated/README.md | # company-cli
create a company from the command line interface
```
company init # if not already in a company repository, asks for:
# * type (choose from abstract-company implementations, e.g. company-uk-ltd)
# * name (check availability)
# * leads through the implementation independent setup process and generates the companies file structure
# * @TODO: use "puke" idea
# * breaks if process needs manual steps to be completed in between and offers:
company continue # instead of `company init` in a repo that has a company with incomplete setup process
# the `company init` or `company continue` are greyed out and `company run` becomes available
company run # starts a dev server with a user interface to interact with the company (might also offer other "run cli usage)
``` | <|file_sep|>original/README.md
# company-cli
create a company from the command line interface
<|file_sep|>current/README.md
# company-cli
create a company from the command line interface
<|file_sep|>updated/README.md
# company-cli
create a company from the command line interface
```
company init # if not already in a company repository, asks for:
# * type (choose from abstract-company implementations, e.g. company-uk-ltd)
# * name (check availability)
# * leads through the implementation independent setup process and generates the companies file structure
# * @TODO: use "puke" idea
# * breaks if process needs manual steps to be completed in between and offers:
company continue # instead of `company init` in a repo that has a company with incomplete setup process
# the `company init` or `company continue` are greyed out and `company run` becomes available
company run # starts a dev server with a user interface to interact with the company (might also offer other "run cli usage)
``` | 6429e994e782bd77fcedd81a253549aeae318829 | README.md | README.md | Markdown |
<|file_sep|>original/lib/options.js
const {cpus} = require('os')
const defaultLogger = require('./logger')
module.exports = {
etag: null,
lastModified: null,
userAgent: 'plunger/1.0',
timeout: {
connection: 2000,
activity: 4000,
download: 0
},
maxDownloadSize: 100 * 1024 * 1024,
digestAlgorithm: 'sha384',
extractArchives: true,
indexOfMatches: [
/Directory of/,
/Index of/,
/Listing of/
],
<|file_sep|>current/lib/options.js
const {cpus} = require('os')
const defaultLogger = require('./logger')
module.exports = {
etag: null,
lastModified: null,
userAgent: 'plunger/1.0',
timeout: {
connection: 2000,
activity: 4000,
download: 0
},
maxDownloadSize: 100 * 1024 * 1024,
digestAlgorithm: 'sha384',
extractArchives: true,
indexOfMatches: [
/Directory of/,
/Index of/,
/Listing of/
],
<|file_sep|>updated/lib/options.js | const {cpus} = require('os')
const defaultLogger = require('./logger')
module.exports = {
etag: null,
lastModified: null,
userAgent: 'plunger/2.0 (+https://github.com/inspireteam/plunger)',
timeout: {
connection: 2000,
activity: 4000,
download: 0
},
maxDownloadSize: 100 * 1024 * 1024,
digestAlgorithm: 'sha384',
extractArchives: true,
indexOfMatches: [
/Directory of/,
/Index of/,
/Listing of/
], | <|file_sep|>original/lib/options.js
const {cpus} = require('os')
const defaultLogger = require('./logger')
module.exports = {
etag: null,
lastModified: null,
userAgent: 'plunger/1.0',
timeout: {
connection: 2000,
activity: 4000,
download: 0
},
maxDownloadSize: 100 * 1024 * 1024,
digestAlgorithm: 'sha384',
extractArchives: true,
indexOfMatches: [
/Directory of/,
/Index of/,
/Listing of/
],
<|file_sep|>current/lib/options.js
const {cpus} = require('os')
const defaultLogger = require('./logger')
module.exports = {
etag: null,
lastModified: null,
userAgent: 'plunger/1.0',
timeout: {
connection: 2000,
activity: 4000,
download: 0
},
maxDownloadSize: 100 * 1024 * 1024,
digestAlgorithm: 'sha384',
extractArchives: true,
indexOfMatches: [
/Directory of/,
/Index of/,
/Listing of/
],
<|file_sep|>updated/lib/options.js
const {cpus} = require('os')
const defaultLogger = require('./logger')
module.exports = {
etag: null,
lastModified: null,
userAgent: 'plunger/2.0 (+https://github.com/inspireteam/plunger)',
timeout: {
connection: 2000,
activity: 4000,
download: 0
},
maxDownloadSize: 100 * 1024 * 1024,
digestAlgorithm: 'sha384',
extractArchives: true,
indexOfMatches: [
/Directory of/,
/Index of/,
/Listing of/
], | 078aab4f58098cc62936aa5cf4fea591db25a7af | lib/options.js | lib/options.js | JavaScript |
<|file_sep|>test/helpers/views/hpp.erb.diff
original:
<head>
<title> HPP Payment</title>
</head>
<body>
<% hpp_client = Adyen::HPP::Client.new(:test, :testing) %>
<form action="<%= hpp_client.url %>" method="post">
<p> Price: <strong>EUR 43.21</strong>. </p>
<p>
<%= hpp_client.new_request.hidden_fields(@payment) %>
<input type="submit" value="Pay" />
</p>
</form>
updated:
<head>
<title> HPP Payment</title>
</head>
<body>
<% hpp_request = Adyen::HPP::Request.new(@payment, environment: :test, skin: :testing) %>
<form action="<%= hpp_request.url %>" method="post">
<p> Price: <strong>EUR 43.21</strong>. </p>
<p>
<%= hpp_request.hidden_fields %>
<input type="submit" value="Pay" />
</p>
</form>
<|file_sep|>test/helpers/views/hpp.erb.diff
original:
<p>If you are using the test account of this library, you will be redirected back
updated:
<p>If you are using the test account of this library, you will be redirected back
<|file_sep|>original/test/helpers/views/hpp.erb
<html>
<head>
<title> HPP Payment</title>
</head>
<body>
<% hpp_client = Adyen::HPP::Client.new(:test, :testing) %>
<form action="<%= hpp_client.url %>" method="post">
<p> Price: <strong>EUR 43.21</strong>. </p>
<p>
<%= hpp_client.new_request.hidden_fields(@payment) %>
<input type="submit" value="Pay" />
</p>
</form>
<p>If you are using the test account of this library, you will be redirected back
to <strong>example.com</strong> after you complete the payment on Adyen's hosted
payment pages. This is due to how the test account is configured.
</p>
</body>
</html>
<|file_sep|>current/test/helpers/views/hpp.erb
<html>
<head>
<title> HPP Payment</title>
</head>
<body>
<% hpp_request = Adyen::HPP::Request.new(@payment, environment: :test, skin: :testing) %>
<form action="<%= hpp_request.url %>" method="post">
<p> Price: <strong>EUR 43.21</strong>. </p>
<p>
<%= hpp_request.hidden_fields %>
<input type="submit" value="Pay" />
</p>
</form>
<p>If you are using the test account of this library, you will be redirected back
to <strong>example.com</strong> after you complete the payment on Adyen's hosted
payment pages. This is due to how the test account is configured.
</p>
</body>
</html>
<|file_sep|>updated/test/helpers/views/hpp.erb | <html>
<head>
<title> HPP Payment</title>
</head>
<body>
<% hpp_request = Adyen::HPP::Request.new(@payment, environment: :test, skin: :testing) %>
<form action="<%= hpp_request.url %>" method="post">
<p> Price: <strong>EUR 43.21</strong>. </p>
<p>
<%= hpp_request.hidden_fields %>
<input type="submit" value="Pay" />
</p>
</form>
<p>If you are using the test account of this library, you will be redirected back
to <strong>example.com</strong> after you complete the payment on Adyen's hosted
payment pages. This is due to how the test account is configured.
</p>
</body>
</html> | <|file_sep|>test/helpers/views/hpp.erb.diff
original:
<head>
<title> HPP Payment</title>
</head>
<body>
<% hpp_client = Adyen::HPP::Client.new(:test, :testing) %>
<form action="<%= hpp_client.url %>" method="post">
<p> Price: <strong>EUR 43.21</strong>. </p>
<p>
<%= hpp_client.new_request.hidden_fields(@payment) %>
<input type="submit" value="Pay" />
</p>
</form>
updated:
<head>
<title> HPP Payment</title>
</head>
<body>
<% hpp_request = Adyen::HPP::Request.new(@payment, environment: :test, skin: :testing) %>
<form action="<%= hpp_request.url %>" method="post">
<p> Price: <strong>EUR 43.21</strong>. </p>
<p>
<%= hpp_request.hidden_fields %>
<input type="submit" value="Pay" />
</p>
</form>
<|file_sep|>test/helpers/views/hpp.erb.diff
original:
<p>If you are using the test account of this library, you will be redirected back
updated:
<p>If you are using the test account of this library, you will be redirected back
<|file_sep|>original/test/helpers/views/hpp.erb
<html>
<head>
<title> HPP Payment</title>
</head>
<body>
<% hpp_client = Adyen::HPP::Client.new(:test, :testing) %>
<form action="<%= hpp_client.url %>" method="post">
<p> Price: <strong>EUR 43.21</strong>. </p>
<p>
<%= hpp_client.new_request.hidden_fields(@payment) %>
<input type="submit" value="Pay" />
</p>
</form>
<p>If you are using the test account of this library, you will be redirected back
to <strong>example.com</strong> after you complete the payment on Adyen's hosted
payment pages. This is due to how the test account is configured.
</p>
</body>
</html>
<|file_sep|>current/test/helpers/views/hpp.erb
<html>
<head>
<title> HPP Payment</title>
</head>
<body>
<% hpp_request = Adyen::HPP::Request.new(@payment, environment: :test, skin: :testing) %>
<form action="<%= hpp_request.url %>" method="post">
<p> Price: <strong>EUR 43.21</strong>. </p>
<p>
<%= hpp_request.hidden_fields %>
<input type="submit" value="Pay" />
</p>
</form>
<p>If you are using the test account of this library, you will be redirected back
to <strong>example.com</strong> after you complete the payment on Adyen's hosted
payment pages. This is due to how the test account is configured.
</p>
</body>
</html>
<|file_sep|>updated/test/helpers/views/hpp.erb
<html>
<head>
<title> HPP Payment</title>
</head>
<body>
<% hpp_request = Adyen::HPP::Request.new(@payment, environment: :test, skin: :testing) %>
<form action="<%= hpp_request.url %>" method="post">
<p> Price: <strong>EUR 43.21</strong>. </p>
<p>
<%= hpp_request.hidden_fields %>
<input type="submit" value="Pay" />
</p>
</form>
<p>If you are using the test account of this library, you will be redirected back
to <strong>example.com</strong> after you complete the payment on Adyen's hosted
payment pages. This is due to how the test account is configured.
</p>
</body>
</html> | cc7f21536f6bb9bc84fd9ed6781876ffba7ff114 | test/helpers/views/hpp.erb | test/helpers/views/hpp.erb | HTML+ERB |
<|file_sep|>original/js/local.js
var match = RegExp('[?&]' + name + '=([^&]*)').exec(window.location.search);
return match && decodeURIComponent(match[1].replace(/\+/g, ' '));
}
function fixPath (path) {
var match = RegExp('https?://[^/]*/(.*?)([?#]|$)').exec(window.location.href);
// If a stub exists, assume secure operation, so:
var stub = match && match[1];
if (stub) {
// (1) remove &admin= value from URL
path = path.replace(/(\?)(?:admin=[^&]*)*(.*?)(?:&)(?:admin=[^&]*)*/,'$1$2');
// (2) if URL begins with '/?', append stub to '/'
path = path.replace(/^(\/)(\?)/, '$1' + stub + '$2');
// (3) remove any port designation from URL
path = path.replace(/(https?:\/\/[^\/]*):[0-9]+/, '$1');
}
return path;
}
function apiRequest (url, obj, returnAsString) {
url = fixPath(url);
<|file_sep|>current/js/local.js
var match = RegExp('[?&]' + name + '=([^&]*)').exec(window.location.search);
return match && decodeURIComponent(match[1].replace(/\+/g, ' '));
}
function fixPath (path) {
var match = RegExp('https?://[^/]*/(.*?)([?#]|$)').exec(window.location.href);
// If a stub exists, assume secure operation, so:
var stub = match && match[1];
if (stub) {
// (1) remove &admin= value from URL
path = path.replace(/(\?)(?:admin=[^&]*)*(.*?)(?:&)(?:admin=[^&]*)*/,'$1$2');
// (2) if URL begins with '/?', append stub to '/'
path = path.replace(/^(\/)(\?)/, '$1' + stub + '$2');
// (3) remove any port designation from URL
path = path.replace(/(https?:\/\/[^\/]*):[0-9]+/, '$1');
}
return path;
}
function apiRequest (url, obj, returnAsString) {
url = fixPath(url);
<|file_sep|>updated/js/local.js | var match = RegExp('[?&]' + name + '=([^&]*)').exec(window.location.search);
return match && decodeURIComponent(match[1].replace(/\+/g, ' '));
}
function fixPath (path) {
var match = RegExp('https?://[^/]*/(.*?)([?#]|$)').exec(window.location.href);
// If a stub exists, assume secure operation, so:
var stub = match && match[1];
if (stub) {
// (1) remove &admin= value from URL
path = path.replace(/(\?)(?:admin=[^&]*)*(.*?)(?:&admin=[^&]*)*/,'$1$2');
// (2) if URL begins with '/?', append stub to '/'
path = path.replace(/^(\/)(\?)/, '$1' + stub + '$2');
// (3) remove any port designation from URL
path = path.replace(/(https?:\/\/[^\/]*):[0-9]+/, '$1');
}
return path;
}
function apiRequest (url, obj, returnAsString) {
url = fixPath(url); | <|file_sep|>original/js/local.js
var match = RegExp('[?&]' + name + '=([^&]*)').exec(window.location.search);
return match && decodeURIComponent(match[1].replace(/\+/g, ' '));
}
function fixPath (path) {
var match = RegExp('https?://[^/]*/(.*?)([?#]|$)').exec(window.location.href);
// If a stub exists, assume secure operation, so:
var stub = match && match[1];
if (stub) {
// (1) remove &admin= value from URL
path = path.replace(/(\?)(?:admin=[^&]*)*(.*?)(?:&)(?:admin=[^&]*)*/,'$1$2');
// (2) if URL begins with '/?', append stub to '/'
path = path.replace(/^(\/)(\?)/, '$1' + stub + '$2');
// (3) remove any port designation from URL
path = path.replace(/(https?:\/\/[^\/]*):[0-9]+/, '$1');
}
return path;
}
function apiRequest (url, obj, returnAsString) {
url = fixPath(url);
<|file_sep|>current/js/local.js
var match = RegExp('[?&]' + name + '=([^&]*)').exec(window.location.search);
return match && decodeURIComponent(match[1].replace(/\+/g, ' '));
}
function fixPath (path) {
var match = RegExp('https?://[^/]*/(.*?)([?#]|$)').exec(window.location.href);
// If a stub exists, assume secure operation, so:
var stub = match && match[1];
if (stub) {
// (1) remove &admin= value from URL
path = path.replace(/(\?)(?:admin=[^&]*)*(.*?)(?:&)(?:admin=[^&]*)*/,'$1$2');
// (2) if URL begins with '/?', append stub to '/'
path = path.replace(/^(\/)(\?)/, '$1' + stub + '$2');
// (3) remove any port designation from URL
path = path.replace(/(https?:\/\/[^\/]*):[0-9]+/, '$1');
}
return path;
}
function apiRequest (url, obj, returnAsString) {
url = fixPath(url);
<|file_sep|>updated/js/local.js
var match = RegExp('[?&]' + name + '=([^&]*)').exec(window.location.search);
return match && decodeURIComponent(match[1].replace(/\+/g, ' '));
}
function fixPath (path) {
var match = RegExp('https?://[^/]*/(.*?)([?#]|$)').exec(window.location.href);
// If a stub exists, assume secure operation, so:
var stub = match && match[1];
if (stub) {
// (1) remove &admin= value from URL
path = path.replace(/(\?)(?:admin=[^&]*)*(.*?)(?:&admin=[^&]*)*/,'$1$2');
// (2) if URL begins with '/?', append stub to '/'
path = path.replace(/^(\/)(\?)/, '$1' + stub + '$2');
// (3) remove any port designation from URL
path = path.replace(/(https?:\/\/[^\/]*):[0-9]+/, '$1');
}
return path;
}
function apiRequest (url, obj, returnAsString) {
url = fixPath(url); | 647f0066ef4b76112093a9484fbe2756ce5a1842 | js/local.js | js/local.js | JavaScript |
<|file_sep|>.travis/run.sh.diff
original:
updated:
# Commented out pep8 until above bug is fixed.
<|file_sep|>original/.travis/run.sh
#!/bin/bash
set -e
set -x
# Commented out py3 specific apps until #2442/3 resolved:
# https://github.com/travis-ci/apt-package-whitelist/issues
#if [[ $TRAVIS_PYTHON_VERSION == 3* ]]; then
# PEP8="python3 /usr/lib/python3/dist-packages/pep8.py"
# PYLINT=/usr/bin/pylint3
#else
PEP8=/usr/bin/pep8
PYLINT=/usr/bin/pylint
#fi
# Ignore imports not at start and line-too-long)
$PEP8 --ignore=E402,E501 clipster
$PYLINT --errors-only clipster
python tests/tests.py
<|file_sep|>current/.travis/run.sh
#!/bin/bash
set -e
set -x
# Commented out py3 specific apps until #2442/3 resolved:
# https://github.com/travis-ci/apt-package-whitelist/issues
#if [[ $TRAVIS_PYTHON_VERSION == 3* ]]; then
# PEP8="python3 /usr/lib/python3/dist-packages/pep8.py"
# PYLINT=/usr/bin/pylint3
#else
PEP8=/usr/bin/pep8
PYLINT=/usr/bin/pylint
#fi
# Commented out pep8 until above bug is fixed.
# Ignore imports not at start and line-too-long)
$PEP8 --ignore=E402,E501 clipster
$PYLINT --errors-only clipster
python tests/tests.py
<|file_sep|>updated/.travis/run.sh | #!/bin/bash
set -e
set -x
# Commented out py3 specific apps until #2442/3 resolved:
# https://github.com/travis-ci/apt-package-whitelist/issues
#if [[ $TRAVIS_PYTHON_VERSION == 3* ]]; then
# PEP8="python3 /usr/lib/python3/dist-packages/pep8.py"
# PYLINT=/usr/bin/pylint3
#else
PEP8=/usr/bin/pep8
PYLINT=/usr/bin/pylint
#fi
# Commented out pep8 until above bug is fixed.
# Ignore imports not at start and line-too-long)
#$PEP8 --ignore=E402,E501 clipster
$PYLINT --errors-only clipster
python tests/tests.py | <|file_sep|>.travis/run.sh.diff
original:
updated:
# Commented out pep8 until above bug is fixed.
<|file_sep|>original/.travis/run.sh
#!/bin/bash
set -e
set -x
# Commented out py3 specific apps until #2442/3 resolved:
# https://github.com/travis-ci/apt-package-whitelist/issues
#if [[ $TRAVIS_PYTHON_VERSION == 3* ]]; then
# PEP8="python3 /usr/lib/python3/dist-packages/pep8.py"
# PYLINT=/usr/bin/pylint3
#else
PEP8=/usr/bin/pep8
PYLINT=/usr/bin/pylint
#fi
# Ignore imports not at start and line-too-long)
$PEP8 --ignore=E402,E501 clipster
$PYLINT --errors-only clipster
python tests/tests.py
<|file_sep|>current/.travis/run.sh
#!/bin/bash
set -e
set -x
# Commented out py3 specific apps until #2442/3 resolved:
# https://github.com/travis-ci/apt-package-whitelist/issues
#if [[ $TRAVIS_PYTHON_VERSION == 3* ]]; then
# PEP8="python3 /usr/lib/python3/dist-packages/pep8.py"
# PYLINT=/usr/bin/pylint3
#else
PEP8=/usr/bin/pep8
PYLINT=/usr/bin/pylint
#fi
# Commented out pep8 until above bug is fixed.
# Ignore imports not at start and line-too-long)
$PEP8 --ignore=E402,E501 clipster
$PYLINT --errors-only clipster
python tests/tests.py
<|file_sep|>updated/.travis/run.sh
#!/bin/bash
set -e
set -x
# Commented out py3 specific apps until #2442/3 resolved:
# https://github.com/travis-ci/apt-package-whitelist/issues
#if [[ $TRAVIS_PYTHON_VERSION == 3* ]]; then
# PEP8="python3 /usr/lib/python3/dist-packages/pep8.py"
# PYLINT=/usr/bin/pylint3
#else
PEP8=/usr/bin/pep8
PYLINT=/usr/bin/pylint
#fi
# Commented out pep8 until above bug is fixed.
# Ignore imports not at start and line-too-long)
#$PEP8 --ignore=E402,E501 clipster
$PYLINT --errors-only clipster
python tests/tests.py | 06d8aebc4204baaf95cd6b154edf82e38814fdda | .travis/run.sh | .travis/run.sh | Shell |
<|file_sep|>.travis.yml.diff
original:
updated:
dist:
trusty
<|file_sep|>original/.travis.yml
language:
- scala
scala:
- 2.11.12
- 2.12.6
jdk:
- oraclejdk8
<|file_sep|>current/.travis.yml
dist:
trusty
language:
- scala
scala:
- 2.11.12
- 2.12.6
jdk:
- oraclejdk8
<|file_sep|>updated/.travis.yml | dist:
trusty
language:
scala
scala:
- 2.11.12
- 2.12.6
jdk:
- oraclejdk8 | <|file_sep|>.travis.yml.diff
original:
updated:
dist:
trusty
<|file_sep|>original/.travis.yml
language:
- scala
scala:
- 2.11.12
- 2.12.6
jdk:
- oraclejdk8
<|file_sep|>current/.travis.yml
dist:
trusty
language:
- scala
scala:
- 2.11.12
- 2.12.6
jdk:
- oraclejdk8
<|file_sep|>updated/.travis.yml
dist:
trusty
language:
scala
scala:
- 2.11.12
- 2.12.6
jdk:
- oraclejdk8 | 2a17e813554379a8b51c14f0a0699ea32e4d0ffa | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/addons/l10n_in_hr_payroll/l10n_in_hr_payroll_report.xml
model="hr.payroll.advice"
name="payroll.advice"
rml="l10n_in_hr_payroll/report/report_payroll_advice.rml"
string="Print Advice" />
<report
id="hr_salary_employee_bymonth"
string="Yearly Salary by Head"
model="hr.salary.employee.month"
report_type="qweb-pdf"
file="Yearly Salary By Head"
name="l10n_in_hr_payroll.employee_salary_bymonth"/>
<report
auto="False"
id="yearly_salary"
model="yearly.salary.detail"
name="salary.detail.byyear"
rml="l10n_in_hr_payroll/report/report_hr_yearly_salary_detail.rml"
string="Yearly Salary by Employee" />
<|file_sep|>current/addons/l10n_in_hr_payroll/l10n_in_hr_payroll_report.xml
model="hr.payroll.advice"
name="payroll.advice"
rml="l10n_in_hr_payroll/report/report_payroll_advice.rml"
string="Print Advice" />
<report
id="hr_salary_employee_bymonth"
string="Yearly Salary by Head"
model="hr.salary.employee.month"
report_type="qweb-pdf"
file="Yearly Salary By Head"
name="l10n_in_hr_payroll.employee_salary_bymonth"/>
<report
auto="False"
id="yearly_salary"
model="yearly.salary.detail"
name="salary.detail.byyear"
rml="l10n_in_hr_payroll/report/report_hr_yearly_salary_detail.rml"
string="Yearly Salary by Employee" />
<|file_sep|>updated/addons/l10n_in_hr_payroll/l10n_in_hr_payroll_report.xml | model="hr.payroll.advice"
name="payroll.advice"
rml="l10n_in_hr_payroll/report/report_payroll_advice.rml"
string="Print Advice" />
<report
id="hr_salary_employee_bymonth"
string="Yearly Salary by Head"
model="hr.salary.employee.month"
report_type="qweb-pdf"
file="l10n_in_hr_payroll.employee_salary_bymonth"
name="l10n_in_hr_payroll.employee_salary_bymonth"/>
<report
auto="False"
id="yearly_salary"
model="yearly.salary.detail"
name="salary.detail.byyear"
rml="l10n_in_hr_payroll/report/report_hr_yearly_salary_detail.rml"
string="Yearly Salary by Employee" />
| <|file_sep|>original/addons/l10n_in_hr_payroll/l10n_in_hr_payroll_report.xml
model="hr.payroll.advice"
name="payroll.advice"
rml="l10n_in_hr_payroll/report/report_payroll_advice.rml"
string="Print Advice" />
<report
id="hr_salary_employee_bymonth"
string="Yearly Salary by Head"
model="hr.salary.employee.month"
report_type="qweb-pdf"
file="Yearly Salary By Head"
name="l10n_in_hr_payroll.employee_salary_bymonth"/>
<report
auto="False"
id="yearly_salary"
model="yearly.salary.detail"
name="salary.detail.byyear"
rml="l10n_in_hr_payroll/report/report_hr_yearly_salary_detail.rml"
string="Yearly Salary by Employee" />
<|file_sep|>current/addons/l10n_in_hr_payroll/l10n_in_hr_payroll_report.xml
model="hr.payroll.advice"
name="payroll.advice"
rml="l10n_in_hr_payroll/report/report_payroll_advice.rml"
string="Print Advice" />
<report
id="hr_salary_employee_bymonth"
string="Yearly Salary by Head"
model="hr.salary.employee.month"
report_type="qweb-pdf"
file="Yearly Salary By Head"
name="l10n_in_hr_payroll.employee_salary_bymonth"/>
<report
auto="False"
id="yearly_salary"
model="yearly.salary.detail"
name="salary.detail.byyear"
rml="l10n_in_hr_payroll/report/report_hr_yearly_salary_detail.rml"
string="Yearly Salary by Employee" />
<|file_sep|>updated/addons/l10n_in_hr_payroll/l10n_in_hr_payroll_report.xml
model="hr.payroll.advice"
name="payroll.advice"
rml="l10n_in_hr_payroll/report/report_payroll_advice.rml"
string="Print Advice" />
<report
id="hr_salary_employee_bymonth"
string="Yearly Salary by Head"
model="hr.salary.employee.month"
report_type="qweb-pdf"
file="l10n_in_hr_payroll.employee_salary_bymonth"
name="l10n_in_hr_payroll.employee_salary_bymonth"/>
<report
auto="False"
id="yearly_salary"
model="yearly.salary.detail"
name="salary.detail.byyear"
rml="l10n_in_hr_payroll/report/report_hr_yearly_salary_detail.rml"
string="Yearly Salary by Employee" />
| 2e67cf77de903671f7af9b20955893687a826c91 | addons/l10n_in_hr_payroll/l10n_in_hr_payroll_report.xml | addons/l10n_in_hr_payroll/l10n_in_hr_payroll_report.xml | XML |
<|file_sep|>original/.travis.yml
bundler_args: --without development
language: ruby
rvm:
- 1.9.2
- 1.9.3
- 2.0.0
- 2.1.0
- rbx
- ruby-head
matrix:
include:
- rvm: jruby-19mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-20mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-21mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-head
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
allow_failures:
- rvm: jruby-head
<|file_sep|>current/.travis.yml
bundler_args: --without development
language: ruby
rvm:
- 1.9.2
- 1.9.3
- 2.0.0
- 2.1.0
- rbx
- ruby-head
matrix:
include:
- rvm: jruby-19mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-20mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-21mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-head
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
allow_failures:
- rvm: jruby-head
<|file_sep|>updated/.travis.yml | before_install:
- gem update bundler
- bundle --version
- gem update --system 2.1.11
- gem --version
bundler_args: --without development
language: ruby
rvm:
- 1.9.2
- 1.9.3
- 2.0.0
- 2.1.0
- rbx
- ruby-head
matrix:
include:
- rvm: jruby-19mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-20mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-21mode | <|file_sep|>original/.travis.yml
bundler_args: --without development
language: ruby
rvm:
- 1.9.2
- 1.9.3
- 2.0.0
- 2.1.0
- rbx
- ruby-head
matrix:
include:
- rvm: jruby-19mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-20mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-21mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-head
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
allow_failures:
- rvm: jruby-head
<|file_sep|>current/.travis.yml
bundler_args: --without development
language: ruby
rvm:
- 1.9.2
- 1.9.3
- 2.0.0
- 2.1.0
- rbx
- ruby-head
matrix:
include:
- rvm: jruby-19mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-20mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-21mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-head
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
allow_failures:
- rvm: jruby-head
<|file_sep|>updated/.travis.yml
before_install:
- gem update bundler
- bundle --version
- gem update --system 2.1.11
- gem --version
bundler_args: --without development
language: ruby
rvm:
- 1.9.2
- 1.9.3
- 2.0.0
- 2.1.0
- rbx
- ruby-head
matrix:
include:
- rvm: jruby-19mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-20mode
env: JRUBY_OPTS="$JRUBY_OPTS --debug"
- rvm: jruby-21mode | 6ffb8de7aae454e4f26b7bbe98e27a82d59cc459 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/src/applications/auth/phid/PhabricatorAuthMessagePHIDType.php
public function getPHIDTypeApplicationClass() {
return 'PhabricatorAuthApplication';
}
protected function buildQueryForObjects(
PhabricatorObjectQuery $query,
array $phids) {
return id(new PhabricatorAuthMessageQuery())
->withPHIDs($phids);
}
public function loadHandles(
PhabricatorHandleQuery $query,
array $handles,
array $objects) {
return;
}
}
<|file_sep|>current/src/applications/auth/phid/PhabricatorAuthMessagePHIDType.php
public function getPHIDTypeApplicationClass() {
return 'PhabricatorAuthApplication';
}
protected function buildQueryForObjects(
PhabricatorObjectQuery $query,
array $phids) {
return id(new PhabricatorAuthMessageQuery())
->withPHIDs($phids);
}
public function loadHandles(
PhabricatorHandleQuery $query,
array $handles,
array $objects) {
return;
}
}
<|file_sep|>updated/src/applications/auth/phid/PhabricatorAuthMessagePHIDType.php |
protected function buildQueryForObjects(
PhabricatorObjectQuery $query,
array $phids) {
return id(new PhabricatorAuthMessageQuery())
->withPHIDs($phids);
}
public function loadHandles(
PhabricatorHandleQuery $query,
array $handles,
array $objects) {
foreach ($handles as $phid => $handle) {
$message = $objects[$phid];
$handle->setURI($message->getURI());
}
}
} | <|file_sep|>original/src/applications/auth/phid/PhabricatorAuthMessagePHIDType.php
public function getPHIDTypeApplicationClass() {
return 'PhabricatorAuthApplication';
}
protected function buildQueryForObjects(
PhabricatorObjectQuery $query,
array $phids) {
return id(new PhabricatorAuthMessageQuery())
->withPHIDs($phids);
}
public function loadHandles(
PhabricatorHandleQuery $query,
array $handles,
array $objects) {
return;
}
}
<|file_sep|>current/src/applications/auth/phid/PhabricatorAuthMessagePHIDType.php
public function getPHIDTypeApplicationClass() {
return 'PhabricatorAuthApplication';
}
protected function buildQueryForObjects(
PhabricatorObjectQuery $query,
array $phids) {
return id(new PhabricatorAuthMessageQuery())
->withPHIDs($phids);
}
public function loadHandles(
PhabricatorHandleQuery $query,
array $handles,
array $objects) {
return;
}
}
<|file_sep|>updated/src/applications/auth/phid/PhabricatorAuthMessagePHIDType.php
protected function buildQueryForObjects(
PhabricatorObjectQuery $query,
array $phids) {
return id(new PhabricatorAuthMessageQuery())
->withPHIDs($phids);
}
public function loadHandles(
PhabricatorHandleQuery $query,
array $handles,
array $objects) {
foreach ($handles as $phid => $handle) {
$message = $objects[$phid];
$handle->setURI($message->getURI());
}
}
} | 648d5ce19076ea5d14b4f2d628811079f8472abd | src/applications/auth/phid/PhabricatorAuthMessagePHIDType.php | src/applications/auth/phid/PhabricatorAuthMessagePHIDType.php | PHP |
<|file_sep|>original/helloworld.py
#This is my hello world program
str1='Hello'
str2='Tarun'
print str1 +' '+ str2
# this is my hello world program
print 'Hello World!'
#This is my Hello world program
str1='Hello'
str2='Akash'
print str1 + ' ' + str2 + '!'
#this is a comment
str1='Hello'
str2='Priyanka'
print str1+' '+str2
<|file_sep|>current/helloworld.py
#This is my hello world program
str1='Hello'
str2='Tarun'
print str1 +' '+ str2
# this is my hello world program
print 'Hello World!'
#This is my Hello world program
str1='Hello'
str2='Akash'
print str1 + ' ' + str2 + '!'
#this is a comment
str1='Hello'
str2='Priyanka'
print str1+' '+str2
<|file_sep|>updated/helloworld.py |
print "helloworld" | <|file_sep|>original/helloworld.py
#This is my hello world program
str1='Hello'
str2='Tarun'
print str1 +' '+ str2
# this is my hello world program
print 'Hello World!'
#This is my Hello world program
str1='Hello'
str2='Akash'
print str1 + ' ' + str2 + '!'
#this is a comment
str1='Hello'
str2='Priyanka'
print str1+' '+str2
<|file_sep|>current/helloworld.py
#This is my hello world program
str1='Hello'
str2='Tarun'
print str1 +' '+ str2
# this is my hello world program
print 'Hello World!'
#This is my Hello world program
str1='Hello'
str2='Akash'
print str1 + ' ' + str2 + '!'
#this is a comment
str1='Hello'
str2='Priyanka'
print str1+' '+str2
<|file_sep|>updated/helloworld.py
print "helloworld" | 1ecbd06083ac65a9520bcf0f87c5f5f1b4a4e532 | helloworld.py | helloworld.py | Python |
<|file_sep|>original/SkyFloatingLabelTextField.podspec
Pod::Spec.new do |s|
s.name = "SkyFloatingLabelTextField"
s.version = "3.8.0"
s.summary = "A beautiful, flexible and customizable textfield that minimizes space when displaying additional context."
s.homepage = "https://github.com/Skyscanner/SkyFloatingLabelTextField"
s.license = { :type => "Apache 2.0", :file => "LICENSE" }
s.authors = "Daniel Langh, Gergely Orosz, Raimon Lapuente"
s.ios.deployment_target = "8.0"
s.source = { :git => "https://github.com/Skyscanner/SkyFloatingLabelTextField.git", :tag => "v#{s.version}" }
s.source_files = 'Sources/*.swift'
s.swift_versions = ['3.1', '3.2', '3.3', '3.4', '4.0', '4.1', '4.2', '4.3', '5.0']
end
<|file_sep|>current/SkyFloatingLabelTextField.podspec
Pod::Spec.new do |s|
s.name = "SkyFloatingLabelTextField"
s.version = "3.8.0"
s.summary = "A beautiful, flexible and customizable textfield that minimizes space when displaying additional context."
s.homepage = "https://github.com/Skyscanner/SkyFloatingLabelTextField"
s.license = { :type => "Apache 2.0", :file => "LICENSE" }
s.authors = "Daniel Langh, Gergely Orosz, Raimon Lapuente"
s.ios.deployment_target = "8.0"
s.source = { :git => "https://github.com/Skyscanner/SkyFloatingLabelTextField.git", :tag => "v#{s.version}" }
s.source_files = 'Sources/*.swift'
s.swift_versions = ['3.1', '3.2', '3.3', '3.4', '4.0', '4.1', '4.2', '4.3', '5.0']
end
<|file_sep|>updated/SkyFloatingLabelTextField.podspec | Pod::Spec.new do |s|
s.name = "SkyFloatingLabelTextField"
s.version = "3.7.0"
s.summary = "A beautiful, flexible and customizable textfield that minimizes space when displaying additional context."
s.homepage = "https://github.com/Skyscanner/SkyFloatingLabelTextField"
s.license = { :type => "Apache 2.0", :file => "LICENSE" }
s.authors = "Daniel Langh, Gergely Orosz, Raimon Lapuente"
s.ios.deployment_target = "8.0"
s.source = { :git => "https://github.com/Skyscanner/SkyFloatingLabelTextField.git", :tag => "v#{s.version}" }
s.source_files = 'Sources/*.swift'
s.swift_versions = ['3.1', '3.2', '3.3', '3.4', '4.0', '4.1', '4.2', '4.3', '5.0']
end | <|file_sep|>original/SkyFloatingLabelTextField.podspec
Pod::Spec.new do |s|
s.name = "SkyFloatingLabelTextField"
s.version = "3.8.0"
s.summary = "A beautiful, flexible and customizable textfield that minimizes space when displaying additional context."
s.homepage = "https://github.com/Skyscanner/SkyFloatingLabelTextField"
s.license = { :type => "Apache 2.0", :file => "LICENSE" }
s.authors = "Daniel Langh, Gergely Orosz, Raimon Lapuente"
s.ios.deployment_target = "8.0"
s.source = { :git => "https://github.com/Skyscanner/SkyFloatingLabelTextField.git", :tag => "v#{s.version}" }
s.source_files = 'Sources/*.swift'
s.swift_versions = ['3.1', '3.2', '3.3', '3.4', '4.0', '4.1', '4.2', '4.3', '5.0']
end
<|file_sep|>current/SkyFloatingLabelTextField.podspec
Pod::Spec.new do |s|
s.name = "SkyFloatingLabelTextField"
s.version = "3.8.0"
s.summary = "A beautiful, flexible and customizable textfield that minimizes space when displaying additional context."
s.homepage = "https://github.com/Skyscanner/SkyFloatingLabelTextField"
s.license = { :type => "Apache 2.0", :file => "LICENSE" }
s.authors = "Daniel Langh, Gergely Orosz, Raimon Lapuente"
s.ios.deployment_target = "8.0"
s.source = { :git => "https://github.com/Skyscanner/SkyFloatingLabelTextField.git", :tag => "v#{s.version}" }
s.source_files = 'Sources/*.swift'
s.swift_versions = ['3.1', '3.2', '3.3', '3.4', '4.0', '4.1', '4.2', '4.3', '5.0']
end
<|file_sep|>updated/SkyFloatingLabelTextField.podspec
Pod::Spec.new do |s|
s.name = "SkyFloatingLabelTextField"
s.version = "3.7.0"
s.summary = "A beautiful, flexible and customizable textfield that minimizes space when displaying additional context."
s.homepage = "https://github.com/Skyscanner/SkyFloatingLabelTextField"
s.license = { :type => "Apache 2.0", :file => "LICENSE" }
s.authors = "Daniel Langh, Gergely Orosz, Raimon Lapuente"
s.ios.deployment_target = "8.0"
s.source = { :git => "https://github.com/Skyscanner/SkyFloatingLabelTextField.git", :tag => "v#{s.version}" }
s.source_files = 'Sources/*.swift'
s.swift_versions = ['3.1', '3.2', '3.3', '3.4', '4.0', '4.1', '4.2', '4.3', '5.0']
end | 24e84c4b5d72130c0ea480a65923cc90d2734aa0 | SkyFloatingLabelTextField.podspec | SkyFloatingLabelTextField.podspec | Ruby |
<|file_sep|>content/events/2018-zurich/program/workshop-with-brian-christner.md.diff
original:
slideshare = ""
updated:
slideshare = "https://www.slideshare.net/BrianChristner/56kcloud-docker-training"
<|file_sep|>original/content/events/2018-zurich/program/workshop-with-brian-christner.md
+++
City = "Zurich"
Year = "2018"
talk_date = ""
talk_start_time = ""
talk_end_time = ""
title = "Introduction to Docker - Hands On"
type = "talk"
speakers = ["brian-christner"]
youtube = ""
vimeo = ""
speakerdeck = ""
slideshare = ""
slides = ""
+++
A hands-on introductory course to Docker and Docker Orchestration. The course will cover
the basics of what are containers, how to create a container, create a Docker Swarm, and
finally, create a web-app and deploy it to a Docker Swarm.
<|file_sep|>current/content/events/2018-zurich/program/workshop-with-brian-christner.md
+++
City = "Zurich"
Year = "2018"
talk_date = ""
talk_start_time = ""
talk_end_time = ""
title = "Introduction to Docker - Hands On"
type = "talk"
speakers = ["brian-christner"]
youtube = ""
vimeo = ""
speakerdeck = ""
slideshare = "https://www.slideshare.net/BrianChristner/56kcloud-docker-training"
slides = ""
+++
A hands-on introductory course to Docker and Docker Orchestration. The course will cover
the basics of what are containers, how to create a container, create a Docker Swarm, and
finally, create a web-app and deploy it to a Docker Swarm.
<|file_sep|>updated/content/events/2018-zurich/program/workshop-with-brian-christner.md | +++
City = "Zurich"
Year = "2018"
talk_date = ""
talk_start_time = ""
talk_end_time = ""
title = "Introduction to Docker - Hands On"
type = "talk"
speakers = ["brian-christner"]
youtube = ""
vimeo = ""
speakerdeck = ""
slideshare = "https://www.slideshare.net/BrianChristner/56kcloud-docker-training"
slides = ""
+++
A hands-on introductory course to Docker and Docker Orchestration. The course will cover
the basics of what are containers, how to create a container, create a Docker Swarm, and
finally, create a web-app and deploy it to a Docker Swarm.
[Training Material](https://github.com/56kcloud/Training/blob/master/DevOpsDays/readme.md) | <|file_sep|>content/events/2018-zurich/program/workshop-with-brian-christner.md.diff
original:
slideshare = ""
updated:
slideshare = "https://www.slideshare.net/BrianChristner/56kcloud-docker-training"
<|file_sep|>original/content/events/2018-zurich/program/workshop-with-brian-christner.md
+++
City = "Zurich"
Year = "2018"
talk_date = ""
talk_start_time = ""
talk_end_time = ""
title = "Introduction to Docker - Hands On"
type = "talk"
speakers = ["brian-christner"]
youtube = ""
vimeo = ""
speakerdeck = ""
slideshare = ""
slides = ""
+++
A hands-on introductory course to Docker and Docker Orchestration. The course will cover
the basics of what are containers, how to create a container, create a Docker Swarm, and
finally, create a web-app and deploy it to a Docker Swarm.
<|file_sep|>current/content/events/2018-zurich/program/workshop-with-brian-christner.md
+++
City = "Zurich"
Year = "2018"
talk_date = ""
talk_start_time = ""
talk_end_time = ""
title = "Introduction to Docker - Hands On"
type = "talk"
speakers = ["brian-christner"]
youtube = ""
vimeo = ""
speakerdeck = ""
slideshare = "https://www.slideshare.net/BrianChristner/56kcloud-docker-training"
slides = ""
+++
A hands-on introductory course to Docker and Docker Orchestration. The course will cover
the basics of what are containers, how to create a container, create a Docker Swarm, and
finally, create a web-app and deploy it to a Docker Swarm.
<|file_sep|>updated/content/events/2018-zurich/program/workshop-with-brian-christner.md
+++
City = "Zurich"
Year = "2018"
talk_date = ""
talk_start_time = ""
talk_end_time = ""
title = "Introduction to Docker - Hands On"
type = "talk"
speakers = ["brian-christner"]
youtube = ""
vimeo = ""
speakerdeck = ""
slideshare = "https://www.slideshare.net/BrianChristner/56kcloud-docker-training"
slides = ""
+++
A hands-on introductory course to Docker and Docker Orchestration. The course will cover
the basics of what are containers, how to create a container, create a Docker Swarm, and
finally, create a web-app and deploy it to a Docker Swarm.
[Training Material](https://github.com/56kcloud/Training/blob/master/DevOpsDays/readme.md) | 5515a03d179639a654f6045f6ee5199ca5e12a92 | content/events/2018-zurich/program/workshop-with-brian-christner.md | content/events/2018-zurich/program/workshop-with-brian-christner.md | Markdown |
<|file_sep|>original/.gitlab-ci.yml
variables:
GIT_SUBMODULE_STRATEGY: recursive
image: registry.local/builder-jdk8
stages:
- build
- upload
build_package:
stage: build
cache:
paths:
- build/
- "addon/*/*/build"
key: "$CI_BUILD_REF_NAME"
untracked: true
script:
- gradle build
- ( cd addon/ObjectTrap/PyScript && gradle build )
- ( cd addon/ObjectTrap/ScalaScript && gradle build )
<|file_sep|>current/.gitlab-ci.yml
variables:
GIT_SUBMODULE_STRATEGY: recursive
image: registry.local/builder-jdk8
stages:
- build
- upload
build_package:
stage: build
cache:
paths:
- build/
- "addon/*/*/build"
key: "$CI_BUILD_REF_NAME"
untracked: true
script:
- gradle build
- ( cd addon/ObjectTrap/PyScript && gradle build )
- ( cd addon/ObjectTrap/ScalaScript && gradle build )
<|file_sep|>updated/.gitlab-ci.yml | variables:
GIT_SUBMODULE_STRATEGY: recursive
image: registry.local/builder-jdk8:latest
stages:
- build
- upload
build_package:
stage: build
cache:
paths:
- build/
- "addon/*/*/build"
key: "$CI_BUILD_REF_NAME"
untracked: true
script:
- gradle build
- ( cd addon/ObjectTrap/PyScript && gradle build )
- ( cd addon/ObjectTrap/ScalaScript && gradle build ) | <|file_sep|>original/.gitlab-ci.yml
variables:
GIT_SUBMODULE_STRATEGY: recursive
image: registry.local/builder-jdk8
stages:
- build
- upload
build_package:
stage: build
cache:
paths:
- build/
- "addon/*/*/build"
key: "$CI_BUILD_REF_NAME"
untracked: true
script:
- gradle build
- ( cd addon/ObjectTrap/PyScript && gradle build )
- ( cd addon/ObjectTrap/ScalaScript && gradle build )
<|file_sep|>current/.gitlab-ci.yml
variables:
GIT_SUBMODULE_STRATEGY: recursive
image: registry.local/builder-jdk8
stages:
- build
- upload
build_package:
stage: build
cache:
paths:
- build/
- "addon/*/*/build"
key: "$CI_BUILD_REF_NAME"
untracked: true
script:
- gradle build
- ( cd addon/ObjectTrap/PyScript && gradle build )
- ( cd addon/ObjectTrap/ScalaScript && gradle build )
<|file_sep|>updated/.gitlab-ci.yml
variables:
GIT_SUBMODULE_STRATEGY: recursive
image: registry.local/builder-jdk8:latest
stages:
- build
- upload
build_package:
stage: build
cache:
paths:
- build/
- "addon/*/*/build"
key: "$CI_BUILD_REF_NAME"
untracked: true
script:
- gradle build
- ( cd addon/ObjectTrap/PyScript && gradle build )
- ( cd addon/ObjectTrap/ScalaScript && gradle build ) | bb62cd6ada08ca322afcfdb3784c6a7ce4ae5de2 | .gitlab-ci.yml | .gitlab-ci.yml | YAML |
<|file_sep|>original/conda-recipe/meta.yaml
- gsw
- numpy
- scipy
- netcdf4
test:
imports:
- gutils
- gutils.ctd
- gutils.gps
- gutils.yo
- gutils.yo.filters
- gutils.nc
- gutils.gbdr
- gutils.gbdr.methods
about:
home: https://github.com/axiom-data-science/GUTILS
license: MIT License
license_file: LICENSE
summary: A set of Python utilities for reading, merging, and post processing Teledyne Webb Slocum Glider data
<|file_sep|>current/conda-recipe/meta.yaml
- gsw
- numpy
- scipy
- netcdf4
test:
imports:
- gutils
- gutils.ctd
- gutils.gps
- gutils.yo
- gutils.yo.filters
- gutils.nc
- gutils.gbdr
- gutils.gbdr.methods
about:
home: https://github.com/axiom-data-science/GUTILS
license: MIT License
license_file: LICENSE
summary: A set of Python utilities for reading, merging, and post processing Teledyne Webb Slocum Glider data
<|file_sep|>updated/conda-recipe/meta.yaml | - netcdf4
test:
imports:
- gutils
- gutils.ctd
- gutils.gps
- gutils.yo
- gutils.yo.filters
- gutils.nc
- gutils.gbdr
- gutils.gbdr.methods
commands:
- create_glider_netcdf.py --help
about:
home: https://github.com/axiom-data-science/GUTILS
license: MIT License
license_file: LICENSE
summary: A set of Python utilities for reading, merging, and post processing Teledyne Webb Slocum Glider data | <|file_sep|>original/conda-recipe/meta.yaml
- gsw
- numpy
- scipy
- netcdf4
test:
imports:
- gutils
- gutils.ctd
- gutils.gps
- gutils.yo
- gutils.yo.filters
- gutils.nc
- gutils.gbdr
- gutils.gbdr.methods
about:
home: https://github.com/axiom-data-science/GUTILS
license: MIT License
license_file: LICENSE
summary: A set of Python utilities for reading, merging, and post processing Teledyne Webb Slocum Glider data
<|file_sep|>current/conda-recipe/meta.yaml
- gsw
- numpy
- scipy
- netcdf4
test:
imports:
- gutils
- gutils.ctd
- gutils.gps
- gutils.yo
- gutils.yo.filters
- gutils.nc
- gutils.gbdr
- gutils.gbdr.methods
about:
home: https://github.com/axiom-data-science/GUTILS
license: MIT License
license_file: LICENSE
summary: A set of Python utilities for reading, merging, and post processing Teledyne Webb Slocum Glider data
<|file_sep|>updated/conda-recipe/meta.yaml
- netcdf4
test:
imports:
- gutils
- gutils.ctd
- gutils.gps
- gutils.yo
- gutils.yo.filters
- gutils.nc
- gutils.gbdr
- gutils.gbdr.methods
commands:
- create_glider_netcdf.py --help
about:
home: https://github.com/axiom-data-science/GUTILS
license: MIT License
license_file: LICENSE
summary: A set of Python utilities for reading, merging, and post processing Teledyne Webb Slocum Glider data | 2a48ddaf121133c574dd318b4eb40263fb3c33f8 | conda-recipe/meta.yaml | conda-recipe/meta.yaml | YAML |
<|file_sep|>original/sample/src/main/AndroidManifest.xml
<?xml version="1.0" encoding="utf-8"?>
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="com.github.pedrovgs.nox.sample">
<application
android:icon="@drawable/ic_launcher"
android:label="@string/app_name"
android:theme="@style/AppTheme">
<activity
android:name=".MainActivity"
android:label="@string/app_name">
</activity>
</application>
</manifest>
<|file_sep|>current/sample/src/main/AndroidManifest.xml
<?xml version="1.0" encoding="utf-8"?>
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="com.github.pedrovgs.nox.sample">
<application
android:icon="@drawable/ic_launcher"
android:label="@string/app_name"
android:theme="@style/AppTheme">
<activity
android:name=".MainActivity"
android:label="@string/app_name">
</activity>
</application>
</manifest>
<|file_sep|>updated/sample/src/main/AndroidManifest.xml | <?xml version="1.0" encoding="utf-8"?>
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="com.github.pedrovgs.nox.sample">
<application
android:icon="@drawable/ic_launcher"
android:label="@string/app_name"
android:theme="@style/AppTheme">
<activity
android:name=".MainActivity"
android:label="@string/app_name">
<intent-filter>
<action android:name="android.intent.action.MAIN"/>
<category android:name="android.intent.category.LAUNCHER"/>
</intent-filter>
</activity>
</application>
</manifest> | <|file_sep|>original/sample/src/main/AndroidManifest.xml
<?xml version="1.0" encoding="utf-8"?>
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="com.github.pedrovgs.nox.sample">
<application
android:icon="@drawable/ic_launcher"
android:label="@string/app_name"
android:theme="@style/AppTheme">
<activity
android:name=".MainActivity"
android:label="@string/app_name">
</activity>
</application>
</manifest>
<|file_sep|>current/sample/src/main/AndroidManifest.xml
<?xml version="1.0" encoding="utf-8"?>
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="com.github.pedrovgs.nox.sample">
<application
android:icon="@drawable/ic_launcher"
android:label="@string/app_name"
android:theme="@style/AppTheme">
<activity
android:name=".MainActivity"
android:label="@string/app_name">
</activity>
</application>
</manifest>
<|file_sep|>updated/sample/src/main/AndroidManifest.xml
<?xml version="1.0" encoding="utf-8"?>
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="com.github.pedrovgs.nox.sample">
<application
android:icon="@drawable/ic_launcher"
android:label="@string/app_name"
android:theme="@style/AppTheme">
<activity
android:name=".MainActivity"
android:label="@string/app_name">
<intent-filter>
<action android:name="android.intent.action.MAIN"/>
<category android:name="android.intent.category.LAUNCHER"/>
</intent-filter>
</activity>
</application>
</manifest> | 9a84b717702b1523a48b3568e5e5e22033b1aa73 | sample/src/main/AndroidManifest.xml | sample/src/main/AndroidManifest.xml | XML |
<|file_sep|>original/app/views/people/_person.html.erb
<% if person.image? %>
<%= cdn_image_tag(person.image, state: person.state, size: "60x60", alt: "") %>
<% else %>
<%= image_tag("placeholder.png", size: "60x60", alt: "") %>
<% end %>
</div>
<div class="person-info">
<h2><%= link_to person.name, person_path(person["state"], person) %> <% if person.verified? -%><i class="icon-ok-sign is-verified"><span class="verified-tooltip">AskThem Verified</span></i><% end -%></h2>
<span class="jurisdiction"><%= person_attributes(person) %></span>
<span class="stats">
<%- if person.questions.count > 0 %>
<span><%= pluralize person.questions.count, "Question" %></span>
<%- if person.questions_answered.count > 0 %>
<span><%= pluralize person.questions_answered.count, "Answer" %></span>
<%- end %>
<%- else %>
<%= link_to "Ask the first question >>", new_question_path(person.state, person: @person) -%>
<%- end %>
</span>
</div>
</li>
<|file_sep|>current/app/views/people/_person.html.erb
<% if person.image? %>
<%= cdn_image_tag(person.image, state: person.state, size: "60x60", alt: "") %>
<% else %>
<%= image_tag("placeholder.png", size: "60x60", alt: "") %>
<% end %>
</div>
<div class="person-info">
<h2><%= link_to person.name, person_path(person["state"], person) %> <% if person.verified? -%><i class="icon-ok-sign is-verified"><span class="verified-tooltip">AskThem Verified</span></i><% end -%></h2>
<span class="jurisdiction"><%= person_attributes(person) %></span>
<span class="stats">
<%- if person.questions.count > 0 %>
<span><%= pluralize person.questions.count, "Question" %></span>
<%- if person.questions_answered.count > 0 %>
<span><%= pluralize person.questions_answered.count, "Answer" %></span>
<%- end %>
<%- else %>
<%= link_to "Ask the first question >>", new_question_path(person.state, person: @person) -%>
<%- end %>
</span>
</div>
</li>
<|file_sep|>updated/app/views/people/_person.html.erb | <% if person.image? %>
<%= cdn_image_tag(person.image, state: person.state, size: "60x60", alt: "") %>
<% else %>
<%= image_tag("placeholder.png", size: "60x60", alt: "") %>
<% end %>
</div>
<div class="person-info">
<h2><%= link_to person.name, person_path(person["state"], person) %> <% if person.verified? -%><i class="icon-ok-sign is-verified"><span class="verified-tooltip">AskThem Verified</span></i><% end -%></h2>
<span class="jurisdiction"><%= person_attributes(person) %></span>
<span class="stats">
<%- if person.questions.count > 0 %>
<span><%= pluralize person.questions.count, "Question" %></span>
<%- if person.questions_answered.count > 0 %>
<span><%= pluralize person.questions_answered.count, "Answer" %></span>
<%- end %>
<%- else %>
<%= link_to "Ask the first question >>", new_question_path(person.state, person: person) -%>
<%- end %>
</span>
</div>
</li> | <|file_sep|>original/app/views/people/_person.html.erb
<% if person.image? %>
<%= cdn_image_tag(person.image, state: person.state, size: "60x60", alt: "") %>
<% else %>
<%= image_tag("placeholder.png", size: "60x60", alt: "") %>
<% end %>
</div>
<div class="person-info">
<h2><%= link_to person.name, person_path(person["state"], person) %> <% if person.verified? -%><i class="icon-ok-sign is-verified"><span class="verified-tooltip">AskThem Verified</span></i><% end -%></h2>
<span class="jurisdiction"><%= person_attributes(person) %></span>
<span class="stats">
<%- if person.questions.count > 0 %>
<span><%= pluralize person.questions.count, "Question" %></span>
<%- if person.questions_answered.count > 0 %>
<span><%= pluralize person.questions_answered.count, "Answer" %></span>
<%- end %>
<%- else %>
<%= link_to "Ask the first question >>", new_question_path(person.state, person: @person) -%>
<%- end %>
</span>
</div>
</li>
<|file_sep|>current/app/views/people/_person.html.erb
<% if person.image? %>
<%= cdn_image_tag(person.image, state: person.state, size: "60x60", alt: "") %>
<% else %>
<%= image_tag("placeholder.png", size: "60x60", alt: "") %>
<% end %>
</div>
<div class="person-info">
<h2><%= link_to person.name, person_path(person["state"], person) %> <% if person.verified? -%><i class="icon-ok-sign is-verified"><span class="verified-tooltip">AskThem Verified</span></i><% end -%></h2>
<span class="jurisdiction"><%= person_attributes(person) %></span>
<span class="stats">
<%- if person.questions.count > 0 %>
<span><%= pluralize person.questions.count, "Question" %></span>
<%- if person.questions_answered.count > 0 %>
<span><%= pluralize person.questions_answered.count, "Answer" %></span>
<%- end %>
<%- else %>
<%= link_to "Ask the first question >>", new_question_path(person.state, person: @person) -%>
<%- end %>
</span>
</div>
</li>
<|file_sep|>updated/app/views/people/_person.html.erb
<% if person.image? %>
<%= cdn_image_tag(person.image, state: person.state, size: "60x60", alt: "") %>
<% else %>
<%= image_tag("placeholder.png", size: "60x60", alt: "") %>
<% end %>
</div>
<div class="person-info">
<h2><%= link_to person.name, person_path(person["state"], person) %> <% if person.verified? -%><i class="icon-ok-sign is-verified"><span class="verified-tooltip">AskThem Verified</span></i><% end -%></h2>
<span class="jurisdiction"><%= person_attributes(person) %></span>
<span class="stats">
<%- if person.questions.count > 0 %>
<span><%= pluralize person.questions.count, "Question" %></span>
<%- if person.questions_answered.count > 0 %>
<span><%= pluralize person.questions_answered.count, "Answer" %></span>
<%- end %>
<%- else %>
<%= link_to "Ask the first question >>", new_question_path(person.state, person: person) -%>
<%- end %>
</span>
</div>
</li> | 74933afcbfe466f04cb32bb9503639bcdaa0b162 | app/views/people/_person.html.erb | app/views/people/_person.html.erb | HTML+ERB |
<|file_sep|>original/src/CMakeLists.txt
LINK_DIRECTORIES( ${LIBEBOOK_LIBRARY_DIRS} ${LIBECAL_LIBRARY_DIRS} ${LIBEDATABOOK_LIBRARY_DIRS} ${LIBEDATACAL_LIBRARY_DIRS} ${LIBEDATASERVER_LIBRARY_DIRS} )
INCLUDE_DIRECTORIES( ${LIBEBOOK_INCLUDE_DIRS} ${LIBECAL_INCLUDE_DIRS} ${LIBEDATABOOK_INCLUDE_DIRS} ${LIBEDATACAL_INCLUDE_DIRS} ${LIBEDATASERVER_INCLUDE_DIRS} ${GCONF2_INCLUDE_DIRS} ${OPENSYNC_INCLUDE_DIR} ${GLIB2_INCLUDE_DIR} ${LIBXML2_INCLUDE_DIRS} )
SET( evo2_sync_LIB_SRCS
evolution2_sync.c
evolution2_ebook.c
evolution2_ecal.c
evolution2_memo.c
evolution2_etodo.c
)
OPENSYNC_PLUGIN_ADD( evo2-sync ${evo2_sync_LIB_SRCS} )
TARGET_LINK_LIBRARIES( evo2-sync ${LIBEBOOK_LIBRARIES} ${LIBECAL_LIBRARIES} ${LIBEDATABOOK_LIBRARIES} ${LIBEDATACAL_LIBRARIES} ${LIBEDATASERVER_LIBRARIES} ${OPENSYNC_LIBRARIES} ${LIBXML2_LIBRARIES} )
###### INSTALL ###################
OPENSYNC_PLUGIN_INSTALL( evo2-sync )
OPENSYNC_PLUGIN_CONFIG( evo2-sync )
<|file_sep|>current/src/CMakeLists.txt
LINK_DIRECTORIES( ${LIBEBOOK_LIBRARY_DIRS} ${LIBECAL_LIBRARY_DIRS} ${LIBEDATABOOK_LIBRARY_DIRS} ${LIBEDATACAL_LIBRARY_DIRS} ${LIBEDATASERVER_LIBRARY_DIRS} )
INCLUDE_DIRECTORIES( ${LIBEBOOK_INCLUDE_DIRS} ${LIBECAL_INCLUDE_DIRS} ${LIBEDATABOOK_INCLUDE_DIRS} ${LIBEDATACAL_INCLUDE_DIRS} ${LIBEDATASERVER_INCLUDE_DIRS} ${GCONF2_INCLUDE_DIRS} ${OPENSYNC_INCLUDE_DIR} ${GLIB2_INCLUDE_DIR} ${LIBXML2_INCLUDE_DIRS} )
SET( evo2_sync_LIB_SRCS
evolution2_sync.c
evolution2_ebook.c
evolution2_ecal.c
evolution2_memo.c
evolution2_etodo.c
)
OPENSYNC_PLUGIN_ADD( evo2-sync ${evo2_sync_LIB_SRCS} )
TARGET_LINK_LIBRARIES( evo2-sync ${LIBEBOOK_LIBRARIES} ${LIBECAL_LIBRARIES} ${LIBEDATABOOK_LIBRARIES} ${LIBEDATACAL_LIBRARIES} ${LIBEDATASERVER_LIBRARIES} ${OPENSYNC_LIBRARIES} ${LIBXML2_LIBRARIES} )
###### INSTALL ###################
OPENSYNC_PLUGIN_INSTALL( evo2-sync )
OPENSYNC_PLUGIN_CONFIG( evo2-sync )
<|file_sep|>updated/src/CMakeLists.txt | LINK_DIRECTORIES( ${LIBEBOOK_LIBRARY_DIRS} ${LIBECAL_LIBRARY_DIRS} ${LIBEDATABOOK_LIBRARY_DIRS} ${LIBEDATACAL_LIBRARY_DIRS} ${LIBEDATASERVER_LIBRARY_DIRS} ${GCONF2_LIBRARY_DIRS} ${GLIB2_LIBRARY_DIR} ${LIBXML2_LIBRARY_DIRS} )
INCLUDE_DIRECTORIES( ${LIBEBOOK_INCLUDE_DIRS} ${LIBECAL_INCLUDE_DIRS} ${LIBEDATABOOK_INCLUDE_DIRS} ${LIBEDATACAL_INCLUDE_DIRS} ${LIBEDATASERVER_INCLUDE_DIRS} ${GCONF2_INCLUDE_DIRS} ${OPENSYNC_INCLUDE_DIR} ${GLIB2_INCLUDE_DIR} ${LIBXML2_INCLUDE_DIRS} )
SET( evo2_sync_LIB_SRCS
evolution2_sync.c
evolution2_ebook.c
evolution2_ecal.c
evolution2_memo.c
evolution2_etodo.c
)
OPENSYNC_PLUGIN_ADD( evo2-sync ${evo2_sync_LIB_SRCS} )
TARGET_LINK_LIBRARIES( evo2-sync ${LIBEBOOK_LIBRARIES} ${LIBECAL_LIBRARIES} ${LIBEDATABOOK_LIBRARIES} ${LIBEDATACAL_LIBRARIES} ${LIBEDATASERVER_LIBRARIES} ${OPENSYNC_LIBRARIES} ${LIBXML2_LIBRARIES} )
###### INSTALL ###################
OPENSYNC_PLUGIN_INSTALL( evo2-sync )
OPENSYNC_PLUGIN_CONFIG( evo2-sync ) | <|file_sep|>original/src/CMakeLists.txt
LINK_DIRECTORIES( ${LIBEBOOK_LIBRARY_DIRS} ${LIBECAL_LIBRARY_DIRS} ${LIBEDATABOOK_LIBRARY_DIRS} ${LIBEDATACAL_LIBRARY_DIRS} ${LIBEDATASERVER_LIBRARY_DIRS} )
INCLUDE_DIRECTORIES( ${LIBEBOOK_INCLUDE_DIRS} ${LIBECAL_INCLUDE_DIRS} ${LIBEDATABOOK_INCLUDE_DIRS} ${LIBEDATACAL_INCLUDE_DIRS} ${LIBEDATASERVER_INCLUDE_DIRS} ${GCONF2_INCLUDE_DIRS} ${OPENSYNC_INCLUDE_DIR} ${GLIB2_INCLUDE_DIR} ${LIBXML2_INCLUDE_DIRS} )
SET( evo2_sync_LIB_SRCS
evolution2_sync.c
evolution2_ebook.c
evolution2_ecal.c
evolution2_memo.c
evolution2_etodo.c
)
OPENSYNC_PLUGIN_ADD( evo2-sync ${evo2_sync_LIB_SRCS} )
TARGET_LINK_LIBRARIES( evo2-sync ${LIBEBOOK_LIBRARIES} ${LIBECAL_LIBRARIES} ${LIBEDATABOOK_LIBRARIES} ${LIBEDATACAL_LIBRARIES} ${LIBEDATASERVER_LIBRARIES} ${OPENSYNC_LIBRARIES} ${LIBXML2_LIBRARIES} )
###### INSTALL ###################
OPENSYNC_PLUGIN_INSTALL( evo2-sync )
OPENSYNC_PLUGIN_CONFIG( evo2-sync )
<|file_sep|>current/src/CMakeLists.txt
LINK_DIRECTORIES( ${LIBEBOOK_LIBRARY_DIRS} ${LIBECAL_LIBRARY_DIRS} ${LIBEDATABOOK_LIBRARY_DIRS} ${LIBEDATACAL_LIBRARY_DIRS} ${LIBEDATASERVER_LIBRARY_DIRS} )
INCLUDE_DIRECTORIES( ${LIBEBOOK_INCLUDE_DIRS} ${LIBECAL_INCLUDE_DIRS} ${LIBEDATABOOK_INCLUDE_DIRS} ${LIBEDATACAL_INCLUDE_DIRS} ${LIBEDATASERVER_INCLUDE_DIRS} ${GCONF2_INCLUDE_DIRS} ${OPENSYNC_INCLUDE_DIR} ${GLIB2_INCLUDE_DIR} ${LIBXML2_INCLUDE_DIRS} )
SET( evo2_sync_LIB_SRCS
evolution2_sync.c
evolution2_ebook.c
evolution2_ecal.c
evolution2_memo.c
evolution2_etodo.c
)
OPENSYNC_PLUGIN_ADD( evo2-sync ${evo2_sync_LIB_SRCS} )
TARGET_LINK_LIBRARIES( evo2-sync ${LIBEBOOK_LIBRARIES} ${LIBECAL_LIBRARIES} ${LIBEDATABOOK_LIBRARIES} ${LIBEDATACAL_LIBRARIES} ${LIBEDATASERVER_LIBRARIES} ${OPENSYNC_LIBRARIES} ${LIBXML2_LIBRARIES} )
###### INSTALL ###################
OPENSYNC_PLUGIN_INSTALL( evo2-sync )
OPENSYNC_PLUGIN_CONFIG( evo2-sync )
<|file_sep|>updated/src/CMakeLists.txt
LINK_DIRECTORIES( ${LIBEBOOK_LIBRARY_DIRS} ${LIBECAL_LIBRARY_DIRS} ${LIBEDATABOOK_LIBRARY_DIRS} ${LIBEDATACAL_LIBRARY_DIRS} ${LIBEDATASERVER_LIBRARY_DIRS} ${GCONF2_LIBRARY_DIRS} ${GLIB2_LIBRARY_DIR} ${LIBXML2_LIBRARY_DIRS} )
INCLUDE_DIRECTORIES( ${LIBEBOOK_INCLUDE_DIRS} ${LIBECAL_INCLUDE_DIRS} ${LIBEDATABOOK_INCLUDE_DIRS} ${LIBEDATACAL_INCLUDE_DIRS} ${LIBEDATASERVER_INCLUDE_DIRS} ${GCONF2_INCLUDE_DIRS} ${OPENSYNC_INCLUDE_DIR} ${GLIB2_INCLUDE_DIR} ${LIBXML2_INCLUDE_DIRS} )
SET( evo2_sync_LIB_SRCS
evolution2_sync.c
evolution2_ebook.c
evolution2_ecal.c
evolution2_memo.c
evolution2_etodo.c
)
OPENSYNC_PLUGIN_ADD( evo2-sync ${evo2_sync_LIB_SRCS} )
TARGET_LINK_LIBRARIES( evo2-sync ${LIBEBOOK_LIBRARIES} ${LIBECAL_LIBRARIES} ${LIBEDATABOOK_LIBRARIES} ${LIBEDATACAL_LIBRARIES} ${LIBEDATASERVER_LIBRARIES} ${OPENSYNC_LIBRARIES} ${LIBXML2_LIBRARIES} )
###### INSTALL ###################
OPENSYNC_PLUGIN_INSTALL( evo2-sync )
OPENSYNC_PLUGIN_CONFIG( evo2-sync ) | bfaeb917f3e407a17d88b4c04deeb2e40e72bfdb | src/CMakeLists.txt | src/CMakeLists.txt | Text |
<|file_sep|>test/karma.js.diff
original:
updated:
/* eslint-disable no-process-env */
<|file_sep|>test/karma.js.diff
original:
updated:
import {expect} from 'chai';
<|file_sep|>test/karma.js.diff
original:
updated:
const KARMA_BROWSER = process.env.KARMA_BROWSER;
afterEach(function() {
if (KARMA_BROWSER) {
process.env.KARMA_BROWSER = KARMA_BROWSER;
} else {
delete process.env.KARMA_BROWSER;
}
});
<|file_sep|>original/test/karma.js
import karma from '../src/karma';
describe('karma config', function() {
it('should generate config', function() {
// Success is not throwing at this point. The simple karma tests
// will do the actual verification
karma({set() {}});
});
});
<|file_sep|>current/test/karma.js
/* eslint-disable no-process-env */
import karma from '../src/karma';
import {expect} from 'chai';
describe('karma config', function() {
const KARMA_BROWSER = process.env.KARMA_BROWSER;
afterEach(function() {
if (KARMA_BROWSER) {
process.env.KARMA_BROWSER = KARMA_BROWSER;
} else {
delete process.env.KARMA_BROWSER;
}
});
it('should generate config', function() {
// Success is not throwing at this point. The simple karma tests
// will do the actual verification
karma({set() {}});
});
});
<|file_sep|>updated/test/karma.js | } else {
delete process.env.KARMA_BROWSER;
}
});
it('should generate config', function() {
// Success is not throwing at this point. The simple karma tests
// will do the actual verification
karma({set() {}});
});
it('should default to chrome browser', function() {
process.env.KARMA_BROWSER = '';
let config;
karma({set(_config) { config = _config; }});
expect(config.browsers).to.eql(['Chrome']);
});
it('should allow custom browser', function() {
process.env.KARMA_BROWSER = 'test!';
| <|file_sep|>test/karma.js.diff
original:
updated:
/* eslint-disable no-process-env */
<|file_sep|>test/karma.js.diff
original:
updated:
import {expect} from 'chai';
<|file_sep|>test/karma.js.diff
original:
updated:
const KARMA_BROWSER = process.env.KARMA_BROWSER;
afterEach(function() {
if (KARMA_BROWSER) {
process.env.KARMA_BROWSER = KARMA_BROWSER;
} else {
delete process.env.KARMA_BROWSER;
}
});
<|file_sep|>original/test/karma.js
import karma from '../src/karma';
describe('karma config', function() {
it('should generate config', function() {
// Success is not throwing at this point. The simple karma tests
// will do the actual verification
karma({set() {}});
});
});
<|file_sep|>current/test/karma.js
/* eslint-disable no-process-env */
import karma from '../src/karma';
import {expect} from 'chai';
describe('karma config', function() {
const KARMA_BROWSER = process.env.KARMA_BROWSER;
afterEach(function() {
if (KARMA_BROWSER) {
process.env.KARMA_BROWSER = KARMA_BROWSER;
} else {
delete process.env.KARMA_BROWSER;
}
});
it('should generate config', function() {
// Success is not throwing at this point. The simple karma tests
// will do the actual verification
karma({set() {}});
});
});
<|file_sep|>updated/test/karma.js
} else {
delete process.env.KARMA_BROWSER;
}
});
it('should generate config', function() {
// Success is not throwing at this point. The simple karma tests
// will do the actual verification
karma({set() {}});
});
it('should default to chrome browser', function() {
process.env.KARMA_BROWSER = '';
let config;
karma({set(_config) { config = _config; }});
expect(config.browsers).to.eql(['Chrome']);
});
it('should allow custom browser', function() {
process.env.KARMA_BROWSER = 'test!';
| 78476fd9e92f986e641c5bb8d07ec37451d04fe9 | test/karma.js | test/karma.js | JavaScript |
<|file_sep|>original/.travis.yml
language: objective-c
xcode_sdk: iphonesimulator
osx_image: xcode8
script:
- xcodebuild test -project Tests/ThirdPartyMailerExample.xcodeproj -scheme ThirdPartyMailerExample -sdk iphonesimulator ONLY_ACTIVE_ARCH=NO
<|file_sep|>current/.travis.yml
language: objective-c
xcode_sdk: iphonesimulator
osx_image: xcode8
script:
- xcodebuild test -project Tests/ThirdPartyMailerExample.xcodeproj -scheme ThirdPartyMailerExample -sdk iphonesimulator ONLY_ACTIVE_ARCH=NO
<|file_sep|>updated/.travis.yml | language: objective-c
xcode_sdk: iphonesimulator
osx_image: xcode8
script:
- xcodebuild test -project Tests/ThirdPartyMailerExample.xcodeproj -scheme ThirdPartyMailerExample -sdk iphonesimulator -destination 'platform=iOS Simulator,name=iPhone 6,OS=10.0' ONLY_ACTIVE_ARCH=NO | <|file_sep|>original/.travis.yml
language: objective-c
xcode_sdk: iphonesimulator
osx_image: xcode8
script:
- xcodebuild test -project Tests/ThirdPartyMailerExample.xcodeproj -scheme ThirdPartyMailerExample -sdk iphonesimulator ONLY_ACTIVE_ARCH=NO
<|file_sep|>current/.travis.yml
language: objective-c
xcode_sdk: iphonesimulator
osx_image: xcode8
script:
- xcodebuild test -project Tests/ThirdPartyMailerExample.xcodeproj -scheme ThirdPartyMailerExample -sdk iphonesimulator ONLY_ACTIVE_ARCH=NO
<|file_sep|>updated/.travis.yml
language: objective-c
xcode_sdk: iphonesimulator
osx_image: xcode8
script:
- xcodebuild test -project Tests/ThirdPartyMailerExample.xcodeproj -scheme ThirdPartyMailerExample -sdk iphonesimulator -destination 'platform=iOS Simulator,name=iPhone 6,OS=10.0' ONLY_ACTIVE_ARCH=NO | 06a2b8063c7dba89212378c2e9dd5afb49b0f8d5 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/requirements-dev.txt
-r requirements.txt
pytest==3.7.4
pytest-mock==1.10.0
sphinx==1.7.9
<|file_sep|>current/requirements-dev.txt
-r requirements.txt
pytest==3.7.4
pytest-mock==1.10.0
sphinx==1.7.9
<|file_sep|>updated/requirements-dev.txt | -r requirements.txt
pytest==3.8.0
pytest-mock==1.10.0
sphinx==1.7.9 | <|file_sep|>original/requirements-dev.txt
-r requirements.txt
pytest==3.7.4
pytest-mock==1.10.0
sphinx==1.7.9
<|file_sep|>current/requirements-dev.txt
-r requirements.txt
pytest==3.7.4
pytest-mock==1.10.0
sphinx==1.7.9
<|file_sep|>updated/requirements-dev.txt
-r requirements.txt
pytest==3.8.0
pytest-mock==1.10.0
sphinx==1.7.9 | 808ce685cdbd6085e6cefd797dcb896a9d4cc0c4 | requirements-dev.txt | requirements-dev.txt | Text |
<|file_sep|>original/app/views/notifications_mailer/email_peek.html.erb
<!DOCTYPE html>
<html>
<body>
<p>
The <%= link_to 'Peek at Our Week(s) and Vocabulary', parents_peek_current_url %> for our upcoming unit has been posted for you to view. Login to see what fun and important learning we will be engaged in during the weeks ahead.
</p>
<%= render 'layouts/mailers/footer' %>
</body>
</html>
<|file_sep|>current/app/views/notifications_mailer/email_peek.html.erb
<!DOCTYPE html>
<html>
<body>
<p>
The <%= link_to 'Peek at Our Week(s) and Vocabulary', parents_peek_current_url %> for our upcoming unit has been posted for you to view. Login to see what fun and important learning we will be engaged in during the weeks ahead.
</p>
<%= render 'layouts/mailers/footer' %>
</body>
</html>
<|file_sep|>updated/app/views/notifications_mailer/email_peek.html.erb | <!DOCTYPE html>
<html>
<body>
<p>
The <%= link_to 'Peek at Our Week(s) and Vocabulary', parents_curriculum_current_peek_url %> for our upcoming unit has been posted for you to view. Login to see what fun and important learning we will be engaged in during the weeks ahead.
</p>
<%= render 'layouts/mailers/footer' %>
</body>
</html> | <|file_sep|>original/app/views/notifications_mailer/email_peek.html.erb
<!DOCTYPE html>
<html>
<body>
<p>
The <%= link_to 'Peek at Our Week(s) and Vocabulary', parents_peek_current_url %> for our upcoming unit has been posted for you to view. Login to see what fun and important learning we will be engaged in during the weeks ahead.
</p>
<%= render 'layouts/mailers/footer' %>
</body>
</html>
<|file_sep|>current/app/views/notifications_mailer/email_peek.html.erb
<!DOCTYPE html>
<html>
<body>
<p>
The <%= link_to 'Peek at Our Week(s) and Vocabulary', parents_peek_current_url %> for our upcoming unit has been posted for you to view. Login to see what fun and important learning we will be engaged in during the weeks ahead.
</p>
<%= render 'layouts/mailers/footer' %>
</body>
</html>
<|file_sep|>updated/app/views/notifications_mailer/email_peek.html.erb
<!DOCTYPE html>
<html>
<body>
<p>
The <%= link_to 'Peek at Our Week(s) and Vocabulary', parents_curriculum_current_peek_url %> for our upcoming unit has been posted for you to view. Login to see what fun and important learning we will be engaged in during the weeks ahead.
</p>
<%= render 'layouts/mailers/footer' %>
</body>
</html> | d1e08960b723682c8db925f3998d45687157c8b7 | app/views/notifications_mailer/email_peek.html.erb | app/views/notifications_mailer/email_peek.html.erb | HTML+ERB |
<|file_sep|>original/src/test/compile-fail/borrowck-multiple-captures.rs
<|file_sep|>current/src/test/compile-fail/borrowck-multiple-captures.rs
<|file_sep|>updated/src/test/compile-fail/borrowck-multiple-captures.rs | // Copyright 2014 The Rust Project Developers. See the COPYRIGHT
// file at the top-level directory of this distribution and at
// http://rust-lang.org/COPYRIGHT.
//
// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
// option. This file may not be copied, modified, or distributed
// except according to those terms.
use std::task;
fn borrow<T>(_: &T) { }
fn different_vars_after_borrows() {
let x1 = box 1;
let p1 = &x1;
let x2 = box 2;
let p2 = &x2;
task::spawn(proc() {
drop(x1); //~ ERROR cannot move `x1` into closure because it is borrowed | <|file_sep|>original/src/test/compile-fail/borrowck-multiple-captures.rs
<|file_sep|>current/src/test/compile-fail/borrowck-multiple-captures.rs
<|file_sep|>updated/src/test/compile-fail/borrowck-multiple-captures.rs
// Copyright 2014 The Rust Project Developers. See the COPYRIGHT
// file at the top-level directory of this distribution and at
// http://rust-lang.org/COPYRIGHT.
//
// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
// option. This file may not be copied, modified, or distributed
// except according to those terms.
use std::task;
fn borrow<T>(_: &T) { }
fn different_vars_after_borrows() {
let x1 = box 1;
let p1 = &x1;
let x2 = box 2;
let p2 = &x2;
task::spawn(proc() {
drop(x1); //~ ERROR cannot move `x1` into closure because it is borrowed | f1542a6a7da088f99bfe2963d29e6ea4b451b432 | src/test/compile-fail/borrowck-multiple-captures.rs | src/test/compile-fail/borrowck-multiple-captures.rs | Rust |
<|file_sep|>original/org.scala-ide.sdt.core/src/org/scalaide/util/internal/SbtUtils.scala
object SbtUtils {
def m2o[S](opt: Maybe[S]): Option[S] = if (opt.isEmpty) None else Some(opt.get)
def readCache(cacheFile: File): Option[(Analysis, MiniSetup)] =
FileBasedStore(cacheFile).get().map(_ match {
case (a: Analysis, i) => (a, i)
case (a, _) => throw new RuntimeException(s"Expected that sbt analysis for $cacheFile is of type ${classOf[Analysis]} but was ${a.getClass}.")
})
def readAnalysis(cacheFile: File): Analysis =
// zinc requires name hashing to be enabled
readCache(cacheFile).map(_._1).getOrElse(Analysis.empty(nameHashing = true))
object NoPosition extends xsbti.Position {
def line(): Maybe[Integer] = Maybe.nothing()
def lineContent(): String = ""
def offset(): Maybe[Integer] = Maybe.nothing[Integer]
def pointer(): Maybe[Integer] = Maybe.nothing[Integer]
def pointerSpace(): Maybe[String] = Maybe.nothing[String]
def sourceFile(): Maybe[File] = Maybe.nothing[File]
def sourcePath(): Maybe[String] = Maybe.nothing[String]
<|file_sep|>current/org.scala-ide.sdt.core/src/org/scalaide/util/internal/SbtUtils.scala
object SbtUtils {
def m2o[S](opt: Maybe[S]): Option[S] = if (opt.isEmpty) None else Some(opt.get)
def readCache(cacheFile: File): Option[(Analysis, MiniSetup)] =
FileBasedStore(cacheFile).get().map(_ match {
case (a: Analysis, i) => (a, i)
case (a, _) => throw new RuntimeException(s"Expected that sbt analysis for $cacheFile is of type ${classOf[Analysis]} but was ${a.getClass}.")
})
def readAnalysis(cacheFile: File): Analysis =
// zinc requires name hashing to be enabled
readCache(cacheFile).map(_._1).getOrElse(Analysis.empty(nameHashing = true))
object NoPosition extends xsbti.Position {
def line(): Maybe[Integer] = Maybe.nothing()
def lineContent(): String = ""
def offset(): Maybe[Integer] = Maybe.nothing[Integer]
def pointer(): Maybe[Integer] = Maybe.nothing[Integer]
def pointerSpace(): Maybe[String] = Maybe.nothing[String]
def sourceFile(): Maybe[File] = Maybe.nothing[File]
def sourcePath(): Maybe[String] = Maybe.nothing[String]
<|file_sep|>updated/org.scala-ide.sdt.core/src/org/scalaide/util/internal/SbtUtils.scala | object SbtUtils {
def m2o[S](opt: Maybe[S]): Option[S] = if (opt.isEmpty) None else Some(opt.get)
def readCache(cacheFile: File): Option[(Analysis, MiniSetup)] =
FileBasedStore(cacheFile).get().map(_ match {
case (a: Analysis, i) => (a, i)
case (a, _) => throw new RuntimeException(s"Expected that sbt analysis for $cacheFile is of type ${classOf[Analysis]} but was ${a.getClass}.")
})
def readAnalysis(cacheFile: File): Analysis =
readCache(cacheFile).map(_._1).getOrElse(Analysis.empty)
object NoPosition extends xsbti.Position {
def line(): Maybe[Integer] = Maybe.nothing()
def lineContent(): String = ""
def offset(): Maybe[Integer] = Maybe.nothing[Integer]
def pointer(): Maybe[Integer] = Maybe.nothing[Integer]
def pointerSpace(): Maybe[String] = Maybe.nothing[String]
def sourceFile(): Maybe[File] = Maybe.nothing[File]
def sourcePath(): Maybe[String] = Maybe.nothing[String]
| <|file_sep|>original/org.scala-ide.sdt.core/src/org/scalaide/util/internal/SbtUtils.scala
object SbtUtils {
def m2o[S](opt: Maybe[S]): Option[S] = if (opt.isEmpty) None else Some(opt.get)
def readCache(cacheFile: File): Option[(Analysis, MiniSetup)] =
FileBasedStore(cacheFile).get().map(_ match {
case (a: Analysis, i) => (a, i)
case (a, _) => throw new RuntimeException(s"Expected that sbt analysis for $cacheFile is of type ${classOf[Analysis]} but was ${a.getClass}.")
})
def readAnalysis(cacheFile: File): Analysis =
// zinc requires name hashing to be enabled
readCache(cacheFile).map(_._1).getOrElse(Analysis.empty(nameHashing = true))
object NoPosition extends xsbti.Position {
def line(): Maybe[Integer] = Maybe.nothing()
def lineContent(): String = ""
def offset(): Maybe[Integer] = Maybe.nothing[Integer]
def pointer(): Maybe[Integer] = Maybe.nothing[Integer]
def pointerSpace(): Maybe[String] = Maybe.nothing[String]
def sourceFile(): Maybe[File] = Maybe.nothing[File]
def sourcePath(): Maybe[String] = Maybe.nothing[String]
<|file_sep|>current/org.scala-ide.sdt.core/src/org/scalaide/util/internal/SbtUtils.scala
object SbtUtils {
def m2o[S](opt: Maybe[S]): Option[S] = if (opt.isEmpty) None else Some(opt.get)
def readCache(cacheFile: File): Option[(Analysis, MiniSetup)] =
FileBasedStore(cacheFile).get().map(_ match {
case (a: Analysis, i) => (a, i)
case (a, _) => throw new RuntimeException(s"Expected that sbt analysis for $cacheFile is of type ${classOf[Analysis]} but was ${a.getClass}.")
})
def readAnalysis(cacheFile: File): Analysis =
// zinc requires name hashing to be enabled
readCache(cacheFile).map(_._1).getOrElse(Analysis.empty(nameHashing = true))
object NoPosition extends xsbti.Position {
def line(): Maybe[Integer] = Maybe.nothing()
def lineContent(): String = ""
def offset(): Maybe[Integer] = Maybe.nothing[Integer]
def pointer(): Maybe[Integer] = Maybe.nothing[Integer]
def pointerSpace(): Maybe[String] = Maybe.nothing[String]
def sourceFile(): Maybe[File] = Maybe.nothing[File]
def sourcePath(): Maybe[String] = Maybe.nothing[String]
<|file_sep|>updated/org.scala-ide.sdt.core/src/org/scalaide/util/internal/SbtUtils.scala
object SbtUtils {
def m2o[S](opt: Maybe[S]): Option[S] = if (opt.isEmpty) None else Some(opt.get)
def readCache(cacheFile: File): Option[(Analysis, MiniSetup)] =
FileBasedStore(cacheFile).get().map(_ match {
case (a: Analysis, i) => (a, i)
case (a, _) => throw new RuntimeException(s"Expected that sbt analysis for $cacheFile is of type ${classOf[Analysis]} but was ${a.getClass}.")
})
def readAnalysis(cacheFile: File): Analysis =
readCache(cacheFile).map(_._1).getOrElse(Analysis.empty)
object NoPosition extends xsbti.Position {
def line(): Maybe[Integer] = Maybe.nothing()
def lineContent(): String = ""
def offset(): Maybe[Integer] = Maybe.nothing[Integer]
def pointer(): Maybe[Integer] = Maybe.nothing[Integer]
def pointerSpace(): Maybe[String] = Maybe.nothing[String]
def sourceFile(): Maybe[File] = Maybe.nothing[File]
def sourcePath(): Maybe[String] = Maybe.nothing[String]
| a38dab1148a8c1b7c83b179d8453fc02e10251a0 | org.scala-ide.sdt.core/src/org/scalaide/util/internal/SbtUtils.scala | org.scala-ide.sdt.core/src/org/scalaide/util/internal/SbtUtils.scala | Scala |
<|file_sep|>rest-servlet/src/main/java/org/rhq/metrics/restServlet/MetricsServiceProducer.java.diff
original:
updated:
@SuppressWarnings("unused")
<|file_sep|>original/rest-servlet/src/main/java/org/rhq/metrics/restServlet/MetricsServiceProducer.java
@Inject
private javax.servlet.ServletContext context;
private MetricsService metricsService;
@Produces
public MetricsService getMetricsService() {
if (metricsService == null) {
try {
String className = context.getInitParameter("rhq-metrics.backend");
Class clazz = Class.forName(className);
metricsService = (MetricsService) clazz.newInstance();
// TODO passs servlet params
metricsService.startUp(Collections.<String, String>emptyMap());
ServiceKeeper.getInstance().service = metricsService;
return metricsService;
} catch (ClassNotFoundException | InstantiationException | IllegalAccessException e) {
throw new RuntimeException("Cannot create MetricsService class", e);
}
<|file_sep|>current/rest-servlet/src/main/java/org/rhq/metrics/restServlet/MetricsServiceProducer.java
@Inject
private javax.servlet.ServletContext context;
private MetricsService metricsService;
@Produces
public MetricsService getMetricsService() {
if (metricsService == null) {
try {
String className = context.getInitParameter("rhq-metrics.backend");
Class clazz = Class.forName(className);
metricsService = (MetricsService) clazz.newInstance();
// TODO passs servlet params
metricsService.startUp(Collections.<String, String>emptyMap());
ServiceKeeper.getInstance().service = metricsService;
return metricsService;
} catch (ClassNotFoundException | InstantiationException | IllegalAccessException e) {
throw new RuntimeException("Cannot create MetricsService class", e);
<|file_sep|>updated/rest-servlet/src/main/java/org/rhq/metrics/restServlet/MetricsServiceProducer.java |
@Inject
private javax.servlet.ServletContext context;
private MetricsService metricsService;
@Produces
public MetricsService getMetricsService() {
if (metricsService == null) {
try {
String className = null;
String backend = System.getProperty("rhq-metrics.backend");
if (backend!=null) {
switch (backend) {
case "mem":
className="org.rhq.metrics.impl.memory.MemoryMetricsService";
break;
case "cass":
className="org.rhq.metrics.impl.cassandra.MetricsServiceCassandra";
break;
} | <|file_sep|>rest-servlet/src/main/java/org/rhq/metrics/restServlet/MetricsServiceProducer.java.diff
original:
updated:
@SuppressWarnings("unused")
<|file_sep|>original/rest-servlet/src/main/java/org/rhq/metrics/restServlet/MetricsServiceProducer.java
@Inject
private javax.servlet.ServletContext context;
private MetricsService metricsService;
@Produces
public MetricsService getMetricsService() {
if (metricsService == null) {
try {
String className = context.getInitParameter("rhq-metrics.backend");
Class clazz = Class.forName(className);
metricsService = (MetricsService) clazz.newInstance();
// TODO passs servlet params
metricsService.startUp(Collections.<String, String>emptyMap());
ServiceKeeper.getInstance().service = metricsService;
return metricsService;
} catch (ClassNotFoundException | InstantiationException | IllegalAccessException e) {
throw new RuntimeException("Cannot create MetricsService class", e);
}
<|file_sep|>current/rest-servlet/src/main/java/org/rhq/metrics/restServlet/MetricsServiceProducer.java
@Inject
private javax.servlet.ServletContext context;
private MetricsService metricsService;
@Produces
public MetricsService getMetricsService() {
if (metricsService == null) {
try {
String className = context.getInitParameter("rhq-metrics.backend");
Class clazz = Class.forName(className);
metricsService = (MetricsService) clazz.newInstance();
// TODO passs servlet params
metricsService.startUp(Collections.<String, String>emptyMap());
ServiceKeeper.getInstance().service = metricsService;
return metricsService;
} catch (ClassNotFoundException | InstantiationException | IllegalAccessException e) {
throw new RuntimeException("Cannot create MetricsService class", e);
<|file_sep|>updated/rest-servlet/src/main/java/org/rhq/metrics/restServlet/MetricsServiceProducer.java
@Inject
private javax.servlet.ServletContext context;
private MetricsService metricsService;
@Produces
public MetricsService getMetricsService() {
if (metricsService == null) {
try {
String className = null;
String backend = System.getProperty("rhq-metrics.backend");
if (backend!=null) {
switch (backend) {
case "mem":
className="org.rhq.metrics.impl.memory.MemoryMetricsService";
break;
case "cass":
className="org.rhq.metrics.impl.cassandra.MetricsServiceCassandra";
break;
} | c1d931550ca3eb73fa6d91d6b38a5bd3e17f5d1e | rest-servlet/src/main/java/org/rhq/metrics/restServlet/MetricsServiceProducer.java | rest-servlet/src/main/java/org/rhq/metrics/restServlet/MetricsServiceProducer.java | Java |
<|file_sep|>original/static/css/site.css
pre {
background: #ffffcc;
padding-left: 5px;
}
div.body {
min-height: 600px;
}
div.header, div.footer {
color: #666666;
font-size: 9pt;
}
<|file_sep|>current/static/css/site.css
pre {
background: #ffffcc;
padding-left: 5px;
}
div.body {
min-height: 600px;
}
div.header, div.footer {
color: #666666;
font-size: 9pt;
}
<|file_sep|>updated/static/css/site.css | pre {
background: #ffeeee;
padding-left: 5px;
}
div.body {
min-height: 600px;
}
div.header, div.footer {
color: #666666;
font-size: 9pt;
} | <|file_sep|>original/static/css/site.css
pre {
background: #ffffcc;
padding-left: 5px;
}
div.body {
min-height: 600px;
}
div.header, div.footer {
color: #666666;
font-size: 9pt;
}
<|file_sep|>current/static/css/site.css
pre {
background: #ffffcc;
padding-left: 5px;
}
div.body {
min-height: 600px;
}
div.header, div.footer {
color: #666666;
font-size: 9pt;
}
<|file_sep|>updated/static/css/site.css
pre {
background: #ffeeee;
padding-left: 5px;
}
div.body {
min-height: 600px;
}
div.header, div.footer {
color: #666666;
font-size: 9pt;
} | 9a782f653c2552d352ddadc4754a3e618db2f67d | static/css/site.css | static/css/site.css | CSS |
<|file_sep|>original/spec/spec_helper.rb
# loaded once.
#
# See http://rubydoc.info/gems/rspec-core/RSpec/Core/Configuration
require 'rubygems'
require 'bundler/setup'
if ENV['COVERAGE']
require 'simplecov'
require 'simplecov-rcov'
SimpleCov.formatter = SimpleCov::Formatter::RcovFormatter
SimpleCov.start
end
require 'rails_core_extensions'
RSpec.configure do |config|
config.treat_symbols_as_metadata_keys_with_true_values = true
config.run_all_when_everything_filtered = true
config.filter_run :focus
end
<|file_sep|>current/spec/spec_helper.rb
# loaded once.
#
# See http://rubydoc.info/gems/rspec-core/RSpec/Core/Configuration
require 'rubygems'
require 'bundler/setup'
if ENV['COVERAGE']
require 'simplecov'
require 'simplecov-rcov'
SimpleCov.formatter = SimpleCov::Formatter::RcovFormatter
SimpleCov.start
end
require 'rails_core_extensions'
RSpec.configure do |config|
config.treat_symbols_as_metadata_keys_with_true_values = true
config.run_all_when_everything_filtered = true
config.filter_run :focus
end
<|file_sep|>updated/spec/spec_helper.rb | #
# See http://rubydoc.info/gems/rspec-core/RSpec/Core/Configuration
require 'rubygems'
require 'bundler/setup'
if ENV['COVERAGE']
require 'simplecov'
require 'simplecov-rcov'
SimpleCov.formatter = SimpleCov::Formatter::RcovFormatter
SimpleCov.start do
add_filter '/vendor/'
add_filter '/spec/'
add_group 'lib', 'lib'
end
end
require 'rails_core_extensions'
RSpec.configure do |config|
config.treat_symbols_as_metadata_keys_with_true_values = true | <|file_sep|>original/spec/spec_helper.rb
# loaded once.
#
# See http://rubydoc.info/gems/rspec-core/RSpec/Core/Configuration
require 'rubygems'
require 'bundler/setup'
if ENV['COVERAGE']
require 'simplecov'
require 'simplecov-rcov'
SimpleCov.formatter = SimpleCov::Formatter::RcovFormatter
SimpleCov.start
end
require 'rails_core_extensions'
RSpec.configure do |config|
config.treat_symbols_as_metadata_keys_with_true_values = true
config.run_all_when_everything_filtered = true
config.filter_run :focus
end
<|file_sep|>current/spec/spec_helper.rb
# loaded once.
#
# See http://rubydoc.info/gems/rspec-core/RSpec/Core/Configuration
require 'rubygems'
require 'bundler/setup'
if ENV['COVERAGE']
require 'simplecov'
require 'simplecov-rcov'
SimpleCov.formatter = SimpleCov::Formatter::RcovFormatter
SimpleCov.start
end
require 'rails_core_extensions'
RSpec.configure do |config|
config.treat_symbols_as_metadata_keys_with_true_values = true
config.run_all_when_everything_filtered = true
config.filter_run :focus
end
<|file_sep|>updated/spec/spec_helper.rb
#
# See http://rubydoc.info/gems/rspec-core/RSpec/Core/Configuration
require 'rubygems'
require 'bundler/setup'
if ENV['COVERAGE']
require 'simplecov'
require 'simplecov-rcov'
SimpleCov.formatter = SimpleCov::Formatter::RcovFormatter
SimpleCov.start do
add_filter '/vendor/'
add_filter '/spec/'
add_group 'lib', 'lib'
end
end
require 'rails_core_extensions'
RSpec.configure do |config|
config.treat_symbols_as_metadata_keys_with_true_values = true | 2c44cb49d7b4d1067c1118fdfb79d64c3464117c | spec/spec_helper.rb | spec/spec_helper.rb | Ruby |
<|file_sep|>original/README.md
# Apache Creadur Whisker
[](https://travis-ci.org/apache/creadur-whisker)
## What is Whisker?
Apache Whisker is developed by the Apache Creadur project of the Apache Software
Foundation. Join us at https://creadur.apache.org and read more about Apache Whisker
at https://creadur.apache.org/whisker.
## License
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements. See the NOTICE file distributed with
this work for additional information regarding copyright ownership.
The ASF licenses this file to You under the Apache License, Version 2.0
(the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
<|file_sep|>current/README.md
# Apache Creadur Whisker
[](https://travis-ci.org/apache/creadur-whisker)
## What is Whisker?
Apache Whisker is developed by the Apache Creadur project of the Apache Software
Foundation. Join us at https://creadur.apache.org and read more about Apache Whisker
at https://creadur.apache.org/whisker.
## License
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements. See the NOTICE file distributed with
this work for additional information regarding copyright ownership.
The ASF licenses this file to You under the Apache License, Version 2.0
(the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
<|file_sep|>updated/README.md | # Apache Creadur Whisker
[](https://ci-builds.apache.org/job/Creadur/job/Creadur-Whisker/)
[](https://travis-ci.org/apache/creadur-whisker)
## What is Whisker?
Apache Whisker is developed by the Apache Creadur project of the Apache Software
Foundation. Join us at https://creadur.apache.org and read more about Apache Whisker
at https://creadur.apache.org/whisker.
## License
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements. See the NOTICE file distributed with
this work for additional information regarding copyright ownership.
The ASF licenses this file to You under the Apache License, Version 2.0
(the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0 | <|file_sep|>original/README.md
# Apache Creadur Whisker
[](https://travis-ci.org/apache/creadur-whisker)
## What is Whisker?
Apache Whisker is developed by the Apache Creadur project of the Apache Software
Foundation. Join us at https://creadur.apache.org and read more about Apache Whisker
at https://creadur.apache.org/whisker.
## License
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements. See the NOTICE file distributed with
this work for additional information regarding copyright ownership.
The ASF licenses this file to You under the Apache License, Version 2.0
(the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
<|file_sep|>current/README.md
# Apache Creadur Whisker
[](https://travis-ci.org/apache/creadur-whisker)
## What is Whisker?
Apache Whisker is developed by the Apache Creadur project of the Apache Software
Foundation. Join us at https://creadur.apache.org and read more about Apache Whisker
at https://creadur.apache.org/whisker.
## License
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements. See the NOTICE file distributed with
this work for additional information regarding copyright ownership.
The ASF licenses this file to You under the Apache License, Version 2.0
(the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
<|file_sep|>updated/README.md
# Apache Creadur Whisker
[](https://ci-builds.apache.org/job/Creadur/job/Creadur-Whisker/)
[](https://travis-ci.org/apache/creadur-whisker)
## What is Whisker?
Apache Whisker is developed by the Apache Creadur project of the Apache Software
Foundation. Join us at https://creadur.apache.org and read more about Apache Whisker
at https://creadur.apache.org/whisker.
## License
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements. See the NOTICE file distributed with
this work for additional information regarding copyright ownership.
The ASF licenses this file to You under the Apache License, Version 2.0
(the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0 | a4302a0d46f9d98d9cdab81fa9a9e136c77b4e1a | README.md | README.md | Markdown |
<|file_sep|>plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt.diff
original:
import circlet.app.*
updated:
<|file_sep|>plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt.diff
original:
import klogging.*
updated:
<|file_sep|>plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt.diff
original:
@Suppress("unused")
private val log = KLoggers.logger("plugin/TestCircletAction.kt")
updated:
<|file_sep|>plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt.diff
original:
val loginModel = e.project?.connection?.loginModel
updated:
<|file_sep|>plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt.diff
original:
loginModel?.meSession?.value is MeSession.Connected &&
loginModel.client.connectionStatus.value == ConnectionStatus.CONNECTED
updated:
e.project?.connection?.loginModel?.clientOrNull?.connectionStatus?.value == ConnectionStatus.CONNECTED
<|file_sep|>original/plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt
import klogging.*
import kotlinx.coroutines.experimental.*
import runtime.*
@Suppress("unused")
private val log = KLoggers.logger("plugin/TestCircletAction.kt")
class TestCircletAction : AnAction() {
override fun update(e: AnActionEvent) {
val loginModel = e.project?.connection?.loginModel
e.presentation.isEnabledAndVisible =
loginModel?.meSession?.value is MeSession.Connected &&
loginModel.client.connectionStatus.value == ConnectionStatus.CONNECTED
}
override fun actionPerformed(e: AnActionEvent) {
async(UiDispatch.coroutineContext) {
val project = e.project!!
val result = project.connection.loginModel!!.client.me.info()
Notification(
<|file_sep|>current/plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt
import runtime.*
class TestCircletAction : AnAction() {
override fun update(e: AnActionEvent) {
e.presentation.isEnabledAndVisible =
e.project?.connection?.loginModel?.clientOrNull?.connectionStatus?.value == ConnectionStatus.CONNECTED
}
override fun actionPerformed(e: AnActionEvent) {
async(UiDispatch.coroutineContext) {
val project = e.project!!
val result = project.connection.loginModel!!.client.me.info()
Notification(
"Circlet",
"Circlet check",
"Me = $result",
NotificationType.INFORMATION
).notify(project)
}
}
}
<|file_sep|>updated/plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt |
class TestCircletAction : AnAction() {
override fun update(e: AnActionEvent) {
e.presentation.isEnabledAndVisible =
e.project?.connection?.loginModel?.clientOrNull?.connectionStatus?.value == ConnectionStatus.CONNECTED
}
override fun actionPerformed(e: AnActionEvent) {
async(UiDispatch.coroutineContext) {
val project = e.project!!
val result = project.connection.loginModel!!.client.me.info()
Notification(
"Circlet",
"Circlet check",
"Me = $result",
NotificationType.INFORMATION
).notify(project)
}
}
} | <|file_sep|>plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt.diff
original:
import circlet.app.*
updated:
<|file_sep|>plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt.diff
original:
import klogging.*
updated:
<|file_sep|>plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt.diff
original:
@Suppress("unused")
private val log = KLoggers.logger("plugin/TestCircletAction.kt")
updated:
<|file_sep|>plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt.diff
original:
val loginModel = e.project?.connection?.loginModel
updated:
<|file_sep|>plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt.diff
original:
loginModel?.meSession?.value is MeSession.Connected &&
loginModel.client.connectionStatus.value == ConnectionStatus.CONNECTED
updated:
e.project?.connection?.loginModel?.clientOrNull?.connectionStatus?.value == ConnectionStatus.CONNECTED
<|file_sep|>original/plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt
import klogging.*
import kotlinx.coroutines.experimental.*
import runtime.*
@Suppress("unused")
private val log = KLoggers.logger("plugin/TestCircletAction.kt")
class TestCircletAction : AnAction() {
override fun update(e: AnActionEvent) {
val loginModel = e.project?.connection?.loginModel
e.presentation.isEnabledAndVisible =
loginModel?.meSession?.value is MeSession.Connected &&
loginModel.client.connectionStatus.value == ConnectionStatus.CONNECTED
}
override fun actionPerformed(e: AnActionEvent) {
async(UiDispatch.coroutineContext) {
val project = e.project!!
val result = project.connection.loginModel!!.client.me.info()
Notification(
<|file_sep|>current/plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt
import runtime.*
class TestCircletAction : AnAction() {
override fun update(e: AnActionEvent) {
e.presentation.isEnabledAndVisible =
e.project?.connection?.loginModel?.clientOrNull?.connectionStatus?.value == ConnectionStatus.CONNECTED
}
override fun actionPerformed(e: AnActionEvent) {
async(UiDispatch.coroutineContext) {
val project = e.project!!
val result = project.connection.loginModel!!.client.me.info()
Notification(
"Circlet",
"Circlet check",
"Me = $result",
NotificationType.INFORMATION
).notify(project)
}
}
}
<|file_sep|>updated/plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt
class TestCircletAction : AnAction() {
override fun update(e: AnActionEvent) {
e.presentation.isEnabledAndVisible =
e.project?.connection?.loginModel?.clientOrNull?.connectionStatus?.value == ConnectionStatus.CONNECTED
}
override fun actionPerformed(e: AnActionEvent) {
async(UiDispatch.coroutineContext) {
val project = e.project!!
val result = project.connection.loginModel!!.client.me.info()
Notification(
"Circlet",
"Circlet check",
"Me = $result",
NotificationType.INFORMATION
).notify(project)
}
}
} | 471609f3796ef5e079e8de82de9a595a0716f95e | plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt | plugins/space/src/main/kotlin/circlet/actions/TestCircletAction.kt | Kotlin |
<|file_sep|>original/src/ssi/google-analytics.html
<script>
(function (document, trackingId) {
if (navigator.userAgent.indexOf('Google Page Speed Insights') === -1) {
var gtagElement = document.createElement('script');
gtagElement.src = 'https://www.googletagmanager.com/gtagElementag/js?id=' + trackingId;
gtagElement.async = true;
var firstScriptElement = document.scripts[0];
firstScriptElement.parentNode.insertBefore(gtagElement, firstScriptElement);
window.dataLayer = window.dataLayer || [];
function gtag() {
dataLayer.push(arguments);
}
gtag('js', new Date());
gtag('config', trackingId);
}
<|file_sep|>current/src/ssi/google-analytics.html
<script>
(function (document, trackingId) {
if (navigator.userAgent.indexOf('Google Page Speed Insights') === -1) {
var gtagElement = document.createElement('script');
gtagElement.src = 'https://www.googletagmanager.com/gtagElementag/js?id=' + trackingId;
gtagElement.async = true;
var firstScriptElement = document.scripts[0];
firstScriptElement.parentNode.insertBefore(gtagElement, firstScriptElement);
window.dataLayer = window.dataLayer || [];
function gtag() {
dataLayer.push(arguments);
}
gtag('js', new Date());
gtag('config', trackingId);
}
<|file_sep|>updated/src/ssi/google-analytics.html | <script>
(function (document, trackingId) {
if (navigator.userAgent.indexOf('Google Page Speed Insights') === -1) {
var gtagElement = document.createElement('script');
gtagElement.src = 'https://www.googletagmanager.com/gtag/js?id=' + trackingId;
gtagElement.async = true;
var firstScriptElement = document.scripts[0];
firstScriptElement.parentNode.insertBefore(gtagElement, firstScriptElement);
window.dataLayer = window.dataLayer || [];
function gtag() {
dataLayer.push(arguments);
}
gtag('js', new Date());
gtag('config', trackingId);
} | <|file_sep|>original/src/ssi/google-analytics.html
<script>
(function (document, trackingId) {
if (navigator.userAgent.indexOf('Google Page Speed Insights') === -1) {
var gtagElement = document.createElement('script');
gtagElement.src = 'https://www.googletagmanager.com/gtagElementag/js?id=' + trackingId;
gtagElement.async = true;
var firstScriptElement = document.scripts[0];
firstScriptElement.parentNode.insertBefore(gtagElement, firstScriptElement);
window.dataLayer = window.dataLayer || [];
function gtag() {
dataLayer.push(arguments);
}
gtag('js', new Date());
gtag('config', trackingId);
}
<|file_sep|>current/src/ssi/google-analytics.html
<script>
(function (document, trackingId) {
if (navigator.userAgent.indexOf('Google Page Speed Insights') === -1) {
var gtagElement = document.createElement('script');
gtagElement.src = 'https://www.googletagmanager.com/gtagElementag/js?id=' + trackingId;
gtagElement.async = true;
var firstScriptElement = document.scripts[0];
firstScriptElement.parentNode.insertBefore(gtagElement, firstScriptElement);
window.dataLayer = window.dataLayer || [];
function gtag() {
dataLayer.push(arguments);
}
gtag('js', new Date());
gtag('config', trackingId);
}
<|file_sep|>updated/src/ssi/google-analytics.html
<script>
(function (document, trackingId) {
if (navigator.userAgent.indexOf('Google Page Speed Insights') === -1) {
var gtagElement = document.createElement('script');
gtagElement.src = 'https://www.googletagmanager.com/gtag/js?id=' + trackingId;
gtagElement.async = true;
var firstScriptElement = document.scripts[0];
firstScriptElement.parentNode.insertBefore(gtagElement, firstScriptElement);
window.dataLayer = window.dataLayer || [];
function gtag() {
dataLayer.push(arguments);
}
gtag('js', new Date());
gtag('config', trackingId);
} | a81e250a20e5c71281ee3611a44d1f8b3df31b4c | src/ssi/google-analytics.html | src/ssi/google-analytics.html | HTML |
<|file_sep|>original/lib/fog/vcloud_director/requests/compute/get_network_complete.rb
# @raise [Fog::Compute::VcloudDirector::Forbidden]
#
# @see http://pubs.vmware.com/vcd-51/topic/com.vmware.vcloud.api.reference.doc_51/doc/operations/GET-Network.html
# @since vCloud API version 0.9
def get_network_complete(id)
request(
:expects => 200,
:idempotent => true,
:method => 'GET',
:parser => Fog::ToHashDocument.new,
:path => "network/#{id}"
)
end
end
class Mock
def get_network_complete(id)
unless network = data[:networks][id]
raise Fog::Compute::VcloudDirector::Forbidden.new(
'This operation is denied.'
)
<|file_sep|>current/lib/fog/vcloud_director/requests/compute/get_network_complete.rb
# @raise [Fog::Compute::VcloudDirector::Forbidden]
#
# @see http://pubs.vmware.com/vcd-51/topic/com.vmware.vcloud.api.reference.doc_51/doc/operations/GET-Network.html
# @since vCloud API version 0.9
def get_network_complete(id)
request(
:expects => 200,
:idempotent => true,
:method => 'GET',
:parser => Fog::ToHashDocument.new,
:path => "network/#{id}"
)
end
end
class Mock
def get_network_complete(id)
unless network = data[:networks][id]
raise Fog::Compute::VcloudDirector::Forbidden.new(
'This operation is denied.'
)
<|file_sep|>updated/lib/fog/vcloud_director/requests/compute/get_network_complete.rb | # @raise [Fog::Compute::VcloudDirector::Forbidden]
#
# @see http://pubs.vmware.com/vcd-51/topic/com.vmware.vcloud.api.reference.doc_51/doc/operations/GET-Network.html
# @since vCloud API version 0.9
def get_network_complete(id)
request(
:expects => 200,
:idempotent => true,
:method => 'GET',
:parser => Fog::ToHashDocument.new,
:path => "admin/network/#{id}"
)
end
end
class Mock
def get_network_complete(id)
unless network = data[:networks][id]
raise Fog::Compute::VcloudDirector::Forbidden.new(
'This operation is denied.'
) | <|file_sep|>original/lib/fog/vcloud_director/requests/compute/get_network_complete.rb
# @raise [Fog::Compute::VcloudDirector::Forbidden]
#
# @see http://pubs.vmware.com/vcd-51/topic/com.vmware.vcloud.api.reference.doc_51/doc/operations/GET-Network.html
# @since vCloud API version 0.9
def get_network_complete(id)
request(
:expects => 200,
:idempotent => true,
:method => 'GET',
:parser => Fog::ToHashDocument.new,
:path => "network/#{id}"
)
end
end
class Mock
def get_network_complete(id)
unless network = data[:networks][id]
raise Fog::Compute::VcloudDirector::Forbidden.new(
'This operation is denied.'
)
<|file_sep|>current/lib/fog/vcloud_director/requests/compute/get_network_complete.rb
# @raise [Fog::Compute::VcloudDirector::Forbidden]
#
# @see http://pubs.vmware.com/vcd-51/topic/com.vmware.vcloud.api.reference.doc_51/doc/operations/GET-Network.html
# @since vCloud API version 0.9
def get_network_complete(id)
request(
:expects => 200,
:idempotent => true,
:method => 'GET',
:parser => Fog::ToHashDocument.new,
:path => "network/#{id}"
)
end
end
class Mock
def get_network_complete(id)
unless network = data[:networks][id]
raise Fog::Compute::VcloudDirector::Forbidden.new(
'This operation is denied.'
)
<|file_sep|>updated/lib/fog/vcloud_director/requests/compute/get_network_complete.rb
# @raise [Fog::Compute::VcloudDirector::Forbidden]
#
# @see http://pubs.vmware.com/vcd-51/topic/com.vmware.vcloud.api.reference.doc_51/doc/operations/GET-Network.html
# @since vCloud API version 0.9
def get_network_complete(id)
request(
:expects => 200,
:idempotent => true,
:method => 'GET',
:parser => Fog::ToHashDocument.new,
:path => "admin/network/#{id}"
)
end
end
class Mock
def get_network_complete(id)
unless network = data[:networks][id]
raise Fog::Compute::VcloudDirector::Forbidden.new(
'This operation is denied.'
) | 27bd0edb26450898c56acc2f7e3b4e0e13121a6b | lib/fog/vcloud_director/requests/compute/get_network_complete.rb | lib/fog/vcloud_director/requests/compute/get_network_complete.rb | Ruby |
<|file_sep|>tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py.diff
original:
updated:
from chainer.training import trigger
<|file_sep|>tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py.diff
original:
updated:
{'trigger': trigger.IntervalTrigger(5, 'epoch')},
{'trigger': trigger.IntervalTrigger(20, 'iteration')},
<|file_sep|>original/tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py
class TestSnapshotObject(unittest.TestCase):
def test_trigger(self):
target = mock.MagicMock()
snapshot_object = extensions.snapshot_object(target, 'myfile.dat',
trigger=self.trigger)
self.assertEqual(snapshot_object.trigger, self.trigger)
@testing.parameterize(
{'trigger': ('epoch', 2)},
{'trigger': ('iteration', 10)},
)
class TestSnapshot(unittest.TestCase):
def test_trigger(self):
snapshot = extensions.snapshot(trigger=self.trigger)
self.assertEqual(snapshot.trigger, self.trigger)
testing.run_module(__name__, __file__)
<|file_sep|>current/tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py
class TestSnapshotObject(unittest.TestCase):
def test_trigger(self):
target = mock.MagicMock()
snapshot_object = extensions.snapshot_object(target, 'myfile.dat',
trigger=self.trigger)
self.assertEqual(snapshot_object.trigger, self.trigger)
@testing.parameterize(
{'trigger': ('epoch', 2)},
{'trigger': ('iteration', 10)},
)
class TestSnapshot(unittest.TestCase):
def test_trigger(self):
snapshot = extensions.snapshot(trigger=self.trigger)
self.assertEqual(snapshot.trigger, self.trigger)
testing.run_module(__name__, __file__)
<|file_sep|>updated/tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py | def test_trigger(self):
target = mock.MagicMock()
snapshot_object = extensions.snapshot_object(target, 'myfile.dat',
trigger=self.trigger)
self.assertEqual(snapshot_object.trigger, self.trigger)
@testing.parameterize(
{'trigger': ('epoch', 2)},
{'trigger': ('iteration', 10)},
{'trigger': trigger.IntervalTrigger(5, 'epoch')},
{'trigger': trigger.IntervalTrigger(20, 'iteration')},
)
class TestSnapshot(unittest.TestCase):
def test_trigger(self):
snapshot = extensions.snapshot(trigger=self.trigger)
self.assertEqual(snapshot.trigger, self.trigger)
testing.run_module(__name__, __file__) | <|file_sep|>tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py.diff
original:
updated:
from chainer.training import trigger
<|file_sep|>tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py.diff
original:
updated:
{'trigger': trigger.IntervalTrigger(5, 'epoch')},
{'trigger': trigger.IntervalTrigger(20, 'iteration')},
<|file_sep|>original/tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py
class TestSnapshotObject(unittest.TestCase):
def test_trigger(self):
target = mock.MagicMock()
snapshot_object = extensions.snapshot_object(target, 'myfile.dat',
trigger=self.trigger)
self.assertEqual(snapshot_object.trigger, self.trigger)
@testing.parameterize(
{'trigger': ('epoch', 2)},
{'trigger': ('iteration', 10)},
)
class TestSnapshot(unittest.TestCase):
def test_trigger(self):
snapshot = extensions.snapshot(trigger=self.trigger)
self.assertEqual(snapshot.trigger, self.trigger)
testing.run_module(__name__, __file__)
<|file_sep|>current/tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py
class TestSnapshotObject(unittest.TestCase):
def test_trigger(self):
target = mock.MagicMock()
snapshot_object = extensions.snapshot_object(target, 'myfile.dat',
trigger=self.trigger)
self.assertEqual(snapshot_object.trigger, self.trigger)
@testing.parameterize(
{'trigger': ('epoch', 2)},
{'trigger': ('iteration', 10)},
)
class TestSnapshot(unittest.TestCase):
def test_trigger(self):
snapshot = extensions.snapshot(trigger=self.trigger)
self.assertEqual(snapshot.trigger, self.trigger)
testing.run_module(__name__, __file__)
<|file_sep|>updated/tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py
def test_trigger(self):
target = mock.MagicMock()
snapshot_object = extensions.snapshot_object(target, 'myfile.dat',
trigger=self.trigger)
self.assertEqual(snapshot_object.trigger, self.trigger)
@testing.parameterize(
{'trigger': ('epoch', 2)},
{'trigger': ('iteration', 10)},
{'trigger': trigger.IntervalTrigger(5, 'epoch')},
{'trigger': trigger.IntervalTrigger(20, 'iteration')},
)
class TestSnapshot(unittest.TestCase):
def test_trigger(self):
snapshot = extensions.snapshot(trigger=self.trigger)
self.assertEqual(snapshot.trigger, self.trigger)
testing.run_module(__name__, __file__) | 8360bebbd4bf2b2e9d51c7aa16bdb9506a91883e | tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py | tests/chainer_tests/training_tests/extensions_tests/test_snapshot.py | Python |
<|file_sep|>original/scripts/explore_params.sh
<|file_sep|>current/scripts/explore_params.sh
<|file_sep|>updated/scripts/explore_params.sh | rosparam set /hector_exploration_node/hector_exploration_planner/dist_for_goal_reached 0.50
#rosparam set /hector_exploration_node/hector_exploration_planner/goal_angle_penalty
#rosparam set /hector_exploration_node/hector_exploration_planner/min_frontier_size
rosparam set /hector_exploration_node/hector_exploration_planner/min_obstacle_dist 20
rosparam set /hector_exploration_node/hector_exploration_planner/obstacle_cutoff_distance 2.0
#rosparam set /hector_exploration_node/hector_exploration_planner/plan_in_unknown
#rosparam set /hector_exploration_node/hector_exploration_planner/same_frontier_distance
#rosparam set /hector_exploration_node/hector_exploration_planner/security_constant
#rosparam set /hector_exploration_node/hector_exploration_planner/use_inflated_obstacles
echo -n "dist for goal reached: "
rosparam get /hector_exploration_node/hector_exploration_planner/dist_for_goal_reached
echo -n "goal_angle_penalty: "
rosparam get /hector_exploration_node/hector_exploration_planner/goal_angle_penalty
echo -n "min_frontier_size: "
rosparam get /hector_exploration_node/hector_exploration_planner/min_frontier_size
echo -n "min_obstacle_dist: "
rosparam get /hector_exploration_node/hector_exploration_planner/min_obstacle_dist
echo -n "obstacle_cutoff_distance: "
rosparam get /hector_exploration_node/hector_exploration_planner/obstacle_cutoff_distance
echo -n "plan_in_unknown: "
rosparam get /hector_exploration_node/hector_exploration_planner/plan_in_unknown | <|file_sep|>original/scripts/explore_params.sh
<|file_sep|>current/scripts/explore_params.sh
<|file_sep|>updated/scripts/explore_params.sh
rosparam set /hector_exploration_node/hector_exploration_planner/dist_for_goal_reached 0.50
#rosparam set /hector_exploration_node/hector_exploration_planner/goal_angle_penalty
#rosparam set /hector_exploration_node/hector_exploration_planner/min_frontier_size
rosparam set /hector_exploration_node/hector_exploration_planner/min_obstacle_dist 20
rosparam set /hector_exploration_node/hector_exploration_planner/obstacle_cutoff_distance 2.0
#rosparam set /hector_exploration_node/hector_exploration_planner/plan_in_unknown
#rosparam set /hector_exploration_node/hector_exploration_planner/same_frontier_distance
#rosparam set /hector_exploration_node/hector_exploration_planner/security_constant
#rosparam set /hector_exploration_node/hector_exploration_planner/use_inflated_obstacles
echo -n "dist for goal reached: "
rosparam get /hector_exploration_node/hector_exploration_planner/dist_for_goal_reached
echo -n "goal_angle_penalty: "
rosparam get /hector_exploration_node/hector_exploration_planner/goal_angle_penalty
echo -n "min_frontier_size: "
rosparam get /hector_exploration_node/hector_exploration_planner/min_frontier_size
echo -n "min_obstacle_dist: "
rosparam get /hector_exploration_node/hector_exploration_planner/min_obstacle_dist
echo -n "obstacle_cutoff_distance: "
rosparam get /hector_exploration_node/hector_exploration_planner/obstacle_cutoff_distance
echo -n "plan_in_unknown: "
rosparam get /hector_exploration_node/hector_exploration_planner/plan_in_unknown | 14451c6a7f31bd4d9f60a1df42a2e49285805bdf | scripts/explore_params.sh | scripts/explore_params.sh | Shell |
<|file_sep|>.travis.yml.diff
original:
sudo: false
updated:
sudo: required
<|file_sep|>.travis.yml.diff
original:
updated:
- docker
<|file_sep|>.travis.yml.diff
original:
- DB_USER=test DB_PASSWORD=test DB_DATABASE=travis_ci_test DB_TEST_DATABASE=travis_ci_test DB_HOST=localhost DB_PORT=5432 TOKEN_SECRET=travis SERVER_PORT=3000
updated:
- DB_USER=test DB_PASSWORD=test DB_DATABASE=travis_ci_test DB_TEST_DATABASE=travis_ci_test DB_HOST=localhost DB_PORT=5432 TOKEN_SECRET=travis SERVER_PORT=3000 DOCKER_COMPOSE_VERSION=1.16.1
before_install:
- sudo rm /usr/local/bin/docker-compose
- curl -L https://github.com/docker/compose/releases/download/${DOCKER_COMPOSE_VERSION}/docker-compose-`uname -s`-`uname -m` > docker-compose
- chmod +x docker-compose
- sudo mv docker-compose /usr/local/bin
<|file_sep|>.travis.yml.diff
original:
updated:
# Database for non-Docker execution
<|file_sep|>original/.travis.yml
dist: trusty
sudo: false
language: node_js
node_js:
- "7"
services:
- postgresql
addons:
postgresql: "9.6"
env:
- DB_USER=test DB_PASSWORD=test DB_DATABASE=travis_ci_test DB_TEST_DATABASE=travis_ci_test DB_HOST=localhost DB_PORT=5432 TOKEN_SECRET=travis SERVER_PORT=3000
before_script:
- psql -c 'CREATE DATABASE travis_ci_test' -U postgres
- psql -c "CREATE USER test WITH PASSWORD 'test';" -U postgres
<|file_sep|>current/.travis.yml
- "7"
services:
- postgresql
- docker
addons:
postgresql: "9.6"
env:
- DB_USER=test DB_PASSWORD=test DB_DATABASE=travis_ci_test DB_TEST_DATABASE=travis_ci_test DB_HOST=localhost DB_PORT=5432 TOKEN_SECRET=travis SERVER_PORT=3000 DOCKER_COMPOSE_VERSION=1.16.1
before_install:
- sudo rm /usr/local/bin/docker-compose
- curl -L https://github.com/docker/compose/releases/download/${DOCKER_COMPOSE_VERSION}/docker-compose-`uname -s`-`uname -m` > docker-compose
- chmod +x docker-compose
- sudo mv docker-compose /usr/local/bin
before_script:
# Database for non-Docker execution
- psql -c 'CREATE DATABASE travis_ci_test' -U postgres
- psql -c "CREATE USER test WITH PASSWORD 'test';" -U postgres
<|file_sep|>updated/.travis.yml | env:
- DB_USER=test DB_PASSWORD=test DB_DATABASE=travis_ci_test DB_TEST_DATABASE=travis_ci_test DB_HOST=localhost DB_PORT=5432 TOKEN_SECRET=travis SERVER_PORT=3000 DOCKER_COMPOSE_VERSION=1.16.1
before_install:
- sudo rm /usr/local/bin/docker-compose
- curl -L https://github.com/docker/compose/releases/download/${DOCKER_COMPOSE_VERSION}/docker-compose-`uname -s`-`uname -m` > docker-compose
- chmod +x docker-compose
- sudo mv docker-compose /usr/local/bin
before_script:
# Database for non-Docker execution
- psql -c 'CREATE DATABASE travis_ci_test' -U postgres
- psql -c "CREATE USER test WITH PASSWORD 'test';" -U postgres
# Database for Docker execution
- docker-compose run --rm web scripts/init.sh NONE
script:
# Test non-Docker
- npm run test
# Test in Docker
- docker-compose run --service-ports --rm web npm run test | <|file_sep|>.travis.yml.diff
original:
sudo: false
updated:
sudo: required
<|file_sep|>.travis.yml.diff
original:
updated:
- docker
<|file_sep|>.travis.yml.diff
original:
- DB_USER=test DB_PASSWORD=test DB_DATABASE=travis_ci_test DB_TEST_DATABASE=travis_ci_test DB_HOST=localhost DB_PORT=5432 TOKEN_SECRET=travis SERVER_PORT=3000
updated:
- DB_USER=test DB_PASSWORD=test DB_DATABASE=travis_ci_test DB_TEST_DATABASE=travis_ci_test DB_HOST=localhost DB_PORT=5432 TOKEN_SECRET=travis SERVER_PORT=3000 DOCKER_COMPOSE_VERSION=1.16.1
before_install:
- sudo rm /usr/local/bin/docker-compose
- curl -L https://github.com/docker/compose/releases/download/${DOCKER_COMPOSE_VERSION}/docker-compose-`uname -s`-`uname -m` > docker-compose
- chmod +x docker-compose
- sudo mv docker-compose /usr/local/bin
<|file_sep|>.travis.yml.diff
original:
updated:
# Database for non-Docker execution
<|file_sep|>original/.travis.yml
dist: trusty
sudo: false
language: node_js
node_js:
- "7"
services:
- postgresql
addons:
postgresql: "9.6"
env:
- DB_USER=test DB_PASSWORD=test DB_DATABASE=travis_ci_test DB_TEST_DATABASE=travis_ci_test DB_HOST=localhost DB_PORT=5432 TOKEN_SECRET=travis SERVER_PORT=3000
before_script:
- psql -c 'CREATE DATABASE travis_ci_test' -U postgres
- psql -c "CREATE USER test WITH PASSWORD 'test';" -U postgres
<|file_sep|>current/.travis.yml
- "7"
services:
- postgresql
- docker
addons:
postgresql: "9.6"
env:
- DB_USER=test DB_PASSWORD=test DB_DATABASE=travis_ci_test DB_TEST_DATABASE=travis_ci_test DB_HOST=localhost DB_PORT=5432 TOKEN_SECRET=travis SERVER_PORT=3000 DOCKER_COMPOSE_VERSION=1.16.1
before_install:
- sudo rm /usr/local/bin/docker-compose
- curl -L https://github.com/docker/compose/releases/download/${DOCKER_COMPOSE_VERSION}/docker-compose-`uname -s`-`uname -m` > docker-compose
- chmod +x docker-compose
- sudo mv docker-compose /usr/local/bin
before_script:
# Database for non-Docker execution
- psql -c 'CREATE DATABASE travis_ci_test' -U postgres
- psql -c "CREATE USER test WITH PASSWORD 'test';" -U postgres
<|file_sep|>updated/.travis.yml
env:
- DB_USER=test DB_PASSWORD=test DB_DATABASE=travis_ci_test DB_TEST_DATABASE=travis_ci_test DB_HOST=localhost DB_PORT=5432 TOKEN_SECRET=travis SERVER_PORT=3000 DOCKER_COMPOSE_VERSION=1.16.1
before_install:
- sudo rm /usr/local/bin/docker-compose
- curl -L https://github.com/docker/compose/releases/download/${DOCKER_COMPOSE_VERSION}/docker-compose-`uname -s`-`uname -m` > docker-compose
- chmod +x docker-compose
- sudo mv docker-compose /usr/local/bin
before_script:
# Database for non-Docker execution
- psql -c 'CREATE DATABASE travis_ci_test' -U postgres
- psql -c "CREATE USER test WITH PASSWORD 'test';" -U postgres
# Database for Docker execution
- docker-compose run --rm web scripts/init.sh NONE
script:
# Test non-Docker
- npm run test
# Test in Docker
- docker-compose run --service-ports --rm web npm run test | ae63aef25b681e423b075f9d3328b3dcbe29b2f7 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/res/values/gtfs_rts_values_gen.xml
<?xml version="1.0" encoding="utf-8"?>
<resources xmlns:tools="http://schemas.android.com/tools" tools:ignore="MissingTranslation">
<integer name="gtfs_rts_agency_type">3</integer>
<bool name="gtfs_rts_schedule_available">true</bool>
<bool name="gtfs_rts_frequency_available">false</bool>
<string name="gtfs_rts_timezone">America/Montreal</string>
<string name="gtfs_rts_area_min_lat">45.4581913018884</string>
<string name="gtfs_rts_area_max_lat">45.4984402900159</string>
<string name="gtfs_rts_area_min_lng">-73.5668389536944</string>
<string name="gtfs_rts_area_max_lng">-73.4418689636485</string>
<string name="gtfs_rts_color">20558A</string>
</resources>
<|file_sep|>current/res/values/gtfs_rts_values_gen.xml
<?xml version="1.0" encoding="utf-8"?>
<resources xmlns:tools="http://schemas.android.com/tools" tools:ignore="MissingTranslation">
<integer name="gtfs_rts_agency_type">3</integer>
<bool name="gtfs_rts_schedule_available">true</bool>
<bool name="gtfs_rts_frequency_available">false</bool>
<string name="gtfs_rts_timezone">America/Montreal</string>
<string name="gtfs_rts_area_min_lat">45.4581913018884</string>
<string name="gtfs_rts_area_max_lat">45.4984402900159</string>
<string name="gtfs_rts_area_min_lng">-73.5668389536944</string>
<string name="gtfs_rts_area_max_lng">-73.4418689636485</string>
<string name="gtfs_rts_color">20558A</string>
</resources>
<|file_sep|>updated/res/values/gtfs_rts_values_gen.xml | <?xml version="1.0" encoding="utf-8"?>
<resources xmlns:tools="http://schemas.android.com/tools" tools:ignore="MissingTranslation">
<integer name="gtfs_rts_agency_type">3</integer>
<bool name="gtfs_rts_schedule_available">true</bool>
<bool name="gtfs_rts_frequency_available">false</bool>
<string name="gtfs_rts_timezone">America/Montreal</string>
<string name="gtfs_rts_area_min_lat">45.4581913018884</string>
<string name="gtfs_rts_area_max_lat">45.4984402900159</string>
<string name="gtfs_rts_area_min_lng">-73.5668389536944</string>
<string name="gtfs_rts_area_max_lng">-73.4418689636485</string>
<string name="gtfs_rts_color">20558A</string>
<integer name="gtfs_rts_first_departure_in_sec">1529924520</integer>
<integer name="gtfs_rts_last_departure_in_sec">1534736820</integer>
</resources> | <|file_sep|>original/res/values/gtfs_rts_values_gen.xml
<?xml version="1.0" encoding="utf-8"?>
<resources xmlns:tools="http://schemas.android.com/tools" tools:ignore="MissingTranslation">
<integer name="gtfs_rts_agency_type">3</integer>
<bool name="gtfs_rts_schedule_available">true</bool>
<bool name="gtfs_rts_frequency_available">false</bool>
<string name="gtfs_rts_timezone">America/Montreal</string>
<string name="gtfs_rts_area_min_lat">45.4581913018884</string>
<string name="gtfs_rts_area_max_lat">45.4984402900159</string>
<string name="gtfs_rts_area_min_lng">-73.5668389536944</string>
<string name="gtfs_rts_area_max_lng">-73.4418689636485</string>
<string name="gtfs_rts_color">20558A</string>
</resources>
<|file_sep|>current/res/values/gtfs_rts_values_gen.xml
<?xml version="1.0" encoding="utf-8"?>
<resources xmlns:tools="http://schemas.android.com/tools" tools:ignore="MissingTranslation">
<integer name="gtfs_rts_agency_type">3</integer>
<bool name="gtfs_rts_schedule_available">true</bool>
<bool name="gtfs_rts_frequency_available">false</bool>
<string name="gtfs_rts_timezone">America/Montreal</string>
<string name="gtfs_rts_area_min_lat">45.4581913018884</string>
<string name="gtfs_rts_area_max_lat">45.4984402900159</string>
<string name="gtfs_rts_area_min_lng">-73.5668389536944</string>
<string name="gtfs_rts_area_max_lng">-73.4418689636485</string>
<string name="gtfs_rts_color">20558A</string>
</resources>
<|file_sep|>updated/res/values/gtfs_rts_values_gen.xml
<?xml version="1.0" encoding="utf-8"?>
<resources xmlns:tools="http://schemas.android.com/tools" tools:ignore="MissingTranslation">
<integer name="gtfs_rts_agency_type">3</integer>
<bool name="gtfs_rts_schedule_available">true</bool>
<bool name="gtfs_rts_frequency_available">false</bool>
<string name="gtfs_rts_timezone">America/Montreal</string>
<string name="gtfs_rts_area_min_lat">45.4581913018884</string>
<string name="gtfs_rts_area_max_lat">45.4984402900159</string>
<string name="gtfs_rts_area_min_lng">-73.5668389536944</string>
<string name="gtfs_rts_area_max_lng">-73.4418689636485</string>
<string name="gtfs_rts_color">20558A</string>
<integer name="gtfs_rts_first_departure_in_sec">1529924520</integer>
<integer name="gtfs_rts_last_departure_in_sec">1534736820</integer>
</resources> | 6ab2b7d7792efc2c8a2c6fa77d1e82b5d470032c | res/values/gtfs_rts_values_gen.xml | res/values/gtfs_rts_values_gen.xml | XML |
<|file_sep|>original/requirements/dev.txt
-r common.txt
pip==10.0.1
bumpversion==0.5.3
wheel==0.31.1
watchdog==0.8.3
flake8==3.5.0
coverage==4.5.1
Sphinx==1.7.6
cryptography==2.2.2
<|file_sep|>current/requirements/dev.txt
-r common.txt
pip==10.0.1
bumpversion==0.5.3
wheel==0.31.1
watchdog==0.8.3
flake8==3.5.0
coverage==4.5.1
Sphinx==1.7.6
cryptography==2.2.2
<|file_sep|>updated/requirements/dev.txt | -r common.txt
pip==10.0.1
bumpversion==0.5.3
wheel==0.31.1
watchdog==0.8.3
flake8==3.5.0
coverage==4.5.1
Sphinx==1.7.6
cryptography==2.3 | <|file_sep|>original/requirements/dev.txt
-r common.txt
pip==10.0.1
bumpversion==0.5.3
wheel==0.31.1
watchdog==0.8.3
flake8==3.5.0
coverage==4.5.1
Sphinx==1.7.6
cryptography==2.2.2
<|file_sep|>current/requirements/dev.txt
-r common.txt
pip==10.0.1
bumpversion==0.5.3
wheel==0.31.1
watchdog==0.8.3
flake8==3.5.0
coverage==4.5.1
Sphinx==1.7.6
cryptography==2.2.2
<|file_sep|>updated/requirements/dev.txt
-r common.txt
pip==10.0.1
bumpversion==0.5.3
wheel==0.31.1
watchdog==0.8.3
flake8==3.5.0
coverage==4.5.1
Sphinx==1.7.6
cryptography==2.3 | c91c63c57623ee2b7aefdf2c372f7205b0cb1a91 | requirements/dev.txt | requirements/dev.txt | Text |
<|file_sep|>original/lib/sync_checker/formats/document_collection_check.rb
class DocumentCollectionCheck < EditionBase
def root_path
"/government/collections/"
end
def rendering_app
Whitehall::RenderingApp::WHITEHALL_FRONTEND
end
def checks_for_live(locale)
super
end
def expected_details_hash(edition)
super.merge(
collection_groups: collection_groups(edition)
)
end
private
<|file_sep|>current/lib/sync_checker/formats/document_collection_check.rb
class DocumentCollectionCheck < EditionBase
def root_path
"/government/collections/"
end
def rendering_app
Whitehall::RenderingApp::WHITEHALL_FRONTEND
end
def checks_for_live(locale)
super
end
def expected_details_hash(edition)
super.merge(
collection_groups: collection_groups(edition)
)
end
private
<|file_sep|>updated/lib/sync_checker/formats/document_collection_check.rb | class DocumentCollectionCheck < EditionBase
def root_path
"/government/collections/"
end
def rendering_app
Whitehall::RenderingApp::WHITEHALL_FRONTEND
end
def checks_for_live(locale)
super << Checks::LinksCheck.new(
"topical_events",
TopicalEvent
.for_edition(edition_expected_in_live.id)
.pluck(:content_id)
)
end
def expected_details_hash(edition)
super.merge(
collection_groups: collection_groups(edition) | <|file_sep|>original/lib/sync_checker/formats/document_collection_check.rb
class DocumentCollectionCheck < EditionBase
def root_path
"/government/collections/"
end
def rendering_app
Whitehall::RenderingApp::WHITEHALL_FRONTEND
end
def checks_for_live(locale)
super
end
def expected_details_hash(edition)
super.merge(
collection_groups: collection_groups(edition)
)
end
private
<|file_sep|>current/lib/sync_checker/formats/document_collection_check.rb
class DocumentCollectionCheck < EditionBase
def root_path
"/government/collections/"
end
def rendering_app
Whitehall::RenderingApp::WHITEHALL_FRONTEND
end
def checks_for_live(locale)
super
end
def expected_details_hash(edition)
super.merge(
collection_groups: collection_groups(edition)
)
end
private
<|file_sep|>updated/lib/sync_checker/formats/document_collection_check.rb
class DocumentCollectionCheck < EditionBase
def root_path
"/government/collections/"
end
def rendering_app
Whitehall::RenderingApp::WHITEHALL_FRONTEND
end
def checks_for_live(locale)
super << Checks::LinksCheck.new(
"topical_events",
TopicalEvent
.for_edition(edition_expected_in_live.id)
.pluck(:content_id)
)
end
def expected_details_hash(edition)
super.merge(
collection_groups: collection_groups(edition) | 9a8c0345c2ad1d140038fe2c4433289c0cc89588 | lib/sync_checker/formats/document_collection_check.rb | lib/sync_checker/formats/document_collection_check.rb | Ruby |
<|file_sep|>original/setup.cfg
[metadata]
name = fuel-agent
version = 6.0.0
author = Mirantis
author-email = fuel-dev@lists.launchpad.net
summary = Fuel agent
classifier =
Development Status :: 4 - Beta
Programming Language :: Python
[files]
packages =
fuel_agent
[entry_points]
console_scripts =
provision = fuel_agent.cmd.agent:provision
partition = fuel_agent.cmd.agent:partition
configdrive = fuel_agent.cmd.agent:configdrive
copyimage = fuel_agent.cmd.agent:copyimage
bootloader = fuel_agent.cmd.agent:bootloader
<|file_sep|>current/setup.cfg
[metadata]
name = fuel-agent
version = 6.0.0
author = Mirantis
author-email = fuel-dev@lists.launchpad.net
summary = Fuel agent
classifier =
Development Status :: 4 - Beta
Programming Language :: Python
[files]
packages =
fuel_agent
[entry_points]
console_scripts =
provision = fuel_agent.cmd.agent:provision
partition = fuel_agent.cmd.agent:partition
configdrive = fuel_agent.cmd.agent:configdrive
copyimage = fuel_agent.cmd.agent:copyimage
bootloader = fuel_agent.cmd.agent:bootloader
<|file_sep|>updated/setup.cfg | [metadata]
name = fuel-agent
version = 0.1.0
author = Mirantis
author-email = fuel-dev@lists.launchpad.net
summary = Fuel agent
classifier =
Development Status :: 4 - Beta
Programming Language :: Python
[files]
packages =
fuel_agent
[entry_points]
console_scripts =
provision = fuel_agent.cmd.agent:provision
partition = fuel_agent.cmd.agent:partition
configdrive = fuel_agent.cmd.agent:configdrive
copyimage = fuel_agent.cmd.agent:copyimage
bootloader = fuel_agent.cmd.agent:bootloader | <|file_sep|>original/setup.cfg
[metadata]
name = fuel-agent
version = 6.0.0
author = Mirantis
author-email = fuel-dev@lists.launchpad.net
summary = Fuel agent
classifier =
Development Status :: 4 - Beta
Programming Language :: Python
[files]
packages =
fuel_agent
[entry_points]
console_scripts =
provision = fuel_agent.cmd.agent:provision
partition = fuel_agent.cmd.agent:partition
configdrive = fuel_agent.cmd.agent:configdrive
copyimage = fuel_agent.cmd.agent:copyimage
bootloader = fuel_agent.cmd.agent:bootloader
<|file_sep|>current/setup.cfg
[metadata]
name = fuel-agent
version = 6.0.0
author = Mirantis
author-email = fuel-dev@lists.launchpad.net
summary = Fuel agent
classifier =
Development Status :: 4 - Beta
Programming Language :: Python
[files]
packages =
fuel_agent
[entry_points]
console_scripts =
provision = fuel_agent.cmd.agent:provision
partition = fuel_agent.cmd.agent:partition
configdrive = fuel_agent.cmd.agent:configdrive
copyimage = fuel_agent.cmd.agent:copyimage
bootloader = fuel_agent.cmd.agent:bootloader
<|file_sep|>updated/setup.cfg
[metadata]
name = fuel-agent
version = 0.1.0
author = Mirantis
author-email = fuel-dev@lists.launchpad.net
summary = Fuel agent
classifier =
Development Status :: 4 - Beta
Programming Language :: Python
[files]
packages =
fuel_agent
[entry_points]
console_scripts =
provision = fuel_agent.cmd.agent:provision
partition = fuel_agent.cmd.agent:partition
configdrive = fuel_agent.cmd.agent:configdrive
copyimage = fuel_agent.cmd.agent:copyimage
bootloader = fuel_agent.cmd.agent:bootloader | c6ab65210d13f8f82befc7dab3806588d1764def | setup.cfg | setup.cfg | INI |
<|file_sep|>original/heroku/heroku-notes.md
# Heroku Notes
Heroku is a PaaS that has support for various languages.
See https://en.wikipedia.org/wiki/Heroku and https://www.heroku.com/what for
details.
The interface to Heroku primarily consists of a web-based dashboard and the
`heroku` CLI.
## Uncategorized
**App's First Deployment**
- `heroku login`
- `cd path/to/app/`
- `heroku create`
- `git push heroku master`
- `heroku open`
<|file_sep|>current/heroku/heroku-notes.md
# Heroku Notes
Heroku is a PaaS that has support for various languages.
See https://en.wikipedia.org/wiki/Heroku and https://www.heroku.com/what for
details.
The interface to Heroku primarily consists of a web-based dashboard and the
`heroku` CLI.
## Uncategorized
**App's First Deployment**
- `heroku login`
- `cd path/to/app/`
- `heroku create`
- `git push heroku master`
- `heroku open`
<|file_sep|>updated/heroku/heroku-notes.md |
## Uncategorized
**App's First Deployment**
- `heroku login`
- `cd path/to/app/`
- `heroku create`
- `git push heroku master`
- `heroku open`
**Associate a Git Repo with Heroku**
- `heroku login`
- `heroku git:remote -a <your-app-name>`
**Scaling a Web App**
- `heroku login`
- `cd path/to/app/`
- `heroku ps:scale web={n}` | <|file_sep|>original/heroku/heroku-notes.md
# Heroku Notes
Heroku is a PaaS that has support for various languages.
See https://en.wikipedia.org/wiki/Heroku and https://www.heroku.com/what for
details.
The interface to Heroku primarily consists of a web-based dashboard and the
`heroku` CLI.
## Uncategorized
**App's First Deployment**
- `heroku login`
- `cd path/to/app/`
- `heroku create`
- `git push heroku master`
- `heroku open`
<|file_sep|>current/heroku/heroku-notes.md
# Heroku Notes
Heroku is a PaaS that has support for various languages.
See https://en.wikipedia.org/wiki/Heroku and https://www.heroku.com/what for
details.
The interface to Heroku primarily consists of a web-based dashboard and the
`heroku` CLI.
## Uncategorized
**App's First Deployment**
- `heroku login`
- `cd path/to/app/`
- `heroku create`
- `git push heroku master`
- `heroku open`
<|file_sep|>updated/heroku/heroku-notes.md
## Uncategorized
**App's First Deployment**
- `heroku login`
- `cd path/to/app/`
- `heroku create`
- `git push heroku master`
- `heroku open`
**Associate a Git Repo with Heroku**
- `heroku login`
- `heroku git:remote -a <your-app-name>`
**Scaling a Web App**
- `heroku login`
- `cd path/to/app/`
- `heroku ps:scale web={n}` | 05956313707776d5d810389ed44965d755dd56d9 | heroku/heroku-notes.md | heroku/heroku-notes.md | Markdown |
<|file_sep|>original/operations/bits-service-alicloud.yml
---
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/app_stash
value:
directory_key: "((resource_directory_key))"
fog_connection: &alicloud-config
aliyun_accesskey_id: "((alibaba_access_key_id))"
aliyun_accesskey_secret: "((alibaba_secret_access_key))"
aliyun_oss_endpoint: "((alibaba_endpoint))"
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/buildpacks
value:
directory_key: "((buildpack_directory_key))"
fog_connection: *alicloud-config
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/droplets
value:
directory_key: "((droplet_directory_key))"
fog_connection: *alicloud-config
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/packages
<|file_sep|>current/operations/bits-service-alicloud.yml
---
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/app_stash
value:
directory_key: "((resource_directory_key))"
fog_connection: &alicloud-config
aliyun_accesskey_id: "((alibaba_access_key_id))"
aliyun_accesskey_secret: "((alibaba_secret_access_key))"
aliyun_oss_endpoint: "((alibaba_endpoint))"
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/buildpacks
value:
directory_key: "((buildpack_directory_key))"
fog_connection: *alicloud-config
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/droplets
value:
directory_key: "((droplet_directory_key))"
fog_connection: *alicloud-config
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/packages
<|file_sep|>updated/operations/bits-service-alicloud.yml | ---
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/app_stash
value:
directory_key: "((resource_directory_key))"
fog_connection: &alicloud-config
provider: aliyun
aliyun_accesskey_id: "((alibaba_access_key_id))"
aliyun_accesskey_secret: "((alibaba_secret_access_key))"
aliyun_oss_endpoint: "((alibaba_endpoint))"
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/buildpacks
value:
directory_key: "((buildpack_directory_key))"
fog_connection: *alicloud-config
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/droplets
value:
directory_key: "((droplet_directory_key))"
fog_connection: *alicloud-config
- type: replace | <|file_sep|>original/operations/bits-service-alicloud.yml
---
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/app_stash
value:
directory_key: "((resource_directory_key))"
fog_connection: &alicloud-config
aliyun_accesskey_id: "((alibaba_access_key_id))"
aliyun_accesskey_secret: "((alibaba_secret_access_key))"
aliyun_oss_endpoint: "((alibaba_endpoint))"
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/buildpacks
value:
directory_key: "((buildpack_directory_key))"
fog_connection: *alicloud-config
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/droplets
value:
directory_key: "((droplet_directory_key))"
fog_connection: *alicloud-config
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/packages
<|file_sep|>current/operations/bits-service-alicloud.yml
---
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/app_stash
value:
directory_key: "((resource_directory_key))"
fog_connection: &alicloud-config
aliyun_accesskey_id: "((alibaba_access_key_id))"
aliyun_accesskey_secret: "((alibaba_secret_access_key))"
aliyun_oss_endpoint: "((alibaba_endpoint))"
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/buildpacks
value:
directory_key: "((buildpack_directory_key))"
fog_connection: *alicloud-config
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/droplets
value:
directory_key: "((droplet_directory_key))"
fog_connection: *alicloud-config
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/packages
<|file_sep|>updated/operations/bits-service-alicloud.yml
---
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/app_stash
value:
directory_key: "((resource_directory_key))"
fog_connection: &alicloud-config
provider: aliyun
aliyun_accesskey_id: "((alibaba_access_key_id))"
aliyun_accesskey_secret: "((alibaba_secret_access_key))"
aliyun_oss_endpoint: "((alibaba_endpoint))"
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/buildpacks
value:
directory_key: "((buildpack_directory_key))"
fog_connection: *alicloud-config
- type: replace
path: /instance_groups/name=bits/jobs/name=bits-service/properties/bits-service/droplets
value:
directory_key: "((droplet_directory_key))"
fog_connection: *alicloud-config
- type: replace | 439e775042a0db88f0b0ae515fea94d7ce986112 | operations/bits-service-alicloud.yml | operations/bits-service-alicloud.yml | YAML |
<|file_sep|>convert_airline_IATA_to_ICAO/run.ps1.diff
original:
$Fields = @{"iatacode" = "$req_query_iata"}
$WebResponse = Invoke-RestMethod -Uri "http://www.airlinecodes.co.uk/airlcoderes.asp" -Method Post -Body $Fields -Verbose
updated:
# status Y means current data
$Fields = @{"status" = "Y"; "iataairl" = "$req_query_iata"}
$WebResponse = Invoke-RestMethod -Uri "http://avcodes.co.uk/airlcoderes.asp" -Method Post -Body $Fields -Verbose
<|file_sep|>convert_airline_IATA_to_ICAO/run.ps1.diff
original:
$regex = '<TD>\D{3}<\/TD>'
updated:
$regex = 'ICAO Code:<br /> \D{3}'
<|file_sep|>convert_airline_IATA_to_ICAO/run.ps1.diff
original:
updated:
<|file_sep|>original/convert_airline_IATA_to_ICAO/run.ps1
$req_query_iata
$Fields = @{"iatacode" = "$req_query_iata"}
$WebResponse = Invoke-RestMethod -Uri "http://www.airlinecodes.co.uk/airlcoderes.asp" -Method Post -Body $Fields -Verbose
$regex = '<TD>\D{3}<\/TD>'
$WebResponse -match $regex
$result = $Matches[0]
Out-File -Encoding Ascii $res -inputObject $result.Substring(4,3)
<|file_sep|>current/convert_airline_IATA_to_ICAO/run.ps1
$req_query_iata
# status Y means current data
$Fields = @{"status" = "Y"; "iataairl" = "$req_query_iata"}
$WebResponse = Invoke-RestMethod -Uri "http://avcodes.co.uk/airlcoderes.asp" -Method Post -Body $Fields -Verbose
$regex = 'ICAO Code:<br /> \D{3}'
$WebResponse -match $regex
$result = $Matches[0]
Out-File -Encoding Ascii $res -inputObject $result.Substring(4,3)
<|file_sep|>updated/convert_airline_IATA_to_ICAO/run.ps1 |
$req_query_iata
# status Y means current data
$Fields = @{"status" = "Y"; "iataairl" = "$req_query_iata"}
$WebResponse = Invoke-RestMethod -Uri "http://avcodes.co.uk/airlcoderes.asp" -Method Post -Body $Fields -Verbose
$regex = 'ICAO Code:<br /> \D{3}'
$WebResponse -match $regex
$result = $Matches[0]
Out-File -Encoding Ascii $res -inputObject $result.Split(';')[1] | <|file_sep|>convert_airline_IATA_to_ICAO/run.ps1.diff
original:
$Fields = @{"iatacode" = "$req_query_iata"}
$WebResponse = Invoke-RestMethod -Uri "http://www.airlinecodes.co.uk/airlcoderes.asp" -Method Post -Body $Fields -Verbose
updated:
# status Y means current data
$Fields = @{"status" = "Y"; "iataairl" = "$req_query_iata"}
$WebResponse = Invoke-RestMethod -Uri "http://avcodes.co.uk/airlcoderes.asp" -Method Post -Body $Fields -Verbose
<|file_sep|>convert_airline_IATA_to_ICAO/run.ps1.diff
original:
$regex = '<TD>\D{3}<\/TD>'
updated:
$regex = 'ICAO Code:<br /> \D{3}'
<|file_sep|>convert_airline_IATA_to_ICAO/run.ps1.diff
original:
updated:
<|file_sep|>original/convert_airline_IATA_to_ICAO/run.ps1
$req_query_iata
$Fields = @{"iatacode" = "$req_query_iata"}
$WebResponse = Invoke-RestMethod -Uri "http://www.airlinecodes.co.uk/airlcoderes.asp" -Method Post -Body $Fields -Verbose
$regex = '<TD>\D{3}<\/TD>'
$WebResponse -match $regex
$result = $Matches[0]
Out-File -Encoding Ascii $res -inputObject $result.Substring(4,3)
<|file_sep|>current/convert_airline_IATA_to_ICAO/run.ps1
$req_query_iata
# status Y means current data
$Fields = @{"status" = "Y"; "iataairl" = "$req_query_iata"}
$WebResponse = Invoke-RestMethod -Uri "http://avcodes.co.uk/airlcoderes.asp" -Method Post -Body $Fields -Verbose
$regex = 'ICAO Code:<br /> \D{3}'
$WebResponse -match $regex
$result = $Matches[0]
Out-File -Encoding Ascii $res -inputObject $result.Substring(4,3)
<|file_sep|>updated/convert_airline_IATA_to_ICAO/run.ps1
$req_query_iata
# status Y means current data
$Fields = @{"status" = "Y"; "iataairl" = "$req_query_iata"}
$WebResponse = Invoke-RestMethod -Uri "http://avcodes.co.uk/airlcoderes.asp" -Method Post -Body $Fields -Verbose
$regex = 'ICAO Code:<br /> \D{3}'
$WebResponse -match $regex
$result = $Matches[0]
Out-File -Encoding Ascii $res -inputObject $result.Split(';')[1] | 8d1b0e15d4afcf6bd5cae4ef05a9510d4be8db00 | convert_airline_IATA_to_ICAO/run.ps1 | convert_airline_IATA_to_ICAO/run.ps1 | PowerShell |
<|file_sep|>original/requirements.txt
beautifulsoup4==4.7.1
croniter==0.3.27
Cython==0.29.6
git+git://github.com/andreasvc/pyre2.git@7146ce3#egg=re2
google-api-python-client==1.7.8
isodate==0.6.0
jsonpath-ng==1.4.3
lxml==4.3.2
numpy==1.16.1
parsedatetime==2.4
psutil==5.5.1
pyhedrals==0.1.2
python-dateutil==2.8.0
pytimeparse==1.1.8
pyxDamerauLevenshtein==1.5.2
requests==2.21.0
ruamel.yaml==0.15.88
Twisted[tls]==18.9.0
<|file_sep|>current/requirements.txt
beautifulsoup4==4.7.1
croniter==0.3.27
Cython==0.29.6
git+git://github.com/andreasvc/pyre2.git@7146ce3#egg=re2
google-api-python-client==1.7.8
isodate==0.6.0
jsonpath-ng==1.4.3
lxml==4.3.2
numpy==1.16.1
parsedatetime==2.4
psutil==5.5.1
pyhedrals==0.1.2
python-dateutil==2.8.0
pytimeparse==1.1.8
pyxDamerauLevenshtein==1.5.2
requests==2.21.0
ruamel.yaml==0.15.88
Twisted[tls]==18.9.0
<|file_sep|>updated/requirements.txt | beautifulsoup4==4.7.1
croniter==0.3.27
Cython==0.29.6
git+git://github.com/andreasvc/pyre2.git@7146ce3#egg=re2
google-api-python-client==1.7.8
isodate==0.6.0
jsonpath-ng==1.4.3
lxml==4.3.2
numpy==1.16.2
parsedatetime==2.4
psutil==5.5.1
pyhedrals==0.1.2
python-dateutil==2.8.0
pytimeparse==1.1.8
pyxDamerauLevenshtein==1.5.2
requests==2.21.0
ruamel.yaml==0.15.88
Twisted[tls]==18.9.0 | <|file_sep|>original/requirements.txt
beautifulsoup4==4.7.1
croniter==0.3.27
Cython==0.29.6
git+git://github.com/andreasvc/pyre2.git@7146ce3#egg=re2
google-api-python-client==1.7.8
isodate==0.6.0
jsonpath-ng==1.4.3
lxml==4.3.2
numpy==1.16.1
parsedatetime==2.4
psutil==5.5.1
pyhedrals==0.1.2
python-dateutil==2.8.0
pytimeparse==1.1.8
pyxDamerauLevenshtein==1.5.2
requests==2.21.0
ruamel.yaml==0.15.88
Twisted[tls]==18.9.0
<|file_sep|>current/requirements.txt
beautifulsoup4==4.7.1
croniter==0.3.27
Cython==0.29.6
git+git://github.com/andreasvc/pyre2.git@7146ce3#egg=re2
google-api-python-client==1.7.8
isodate==0.6.0
jsonpath-ng==1.4.3
lxml==4.3.2
numpy==1.16.1
parsedatetime==2.4
psutil==5.5.1
pyhedrals==0.1.2
python-dateutil==2.8.0
pytimeparse==1.1.8
pyxDamerauLevenshtein==1.5.2
requests==2.21.0
ruamel.yaml==0.15.88
Twisted[tls]==18.9.0
<|file_sep|>updated/requirements.txt
beautifulsoup4==4.7.1
croniter==0.3.27
Cython==0.29.6
git+git://github.com/andreasvc/pyre2.git@7146ce3#egg=re2
google-api-python-client==1.7.8
isodate==0.6.0
jsonpath-ng==1.4.3
lxml==4.3.2
numpy==1.16.2
parsedatetime==2.4
psutil==5.5.1
pyhedrals==0.1.2
python-dateutil==2.8.0
pytimeparse==1.1.8
pyxDamerauLevenshtein==1.5.2
requests==2.21.0
ruamel.yaml==0.15.88
Twisted[tls]==18.9.0 | cf06ad0189cab817dc0d4288e0e35005ee568201 | requirements.txt | requirements.txt | Text |
<|file_sep|>original/lib/vagrant-butcher/action/copy_guest_key.rb
end
unless File.exists?(cache_dir(env))
env[:butcher].ui.info "Creating #{cache_dir(env)}"
Dir.mkdir(cache_dir(env))
end
begin
env[:machine].communicate.execute "cp #{guest_key_path(env)} #{guest_client_key_path(env)}", :sudo => true
rescue Exception => e
env[:butcher].ui.error "Failed to copy #{guest_key_path(env)} to #{client_key_path(env)}"
env[:butcher].ui.error e
return false
end
env[:butcher].ui.info "Copied #{guest_key_path(env)} to #{client_key_path(env)}"
return true
end
def call(env)
<|file_sep|>current/lib/vagrant-butcher/action/copy_guest_key.rb
end
unless File.exists?(cache_dir(env))
env[:butcher].ui.info "Creating #{cache_dir(env)}"
Dir.mkdir(cache_dir(env))
end
begin
env[:machine].communicate.execute "cp #{guest_key_path(env)} #{guest_client_key_path(env)}", :sudo => true
rescue Exception => e
env[:butcher].ui.error "Failed to copy #{guest_key_path(env)} to #{client_key_path(env)}"
env[:butcher].ui.error e
return false
end
env[:butcher].ui.info "Copied #{guest_key_path(env)} to #{client_key_path(env)}"
return true
end
def call(env)
<|file_sep|>updated/lib/vagrant-butcher/action/copy_guest_key.rb | end
unless File.exists?(cache_dir(env))
env[:butcher].ui.info "Creating #{cache_dir(env)}"
Dir.mkdir(cache_dir(env))
end
begin
env[:machine].communicate.execute "cp #{guest_key_path(env)} #{guest_client_key_path(env)}", :sudo => true
rescue Exception => e
env[:butcher].ui.error "Failed to copy #{guest_key_path(env)} to #{client_key_path(env)}"
env[:butcher].ui.error e
return false
end
env[:butcher].ui.info "Copied #{guest_key_path(env)} to #{client_key_path(env)}"
return true
end
def call(env) | <|file_sep|>original/lib/vagrant-butcher/action/copy_guest_key.rb
end
unless File.exists?(cache_dir(env))
env[:butcher].ui.info "Creating #{cache_dir(env)}"
Dir.mkdir(cache_dir(env))
end
begin
env[:machine].communicate.execute "cp #{guest_key_path(env)} #{guest_client_key_path(env)}", :sudo => true
rescue Exception => e
env[:butcher].ui.error "Failed to copy #{guest_key_path(env)} to #{client_key_path(env)}"
env[:butcher].ui.error e
return false
end
env[:butcher].ui.info "Copied #{guest_key_path(env)} to #{client_key_path(env)}"
return true
end
def call(env)
<|file_sep|>current/lib/vagrant-butcher/action/copy_guest_key.rb
end
unless File.exists?(cache_dir(env))
env[:butcher].ui.info "Creating #{cache_dir(env)}"
Dir.mkdir(cache_dir(env))
end
begin
env[:machine].communicate.execute "cp #{guest_key_path(env)} #{guest_client_key_path(env)}", :sudo => true
rescue Exception => e
env[:butcher].ui.error "Failed to copy #{guest_key_path(env)} to #{client_key_path(env)}"
env[:butcher].ui.error e
return false
end
env[:butcher].ui.info "Copied #{guest_key_path(env)} to #{client_key_path(env)}"
return true
end
def call(env)
<|file_sep|>updated/lib/vagrant-butcher/action/copy_guest_key.rb
end
unless File.exists?(cache_dir(env))
env[:butcher].ui.info "Creating #{cache_dir(env)}"
Dir.mkdir(cache_dir(env))
end
begin
env[:machine].communicate.execute "cp #{guest_key_path(env)} #{guest_client_key_path(env)}", :sudo => true
rescue Exception => e
env[:butcher].ui.error "Failed to copy #{guest_key_path(env)} to #{client_key_path(env)}"
env[:butcher].ui.error e
return false
end
env[:butcher].ui.info "Copied #{guest_key_path(env)} to #{client_key_path(env)}"
return true
end
def call(env) | 4f0b9f26239c06fbeffb3f5e23aeed06c665d2ad | lib/vagrant-butcher/action/copy_guest_key.rb | lib/vagrant-butcher/action/copy_guest_key.rb | Ruby |
<|file_sep|>setup.py.diff
original:
updated:
import os
<|file_sep|>setup.py.diff
original:
updated:
with open(os.path.join(os.path.dirname(__file__), 'README.rst')) as fh:
readme = fh.read()
<|file_sep|>original/setup.py
}
setup(
name='huey',
version=__import__('huey').__version__,
description='huey, a little task queue',
author='Charles Leifer',
author_email='coleifer@gmail.com',
url='http://github.com/coleifer/huey/',
packages=find_packages(),
extras_require=extras_require,
package_data={
'huey': [
],
},
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
<|file_sep|>current/setup.py
readme = fh.read()
extras_require = {
'backends': ('peewee', "redis"),
}
setup(
name='huey',
version=__import__('huey').__version__,
description='huey, a little task queue',
author='Charles Leifer',
author_email='coleifer@gmail.com',
url='http://github.com/coleifer/huey/',
packages=find_packages(),
extras_require=extras_require,
package_data={
'huey': [
],
},
classifiers=[
'Development Status :: 4 - Beta',
<|file_sep|>updated/setup.py | readme = fh.read()
extras_require = {
'backends': ('peewee', "redis"),
}
setup(
name='huey',
version=__import__('huey').__version__,
description='huey, a little task queue',
long_description=readme,
author='Charles Leifer',
author_email='coleifer@gmail.com',
url='http://github.com/coleifer/huey/',
packages=find_packages(),
extras_require=extras_require,
package_data={
'huey': [
],
},
classifiers=[ | <|file_sep|>setup.py.diff
original:
updated:
import os
<|file_sep|>setup.py.diff
original:
updated:
with open(os.path.join(os.path.dirname(__file__), 'README.rst')) as fh:
readme = fh.read()
<|file_sep|>original/setup.py
}
setup(
name='huey',
version=__import__('huey').__version__,
description='huey, a little task queue',
author='Charles Leifer',
author_email='coleifer@gmail.com',
url='http://github.com/coleifer/huey/',
packages=find_packages(),
extras_require=extras_require,
package_data={
'huey': [
],
},
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
<|file_sep|>current/setup.py
readme = fh.read()
extras_require = {
'backends': ('peewee', "redis"),
}
setup(
name='huey',
version=__import__('huey').__version__,
description='huey, a little task queue',
author='Charles Leifer',
author_email='coleifer@gmail.com',
url='http://github.com/coleifer/huey/',
packages=find_packages(),
extras_require=extras_require,
package_data={
'huey': [
],
},
classifiers=[
'Development Status :: 4 - Beta',
<|file_sep|>updated/setup.py
readme = fh.read()
extras_require = {
'backends': ('peewee', "redis"),
}
setup(
name='huey',
version=__import__('huey').__version__,
description='huey, a little task queue',
long_description=readme,
author='Charles Leifer',
author_email='coleifer@gmail.com',
url='http://github.com/coleifer/huey/',
packages=find_packages(),
extras_require=extras_require,
package_data={
'huey': [
],
},
classifiers=[ | 32f6acb989ea5bcf7df1ad1dbc6385c48f443d3e | setup.py | setup.py | Python |
<|file_sep|>config/core.entity_view_display.node.insert_box.default.yml.diff
original:
field_insert_type:
label: hidden
type: entity_reference_label
weight: 20
settings:
link: true
third_party_settings: { }
region: content
field_topic:
label: hidden
type: entity_reference_label
weight: 20
settings:
link: true
third_party_settings: { }
region: content
updated:
<|file_sep|>original/config/core.entity_view_display.node.insert_box.default.yml
- text
- user
id: node.insert_box.default
targetEntityType: node
bundle: insert_box
mode: default
content:
body:
label: hidden
type: text_default
weight: 0
settings: { }
third_party_settings: { }
region: content
field_insert_type:
label: hidden
type: entity_reference_label
weight: 20
settings:
link: true
third_party_settings: { }
<|file_sep|>current/config/core.entity_view_display.node.insert_box.default.yml
- text
- user
id: node.insert_box.default
targetEntityType: node
bundle: insert_box
mode: default
content:
body:
label: hidden
type: text_default
weight: 0
settings: { }
third_party_settings: { }
region: content
links:
weight: 1
region: content
settings: { }
third_party_settings: { }
hidden:
field_header_text: true
<|file_sep|>updated/config/core.entity_view_display.node.insert_box.default.yml | id: node.insert_box.default
targetEntityType: node
bundle: insert_box
mode: default
content:
body:
label: hidden
type: text_default
weight: 0
settings: { }
third_party_settings: { }
region: content
links:
weight: 1
region: content
settings: { }
third_party_settings: { }
hidden:
field_header_text: true
field_insert_type: true
field_topic: true | <|file_sep|>config/core.entity_view_display.node.insert_box.default.yml.diff
original:
field_insert_type:
label: hidden
type: entity_reference_label
weight: 20
settings:
link: true
third_party_settings: { }
region: content
field_topic:
label: hidden
type: entity_reference_label
weight: 20
settings:
link: true
third_party_settings: { }
region: content
updated:
<|file_sep|>original/config/core.entity_view_display.node.insert_box.default.yml
- text
- user
id: node.insert_box.default
targetEntityType: node
bundle: insert_box
mode: default
content:
body:
label: hidden
type: text_default
weight: 0
settings: { }
third_party_settings: { }
region: content
field_insert_type:
label: hidden
type: entity_reference_label
weight: 20
settings:
link: true
third_party_settings: { }
<|file_sep|>current/config/core.entity_view_display.node.insert_box.default.yml
- text
- user
id: node.insert_box.default
targetEntityType: node
bundle: insert_box
mode: default
content:
body:
label: hidden
type: text_default
weight: 0
settings: { }
third_party_settings: { }
region: content
links:
weight: 1
region: content
settings: { }
third_party_settings: { }
hidden:
field_header_text: true
<|file_sep|>updated/config/core.entity_view_display.node.insert_box.default.yml
id: node.insert_box.default
targetEntityType: node
bundle: insert_box
mode: default
content:
body:
label: hidden
type: text_default
weight: 0
settings: { }
third_party_settings: { }
region: content
links:
weight: 1
region: content
settings: { }
third_party_settings: { }
hidden:
field_header_text: true
field_insert_type: true
field_topic: true | 3cb623c53045e23a3b604fcbc3f6a1513528c9ad | config/core.entity_view_display.node.insert_box.default.yml | config/core.entity_view_display.node.insert_box.default.yml | YAML |
<|file_sep|>original/website/static/js/growlBox.js
}(this, function($) { // named dependencies here
'use strict';
// Private methods go up here
// This is the public API
// The constructor
function GrowlBox (title, message) {
var self = this;
if (typeof title === 'undefined'){
title = '';
}
self.title = title;
self.message = message;
self.init(self);
}
// Methods
GrowlBox.prototype.init = function(self) {
$.growl({
title: '<strong>' + self.title + '<strong><br />',
message: self.message
<|file_sep|>current/website/static/js/growlBox.js
}(this, function($) { // named dependencies here
'use strict';
// Private methods go up here
// This is the public API
// The constructor
function GrowlBox (title, message) {
var self = this;
if (typeof title === 'undefined'){
title = '';
}
self.title = title;
self.message = message;
self.init(self);
}
// Methods
GrowlBox.prototype.init = function(self) {
$.growl({
title: '<strong>' + self.title + '<strong><br />',
message: self.message
<|file_sep|>updated/website/static/js/growlBox.js | }(this, function($) { // named dependencies here
'use strict';
// Private methods go up here
// This is the public API
// The constructor
function GrowlBox (title, message) {
var self = this;
self.title = title;
self.message = message;
self.init(self);
}
// Methods
GrowlBox.prototype.init = function(self) {
$.growl({
title: '<strong>' + self.title + '<strong><br />',
message: self.message
},{
type: 'danger', | <|file_sep|>original/website/static/js/growlBox.js
}(this, function($) { // named dependencies here
'use strict';
// Private methods go up here
// This is the public API
// The constructor
function GrowlBox (title, message) {
var self = this;
if (typeof title === 'undefined'){
title = '';
}
self.title = title;
self.message = message;
self.init(self);
}
// Methods
GrowlBox.prototype.init = function(self) {
$.growl({
title: '<strong>' + self.title + '<strong><br />',
message: self.message
<|file_sep|>current/website/static/js/growlBox.js
}(this, function($) { // named dependencies here
'use strict';
// Private methods go up here
// This is the public API
// The constructor
function GrowlBox (title, message) {
var self = this;
if (typeof title === 'undefined'){
title = '';
}
self.title = title;
self.message = message;
self.init(self);
}
// Methods
GrowlBox.prototype.init = function(self) {
$.growl({
title: '<strong>' + self.title + '<strong><br />',
message: self.message
<|file_sep|>updated/website/static/js/growlBox.js
}(this, function($) { // named dependencies here
'use strict';
// Private methods go up here
// This is the public API
// The constructor
function GrowlBox (title, message) {
var self = this;
self.title = title;
self.message = message;
self.init(self);
}
// Methods
GrowlBox.prototype.init = function(self) {
$.growl({
title: '<strong>' + self.title + '<strong><br />',
message: self.message
},{
type: 'danger', | c1f06e014582a4355616bd3b8a4e216ec21c600a | website/static/js/growlBox.js | website/static/js/growlBox.js | JavaScript |
<|file_sep|>original/src/requirements.txt
chart_studio==1.1.0
coverage==5.1
decorator==4.4.2
jinja2==2.11.2
jsonschema==3.2.0
markupsafe==1.1.1
mock==4.0.2
nose>=1.1.2
pexpect==4.8.0
pickleshare==0.7.5
pillow==7.1.2
plotly==4.8.2
pyasn1==0.4.8
pycrypto>=2.6
pygments==2.6.1
pymongo==3.10.1
pyyaml==5.3.1
requests==2.24.0
rsa==4.6
semantic_version==2.8.5
setuptools==47.3.1
<|file_sep|>current/src/requirements.txt
chart_studio==1.1.0
coverage==5.1
decorator==4.4.2
jinja2==2.11.2
jsonschema==3.2.0
markupsafe==1.1.1
mock==4.0.2
nose>=1.1.2
pexpect==4.8.0
pickleshare==0.7.5
pillow==7.1.2
plotly==4.8.2
pyasn1==0.4.8
pycrypto>=2.6
pygments==2.6.1
pymongo==3.10.1
pyyaml==5.3.1
requests==2.24.0
rsa==4.6
semantic_version==2.8.5
setuptools==47.3.1
<|file_sep|>updated/src/requirements.txt | chart_studio==1.1.0
coverage==5.1
decorator==4.4.2
jinja2==2.11.2
jsonschema==3.2.0
markupsafe==1.1.1
mock==4.0.2
nose>=1.1.2
pexpect==4.8.0
pickleshare==0.7.5
pillow==7.2.0
plotly==4.8.2
pyasn1==0.4.8
pycrypto>=2.6
pygments==2.6.1
pymongo==3.10.1
pyyaml==5.3.1
requests==2.24.0
rsa==4.6
semantic_version==2.8.5
setuptools==47.3.1 | <|file_sep|>original/src/requirements.txt
chart_studio==1.1.0
coverage==5.1
decorator==4.4.2
jinja2==2.11.2
jsonschema==3.2.0
markupsafe==1.1.1
mock==4.0.2
nose>=1.1.2
pexpect==4.8.0
pickleshare==0.7.5
pillow==7.1.2
plotly==4.8.2
pyasn1==0.4.8
pycrypto>=2.6
pygments==2.6.1
pymongo==3.10.1
pyyaml==5.3.1
requests==2.24.0
rsa==4.6
semantic_version==2.8.5
setuptools==47.3.1
<|file_sep|>current/src/requirements.txt
chart_studio==1.1.0
coverage==5.1
decorator==4.4.2
jinja2==2.11.2
jsonschema==3.2.0
markupsafe==1.1.1
mock==4.0.2
nose>=1.1.2
pexpect==4.8.0
pickleshare==0.7.5
pillow==7.1.2
plotly==4.8.2
pyasn1==0.4.8
pycrypto>=2.6
pygments==2.6.1
pymongo==3.10.1
pyyaml==5.3.1
requests==2.24.0
rsa==4.6
semantic_version==2.8.5
setuptools==47.3.1
<|file_sep|>updated/src/requirements.txt
chart_studio==1.1.0
coverage==5.1
decorator==4.4.2
jinja2==2.11.2
jsonschema==3.2.0
markupsafe==1.1.1
mock==4.0.2
nose>=1.1.2
pexpect==4.8.0
pickleshare==0.7.5
pillow==7.2.0
plotly==4.8.2
pyasn1==0.4.8
pycrypto>=2.6
pygments==2.6.1
pymongo==3.10.1
pyyaml==5.3.1
requests==2.24.0
rsa==4.6
semantic_version==2.8.5
setuptools==47.3.1 | 5b982d72cc68c323d46299f0a2c896d5dfdd5ec9 | src/requirements.txt | src/requirements.txt | Text |
<|file_sep|>original/.travis.yml
sudo: false
language: node_js
node_js:
- iojs
- "0.12"
- "0.10"
<|file_sep|>current/.travis.yml
sudo: false
language: node_js
node_js:
- iojs
- "0.12"
- "0.10"
<|file_sep|>updated/.travis.yml | sudo: false
language: node_js
node_js:
- iojs
- "0.12" | <|file_sep|>original/.travis.yml
sudo: false
language: node_js
node_js:
- iojs
- "0.12"
- "0.10"
<|file_sep|>current/.travis.yml
sudo: false
language: node_js
node_js:
- iojs
- "0.12"
- "0.10"
<|file_sep|>updated/.travis.yml
sudo: false
language: node_js
node_js:
- iojs
- "0.12" | 5798d23fc9574b9798d3cea4004a2eee18807861 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/unsupported/test/CMakeLists.txt
include_directories(${ADOLC_INCLUDES})
ei_add_property(EIGEN_TESTED_BACKENDS "Adolc")
ei_add_test(forward_adolc " " ${ADOLC_LIBRARIES})
else(ADOLC_FOUND)
ei_add_property(EIGEN_MISSING_BACKENDS "Adolc")
endif(ADOLC_FOUND)
ei_add_test(NonLinear)
ei_add_test(NumericalDiff)
ei_add_test(autodiff)
ei_add_test(BVH)
#ei_add_test(matrixExponential)
ei_add_test(alignedvector3)
ei_add_test(FFT)
find_package(FFTW)
if(FFTW_FOUND)
ei_add_test(FFTW "-DEIGEN_FFTW_DEFAULT " "-lfftw3 -lfftw3f -lfftw3l" )
endif(FFTW_FOUND)
ei_add_test(Complex)
<|file_sep|>current/unsupported/test/CMakeLists.txt
include_directories(${ADOLC_INCLUDES})
ei_add_property(EIGEN_TESTED_BACKENDS "Adolc")
ei_add_test(forward_adolc " " ${ADOLC_LIBRARIES})
else(ADOLC_FOUND)
ei_add_property(EIGEN_MISSING_BACKENDS "Adolc")
endif(ADOLC_FOUND)
ei_add_test(NonLinear)
ei_add_test(NumericalDiff)
ei_add_test(autodiff)
ei_add_test(BVH)
#ei_add_test(matrixExponential)
ei_add_test(alignedvector3)
ei_add_test(FFT)
find_package(FFTW)
if(FFTW_FOUND)
ei_add_test(FFTW "-DEIGEN_FFTW_DEFAULT " "-lfftw3 -lfftw3f -lfftw3l" )
endif(FFTW_FOUND)
ei_add_test(Complex)
<|file_sep|>updated/unsupported/test/CMakeLists.txt | include_directories(${ADOLC_INCLUDES})
ei_add_property(EIGEN_TESTED_BACKENDS "Adolc")
ei_add_test(forward_adolc " " ${ADOLC_LIBRARIES})
else(ADOLC_FOUND)
ei_add_property(EIGEN_MISSING_BACKENDS "Adolc")
endif(ADOLC_FOUND)
ei_add_test(NonLinear)
ei_add_test(NumericalDiff)
ei_add_test(autodiff)
ei_add_test(BVH)
ei_add_test(matrixExponential)
ei_add_test(alignedvector3)
ei_add_test(FFT)
find_package(FFTW)
if(FFTW_FOUND)
ei_add_test(FFTW "-DEIGEN_FFTW_DEFAULT " "-lfftw3 -lfftw3f -lfftw3l" )
endif(FFTW_FOUND)
ei_add_test(Complex) | <|file_sep|>original/unsupported/test/CMakeLists.txt
include_directories(${ADOLC_INCLUDES})
ei_add_property(EIGEN_TESTED_BACKENDS "Adolc")
ei_add_test(forward_adolc " " ${ADOLC_LIBRARIES})
else(ADOLC_FOUND)
ei_add_property(EIGEN_MISSING_BACKENDS "Adolc")
endif(ADOLC_FOUND)
ei_add_test(NonLinear)
ei_add_test(NumericalDiff)
ei_add_test(autodiff)
ei_add_test(BVH)
#ei_add_test(matrixExponential)
ei_add_test(alignedvector3)
ei_add_test(FFT)
find_package(FFTW)
if(FFTW_FOUND)
ei_add_test(FFTW "-DEIGEN_FFTW_DEFAULT " "-lfftw3 -lfftw3f -lfftw3l" )
endif(FFTW_FOUND)
ei_add_test(Complex)
<|file_sep|>current/unsupported/test/CMakeLists.txt
include_directories(${ADOLC_INCLUDES})
ei_add_property(EIGEN_TESTED_BACKENDS "Adolc")
ei_add_test(forward_adolc " " ${ADOLC_LIBRARIES})
else(ADOLC_FOUND)
ei_add_property(EIGEN_MISSING_BACKENDS "Adolc")
endif(ADOLC_FOUND)
ei_add_test(NonLinear)
ei_add_test(NumericalDiff)
ei_add_test(autodiff)
ei_add_test(BVH)
#ei_add_test(matrixExponential)
ei_add_test(alignedvector3)
ei_add_test(FFT)
find_package(FFTW)
if(FFTW_FOUND)
ei_add_test(FFTW "-DEIGEN_FFTW_DEFAULT " "-lfftw3 -lfftw3f -lfftw3l" )
endif(FFTW_FOUND)
ei_add_test(Complex)
<|file_sep|>updated/unsupported/test/CMakeLists.txt
include_directories(${ADOLC_INCLUDES})
ei_add_property(EIGEN_TESTED_BACKENDS "Adolc")
ei_add_test(forward_adolc " " ${ADOLC_LIBRARIES})
else(ADOLC_FOUND)
ei_add_property(EIGEN_MISSING_BACKENDS "Adolc")
endif(ADOLC_FOUND)
ei_add_test(NonLinear)
ei_add_test(NumericalDiff)
ei_add_test(autodiff)
ei_add_test(BVH)
ei_add_test(matrixExponential)
ei_add_test(alignedvector3)
ei_add_test(FFT)
find_package(FFTW)
if(FFTW_FOUND)
ei_add_test(FFTW "-DEIGEN_FFTW_DEFAULT " "-lfftw3 -lfftw3f -lfftw3l" )
endif(FFTW_FOUND)
ei_add_test(Complex) | e193ff7979865f025cbbe0756c86de2b18447324 | unsupported/test/CMakeLists.txt | unsupported/test/CMakeLists.txt | Text |
<|file_sep|>lib/search_object/base.rb.diff
original:
@scope, @filters = self.class.scope_and_filters(args)
updated:
@search = self.class.search args
<|file_sep|>lib/search_object/base.rb.diff
original:
@count ||= _fetch_results.count
updated:
@count ||= @search.count self
<|file_sep|>lib/search_object/base.rb.diff
original:
@filters
updated:
@search.params
<|file_sep|>lib/search_object/base.rb.diff
original:
@filters.merge Helper.stringify_keys(additions)
updated:
@search.params.merge Helper.stringify_keys(additions)
<|file_sep|>lib/search_object/base.rb.diff
original:
_fetch_results
end
def _fetch_results
self.class.fetch_results_for @scope, self
updated:
@search.query self
<|file_sep|>original/lib/search_object/base.rb
end
module ClassMethods
def scope_and_filters(args)
scope = (@scope && @scope.call) || args.shift
params = @defaults.merge(Helper.select_keys Helper.stringify_keys(args.shift || {}), @actions.keys)
[scope, params]
end
def fetch_results_for(scope, search)
search.params.inject(scope) do |scope, (name, value)|
new_scope = search.instance_exec scope, value, &@actions[name]
new_scope || scope
end
end
def scope(&block)
@scope = block
end
<|file_sep|>current/lib/search_object/base.rb
scope = (@scope && @scope.call) || args.shift
params = @defaults.merge(Helper.select_keys Helper.stringify_keys(args.shift || {}), @actions.keys)
Search.new scope, params, @actions
end
def scope(&block)
@scope = block
end
def option(name, default = nil, &block)
name = name.to_s
@defaults[name] = default unless default.nil?
@actions[name] = block || ->(scope, value) { scope.where name => value }
define_method(name) { @filters[name] }
end
end
end
end
<|file_sep|>updated/lib/search_object/base.rb | scope = (@scope && @scope.call) || args.shift
params = @defaults.merge(Helper.select_keys Helper.stringify_keys(args.shift || {}), @actions.keys)
Search.new scope, params, @actions
end
def scope(&block)
@scope = block
end
def option(name, default = nil, &block)
name = name.to_s
@defaults[name] = default unless default.nil?
@actions[name] = block || ->(scope, value) { scope.where name => value }
define_method(name) { @search.param name }
end
end
end
end | <|file_sep|>lib/search_object/base.rb.diff
original:
@scope, @filters = self.class.scope_and_filters(args)
updated:
@search = self.class.search args
<|file_sep|>lib/search_object/base.rb.diff
original:
@count ||= _fetch_results.count
updated:
@count ||= @search.count self
<|file_sep|>lib/search_object/base.rb.diff
original:
@filters
updated:
@search.params
<|file_sep|>lib/search_object/base.rb.diff
original:
@filters.merge Helper.stringify_keys(additions)
updated:
@search.params.merge Helper.stringify_keys(additions)
<|file_sep|>lib/search_object/base.rb.diff
original:
_fetch_results
end
def _fetch_results
self.class.fetch_results_for @scope, self
updated:
@search.query self
<|file_sep|>original/lib/search_object/base.rb
end
module ClassMethods
def scope_and_filters(args)
scope = (@scope && @scope.call) || args.shift
params = @defaults.merge(Helper.select_keys Helper.stringify_keys(args.shift || {}), @actions.keys)
[scope, params]
end
def fetch_results_for(scope, search)
search.params.inject(scope) do |scope, (name, value)|
new_scope = search.instance_exec scope, value, &@actions[name]
new_scope || scope
end
end
def scope(&block)
@scope = block
end
<|file_sep|>current/lib/search_object/base.rb
scope = (@scope && @scope.call) || args.shift
params = @defaults.merge(Helper.select_keys Helper.stringify_keys(args.shift || {}), @actions.keys)
Search.new scope, params, @actions
end
def scope(&block)
@scope = block
end
def option(name, default = nil, &block)
name = name.to_s
@defaults[name] = default unless default.nil?
@actions[name] = block || ->(scope, value) { scope.where name => value }
define_method(name) { @filters[name] }
end
end
end
end
<|file_sep|>updated/lib/search_object/base.rb
scope = (@scope && @scope.call) || args.shift
params = @defaults.merge(Helper.select_keys Helper.stringify_keys(args.shift || {}), @actions.keys)
Search.new scope, params, @actions
end
def scope(&block)
@scope = block
end
def option(name, default = nil, &block)
name = name.to_s
@defaults[name] = default unless default.nil?
@actions[name] = block || ->(scope, value) { scope.where name => value }
define_method(name) { @search.param name }
end
end
end
end | 0e99905e1a420edfae2b694479629552d33ef420 | lib/search_object/base.rb | lib/search_object/base.rb | Ruby |
<|file_sep|>original/lib/tritium/parser/instructions/literal.rb
module Tritium
module Parser
module Instructions
class Literal < Instruction
def initialize(filename, line_num, value)
super(filename, line_num)
@value = value
end
def to_s(depth = 0)
"#{@@tab * depth}#{@value.inspect}"
end
def unquote
val = eval(self.to_s)
Regexp === val ? val.inspect : val
end
end
end
end
end
<|file_sep|>current/lib/tritium/parser/instructions/literal.rb
module Tritium
module Parser
module Instructions
class Literal < Instruction
def initialize(filename, line_num, value)
super(filename, line_num)
@value = value
end
def to_s(depth = 0)
"#{@@tab * depth}#{@value.inspect}"
end
def unquote
val = eval(self.to_s)
Regexp === val ? val.inspect : val
end
end
end
end
end
<|file_sep|>updated/lib/tritium/parser/instructions/literal.rb | module Tritium
module Parser
module Instructions
class Literal < Instruction
attr :value
def initialize(filename, line_num, value)
super(filename, line_num)
@value = value
end
def to_s(depth = 0)
"#{@@tab * depth}#{@value.inspect}"
end
def unquote
val = eval(self.to_s)
Regexp === val ? val.inspect : val
end
end
end
end | <|file_sep|>original/lib/tritium/parser/instructions/literal.rb
module Tritium
module Parser
module Instructions
class Literal < Instruction
def initialize(filename, line_num, value)
super(filename, line_num)
@value = value
end
def to_s(depth = 0)
"#{@@tab * depth}#{@value.inspect}"
end
def unquote
val = eval(self.to_s)
Regexp === val ? val.inspect : val
end
end
end
end
end
<|file_sep|>current/lib/tritium/parser/instructions/literal.rb
module Tritium
module Parser
module Instructions
class Literal < Instruction
def initialize(filename, line_num, value)
super(filename, line_num)
@value = value
end
def to_s(depth = 0)
"#{@@tab * depth}#{@value.inspect}"
end
def unquote
val = eval(self.to_s)
Regexp === val ? val.inspect : val
end
end
end
end
end
<|file_sep|>updated/lib/tritium/parser/instructions/literal.rb
module Tritium
module Parser
module Instructions
class Literal < Instruction
attr :value
def initialize(filename, line_num, value)
super(filename, line_num)
@value = value
end
def to_s(depth = 0)
"#{@@tab * depth}#{@value.inspect}"
end
def unquote
val = eval(self.to_s)
Regexp === val ? val.inspect : val
end
end
end
end | b42381968ae84c2bab5faf41b9f12bd63f5c8927 | lib/tritium/parser/instructions/literal.rb | lib/tritium/parser/instructions/literal.rb | Ruby |
<|file_sep|>original/after/plugin/after_bundles.vim
" You can make commands run after plugins are loaded from the .vimrc (and
" init/ directory) like so:
"
" autocmd User AfterBundles SomePluginCommand
"
" This lets you use commands defined in plugins, which haven't been created at
" .vimrc-time.
doautocmd User AfterBundles
<|file_sep|>current/after/plugin/after_bundles.vim
" You can make commands run after plugins are loaded from the .vimrc (and
" init/ directory) like so:
"
" autocmd User AfterBundles SomePluginCommand
"
" This lets you use commands defined in plugins, which haven't been created at
" .vimrc-time.
doautocmd User AfterBundles
<|file_sep|>updated/after/plugin/after_bundles.vim | " You can make commands run after plugins are loaded from the .vimrc (and
" init/ directory) like so:
"
" autocmd User AfterBundles SomePluginCommand
"
" This lets you use commands defined in plugins, which haven't been created at
" .vimrc-time.
silent doautocmd User AfterBundles | <|file_sep|>original/after/plugin/after_bundles.vim
" You can make commands run after plugins are loaded from the .vimrc (and
" init/ directory) like so:
"
" autocmd User AfterBundles SomePluginCommand
"
" This lets you use commands defined in plugins, which haven't been created at
" .vimrc-time.
doautocmd User AfterBundles
<|file_sep|>current/after/plugin/after_bundles.vim
" You can make commands run after plugins are loaded from the .vimrc (and
" init/ directory) like so:
"
" autocmd User AfterBundles SomePluginCommand
"
" This lets you use commands defined in plugins, which haven't been created at
" .vimrc-time.
doautocmd User AfterBundles
<|file_sep|>updated/after/plugin/after_bundles.vim
" You can make commands run after plugins are loaded from the .vimrc (and
" init/ directory) like so:
"
" autocmd User AfterBundles SomePluginCommand
"
" This lets you use commands defined in plugins, which haven't been created at
" .vimrc-time.
silent doautocmd User AfterBundles | 1d07400c01dfe3e2c95acac20269c9f45f9842e0 | after/plugin/after_bundles.vim | after/plugin/after_bundles.vim | VimL |
<|file_sep|>original/scripts/import/AncestralAlleleData/ancestral_allele_table.sql
<|file_sep|>current/scripts/import/AncestralAlleleData/ancestral_allele_table.sql
<|file_sep|>updated/scripts/import/AncestralAlleleData/ancestral_allele_table.sql | CREATE TABLE ancestral_alleles (
ancestral_allele_id int(11) NOT NULL AUTO_INCREMENT,
variation_id int(10) unsigned NOT NULL,
ancestral_allele varchar(255) DEFAULT NULL,
PRIMARY KEY (ancestral_allele_id)
); | <|file_sep|>original/scripts/import/AncestralAlleleData/ancestral_allele_table.sql
<|file_sep|>current/scripts/import/AncestralAlleleData/ancestral_allele_table.sql
<|file_sep|>updated/scripts/import/AncestralAlleleData/ancestral_allele_table.sql
CREATE TABLE ancestral_alleles (
ancestral_allele_id int(11) NOT NULL AUTO_INCREMENT,
variation_id int(10) unsigned NOT NULL,
ancestral_allele varchar(255) DEFAULT NULL,
PRIMARY KEY (ancestral_allele_id)
); | fc6900e6206fc26d2060fb99fa723bb513e1f1c4 | scripts/import/AncestralAlleleData/ancestral_allele_table.sql | scripts/import/AncestralAlleleData/ancestral_allele_table.sql | SQL |
<|file_sep|>metadata/de.vier_bier.habpanelviewer.txt.diff
original:
updated:
Build:0.9.20,3
commit=0.9.20
subdir=app
gradle=yes
<|file_sep|>original/metadata/de.vier_bier.habpanelviewer.txt
Categories:Internet
License:GPL-3.0-only
Web Site:https://github.com/vbier/habpanelviewer
Source Code:https://github.com/vbier/habpanelviewer
Issue Tracker:https://github.com/vbier/habpanelviewer/issues
Auto Name:HABPanelViewer
Repo Type:git
Repo:https://github.com/vbier/habpanelviewer
Build:0.9.19,2
commit=0.9.19
subdir=app
gradle=yes
Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.9.19
Current Version Code:2
<|file_sep|>current/metadata/de.vier_bier.habpanelviewer.txt
Issue Tracker:https://github.com/vbier/habpanelviewer/issues
Auto Name:HABPanelViewer
Repo Type:git
Repo:https://github.com/vbier/habpanelviewer
Build:0.9.19,2
commit=0.9.19
subdir=app
gradle=yes
Build:0.9.20,3
commit=0.9.20
subdir=app
gradle=yes
Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.9.19
Current Version Code:2
<|file_sep|>updated/metadata/de.vier_bier.habpanelviewer.txt | Issue Tracker:https://github.com/vbier/habpanelviewer/issues
Auto Name:HABPanelViewer
Repo Type:git
Repo:https://github.com/vbier/habpanelviewer
Build:0.9.19,2
commit=0.9.19
subdir=app
gradle=yes
Build:0.9.20,3
commit=0.9.20
subdir=app
gradle=yes
Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.9.20
Current Version Code:3 | <|file_sep|>metadata/de.vier_bier.habpanelviewer.txt.diff
original:
updated:
Build:0.9.20,3
commit=0.9.20
subdir=app
gradle=yes
<|file_sep|>original/metadata/de.vier_bier.habpanelviewer.txt
Categories:Internet
License:GPL-3.0-only
Web Site:https://github.com/vbier/habpanelviewer
Source Code:https://github.com/vbier/habpanelviewer
Issue Tracker:https://github.com/vbier/habpanelviewer/issues
Auto Name:HABPanelViewer
Repo Type:git
Repo:https://github.com/vbier/habpanelviewer
Build:0.9.19,2
commit=0.9.19
subdir=app
gradle=yes
Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.9.19
Current Version Code:2
<|file_sep|>current/metadata/de.vier_bier.habpanelviewer.txt
Issue Tracker:https://github.com/vbier/habpanelviewer/issues
Auto Name:HABPanelViewer
Repo Type:git
Repo:https://github.com/vbier/habpanelviewer
Build:0.9.19,2
commit=0.9.19
subdir=app
gradle=yes
Build:0.9.20,3
commit=0.9.20
subdir=app
gradle=yes
Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.9.19
Current Version Code:2
<|file_sep|>updated/metadata/de.vier_bier.habpanelviewer.txt
Issue Tracker:https://github.com/vbier/habpanelviewer/issues
Auto Name:HABPanelViewer
Repo Type:git
Repo:https://github.com/vbier/habpanelviewer
Build:0.9.19,2
commit=0.9.19
subdir=app
gradle=yes
Build:0.9.20,3
commit=0.9.20
subdir=app
gradle=yes
Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.9.20
Current Version Code:3 | 085f112a32fbd2ac88f57d1d19d240e906877ec6 | metadata/de.vier_bier.habpanelviewer.txt | metadata/de.vier_bier.habpanelviewer.txt | Text |
<|file_sep|>original/simples3/gae.py
<|file_sep|>current/simples3/gae.py
<|file_sep|>updated/simples3/gae.py | """Compatibility layer for Google App Engine
Use as you would normally do with :mod:`simples3`, only instead of
:class:`simples3.S3Bucket`, use :class:`simples3.gae.AppEngineS3Bucket`.
"""
import urllib2
from StringIO import StringIO
from urllib import addinfourl
from google.appengine.api import urlfetch
from simples3.bucket import S3Bucket
class _FakeDict(list):
def iteritems(self):
return self
def _http_open(req):
resp = urlfetch.fetch(req.get_full_url(),
payload=req.get_data(),
method=req.get_method(),
headers=_FakeDict(req.header_items())) | <|file_sep|>original/simples3/gae.py
<|file_sep|>current/simples3/gae.py
<|file_sep|>updated/simples3/gae.py
"""Compatibility layer for Google App Engine
Use as you would normally do with :mod:`simples3`, only instead of
:class:`simples3.S3Bucket`, use :class:`simples3.gae.AppEngineS3Bucket`.
"""
import urllib2
from StringIO import StringIO
from urllib import addinfourl
from google.appengine.api import urlfetch
from simples3.bucket import S3Bucket
class _FakeDict(list):
def iteritems(self):
return self
def _http_open(req):
resp = urlfetch.fetch(req.get_full_url(),
payload=req.get_data(),
method=req.get_method(),
headers=_FakeDict(req.header_items())) | f22d1e7d39a282f12f92140b552c3d6751135ae1 | simples3/gae.py | simples3/gae.py | Python |
<|file_sep|>app/controllers/news_items_controller.rb.diff
original:
updated:
before_filter :load_networks, :only => [:index, :search]
<|file_sep|>app/controllers/news_items_controller.rb.diff
original:
@networks = Network.all
updated:
<|file_sep|>original/app/controllers/news_items_controller.rb
def sort_options
render :text => options_for_sorting(
params.fetch(:news_item_type, "news_items"),
params.fetch(:sort_by, "desc")
)
end
def search
get_news_items
render :action => 'index'
end
protected
def get_news_items
model_name = params[:news_item_type]
model_name = 'news_item' unless %w(tips pitches news_items).include?(model_name)
model = model_name.classify.constantize
@news_items = model.with_sort(params[:sort_by]).by_network(current_network).paginate(:page => params[:page])
end
end
<|file_sep|>current/app/controllers/news_items_controller.rb
def sort_options
render :text => options_for_sorting(
params.fetch(:news_item_type, "news_items"),
params.fetch(:sort_by, "desc")
)
end
def search
get_news_items
render :action => 'index'
end
protected
def get_news_items
model_name = params[:news_item_type]
model_name = 'news_item' unless %w(tips pitches news_items).include?(model_name)
model = model_name.classify.constantize
@news_items = model.with_sort(params[:sort_by]).by_network(current_network).paginate(:page => params[:page])
end
end
<|file_sep|>updated/app/controllers/news_items_controller.rb | )
end
def search
get_news_items
render :action => 'index'
end
protected
def get_news_items
model_name = params[:news_item_type]
model_name = 'news_item' unless %w(tips pitches news_items).include?(model_name)
model = model_name.classify.constantize
@news_items = model.with_sort(params[:sort_by]).by_network(current_network).paginate(:page => params[:page])
end
def load_networks
@networks = Network.all
end
end | <|file_sep|>app/controllers/news_items_controller.rb.diff
original:
updated:
before_filter :load_networks, :only => [:index, :search]
<|file_sep|>app/controllers/news_items_controller.rb.diff
original:
@networks = Network.all
updated:
<|file_sep|>original/app/controllers/news_items_controller.rb
def sort_options
render :text => options_for_sorting(
params.fetch(:news_item_type, "news_items"),
params.fetch(:sort_by, "desc")
)
end
def search
get_news_items
render :action => 'index'
end
protected
def get_news_items
model_name = params[:news_item_type]
model_name = 'news_item' unless %w(tips pitches news_items).include?(model_name)
model = model_name.classify.constantize
@news_items = model.with_sort(params[:sort_by]).by_network(current_network).paginate(:page => params[:page])
end
end
<|file_sep|>current/app/controllers/news_items_controller.rb
def sort_options
render :text => options_for_sorting(
params.fetch(:news_item_type, "news_items"),
params.fetch(:sort_by, "desc")
)
end
def search
get_news_items
render :action => 'index'
end
protected
def get_news_items
model_name = params[:news_item_type]
model_name = 'news_item' unless %w(tips pitches news_items).include?(model_name)
model = model_name.classify.constantize
@news_items = model.with_sort(params[:sort_by]).by_network(current_network).paginate(:page => params[:page])
end
end
<|file_sep|>updated/app/controllers/news_items_controller.rb
)
end
def search
get_news_items
render :action => 'index'
end
protected
def get_news_items
model_name = params[:news_item_type]
model_name = 'news_item' unless %w(tips pitches news_items).include?(model_name)
model = model_name.classify.constantize
@news_items = model.with_sort(params[:sort_by]).by_network(current_network).paginate(:page => params[:page])
end
def load_networks
@networks = Network.all
end
end | b245c80e6458915d1a18fc809145cb1ecf5080e7 | app/controllers/news_items_controller.rb | app/controllers/news_items_controller.rb | Ruby |
<|file_sep|>components/HeaderImg.js.diff
original:
updated:
const style = {
width: '98%',
height: '500px',
clear: 'both',
margin: '0 1% .5em',
text-align: 'center',
overflow: 'hidden'
};
const src = 'https://wordanddeedindia.imgix.net/images/child.jpg?fit=crop&crop=faces';
<|file_sep|>original/components/HeaderImg.js
import React from 'react';
class HeaderImg extends React.Component {
render() {
return (
<div className="header-img">
<img className="imgix-fluid" data-src="https://wordanddeedindia.imgix.net/images/child.jpg?fit=crop&crop=faces" />
</div>
);
}
}
export default HeaderImg;
<|file_sep|>current/components/HeaderImg.js
render() {
const style = {
width: '98%',
height: '500px',
clear: 'both',
margin: '0 1% .5em',
text-align: 'center',
overflow: 'hidden'
};
const src = 'https://wordanddeedindia.imgix.net/images/child.jpg?fit=crop&crop=faces';
return (
<div className="header-img">
<img className="imgix-fluid" data-src="https://wordanddeedindia.imgix.net/images/child.jpg?fit=crop&crop=faces" />
</div>
);
}
}
export default HeaderImg;
<|file_sep|>updated/components/HeaderImg.js | render() {
const style = {
width: '98%',
height: '500px',
clear: 'both',
margin: '0 1% .5em',
text-align: 'center',
overflow: 'hidden'
};
const src = 'https://wordanddeedindia.imgix.net/images/child.jpg?fit=crop&crop=faces';
return (
<div style={style}>
<img className="imgix-fluid" data-src={src} />
</div>
);
}
}
export default HeaderImg; | <|file_sep|>components/HeaderImg.js.diff
original:
updated:
const style = {
width: '98%',
height: '500px',
clear: 'both',
margin: '0 1% .5em',
text-align: 'center',
overflow: 'hidden'
};
const src = 'https://wordanddeedindia.imgix.net/images/child.jpg?fit=crop&crop=faces';
<|file_sep|>original/components/HeaderImg.js
import React from 'react';
class HeaderImg extends React.Component {
render() {
return (
<div className="header-img">
<img className="imgix-fluid" data-src="https://wordanddeedindia.imgix.net/images/child.jpg?fit=crop&crop=faces" />
</div>
);
}
}
export default HeaderImg;
<|file_sep|>current/components/HeaderImg.js
render() {
const style = {
width: '98%',
height: '500px',
clear: 'both',
margin: '0 1% .5em',
text-align: 'center',
overflow: 'hidden'
};
const src = 'https://wordanddeedindia.imgix.net/images/child.jpg?fit=crop&crop=faces';
return (
<div className="header-img">
<img className="imgix-fluid" data-src="https://wordanddeedindia.imgix.net/images/child.jpg?fit=crop&crop=faces" />
</div>
);
}
}
export default HeaderImg;
<|file_sep|>updated/components/HeaderImg.js
render() {
const style = {
width: '98%',
height: '500px',
clear: 'both',
margin: '0 1% .5em',
text-align: 'center',
overflow: 'hidden'
};
const src = 'https://wordanddeedindia.imgix.net/images/child.jpg?fit=crop&crop=faces';
return (
<div style={style}>
<img className="imgix-fluid" data-src={src} />
</div>
);
}
}
export default HeaderImg; | 4a57554050ef2e00188d6f952a8f7eb3891e710f | components/HeaderImg.js | components/HeaderImg.js | JavaScript |
<|file_sep|>original/.github/workflows/test.yml
uses: actions/setup-node@v2.1.2
with:
node-version: ${{ matrix.node-version }}
- name: install build dependencies
run: |
sudo apt-get install -y build-essential g++-4.8 libcairo2-dev libjpeg8-dev libpango1.0-dev libgif-dev
- name: npm install and test
run: |
npm install
npm test
env:
CI: true
merge-me:
name: Auto-merge dependency updates
needs: test
runs-on: ubuntu-latest
steps:
- uses: ridedott/merge-me-action@v1.8.3
with:
GITHUB_LOGIN: 'dependabot[bot]'
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
<|file_sep|>current/.github/workflows/test.yml
uses: actions/setup-node@v2.1.2
with:
node-version: ${{ matrix.node-version }}
- name: install build dependencies
run: |
sudo apt-get install -y build-essential g++-4.8 libcairo2-dev libjpeg8-dev libpango1.0-dev libgif-dev
- name: npm install and test
run: |
npm install
npm test
env:
CI: true
merge-me:
name: Auto-merge dependency updates
needs: test
runs-on: ubuntu-latest
steps:
- uses: ridedott/merge-me-action@v1.8.3
with:
GITHUB_LOGIN: 'dependabot[bot]'
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
<|file_sep|>updated/.github/workflows/test.yml | uses: actions/setup-node@v2.1.2
with:
node-version: ${{ matrix.node-version }}
- name: install build dependencies
run: |
sudo apt-get install -y build-essential g++-4.8 libcairo2-dev libjpeg8-dev libpango1.0-dev libgif-dev
- name: npm install and test
run: |
npm install
npm test
env:
CI: true
merge-me:
name: Auto-merge dependency updates
needs: test
runs-on: ubuntu-latest
steps:
- uses: ridedott/merge-me-action@v1.8.8
with:
GITHUB_LOGIN: 'dependabot[bot]'
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} | <|file_sep|>original/.github/workflows/test.yml
uses: actions/setup-node@v2.1.2
with:
node-version: ${{ matrix.node-version }}
- name: install build dependencies
run: |
sudo apt-get install -y build-essential g++-4.8 libcairo2-dev libjpeg8-dev libpango1.0-dev libgif-dev
- name: npm install and test
run: |
npm install
npm test
env:
CI: true
merge-me:
name: Auto-merge dependency updates
needs: test
runs-on: ubuntu-latest
steps:
- uses: ridedott/merge-me-action@v1.8.3
with:
GITHUB_LOGIN: 'dependabot[bot]'
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
<|file_sep|>current/.github/workflows/test.yml
uses: actions/setup-node@v2.1.2
with:
node-version: ${{ matrix.node-version }}
- name: install build dependencies
run: |
sudo apt-get install -y build-essential g++-4.8 libcairo2-dev libjpeg8-dev libpango1.0-dev libgif-dev
- name: npm install and test
run: |
npm install
npm test
env:
CI: true
merge-me:
name: Auto-merge dependency updates
needs: test
runs-on: ubuntu-latest
steps:
- uses: ridedott/merge-me-action@v1.8.3
with:
GITHUB_LOGIN: 'dependabot[bot]'
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
<|file_sep|>updated/.github/workflows/test.yml
uses: actions/setup-node@v2.1.2
with:
node-version: ${{ matrix.node-version }}
- name: install build dependencies
run: |
sudo apt-get install -y build-essential g++-4.8 libcairo2-dev libjpeg8-dev libpango1.0-dev libgif-dev
- name: npm install and test
run: |
npm install
npm test
env:
CI: true
merge-me:
name: Auto-merge dependency updates
needs: test
runs-on: ubuntu-latest
steps:
- uses: ridedott/merge-me-action@v1.8.8
with:
GITHUB_LOGIN: 'dependabot[bot]'
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} | a22e5d508f5db695b8e7d6799415923da8624abd | .github/workflows/test.yml | .github/workflows/test.yml | YAML |
<|file_sep|>original/event_store-entity_store.gemspec
# -*- encoding: utf-8 -*-
Gem::Specification.new do |s|
s.name = 'event_store-entity_store'
s.version = '0.1.1'
s.summary = 'Store of entities that are projected from EventStore streams'
s.description = ' '
s.authors = ['Obsidian Software, Inc']
s.email = 'opensource@obsidianexchange.com'
s.homepage = 'https://github.com/obsidian-btc/error_data'
s.licenses = ['MIT']
s.require_paths = ['lib']
s.files = Dir.glob('{lib}/**/*')
s.platform = Gem::Platform::RUBY
s.required_ruby_version = '>= 2.2.3'
s.add_runtime_dependency 'event_store-entity_projection', '~> 0'
s.add_development_dependency 'minitest'
s.add_development_dependency 'minitest-spec-context', '~> 0'
<|file_sep|>current/event_store-entity_store.gemspec
# -*- encoding: utf-8 -*-
Gem::Specification.new do |s|
s.name = 'event_store-entity_store'
s.version = '0.1.1'
s.summary = 'Store of entities that are projected from EventStore streams'
s.description = ' '
s.authors = ['Obsidian Software, Inc']
s.email = 'opensource@obsidianexchange.com'
s.homepage = 'https://github.com/obsidian-btc/error_data'
s.licenses = ['MIT']
s.require_paths = ['lib']
s.files = Dir.glob('{lib}/**/*')
s.platform = Gem::Platform::RUBY
s.required_ruby_version = '>= 2.2.3'
s.add_runtime_dependency 'event_store-entity_projection', '~> 0'
s.add_development_dependency 'minitest'
s.add_development_dependency 'minitest-spec-context', '~> 0'
<|file_sep|>updated/event_store-entity_store.gemspec | # -*- encoding: utf-8 -*-
Gem::Specification.new do |s|
s.name = 'event_store-entity_store'
s.version = '0.1.2'
s.summary = 'Store of entities that are projected from EventStore streams'
s.description = ' '
s.authors = ['Obsidian Software, Inc']
s.email = 'opensource@obsidianexchange.com'
s.homepage = 'https://github.com/obsidian-btc/error_data'
s.licenses = ['MIT']
s.require_paths = ['lib']
s.files = Dir.glob('{lib}/**/*')
s.platform = Gem::Platform::RUBY
s.required_ruby_version = '>= 2.2.3'
s.add_runtime_dependency 'event_store-entity_projection', '~> 0'
s.add_development_dependency 'minitest'
s.add_development_dependency 'minitest-spec-context', '~> 0' | <|file_sep|>original/event_store-entity_store.gemspec
# -*- encoding: utf-8 -*-
Gem::Specification.new do |s|
s.name = 'event_store-entity_store'
s.version = '0.1.1'
s.summary = 'Store of entities that are projected from EventStore streams'
s.description = ' '
s.authors = ['Obsidian Software, Inc']
s.email = 'opensource@obsidianexchange.com'
s.homepage = 'https://github.com/obsidian-btc/error_data'
s.licenses = ['MIT']
s.require_paths = ['lib']
s.files = Dir.glob('{lib}/**/*')
s.platform = Gem::Platform::RUBY
s.required_ruby_version = '>= 2.2.3'
s.add_runtime_dependency 'event_store-entity_projection', '~> 0'
s.add_development_dependency 'minitest'
s.add_development_dependency 'minitest-spec-context', '~> 0'
<|file_sep|>current/event_store-entity_store.gemspec
# -*- encoding: utf-8 -*-
Gem::Specification.new do |s|
s.name = 'event_store-entity_store'
s.version = '0.1.1'
s.summary = 'Store of entities that are projected from EventStore streams'
s.description = ' '
s.authors = ['Obsidian Software, Inc']
s.email = 'opensource@obsidianexchange.com'
s.homepage = 'https://github.com/obsidian-btc/error_data'
s.licenses = ['MIT']
s.require_paths = ['lib']
s.files = Dir.glob('{lib}/**/*')
s.platform = Gem::Platform::RUBY
s.required_ruby_version = '>= 2.2.3'
s.add_runtime_dependency 'event_store-entity_projection', '~> 0'
s.add_development_dependency 'minitest'
s.add_development_dependency 'minitest-spec-context', '~> 0'
<|file_sep|>updated/event_store-entity_store.gemspec
# -*- encoding: utf-8 -*-
Gem::Specification.new do |s|
s.name = 'event_store-entity_store'
s.version = '0.1.2'
s.summary = 'Store of entities that are projected from EventStore streams'
s.description = ' '
s.authors = ['Obsidian Software, Inc']
s.email = 'opensource@obsidianexchange.com'
s.homepage = 'https://github.com/obsidian-btc/error_data'
s.licenses = ['MIT']
s.require_paths = ['lib']
s.files = Dir.glob('{lib}/**/*')
s.platform = Gem::Platform::RUBY
s.required_ruby_version = '>= 2.2.3'
s.add_runtime_dependency 'event_store-entity_projection', '~> 0'
s.add_development_dependency 'minitest'
s.add_development_dependency 'minitest-spec-context', '~> 0' | 0b2ea65e1c94fcf0926336c46bfe77dd86e6117d | event_store-entity_store.gemspec | event_store-entity_store.gemspec | Ruby |
<|file_sep|>osu.Game/Skinning/LegacySkinDecoder.cs.diff
original:
public class LegacySkinDecoder : LegacyDecoder<SkinConfiguration>
updated:
public class LegacySkinDecoder : LegacyDecoder<DefaultSkinConfiguration>
<|file_sep|>original/osu.Game/Skinning/LegacySkinDecoder.cs
namespace osu.Game.Skinning
{
public class LegacySkinDecoder : LegacyDecoder<SkinConfiguration>
{
public LegacySkinDecoder()
: base(1)
{
}
protected override void ParseLine(SkinConfiguration skin, Section section, string line)
{
line = StripComments(line);
var pair = SplitKeyVal(line);
switch (section)
{
case Section.General:
switch (pair.Key)
{
<|file_sep|>current/osu.Game/Skinning/LegacySkinDecoder.cs
namespace osu.Game.Skinning
{
public class LegacySkinDecoder : LegacyDecoder<DefaultSkinConfiguration>
{
public LegacySkinDecoder()
: base(1)
{
}
protected override void ParseLine(SkinConfiguration skin, Section section, string line)
{
line = StripComments(line);
var pair = SplitKeyVal(line);
switch (section)
{
case Section.General:
switch (pair.Key)
{
<|file_sep|>updated/osu.Game/Skinning/LegacySkinDecoder.cs |
namespace osu.Game.Skinning
{
public class LegacySkinDecoder : LegacyDecoder<DefaultSkinConfiguration>
{
public LegacySkinDecoder()
: base(1)
{
}
protected override void ParseLine(DefaultSkinConfiguration skin, Section section, string line)
{
line = StripComments(line);
var pair = SplitKeyVal(line);
switch (section)
{
case Section.General:
switch (pair.Key)
{ | <|file_sep|>osu.Game/Skinning/LegacySkinDecoder.cs.diff
original:
public class LegacySkinDecoder : LegacyDecoder<SkinConfiguration>
updated:
public class LegacySkinDecoder : LegacyDecoder<DefaultSkinConfiguration>
<|file_sep|>original/osu.Game/Skinning/LegacySkinDecoder.cs
namespace osu.Game.Skinning
{
public class LegacySkinDecoder : LegacyDecoder<SkinConfiguration>
{
public LegacySkinDecoder()
: base(1)
{
}
protected override void ParseLine(SkinConfiguration skin, Section section, string line)
{
line = StripComments(line);
var pair = SplitKeyVal(line);
switch (section)
{
case Section.General:
switch (pair.Key)
{
<|file_sep|>current/osu.Game/Skinning/LegacySkinDecoder.cs
namespace osu.Game.Skinning
{
public class LegacySkinDecoder : LegacyDecoder<DefaultSkinConfiguration>
{
public LegacySkinDecoder()
: base(1)
{
}
protected override void ParseLine(SkinConfiguration skin, Section section, string line)
{
line = StripComments(line);
var pair = SplitKeyVal(line);
switch (section)
{
case Section.General:
switch (pair.Key)
{
<|file_sep|>updated/osu.Game/Skinning/LegacySkinDecoder.cs
namespace osu.Game.Skinning
{
public class LegacySkinDecoder : LegacyDecoder<DefaultSkinConfiguration>
{
public LegacySkinDecoder()
: base(1)
{
}
protected override void ParseLine(DefaultSkinConfiguration skin, Section section, string line)
{
line = StripComments(line);
var pair = SplitKeyVal(line);
switch (section)
{
case Section.General:
switch (pair.Key)
{ | d2a3e0581b750c030d197573e6923815ca2e60f8 | osu.Game/Skinning/LegacySkinDecoder.cs | osu.Game/Skinning/LegacySkinDecoder.cs | C# |
<|file_sep|>original/src/items.h
<|file_sep|>current/src/items.h
<|file_sep|>updated/src/items.h | /*------------------------------------------------------------------------------
| NuCTex | items.h
| Author | Benjamin A - Nullsrc
| Created | 17 January, 2016
| Changed | 17 January, 2016
|-------------------------------------------------------------------------------
| Overview | Declare item structures used in the code
\-----------------------------------------------------------------------------*/
#ifndef NULLSRC_ITEMS_HEADER
#define NULLSRC_ITEMS_HEADER
typedef struct Item {
char* name;
char* description;
int id;
float size;
float weight;
int strength;
int agility;
int intelligence; | <|file_sep|>original/src/items.h
<|file_sep|>current/src/items.h
<|file_sep|>updated/src/items.h
/*------------------------------------------------------------------------------
| NuCTex | items.h
| Author | Benjamin A - Nullsrc
| Created | 17 January, 2016
| Changed | 17 January, 2016
|-------------------------------------------------------------------------------
| Overview | Declare item structures used in the code
\-----------------------------------------------------------------------------*/
#ifndef NULLSRC_ITEMS_HEADER
#define NULLSRC_ITEMS_HEADER
typedef struct Item {
char* name;
char* description;
int id;
float size;
float weight;
int strength;
int agility;
int intelligence; | 0c4431765bbcdc19826068ce34c2de2b2fca22d5 | src/items.h | src/items.h | C |
<|file_sep|>recipes/nested-lookup/meta.yaml.diff
original:
- python
updated:
- python >=3.5
<|file_sep|>original/recipes/nested-lookup/meta.yaml
build:
noarch: python
number: 0
script: "{{ PYTHON }} -m pip install . -vv"
requirements:
host:
- python
- pip
run:
- python
- six
test:
imports:
- nested_lookup
about:
home: https://github.com/russellballestrini/nested-lookup
# NOTE : The author of nested-lookup releases the code to the public domain,
# so the source is licensed as MIT for the purposes of use in Conda Forge.
<|file_sep|>current/recipes/nested-lookup/meta.yaml
build:
noarch: python
number: 0
script: "{{ PYTHON }} -m pip install . -vv"
requirements:
host:
- python >=3.5
- pip
run:
- python
- six
test:
imports:
- nested_lookup
about:
home: https://github.com/russellballestrini/nested-lookup
# NOTE : The author of nested-lookup releases the code to the public domain,
# so the source is licensed as MIT for the purposes of use in Conda Forge.
<|file_sep|>updated/recipes/nested-lookup/meta.yaml | build:
noarch: python
number: 0
script: "{{ PYTHON }} -m pip install . -vv"
requirements:
host:
- python >=3.5
- pip
run:
- python >=3.5
- six
test:
imports:
- nested_lookup
about:
home: https://github.com/russellballestrini/nested-lookup
# NOTE : The author of nested-lookup releases the code to the public domain,
# so the source is licensed as MIT for the purposes of use in Conda Forge. | <|file_sep|>recipes/nested-lookup/meta.yaml.diff
original:
- python
updated:
- python >=3.5
<|file_sep|>original/recipes/nested-lookup/meta.yaml
build:
noarch: python
number: 0
script: "{{ PYTHON }} -m pip install . -vv"
requirements:
host:
- python
- pip
run:
- python
- six
test:
imports:
- nested_lookup
about:
home: https://github.com/russellballestrini/nested-lookup
# NOTE : The author of nested-lookup releases the code to the public domain,
# so the source is licensed as MIT for the purposes of use in Conda Forge.
<|file_sep|>current/recipes/nested-lookup/meta.yaml
build:
noarch: python
number: 0
script: "{{ PYTHON }} -m pip install . -vv"
requirements:
host:
- python >=3.5
- pip
run:
- python
- six
test:
imports:
- nested_lookup
about:
home: https://github.com/russellballestrini/nested-lookup
# NOTE : The author of nested-lookup releases the code to the public domain,
# so the source is licensed as MIT for the purposes of use in Conda Forge.
<|file_sep|>updated/recipes/nested-lookup/meta.yaml
build:
noarch: python
number: 0
script: "{{ PYTHON }} -m pip install . -vv"
requirements:
host:
- python >=3.5
- pip
run:
- python >=3.5
- six
test:
imports:
- nested_lookup
about:
home: https://github.com/russellballestrini/nested-lookup
# NOTE : The author of nested-lookup releases the code to the public domain,
# so the source is licensed as MIT for the purposes of use in Conda Forge. | f2c8691af1b48eb4b5fcf72d035e3a97d7333da8 | recipes/nested-lookup/meta.yaml | recipes/nested-lookup/meta.yaml | YAML |
<|file_sep|>original/requirements-dev.txt
flake8==3.3.0
pytest==3.0.6
pytest-cov==2.4.0
pytest-catchlog==1.2.2
docker-py==1.10.6
coveralls==1.1
Sphinx==1.5.2
lz4tools==1.3.1.2
xxhash==1.0.0
sphinxcontrib-asyncio>=0.2.0
sphinx-rtd-theme==0.1.9
diff-cover==0.9.9
setuptools==34.2.0
<|file_sep|>current/requirements-dev.txt
flake8==3.3.0
pytest==3.0.6
pytest-cov==2.4.0
pytest-catchlog==1.2.2
docker-py==1.10.6
coveralls==1.1
Sphinx==1.5.2
lz4tools==1.3.1.2
xxhash==1.0.0
sphinxcontrib-asyncio>=0.2.0
sphinx-rtd-theme==0.1.9
diff-cover==0.9.9
setuptools==34.2.0
<|file_sep|>updated/requirements-dev.txt | flake8==3.3.0
pytest==3.0.6
pytest-cov==2.4.0
pytest-catchlog==1.2.2
docker-py==1.10.6
coveralls==1.1
Sphinx==1.5.2
lz4tools==1.3.1.2
xxhash==1.0.0
sphinxcontrib-asyncio>=0.2.0
sphinx-rtd-theme==0.1.9
diff-cover==0.9.9
setuptools==34.3.0 | <|file_sep|>original/requirements-dev.txt
flake8==3.3.0
pytest==3.0.6
pytest-cov==2.4.0
pytest-catchlog==1.2.2
docker-py==1.10.6
coveralls==1.1
Sphinx==1.5.2
lz4tools==1.3.1.2
xxhash==1.0.0
sphinxcontrib-asyncio>=0.2.0
sphinx-rtd-theme==0.1.9
diff-cover==0.9.9
setuptools==34.2.0
<|file_sep|>current/requirements-dev.txt
flake8==3.3.0
pytest==3.0.6
pytest-cov==2.4.0
pytest-catchlog==1.2.2
docker-py==1.10.6
coveralls==1.1
Sphinx==1.5.2
lz4tools==1.3.1.2
xxhash==1.0.0
sphinxcontrib-asyncio>=0.2.0
sphinx-rtd-theme==0.1.9
diff-cover==0.9.9
setuptools==34.2.0
<|file_sep|>updated/requirements-dev.txt
flake8==3.3.0
pytest==3.0.6
pytest-cov==2.4.0
pytest-catchlog==1.2.2
docker-py==1.10.6
coveralls==1.1
Sphinx==1.5.2
lz4tools==1.3.1.2
xxhash==1.0.0
sphinxcontrib-asyncio>=0.2.0
sphinx-rtd-theme==0.1.9
diff-cover==0.9.9
setuptools==34.3.0 | 81a1be9b4011d0e52bcc5ba6f37c23bddd0f5f16 | requirements-dev.txt | requirements-dev.txt | Text |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.