prompt stringlengths 77 6.9k | completion stringlengths 1 3.38k | text stringlengths 163 9.03k | commit stringlengths 40 40 | old_file stringlengths 4 264 | new_file stringlengths 4 264 | lang stringclasses 277
values |
|---|---|---|---|---|---|---|
<|file_sep|>original/.github/workflows/visual-regression.yml
name: Visual Regresion Testing
on:
pull_request:
types:
- opened
- reopened
- synchronize
- ready_for_review
jobs:
visual-regression:
name: Run Tests
runs-on: ubuntu-latest
if: github.event.pull_request.draft == false
steps:
- uses: actions/checkout@v2
- name: Read .nvmrc
run: echo "##[set-output name=NVMRC;]$(cat .nvmrc)"
id: nvm
- name: Setup Node.js (.nvmrc)
<|file_sep|>current/.github/workflows/visual-regression.yml
name: Visual Regresion Testing
on:
pull_request:
types:
- opened
- reopened
- synchronize
- ready_for_review
jobs:
visual-regression:
name: Run Tests
runs-on: ubuntu-latest
if: github.event.pull_request.draft == false
steps:
- uses: actions/checkout@v2
- name: Read .nvmrc
run: echo "##[set-output name=NVMRC;]$(cat .nvmrc)"
id: nvm
- name: Setup Node.js (.nvmrc)
<|file_sep|>updated/.github/workflows/visual-regression.yml | name: Visual Regresion Testing
on:
pull_request:
branches:
- master
- develop
types:
- opened
- reopened
- synchronize
- ready_for_review
jobs:
visual-regression:
name: Run Tests
runs-on: ubuntu-latest
if: github.event.pull_request.draft == false
steps:
- uses: actions/checkout@v2
- name: Read .nvmrc | <|file_sep|>original/.github/workflows/visual-regression.yml
name: Visual Regresion Testing
on:
pull_request:
types:
- opened
- reopened
- synchronize
- ready_for_review
jobs:
visual-regression:
name: Run Tests
runs-on: ubuntu-latest
if: github.event.pull_request.draft == false
steps:
- uses: actions/checkout@v2
- name: Read .nvmrc
run: echo "##[set-output name=NVMRC;]$(cat .nvmrc)"
id: nvm
- name: Setup Node.js (.nvmrc)
<|file_sep|>current/.github/workflows/visual-regression.yml
name: Visual Regresion Testing
on:
pull_request:
types:
- opened
- reopened
- synchronize
- ready_for_review
jobs:
visual-regression:
name: Run Tests
runs-on: ubuntu-latest
if: github.event.pull_request.draft == false
steps:
- uses: actions/checkout@v2
- name: Read .nvmrc
run: echo "##[set-output name=NVMRC;]$(cat .nvmrc)"
id: nvm
- name: Setup Node.js (.nvmrc)
<|file_sep|>updated/.github/workflows/visual-regression.yml
name: Visual Regresion Testing
on:
pull_request:
branches:
- master
- develop
types:
- opened
- reopened
- synchronize
- ready_for_review
jobs:
visual-regression:
name: Run Tests
runs-on: ubuntu-latest
if: github.event.pull_request.draft == false
steps:
- uses: actions/checkout@v2
- name: Read .nvmrc | ed0519ad6603ee7c6c93babf0df97e457554aea0 | .github/workflows/visual-regression.yml | .github/workflows/visual-regression.yml | YAML |
<|file_sep|>original/client/app/helpers.js
import 'whatwg-fetch';
var helpers = {
checkStatus,
parseJSON
};
function checkStatus(res) {
if (res.status >= 200 && res.status < 300) {
return res;
} else {
var error = new Error(res.statusText);
error.res = res;
throw error;
}
}
function parseJSON(res) {
return res.json();
}
<|file_sep|>current/client/app/helpers.js
import 'whatwg-fetch';
var helpers = {
checkStatus,
parseJSON
};
function checkStatus(res) {
if (res.status >= 200 && res.status < 300) {
return res;
} else {
var error = new Error(res.statusText);
error.res = res;
throw error;
}
}
function parseJSON(res) {
return res.json();
}
<|file_sep|>updated/client/app/helpers.js | import 'whatwg-fetch';
var helpers = {
checkStatus,
parseJSON,
headers: {
'Accept': 'application/json',
'Content-Type': 'application/json'
}
};
function checkStatus(res) {
if (res.status >= 200 && res.status < 300) {
return res;
} else {
var error = new Error(res.statusText);
error.res = res;
throw error;
}
}
| <|file_sep|>original/client/app/helpers.js
import 'whatwg-fetch';
var helpers = {
checkStatus,
parseJSON
};
function checkStatus(res) {
if (res.status >= 200 && res.status < 300) {
return res;
} else {
var error = new Error(res.statusText);
error.res = res;
throw error;
}
}
function parseJSON(res) {
return res.json();
}
<|file_sep|>current/client/app/helpers.js
import 'whatwg-fetch';
var helpers = {
checkStatus,
parseJSON
};
function checkStatus(res) {
if (res.status >= 200 && res.status < 300) {
return res;
} else {
var error = new Error(res.statusText);
error.res = res;
throw error;
}
}
function parseJSON(res) {
return res.json();
}
<|file_sep|>updated/client/app/helpers.js
import 'whatwg-fetch';
var helpers = {
checkStatus,
parseJSON,
headers: {
'Accept': 'application/json',
'Content-Type': 'application/json'
}
};
function checkStatus(res) {
if (res.status >= 200 && res.status < 300) {
return res;
} else {
var error = new Error(res.statusText);
error.res = res;
throw error;
}
}
| 21fb9d7af26c22999b283c1e18cda29fdd33048d | client/app/helpers.js | client/app/helpers.js | JavaScript |
<|file_sep|>original/lib/gyro/liquidgen/infos.rb
module LiquidGen
module Infos
def self.listTemplates
Gyro.templates_dir.children.select(&:directory?).each do |entry|
puts " - #{entry.basename}"
end
end
def self.showInfos(template)
if template.include?('/')
readme = template + 'README.md' unless readme.exist?
else
readme = Gyro.templates_dir + template + 'README.md'
end
Gyro::Error.exit_with_error("No README.md found for template #{template}.") unless readme.exist?
puts readme.read
end
end
end
end
<|file_sep|>current/lib/gyro/liquidgen/infos.rb
module LiquidGen
module Infos
def self.listTemplates
Gyro.templates_dir.children.select(&:directory?).each do |entry|
puts " - #{entry.basename}"
end
end
def self.showInfos(template)
if template.include?('/')
readme = template + 'README.md' unless readme.exist?
else
readme = Gyro.templates_dir + template + 'README.md'
end
Gyro::Error.exit_with_error("No README.md found for template #{template}.") unless readme.exist?
puts readme.read
end
end
end
end
<|file_sep|>updated/lib/gyro/liquidgen/infos.rb | module LiquidGen
module Infos
def self.listTemplates
Gyro.templates_dir.children.select(&:directory?).each do |entry|
puts " - #{entry.basename}"
end
end
def self.showInfos(template)
if template.include?('/')
readme = Pathname.new(template) + 'README.md'
else
readme = Gyro.templates_dir + template + 'README.md'
end
Gyro::Error.exit_with_error("No README.md found for template #{template}.") unless readme.exist?
puts readme.read
end
end
end
end | <|file_sep|>original/lib/gyro/liquidgen/infos.rb
module LiquidGen
module Infos
def self.listTemplates
Gyro.templates_dir.children.select(&:directory?).each do |entry|
puts " - #{entry.basename}"
end
end
def self.showInfos(template)
if template.include?('/')
readme = template + 'README.md' unless readme.exist?
else
readme = Gyro.templates_dir + template + 'README.md'
end
Gyro::Error.exit_with_error("No README.md found for template #{template}.") unless readme.exist?
puts readme.read
end
end
end
end
<|file_sep|>current/lib/gyro/liquidgen/infos.rb
module LiquidGen
module Infos
def self.listTemplates
Gyro.templates_dir.children.select(&:directory?).each do |entry|
puts " - #{entry.basename}"
end
end
def self.showInfos(template)
if template.include?('/')
readme = template + 'README.md' unless readme.exist?
else
readme = Gyro.templates_dir + template + 'README.md'
end
Gyro::Error.exit_with_error("No README.md found for template #{template}.") unless readme.exist?
puts readme.read
end
end
end
end
<|file_sep|>updated/lib/gyro/liquidgen/infos.rb
module LiquidGen
module Infos
def self.listTemplates
Gyro.templates_dir.children.select(&:directory?).each do |entry|
puts " - #{entry.basename}"
end
end
def self.showInfos(template)
if template.include?('/')
readme = Pathname.new(template) + 'README.md'
else
readme = Gyro.templates_dir + template + 'README.md'
end
Gyro::Error.exit_with_error("No README.md found for template #{template}.") unless readme.exist?
puts readme.read
end
end
end
end | a2460f66d026773615553c4a4173197516b4137c | lib/gyro/liquidgen/infos.rb | lib/gyro/liquidgen/infos.rb | Ruby |
<|file_sep|>app/client/src/dl-tools/components/repository/repo.tpl.html.diff
original:
<p>
<code>{{ project.ssh_url_to_repo }}</code>
<a class="pull-right" href="{{ project.web_url }}">Open on Gitlab</a>
updated:
<p class="pull-right">
<code>{{ project.ssh_url_to_repo }}</code> |
<a href="{{ project.web_url }}">Open on Gitlab</a>
<|file_sep|>original/app/client/src/dl-tools/components/repository/repo.tpl.html
<span loading-indicator data-flag="!project" data-type="inline"></span>
{{ project.name_with_namespace }}
</li>
</ol>
<div ng-if="project">
<h3>{{ project.name_with_namespace }}</h3>
<p>
<code>{{ project.ssh_url_to_repo }}</code>
<a class="pull-right" href="{{ project.web_url }}">Open on Gitlab</a>
</p>
<h4>Manage Branches</h4>
<repository-branches-management project="project" />
</div>
<div loading-indicator data-flag="!project" data-type="big"></div>
</div>
<|file_sep|>current/app/client/src/dl-tools/components/repository/repo.tpl.html
<span loading-indicator data-flag="!project" data-type="inline"></span>
{{ project.name_with_namespace }}
</li>
</ol>
<div ng-if="project">
<h3>{{ project.name_with_namespace }}</h3>
<p class="pull-right">
<code>{{ project.ssh_url_to_repo }}</code> |
<a href="{{ project.web_url }}">Open on Gitlab</a>
</p>
<h4>Manage Branches</h4>
<repository-branches-management project="project" />
</div>
<div loading-indicator data-flag="!project" data-type="big"></div>
</div>
<|file_sep|>updated/app/client/src/dl-tools/components/repository/repo.tpl.html | {{ project.name_with_namespace }}
</li>
</ol>
<div ng-if="project">
<h3>{{ project.name_with_namespace }}</h3>
<p class="pull-right">
<code>{{ project.ssh_url_to_repo }}</code> |
<a href="{{ project.web_url }}">Open on Gitlab</a>
</p>
<h4>Manage Branches</h4>
<repository-branches-management project="project" />
</div>
<div loading-indicator data-flag="!project" data-type="big"></div>
</div> | <|file_sep|>app/client/src/dl-tools/components/repository/repo.tpl.html.diff
original:
<p>
<code>{{ project.ssh_url_to_repo }}</code>
<a class="pull-right" href="{{ project.web_url }}">Open on Gitlab</a>
updated:
<p class="pull-right">
<code>{{ project.ssh_url_to_repo }}</code> |
<a href="{{ project.web_url }}">Open on Gitlab</a>
<|file_sep|>original/app/client/src/dl-tools/components/repository/repo.tpl.html
<span loading-indicator data-flag="!project" data-type="inline"></span>
{{ project.name_with_namespace }}
</li>
</ol>
<div ng-if="project">
<h3>{{ project.name_with_namespace }}</h3>
<p>
<code>{{ project.ssh_url_to_repo }}</code>
<a class="pull-right" href="{{ project.web_url }}">Open on Gitlab</a>
</p>
<h4>Manage Branches</h4>
<repository-branches-management project="project" />
</div>
<div loading-indicator data-flag="!project" data-type="big"></div>
</div>
<|file_sep|>current/app/client/src/dl-tools/components/repository/repo.tpl.html
<span loading-indicator data-flag="!project" data-type="inline"></span>
{{ project.name_with_namespace }}
</li>
</ol>
<div ng-if="project">
<h3>{{ project.name_with_namespace }}</h3>
<p class="pull-right">
<code>{{ project.ssh_url_to_repo }}</code> |
<a href="{{ project.web_url }}">Open on Gitlab</a>
</p>
<h4>Manage Branches</h4>
<repository-branches-management project="project" />
</div>
<div loading-indicator data-flag="!project" data-type="big"></div>
</div>
<|file_sep|>updated/app/client/src/dl-tools/components/repository/repo.tpl.html
{{ project.name_with_namespace }}
</li>
</ol>
<div ng-if="project">
<h3>{{ project.name_with_namespace }}</h3>
<p class="pull-right">
<code>{{ project.ssh_url_to_repo }}</code> |
<a href="{{ project.web_url }}">Open on Gitlab</a>
</p>
<h4>Manage Branches</h4>
<repository-branches-management project="project" />
</div>
<div loading-indicator data-flag="!project" data-type="big"></div>
</div> | 9388ffb93d677b55ec37978a1ed0477bb5c15b9d | app/client/src/dl-tools/components/repository/repo.tpl.html | app/client/src/dl-tools/components/repository/repo.tpl.html | HTML |
<|file_sep|>original/resources/views/root/dashboard.blade.php
@extends('root.app')
@section('content')
<div class="container-fluid">
<h1>Registered Users</h1>
{!! Table::withContents($users->toArray())->striped()->condensed()->hover() !!}
</div>
@endsection
<|file_sep|>current/resources/views/root/dashboard.blade.php
@extends('root.app')
@section('content')
<div class="container-fluid">
<h1>Registered Users</h1>
{!! Table::withContents($users->toArray())->striped()->condensed()->hover() !!}
</div>
@endsection
<|file_sep|>updated/resources/views/root/dashboard.blade.php | @extends('layouts.user')
@section('content')
<div class="container-fluid">
<h1>Registered Users</h1>
{!! Table::withContents($users->toArray())->striped()->condensed()->hover() !!}
</div>
@endsection | <|file_sep|>original/resources/views/root/dashboard.blade.php
@extends('root.app')
@section('content')
<div class="container-fluid">
<h1>Registered Users</h1>
{!! Table::withContents($users->toArray())->striped()->condensed()->hover() !!}
</div>
@endsection
<|file_sep|>current/resources/views/root/dashboard.blade.php
@extends('root.app')
@section('content')
<div class="container-fluid">
<h1>Registered Users</h1>
{!! Table::withContents($users->toArray())->striped()->condensed()->hover() !!}
</div>
@endsection
<|file_sep|>updated/resources/views/root/dashboard.blade.php
@extends('layouts.user')
@section('content')
<div class="container-fluid">
<h1>Registered Users</h1>
{!! Table::withContents($users->toArray())->striped()->condensed()->hover() !!}
</div>
@endsection | fd5b128784fb905d9bc3728b2c0ce016bcafb1f2 | resources/views/root/dashboard.blade.php | resources/views/root/dashboard.blade.php | PHP |
<|file_sep|>byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html.diff
original:
<tr>
updated:
{%- with is_available = (item.article.quantity > 0) %}
<tr{% if not is_available %} class="dimmed"{% endif %}>
<|file_sep|>byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html.diff
original:
{{ field() }}
updated:
{{ field(hidden=not is_available) }}
<|file_sep|>byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html.diff
original:
<td><strong>{{ item.article.description }}</strong></td>
updated:
<td><strong>{{ item.article.description }}</strong>
{%- if not is_available %}
<br><em>derzeit nicht verfügbar</em>
{%- endif -%}
</td>
<|file_sep|>original/byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html
<td class="number">{{ item.fixed_quantity }}</td>
{%- else %}
<td class="number">
{%- set field = form.get_field_for_article(item.article) %}
{{ field() }}
{%- if field.errors %}
<ol class="errors" style="text-align: left;">
{%- for error in field.errors %}
<li><strong>Fehler:</strong> <span>{{ error }}</span></li>
{%- endfor %}
</ol>
{%- endif -%}
</td>
{%- endif %}
<td><strong>{{ item.article.description }}</strong></td>
<td class="number">{{ item.article.tax_rate_as_percentage }} %</td>
<td class="number">{{ item.article.price|format_euro_amount }}</td>
</tr>
{%- endfor %}
</tbody>
</table>
<|file_sep|>current/byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html
{{ field(hidden=not is_available) }}
{%- if field.errors %}
<ol class="errors" style="text-align: left;">
{%- for error in field.errors %}
<li><strong>Fehler:</strong> <span>{{ error }}</span></li>
{%- endfor %}
</ol>
{%- endif -%}
</td>
{%- endif %}
<td><strong>{{ item.article.description }}</strong>
{%- if not is_available %}
<br><em>derzeit nicht verfügbar</em>
{%- endif -%}
</td>
<td class="number">{{ item.article.tax_rate_as_percentage }} %</td>
<td class="number">{{ item.article.price|format_euro_amount }}</td>
</tr>
{%- endfor %}
</tbody>
</table>
<|file_sep|>updated/byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html | {%- if field.errors %}
<ol class="errors" style="text-align: left;">
{%- for error in field.errors %}
<li><strong>Fehler:</strong> <span>{{ error }}</span></li>
{%- endfor %}
</ol>
{%- endif -%}
</td>
{%- endif %}
<td><strong>{{ item.article.description }}</strong>
{%- if not is_available %}
<br><em>derzeit nicht verfügbar</em>
{%- endif -%}
</td>
<td class="number">{{ item.article.tax_rate_as_percentage }} %</td>
<td class="number">{{ item.article.price|format_euro_amount }}</td>
</tr>
{%- endwith %}
{%- endfor %}
</tbody>
</table> | <|file_sep|>byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html.diff
original:
<tr>
updated:
{%- with is_available = (item.article.quantity > 0) %}
<tr{% if not is_available %} class="dimmed"{% endif %}>
<|file_sep|>byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html.diff
original:
{{ field() }}
updated:
{{ field(hidden=not is_available) }}
<|file_sep|>byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html.diff
original:
<td><strong>{{ item.article.description }}</strong></td>
updated:
<td><strong>{{ item.article.description }}</strong>
{%- if not is_available %}
<br><em>derzeit nicht verfügbar</em>
{%- endif -%}
</td>
<|file_sep|>original/byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html
<td class="number">{{ item.fixed_quantity }}</td>
{%- else %}
<td class="number">
{%- set field = form.get_field_for_article(item.article) %}
{{ field() }}
{%- if field.errors %}
<ol class="errors" style="text-align: left;">
{%- for error in field.errors %}
<li><strong>Fehler:</strong> <span>{{ error }}</span></li>
{%- endfor %}
</ol>
{%- endif -%}
</td>
{%- endif %}
<td><strong>{{ item.article.description }}</strong></td>
<td class="number">{{ item.article.tax_rate_as_percentage }} %</td>
<td class="number">{{ item.article.price|format_euro_amount }}</td>
</tr>
{%- endfor %}
</tbody>
</table>
<|file_sep|>current/byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html
{{ field(hidden=not is_available) }}
{%- if field.errors %}
<ol class="errors" style="text-align: left;">
{%- for error in field.errors %}
<li><strong>Fehler:</strong> <span>{{ error }}</span></li>
{%- endfor %}
</ol>
{%- endif -%}
</td>
{%- endif %}
<td><strong>{{ item.article.description }}</strong>
{%- if not is_available %}
<br><em>derzeit nicht verfügbar</em>
{%- endif -%}
</td>
<td class="number">{{ item.article.tax_rate_as_percentage }} %</td>
<td class="number">{{ item.article.price|format_euro_amount }}</td>
</tr>
{%- endfor %}
</tbody>
</table>
<|file_sep|>updated/byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html
{%- if field.errors %}
<ol class="errors" style="text-align: left;">
{%- for error in field.errors %}
<li><strong>Fehler:</strong> <span>{{ error }}</span></li>
{%- endfor %}
</ol>
{%- endif -%}
</td>
{%- endif %}
<td><strong>{{ item.article.description }}</strong>
{%- if not is_available %}
<br><em>derzeit nicht verfügbar</em>
{%- endif -%}
</td>
<td class="number">{{ item.article.tax_rate_as_percentage }} %</td>
<td class="number">{{ item.article.price|format_euro_amount }}</td>
</tr>
{%- endwith %}
{%- endfor %}
</tbody>
</table> | a2fd89dcda80a389006389b4b80f62d55a728907 | byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html | byceps/blueprints/shop_order/templates/shop_order/_order_form_articles.html | HTML |
<|file_sep|>appveyor.yml.diff
original:
CONFIG: Debug
updated:
BUILD_TYPE: Debug
<|file_sep|>appveyor.yml.diff
original:
CONFIG: Release
updated:
BUILD_TYPE: Release
<|file_sep|>appveyor.yml.diff
original:
CONFIG: Debug
updated:
BUILD_TYPE: Debug
<|file_sep|>appveyor.yml.diff
original:
CONFIG: Release
updated:
BUILD_TYPE: Release
<|file_sep|>original/appveyor.yml
environment:
matrix:
- GENERATOR: "Visual Studio 14"
CONFIG: Debug
- GENERATOR: "Visual Studio 14"
CONFIG: Release
- GENERATOR: "Visual Studio 14 Win64"
CONFIG: Debug
- GENERATOR: "Visual Studio 14 Win64"
CONFIG: Release
build_script:
- mkdir _build
- cd _build
- cmake "-G%GENERATOR%" "-DCMAKE_BUILD_TYPE=%BUILD_TYPE%" ..
- cmake --build .
<|file_sep|>current/appveyor.yml
environment:
matrix:
- GENERATOR: "Visual Studio 14"
BUILD_TYPE: Debug
- GENERATOR: "Visual Studio 14"
BUILD_TYPE: Release
- GENERATOR: "Visual Studio 14 Win64"
BUILD_TYPE: Debug
- GENERATOR: "Visual Studio 14 Win64"
BUILD_TYPE: Release
build_script:
- mkdir _build
- cd _build
- cmake "-G%GENERATOR%" "-DCMAKE_BUILD_TYPE=%BUILD_TYPE%" ..
- cmake --build .
<|file_sep|>updated/appveyor.yml |
environment:
matrix:
- GENERATOR: "Visual Studio 14"
BUILD_TYPE: Debug
- GENERATOR: "Visual Studio 14"
BUILD_TYPE: Release
- GENERATOR: "Visual Studio 14 Win64"
BUILD_TYPE: Debug
- GENERATOR: "Visual Studio 14 Win64"
BUILD_TYPE: Release
build_script:
- mkdir _build
- cd _build
- cmake "-G%GENERATOR%" ..
- cmake --build . --config "%CONFIG%" | <|file_sep|>appveyor.yml.diff
original:
CONFIG: Debug
updated:
BUILD_TYPE: Debug
<|file_sep|>appveyor.yml.diff
original:
CONFIG: Release
updated:
BUILD_TYPE: Release
<|file_sep|>appveyor.yml.diff
original:
CONFIG: Debug
updated:
BUILD_TYPE: Debug
<|file_sep|>appveyor.yml.diff
original:
CONFIG: Release
updated:
BUILD_TYPE: Release
<|file_sep|>original/appveyor.yml
environment:
matrix:
- GENERATOR: "Visual Studio 14"
CONFIG: Debug
- GENERATOR: "Visual Studio 14"
CONFIG: Release
- GENERATOR: "Visual Studio 14 Win64"
CONFIG: Debug
- GENERATOR: "Visual Studio 14 Win64"
CONFIG: Release
build_script:
- mkdir _build
- cd _build
- cmake "-G%GENERATOR%" "-DCMAKE_BUILD_TYPE=%BUILD_TYPE%" ..
- cmake --build .
<|file_sep|>current/appveyor.yml
environment:
matrix:
- GENERATOR: "Visual Studio 14"
BUILD_TYPE: Debug
- GENERATOR: "Visual Studio 14"
BUILD_TYPE: Release
- GENERATOR: "Visual Studio 14 Win64"
BUILD_TYPE: Debug
- GENERATOR: "Visual Studio 14 Win64"
BUILD_TYPE: Release
build_script:
- mkdir _build
- cd _build
- cmake "-G%GENERATOR%" "-DCMAKE_BUILD_TYPE=%BUILD_TYPE%" ..
- cmake --build .
<|file_sep|>updated/appveyor.yml
environment:
matrix:
- GENERATOR: "Visual Studio 14"
BUILD_TYPE: Debug
- GENERATOR: "Visual Studio 14"
BUILD_TYPE: Release
- GENERATOR: "Visual Studio 14 Win64"
BUILD_TYPE: Debug
- GENERATOR: "Visual Studio 14 Win64"
BUILD_TYPE: Release
build_script:
- mkdir _build
- cd _build
- cmake "-G%GENERATOR%" ..
- cmake --build . --config "%CONFIG%" | 885596d696b547731610dc01e0037c48d339f45c | appveyor.yml | appveyor.yml | YAML |
<|file_sep|>original/.github/workflows/ci.yml
<|file_sep|>current/.github/workflows/ci.yml
<|file_sep|>updated/.github/workflows/ci.yml | name: Django CI
on:
push:
branches: ['**']
pull_request:
branches: ['**']
jobs:
build:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- name: Build docker image
run: docker build -t p2pu/learning-circles .
- name: start postgres container
run: docker run -d --name=postgres -e POSTGRES_PASSWORD=password postgres:11
- name: start selenium container
run: docker run -d --name=selenium selenium/standalone-chrome
- name: run tests
run: docker run --rm -it --link postgres --link selenium -e DATABASE_URL=postgres://postgres:password@postgres/lc -e SECRET_KEY=abc -e ADMIN_EMAIL=admin@localhost p2pu/learning-circles dockerize -wait tcp://postgres:5432 /opt/django-venv/bin/python /opt/app/manage.py test | <|file_sep|>original/.github/workflows/ci.yml
<|file_sep|>current/.github/workflows/ci.yml
<|file_sep|>updated/.github/workflows/ci.yml
name: Django CI
on:
push:
branches: ['**']
pull_request:
branches: ['**']
jobs:
build:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- name: Build docker image
run: docker build -t p2pu/learning-circles .
- name: start postgres container
run: docker run -d --name=postgres -e POSTGRES_PASSWORD=password postgres:11
- name: start selenium container
run: docker run -d --name=selenium selenium/standalone-chrome
- name: run tests
run: docker run --rm -it --link postgres --link selenium -e DATABASE_URL=postgres://postgres:password@postgres/lc -e SECRET_KEY=abc -e ADMIN_EMAIL=admin@localhost p2pu/learning-circles dockerize -wait tcp://postgres:5432 /opt/django-venv/bin/python /opt/app/manage.py test | a0259fbc5120c518a77be321dc01e2b785a8e292 | .github/workflows/ci.yml | .github/workflows/ci.yml | YAML |
<|file_sep|>original/buildout.cfg
[buildout]
develop = .
parts = test scripts data
[test]
recipe = zc.recipe.testrunner
eggs = pycountry
defaults = ['-vv', '--all', '--color']
[scripts]
recipe = zc.recipe.egg
eggs = pycountry
interpreter = py
[data]
recipe = zerokspot.recipe.git
repository = git://git.debian.org/git/iso-codes/iso-codes.git
rev =
<|file_sep|>current/buildout.cfg
[buildout]
develop = .
parts = test scripts data
[test]
recipe = zc.recipe.testrunner
eggs = pycountry
defaults = ['-vv', '--all', '--color']
[scripts]
recipe = zc.recipe.egg
eggs = pycountry
interpreter = py
[data]
recipe = zerokspot.recipe.git
repository = git://git.debian.org/git/iso-codes/iso-codes.git
rev =
<|file_sep|>updated/buildout.cfg | [buildout]
develop = .
parts = test scripts data
[test]
recipe = zc.recipe.testrunner
eggs = pycountry
defaults = ['-vv', '--all', '--color']
[scripts]
recipe = zc.recipe.egg
eggs = pycountry
interpreter = py
[data]
recipe = zerokspot.recipe.git
repository = git://git.debian.org/git/iso-codes/iso-codes.git
rev = 770fa9cd603f90f9fb982b32fe6f45d253f1d33e | <|file_sep|>original/buildout.cfg
[buildout]
develop = .
parts = test scripts data
[test]
recipe = zc.recipe.testrunner
eggs = pycountry
defaults = ['-vv', '--all', '--color']
[scripts]
recipe = zc.recipe.egg
eggs = pycountry
interpreter = py
[data]
recipe = zerokspot.recipe.git
repository = git://git.debian.org/git/iso-codes/iso-codes.git
rev =
<|file_sep|>current/buildout.cfg
[buildout]
develop = .
parts = test scripts data
[test]
recipe = zc.recipe.testrunner
eggs = pycountry
defaults = ['-vv', '--all', '--color']
[scripts]
recipe = zc.recipe.egg
eggs = pycountry
interpreter = py
[data]
recipe = zerokspot.recipe.git
repository = git://git.debian.org/git/iso-codes/iso-codes.git
rev =
<|file_sep|>updated/buildout.cfg
[buildout]
develop = .
parts = test scripts data
[test]
recipe = zc.recipe.testrunner
eggs = pycountry
defaults = ['-vv', '--all', '--color']
[scripts]
recipe = zc.recipe.egg
eggs = pycountry
interpreter = py
[data]
recipe = zerokspot.recipe.git
repository = git://git.debian.org/git/iso-codes/iso-codes.git
rev = 770fa9cd603f90f9fb982b32fe6f45d253f1d33e | d8c60a7d7f90d54ed2ef9cb0cb1ed0af413f947e | buildout.cfg | buildout.cfg | INI |
<|file_sep|>original/_posts/2017-08-28-lsof.md
<|file_sep|>current/_posts/2017-08-28-lsof.md
<|file_sep|>updated/_posts/2017-08-28-lsof.md | ---
layout: post
title: Trouver un port utilisé avec lsof
---
[lsof] est une commande très pratique permettant de trouver les utilisateurs et les commandes utilisant certains fichiers. Par exemple on peut s'en servir pour voir quel programme utilise un port donné, un port étant considéré comme un fichier dans les systèmes Unix.
Pour trouver le processus utilisant le port 8080 :
lsof -i :8080
[lsof] possède de nombreuses options, consultables dans son [man]uel.
[lsof]: https://fr.wikipedia.org/wiki/Lsof
[man]: https://man.cx/lsof | <|file_sep|>original/_posts/2017-08-28-lsof.md
<|file_sep|>current/_posts/2017-08-28-lsof.md
<|file_sep|>updated/_posts/2017-08-28-lsof.md
---
layout: post
title: Trouver un port utilisé avec lsof
---
[lsof] est une commande très pratique permettant de trouver les utilisateurs et les commandes utilisant certains fichiers. Par exemple on peut s'en servir pour voir quel programme utilise un port donné, un port étant considéré comme un fichier dans les systèmes Unix.
Pour trouver le processus utilisant le port 8080 :
lsof -i :8080
[lsof] possède de nombreuses options, consultables dans son [man]uel.
[lsof]: https://fr.wikipedia.org/wiki/Lsof
[man]: https://man.cx/lsof | c3dbbf71d76230fc11ff97ce4c90ec8c8907373f | _posts/2017-08-28-lsof.md | _posts/2017-08-28-lsof.md | Markdown |
<|file_sep|>original/requirements_dev.txt
pip==20.2.4
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.4.1
tox==2.9.1
coverage==4.4.1
Sphinx==3.2.1
cryptography==3.2.1
PyYAML==5.3.1
pytest==3.2.3
pytest-runner==2.12.1
pytest-catchlog==1.2.2
piexif
filetype
jmespath
python-box
python-frontmatter
<|file_sep|>current/requirements_dev.txt
pip==20.2.4
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.4.1
tox==2.9.1
coverage==4.4.1
Sphinx==3.2.1
cryptography==3.2.1
PyYAML==5.3.1
pytest==3.2.3
pytest-runner==2.12.1
pytest-catchlog==1.2.2
piexif
filetype
jmespath
python-box
python-frontmatter
<|file_sep|>updated/requirements_dev.txt | pip==20.2.4
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.4.1
tox==2.9.1
coverage==4.4.1
Sphinx==3.2.1
cryptography==3.3.1
PyYAML==5.3.1
pytest==3.2.3
pytest-runner==2.12.1
pytest-catchlog==1.2.2
piexif
filetype
jmespath
python-box
python-frontmatter | <|file_sep|>original/requirements_dev.txt
pip==20.2.4
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.4.1
tox==2.9.1
coverage==4.4.1
Sphinx==3.2.1
cryptography==3.2.1
PyYAML==5.3.1
pytest==3.2.3
pytest-runner==2.12.1
pytest-catchlog==1.2.2
piexif
filetype
jmespath
python-box
python-frontmatter
<|file_sep|>current/requirements_dev.txt
pip==20.2.4
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.4.1
tox==2.9.1
coverage==4.4.1
Sphinx==3.2.1
cryptography==3.2.1
PyYAML==5.3.1
pytest==3.2.3
pytest-runner==2.12.1
pytest-catchlog==1.2.2
piexif
filetype
jmespath
python-box
python-frontmatter
<|file_sep|>updated/requirements_dev.txt
pip==20.2.4
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.4.1
tox==2.9.1
coverage==4.4.1
Sphinx==3.2.1
cryptography==3.3.1
PyYAML==5.3.1
pytest==3.2.3
pytest-runner==2.12.1
pytest-catchlog==1.2.2
piexif
filetype
jmespath
python-box
python-frontmatter | 0a6ef610ab5ee8a964d3e5f2e9a5237c03002b2a | requirements_dev.txt | requirements_dev.txt | Text |
<|file_sep|>original/.travis.yml
python:
- "3.4"
- "3.3"
- "3.2"
- "2.7"
# install pgpdump - it's needed for testing
before_install:
- sudo apt-get update -qq
- sudo apt-get install gnupg pgpdump -qq
# command to install dependencies
install:
- pip install -r requirements.txt
- pip install -r requirements-test.txt
- pip install python-coveralls
- pip install .
# use setup.py to invoke testing via coveralls
script:
py.test -vv --cov pgpy --cov-report term-missing tests/
# and report coverage to coveralls
after_success:
coveralls
<|file_sep|>current/.travis.yml
python:
- "3.4"
- "3.3"
- "3.2"
- "2.7"
# install pgpdump - it's needed for testing
before_install:
- sudo apt-get update -qq
- sudo apt-get install gnupg pgpdump -qq
# command to install dependencies
install:
- pip install -r requirements.txt
- pip install -r requirements-test.txt
- pip install python-coveralls
- pip install .
# use setup.py to invoke testing via coveralls
script:
py.test -vv --cov pgpy --cov-report term-missing tests/
# and report coverage to coveralls
after_success:
coveralls
<|file_sep|>updated/.travis.yml | python:
- "3.4"
- "3.3"
- "3.2"
- "2.7"
# install pgpdump - it's needed for testing
before_install:
- sudo apt-get update -qq
- sudo apt-get install gnupg pgpdump -qq
# command to install dependencies
install:
- pip install -r requirements.txt
- pip install -r requirements-test.txt
- pip install python-coveralls
- pip install .
# use setup.py to invoke testing via coveralls
script:
PYTHONPATH=. py.test -vv --cov pgpy --cov-report term-missing tests/
# and report coverage to coveralls
after_success:
coveralls | <|file_sep|>original/.travis.yml
python:
- "3.4"
- "3.3"
- "3.2"
- "2.7"
# install pgpdump - it's needed for testing
before_install:
- sudo apt-get update -qq
- sudo apt-get install gnupg pgpdump -qq
# command to install dependencies
install:
- pip install -r requirements.txt
- pip install -r requirements-test.txt
- pip install python-coveralls
- pip install .
# use setup.py to invoke testing via coveralls
script:
py.test -vv --cov pgpy --cov-report term-missing tests/
# and report coverage to coveralls
after_success:
coveralls
<|file_sep|>current/.travis.yml
python:
- "3.4"
- "3.3"
- "3.2"
- "2.7"
# install pgpdump - it's needed for testing
before_install:
- sudo apt-get update -qq
- sudo apt-get install gnupg pgpdump -qq
# command to install dependencies
install:
- pip install -r requirements.txt
- pip install -r requirements-test.txt
- pip install python-coveralls
- pip install .
# use setup.py to invoke testing via coveralls
script:
py.test -vv --cov pgpy --cov-report term-missing tests/
# and report coverage to coveralls
after_success:
coveralls
<|file_sep|>updated/.travis.yml
python:
- "3.4"
- "3.3"
- "3.2"
- "2.7"
# install pgpdump - it's needed for testing
before_install:
- sudo apt-get update -qq
- sudo apt-get install gnupg pgpdump -qq
# command to install dependencies
install:
- pip install -r requirements.txt
- pip install -r requirements-test.txt
- pip install python-coveralls
- pip install .
# use setup.py to invoke testing via coveralls
script:
PYTHONPATH=. py.test -vv --cov pgpy --cov-report term-missing tests/
# and report coverage to coveralls
after_success:
coveralls | c835d8d9d1bd269b187df50ec1ba81035ddf1a3c | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/README.md
- No down time to relad configuration
- configuration should be shared by all the nodes but no context is shared between nodes (No node to node communication)
- ...
### Current solution:
The current solution use openresty (NGINX based scriptable gateway) with a redis as a Database. It's based on dns
The NGINX have 2 servers configured:
- server listening on 8888 is for configuration and is used to push an upstream to a redis List ( the key is the desired location/API)
- server listening on 8666 have a wildcard location that will execute a lua (rewrite) script to find the upstream, based on the path
and rewrite it to the configured upstream
The current solution is based on https://github.com/tes/lua-resty-router
### TODO
- Update the README to add a demo and how to deploy
- Add the ability to remove an upstream
- Add the ability to configure multiple end points for one upstream
- Add a pipeling with tests (specially for LUA scripting)
- ...
<|file_sep|>current/README.md
- No down time to relad configuration
- configuration should be shared by all the nodes but no context is shared between nodes (No node to node communication)
- ...
### Current solution:
The current solution use openresty (NGINX based scriptable gateway) with a redis as a Database. It's based on dns
The NGINX have 2 servers configured:
- server listening on 8888 is for configuration and is used to push an upstream to a redis List ( the key is the desired location/API)
- server listening on 8666 have a wildcard location that will execute a lua (rewrite) script to find the upstream, based on the path
and rewrite it to the configured upstream
The current solution is based on https://github.com/tes/lua-resty-router
### TODO
- Update the README to add a demo and how to deploy
- Add the ability to remove an upstream
- Add the ability to configure multiple end points for one upstream
- Add a pipeling with tests (specially for LUA scripting)
- ...
<|file_sep|>updated/README.md |
### Current solution:
The current solution use openresty (NGINX based scriptable gateway) with a redis as a Database. It's based on dns
The NGINX have 2 servers configured:
- server listening on 8888 is for configuration and is used to push an upstream to a redis List ( the key is the desired location/API)
- server listening on 8666 have a wildcard location that will execute a lua (rewrite) script to find the upstream, based on the path
and rewrite it to the configured upstream
The current solution is based on https://github.com/tes/lua-resty-router
### TODO
- Update the README to add a demo and how to deploy
- ~~Add the ability to remove a location~~
- Add the ability to configure multiple upstream for one location (load balancing)
- ~~Add a pipeling with tests (specially for LUA scripting)~~
- Add the ability to delete all locations
- Add unit test framework for the lua libraries
- Write documentation with examples and use cases
- ...
| <|file_sep|>original/README.md
- No down time to relad configuration
- configuration should be shared by all the nodes but no context is shared between nodes (No node to node communication)
- ...
### Current solution:
The current solution use openresty (NGINX based scriptable gateway) with a redis as a Database. It's based on dns
The NGINX have 2 servers configured:
- server listening on 8888 is for configuration and is used to push an upstream to a redis List ( the key is the desired location/API)
- server listening on 8666 have a wildcard location that will execute a lua (rewrite) script to find the upstream, based on the path
and rewrite it to the configured upstream
The current solution is based on https://github.com/tes/lua-resty-router
### TODO
- Update the README to add a demo and how to deploy
- Add the ability to remove an upstream
- Add the ability to configure multiple end points for one upstream
- Add a pipeling with tests (specially for LUA scripting)
- ...
<|file_sep|>current/README.md
- No down time to relad configuration
- configuration should be shared by all the nodes but no context is shared between nodes (No node to node communication)
- ...
### Current solution:
The current solution use openresty (NGINX based scriptable gateway) with a redis as a Database. It's based on dns
The NGINX have 2 servers configured:
- server listening on 8888 is for configuration and is used to push an upstream to a redis List ( the key is the desired location/API)
- server listening on 8666 have a wildcard location that will execute a lua (rewrite) script to find the upstream, based on the path
and rewrite it to the configured upstream
The current solution is based on https://github.com/tes/lua-resty-router
### TODO
- Update the README to add a demo and how to deploy
- Add the ability to remove an upstream
- Add the ability to configure multiple end points for one upstream
- Add a pipeling with tests (specially for LUA scripting)
- ...
<|file_sep|>updated/README.md
### Current solution:
The current solution use openresty (NGINX based scriptable gateway) with a redis as a Database. It's based on dns
The NGINX have 2 servers configured:
- server listening on 8888 is for configuration and is used to push an upstream to a redis List ( the key is the desired location/API)
- server listening on 8666 have a wildcard location that will execute a lua (rewrite) script to find the upstream, based on the path
and rewrite it to the configured upstream
The current solution is based on https://github.com/tes/lua-resty-router
### TODO
- Update the README to add a demo and how to deploy
- ~~Add the ability to remove a location~~
- Add the ability to configure multiple upstream for one location (load balancing)
- ~~Add a pipeling with tests (specially for LUA scripting)~~
- Add the ability to delete all locations
- Add unit test framework for the lua libraries
- Write documentation with examples and use cases
- ...
| 2236f9a0df299f320c52802dd8b75faf27961564 | README.md | README.md | Markdown |
<|file_sep|>example/fig.yml.diff
original:
CONTACT:
ADVERTISE:
updated:
PEER:
IP:
SERF_PORT:
<|file_sep|>original/example/fig.yml
---
example:
image: serfnode_example
environment:
ROLE:
CONTACT:
ADVERTISE:
RPC_PORT:
NODE:
IP:
<|file_sep|>current/example/fig.yml
---
example:
image: serfnode_example
environment:
ROLE:
PEER:
IP:
SERF_PORT:
RPC_PORT:
NODE:
IP:
<|file_sep|>updated/example/fig.yml | ---
example:
image: serfnode_example
environment:
ROLE:
PEER:
IP:
SERF_PORT:
RPC_PORT:
NODE_NAME: | <|file_sep|>example/fig.yml.diff
original:
CONTACT:
ADVERTISE:
updated:
PEER:
IP:
SERF_PORT:
<|file_sep|>original/example/fig.yml
---
example:
image: serfnode_example
environment:
ROLE:
CONTACT:
ADVERTISE:
RPC_PORT:
NODE:
IP:
<|file_sep|>current/example/fig.yml
---
example:
image: serfnode_example
environment:
ROLE:
PEER:
IP:
SERF_PORT:
RPC_PORT:
NODE:
IP:
<|file_sep|>updated/example/fig.yml
---
example:
image: serfnode_example
environment:
ROLE:
PEER:
IP:
SERF_PORT:
RPC_PORT:
NODE_NAME: | ae8d918e662230325fc26925f8da3e4272b77c03 | example/fig.yml | example/fig.yml | YAML |
<|file_sep|>package.json.diff
original:
"gm":"*"
updated:
"gm": "*"
<|file_sep|>original/package.json
},
"scripts": {
"test": "grunt test"
},
"dependencies": {
"spritesmith": "0.7.1"
},
"devDependencies": {
"grunt": "~0.4.1",
"gm":"*"
},
"peerDependencies": {
"grunt": "~0.4.1"
},
"keywords": [
"gruntplugin",
"sprite",
"css",
"autosprite"
]
}
<|file_sep|>current/package.json
},
"scripts": {
"test": "grunt test"
},
"dependencies": {
"spritesmith": "0.7.1"
},
"devDependencies": {
"grunt": "~0.4.1",
"gm": "*"
},
"peerDependencies": {
"grunt": "~0.4.1"
},
"keywords": [
"gruntplugin",
"sprite",
"css",
"autosprite"
]
}
<|file_sep|>updated/package.json | },
"scripts": {
"test": "grunt test"
},
"dependencies": {
"spritesmith": "0.7.1"
},
"devDependencies": {
"grunt": "~0.4.1",
"gm": "*"
},
"peerDependencies": {
"grunt": ">=0.4.0"
},
"keywords": [
"gruntplugin",
"sprite",
"css",
"autosprite"
]
} | <|file_sep|>package.json.diff
original:
"gm":"*"
updated:
"gm": "*"
<|file_sep|>original/package.json
},
"scripts": {
"test": "grunt test"
},
"dependencies": {
"spritesmith": "0.7.1"
},
"devDependencies": {
"grunt": "~0.4.1",
"gm":"*"
},
"peerDependencies": {
"grunt": "~0.4.1"
},
"keywords": [
"gruntplugin",
"sprite",
"css",
"autosprite"
]
}
<|file_sep|>current/package.json
},
"scripts": {
"test": "grunt test"
},
"dependencies": {
"spritesmith": "0.7.1"
},
"devDependencies": {
"grunt": "~0.4.1",
"gm": "*"
},
"peerDependencies": {
"grunt": "~0.4.1"
},
"keywords": [
"gruntplugin",
"sprite",
"css",
"autosprite"
]
}
<|file_sep|>updated/package.json
},
"scripts": {
"test": "grunt test"
},
"dependencies": {
"spritesmith": "0.7.1"
},
"devDependencies": {
"grunt": "~0.4.1",
"gm": "*"
},
"peerDependencies": {
"grunt": ">=0.4.0"
},
"keywords": [
"gruntplugin",
"sprite",
"css",
"autosprite"
]
} | 624f40ef56c04f589b07aa1f85d66feb9aa0e06c | package.json | package.json | JSON |
<|file_sep|>original/input/docs/issue-provider/inspectcode.md
---
Order: 20
Title: JetBrains InspectCode
---
Support for reading issues reported by [JetBrains InsepectCode] is implemented in the [Cake.Prca.Issues.InspectCode addin].
## Requirements
* Cake 0.16.2 or newer.
## Features
* Reads warnings from JetBrains InspectCode log files.
[JetBrains InsepectCode]: https://www.jetbrains.com/help/resharper/2017.1/InspectCode.html
[Cake.Prca.Issues.InspectCode addin]: https://www.nuget.org/packages/Cake.Prca.Issues.InspectCode
<|file_sep|>current/input/docs/issue-provider/inspectcode.md
---
Order: 20
Title: JetBrains InspectCode
---
Support for reading issues reported by [JetBrains InsepectCode] is implemented in the [Cake.Prca.Issues.InspectCode addin].
## Requirements
* Cake 0.16.2 or newer.
## Features
* Reads warnings from JetBrains InspectCode log files.
[JetBrains InsepectCode]: https://www.jetbrains.com/help/resharper/2017.1/InspectCode.html
[Cake.Prca.Issues.InspectCode addin]: https://www.nuget.org/packages/Cake.Prca.Issues.InspectCode
<|file_sep|>updated/input/docs/issue-provider/inspectcode.md | ---
Order: 20
Title: JetBrains InspectCode
---
Support for reading issues reported by [JetBrains InsepectCode] is implemented in the [Cake.Prca.Issues.InspectCode addin].
## Requirements
* Cake 0.16.2 or newer.
## Features
* Reads warnings from JetBrains InspectCode log files.
* Provides URLs for issues containing a Wiki URL.
[JetBrains InsepectCode]: https://www.jetbrains.com/help/resharper/2017.1/InspectCode.html
[Cake.Prca.Issues.InspectCode addin]: https://www.nuget.org/packages/Cake.Prca.Issues.InspectCode | <|file_sep|>original/input/docs/issue-provider/inspectcode.md
---
Order: 20
Title: JetBrains InspectCode
---
Support for reading issues reported by [JetBrains InsepectCode] is implemented in the [Cake.Prca.Issues.InspectCode addin].
## Requirements
* Cake 0.16.2 or newer.
## Features
* Reads warnings from JetBrains InspectCode log files.
[JetBrains InsepectCode]: https://www.jetbrains.com/help/resharper/2017.1/InspectCode.html
[Cake.Prca.Issues.InspectCode addin]: https://www.nuget.org/packages/Cake.Prca.Issues.InspectCode
<|file_sep|>current/input/docs/issue-provider/inspectcode.md
---
Order: 20
Title: JetBrains InspectCode
---
Support for reading issues reported by [JetBrains InsepectCode] is implemented in the [Cake.Prca.Issues.InspectCode addin].
## Requirements
* Cake 0.16.2 or newer.
## Features
* Reads warnings from JetBrains InspectCode log files.
[JetBrains InsepectCode]: https://www.jetbrains.com/help/resharper/2017.1/InspectCode.html
[Cake.Prca.Issues.InspectCode addin]: https://www.nuget.org/packages/Cake.Prca.Issues.InspectCode
<|file_sep|>updated/input/docs/issue-provider/inspectcode.md
---
Order: 20
Title: JetBrains InspectCode
---
Support for reading issues reported by [JetBrains InsepectCode] is implemented in the [Cake.Prca.Issues.InspectCode addin].
## Requirements
* Cake 0.16.2 or newer.
## Features
* Reads warnings from JetBrains InspectCode log files.
* Provides URLs for issues containing a Wiki URL.
[JetBrains InsepectCode]: https://www.jetbrains.com/help/resharper/2017.1/InspectCode.html
[Cake.Prca.Issues.InspectCode addin]: https://www.nuget.org/packages/Cake.Prca.Issues.InspectCode | 7dba182f3d30dfc030fb111deb46206e9eb13928 | input/docs/issue-provider/inspectcode.md | input/docs/issue-provider/inspectcode.md | Markdown |
<|file_sep|>lib/content_manager/class_resolution.rb.diff
original:
updated:
<|file_sep|>original/lib/content_manager/class_resolution.rb
if Object.const_defined?(klass)
return klass
elsif file_path = rails_constant_file_path(class_name)
require_dependency file_path
return klass
else
cm_class_name = "content_manager/#{class_name}".classify
begin
cm_class_name.constantize
return cm_class_name
rescue NameError
raise "Couldn't find constant definition #{name}, it should be in a file called #{name}.rb"
end
end
end
def self.rails_constant_file_path(class_name)
Dir["#{Rails.root}/{app,lib}/**/#{class_name}.rb"].first
end
end
<|file_sep|>current/lib/content_manager/class_resolution.rb
if Object.const_defined?(klass)
return klass
elsif file_path = rails_constant_file_path(class_name)
require_dependency file_path
return klass
else
cm_class_name = "content_manager/#{class_name}".classify
begin
cm_class_name.constantize
return cm_class_name
rescue NameError
raise "Couldn't find constant definition #{name}, it should be in a file called #{name}.rb"
end
end
end
def self.rails_constant_file_path(class_name)
Dir["#{Rails.root}/{app,lib}/**/#{class_name}.rb"].first
end
end
end
<|file_sep|>updated/lib/content_manager/class_resolution.rb | if Object.const_defined?(klass)
return klass
elsif file_path = rails_constant_file_path(class_name)
require_dependency file_path
return klass
else
cm_class_name = "content_manager/#{class_name}".classify
begin
cm_class_name.constantize
return cm_class_name
rescue NameError
raise "Couldn't find constant definition #{class_name}, it should be in a file called #{class_name}.rb"
end
end
end
def self.rails_constant_file_path(class_name)
Dir["#{Rails.root}/{app,lib}/**/#{class_name}.rb"].first
end
end
end | <|file_sep|>lib/content_manager/class_resolution.rb.diff
original:
updated:
<|file_sep|>original/lib/content_manager/class_resolution.rb
if Object.const_defined?(klass)
return klass
elsif file_path = rails_constant_file_path(class_name)
require_dependency file_path
return klass
else
cm_class_name = "content_manager/#{class_name}".classify
begin
cm_class_name.constantize
return cm_class_name
rescue NameError
raise "Couldn't find constant definition #{name}, it should be in a file called #{name}.rb"
end
end
end
def self.rails_constant_file_path(class_name)
Dir["#{Rails.root}/{app,lib}/**/#{class_name}.rb"].first
end
end
<|file_sep|>current/lib/content_manager/class_resolution.rb
if Object.const_defined?(klass)
return klass
elsif file_path = rails_constant_file_path(class_name)
require_dependency file_path
return klass
else
cm_class_name = "content_manager/#{class_name}".classify
begin
cm_class_name.constantize
return cm_class_name
rescue NameError
raise "Couldn't find constant definition #{name}, it should be in a file called #{name}.rb"
end
end
end
def self.rails_constant_file_path(class_name)
Dir["#{Rails.root}/{app,lib}/**/#{class_name}.rb"].first
end
end
end
<|file_sep|>updated/lib/content_manager/class_resolution.rb
if Object.const_defined?(klass)
return klass
elsif file_path = rails_constant_file_path(class_name)
require_dependency file_path
return klass
else
cm_class_name = "content_manager/#{class_name}".classify
begin
cm_class_name.constantize
return cm_class_name
rescue NameError
raise "Couldn't find constant definition #{class_name}, it should be in a file called #{class_name}.rb"
end
end
end
def self.rails_constant_file_path(class_name)
Dir["#{Rails.root}/{app,lib}/**/#{class_name}.rb"].first
end
end
end | 4ac134803b5b069311e76098900e996bf2a0e151 | lib/content_manager/class_resolution.rb | lib/content_manager/class_resolution.rb | Ruby |
<|file_sep|>original/fedopt_guide/README.md
<|file_sep|>current/fedopt_guide/README.md
<|file_sep|>updated/fedopt_guide/README.md | # FedOpt Guide experiments
Note: This directory is a work-in-progress.
This is a shared folder for developing a field guide to federated optimization.
* The code in this folder should mimic the structure in `optimization/`.
* Minimize the number of PRs when you try to commit your code. Note that you
do not need to commit your code at this point if it will not be used by
others.
* Consider creating a subfolder for each task by `dataset_model`, for example,
`fedopt_guide/gld23k_mobilenet`.
* Re-use the shared code in `optimization/shared/` and `utils/` whenever
possible, for example, `utils.training_loop` is recommended for the training
loop. Use `tff.learning.build_federated_averaging_process` instead of
`optimization.shared.fed_avg_schedule` to build your TFF iterative process. | <|file_sep|>original/fedopt_guide/README.md
<|file_sep|>current/fedopt_guide/README.md
<|file_sep|>updated/fedopt_guide/README.md
# FedOpt Guide experiments
Note: This directory is a work-in-progress.
This is a shared folder for developing a field guide to federated optimization.
* The code in this folder should mimic the structure in `optimization/`.
* Minimize the number of PRs when you try to commit your code. Note that you
do not need to commit your code at this point if it will not be used by
others.
* Consider creating a subfolder for each task by `dataset_model`, for example,
`fedopt_guide/gld23k_mobilenet`.
* Re-use the shared code in `optimization/shared/` and `utils/` whenever
possible, for example, `utils.training_loop` is recommended for the training
loop. Use `tff.learning.build_federated_averaging_process` instead of
`optimization.shared.fed_avg_schedule` to build your TFF iterative process. | e190f453f5e989028dfc7189f86c0ec64564fef5 | fedopt_guide/README.md | fedopt_guide/README.md | Markdown |
<|file_sep|>src/Group/Form/GroupFormBuilder.php.diff
original:
updated:
'searchable',
<|file_sep|>original/src/Group/Form/GroupFormBuilder.php
*/
protected $prefix = 'variables_';
/**
* The stream namespace.
*
* @var string
*/
protected $namespace = 'variables';
/**
* The skipped fields.
*
* @var array
*/
protected $skips = [
'title_column',
'trashable',
'sortable',
];
}
<|file_sep|>current/src/Group/Form/GroupFormBuilder.php
protected $prefix = 'variables_';
/**
* The stream namespace.
*
* @var string
*/
protected $namespace = 'variables';
/**
* The skipped fields.
*
* @var array
*/
protected $skips = [
'title_column',
'searchable',
'trashable',
'sortable',
];
}
<|file_sep|>updated/src/Group/Form/GroupFormBuilder.php |
/**
* The stream namespace.
*
* @var string
*/
protected $namespace = 'variables';
/**
* The skipped fields.
*
* @var array
*/
protected $skips = [
'title_column',
'searchable',
'trashable',
'sortable',
'config',
];
} | <|file_sep|>src/Group/Form/GroupFormBuilder.php.diff
original:
updated:
'searchable',
<|file_sep|>original/src/Group/Form/GroupFormBuilder.php
*/
protected $prefix = 'variables_';
/**
* The stream namespace.
*
* @var string
*/
protected $namespace = 'variables';
/**
* The skipped fields.
*
* @var array
*/
protected $skips = [
'title_column',
'trashable',
'sortable',
];
}
<|file_sep|>current/src/Group/Form/GroupFormBuilder.php
protected $prefix = 'variables_';
/**
* The stream namespace.
*
* @var string
*/
protected $namespace = 'variables';
/**
* The skipped fields.
*
* @var array
*/
protected $skips = [
'title_column',
'searchable',
'trashable',
'sortable',
];
}
<|file_sep|>updated/src/Group/Form/GroupFormBuilder.php
/**
* The stream namespace.
*
* @var string
*/
protected $namespace = 'variables';
/**
* The skipped fields.
*
* @var array
*/
protected $skips = [
'title_column',
'searchable',
'trashable',
'sortable',
'config',
];
} | c8684e638fdab86508ba5d149fdc5b32b8d9811e | src/Group/Form/GroupFormBuilder.php | src/Group/Form/GroupFormBuilder.php | PHP |
<|file_sep|>original/maas/client/viscera/tests/test_sshkeys.py
<|file_sep|>current/maas/client/viscera/tests/test_sshkeys.py
<|file_sep|>updated/maas/client/viscera/tests/test_sshkeys.py | """Test for `maas.client.viscera.sshkeys`."""
from maas.client.viscera import Origin
from .. import sshkeys
from ...testing import TestCase
from ..testing import bind
def make_origin():
return bind(sshkeys.SSHKeys, sshkeys.SSHKey)
class TestSSHKeys(TestCase):
def test__sshkeys_read(self):
""" SSHKeys.read() returns all SSH keys. """
SSHKeys = make_origin().SSHKeys
# create a list of dicts for read() to return
self.assertThat(SSHKeys.read(), Equals())
def test__sshkeys_create(self): | <|file_sep|>original/maas/client/viscera/tests/test_sshkeys.py
<|file_sep|>current/maas/client/viscera/tests/test_sshkeys.py
<|file_sep|>updated/maas/client/viscera/tests/test_sshkeys.py
"""Test for `maas.client.viscera.sshkeys`."""
from maas.client.viscera import Origin
from .. import sshkeys
from ...testing import TestCase
from ..testing import bind
def make_origin():
return bind(sshkeys.SSHKeys, sshkeys.SSHKey)
class TestSSHKeys(TestCase):
def test__sshkeys_read(self):
""" SSHKeys.read() returns all SSH keys. """
SSHKeys = make_origin().SSHKeys
# create a list of dicts for read() to return
self.assertThat(SSHKeys.read(), Equals())
def test__sshkeys_create(self): | 4a83a84f5358482fab7d1b7b7e259aed2b958e5d | maas/client/viscera/tests/test_sshkeys.py | maas/client/viscera/tests/test_sshkeys.py | Python |
<|file_sep|>original/examples/trace_example.js
const trace = require('./lib/trace')();
// register a JS listener
trace.on('trace', (category, trace_obj) => {
console.log(category, trace_obj);
});
// send a trace
trace.trace('joshgav-trace', { message: "stuff happened" });
<|file_sep|>current/examples/trace_example.js
const trace = require('./lib/trace')();
// register a JS listener
trace.on('trace', (category, trace_obj) => {
console.log(category, trace_obj);
});
// send a trace
trace.trace('joshgav-trace', { message: "stuff happened" });
<|file_sep|>updated/examples/trace_example.js | const trace = require('../lib/trace')();
// register a JS listener
trace.on('trace', (category, trace_obj) => {
console.log(category, trace_obj);
});
// send a trace
trace.trace('joshgav-trace', { message: "stuff happened" });
| <|file_sep|>original/examples/trace_example.js
const trace = require('./lib/trace')();
// register a JS listener
trace.on('trace', (category, trace_obj) => {
console.log(category, trace_obj);
});
// send a trace
trace.trace('joshgav-trace', { message: "stuff happened" });
<|file_sep|>current/examples/trace_example.js
const trace = require('./lib/trace')();
// register a JS listener
trace.on('trace', (category, trace_obj) => {
console.log(category, trace_obj);
});
// send a trace
trace.trace('joshgav-trace', { message: "stuff happened" });
<|file_sep|>updated/examples/trace_example.js
const trace = require('../lib/trace')();
// register a JS listener
trace.on('trace', (category, trace_obj) => {
console.log(category, trace_obj);
});
// send a trace
trace.trace('joshgav-trace', { message: "stuff happened" });
| b32ced24dd10d223902497798fa0950e1f157e84 | examples/trace_example.js | examples/trace_example.js | JavaScript |
<|file_sep|>original/.travis.yml
language: ruby
rvm:
- 1.9.3
- 2.0.0
- 2.1.1
<|file_sep|>current/.travis.yml
language: ruby
rvm:
- 1.9.3
- 2.0.0
- 2.1.1
<|file_sep|>updated/.travis.yml | language: ruby
rvm:
- 1.9.3
- 2.0.0
- 2.1.3 | <|file_sep|>original/.travis.yml
language: ruby
rvm:
- 1.9.3
- 2.0.0
- 2.1.1
<|file_sep|>current/.travis.yml
language: ruby
rvm:
- 1.9.3
- 2.0.0
- 2.1.1
<|file_sep|>updated/.travis.yml
language: ruby
rvm:
- 1.9.3
- 2.0.0
- 2.1.3 | 9cbc501cd57313aef83511cdb3e1707f84047900 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/romana-install/roles/romana/postinstall/devstack/master/tasks/main.yml
---
# Make sure the bin directory exists
- file: path="/home/{{ ansible_ssh_user }}/bin/" mode=0775 state=directory
# Ensure the jq tool is installed
- include: "{{ item }}"
with_first_found:
- files:
- "packages_{{ ansible_distribution|lower }}_{{ ansible_distribution_release|lower }}.yml"
- "packages_{{ ansible_distribution|lower }}.yml"
- "packages_{{ ansible_os_family|lower }}.yml"
skip: true
become: true
become_user: root
- name: Update user profile
patch: src="{{ ansible_distribution|lower }}_profile_add_openrc.patch" basedir="/home/{{ ansible_ssh_user }}"
- name: Install romana post-install script
template: src="romana-post-install.sh" dest="/var/tmp/romana-post-install.sh" mode=0755
<|file_sep|>current/romana-install/roles/romana/postinstall/devstack/master/tasks/main.yml
---
# Make sure the bin directory exists
- file: path="/home/{{ ansible_ssh_user }}/bin/" mode=0775 state=directory
# Ensure the jq tool is installed
- include: "{{ item }}"
with_first_found:
- files:
- "packages_{{ ansible_distribution|lower }}_{{ ansible_distribution_release|lower }}.yml"
- "packages_{{ ansible_distribution|lower }}.yml"
- "packages_{{ ansible_os_family|lower }}.yml"
skip: true
become: true
become_user: root
- name: Update user profile
patch: src="{{ ansible_distribution|lower }}_profile_add_openrc.patch" basedir="/home/{{ ansible_ssh_user }}"
- name: Install romana post-install script
template: src="romana-post-install.sh" dest="/var/tmp/romana-post-install.sh" mode=0755
<|file_sep|>updated/romana-install/roles/romana/postinstall/devstack/master/tasks/main.yml | ---
# Ensure the jq tool is installed
- include: "{{ item }}"
with_first_found:
- files:
- "packages_{{ ansible_distribution|lower }}_{{ ansible_distribution_release|lower }}.yml"
- "packages_{{ ansible_distribution|lower }}.yml"
- "packages_{{ ansible_os_family|lower }}.yml"
skip: true
become: true
become_user: root
- name: Update user profile
patch: src="{{ ansible_distribution|lower }}_profile_add_openrc.patch" basedir="/home/{{ ansible_ssh_user }}"
- name: Install romana post-install script
template: src="romana-post-install.sh" dest="/var/tmp/romana-post-install.sh" mode=0755
- name: Execute romana post-install script
shell: /var/tmp/romana-post-install.sh | <|file_sep|>original/romana-install/roles/romana/postinstall/devstack/master/tasks/main.yml
---
# Make sure the bin directory exists
- file: path="/home/{{ ansible_ssh_user }}/bin/" mode=0775 state=directory
# Ensure the jq tool is installed
- include: "{{ item }}"
with_first_found:
- files:
- "packages_{{ ansible_distribution|lower }}_{{ ansible_distribution_release|lower }}.yml"
- "packages_{{ ansible_distribution|lower }}.yml"
- "packages_{{ ansible_os_family|lower }}.yml"
skip: true
become: true
become_user: root
- name: Update user profile
patch: src="{{ ansible_distribution|lower }}_profile_add_openrc.patch" basedir="/home/{{ ansible_ssh_user }}"
- name: Install romana post-install script
template: src="romana-post-install.sh" dest="/var/tmp/romana-post-install.sh" mode=0755
<|file_sep|>current/romana-install/roles/romana/postinstall/devstack/master/tasks/main.yml
---
# Make sure the bin directory exists
- file: path="/home/{{ ansible_ssh_user }}/bin/" mode=0775 state=directory
# Ensure the jq tool is installed
- include: "{{ item }}"
with_first_found:
- files:
- "packages_{{ ansible_distribution|lower }}_{{ ansible_distribution_release|lower }}.yml"
- "packages_{{ ansible_distribution|lower }}.yml"
- "packages_{{ ansible_os_family|lower }}.yml"
skip: true
become: true
become_user: root
- name: Update user profile
patch: src="{{ ansible_distribution|lower }}_profile_add_openrc.patch" basedir="/home/{{ ansible_ssh_user }}"
- name: Install romana post-install script
template: src="romana-post-install.sh" dest="/var/tmp/romana-post-install.sh" mode=0755
<|file_sep|>updated/romana-install/roles/romana/postinstall/devstack/master/tasks/main.yml
---
# Ensure the jq tool is installed
- include: "{{ item }}"
with_first_found:
- files:
- "packages_{{ ansible_distribution|lower }}_{{ ansible_distribution_release|lower }}.yml"
- "packages_{{ ansible_distribution|lower }}.yml"
- "packages_{{ ansible_os_family|lower }}.yml"
skip: true
become: true
become_user: root
- name: Update user profile
patch: src="{{ ansible_distribution|lower }}_profile_add_openrc.patch" basedir="/home/{{ ansible_ssh_user }}"
- name: Install romana post-install script
template: src="romana-post-install.sh" dest="/var/tmp/romana-post-install.sh" mode=0755
- name: Execute romana post-install script
shell: /var/tmp/romana-post-install.sh | 0d3ecd3c6be0204d37391835dddce50d21e6a867 | romana-install/roles/romana/postinstall/devstack/master/tasks/main.yml | romana-install/roles/romana/postinstall/devstack/master/tasks/main.yml | YAML |
<|file_sep|>original/srv/handler/helper.go
import (
elasticseach "github.com/Rakanixu/elasticsearch/srv/proto/elasticsearch"
"github.com/micro/go-micro/errors"
)
// RequiredRecordFieldsExists returns an error if DocRef struct has zero value
func DocRefFieldsExists(dr *elasticseach.DocRef) error {
if len(dr.Index) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Index required")
}
if len(dr.Type) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Type required")
}
if len(dr.Id) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Id required")
}
return nil
}
<|file_sep|>current/srv/handler/helper.go
import (
elasticseach "github.com/Rakanixu/elasticsearch/srv/proto/elasticsearch"
"github.com/micro/go-micro/errors"
)
// RequiredRecordFieldsExists returns an error if DocRef struct has zero value
func DocRefFieldsExists(dr *elasticseach.DocRef) error {
if len(dr.Index) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Index required")
}
if len(dr.Type) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Type required")
}
if len(dr.Id) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Id required")
}
return nil
}
<|file_sep|>updated/srv/handler/helper.go | package handler
import (
elasticseach "github.com/Rakanixu/elasticsearch/srv/proto/elasticsearch"
"github.com/micro/go-micro/errors"
)
// RequiredRecordFieldsExists returns an error if DocRef struct has zero value
func DocRefFieldsExists(dr *elasticseach.DocRef) error {
if len(dr.Index) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Index required")
}
if len(dr.Type) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Type required")
}
return nil
} | <|file_sep|>original/srv/handler/helper.go
import (
elasticseach "github.com/Rakanixu/elasticsearch/srv/proto/elasticsearch"
"github.com/micro/go-micro/errors"
)
// RequiredRecordFieldsExists returns an error if DocRef struct has zero value
func DocRefFieldsExists(dr *elasticseach.DocRef) error {
if len(dr.Index) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Index required")
}
if len(dr.Type) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Type required")
}
if len(dr.Id) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Id required")
}
return nil
}
<|file_sep|>current/srv/handler/helper.go
import (
elasticseach "github.com/Rakanixu/elasticsearch/srv/proto/elasticsearch"
"github.com/micro/go-micro/errors"
)
// RequiredRecordFieldsExists returns an error if DocRef struct has zero value
func DocRefFieldsExists(dr *elasticseach.DocRef) error {
if len(dr.Index) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Index required")
}
if len(dr.Type) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Type required")
}
if len(dr.Id) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Id required")
}
return nil
}
<|file_sep|>updated/srv/handler/helper.go
package handler
import (
elasticseach "github.com/Rakanixu/elasticsearch/srv/proto/elasticsearch"
"github.com/micro/go-micro/errors"
)
// RequiredRecordFieldsExists returns an error if DocRef struct has zero value
func DocRefFieldsExists(dr *elasticseach.DocRef) error {
if len(dr.Index) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Index required")
}
if len(dr.Type) <= 0 {
return errors.BadRequest("go.micro.srv.elasticsearch", "Type required")
}
return nil
} | 62ef15c40b2391cc28188da04ed7fb4f3c4c8f4e | srv/handler/helper.go | srv/handler/helper.go | Go |
<|file_sep|>package.json.diff
original:
"name": "signature-pad",
updated:
"name": "@envoy/react-native-signature-pad",
<|file_sep|>original/package.json
{
"name": "signature-pad",
"version": "0.0.1",
"main": "dist/index.js",
"types": "dist/index.d.ts",
"keywords": [
"react-native"
],
"license": "MIT",
"author": {
"name": "Fang-Pen Lin",
"email": "hello@fangpenlin.com",
"url": "https://fangpenlin.com"
},
"rnpm": {},
"devDependencies": {
"@types/node": "^8.0.26",
"@types/react": "^16.0.5",
"@types/react-native": "^0.47.8",
"tslint": "^5.7.0",
"tslint-config-standard": "^6.0.1",
<|file_sep|>current/package.json
{
"name": "@envoy/react-native-signature-pad",
"version": "0.0.1",
"main": "dist/index.js",
"types": "dist/index.d.ts",
"keywords": [
"react-native"
],
"license": "MIT",
"author": {
"name": "Fang-Pen Lin",
"email": "hello@fangpenlin.com",
"url": "https://fangpenlin.com"
},
"rnpm": {},
"devDependencies": {
"@types/node": "^8.0.26",
"@types/react": "^16.0.5",
"@types/react-native": "^0.47.8",
"tslint": "^5.7.0",
"tslint-config-standard": "^6.0.1",
<|file_sep|>updated/package.json | {
"name": "@envoy/react-native-signature-pad",
"version": "0.0.1",
"main": "dist/index.js",
"types": "dist/index.d.ts",
"keywords": ["react-native"],
"license": "MIT",
"author": {
"name": "Fang-Pen Lin",
"email": "hello@fangpenlin.com",
"url": "https://fangpenlin.com"
},
"rnpm": {},
"devDependencies": {
"@types/node": "^8.0.26",
"@types/react": "^16.0.5",
"@types/react-native": "^0.47.8",
"tslint": "^5.7.0",
"tslint-config-standard": "^6.0.1",
"tslint-react": "^3.2.0",
"typescript": "^2.4.2" | <|file_sep|>package.json.diff
original:
"name": "signature-pad",
updated:
"name": "@envoy/react-native-signature-pad",
<|file_sep|>original/package.json
{
"name": "signature-pad",
"version": "0.0.1",
"main": "dist/index.js",
"types": "dist/index.d.ts",
"keywords": [
"react-native"
],
"license": "MIT",
"author": {
"name": "Fang-Pen Lin",
"email": "hello@fangpenlin.com",
"url": "https://fangpenlin.com"
},
"rnpm": {},
"devDependencies": {
"@types/node": "^8.0.26",
"@types/react": "^16.0.5",
"@types/react-native": "^0.47.8",
"tslint": "^5.7.0",
"tslint-config-standard": "^6.0.1",
<|file_sep|>current/package.json
{
"name": "@envoy/react-native-signature-pad",
"version": "0.0.1",
"main": "dist/index.js",
"types": "dist/index.d.ts",
"keywords": [
"react-native"
],
"license": "MIT",
"author": {
"name": "Fang-Pen Lin",
"email": "hello@fangpenlin.com",
"url": "https://fangpenlin.com"
},
"rnpm": {},
"devDependencies": {
"@types/node": "^8.0.26",
"@types/react": "^16.0.5",
"@types/react-native": "^0.47.8",
"tslint": "^5.7.0",
"tslint-config-standard": "^6.0.1",
<|file_sep|>updated/package.json
{
"name": "@envoy/react-native-signature-pad",
"version": "0.0.1",
"main": "dist/index.js",
"types": "dist/index.d.ts",
"keywords": ["react-native"],
"license": "MIT",
"author": {
"name": "Fang-Pen Lin",
"email": "hello@fangpenlin.com",
"url": "https://fangpenlin.com"
},
"rnpm": {},
"devDependencies": {
"@types/node": "^8.0.26",
"@types/react": "^16.0.5",
"@types/react-native": "^0.47.8",
"tslint": "^5.7.0",
"tslint-config-standard": "^6.0.1",
"tslint-react": "^3.2.0",
"typescript": "^2.4.2" | 58d4d2ae39a3985480b3898cc365522f7f6cd616 | package.json | package.json | JSON |
<|file_sep|>lib/nehm/artwork.rb.diff
original:
end
def url
hash = @track.hash
url =
if hash['artwork_url'].nil?
hash['user']['avatar_url']
else
hash['artwork_url']
end
url.sub('large', 't500x500')
updated:
<|file_sep|>original/lib/nehm/artwork.rb
# Artwork objects contains all needed information of track's artwork
class Artwork
def initialize(track)
@track = track
end
def url
hash = @track.hash
url =
if hash['artwork_url'].nil?
hash['user']['avatar_url']
else
hash['artwork_url']
end
url.sub('large', 't500x500')
end
# Use in Get.dl
def name
'artwork'
end
<|file_sep|>current/lib/nehm/artwork.rb
# Artwork objects contains all needed information of track's artwork
class Artwork
def initialize(track)
@track = track
end
# Use in Get.dl
def name
'artwork'
end
def file_path
File.join('/tmp', "#{@track.id}.jpg")
end
def suicide
File.delete(file_path)
end
end
<|file_sep|>updated/lib/nehm/artwork.rb | 'artwork'
end
def file_path
File.join('/tmp', "#{@track.id}.jpg")
end
def suicide
File.delete(file_path)
end
def url
hash = @track.hash
url =
if hash['artwork_url'].nil?
hash['user']['avatar_url']
else
hash['artwork_url']
end
url.sub('large', 't500x500')
end | <|file_sep|>lib/nehm/artwork.rb.diff
original:
end
def url
hash = @track.hash
url =
if hash['artwork_url'].nil?
hash['user']['avatar_url']
else
hash['artwork_url']
end
url.sub('large', 't500x500')
updated:
<|file_sep|>original/lib/nehm/artwork.rb
# Artwork objects contains all needed information of track's artwork
class Artwork
def initialize(track)
@track = track
end
def url
hash = @track.hash
url =
if hash['artwork_url'].nil?
hash['user']['avatar_url']
else
hash['artwork_url']
end
url.sub('large', 't500x500')
end
# Use in Get.dl
def name
'artwork'
end
<|file_sep|>current/lib/nehm/artwork.rb
# Artwork objects contains all needed information of track's artwork
class Artwork
def initialize(track)
@track = track
end
# Use in Get.dl
def name
'artwork'
end
def file_path
File.join('/tmp', "#{@track.id}.jpg")
end
def suicide
File.delete(file_path)
end
end
<|file_sep|>updated/lib/nehm/artwork.rb
'artwork'
end
def file_path
File.join('/tmp', "#{@track.id}.jpg")
end
def suicide
File.delete(file_path)
end
def url
hash = @track.hash
url =
if hash['artwork_url'].nil?
hash['user']['avatar_url']
else
hash['artwork_url']
end
url.sub('large', 't500x500')
end | d162ee18b9a250399a178725147262bf635e557d | lib/nehm/artwork.rb | lib/nehm/artwork.rb | Ruby |
<|file_sep|>original/src/zeit/push/tests/fixtures/authors.json
<|file_sep|>current/src/zeit/push/tests/fixtures/authors.json
<|file_sep|>updated/src/zeit/push/tests/fixtures/authors.json | {% macro author_tags() -%}
{% for uuid in author_push_uuids %}
{% if loop.last %}
{"tag": "{{ uuid }}", "group": "authors"}
{% else %}
{"tag": "{{ uuid }}", "group": "authors"},
{% endif %}
{% endfor %}
{%- endmacro %}
{
"default_title": "Autorenpush",
"messages": [
{
"device_types": [
"android"
],
"notification": {
"android": {
"actions": {
"open": {
"content": "{{app_link}}?wt_zmc=fix.int.zonaudev.push.authorpush.zeitde.andpush.link.x&utm_campaign=authorpush&utm_medium=fix&utm_source=push_zonaudev_int&utm_content=zeitde_andpush_link_x", | <|file_sep|>original/src/zeit/push/tests/fixtures/authors.json
<|file_sep|>current/src/zeit/push/tests/fixtures/authors.json
<|file_sep|>updated/src/zeit/push/tests/fixtures/authors.json
{% macro author_tags() -%}
{% for uuid in author_push_uuids %}
{% if loop.last %}
{"tag": "{{ uuid }}", "group": "authors"}
{% else %}
{"tag": "{{ uuid }}", "group": "authors"},
{% endif %}
{% endfor %}
{%- endmacro %}
{
"default_title": "Autorenpush",
"messages": [
{
"device_types": [
"android"
],
"notification": {
"android": {
"actions": {
"open": {
"content": "{{app_link}}?wt_zmc=fix.int.zonaudev.push.authorpush.zeitde.andpush.link.x&utm_campaign=authorpush&utm_medium=fix&utm_source=push_zonaudev_int&utm_content=zeitde_andpush_link_x", | cc8218c87ba561ed6428a01fc89ae48abab65508 | src/zeit/push/tests/fixtures/authors.json | src/zeit/push/tests/fixtures/authors.json | JSON |
<|file_sep|>original/metadata/com.aurora.corona.yml
gradle:
- yes
- versionName: 1.0.2
versionCode: 2
commit: 1.0.2
subdir: app
gradle:
- yes
- versionName: 1.0.3
versionCode: 3
commit: 1.0.3
subdir: app
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 1.0.2
CurrentVersionCode: 2
<|file_sep|>current/metadata/com.aurora.corona.yml
gradle:
- yes
- versionName: 1.0.2
versionCode: 2
commit: 1.0.2
subdir: app
gradle:
- yes
- versionName: 1.0.3
versionCode: 3
commit: 1.0.3
subdir: app
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 1.0.2
CurrentVersionCode: 2
<|file_sep|>updated/metadata/com.aurora.corona.yml | gradle:
- yes
- versionName: 1.0.2
versionCode: 2
commit: 1.0.2
subdir: app
gradle:
- yes
- versionName: 1.0.3
versionCode: 3
commit: 1.0.3
subdir: app
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 1.0.3
CurrentVersionCode: 3 | <|file_sep|>original/metadata/com.aurora.corona.yml
gradle:
- yes
- versionName: 1.0.2
versionCode: 2
commit: 1.0.2
subdir: app
gradle:
- yes
- versionName: 1.0.3
versionCode: 3
commit: 1.0.3
subdir: app
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 1.0.2
CurrentVersionCode: 2
<|file_sep|>current/metadata/com.aurora.corona.yml
gradle:
- yes
- versionName: 1.0.2
versionCode: 2
commit: 1.0.2
subdir: app
gradle:
- yes
- versionName: 1.0.3
versionCode: 3
commit: 1.0.3
subdir: app
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 1.0.2
CurrentVersionCode: 2
<|file_sep|>updated/metadata/com.aurora.corona.yml
gradle:
- yes
- versionName: 1.0.2
versionCode: 2
commit: 1.0.2
subdir: app
gradle:
- yes
- versionName: 1.0.3
versionCode: 3
commit: 1.0.3
subdir: app
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 1.0.3
CurrentVersionCode: 3 | fbaf604589b18e920513cc5567c1dda839ad678e | metadata/com.aurora.corona.yml | metadata/com.aurora.corona.yml | YAML |
<|file_sep|>original/.travis.yml
language: node_js
node_js:
- '7'
- '8'
- '9'
- '10'
after_success:
- npm run coveralls
<|file_sep|>current/.travis.yml
language: node_js
node_js:
- '7'
- '8'
- '9'
- '10'
after_success:
- npm run coveralls
<|file_sep|>updated/.travis.yml | language: node_js
node_js:
- '7'
- '8'
- '9'
- '10'
- '11'
after_success:
- npm run coveralls | <|file_sep|>original/.travis.yml
language: node_js
node_js:
- '7'
- '8'
- '9'
- '10'
after_success:
- npm run coveralls
<|file_sep|>current/.travis.yml
language: node_js
node_js:
- '7'
- '8'
- '9'
- '10'
after_success:
- npm run coveralls
<|file_sep|>updated/.travis.yml
language: node_js
node_js:
- '7'
- '8'
- '9'
- '10'
- '11'
after_success:
- npm run coveralls | 5a1442981e197dee7363a0ae3936aaf0942bb1f5 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/theme/partials/header.hbs
<div role='banner' class="Yayul_header {{#if @blog.cover}}" style="background-image: url({{@blog.cover}}){{else}}no-cover{{/if}}">
<div class="Yayul_header__content">
{{#if @blog.logo}}
<a class="Yayul_header__logo" href="{{@blog.url}}">
<img src="{{@blog.logo}}" alt="{{@blog.title}}" />
</a>
{{/if}}
<h1 class="Yayul_header__title">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{{@blog.title}}}
</a>
</h1>
<p class="Yayul_header__description">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{@blog.description}}
</a>
</p>
</div>
<|file_sep|>current/theme/partials/header.hbs
<div role='banner' class="Yayul_header {{#if @blog.cover}}" style="background-image: url({{@blog.cover}}){{else}}no-cover{{/if}}">
<div class="Yayul_header__content">
{{#if @blog.logo}}
<a class="Yayul_header__logo" href="{{@blog.url}}">
<img src="{{@blog.logo}}" alt="{{@blog.title}}" />
</a>
{{/if}}
<h1 class="Yayul_header__title">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{{@blog.title}}}
</a>
</h1>
<p class="Yayul_header__description">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{@blog.description}}
</a>
</p>
</div>
<|file_sep|>updated/theme/partials/header.hbs | <div role='banner' class="Yayul_header
{{#if image}} style="background-image: url({{image}})
{{else if @blog.cover}} style="background-image: url({{@blog.cover}})
{{else}}no-cover{{/if}}">
<div class="Yayul_header__content">
{{#if @blog.logo}}
<a class="Yayul_header__logo" href="{{@blog.url}}">
<img src="{{@blog.logo}}" alt="{{@blog.title}}" />
</a>
{{/if}}
<h1 class="Yayul_header__title">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{{@blog.title}}}
</a>
</h1>
<p class="Yayul_header__description">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{@blog.description}}
</a> | <|file_sep|>original/theme/partials/header.hbs
<div role='banner' class="Yayul_header {{#if @blog.cover}}" style="background-image: url({{@blog.cover}}){{else}}no-cover{{/if}}">
<div class="Yayul_header__content">
{{#if @blog.logo}}
<a class="Yayul_header__logo" href="{{@blog.url}}">
<img src="{{@blog.logo}}" alt="{{@blog.title}}" />
</a>
{{/if}}
<h1 class="Yayul_header__title">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{{@blog.title}}}
</a>
</h1>
<p class="Yayul_header__description">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{@blog.description}}
</a>
</p>
</div>
<|file_sep|>current/theme/partials/header.hbs
<div role='banner' class="Yayul_header {{#if @blog.cover}}" style="background-image: url({{@blog.cover}}){{else}}no-cover{{/if}}">
<div class="Yayul_header__content">
{{#if @blog.logo}}
<a class="Yayul_header__logo" href="{{@blog.url}}">
<img src="{{@blog.logo}}" alt="{{@blog.title}}" />
</a>
{{/if}}
<h1 class="Yayul_header__title">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{{@blog.title}}}
</a>
</h1>
<p class="Yayul_header__description">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{@blog.description}}
</a>
</p>
</div>
<|file_sep|>updated/theme/partials/header.hbs
<div role='banner' class="Yayul_header
{{#if image}} style="background-image: url({{image}})
{{else if @blog.cover}} style="background-image: url({{@blog.cover}})
{{else}}no-cover{{/if}}">
<div class="Yayul_header__content">
{{#if @blog.logo}}
<a class="Yayul_header__logo" href="{{@blog.url}}">
<img src="{{@blog.logo}}" alt="{{@blog.title}}" />
</a>
{{/if}}
<h1 class="Yayul_header__title">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{{@blog.title}}}
</a>
</h1>
<p class="Yayul_header__description">
<a title="{{@blog.title}}" href='{{@blog.url}}'>
{{@blog.description}}
</a> | 2ba81a317be4844520ee3fdcf9ec1f6af510ae95 | theme/partials/header.hbs | theme/partials/header.hbs | Handlebars |
<|file_sep|>original/common/model/host.go
package model
import (
"fmt"
)
type Host struct {
Id int
Name string
}
func (this *Host) String() string {
return fmt.Sprintf(
"<id:%s,name:%s>",
this.Id,
this.Name,
)
}
<|file_sep|>current/common/model/host.go
package model
import (
"fmt"
)
type Host struct {
Id int
Name string
}
func (this *Host) String() string {
return fmt.Sprintf(
"<id:%s,name:%s>",
this.Id,
this.Name,
)
}
<|file_sep|>updated/common/model/host.go | package model
import (
"fmt"
)
type Host struct {
Id int
Name string
}
func (this *Host) String() string {
return fmt.Sprintf(
"<id:%d,name:%s>",
this.Id,
this.Name,
)
} | <|file_sep|>original/common/model/host.go
package model
import (
"fmt"
)
type Host struct {
Id int
Name string
}
func (this *Host) String() string {
return fmt.Sprintf(
"<id:%s,name:%s>",
this.Id,
this.Name,
)
}
<|file_sep|>current/common/model/host.go
package model
import (
"fmt"
)
type Host struct {
Id int
Name string
}
func (this *Host) String() string {
return fmt.Sprintf(
"<id:%s,name:%s>",
this.Id,
this.Name,
)
}
<|file_sep|>updated/common/model/host.go
package model
import (
"fmt"
)
type Host struct {
Id int
Name string
}
func (this *Host) String() string {
return fmt.Sprintf(
"<id:%d,name:%s>",
this.Id,
this.Name,
)
} | 7b08eaef24cbded87eea3a47dc76cabcec73bbfe | common/model/host.go | common/model/host.go | Go |
<|file_sep|>original/Formula/barty_crouch.rb
<|file_sep|>current/Formula/barty_crouch.rb
<|file_sep|>updated/Formula/barty_crouch.rb | class BartyCrouch < Formula
desc "Localization/I18n: Incrementally update/translate your Strings files from .swift, .h, .m(m), .storyboard or .xib files."
homepage "https://github.com/Flinesoft/BartyCrouch"
url "https://github.com/Flinesoft/BartyCrouch.git", :tag => "4.0.0-alpha.1", :revision => "4c27662f0800bea9263748fe4e62c163ea9de7f6"
head "https://github.com/Flinesoft/BartyCrouch.git"
depends_on :xcode => ["10.0", :build]
def install
system "make", "install", "prefix=#{prefix}"
end
test do
system "#{bin}/bartycrouch"
end
end | <|file_sep|>original/Formula/barty_crouch.rb
<|file_sep|>current/Formula/barty_crouch.rb
<|file_sep|>updated/Formula/barty_crouch.rb
class BartyCrouch < Formula
desc "Localization/I18n: Incrementally update/translate your Strings files from .swift, .h, .m(m), .storyboard or .xib files."
homepage "https://github.com/Flinesoft/BartyCrouch"
url "https://github.com/Flinesoft/BartyCrouch.git", :tag => "4.0.0-alpha.1", :revision => "4c27662f0800bea9263748fe4e62c163ea9de7f6"
head "https://github.com/Flinesoft/BartyCrouch.git"
depends_on :xcode => ["10.0", :build]
def install
system "make", "install", "prefix=#{prefix}"
end
test do
system "#{bin}/bartycrouch"
end
end | 6a784a28036e48fd967dd0d0d72ea871af5d3db7 | Formula/barty_crouch.rb | Formula/barty_crouch.rb | Ruby |
<|file_sep|>original/.travis.yml
language: node_js
node_js:
- "0.11"
- "0.10"
- "0.8"
- "0.6"
<|file_sep|>current/.travis.yml
language: node_js
node_js:
- "0.11"
- "0.10"
- "0.8"
- "0.6"
<|file_sep|>updated/.travis.yml | language: node_js
node_js:
- "0.11"
- "0.10"
- "0.8" | <|file_sep|>original/.travis.yml
language: node_js
node_js:
- "0.11"
- "0.10"
- "0.8"
- "0.6"
<|file_sep|>current/.travis.yml
language: node_js
node_js:
- "0.11"
- "0.10"
- "0.8"
- "0.6"
<|file_sep|>updated/.travis.yml
language: node_js
node_js:
- "0.11"
- "0.10"
- "0.8" | 86e9902dd883461fa9df10df8a0e93a945251564 | .travis.yml | .travis.yml | YAML |
<|file_sep|>test/src/index.js.diff
original:
it('should add route', function() {
updated:
it('should add route without name', function() {
<|file_sep|>test/src/index.js.diff
original:
updated:
test.string(route.name).is(method + '_' + path);
<|file_sep|>original/test/src/index.js
let test = require('unit.js');
let router = require('../../src');
describe('Router ', function() {
it('should add route', function() {
let route;
let fn = function(){};
let path = '/my-path';
let method = 'GET';
router.use({method, path, fn});
test.number(Object.keys(router._routes).length).is(1);
test.bool(router._routes.hasOwnProperty(method + '_' + path)).isTrue();
route = router._routes[method + '_' + path];
test.string(route.method).is(method);
test.string(route.path).is(path);
test.function(route.fn).is(fn);
});
});
<|file_sep|>current/test/src/index.js
let test = require('unit.js');
let router = require('../../src');
describe('Router ', function() {
it('should add route without name', function() {
let route;
let fn = function(){};
let path = '/my-path';
let method = 'GET';
router.use({method, path, fn});
test.number(Object.keys(router._routes).length).is(1);
test.bool(router._routes.hasOwnProperty(method + '_' + path)).isTrue();
route = router._routes[method + '_' + path];
test.string(route.method).is(method);
test.string(route.path).is(path);
test.string(route.name).is(method + '_' + path);
test.function(route.fn).is(fn);
});
});
<|file_sep|>updated/test/src/index.js |
test.number(Object.keys(router._routes).length).is(1);
test.bool(router._routes.hasOwnProperty(method + '_' + path)).isTrue();
route = router._routes[method + '_' + path];
test.string(route.method).is(method);
test.string(route.path).is(path);
test.string(route.name).is(method + '_' + path);
test.function(route.fn).is(fn);
});
it('should add route with name', function() {
let route;
let fn = function(){};
let path = '/my-path2';
let method = 'POST';
let name = 'myRouteName';
router.use({method, path, fn, name});
test.number(Object.keys(router._routes).length).is(2);
test.bool(router._routes.hasOwnProperty(method + '_' + path)).isTrue(); | <|file_sep|>test/src/index.js.diff
original:
it('should add route', function() {
updated:
it('should add route without name', function() {
<|file_sep|>test/src/index.js.diff
original:
updated:
test.string(route.name).is(method + '_' + path);
<|file_sep|>original/test/src/index.js
let test = require('unit.js');
let router = require('../../src');
describe('Router ', function() {
it('should add route', function() {
let route;
let fn = function(){};
let path = '/my-path';
let method = 'GET';
router.use({method, path, fn});
test.number(Object.keys(router._routes).length).is(1);
test.bool(router._routes.hasOwnProperty(method + '_' + path)).isTrue();
route = router._routes[method + '_' + path];
test.string(route.method).is(method);
test.string(route.path).is(path);
test.function(route.fn).is(fn);
});
});
<|file_sep|>current/test/src/index.js
let test = require('unit.js');
let router = require('../../src');
describe('Router ', function() {
it('should add route without name', function() {
let route;
let fn = function(){};
let path = '/my-path';
let method = 'GET';
router.use({method, path, fn});
test.number(Object.keys(router._routes).length).is(1);
test.bool(router._routes.hasOwnProperty(method + '_' + path)).isTrue();
route = router._routes[method + '_' + path];
test.string(route.method).is(method);
test.string(route.path).is(path);
test.string(route.name).is(method + '_' + path);
test.function(route.fn).is(fn);
});
});
<|file_sep|>updated/test/src/index.js
test.number(Object.keys(router._routes).length).is(1);
test.bool(router._routes.hasOwnProperty(method + '_' + path)).isTrue();
route = router._routes[method + '_' + path];
test.string(route.method).is(method);
test.string(route.path).is(path);
test.string(route.name).is(method + '_' + path);
test.function(route.fn).is(fn);
});
it('should add route with name', function() {
let route;
let fn = function(){};
let path = '/my-path2';
let method = 'POST';
let name = 'myRouteName';
router.use({method, path, fn, name});
test.number(Object.keys(router._routes).length).is(2);
test.bool(router._routes.hasOwnProperty(method + '_' + path)).isTrue(); | 26a034f0c0c016e932e84946c4040c9215b9e4c2 | test/src/index.js | test/src/index.js | JavaScript |
<|file_sep|>original/app/views/posts/_image_upload.html.erb
<%= form_tag upload_path, multipart: true, class: 'form-inline upload-form js-upload-form' do %>
<%= hidden_field_tag :authenticity_token, form_authenticity_token %>
<div class="form-group">
<%= label_tag :file, 'Insert an image' %>
<%= file_field_tag :file %>
<span class="form-text text-muted">Max file size <%= SiteSetting['MaxUploadSize'] %>.</span>
</div>
<% end %>
<|file_sep|>current/app/views/posts/_image_upload.html.erb
<%= form_tag upload_path, multipart: true, class: 'form-inline upload-form js-upload-form' do %>
<%= hidden_field_tag :authenticity_token, form_authenticity_token %>
<div class="form-group">
<%= label_tag :file, 'Insert an image' %>
<%= file_field_tag :file %>
<span class="form-text text-muted">Max file size <%= SiteSetting['MaxUploadSize'] %>.</span>
</div>
<% end %>
<|file_sep|>updated/app/views/posts/_image_upload.html.erb | <%= form_tag upload_path, multipart: true, class: 'form-inline upload-form js-upload-form' do %>
<%= hidden_field_tag :authenticity_token, form_authenticity_token %>
<div class="form-group">
<%= label_tag :file, 'Insert an image', class: "form-element" %>
<div class="has-color-tertiary-500 has-font-size-caption">You can upload files, which are smaller than <%= SiteSetting['MaxUploadSize'] %>.</div>
<%= file_field_tag :file, class: "form-element" %>
</div>
<% end %> | <|file_sep|>original/app/views/posts/_image_upload.html.erb
<%= form_tag upload_path, multipart: true, class: 'form-inline upload-form js-upload-form' do %>
<%= hidden_field_tag :authenticity_token, form_authenticity_token %>
<div class="form-group">
<%= label_tag :file, 'Insert an image' %>
<%= file_field_tag :file %>
<span class="form-text text-muted">Max file size <%= SiteSetting['MaxUploadSize'] %>.</span>
</div>
<% end %>
<|file_sep|>current/app/views/posts/_image_upload.html.erb
<%= form_tag upload_path, multipart: true, class: 'form-inline upload-form js-upload-form' do %>
<%= hidden_field_tag :authenticity_token, form_authenticity_token %>
<div class="form-group">
<%= label_tag :file, 'Insert an image' %>
<%= file_field_tag :file %>
<span class="form-text text-muted">Max file size <%= SiteSetting['MaxUploadSize'] %>.</span>
</div>
<% end %>
<|file_sep|>updated/app/views/posts/_image_upload.html.erb
<%= form_tag upload_path, multipart: true, class: 'form-inline upload-form js-upload-form' do %>
<%= hidden_field_tag :authenticity_token, form_authenticity_token %>
<div class="form-group">
<%= label_tag :file, 'Insert an image', class: "form-element" %>
<div class="has-color-tertiary-500 has-font-size-caption">You can upload files, which are smaller than <%= SiteSetting['MaxUploadSize'] %>.</div>
<%= file_field_tag :file, class: "form-element" %>
</div>
<% end %> | 407e3244160773f0824d508c46a042c85527af14 | app/views/posts/_image_upload.html.erb | app/views/posts/_image_upload.html.erb | HTML+ERB |
<|file_sep|>original/.travis.yml
language: python
python:
- "2.7"
env:
global:
- SECRET_KEY=blajblajbasdf
matrix:
- DJANGO=1.5
- DJANGO=1.6
install:
- pip install -r requirements.txt --use-mirrors
script: python manage.py test classphoto groups mail sequence signup unsubscribe
notifications:
email:
on_success: change
on_failure: change
<|file_sep|>current/.travis.yml
language: python
python:
- "2.7"
env:
global:
- SECRET_KEY=blajblajbasdf
matrix:
- DJANGO=1.5
- DJANGO=1.6
install:
- pip install -r requirements.txt --use-mirrors
script: python manage.py test classphoto groups mail sequence signup unsubscribe
notifications:
email:
on_success: change
on_failure: change
<|file_sep|>updated/.travis.yml | language: python
python:
- "2.7"
env:
global:
- SECRET_KEY=blajblajbasdf
matrix:
- DJANGO=1.5
- DJANGO=1.6
install:
- pip install -r requirements.txt
script: python manage.py test classphoto groups mail sequence signup unsubscribe
notifications:
email:
on_success: change
on_failure: change | <|file_sep|>original/.travis.yml
language: python
python:
- "2.7"
env:
global:
- SECRET_KEY=blajblajbasdf
matrix:
- DJANGO=1.5
- DJANGO=1.6
install:
- pip install -r requirements.txt --use-mirrors
script: python manage.py test classphoto groups mail sequence signup unsubscribe
notifications:
email:
on_success: change
on_failure: change
<|file_sep|>current/.travis.yml
language: python
python:
- "2.7"
env:
global:
- SECRET_KEY=blajblajbasdf
matrix:
- DJANGO=1.5
- DJANGO=1.6
install:
- pip install -r requirements.txt --use-mirrors
script: python manage.py test classphoto groups mail sequence signup unsubscribe
notifications:
email:
on_success: change
on_failure: change
<|file_sep|>updated/.travis.yml
language: python
python:
- "2.7"
env:
global:
- SECRET_KEY=blajblajbasdf
matrix:
- DJANGO=1.5
- DJANGO=1.6
install:
- pip install -r requirements.txt
script: python manage.py test classphoto groups mail sequence signup unsubscribe
notifications:
email:
on_success: change
on_failure: change | 43105e2abef9fcf259a40c4c1f926fbd893580ba | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/core/examples/dcmdump.rs
//! WIP
extern crate dicom_core;
use dicom_core::{DicomElement, DicomObject};
use std::fs::File;
use std::path::Path;
fn main() {
println!("Sorry, this tool is not implemented yet.");
}
fn dump<P: AsRef<Path>>(path: P) {
let file = File::open(path).unwrap();
//let obj = DicomLoader.load(file).unwrap();
//for elem in obj {
// dump_element(&elem);
//}
}
fn dump_element(elem: &DicomElement) {}
<|file_sep|>current/core/examples/dcmdump.rs
//! WIP
extern crate dicom_core;
use dicom_core::{DicomElement, DicomObject};
use std::fs::File;
use std::path::Path;
fn main() {
println!("Sorry, this tool is not implemented yet.");
}
fn dump<P: AsRef<Path>>(path: P) {
let file = File::open(path).unwrap();
//let obj = DicomLoader.load(file).unwrap();
//for elem in obj {
// dump_element(&elem);
//}
}
fn dump_element(elem: &DicomElement) {}
<|file_sep|>updated/core/examples/dcmdump.rs | //! A reimplementation of dcmdump in Rust
//! WIP
extern crate dicom_core;
use dicom_core::{DicomElement, DicomObject};
use std::fs::File;
use std::path::Path;
fn main() {
println!("Sorry, this tool is not implemented yet.");
}
fn dump<P: AsRef<Path>>(path: P) {
let file = File::open(path).unwrap();
//let obj = DicomLoader.load(file).unwrap();
//for elem in obj {
// dump_element(&elem);
//}
} | <|file_sep|>original/core/examples/dcmdump.rs
//! WIP
extern crate dicom_core;
use dicom_core::{DicomElement, DicomObject};
use std::fs::File;
use std::path::Path;
fn main() {
println!("Sorry, this tool is not implemented yet.");
}
fn dump<P: AsRef<Path>>(path: P) {
let file = File::open(path).unwrap();
//let obj = DicomLoader.load(file).unwrap();
//for elem in obj {
// dump_element(&elem);
//}
}
fn dump_element(elem: &DicomElement) {}
<|file_sep|>current/core/examples/dcmdump.rs
//! WIP
extern crate dicom_core;
use dicom_core::{DicomElement, DicomObject};
use std::fs::File;
use std::path::Path;
fn main() {
println!("Sorry, this tool is not implemented yet.");
}
fn dump<P: AsRef<Path>>(path: P) {
let file = File::open(path).unwrap();
//let obj = DicomLoader.load(file).unwrap();
//for elem in obj {
// dump_element(&elem);
//}
}
fn dump_element(elem: &DicomElement) {}
<|file_sep|>updated/core/examples/dcmdump.rs
//! A reimplementation of dcmdump in Rust
//! WIP
extern crate dicom_core;
use dicom_core::{DicomElement, DicomObject};
use std::fs::File;
use std::path::Path;
fn main() {
println!("Sorry, this tool is not implemented yet.");
}
fn dump<P: AsRef<Path>>(path: P) {
let file = File::open(path).unwrap();
//let obj = DicomLoader.load(file).unwrap();
//for elem in obj {
// dump_element(&elem);
//}
} | 428b457c5878261baef5311eb11198a45d12e111 | core/examples/dcmdump.rs | core/examples/dcmdump.rs | Rust |
<|file_sep|>db/migrate/20151124210620_create_oauth_client_redirect_uris.rb.diff
original:
t.integer :client_id
updated:
<|file_sep|>original/db/migrate/20151124210620_create_oauth_client_redirect_uris.rb
class CreateOauthClientRedirectUris < ActiveRecord::Migration
def change
create_table :oauth_client_redirect_uris do |t|
t.integer :client_id
t.string :uri
t.timestamps
end
add_index :oauth_client_redirect_uris, :client_id
end
end
<|file_sep|>current/db/migrate/20151124210620_create_oauth_client_redirect_uris.rb
class CreateOauthClientRedirectUris < ActiveRecord::Migration
def change
create_table :oauth_client_redirect_uris do |t|
t.string :uri
t.timestamps
end
add_index :oauth_client_redirect_uris, :client_id
end
end
<|file_sep|>updated/db/migrate/20151124210620_create_oauth_client_redirect_uris.rb | class CreateOauthClientRedirectUris < ActiveRecord::Migration
def change
create_table :oauth_client_redirect_uris do |t|
t.string :uri
t.timestamps
end
add_reference :oauth_client_redirect_uris, :client, index: true
end
end | <|file_sep|>db/migrate/20151124210620_create_oauth_client_redirect_uris.rb.diff
original:
t.integer :client_id
updated:
<|file_sep|>original/db/migrate/20151124210620_create_oauth_client_redirect_uris.rb
class CreateOauthClientRedirectUris < ActiveRecord::Migration
def change
create_table :oauth_client_redirect_uris do |t|
t.integer :client_id
t.string :uri
t.timestamps
end
add_index :oauth_client_redirect_uris, :client_id
end
end
<|file_sep|>current/db/migrate/20151124210620_create_oauth_client_redirect_uris.rb
class CreateOauthClientRedirectUris < ActiveRecord::Migration
def change
create_table :oauth_client_redirect_uris do |t|
t.string :uri
t.timestamps
end
add_index :oauth_client_redirect_uris, :client_id
end
end
<|file_sep|>updated/db/migrate/20151124210620_create_oauth_client_redirect_uris.rb
class CreateOauthClientRedirectUris < ActiveRecord::Migration
def change
create_table :oauth_client_redirect_uris do |t|
t.string :uri
t.timestamps
end
add_reference :oauth_client_redirect_uris, :client, index: true
end
end | 79785de430571443671c443fc1ad2d8f9de966e1 | db/migrate/20151124210620_create_oauth_client_redirect_uris.rb | db/migrate/20151124210620_create_oauth_client_redirect_uris.rb | Ruby |
<|file_sep|>original/site_scons/get_libs.py
def get_lib_paths():
if sys.platform == 'win32':
lib_paths = set(os.environ['PATH'].split(';'))
else:
lib_paths = set()
if os.environ.has_key('LIBRARY_PATH'):
lib_paths.update(os.environ['LIBRARY_PATH'].split(':'))
if os.environ.has_key('LD_LIBRARY_PATH'):
lib_paths.update(os.environ['LD_LIBRARY_PATH'].split(':'))
lib_paths = (['/usr/lib', '/usr/lib/x86_64-linux-gnu',
'/usr/local/lib'] + list(lib_paths))
return lib_paths
def get_lib(lib_name, LIBPATH=None):
if not LIBPATH:
LIBPATH = []
else:
LIBPATH = LIBPATH[:]
LIBPATH += get_lib_paths()
<|file_sep|>current/site_scons/get_libs.py
def get_lib_paths():
if sys.platform == 'win32':
lib_paths = set(os.environ['PATH'].split(';'))
else:
lib_paths = set()
if os.environ.has_key('LIBRARY_PATH'):
lib_paths.update(os.environ['LIBRARY_PATH'].split(':'))
if os.environ.has_key('LD_LIBRARY_PATH'):
lib_paths.update(os.environ['LD_LIBRARY_PATH'].split(':'))
lib_paths = (['/usr/lib', '/usr/lib/x86_64-linux-gnu',
'/usr/local/lib'] + list(lib_paths))
return lib_paths
def get_lib(lib_name, LIBPATH=None):
if not LIBPATH:
LIBPATH = []
else:
LIBPATH = LIBPATH[:]
LIBPATH += get_lib_paths()
<|file_sep|>updated/site_scons/get_libs.py |
def get_lib_paths():
if sys.platform == 'win32':
lib_paths = set(os.environ['PATH'].split(';'))
else:
lib_paths = set()
if os.environ.has_key('LIBRARY_PATH'):
lib_paths.update(os.environ['LIBRARY_PATH'].split(':'))
if os.environ.has_key('LD_LIBRARY_PATH'):
lib_paths.update(os.environ['LD_LIBRARY_PATH'].split(':'))
lib_paths = (['/usr/lib', '/usr/lib/i386-linux-gnu',
'/usr/lib/x86_64-linux-gnu', '/usr/local/lib'] +
list(lib_paths))
return lib_paths
def get_lib(lib_name, LIBPATH=None):
if not LIBPATH:
LIBPATH = []
else:
LIBPATH = LIBPATH[:] | <|file_sep|>original/site_scons/get_libs.py
def get_lib_paths():
if sys.platform == 'win32':
lib_paths = set(os.environ['PATH'].split(';'))
else:
lib_paths = set()
if os.environ.has_key('LIBRARY_PATH'):
lib_paths.update(os.environ['LIBRARY_PATH'].split(':'))
if os.environ.has_key('LD_LIBRARY_PATH'):
lib_paths.update(os.environ['LD_LIBRARY_PATH'].split(':'))
lib_paths = (['/usr/lib', '/usr/lib/x86_64-linux-gnu',
'/usr/local/lib'] + list(lib_paths))
return lib_paths
def get_lib(lib_name, LIBPATH=None):
if not LIBPATH:
LIBPATH = []
else:
LIBPATH = LIBPATH[:]
LIBPATH += get_lib_paths()
<|file_sep|>current/site_scons/get_libs.py
def get_lib_paths():
if sys.platform == 'win32':
lib_paths = set(os.environ['PATH'].split(';'))
else:
lib_paths = set()
if os.environ.has_key('LIBRARY_PATH'):
lib_paths.update(os.environ['LIBRARY_PATH'].split(':'))
if os.environ.has_key('LD_LIBRARY_PATH'):
lib_paths.update(os.environ['LD_LIBRARY_PATH'].split(':'))
lib_paths = (['/usr/lib', '/usr/lib/x86_64-linux-gnu',
'/usr/local/lib'] + list(lib_paths))
return lib_paths
def get_lib(lib_name, LIBPATH=None):
if not LIBPATH:
LIBPATH = []
else:
LIBPATH = LIBPATH[:]
LIBPATH += get_lib_paths()
<|file_sep|>updated/site_scons/get_libs.py
def get_lib_paths():
if sys.platform == 'win32':
lib_paths = set(os.environ['PATH'].split(';'))
else:
lib_paths = set()
if os.environ.has_key('LIBRARY_PATH'):
lib_paths.update(os.environ['LIBRARY_PATH'].split(':'))
if os.environ.has_key('LD_LIBRARY_PATH'):
lib_paths.update(os.environ['LD_LIBRARY_PATH'].split(':'))
lib_paths = (['/usr/lib', '/usr/lib/i386-linux-gnu',
'/usr/lib/x86_64-linux-gnu', '/usr/local/lib'] +
list(lib_paths))
return lib_paths
def get_lib(lib_name, LIBPATH=None):
if not LIBPATH:
LIBPATH = []
else:
LIBPATH = LIBPATH[:] | 40e9375f6b35b4a05ad311822705b7a7efe46b56 | site_scons/get_libs.py | site_scons/get_libs.py | Python |
<|file_sep|>original/shopify_auth/models.py
<|file_sep|>current/shopify_auth/models.py
<|file_sep|>updated/shopify_auth/models.py | from django.conf import settings
from django.db import models
from django.contrib.auth.models import AbstractBaseUser, BaseUserManager
class ShopUserManager(BaseUserManager):
def create_user(self, myshopify_domain, domain, password = None):
"""
Creates and saves a ShopUser with the given domains and password.
"""
if not myshopify_domain:
raise ValueError('ShopUsers must have a myshopify domain')
user = self.model(
myshopify_domain = myshopify_domain,
domain = domain,
)
# Never want to be able to log on externally.
# Authentication will be taken care of by Shopify OAuth. | <|file_sep|>original/shopify_auth/models.py
<|file_sep|>current/shopify_auth/models.py
<|file_sep|>updated/shopify_auth/models.py
from django.conf import settings
from django.db import models
from django.contrib.auth.models import AbstractBaseUser, BaseUserManager
class ShopUserManager(BaseUserManager):
def create_user(self, myshopify_domain, domain, password = None):
"""
Creates and saves a ShopUser with the given domains and password.
"""
if not myshopify_domain:
raise ValueError('ShopUsers must have a myshopify domain')
user = self.model(
myshopify_domain = myshopify_domain,
domain = domain,
)
# Never want to be able to log on externally.
# Authentication will be taken care of by Shopify OAuth. | a8b2930754fd4c62f2393e60826c2e089c6761a4 | shopify_auth/models.py | shopify_auth/models.py | Python |
<|file_sep|>original/requirements/production.txt
# Pro-tip: Try not to put anything here. There should be no dependency in
# production that isn't in development.
-r base.txt
# WSGI Handler
# ------------------------------------------------
gevent==20.5.0
gunicorn==19.7.1
# Static and Media Storage
# ------------------------------------------------
boto==2.48.0
django-storages-redux==1.3.3
Collectfast==0.6.0
# Mailgun Support
# ---------------
django-mailgun==0.9.1
# Redis
django-redis-cache==1.7.1
<|file_sep|>current/requirements/production.txt
# Pro-tip: Try not to put anything here. There should be no dependency in
# production that isn't in development.
-r base.txt
# WSGI Handler
# ------------------------------------------------
gevent==20.5.0
gunicorn==19.7.1
# Static and Media Storage
# ------------------------------------------------
boto==2.48.0
django-storages-redux==1.3.3
Collectfast==0.6.0
# Mailgun Support
# ---------------
django-mailgun==0.9.1
# Redis
django-redis-cache==1.7.1
<|file_sep|>updated/requirements/production.txt | # Pro-tip: Try not to put anything here. There should be no dependency in
# production that isn't in development.
-r base.txt
# WSGI Handler
# ------------------------------------------------
gevent==20.5.0
gunicorn==20.0.4
# Static and Media Storage
# ------------------------------------------------
boto==2.48.0
django-storages-redux==1.3.3
Collectfast==0.6.0
# Mailgun Support
# ---------------
django-mailgun==0.9.1
# Redis
django-redis-cache==1.7.1 | <|file_sep|>original/requirements/production.txt
# Pro-tip: Try not to put anything here. There should be no dependency in
# production that isn't in development.
-r base.txt
# WSGI Handler
# ------------------------------------------------
gevent==20.5.0
gunicorn==19.7.1
# Static and Media Storage
# ------------------------------------------------
boto==2.48.0
django-storages-redux==1.3.3
Collectfast==0.6.0
# Mailgun Support
# ---------------
django-mailgun==0.9.1
# Redis
django-redis-cache==1.7.1
<|file_sep|>current/requirements/production.txt
# Pro-tip: Try not to put anything here. There should be no dependency in
# production that isn't in development.
-r base.txt
# WSGI Handler
# ------------------------------------------------
gevent==20.5.0
gunicorn==19.7.1
# Static and Media Storage
# ------------------------------------------------
boto==2.48.0
django-storages-redux==1.3.3
Collectfast==0.6.0
# Mailgun Support
# ---------------
django-mailgun==0.9.1
# Redis
django-redis-cache==1.7.1
<|file_sep|>updated/requirements/production.txt
# Pro-tip: Try not to put anything here. There should be no dependency in
# production that isn't in development.
-r base.txt
# WSGI Handler
# ------------------------------------------------
gevent==20.5.0
gunicorn==20.0.4
# Static and Media Storage
# ------------------------------------------------
boto==2.48.0
django-storages-redux==1.3.3
Collectfast==0.6.0
# Mailgun Support
# ---------------
django-mailgun==0.9.1
# Redis
django-redis-cache==1.7.1 | 647833bc00e0a7653d22686025c5acbf8f5d0d09 | requirements/production.txt | requirements/production.txt | Text |
<|file_sep|>original/composer.json
{
"require": {
"drush/drush": "dev-master",
"fabpot/php-cs-fixer": "dev-master",
"phpunit/phpunit": "~4.1",
"sami/sami": "dev-master",
"drupal/coder": "~8.2"
}
}
<|file_sep|>current/composer.json
{
"require": {
"drush/drush": "dev-master",
"fabpot/php-cs-fixer": "dev-master",
"phpunit/phpunit": "~4.1",
"sami/sami": "dev-master",
"drupal/coder": "~8.2"
}
}
<|file_sep|>updated/composer.json | {
"require": {
"drush/drush": "~6.0",
"fabpot/php-cs-fixer": "dev-master",
"phpunit/phpunit": "~4.1",
"sami/sami": "dev-master",
"drupal/coder": "~8.2"
}
} | <|file_sep|>original/composer.json
{
"require": {
"drush/drush": "dev-master",
"fabpot/php-cs-fixer": "dev-master",
"phpunit/phpunit": "~4.1",
"sami/sami": "dev-master",
"drupal/coder": "~8.2"
}
}
<|file_sep|>current/composer.json
{
"require": {
"drush/drush": "dev-master",
"fabpot/php-cs-fixer": "dev-master",
"phpunit/phpunit": "~4.1",
"sami/sami": "dev-master",
"drupal/coder": "~8.2"
}
}
<|file_sep|>updated/composer.json
{
"require": {
"drush/drush": "~6.0",
"fabpot/php-cs-fixer": "dev-master",
"phpunit/phpunit": "~4.1",
"sami/sami": "dev-master",
"drupal/coder": "~8.2"
}
} | a37bd6448e71c49146247200acf846354d64034b | composer.json | composer.json | JSON |
<|file_sep|>original/requirements_dev.txt
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.5.0
coverage==4.4.1
tox==2.9.1
Sphinx==1.6.4
cryptography==2.1.2
pytest==3.2.3
PyYAML==3.12
pytest-runner==2.12.1
<|file_sep|>current/requirements_dev.txt
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.5.0
coverage==4.4.1
tox==2.9.1
Sphinx==1.6.4
cryptography==2.1.2
pytest==3.2.3
PyYAML==3.12
pytest-runner==2.12.1
<|file_sep|>updated/requirements_dev.txt | bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.5.0
coverage==4.4.1
tox==2.9.1
Sphinx==1.6.4
cryptography==2.1.3
pytest==3.2.3
PyYAML==3.12
pytest-runner==2.12.1 | <|file_sep|>original/requirements_dev.txt
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.5.0
coverage==4.4.1
tox==2.9.1
Sphinx==1.6.4
cryptography==2.1.2
pytest==3.2.3
PyYAML==3.12
pytest-runner==2.12.1
<|file_sep|>current/requirements_dev.txt
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.5.0
coverage==4.4.1
tox==2.9.1
Sphinx==1.6.4
cryptography==2.1.2
pytest==3.2.3
PyYAML==3.12
pytest-runner==2.12.1
<|file_sep|>updated/requirements_dev.txt
bumpversion==0.5.3
wheel==0.30.0
watchdog==0.8.3
flake8==3.5.0
coverage==4.4.1
tox==2.9.1
Sphinx==1.6.4
cryptography==2.1.3
pytest==3.2.3
PyYAML==3.12
pytest-runner==2.12.1 | b1568d4e67ec1e6a58375cbf15c1beb96597a916 | requirements_dev.txt | requirements_dev.txt | Text |
<|file_sep|>original/locales/da/webextension.properties
# Clip in this context refers to a portion of an image
# Also used for the title on the toolbar button
# The string "{meta_key}-V" should be translated to the region-specific
# shorthand for the Paste keyboard shortcut. {meta_key} is a placeholder for the
# modifier key used in the shortcut (do not translate it): for example, Ctrl-V
# on Windows systems.
# Section for error strings
# "Out of order" is an humorous way to indicate that the service is not working
# properly.
# Section for onboarding strings
# Do not translate {termsAndPrivacyNoticeTermsLink} and
# {termsAndPrivacyNoticyPrivacyLink}. They're placeholders replaced by links
# using the corresponding translated strings as text.
# This string is used as the text for a link in termsAndPrivacyNotice
# This string is used as the text for a link in termsAndPrivacyNotice
<|file_sep|>current/locales/da/webextension.properties
# Clip in this context refers to a portion of an image
# Also used for the title on the toolbar button
# The string "{meta_key}-V" should be translated to the region-specific
# shorthand for the Paste keyboard shortcut. {meta_key} is a placeholder for the
# modifier key used in the shortcut (do not translate it): for example, Ctrl-V
# on Windows systems.
# Section for error strings
# "Out of order" is an humorous way to indicate that the service is not working
# properly.
# Section for onboarding strings
# Do not translate {termsAndPrivacyNoticeTermsLink} and
# {termsAndPrivacyNoticyPrivacyLink}. They're placeholders replaced by links
# using the corresponding translated strings as text.
# This string is used as the text for a link in termsAndPrivacyNotice
# This string is used as the text for a link in termsAndPrivacyNotice
<|file_sep|>updated/locales/da/webextension.properties | # Clip in this context refers to a portion of an image
# Also used for the title on the toolbar button
# The string "{meta_key}-V" should be translated to the region-specific
# shorthand for the Paste keyboard shortcut. {meta_key} is a placeholder for the
# modifier key used in the shortcut (do not translate it): for example, Ctrl-V
# on Windows systems.
# Section for error strings
# "Out of order" is an humorous way to indicate that the service is not working
# properly.
# Section for onboarding strings
# Do not translate {termsAndPrivacyNoticeTermsLink} and
# {termsAndPrivacyNoticyPrivacyLink}. They're placeholders replaced by links
# using the corresponding translated strings as text.
# This string is used as the text for a link in termsAndPrivacyNoticeCloudServices
# This string is used as the text for a link in termsAndPrivacyNoticeCloudServices | <|file_sep|>original/locales/da/webextension.properties
# Clip in this context refers to a portion of an image
# Also used for the title on the toolbar button
# The string "{meta_key}-V" should be translated to the region-specific
# shorthand for the Paste keyboard shortcut. {meta_key} is a placeholder for the
# modifier key used in the shortcut (do not translate it): for example, Ctrl-V
# on Windows systems.
# Section for error strings
# "Out of order" is an humorous way to indicate that the service is not working
# properly.
# Section for onboarding strings
# Do not translate {termsAndPrivacyNoticeTermsLink} and
# {termsAndPrivacyNoticyPrivacyLink}. They're placeholders replaced by links
# using the corresponding translated strings as text.
# This string is used as the text for a link in termsAndPrivacyNotice
# This string is used as the text for a link in termsAndPrivacyNotice
<|file_sep|>current/locales/da/webextension.properties
# Clip in this context refers to a portion of an image
# Also used for the title on the toolbar button
# The string "{meta_key}-V" should be translated to the region-specific
# shorthand for the Paste keyboard shortcut. {meta_key} is a placeholder for the
# modifier key used in the shortcut (do not translate it): for example, Ctrl-V
# on Windows systems.
# Section for error strings
# "Out of order" is an humorous way to indicate that the service is not working
# properly.
# Section for onboarding strings
# Do not translate {termsAndPrivacyNoticeTermsLink} and
# {termsAndPrivacyNoticyPrivacyLink}. They're placeholders replaced by links
# using the corresponding translated strings as text.
# This string is used as the text for a link in termsAndPrivacyNotice
# This string is used as the text for a link in termsAndPrivacyNotice
<|file_sep|>updated/locales/da/webextension.properties
# Clip in this context refers to a portion of an image
# Also used for the title on the toolbar button
# The string "{meta_key}-V" should be translated to the region-specific
# shorthand for the Paste keyboard shortcut. {meta_key} is a placeholder for the
# modifier key used in the shortcut (do not translate it): for example, Ctrl-V
# on Windows systems.
# Section for error strings
# "Out of order" is an humorous way to indicate that the service is not working
# properly.
# Section for onboarding strings
# Do not translate {termsAndPrivacyNoticeTermsLink} and
# {termsAndPrivacyNoticyPrivacyLink}. They're placeholders replaced by links
# using the corresponding translated strings as text.
# This string is used as the text for a link in termsAndPrivacyNoticeCloudServices
# This string is used as the text for a link in termsAndPrivacyNoticeCloudServices | 6865dea8eafcd3cfc5edf1f1f9a409ef31ad1687 | locales/da/webextension.properties | locales/da/webextension.properties | INI |
<|file_sep|>CardboardControl/Scripts/ParsedTouchData.cs.diff
original:
public ParsedTouchData() {}
updated:
public ParsedTouchData() {
Cardboard cardboard = CardboardGameObject().GetComponent<Cardboard>();
cardboard.TapIsTrigger = false;
}
private GameObject CardboardGameObject() {
GameObject gameObject = Camera.main.gameObject;
return gameObject.transform.parent.parent.gameObject;
}
<|file_sep|>CardboardControl/Scripts/ParsedTouchData.cs.diff
original:
wasTouched |= this.IsDown();
updated:
wasTouched |= IsDown();
<|file_sep|>original/CardboardControl/Scripts/ParsedTouchData.cs
public class ParsedTouchData {
private bool wasTouched = false;
public ParsedTouchData() {}
public void Update() {
wasTouched |= this.IsDown();
}
public bool IsDown() {
return Input.touchCount > 0;
}
public bool IsUp() {
if (!this.IsDown() && wasTouched) {
wasTouched = false;
return true;
}
return false;
}
}
<|file_sep|>current/CardboardControl/Scripts/ParsedTouchData.cs
private GameObject CardboardGameObject() {
GameObject gameObject = Camera.main.gameObject;
return gameObject.transform.parent.parent.gameObject;
}
public void Update() {
wasTouched |= IsDown();
}
public bool IsDown() {
return Input.touchCount > 0;
}
public bool IsUp() {
if (!this.IsDown() && wasTouched) {
wasTouched = false;
return true;
}
return false;
}
}
<|file_sep|>updated/CardboardControl/Scripts/ParsedTouchData.cs | private GameObject CardboardGameObject() {
GameObject gameObject = Camera.main.gameObject;
return gameObject.transform.parent.parent.gameObject;
}
public void Update() {
wasTouched |= IsDown();
}
public bool IsDown() {
return Input.touchCount > 0;
}
public bool IsUp() {
if (!IsDown() && wasTouched) {
wasTouched = false;
return true;
}
return false;
}
} | <|file_sep|>CardboardControl/Scripts/ParsedTouchData.cs.diff
original:
public ParsedTouchData() {}
updated:
public ParsedTouchData() {
Cardboard cardboard = CardboardGameObject().GetComponent<Cardboard>();
cardboard.TapIsTrigger = false;
}
private GameObject CardboardGameObject() {
GameObject gameObject = Camera.main.gameObject;
return gameObject.transform.parent.parent.gameObject;
}
<|file_sep|>CardboardControl/Scripts/ParsedTouchData.cs.diff
original:
wasTouched |= this.IsDown();
updated:
wasTouched |= IsDown();
<|file_sep|>original/CardboardControl/Scripts/ParsedTouchData.cs
public class ParsedTouchData {
private bool wasTouched = false;
public ParsedTouchData() {}
public void Update() {
wasTouched |= this.IsDown();
}
public bool IsDown() {
return Input.touchCount > 0;
}
public bool IsUp() {
if (!this.IsDown() && wasTouched) {
wasTouched = false;
return true;
}
return false;
}
}
<|file_sep|>current/CardboardControl/Scripts/ParsedTouchData.cs
private GameObject CardboardGameObject() {
GameObject gameObject = Camera.main.gameObject;
return gameObject.transform.parent.parent.gameObject;
}
public void Update() {
wasTouched |= IsDown();
}
public bool IsDown() {
return Input.touchCount > 0;
}
public bool IsUp() {
if (!this.IsDown() && wasTouched) {
wasTouched = false;
return true;
}
return false;
}
}
<|file_sep|>updated/CardboardControl/Scripts/ParsedTouchData.cs
private GameObject CardboardGameObject() {
GameObject gameObject = Camera.main.gameObject;
return gameObject.transform.parent.parent.gameObject;
}
public void Update() {
wasTouched |= IsDown();
}
public bool IsDown() {
return Input.touchCount > 0;
}
public bool IsUp() {
if (!IsDown() && wasTouched) {
wasTouched = false;
return true;
}
return false;
}
} | 7a6b4aefb7d0750d590fd7fe16b4afdb78415a64 | CardboardControl/Scripts/ParsedTouchData.cs | CardboardControl/Scripts/ParsedTouchData.cs | C# |
<|file_sep|>original/package.json
"nan": "^2.1.0"
},
"devDependencies": {
"ninja": "^3.0.7",
"node-gyp": "^3.2.1",
"pre-commit": "^1.1.2"
},
"scripts": {
"test": "make test",
"lint": "./scripts/cpplint.py",
"preinstall": "npm install nan && npm install cppunitlite",
"install": "node-gyp rebuild"
},
"repository": {
"type": "git",
"url": "git+https://github.com/bbondy/BloomFilter.cpp.git"
},
"author": "Brian R. Bondy <netzen@gmail.com> (http://www.brianbondy.com)",
"license": "MPL-2.0",
"gypfile": true,
"bugs": {
<|file_sep|>current/package.json
"nan": "^2.1.0"
},
"devDependencies": {
"ninja": "^3.0.7",
"node-gyp": "^3.2.1",
"pre-commit": "^1.1.2"
},
"scripts": {
"test": "make test",
"lint": "./scripts/cpplint.py",
"preinstall": "npm install nan && npm install cppunitlite",
"install": "node-gyp rebuild"
},
"repository": {
"type": "git",
"url": "git+https://github.com/bbondy/BloomFilter.cpp.git"
},
"author": "Brian R. Bondy <netzen@gmail.com> (http://www.brianbondy.com)",
"license": "MPL-2.0",
"gypfile": true,
"bugs": {
<|file_sep|>updated/package.json | "nan": "^2.1.0"
},
"devDependencies": {
"ninja": "^3.0.7",
"node-gyp": "^3.2.1",
"pre-commit": "^1.1.2"
},
"scripts": {
"test": "make test",
"lint": "./scripts/cpplint.py",
"install": "node-gyp rebuild"
},
"repository": {
"type": "git",
"url": "git+https://github.com/bbondy/BloomFilter.cpp.git"
},
"author": "Brian R. Bondy <netzen@gmail.com> (http://www.brianbondy.com)",
"license": "MPL-2.0",
"gypfile": true,
"bugs": {
"url": "https://github.com/bbondy/BloomFilter.cpp/issues" | <|file_sep|>original/package.json
"nan": "^2.1.0"
},
"devDependencies": {
"ninja": "^3.0.7",
"node-gyp": "^3.2.1",
"pre-commit": "^1.1.2"
},
"scripts": {
"test": "make test",
"lint": "./scripts/cpplint.py",
"preinstall": "npm install nan && npm install cppunitlite",
"install": "node-gyp rebuild"
},
"repository": {
"type": "git",
"url": "git+https://github.com/bbondy/BloomFilter.cpp.git"
},
"author": "Brian R. Bondy <netzen@gmail.com> (http://www.brianbondy.com)",
"license": "MPL-2.0",
"gypfile": true,
"bugs": {
<|file_sep|>current/package.json
"nan": "^2.1.0"
},
"devDependencies": {
"ninja": "^3.0.7",
"node-gyp": "^3.2.1",
"pre-commit": "^1.1.2"
},
"scripts": {
"test": "make test",
"lint": "./scripts/cpplint.py",
"preinstall": "npm install nan && npm install cppunitlite",
"install": "node-gyp rebuild"
},
"repository": {
"type": "git",
"url": "git+https://github.com/bbondy/BloomFilter.cpp.git"
},
"author": "Brian R. Bondy <netzen@gmail.com> (http://www.brianbondy.com)",
"license": "MPL-2.0",
"gypfile": true,
"bugs": {
<|file_sep|>updated/package.json
"nan": "^2.1.0"
},
"devDependencies": {
"ninja": "^3.0.7",
"node-gyp": "^3.2.1",
"pre-commit": "^1.1.2"
},
"scripts": {
"test": "make test",
"lint": "./scripts/cpplint.py",
"install": "node-gyp rebuild"
},
"repository": {
"type": "git",
"url": "git+https://github.com/bbondy/BloomFilter.cpp.git"
},
"author": "Brian R. Bondy <netzen@gmail.com> (http://www.brianbondy.com)",
"license": "MPL-2.0",
"gypfile": true,
"bugs": {
"url": "https://github.com/bbondy/BloomFilter.cpp/issues" | 7066ef8ffa0395569c7d7b8d610e4469ec62595c | package.json | package.json | JSON |
<|file_sep|>original/README.rst
* Bugs: http://bugs.launchpad.net/delorean
Features
--------
Builds and maintains yum repositories following openstacks uptream repositories
Setup
-----
::
yum install docker-io git createrepo python-virtualenv
systemctl start httpd
systemctl start docker
# Add the user you intend to run as to the docker group and login again
git clone https://github.com/derekhiggins/delorean.git
cd delorean
./scripts/create_build_image.sh
virtualenv ../delorean-venv
. ../delorean-venv/bin/activate
pip install -r requirements.txt
python setup.py develop
<|file_sep|>current/README.rst
* Bugs: http://bugs.launchpad.net/delorean
Features
--------
Builds and maintains yum repositories following openstacks uptream repositories
Setup
-----
::
yum install docker-io git createrepo python-virtualenv
systemctl start httpd
systemctl start docker
# Add the user you intend to run as to the docker group and login again
git clone https://github.com/derekhiggins/delorean.git
cd delorean
./scripts/create_build_image.sh
virtualenv ../delorean-venv
. ../delorean-venv/bin/activate
pip install -r requirements.txt
python setup.py develop
<|file_sep|>updated/README.rst | * Bugs: http://bugs.launchpad.net/delorean
Features
--------
Builds and maintains yum repositories following openstacks uptream repositories
Setup
-----
::
yum install docker-io git createrepo python-virtualenv git-hg
systemctl start httpd
systemctl start docker
# Add the user you intend to run as to the docker group and login again
git clone https://github.com/derekhiggins/delorean.git
cd delorean
./scripts/create_build_image.sh
virtualenv ../delorean-venv
. ../delorean-venv/bin/activate
pip install -r requirements.txt
python setup.py develop | <|file_sep|>original/README.rst
* Bugs: http://bugs.launchpad.net/delorean
Features
--------
Builds and maintains yum repositories following openstacks uptream repositories
Setup
-----
::
yum install docker-io git createrepo python-virtualenv
systemctl start httpd
systemctl start docker
# Add the user you intend to run as to the docker group and login again
git clone https://github.com/derekhiggins/delorean.git
cd delorean
./scripts/create_build_image.sh
virtualenv ../delorean-venv
. ../delorean-venv/bin/activate
pip install -r requirements.txt
python setup.py develop
<|file_sep|>current/README.rst
* Bugs: http://bugs.launchpad.net/delorean
Features
--------
Builds and maintains yum repositories following openstacks uptream repositories
Setup
-----
::
yum install docker-io git createrepo python-virtualenv
systemctl start httpd
systemctl start docker
# Add the user you intend to run as to the docker group and login again
git clone https://github.com/derekhiggins/delorean.git
cd delorean
./scripts/create_build_image.sh
virtualenv ../delorean-venv
. ../delorean-venv/bin/activate
pip install -r requirements.txt
python setup.py develop
<|file_sep|>updated/README.rst
* Bugs: http://bugs.launchpad.net/delorean
Features
--------
Builds and maintains yum repositories following openstacks uptream repositories
Setup
-----
::
yum install docker-io git createrepo python-virtualenv git-hg
systemctl start httpd
systemctl start docker
# Add the user you intend to run as to the docker group and login again
git clone https://github.com/derekhiggins/delorean.git
cd delorean
./scripts/create_build_image.sh
virtualenv ../delorean-venv
. ../delorean-venv/bin/activate
pip install -r requirements.txt
python setup.py develop | 551c0d79f03366c1414a5c481dcb3f229e0e5edd | README.rst | README.rst | reStructuredText |
<|file_sep|>original/bookmarks/core/models.py
from django.db import models
from django.utils import timezone
from taggit.managers import TaggableManager
class Bookmark(models.Model):
title = models.CharField(max_length=200, blank=True, null=True)
description = models.TextField(blank=True, null=True)
date_added = models.DateTimeField(default=timezone.now, blank=True)
tags = TaggableManager(blank=True)
private = models.BooleanField(default=False)
url = models.URLField()
def __unicode__(self):
return "{}: {} [{}]".format(
self.pk,
self.title[:40],
self.date_added
)
<|file_sep|>current/bookmarks/core/models.py
from django.db import models
from django.utils import timezone
from taggit.managers import TaggableManager
class Bookmark(models.Model):
title = models.CharField(max_length=200, blank=True, null=True)
description = models.TextField(blank=True, null=True)
date_added = models.DateTimeField(default=timezone.now, blank=True)
tags = TaggableManager(blank=True)
private = models.BooleanField(default=False)
url = models.URLField()
def __unicode__(self):
return "{}: {} [{}]".format(
self.pk,
self.title[:40],
self.date_added
)
<|file_sep|>updated/bookmarks/core/models.py |
from django.db import models
from django.utils import timezone
from taggit.managers import TaggableManager
class Bookmark(models.Model):
title = models.CharField(max_length=200, blank=True, null=True)
description = models.TextField(blank=True, null=True)
date_added = models.DateTimeField(default=timezone.now, blank=True)
tags = TaggableManager(blank=True)
private = models.BooleanField(default=False)
url = models.URLField(max_length=500)
def __unicode__(self):
return "{}: {} [{}]".format(
self.pk,
self.title[:40],
self.date_added
) | <|file_sep|>original/bookmarks/core/models.py
from django.db import models
from django.utils import timezone
from taggit.managers import TaggableManager
class Bookmark(models.Model):
title = models.CharField(max_length=200, blank=True, null=True)
description = models.TextField(blank=True, null=True)
date_added = models.DateTimeField(default=timezone.now, blank=True)
tags = TaggableManager(blank=True)
private = models.BooleanField(default=False)
url = models.URLField()
def __unicode__(self):
return "{}: {} [{}]".format(
self.pk,
self.title[:40],
self.date_added
)
<|file_sep|>current/bookmarks/core/models.py
from django.db import models
from django.utils import timezone
from taggit.managers import TaggableManager
class Bookmark(models.Model):
title = models.CharField(max_length=200, blank=True, null=True)
description = models.TextField(blank=True, null=True)
date_added = models.DateTimeField(default=timezone.now, blank=True)
tags = TaggableManager(blank=True)
private = models.BooleanField(default=False)
url = models.URLField()
def __unicode__(self):
return "{}: {} [{}]".format(
self.pk,
self.title[:40],
self.date_added
)
<|file_sep|>updated/bookmarks/core/models.py
from django.db import models
from django.utils import timezone
from taggit.managers import TaggableManager
class Bookmark(models.Model):
title = models.CharField(max_length=200, blank=True, null=True)
description = models.TextField(blank=True, null=True)
date_added = models.DateTimeField(default=timezone.now, blank=True)
tags = TaggableManager(blank=True)
private = models.BooleanField(default=False)
url = models.URLField(max_length=500)
def __unicode__(self):
return "{}: {} [{}]".format(
self.pk,
self.title[:40],
self.date_added
) | db8dea37028432c89e098728970fbaa265e49359 | bookmarks/core/models.py | bookmarks/core/models.py | Python |
<|file_sep|>original/src/supp_classes.js
toString(){
return ("date: " + this.day + "-- " + this.hour + ":" + this.min);
}
/*
if a < b : return -1
if a == b : return 0
if a > b : return 1
*/
static compare(a, b){
if (a.day<b.day)
return -1;
if (a.day>b.day)
return 1;
if (a.hour<b.hour)
return -1;
if (a.hour>b.hour)
return 1;
if (a.min<b.min)
return -1;
if (a.min>b.min)
<|file_sep|>current/src/supp_classes.js
toString(){
return ("date: " + this.day + "-- " + this.hour + ":" + this.min);
}
/*
if a < b : return -1
if a == b : return 0
if a > b : return 1
*/
static compare(a, b){
if (a.day<b.day)
return -1;
if (a.day>b.day)
return 1;
if (a.hour<b.hour)
return -1;
if (a.hour>b.hour)
return 1;
if (a.min<b.min)
return -1;
if (a.min>b.min)
<|file_sep|>updated/src/supp_classes.js |
toString(){
return ("date: " + this.day + "-- " + this.hour + ":" + this.min);
}
/*
if a < b : return -1
if a == b : return 0
if a > b : return 1
*/
static compare(a, b){
if (a.day == 6 && b.day == 0)
return -1;
if (a.day == 0 && b.day == 6)
return 1;
if (a.day<b.day)
return -1;
if (a.day>b.day)
return 1;
if (a.hour<b.hour)
return -1;
if (a.hour>b.hour) | <|file_sep|>original/src/supp_classes.js
toString(){
return ("date: " + this.day + "-- " + this.hour + ":" + this.min);
}
/*
if a < b : return -1
if a == b : return 0
if a > b : return 1
*/
static compare(a, b){
if (a.day<b.day)
return -1;
if (a.day>b.day)
return 1;
if (a.hour<b.hour)
return -1;
if (a.hour>b.hour)
return 1;
if (a.min<b.min)
return -1;
if (a.min>b.min)
<|file_sep|>current/src/supp_classes.js
toString(){
return ("date: " + this.day + "-- " + this.hour + ":" + this.min);
}
/*
if a < b : return -1
if a == b : return 0
if a > b : return 1
*/
static compare(a, b){
if (a.day<b.day)
return -1;
if (a.day>b.day)
return 1;
if (a.hour<b.hour)
return -1;
if (a.hour>b.hour)
return 1;
if (a.min<b.min)
return -1;
if (a.min>b.min)
<|file_sep|>updated/src/supp_classes.js
toString(){
return ("date: " + this.day + "-- " + this.hour + ":" + this.min);
}
/*
if a < b : return -1
if a == b : return 0
if a > b : return 1
*/
static compare(a, b){
if (a.day == 6 && b.day == 0)
return -1;
if (a.day == 0 && b.day == 6)
return 1;
if (a.day<b.day)
return -1;
if (a.day>b.day)
return 1;
if (a.hour<b.hour)
return -1;
if (a.hour>b.hour) | 0ef21e977f51aeffa279442326785704a16347fe | src/supp_classes.js | src/supp_classes.js | JavaScript |
<|file_sep|>original/.github/workflows/ci.yml
dokken:
needs: [delivery]
runs-on: ubuntu-latest
strategy:
matrix:
os:
- 'amazonlinux-2'
- 'centos-6'
- 'centos-7'
- 'debian-9'
- 'debian-10'
- 'ubuntu-1604'
- 'ubuntu-1804'
suite:
- 'source'
- 'package'
fail-fast: false
steps:
- name: Check out code
<|file_sep|>current/.github/workflows/ci.yml
dokken:
needs: [delivery]
runs-on: ubuntu-latest
strategy:
matrix:
os:
- 'amazonlinux-2'
- 'centos-6'
- 'centos-7'
- 'debian-9'
- 'debian-10'
- 'ubuntu-1604'
- 'ubuntu-1804'
suite:
- 'source'
- 'package'
fail-fast: false
steps:
- name: Check out code
<|file_sep|>updated/.github/workflows/ci.yml |
dokken:
needs: [delivery]
runs-on: ubuntu-latest
strategy:
matrix:
os:
- 'amazonlinux-2'
- 'centos-6'
- 'centos-7'
- 'centos-8'
- 'debian-9'
- 'debian-10'
- 'ubuntu-1604'
- 'ubuntu-1804'
suite:
- 'source'
- 'package'
fail-fast: false
steps: | <|file_sep|>original/.github/workflows/ci.yml
dokken:
needs: [delivery]
runs-on: ubuntu-latest
strategy:
matrix:
os:
- 'amazonlinux-2'
- 'centos-6'
- 'centos-7'
- 'debian-9'
- 'debian-10'
- 'ubuntu-1604'
- 'ubuntu-1804'
suite:
- 'source'
- 'package'
fail-fast: false
steps:
- name: Check out code
<|file_sep|>current/.github/workflows/ci.yml
dokken:
needs: [delivery]
runs-on: ubuntu-latest
strategy:
matrix:
os:
- 'amazonlinux-2'
- 'centos-6'
- 'centos-7'
- 'debian-9'
- 'debian-10'
- 'ubuntu-1604'
- 'ubuntu-1804'
suite:
- 'source'
- 'package'
fail-fast: false
steps:
- name: Check out code
<|file_sep|>updated/.github/workflows/ci.yml
dokken:
needs: [delivery]
runs-on: ubuntu-latest
strategy:
matrix:
os:
- 'amazonlinux-2'
- 'centos-6'
- 'centos-7'
- 'centos-8'
- 'debian-9'
- 'debian-10'
- 'ubuntu-1604'
- 'ubuntu-1804'
suite:
- 'source'
- 'package'
fail-fast: false
steps: | fc30a239aea2ff4532fe115aaafae03dc1186ad6 | .github/workflows/ci.yml | .github/workflows/ci.yml | YAML |
<|file_sep|>test/setupTests.js.diff
original:
updated:
import { TextEncoder, TextDecoder } from 'util';
<|file_sep|>original/test/setupTests.js
import * as languageHandler from "../src/languageHandler";
languageHandler.setLanguage('en');
languageHandler.setMissingEntryGenerator(key => key.split("|", 2)[1]);
require('jest-fetch-mock').enableMocks();
<|file_sep|>current/test/setupTests.js
import * as languageHandler from "../src/languageHandler";
import { TextEncoder, TextDecoder } from 'util';
languageHandler.setLanguage('en');
languageHandler.setMissingEntryGenerator(key => key.split("|", 2)[1]);
require('jest-fetch-mock').enableMocks();
<|file_sep|>updated/test/setupTests.js | import * as languageHandler from "../src/languageHandler";
import { TextEncoder, TextDecoder } from 'util';
languageHandler.setLanguage('en');
languageHandler.setMissingEntryGenerator(key => key.split("|", 2)[1]);
require('jest-fetch-mock').enableMocks();
// polyfilling TextEncoder as it is not available on JSDOM
// view https://github.com/facebook/jest/issues/9983
global.TextEncoder = TextEncoder;
global.TextDecoder = TextDecoder; | <|file_sep|>test/setupTests.js.diff
original:
updated:
import { TextEncoder, TextDecoder } from 'util';
<|file_sep|>original/test/setupTests.js
import * as languageHandler from "../src/languageHandler";
languageHandler.setLanguage('en');
languageHandler.setMissingEntryGenerator(key => key.split("|", 2)[1]);
require('jest-fetch-mock').enableMocks();
<|file_sep|>current/test/setupTests.js
import * as languageHandler from "../src/languageHandler";
import { TextEncoder, TextDecoder } from 'util';
languageHandler.setLanguage('en');
languageHandler.setMissingEntryGenerator(key => key.split("|", 2)[1]);
require('jest-fetch-mock').enableMocks();
<|file_sep|>updated/test/setupTests.js
import * as languageHandler from "../src/languageHandler";
import { TextEncoder, TextDecoder } from 'util';
languageHandler.setLanguage('en');
languageHandler.setMissingEntryGenerator(key => key.split("|", 2)[1]);
require('jest-fetch-mock').enableMocks();
// polyfilling TextEncoder as it is not available on JSDOM
// view https://github.com/facebook/jest/issues/9983
global.TextEncoder = TextEncoder;
global.TextDecoder = TextDecoder; | 37484e3fc465ef5a05a5001580facbea37d10bb5 | test/setupTests.js | test/setupTests.js | JavaScript |
<|file_sep|>original/ansibleplay/tools/install_ansible.sh
<|file_sep|>current/ansibleplay/tools/install_ansible.sh
<|file_sep|>updated/ansibleplay/tools/install_ansible.sh | #!/usr/bin/env sh
ansible_repo="http://github.com/ansible/ansible.git"
ansible_branch="stable-1.9"
ansible_localdir="/home/vagrant/ansible_stable19"
curdir=$(pwd)
# Install Ansible (git clone)
git clone ${ansible_repo} -b ${ansible_branch} ${ansible_localdir}
# Install Ansible modules
cd ${ansible_localdir}
git submodule update --init --recursive
cd ${curdir}
source ${ansible_localdir}/hacking/env-setup
export ANSIBLE_HOST_KEY_CHECKING=False
pip install jinja2
pip install pyYaml | <|file_sep|>original/ansibleplay/tools/install_ansible.sh
<|file_sep|>current/ansibleplay/tools/install_ansible.sh
<|file_sep|>updated/ansibleplay/tools/install_ansible.sh
#!/usr/bin/env sh
ansible_repo="http://github.com/ansible/ansible.git"
ansible_branch="stable-1.9"
ansible_localdir="/home/vagrant/ansible_stable19"
curdir=$(pwd)
# Install Ansible (git clone)
git clone ${ansible_repo} -b ${ansible_branch} ${ansible_localdir}
# Install Ansible modules
cd ${ansible_localdir}
git submodule update --init --recursive
cd ${curdir}
source ${ansible_localdir}/hacking/env-setup
export ANSIBLE_HOST_KEY_CHECKING=False
pip install jinja2
pip install pyYaml | 2d5a98af90118b4be0f4939b9752c62701916329 | ansibleplay/tools/install_ansible.sh | ansibleplay/tools/install_ansible.sh | Shell |
<|file_sep|>original/packages/pr/prof-flamegraph.yaml
<|file_sep|>current/packages/pr/prof-flamegraph.yaml
<|file_sep|>updated/packages/pr/prof-flamegraph.yaml | homepage: ''
changelog-type: ''
hash: 13f0dd468f03f3d46a3f4c13fdb271aa4a80d57038f302ec8d7dbe317e50b5b4
test-bench-deps: {}
maintainer: Sam Halliday
synopsis: Generate flamegraphs from ghc RTS .prof files
changelog: ''
basic-deps:
base: ^>=4.11.1.0 || ^>=4.12.0.0
optparse-applicative: ^>=0.14.3.0
all-versions:
- '1.0.0'
author: Sam Halliday
latest: '1.0.0'
description-type: haddock
description: ! 'This is a small tool to convert ghc `.prof` files into a format consumable
by [FlameGraph](https://github.com/brendangregg/FlameGraph).
Obtain `.prof` files by compiling with `--enable-profiling` and running with | <|file_sep|>original/packages/pr/prof-flamegraph.yaml
<|file_sep|>current/packages/pr/prof-flamegraph.yaml
<|file_sep|>updated/packages/pr/prof-flamegraph.yaml
homepage: ''
changelog-type: ''
hash: 13f0dd468f03f3d46a3f4c13fdb271aa4a80d57038f302ec8d7dbe317e50b5b4
test-bench-deps: {}
maintainer: Sam Halliday
synopsis: Generate flamegraphs from ghc RTS .prof files
changelog: ''
basic-deps:
base: ^>=4.11.1.0 || ^>=4.12.0.0
optparse-applicative: ^>=0.14.3.0
all-versions:
- '1.0.0'
author: Sam Halliday
latest: '1.0.0'
description-type: haddock
description: ! 'This is a small tool to convert ghc `.prof` files into a format consumable
by [FlameGraph](https://github.com/brendangregg/FlameGraph).
Obtain `.prof` files by compiling with `--enable-profiling` and running with | 9fba07100fe056febfb144304be1320d3e807951 | packages/pr/prof-flamegraph.yaml | packages/pr/prof-flamegraph.yaml | YAML |
<|file_sep|>original/README.md
# typescript-database-abstraction
This is an example repository created for the following blog article: <LINK>.
## Getting started
- `yarn start` Runs the TypeScript compiler in watch mode.
## Build
- `yarn run build-app`
## Run the demo code
- `node app/index.js`
## Lint
- `yarn run lint`
## Test
- `yarn test`
## About
### Author
Markus Oberlehner
<|file_sep|>current/README.md
# typescript-database-abstraction
This is an example repository created for the following blog article: <LINK>.
## Getting started
- `yarn start` Runs the TypeScript compiler in watch mode.
## Build
- `yarn run build-app`
## Run the demo code
- `node app/index.js`
## Lint
- `yarn run lint`
## Test
- `yarn test`
## About
### Author
Markus Oberlehner
<|file_sep|>updated/README.md | # typescript-database-abstraction
This is an example repository created for the following blog article: [Building a simple (but overengineered) database abstraction with TypeScript](https://markus.oberlehner.net/blog/2017/03/building-a-simple-database-abstraction-with-typescript/).
## Getting started
- `yarn start` Runs the TypeScript compiler in watch mode.
## Build
- `yarn run build-app`
## Run the demo code
- `node app/index.js`
## Lint
- `yarn run lint`
## Test
- `yarn test`
## About
### Author
Markus Oberlehner | <|file_sep|>original/README.md
# typescript-database-abstraction
This is an example repository created for the following blog article: <LINK>.
## Getting started
- `yarn start` Runs the TypeScript compiler in watch mode.
## Build
- `yarn run build-app`
## Run the demo code
- `node app/index.js`
## Lint
- `yarn run lint`
## Test
- `yarn test`
## About
### Author
Markus Oberlehner
<|file_sep|>current/README.md
# typescript-database-abstraction
This is an example repository created for the following blog article: <LINK>.
## Getting started
- `yarn start` Runs the TypeScript compiler in watch mode.
## Build
- `yarn run build-app`
## Run the demo code
- `node app/index.js`
## Lint
- `yarn run lint`
## Test
- `yarn test`
## About
### Author
Markus Oberlehner
<|file_sep|>updated/README.md
# typescript-database-abstraction
This is an example repository created for the following blog article: [Building a simple (but overengineered) database abstraction with TypeScript](https://markus.oberlehner.net/blog/2017/03/building-a-simple-database-abstraction-with-typescript/).
## Getting started
- `yarn start` Runs the TypeScript compiler in watch mode.
## Build
- `yarn run build-app`
## Run the demo code
- `node app/index.js`
## Lint
- `yarn run lint`
## Test
- `yarn test`
## About
### Author
Markus Oberlehner | c3fdb07f0c5841bee2e18adfc991247afd00562c | README.md | README.md | Markdown |
<|file_sep|>Formula/alt.rb.diff
original:
url "https://github.com/uptech/alt/releases/download/v2.3.0/alt-2.3.0-x86_64-apple-darwin"
updated:
url "https://github.com/uptech/alt/archive/v2.3.0.tar.gz"
<|file_sep|>Formula/alt.rb.diff
original:
sha256 "255049848f8da5cfb387f80b63fb264f0028a8ce1b8035ae659cc3e3499581eb"
updated:
sha256 "f6f4393a85f0fa0ecc9c65c2a7b0bafc2b5863c632c00c33239dd98ec05433d7"
head "https://github.com/uptech/alt.git"
depends_on "rust" => :build
<|file_sep|>original/Formula/alt.rb
class Alt < Formula
desc "command-line utility to find alternate file"
homepage "https://github.com/uptech/alt"
url "https://github.com/uptech/alt/releases/download/v2.3.0/alt-2.3.0-x86_64-apple-darwin"
version "2.3.0"
sha256 "255049848f8da5cfb387f80b63fb264f0028a8ce1b8035ae659cc3e3499581eb"
def install
mv "alt-2.3.0-x86_64-apple-darwin", "alt"
bin.install "alt"
end
test do
assert_match version.to_s, shell_output("#{bin}/alt --version")
end
end
<|file_sep|>current/Formula/alt.rb
class Alt < Formula
desc "command-line utility to find alternate file"
homepage "https://github.com/uptech/alt"
url "https://github.com/uptech/alt/archive/v2.3.0.tar.gz"
version "2.3.0"
sha256 "f6f4393a85f0fa0ecc9c65c2a7b0bafc2b5863c632c00c33239dd98ec05433d7"
head "https://github.com/uptech/alt.git"
depends_on "rust" => :build
def install
mv "alt-2.3.0-x86_64-apple-darwin", "alt"
bin.install "alt"
end
test do
assert_match version.to_s, shell_output("#{bin}/alt --version")
end
end
<|file_sep|>updated/Formula/alt.rb | class Alt < Formula
desc "command-line utility to find alternate file"
homepage "https://github.com/uptech/alt"
url "https://github.com/uptech/alt/archive/v2.3.0.tar.gz"
version "2.3.0"
sha256 "f6f4393a85f0fa0ecc9c65c2a7b0bafc2b5863c632c00c33239dd98ec05433d7"
head "https://github.com/uptech/alt.git"
depends_on "rust" => :build
def install
system "cargo", "build", "--release"
bin.install "target/release/alt"
# man1.install "doc/alt.1" # only in versions > 2.3.0
end
test do
assert_match version.to_s, shell_output("#{bin}/alt --version")
end
end | <|file_sep|>Formula/alt.rb.diff
original:
url "https://github.com/uptech/alt/releases/download/v2.3.0/alt-2.3.0-x86_64-apple-darwin"
updated:
url "https://github.com/uptech/alt/archive/v2.3.0.tar.gz"
<|file_sep|>Formula/alt.rb.diff
original:
sha256 "255049848f8da5cfb387f80b63fb264f0028a8ce1b8035ae659cc3e3499581eb"
updated:
sha256 "f6f4393a85f0fa0ecc9c65c2a7b0bafc2b5863c632c00c33239dd98ec05433d7"
head "https://github.com/uptech/alt.git"
depends_on "rust" => :build
<|file_sep|>original/Formula/alt.rb
class Alt < Formula
desc "command-line utility to find alternate file"
homepage "https://github.com/uptech/alt"
url "https://github.com/uptech/alt/releases/download/v2.3.0/alt-2.3.0-x86_64-apple-darwin"
version "2.3.0"
sha256 "255049848f8da5cfb387f80b63fb264f0028a8ce1b8035ae659cc3e3499581eb"
def install
mv "alt-2.3.0-x86_64-apple-darwin", "alt"
bin.install "alt"
end
test do
assert_match version.to_s, shell_output("#{bin}/alt --version")
end
end
<|file_sep|>current/Formula/alt.rb
class Alt < Formula
desc "command-line utility to find alternate file"
homepage "https://github.com/uptech/alt"
url "https://github.com/uptech/alt/archive/v2.3.0.tar.gz"
version "2.3.0"
sha256 "f6f4393a85f0fa0ecc9c65c2a7b0bafc2b5863c632c00c33239dd98ec05433d7"
head "https://github.com/uptech/alt.git"
depends_on "rust" => :build
def install
mv "alt-2.3.0-x86_64-apple-darwin", "alt"
bin.install "alt"
end
test do
assert_match version.to_s, shell_output("#{bin}/alt --version")
end
end
<|file_sep|>updated/Formula/alt.rb
class Alt < Formula
desc "command-line utility to find alternate file"
homepage "https://github.com/uptech/alt"
url "https://github.com/uptech/alt/archive/v2.3.0.tar.gz"
version "2.3.0"
sha256 "f6f4393a85f0fa0ecc9c65c2a7b0bafc2b5863c632c00c33239dd98ec05433d7"
head "https://github.com/uptech/alt.git"
depends_on "rust" => :build
def install
system "cargo", "build", "--release"
bin.install "target/release/alt"
# man1.install "doc/alt.1" # only in versions > 2.3.0
end
test do
assert_match version.to_s, shell_output("#{bin}/alt --version")
end
end | 4311e35928675ad65c33f6d3379f9804f4ab630e | Formula/alt.rb | Formula/alt.rb | Ruby |
<|file_sep|>original/upgrades/15.7-to-15.11-upgrade.toml
Exclude = "target|.settings|.git|.svn|.idea"
[[Transformations]]
Filter = "pom.xml"
[[Transformations.Proc]]
Name = "ReplaceMavenDependency"
Params = [
"org.seedstack:seedstack-bom:15.7",
"org.seedstack:seed-bom:15.11-M1-SNAPSHOT"
]
<|file_sep|>current/upgrades/15.7-to-15.11-upgrade.toml
Exclude = "target|.settings|.git|.svn|.idea"
[[Transformations]]
Filter = "pom.xml"
[[Transformations.Proc]]
Name = "ReplaceMavenDependency"
Params = [
"org.seedstack:seedstack-bom:15.7",
"org.seedstack:seed-bom:15.11-M1-SNAPSHOT"
]
<|file_sep|>updated/upgrades/15.7-to-15.11-upgrade.toml |
[[Transformations]]
Filter = "pom.xml"
[[Transformations.Proc]]
Name = "ReplaceMavenDependency"
Params = [
"org.seedstack:seedstack-bom:15.7",
"org.seedstack:seed-bom:15.11-M1-SNAPSHOT"
]
[[Transformations]]
Filter = "*.java"
[[Transformations.Proc]]
Name = "Replace"
Params = [
"org.seedstack.seed.core.api",
"org.seedstack.seed",
"org.seedstack.seed.core.api", | <|file_sep|>original/upgrades/15.7-to-15.11-upgrade.toml
Exclude = "target|.settings|.git|.svn|.idea"
[[Transformations]]
Filter = "pom.xml"
[[Transformations.Proc]]
Name = "ReplaceMavenDependency"
Params = [
"org.seedstack:seedstack-bom:15.7",
"org.seedstack:seed-bom:15.11-M1-SNAPSHOT"
]
<|file_sep|>current/upgrades/15.7-to-15.11-upgrade.toml
Exclude = "target|.settings|.git|.svn|.idea"
[[Transformations]]
Filter = "pom.xml"
[[Transformations.Proc]]
Name = "ReplaceMavenDependency"
Params = [
"org.seedstack:seedstack-bom:15.7",
"org.seedstack:seed-bom:15.11-M1-SNAPSHOT"
]
<|file_sep|>updated/upgrades/15.7-to-15.11-upgrade.toml
[[Transformations]]
Filter = "pom.xml"
[[Transformations.Proc]]
Name = "ReplaceMavenDependency"
Params = [
"org.seedstack:seedstack-bom:15.7",
"org.seedstack:seed-bom:15.11-M1-SNAPSHOT"
]
[[Transformations]]
Filter = "*.java"
[[Transformations.Proc]]
Name = "Replace"
Params = [
"org.seedstack.seed.core.api",
"org.seedstack.seed",
"org.seedstack.seed.core.api", | 0c4b7349954b0d245052bbfaa695a7c82b1097ce | upgrades/15.7-to-15.11-upgrade.toml | upgrades/15.7-to-15.11-upgrade.toml | TOML |
<|file_sep|>droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt.diff
original:
updated:
/**
*
* Set a fixed height for the view dropdown popup
*
* @param view The spinner view that need a custom dropdown popup height
* @param height The height value in dp for the dropdown popup
*
*/
<|file_sep|>droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt.diff
original:
fun bindSpinnerPopupHeight(spinner: Spinner, height: Float) {
updated:
fun bindSpinnerPopupHeight(view: Spinner, height: Float) {
<|file_sep|>droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt.diff
original:
val popupWindow = popup.get(spinner)
updated:
val popupWindow = popup.get(view)
<|file_sep|>original/droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt
*/
@BindingAdapter("popupHeight")
fun bindSpinnerPopupHeight(spinner: Spinner, height: Float) {
try {
val popup = Spinner::class.java.getDeclaredField("mPopup")
popup.isAccessible = true
val popupWindow = popup.get(spinner)
if (popupWindow is ListPopupWindow) {
popupWindow.height = dpToPx(height, spinner.context).toInt()
}
} catch (e: NoSuchFieldException) {
e.printStackTrace()
} catch (e: IllegalAccessException) {
e.printStackTrace()
}
}
<|file_sep|>current/droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt
*
*/
@BindingAdapter("popupHeight")
fun bindSpinnerPopupHeight(view: Spinner, height: Float) {
try {
val popup = Spinner::class.java.getDeclaredField("mPopup")
popup.isAccessible = true
val popupWindow = popup.get(view)
if (popupWindow is ListPopupWindow) {
popupWindow.height = dpToPx(height, spinner.context).toInt()
}
} catch (e: NoSuchFieldException) {
e.printStackTrace()
} catch (e: IllegalAccessException) {
e.printStackTrace()
}
}
<|file_sep|>updated/droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt | *
*/
@BindingAdapter("popupHeight")
fun bindSpinnerPopupHeight(view: Spinner, height: Float) {
try {
val popup = Spinner::class.java.getDeclaredField("mPopup")
popup.isAccessible = true
val popupWindow = popup.get(view)
if (popupWindow is ListPopupWindow) {
popupWindow.height = dpToPx(height, view.context).toInt()
}
} catch (e: NoSuchFieldException) {
e.printStackTrace()
} catch (e: IllegalAccessException) {
e.printStackTrace()
}
} | <|file_sep|>droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt.diff
original:
updated:
/**
*
* Set a fixed height for the view dropdown popup
*
* @param view The spinner view that need a custom dropdown popup height
* @param height The height value in dp for the dropdown popup
*
*/
<|file_sep|>droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt.diff
original:
fun bindSpinnerPopupHeight(spinner: Spinner, height: Float) {
updated:
fun bindSpinnerPopupHeight(view: Spinner, height: Float) {
<|file_sep|>droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt.diff
original:
val popupWindow = popup.get(spinner)
updated:
val popupWindow = popup.get(view)
<|file_sep|>original/droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt
*/
@BindingAdapter("popupHeight")
fun bindSpinnerPopupHeight(spinner: Spinner, height: Float) {
try {
val popup = Spinner::class.java.getDeclaredField("mPopup")
popup.isAccessible = true
val popupWindow = popup.get(spinner)
if (popupWindow is ListPopupWindow) {
popupWindow.height = dpToPx(height, spinner.context).toInt()
}
} catch (e: NoSuchFieldException) {
e.printStackTrace()
} catch (e: IllegalAccessException) {
e.printStackTrace()
}
}
<|file_sep|>current/droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt
*
*/
@BindingAdapter("popupHeight")
fun bindSpinnerPopupHeight(view: Spinner, height: Float) {
try {
val popup = Spinner::class.java.getDeclaredField("mPopup")
popup.isAccessible = true
val popupWindow = popup.get(view)
if (popupWindow is ListPopupWindow) {
popupWindow.height = dpToPx(height, spinner.context).toInt()
}
} catch (e: NoSuchFieldException) {
e.printStackTrace()
} catch (e: IllegalAccessException) {
e.printStackTrace()
}
}
<|file_sep|>updated/droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt
*
*/
@BindingAdapter("popupHeight")
fun bindSpinnerPopupHeight(view: Spinner, height: Float) {
try {
val popup = Spinner::class.java.getDeclaredField("mPopup")
popup.isAccessible = true
val popupWindow = popup.get(view)
if (popupWindow is ListPopupWindow) {
popupWindow.height = dpToPx(height, view.context).toInt()
}
} catch (e: NoSuchFieldException) {
e.printStackTrace()
} catch (e: IllegalAccessException) {
e.printStackTrace()
}
} | 507ac3e9b77393645f9ef6eeb7baa7c1d8c70817 | droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt | droidbox/src/main/java/com/github/jackparisi/droidbox/binding/SpinnerAdapters.kt | Kotlin |
<|file_sep|>original/skimage/io/tests/test_io.py
def test_stack_basic():
x = np.arange(12).reshape(3, 4)
io.push(x)
assert_array_equal(io.pop(), x)
@raises(ValueError)
def test_stack_non_array():
io.push([[1, 2, 3]])
def test_imread_url():
image_url = 'file:{0}{0}{1}{0}camera.png'.format(os.path.sep, data_dir)
image = io.imread(image_url)
assert image.shape == (512, 512)
if __name__ == "__main__":
run_module_suite()
<|file_sep|>current/skimage/io/tests/test_io.py
def test_stack_basic():
x = np.arange(12).reshape(3, 4)
io.push(x)
assert_array_equal(io.pop(), x)
@raises(ValueError)
def test_stack_non_array():
io.push([[1, 2, 3]])
def test_imread_url():
image_url = 'file:{0}{0}{1}{0}camera.png'.format(os.path.sep, data_dir)
image = io.imread(image_url)
assert image.shape == (512, 512)
if __name__ == "__main__":
run_module_suite()
<|file_sep|>updated/skimage/io/tests/test_io.py | io.push(x)
assert_array_equal(io.pop(), x)
@raises(ValueError)
def test_stack_non_array():
io.push([[1, 2, 3]])
def test_imread_url():
# tweak data path so that file URI works on both unix and windows.
data_path = data_dir.lstrip(os.path.sep)
data_path = data_path.replace(os.path.sep, '/')
image_url = 'file:///{0}/camera.png'.format(data_path)
image = io.imread(image_url)
assert image.shape == (512, 512)
if __name__ == "__main__":
run_module_suite() | <|file_sep|>original/skimage/io/tests/test_io.py
def test_stack_basic():
x = np.arange(12).reshape(3, 4)
io.push(x)
assert_array_equal(io.pop(), x)
@raises(ValueError)
def test_stack_non_array():
io.push([[1, 2, 3]])
def test_imread_url():
image_url = 'file:{0}{0}{1}{0}camera.png'.format(os.path.sep, data_dir)
image = io.imread(image_url)
assert image.shape == (512, 512)
if __name__ == "__main__":
run_module_suite()
<|file_sep|>current/skimage/io/tests/test_io.py
def test_stack_basic():
x = np.arange(12).reshape(3, 4)
io.push(x)
assert_array_equal(io.pop(), x)
@raises(ValueError)
def test_stack_non_array():
io.push([[1, 2, 3]])
def test_imread_url():
image_url = 'file:{0}{0}{1}{0}camera.png'.format(os.path.sep, data_dir)
image = io.imread(image_url)
assert image.shape == (512, 512)
if __name__ == "__main__":
run_module_suite()
<|file_sep|>updated/skimage/io/tests/test_io.py
io.push(x)
assert_array_equal(io.pop(), x)
@raises(ValueError)
def test_stack_non_array():
io.push([[1, 2, 3]])
def test_imread_url():
# tweak data path so that file URI works on both unix and windows.
data_path = data_dir.lstrip(os.path.sep)
data_path = data_path.replace(os.path.sep, '/')
image_url = 'file:///{0}/camera.png'.format(data_path)
image = io.imread(image_url)
assert image.shape == (512, 512)
if __name__ == "__main__":
run_module_suite() | abc1d2095f18f6c7ff129f3b8bf9eae2d7a6239a | skimage/io/tests/test_io.py | skimage/io/tests/test_io.py | Python |
<|file_sep|>original/packages/apollo-cache-control/package.json
{
"name": "apollo-cache-control",
"version": "0.6.0-alpha.3",
"description": "A GraphQL extension for cache control",
"main": "./dist/index.js",
"types": "./dist/index.d.ts",
"license": "MIT",
"repository": "apollographql/apollo-cache-control-js",
"author": "Martijn Walraven <martijn@martijnwalraven.com>",
"engines": {
"node": ">=6.0"
},
"dependencies": {
"apollo-server-env": "file:../apollo-server-env",
"graphql-extensions": "file:../graphql-extensions"
},
"peerDependencies": {
"graphql": "^0.12.0 || ^0.13.0 || ^14.0.0"
}
}
<|file_sep|>current/packages/apollo-cache-control/package.json
{
"name": "apollo-cache-control",
"version": "0.6.0-alpha.3",
"description": "A GraphQL extension for cache control",
"main": "./dist/index.js",
"types": "./dist/index.d.ts",
"license": "MIT",
"repository": "apollographql/apollo-cache-control-js",
"author": "Martijn Walraven <martijn@martijnwalraven.com>",
"engines": {
"node": ">=6.0"
},
"dependencies": {
"apollo-server-env": "file:../apollo-server-env",
"graphql-extensions": "file:../graphql-extensions"
},
"peerDependencies": {
"graphql": "^0.12.0 || ^0.13.0 || ^14.0.0"
}
}
<|file_sep|>updated/packages/apollo-cache-control/package.json | {
"name": "apollo-cache-control",
"version": "0.6.0-alpha.6",
"description": "A GraphQL extension for cache control",
"main": "./dist/index.js",
"types": "./dist/index.d.ts",
"license": "MIT",
"repository": "apollographql/apollo-cache-control-js",
"author": "Martijn Walraven <martijn@martijnwalraven.com>",
"engines": {
"node": ">=6.0"
},
"dependencies": {
"apollo-server-env": "file:../apollo-server-env",
"graphql-extensions": "file:../graphql-extensions"
},
"peerDependencies": {
"graphql": "^0.12.0 || ^0.13.0 || ^14.0.0"
}
} | <|file_sep|>original/packages/apollo-cache-control/package.json
{
"name": "apollo-cache-control",
"version": "0.6.0-alpha.3",
"description": "A GraphQL extension for cache control",
"main": "./dist/index.js",
"types": "./dist/index.d.ts",
"license": "MIT",
"repository": "apollographql/apollo-cache-control-js",
"author": "Martijn Walraven <martijn@martijnwalraven.com>",
"engines": {
"node": ">=6.0"
},
"dependencies": {
"apollo-server-env": "file:../apollo-server-env",
"graphql-extensions": "file:../graphql-extensions"
},
"peerDependencies": {
"graphql": "^0.12.0 || ^0.13.0 || ^14.0.0"
}
}
<|file_sep|>current/packages/apollo-cache-control/package.json
{
"name": "apollo-cache-control",
"version": "0.6.0-alpha.3",
"description": "A GraphQL extension for cache control",
"main": "./dist/index.js",
"types": "./dist/index.d.ts",
"license": "MIT",
"repository": "apollographql/apollo-cache-control-js",
"author": "Martijn Walraven <martijn@martijnwalraven.com>",
"engines": {
"node": ">=6.0"
},
"dependencies": {
"apollo-server-env": "file:../apollo-server-env",
"graphql-extensions": "file:../graphql-extensions"
},
"peerDependencies": {
"graphql": "^0.12.0 || ^0.13.0 || ^14.0.0"
}
}
<|file_sep|>updated/packages/apollo-cache-control/package.json
{
"name": "apollo-cache-control",
"version": "0.6.0-alpha.6",
"description": "A GraphQL extension for cache control",
"main": "./dist/index.js",
"types": "./dist/index.d.ts",
"license": "MIT",
"repository": "apollographql/apollo-cache-control-js",
"author": "Martijn Walraven <martijn@martijnwalraven.com>",
"engines": {
"node": ">=6.0"
},
"dependencies": {
"apollo-server-env": "file:../apollo-server-env",
"graphql-extensions": "file:../graphql-extensions"
},
"peerDependencies": {
"graphql": "^0.12.0 || ^0.13.0 || ^14.0.0"
}
} | 630217d93bbc411690456752b01bca75e46269ea | packages/apollo-cache-control/package.json | packages/apollo-cache-control/package.json | JSON |
<|file_sep|>original/README.md
Braintree Hosted Fields + Bootstrap
===================================
Combine [Braintree's Hosted Fields](https://www.braintreepayments.com/features/hosted-fields) and [Bootstrap](https://getbootstrap.com/) for a slick form.
Take a look at the [example checkout](https://evanhahn.github.io/braintree-hosted-fields-bootstrap/) for more!
<|file_sep|>current/README.md
Braintree Hosted Fields + Bootstrap
===================================
Combine [Braintree's Hosted Fields](https://www.braintreepayments.com/features/hosted-fields) and [Bootstrap](https://getbootstrap.com/) for a slick form.
Take a look at the [example checkout](https://evanhahn.github.io/braintree-hosted-fields-bootstrap/) for more!
<|file_sep|>updated/README.md | Braintree Hosted Fields + Bootstrap
===================================
Combine [Braintree's Hosted Fields](https://www.braintreepayments.com/features/hosted-fields) and [Bootstrap](https://getbootstrap.com/) for a slick form.
Usage:
```js
$('#my-checkout-form').hostedFields({
authorization: 'YOUR BRAINTREE CLIENT TOKEN',
number: '#my-hosted-field-number',
cvv: '#my-hosted-field-cvv',
expirationDate: '#my-hosted-field-expiration-date'
})
```
Take a look at the [example checkout](https://evanhahn.github.io/braintree-hosted-fields-bootstrap/) for more! | <|file_sep|>original/README.md
Braintree Hosted Fields + Bootstrap
===================================
Combine [Braintree's Hosted Fields](https://www.braintreepayments.com/features/hosted-fields) and [Bootstrap](https://getbootstrap.com/) for a slick form.
Take a look at the [example checkout](https://evanhahn.github.io/braintree-hosted-fields-bootstrap/) for more!
<|file_sep|>current/README.md
Braintree Hosted Fields + Bootstrap
===================================
Combine [Braintree's Hosted Fields](https://www.braintreepayments.com/features/hosted-fields) and [Bootstrap](https://getbootstrap.com/) for a slick form.
Take a look at the [example checkout](https://evanhahn.github.io/braintree-hosted-fields-bootstrap/) for more!
<|file_sep|>updated/README.md
Braintree Hosted Fields + Bootstrap
===================================
Combine [Braintree's Hosted Fields](https://www.braintreepayments.com/features/hosted-fields) and [Bootstrap](https://getbootstrap.com/) for a slick form.
Usage:
```js
$('#my-checkout-form').hostedFields({
authorization: 'YOUR BRAINTREE CLIENT TOKEN',
number: '#my-hosted-field-number',
cvv: '#my-hosted-field-cvv',
expirationDate: '#my-hosted-field-expiration-date'
})
```
Take a look at the [example checkout](https://evanhahn.github.io/braintree-hosted-fields-bootstrap/) for more! | e6a2797078853dd0feedf0ad101675b8c58c60de | README.md | README.md | Markdown |
<|file_sep|>original/app/src/test/java/com/farmerbb/taskbar/service/UIHostServiceNewControllerTest.java
<|file_sep|>current/app/src/test/java/com/farmerbb/taskbar/service/UIHostServiceNewControllerTest.java
<|file_sep|>updated/app/src/test/java/com/farmerbb/taskbar/service/UIHostServiceNewControllerTest.java | package com.farmerbb.taskbar.service;
import com.farmerbb.taskbar.ui.DashboardController;
import com.farmerbb.taskbar.ui.StartMenuController;
import com.farmerbb.taskbar.ui.TaskbarController;
import org.junit.Test;
import org.junit.runner.RunWith;
import org.robolectric.Robolectric;
import org.robolectric.RobolectricTestRunner;
import org.robolectric.android.controller.ServiceController;
import static org.junit.Assert.assertTrue;
@RunWith(RobolectricTestRunner.class)
public class UIHostServiceNewControllerTest {
@Test
public void testDashboardService() {
ServiceController<DashboardService> dashboardService =
Robolectric.buildService(DashboardService.class);
assertTrue(dashboardService.get().newController() instanceof DashboardController); | <|file_sep|>original/app/src/test/java/com/farmerbb/taskbar/service/UIHostServiceNewControllerTest.java
<|file_sep|>current/app/src/test/java/com/farmerbb/taskbar/service/UIHostServiceNewControllerTest.java
<|file_sep|>updated/app/src/test/java/com/farmerbb/taskbar/service/UIHostServiceNewControllerTest.java
package com.farmerbb.taskbar.service;
import com.farmerbb.taskbar.ui.DashboardController;
import com.farmerbb.taskbar.ui.StartMenuController;
import com.farmerbb.taskbar.ui.TaskbarController;
import org.junit.Test;
import org.junit.runner.RunWith;
import org.robolectric.Robolectric;
import org.robolectric.RobolectricTestRunner;
import org.robolectric.android.controller.ServiceController;
import static org.junit.Assert.assertTrue;
@RunWith(RobolectricTestRunner.class)
public class UIHostServiceNewControllerTest {
@Test
public void testDashboardService() {
ServiceController<DashboardService> dashboardService =
Robolectric.buildService(DashboardService.class);
assertTrue(dashboardService.get().newController() instanceof DashboardController); | bec84dec621824dc604fd41cea8b4dce36104a5b | app/src/test/java/com/farmerbb/taskbar/service/UIHostServiceNewControllerTest.java | app/src/test/java/com/farmerbb/taskbar/service/UIHostServiceNewControllerTest.java | Java |
<|file_sep|>original/js/tests.js
window.addEventListener('DOMContentLoaded', function() {
QUnit.test("Hello test", function(assert) {
assert.ok(1 == "1", "Passed!");
});
});
<|file_sep|>current/js/tests.js
window.addEventListener('DOMContentLoaded', function() {
QUnit.test("Hello test", function(assert) {
assert.ok(1 == "1", "Passed!");
});
});
<|file_sep|>updated/js/tests.js | window.addEventListener('DOMContentLoaded', function() {
QUnit.test("Opened server", function(assert){
assert.ok(1 == "1", "Passed!");
});
QUnit.test("Closed server", function(assert){
assert.ok(1 == "1", "Passed!");
});
QUnit.test("...", function(assert){
assert.ok(1 == "1", "Passed!");
});
}); | <|file_sep|>original/js/tests.js
window.addEventListener('DOMContentLoaded', function() {
QUnit.test("Hello test", function(assert) {
assert.ok(1 == "1", "Passed!");
});
});
<|file_sep|>current/js/tests.js
window.addEventListener('DOMContentLoaded', function() {
QUnit.test("Hello test", function(assert) {
assert.ok(1 == "1", "Passed!");
});
});
<|file_sep|>updated/js/tests.js
window.addEventListener('DOMContentLoaded', function() {
QUnit.test("Opened server", function(assert){
assert.ok(1 == "1", "Passed!");
});
QUnit.test("Closed server", function(assert){
assert.ok(1 == "1", "Passed!");
});
QUnit.test("...", function(assert){
assert.ok(1 == "1", "Passed!");
});
}); | 526539c3706dd3d4629ff9f421dc4ba59bc36647 | js/tests.js | js/tests.js | JavaScript |
<|file_sep|>.github/workflows/maven.yml.diff
original:
- name: Checkout
updated:
- name: "Checkout ${{ github.ref }} ( ${{ github.sha }} )"
<|file_sep|>.github/workflows/maven.yml.diff
original:
updated:
with:
persist-ccredentials: false
<|file_sep|>.github/workflows/maven.yml.diff
original:
run: mvn clean install -DskipTests=true -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V --toolchains .github/workflows/.toolchains.xml
updated:
run: mvn clean install -DskipTests=true -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V -q --toolchains .github/workflows/.toolchains.xml
<|file_sep|>original/.github/workflows/maven.yml
uses: actions/cache@v2
with:
path: ~/.m2/repository
key: build-${{ runner.os }}-maven-${{ hashFiles('**/pom.xml') }}
restore-keys: |
build-${{ runner.os }}-maven-
- name: Install JDK
uses: AdoptOpenJDK/install-jdk@v1
with:
version: '8'
architecture: x64
impl: hotspot
targets: 'JDK_8;JAVA_HOME'
- name: Install Dependencies
run: mvn clean install -DskipTests=true -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V --toolchains .github/workflows/.toolchains.xml
- name: Test & Report
if: ${{ success() }}
run: mvn verify coveralls:report -Dcoveralls-repo-token=${repo_token} -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V --toolchains .github/workflows/.toolchains.xml
<|file_sep|>current/.github/workflows/maven.yml
uses: actions/cache@v2
with:
path: ~/.m2/repository
key: build-${{ runner.os }}-maven-${{ hashFiles('**/pom.xml') }}
restore-keys: |
build-${{ runner.os }}-maven-
- name: Install JDK
uses: AdoptOpenJDK/install-jdk@v1
with:
version: '8'
architecture: x64
impl: hotspot
targets: 'JDK_8;JAVA_HOME'
- name: Install Dependencies
run: mvn clean install -DskipTests=true -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V -q --toolchains .github/workflows/.toolchains.xml
- name: Test & Report
if: ${{ success() }}
run: mvn verify coveralls:report -Dcoveralls-repo-token=${repo_token} -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V --toolchains .github/workflows/.toolchains.xml
<|file_sep|>updated/.github/workflows/maven.yml | uses: actions/cache@v2
with:
path: ~/.m2/repository
key: build-${{ runner.os }}-maven-${{ hashFiles('**/pom.xml') }}
restore-keys: |
build-${{ runner.os }}-maven-
- name: Install JDK
uses: AdoptOpenJDK/install-jdk@v1
with:
version: '8'
architecture: x64
impl: hotspot
targets: 'JDK_8;JAVA_HOME'
- name: Install Dependencies
run: mvn clean install -DskipTests=true -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V -q --toolchains .github/workflows/.toolchains.xml
- name: Test & Report
if: ${{ success() }}
run: mvn verify coveralls:report -Dcoveralls-repo-token=${repo_token} -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V -q --toolchains .github/workflows/.toolchains.xml | <|file_sep|>.github/workflows/maven.yml.diff
original:
- name: Checkout
updated:
- name: "Checkout ${{ github.ref }} ( ${{ github.sha }} )"
<|file_sep|>.github/workflows/maven.yml.diff
original:
updated:
with:
persist-ccredentials: false
<|file_sep|>.github/workflows/maven.yml.diff
original:
run: mvn clean install -DskipTests=true -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V --toolchains .github/workflows/.toolchains.xml
updated:
run: mvn clean install -DskipTests=true -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V -q --toolchains .github/workflows/.toolchains.xml
<|file_sep|>original/.github/workflows/maven.yml
uses: actions/cache@v2
with:
path: ~/.m2/repository
key: build-${{ runner.os }}-maven-${{ hashFiles('**/pom.xml') }}
restore-keys: |
build-${{ runner.os }}-maven-
- name: Install JDK
uses: AdoptOpenJDK/install-jdk@v1
with:
version: '8'
architecture: x64
impl: hotspot
targets: 'JDK_8;JAVA_HOME'
- name: Install Dependencies
run: mvn clean install -DskipTests=true -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V --toolchains .github/workflows/.toolchains.xml
- name: Test & Report
if: ${{ success() }}
run: mvn verify coveralls:report -Dcoveralls-repo-token=${repo_token} -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V --toolchains .github/workflows/.toolchains.xml
<|file_sep|>current/.github/workflows/maven.yml
uses: actions/cache@v2
with:
path: ~/.m2/repository
key: build-${{ runner.os }}-maven-${{ hashFiles('**/pom.xml') }}
restore-keys: |
build-${{ runner.os }}-maven-
- name: Install JDK
uses: AdoptOpenJDK/install-jdk@v1
with:
version: '8'
architecture: x64
impl: hotspot
targets: 'JDK_8;JAVA_HOME'
- name: Install Dependencies
run: mvn clean install -DskipTests=true -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V -q --toolchains .github/workflows/.toolchains.xml
- name: Test & Report
if: ${{ success() }}
run: mvn verify coveralls:report -Dcoveralls-repo-token=${repo_token} -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V --toolchains .github/workflows/.toolchains.xml
<|file_sep|>updated/.github/workflows/maven.yml
uses: actions/cache@v2
with:
path: ~/.m2/repository
key: build-${{ runner.os }}-maven-${{ hashFiles('**/pom.xml') }}
restore-keys: |
build-${{ runner.os }}-maven-
- name: Install JDK
uses: AdoptOpenJDK/install-jdk@v1
with:
version: '8'
architecture: x64
impl: hotspot
targets: 'JDK_8;JAVA_HOME'
- name: Install Dependencies
run: mvn clean install -DskipTests=true -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V -q --toolchains .github/workflows/.toolchains.xml
- name: Test & Report
if: ${{ success() }}
run: mvn verify coveralls:report -Dcoveralls-repo-token=${repo_token} -Dmaven.javadoc.skip=true -Dgpg.skip=true -B -V -q --toolchains .github/workflows/.toolchains.xml | 0dc2035ce98bf88de811acc7988fb2adc2b88faa | .github/workflows/maven.yml | .github/workflows/maven.yml | YAML |
<|file_sep|>swgohBotShard.js.diff
original:
Manager.spawn();
updated:
// Give it a large timeout since it refuses to work otherwise
Manager.spawn({timeout: 60000});
<|file_sep|>original/swgohBotShard.js
// At max, each shard can have 2500 servers
const config = require("./config.js");
const Discord = require("discord.js");
const Manager = new Discord.ShardingManager("./swgohBot.js",{
totalShards: config.shardCount // Tell it how many shards we want (Approx. 1100 servers per shard)
});
Manager.spawn();
Manager.on("shardCreate", (shard) => {
console.log(`Creating Shard ${shard.id + 1}/${Manager.totalShards}`);
});
<|file_sep|>current/swgohBotShard.js
// At max, each shard can have 2500 servers
const config = require("./config.js");
const Discord = require("discord.js");
const Manager = new Discord.ShardingManager("./swgohBot.js",{
totalShards: config.shardCount // Tell it how many shards we want (Approx. 1100 servers per shard)
});
// Give it a large timeout since it refuses to work otherwise
Manager.spawn({timeout: 60000});
Manager.on("shardCreate", (shard) => {
console.log(`Creating Shard ${shard.id + 1}/${Manager.totalShards}`);
});
<|file_sep|>updated/swgohBotShard.js |
const Discord = require("discord.js");
const Manager = new Discord.ShardingManager("./swgohBot.js",{
totalShards: config.shardCount // Tell it how many shards we want (Approx. 1100 servers per shard)
});
// Give it a large timeout since it refuses to work otherwise
Manager.spawn({timeout: 60000});
Manager.on("shardCreate", (shard) => {
shard.on("reconnecting", () => {
console.log(`Reconnecting shard: [${shard.id}]`);
});
shard.on("spawn", () => {
console.log(`Spawned shard: [${shard.id}]`);
});
// shard.on("ready", () => {
// console.log(`Shard [${shard.id}] is ready`);
// });
shard.on("death", () => {
console.log(`Shard Died: [${shard.id}]`); | <|file_sep|>swgohBotShard.js.diff
original:
Manager.spawn();
updated:
// Give it a large timeout since it refuses to work otherwise
Manager.spawn({timeout: 60000});
<|file_sep|>original/swgohBotShard.js
// At max, each shard can have 2500 servers
const config = require("./config.js");
const Discord = require("discord.js");
const Manager = new Discord.ShardingManager("./swgohBot.js",{
totalShards: config.shardCount // Tell it how many shards we want (Approx. 1100 servers per shard)
});
Manager.spawn();
Manager.on("shardCreate", (shard) => {
console.log(`Creating Shard ${shard.id + 1}/${Manager.totalShards}`);
});
<|file_sep|>current/swgohBotShard.js
// At max, each shard can have 2500 servers
const config = require("./config.js");
const Discord = require("discord.js");
const Manager = new Discord.ShardingManager("./swgohBot.js",{
totalShards: config.shardCount // Tell it how many shards we want (Approx. 1100 servers per shard)
});
// Give it a large timeout since it refuses to work otherwise
Manager.spawn({timeout: 60000});
Manager.on("shardCreate", (shard) => {
console.log(`Creating Shard ${shard.id + 1}/${Manager.totalShards}`);
});
<|file_sep|>updated/swgohBotShard.js
const Discord = require("discord.js");
const Manager = new Discord.ShardingManager("./swgohBot.js",{
totalShards: config.shardCount // Tell it how many shards we want (Approx. 1100 servers per shard)
});
// Give it a large timeout since it refuses to work otherwise
Manager.spawn({timeout: 60000});
Manager.on("shardCreate", (shard) => {
shard.on("reconnecting", () => {
console.log(`Reconnecting shard: [${shard.id}]`);
});
shard.on("spawn", () => {
console.log(`Spawned shard: [${shard.id}]`);
});
// shard.on("ready", () => {
// console.log(`Shard [${shard.id}] is ready`);
// });
shard.on("death", () => {
console.log(`Shard Died: [${shard.id}]`); | d9c16a5304f77848f4b8f2c9344e2f08699ee38e | swgohBotShard.js | swgohBotShard.js | JavaScript |
<|file_sep|>nim.nim.diff
original:
# Set cc = clang in compiler's config/nim.cfg for better performance
updated:
<|file_sep|>nim.nim.diff
original:
proc readPlaces: tuple[nodes: seq[Node], num: int] =
let lines = toSeq("agraph".lines)
let numNodes = lines[0].parseInt
var nodes = newSeqWith(numNodes, Node(neighbours: newSeq[Route]()))
updated:
let lines = toSeq("agraph".lines)
let numNodes = lines[0].parseInt
var nodes = newSeqWith(numNodes, Node(neighbours: newSeq[Route]()))
<|file_sep|>nim.nim.diff
original:
for i in 1.. < lines.len:
let nums = lines[i].split(' ')
if nums.len < 3:
break
updated:
for i in 1.. < lines.len:
let nums = lines[i].split(' ')
if nums.len < 3:
break
<|file_sep|>nim.nim.diff
original:
let node = nums[0].parseInt
let neighbour = nums[1].parseInt
let cost = nums[2].parseInt
updated:
let node = nums[0].parseInt
let neighbour = nums[1].parseInt
let cost = nums[2].parseInt
<|file_sep|>nim.nim.diff
original:
nodes[node].neighbours.add(
Route(dest: neighbour, cost: cost))
updated:
nodes[node].neighbours.add(Route(dest: neighbour, cost: cost))
<|file_sep|>original/nim.nim
let cost = nums[2].parseInt
nodes[node].neighbours.add(
Route(dest: neighbour, cost: cost))
return (nodes, numNodes)
proc getLongestPath(nodes: seq[Node], nodeId: int, visited: var seq[bool]): int =
visited[nodeId] = true
for neighbour in nodes[nodeId].neighbours:
if not visited[neighbour.dest]:
let dist = neighbour.cost + getLongestPath(nodes, neighbour.dest, visited)
if dist > result:
result = dist
visited[nodeId] = false
proc main =
let (nodes, numNodes) = readPlaces()
var visited = newSeqWith(numNodes, false)
let start = cpuTime()
<|file_sep|>current/nim.nim
nodes[node].neighbours.add(Route(dest: neighbour, cost: cost))
var visited = newSeq[bool](numNodes)
let start = cpuTime()
proc getLongestPath(nodeId: int): int =
visited[nodeId] = true
for neighbour in nodes[nodeId].neighbours:
if not visited[neighbour.dest]:
let dist = neighbour.cost + getLongestPath(neighbour.dest)
if dist > result:
result = dist
visited[nodeId] = false
let result = getLongestPath(0)
let duration = cpuTime() - start
echo result, " LANGUAGE NIM ", int(duration * 1000)
main()
<|file_sep|>updated/nim.nim | let cost = nums[2].parseInt
nodes[node].neighbours.add(Route(dest: neighbour, cost: cost))
var visited = newSeq[bool](numNodes)
let start = cpuTime()
proc getLongestPath(nodeId: int): int =
visited[nodeId] = true
for neighbour in nodes[nodeId].neighbours:
if not visited[neighbour.dest]:
let dist = neighbour.cost + getLongestPath(neighbour.dest)
if dist > result:
result = dist
visited[nodeId] = false
let result = getLongestPath(0)
let duration = cpuTime() - start
echo result, " LANGUAGE NIM ", int(duration * 1000) | <|file_sep|>nim.nim.diff
original:
# Set cc = clang in compiler's config/nim.cfg for better performance
updated:
<|file_sep|>nim.nim.diff
original:
proc readPlaces: tuple[nodes: seq[Node], num: int] =
let lines = toSeq("agraph".lines)
let numNodes = lines[0].parseInt
var nodes = newSeqWith(numNodes, Node(neighbours: newSeq[Route]()))
updated:
let lines = toSeq("agraph".lines)
let numNodes = lines[0].parseInt
var nodes = newSeqWith(numNodes, Node(neighbours: newSeq[Route]()))
<|file_sep|>nim.nim.diff
original:
for i in 1.. < lines.len:
let nums = lines[i].split(' ')
if nums.len < 3:
break
updated:
for i in 1.. < lines.len:
let nums = lines[i].split(' ')
if nums.len < 3:
break
<|file_sep|>nim.nim.diff
original:
let node = nums[0].parseInt
let neighbour = nums[1].parseInt
let cost = nums[2].parseInt
updated:
let node = nums[0].parseInt
let neighbour = nums[1].parseInt
let cost = nums[2].parseInt
<|file_sep|>nim.nim.diff
original:
nodes[node].neighbours.add(
Route(dest: neighbour, cost: cost))
updated:
nodes[node].neighbours.add(Route(dest: neighbour, cost: cost))
<|file_sep|>original/nim.nim
let cost = nums[2].parseInt
nodes[node].neighbours.add(
Route(dest: neighbour, cost: cost))
return (nodes, numNodes)
proc getLongestPath(nodes: seq[Node], nodeId: int, visited: var seq[bool]): int =
visited[nodeId] = true
for neighbour in nodes[nodeId].neighbours:
if not visited[neighbour.dest]:
let dist = neighbour.cost + getLongestPath(nodes, neighbour.dest, visited)
if dist > result:
result = dist
visited[nodeId] = false
proc main =
let (nodes, numNodes) = readPlaces()
var visited = newSeqWith(numNodes, false)
let start = cpuTime()
<|file_sep|>current/nim.nim
nodes[node].neighbours.add(Route(dest: neighbour, cost: cost))
var visited = newSeq[bool](numNodes)
let start = cpuTime()
proc getLongestPath(nodeId: int): int =
visited[nodeId] = true
for neighbour in nodes[nodeId].neighbours:
if not visited[neighbour.dest]:
let dist = neighbour.cost + getLongestPath(neighbour.dest)
if dist > result:
result = dist
visited[nodeId] = false
let result = getLongestPath(0)
let duration = cpuTime() - start
echo result, " LANGUAGE NIM ", int(duration * 1000)
main()
<|file_sep|>updated/nim.nim
let cost = nums[2].parseInt
nodes[node].neighbours.add(Route(dest: neighbour, cost: cost))
var visited = newSeq[bool](numNodes)
let start = cpuTime()
proc getLongestPath(nodeId: int): int =
visited[nodeId] = true
for neighbour in nodes[nodeId].neighbours:
if not visited[neighbour.dest]:
let dist = neighbour.cost + getLongestPath(neighbour.dest)
if dist > result:
result = dist
visited[nodeId] = false
let result = getLongestPath(0)
let duration = cpuTime() - start
echo result, " LANGUAGE NIM ", int(duration * 1000) | 2278a4d52f5fb0ab5a39efee85a9efd0cd1d5d89 | nim.nim | nim.nim | Nimrod |
<|file_sep|>original/lib/facter/xcodeversion.rb
Facter.add(:xcodeversion) do
confine :operatingsystem => :darwin
setcode do
return "" if ! File.exists('/usr/bin/xcodebuild')
results = %x{ /usr/bin/xcodebuild -version 2>&1 }
return "" if ! results =~ /^Xcode\s((?:\d+\.)?(?:\d+\.)?\d+)/
$1
end
end
<|file_sep|>current/lib/facter/xcodeversion.rb
Facter.add(:xcodeversion) do
confine :operatingsystem => :darwin
setcode do
return "" if ! File.exists('/usr/bin/xcodebuild')
results = %x{ /usr/bin/xcodebuild -version 2>&1 }
return "" if ! results =~ /^Xcode\s((?:\d+\.)?(?:\d+\.)?\d+)/
$1
end
end
<|file_sep|>updated/lib/facter/xcodeversion.rb | Facter.add(:xcodeversion) do
confine :operatingsystem => :darwin
setcode do
return "" if ! File.exists?('/usr/bin/xcodebuild')
results = %x{ /usr/bin/xcodebuild -version 2>&1 }
return "" if ! results =~ /^Xcode\s((?:\d+\.)?(?:\d+\.)?\d+)/
$1
end
end
| <|file_sep|>original/lib/facter/xcodeversion.rb
Facter.add(:xcodeversion) do
confine :operatingsystem => :darwin
setcode do
return "" if ! File.exists('/usr/bin/xcodebuild')
results = %x{ /usr/bin/xcodebuild -version 2>&1 }
return "" if ! results =~ /^Xcode\s((?:\d+\.)?(?:\d+\.)?\d+)/
$1
end
end
<|file_sep|>current/lib/facter/xcodeversion.rb
Facter.add(:xcodeversion) do
confine :operatingsystem => :darwin
setcode do
return "" if ! File.exists('/usr/bin/xcodebuild')
results = %x{ /usr/bin/xcodebuild -version 2>&1 }
return "" if ! results =~ /^Xcode\s((?:\d+\.)?(?:\d+\.)?\d+)/
$1
end
end
<|file_sep|>updated/lib/facter/xcodeversion.rb
Facter.add(:xcodeversion) do
confine :operatingsystem => :darwin
setcode do
return "" if ! File.exists?('/usr/bin/xcodebuild')
results = %x{ /usr/bin/xcodebuild -version 2>&1 }
return "" if ! results =~ /^Xcode\s((?:\d+\.)?(?:\d+\.)?\d+)/
$1
end
end
| f2ed62fe6d691c2c1a928357cee4eb1a7fcf2e1e | lib/facter/xcodeversion.rb | lib/facter/xcodeversion.rb | Ruby |
<|file_sep|>original/app/models/budget/group.rb
validates_translation :name, presence: true
validates :budget_id, presence: true
validates :slug, presence: true, format: /\A[a-z0-9\-_]+\z/
def self.sort_by_name
all.sort_by(&:name)
end
def single_heading_group?
headings.count == 1
end
private
def generate_slug?
slug.nil? || budget.drafting?
end
end
end
<|file_sep|>current/app/models/budget/group.rb
validates_translation :name, presence: true
validates :budget_id, presence: true
validates :slug, presence: true, format: /\A[a-z0-9\-_]+\z/
def self.sort_by_name
all.sort_by(&:name)
end
def single_heading_group?
headings.count == 1
end
private
def generate_slug?
slug.nil? || budget.drafting?
end
end
end
<|file_sep|>updated/app/models/budget/group.rb | def single_heading_group?
headings.count == 1
end
private
def generate_slug?
slug.nil? || budget.drafting?
end
class Translation < Globalize::ActiveRecord::Translation
delegate :budget, to: :globalized_model
validate :name_uniqueness_by_budget
def name_uniqueness_by_budget
if budget.groups.joins(:translations)
.where(name: name)
.where.not("budget_group_translations.budget_group_id": budget_group_id).any?
errors.add(:name, I18n.t("errors.messages.taken"))
end | <|file_sep|>original/app/models/budget/group.rb
validates_translation :name, presence: true
validates :budget_id, presence: true
validates :slug, presence: true, format: /\A[a-z0-9\-_]+\z/
def self.sort_by_name
all.sort_by(&:name)
end
def single_heading_group?
headings.count == 1
end
private
def generate_slug?
slug.nil? || budget.drafting?
end
end
end
<|file_sep|>current/app/models/budget/group.rb
validates_translation :name, presence: true
validates :budget_id, presence: true
validates :slug, presence: true, format: /\A[a-z0-9\-_]+\z/
def self.sort_by_name
all.sort_by(&:name)
end
def single_heading_group?
headings.count == 1
end
private
def generate_slug?
slug.nil? || budget.drafting?
end
end
end
<|file_sep|>updated/app/models/budget/group.rb
def single_heading_group?
headings.count == 1
end
private
def generate_slug?
slug.nil? || budget.drafting?
end
class Translation < Globalize::ActiveRecord::Translation
delegate :budget, to: :globalized_model
validate :name_uniqueness_by_budget
def name_uniqueness_by_budget
if budget.groups.joins(:translations)
.where(name: name)
.where.not("budget_group_translations.budget_group_id": budget_group_id).any?
errors.add(:name, I18n.t("errors.messages.taken"))
end | c2f393276af6acc99594b89a1788b6336e4fa421 | app/models/budget/group.rb | app/models/budget/group.rb | Ruby |
<|file_sep|>original/.travis.yml
language: go
go:
- 1.3
- 1.4
- 1.5
script:
- go test ./...
<|file_sep|>current/.travis.yml
language: go
go:
- 1.3
- 1.4
- 1.5
script:
- go test ./...
<|file_sep|>updated/.travis.yml | language: go
sudo: false
go:
- 1.3
- 1.4
- 1.5
script:
- go test ./... | <|file_sep|>original/.travis.yml
language: go
go:
- 1.3
- 1.4
- 1.5
script:
- go test ./...
<|file_sep|>current/.travis.yml
language: go
go:
- 1.3
- 1.4
- 1.5
script:
- go test ./...
<|file_sep|>updated/.travis.yml
language: go
sudo: false
go:
- 1.3
- 1.4
- 1.5
script:
- go test ./... | ac1c0467274bf9644a71d42d507b5b2d6f67e9c4 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/sidekiq-middleware.gemspec
# -*- encoding: utf-8 -*-
require File.expand_path('../lib/sidekiq-middleware/version', __FILE__)
Gem::Specification.new do |gem|
gem.authors = ["Dmitry Krasnoukhov"]
gem.email = ["dmitry@krasnoukhov.com"]
gem.description = gem.summary = "Additional sidekiq middleware"
gem.homepage = "http://github.com/krasnoukhov/sidekiq-middleware"
gem.files = `git ls-files`.split($\)
gem.executables = gem.files.grep(%r{^bin/}).map{ |f| File.basename(f) }
gem.test_files = gem.files.grep(%r{^(test|spec|features)/})
gem.name = "sidekiq-middleware"
gem.require_paths = ["lib"]
gem.version = Sidekiq::Middleware::VERSION
gem.add_dependency 'sidekiq', '~> 2.12.4'
gem.add_development_dependency 'rake'
gem.add_development_dependency 'bundler', '~> 1.0'
gem.add_development_dependency 'minitest', '~> 3'
end
<|file_sep|>current/sidekiq-middleware.gemspec
# -*- encoding: utf-8 -*-
require File.expand_path('../lib/sidekiq-middleware/version', __FILE__)
Gem::Specification.new do |gem|
gem.authors = ["Dmitry Krasnoukhov"]
gem.email = ["dmitry@krasnoukhov.com"]
gem.description = gem.summary = "Additional sidekiq middleware"
gem.homepage = "http://github.com/krasnoukhov/sidekiq-middleware"
gem.files = `git ls-files`.split($\)
gem.executables = gem.files.grep(%r{^bin/}).map{ |f| File.basename(f) }
gem.test_files = gem.files.grep(%r{^(test|spec|features)/})
gem.name = "sidekiq-middleware"
gem.require_paths = ["lib"]
gem.version = Sidekiq::Middleware::VERSION
gem.add_dependency 'sidekiq', '~> 2.12.4'
gem.add_development_dependency 'rake'
gem.add_development_dependency 'bundler', '~> 1.0'
gem.add_development_dependency 'minitest', '~> 3'
end
<|file_sep|>updated/sidekiq-middleware.gemspec | # -*- encoding: utf-8 -*-
require File.expand_path('../lib/sidekiq-middleware/version', __FILE__)
Gem::Specification.new do |gem|
gem.authors = ["Dmitry Krasnoukhov"]
gem.email = ["dmitry@krasnoukhov.com"]
gem.description = gem.summary = "Additional sidekiq middleware"
gem.homepage = "http://github.com/krasnoukhov/sidekiq-middleware"
gem.files = `git ls-files`.split($\)
gem.executables = gem.files.grep(%r{^bin/}).map{ |f| File.basename(f) }
gem.test_files = gem.files.grep(%r{^(test|spec|features)/})
gem.name = "sidekiq-middleware"
gem.require_paths = ["lib"]
gem.version = Sidekiq::Middleware::VERSION
gem.add_dependency 'sidekiq', '>= 2.12.4'
gem.add_development_dependency 'rake'
gem.add_development_dependency 'bundler', '~> 1.0'
gem.add_development_dependency 'minitest', '~> 3'
end | <|file_sep|>original/sidekiq-middleware.gemspec
# -*- encoding: utf-8 -*-
require File.expand_path('../lib/sidekiq-middleware/version', __FILE__)
Gem::Specification.new do |gem|
gem.authors = ["Dmitry Krasnoukhov"]
gem.email = ["dmitry@krasnoukhov.com"]
gem.description = gem.summary = "Additional sidekiq middleware"
gem.homepage = "http://github.com/krasnoukhov/sidekiq-middleware"
gem.files = `git ls-files`.split($\)
gem.executables = gem.files.grep(%r{^bin/}).map{ |f| File.basename(f) }
gem.test_files = gem.files.grep(%r{^(test|spec|features)/})
gem.name = "sidekiq-middleware"
gem.require_paths = ["lib"]
gem.version = Sidekiq::Middleware::VERSION
gem.add_dependency 'sidekiq', '~> 2.12.4'
gem.add_development_dependency 'rake'
gem.add_development_dependency 'bundler', '~> 1.0'
gem.add_development_dependency 'minitest', '~> 3'
end
<|file_sep|>current/sidekiq-middleware.gemspec
# -*- encoding: utf-8 -*-
require File.expand_path('../lib/sidekiq-middleware/version', __FILE__)
Gem::Specification.new do |gem|
gem.authors = ["Dmitry Krasnoukhov"]
gem.email = ["dmitry@krasnoukhov.com"]
gem.description = gem.summary = "Additional sidekiq middleware"
gem.homepage = "http://github.com/krasnoukhov/sidekiq-middleware"
gem.files = `git ls-files`.split($\)
gem.executables = gem.files.grep(%r{^bin/}).map{ |f| File.basename(f) }
gem.test_files = gem.files.grep(%r{^(test|spec|features)/})
gem.name = "sidekiq-middleware"
gem.require_paths = ["lib"]
gem.version = Sidekiq::Middleware::VERSION
gem.add_dependency 'sidekiq', '~> 2.12.4'
gem.add_development_dependency 'rake'
gem.add_development_dependency 'bundler', '~> 1.0'
gem.add_development_dependency 'minitest', '~> 3'
end
<|file_sep|>updated/sidekiq-middleware.gemspec
# -*- encoding: utf-8 -*-
require File.expand_path('../lib/sidekiq-middleware/version', __FILE__)
Gem::Specification.new do |gem|
gem.authors = ["Dmitry Krasnoukhov"]
gem.email = ["dmitry@krasnoukhov.com"]
gem.description = gem.summary = "Additional sidekiq middleware"
gem.homepage = "http://github.com/krasnoukhov/sidekiq-middleware"
gem.files = `git ls-files`.split($\)
gem.executables = gem.files.grep(%r{^bin/}).map{ |f| File.basename(f) }
gem.test_files = gem.files.grep(%r{^(test|spec|features)/})
gem.name = "sidekiq-middleware"
gem.require_paths = ["lib"]
gem.version = Sidekiq::Middleware::VERSION
gem.add_dependency 'sidekiq', '>= 2.12.4'
gem.add_development_dependency 'rake'
gem.add_development_dependency 'bundler', '~> 1.0'
gem.add_development_dependency 'minitest', '~> 3'
end | 6e942250df1754594c9ed582f9153384a7936920 | sidekiq-middleware.gemspec | sidekiq-middleware.gemspec | Ruby |
<|file_sep|>original/FileProcessingApp/src/main/java/com/atanasg/fileprocessingapp/FileProcessingAppLauncher.java
<|file_sep|>current/FileProcessingApp/src/main/java/com/atanasg/fileprocessingapp/FileProcessingAppLauncher.java
<|file_sep|>updated/FileProcessingApp/src/main/java/com/atanasg/fileprocessingapp/FileProcessingAppLauncher.java | /*
* Copyright (C) 2015 Atanas Gegov
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.atanasg.fileprocessingapp;
import com.atanasg.fileprocessingapp.mvc.controller.FileProcessingAppController;
import com.atanasg.fileprocessingapp.mvc.controller.FileProcessingAppControllerImpl;
import com.atanasg.fileprocessingapp.mvc.model.FileContentModel;
import com.atanasg.fileprocessingapp.mvc.model.FileContentModelImpl; | <|file_sep|>original/FileProcessingApp/src/main/java/com/atanasg/fileprocessingapp/FileProcessingAppLauncher.java
<|file_sep|>current/FileProcessingApp/src/main/java/com/atanasg/fileprocessingapp/FileProcessingAppLauncher.java
<|file_sep|>updated/FileProcessingApp/src/main/java/com/atanasg/fileprocessingapp/FileProcessingAppLauncher.java
/*
* Copyright (C) 2015 Atanas Gegov
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.atanasg.fileprocessingapp;
import com.atanasg.fileprocessingapp.mvc.controller.FileProcessingAppController;
import com.atanasg.fileprocessingapp.mvc.controller.FileProcessingAppControllerImpl;
import com.atanasg.fileprocessingapp.mvc.model.FileContentModel;
import com.atanasg.fileprocessingapp.mvc.model.FileContentModelImpl; | 28c989143ca44e365128de7642ec8bb830d5598c | FileProcessingApp/src/main/java/com/atanasg/fileprocessingapp/FileProcessingAppLauncher.java | FileProcessingApp/src/main/java/com/atanasg/fileprocessingapp/FileProcessingAppLauncher.java | Java |
<|file_sep|>app/oauthtest.html.diff
original:
<script src="/path/to/oauth.js"></script>
updated:
<script src="/bower_components/oauth.io/oauth.js"></script>
<script src="/bower_components/jquery/dist/jquery.js"></script>
<|file_sep|>app/oauthtest.html.diff
original:
OAuth.initialize('d1RnDSEcEeSuMwhQCxUiAy_lN5w');
$.get( "", function( data ) {
OAuth.popup('google_plus', function (err, res) {
res.get('/plus/v1/people/me').done(function (me) {
$('#connect').slideUp('fast')
$('#res').html(template({
data: me
})).slideDown('fast')
res.get('/plus/v1/people/me/activities/public').done(function(activities) {
$('#activities').html(activitiesTemplate({
data: activities
}))
});
})
})
updated:
var access_token;
OAuth.initialize('BLyjDMMIj7nRjs2m3bCFWB0ZsTA');
//Using popup (option 1)
OAuth.popup('google', function(error, result) {
console.log(result);
// access_token = result.access_token;
console.log(error);
//handle error with error
//use result.access_token in your API request
<|file_sep|>app/oauthtest.html.diff
original:
updated:
jQuery.support.cors = true;
<|file_sep|>original/app/oauthtest.html
OAuth.popup('google_plus', function (err, res) {
res.get('/plus/v1/people/me').done(function (me) {
$('#connect').slideUp('fast')
$('#res').html(template({
data: me
})).slideDown('fast')
res.get('/plus/v1/people/me/activities/public').done(function(activities) {
$('#activities').html(activitiesTemplate({
data: activities
}))
});
})
})
});
</script>
<meta charset="UTF-8">
<title></title>
</head>
<|file_sep|>current/app/oauthtest.html
var access_token;
OAuth.initialize('BLyjDMMIj7nRjs2m3bCFWB0ZsTA');
//Using popup (option 1)
OAuth.popup('google', function(error, result) {
console.log(result);
// access_token = result.access_token;
console.log(error);
//handle error with error
//use result.access_token in your API request
});
jQuery.support.cors = true;
</script>
<meta charset="UTF-8">
<title></title>
</head>
<body>
</body>
</html>
<|file_sep|>updated/app/oauthtest.html | OAuth.popup('google', function(error, result) {
console.log(result);
// access_token = result.access_token;
console.log(error);
//handle error with error
//use result.access_token in your API request
});
jQuery.support.cors = true;
$.ajax({
type: 'GET',
url: 'https://picasaweb.google.com/105814678861633692185/',
xhrFields: {
withCredentials: true
},
crossDomain: true,
success: function(response, status, xhr) {
console.log('Jupiter client call success ' );
},
error: function(xhr, status, err) { | <|file_sep|>app/oauthtest.html.diff
original:
<script src="/path/to/oauth.js"></script>
updated:
<script src="/bower_components/oauth.io/oauth.js"></script>
<script src="/bower_components/jquery/dist/jquery.js"></script>
<|file_sep|>app/oauthtest.html.diff
original:
OAuth.initialize('d1RnDSEcEeSuMwhQCxUiAy_lN5w');
$.get( "", function( data ) {
OAuth.popup('google_plus', function (err, res) {
res.get('/plus/v1/people/me').done(function (me) {
$('#connect').slideUp('fast')
$('#res').html(template({
data: me
})).slideDown('fast')
res.get('/plus/v1/people/me/activities/public').done(function(activities) {
$('#activities').html(activitiesTemplate({
data: activities
}))
});
})
})
updated:
var access_token;
OAuth.initialize('BLyjDMMIj7nRjs2m3bCFWB0ZsTA');
//Using popup (option 1)
OAuth.popup('google', function(error, result) {
console.log(result);
// access_token = result.access_token;
console.log(error);
//handle error with error
//use result.access_token in your API request
<|file_sep|>app/oauthtest.html.diff
original:
updated:
jQuery.support.cors = true;
<|file_sep|>original/app/oauthtest.html
OAuth.popup('google_plus', function (err, res) {
res.get('/plus/v1/people/me').done(function (me) {
$('#connect').slideUp('fast')
$('#res').html(template({
data: me
})).slideDown('fast')
res.get('/plus/v1/people/me/activities/public').done(function(activities) {
$('#activities').html(activitiesTemplate({
data: activities
}))
});
})
})
});
</script>
<meta charset="UTF-8">
<title></title>
</head>
<|file_sep|>current/app/oauthtest.html
var access_token;
OAuth.initialize('BLyjDMMIj7nRjs2m3bCFWB0ZsTA');
//Using popup (option 1)
OAuth.popup('google', function(error, result) {
console.log(result);
// access_token = result.access_token;
console.log(error);
//handle error with error
//use result.access_token in your API request
});
jQuery.support.cors = true;
</script>
<meta charset="UTF-8">
<title></title>
</head>
<body>
</body>
</html>
<|file_sep|>updated/app/oauthtest.html
OAuth.popup('google', function(error, result) {
console.log(result);
// access_token = result.access_token;
console.log(error);
//handle error with error
//use result.access_token in your API request
});
jQuery.support.cors = true;
$.ajax({
type: 'GET',
url: 'https://picasaweb.google.com/105814678861633692185/',
xhrFields: {
withCredentials: true
},
crossDomain: true,
success: function(response, status, xhr) {
console.log('Jupiter client call success ' );
},
error: function(xhr, status, err) { | 9ff68a210023c96f16a5266e6dfb3798bd34aa29 | app/oauthtest.html | app/oauthtest.html | HTML |
<|file_sep|>original/Resources/views/icons.html.twig
{% macro icon(name, inverted) %}
<i class="icon-{{ name }}{% if inverted|default(false) %} icon-white{% endif %}"></i>
{% endmacro %}
<|file_sep|>current/Resources/views/icons.html.twig
{% macro icon(name, inverted) %}
<i class="icon-{{ name }}{% if inverted|default(false) %} icon-white{% endif %}"></i>
{% endmacro %}
<|file_sep|>updated/Resources/views/icons.html.twig | {% macro icon(name, inverted) %}
<span class="glyphicon glyphicon-{{ name }}"{% if inverted|default(false) %} style="color: white;"{% endif %}></span>
{% endmacro %} | <|file_sep|>original/Resources/views/icons.html.twig
{% macro icon(name, inverted) %}
<i class="icon-{{ name }}{% if inverted|default(false) %} icon-white{% endif %}"></i>
{% endmacro %}
<|file_sep|>current/Resources/views/icons.html.twig
{% macro icon(name, inverted) %}
<i class="icon-{{ name }}{% if inverted|default(false) %} icon-white{% endif %}"></i>
{% endmacro %}
<|file_sep|>updated/Resources/views/icons.html.twig
{% macro icon(name, inverted) %}
<span class="glyphicon glyphicon-{{ name }}"{% if inverted|default(false) %} style="color: white;"{% endif %}></span>
{% endmacro %} | a6e2035f5d92d2c540b89b795bc756ddb3d10dea | Resources/views/icons.html.twig | Resources/views/icons.html.twig | Twig |
<|file_sep|>.github/workflows/failure-cases.yml.diff
original:
continue-on-error: true
updated:
<|file_sep|>.github/workflows/failure-cases.yml.diff
original:
updated:
continue-on-error: true
<|file_sep|>.github/workflows/failure-cases.yml.diff
original:
continue-on-error: true
updated:
<|file_sep|>original/.github/workflows/failure-cases.yml
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v2
- name: Test wrapper missing
uses: ./
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
bad-configuration:
continue-on-error: true
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v2
- name: Test bad config value
uses: ./
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
<|file_sep|>current/.github/workflows/failure-cases.yml
steps:
- name: Checkout sources
uses: actions/checkout@v2
- name: Test wrapper missing
uses: ./
continue-on-error: true
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
bad-configuration:
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v2
- name: Test bad config value
uses: ./
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
cache-disabled: yes
<|file_sep|>updated/.github/workflows/failure-cases.yml | - name: Checkout sources
uses: actions/checkout@v2
- name: Test wrapper missing
uses: ./
continue-on-error: true
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
bad-configuration:
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v2
- name: Test bad config value
uses: ./
continue-on-error: true
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
cache-disabled: yes | <|file_sep|>.github/workflows/failure-cases.yml.diff
original:
continue-on-error: true
updated:
<|file_sep|>.github/workflows/failure-cases.yml.diff
original:
updated:
continue-on-error: true
<|file_sep|>.github/workflows/failure-cases.yml.diff
original:
continue-on-error: true
updated:
<|file_sep|>original/.github/workflows/failure-cases.yml
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v2
- name: Test wrapper missing
uses: ./
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
bad-configuration:
continue-on-error: true
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v2
- name: Test bad config value
uses: ./
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
<|file_sep|>current/.github/workflows/failure-cases.yml
steps:
- name: Checkout sources
uses: actions/checkout@v2
- name: Test wrapper missing
uses: ./
continue-on-error: true
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
bad-configuration:
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v2
- name: Test bad config value
uses: ./
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
cache-disabled: yes
<|file_sep|>updated/.github/workflows/failure-cases.yml
- name: Checkout sources
uses: actions/checkout@v2
- name: Test wrapper missing
uses: ./
continue-on-error: true
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
bad-configuration:
runs-on: ubuntu-latest
steps:
- name: Checkout sources
uses: actions/checkout@v2
- name: Test bad config value
uses: ./
continue-on-error: true
with:
build-root-directory: __tests__/samples/no-wrapper
arguments: help
cache-disabled: yes | 656ad4b5f2a83f6f2d44a36de37837a4fadbff39 | .github/workflows/failure-cases.yml | .github/workflows/failure-cases.yml | YAML |
<|file_sep|>original/bin/restructure.sh
<|file_sep|>current/bin/restructure.sh
<|file_sep|>updated/bin/restructure.sh | #!/bin/bash
# This is a migration script to migrate the current Fork CMS code to the new structure that will allow installing
# things with composer and give the module developers more freedom of how they want to structure the extra code within
# their module
#
# src
# | Files
# | Core
# | | Backend
# | | Frontend
# | Themes
# | | ThemeName
# | Modules
# | | ModuleName
# | | | Backend
# | | | Frontend
# | Installer
# | Console
while true; do
read -p "Is this a dry run (y/n)?" yn
case $yn in | <|file_sep|>original/bin/restructure.sh
<|file_sep|>current/bin/restructure.sh
<|file_sep|>updated/bin/restructure.sh
#!/bin/bash
# This is a migration script to migrate the current Fork CMS code to the new structure that will allow installing
# things with composer and give the module developers more freedom of how they want to structure the extra code within
# their module
#
# src
# | Files
# | Core
# | | Backend
# | | Frontend
# | Themes
# | | ThemeName
# | Modules
# | | ModuleName
# | | | Backend
# | | | Frontend
# | Installer
# | Console
while true; do
read -p "Is this a dry run (y/n)?" yn
case $yn in | a66bd758f1358e632f0b2a78c3f097b154bb5e91 | bin/restructure.sh | bin/restructure.sh | Shell |
<|file_sep|>original/requirements-test.txt
coverage==6.4.3
freezegun==1.2.2
pytest==7.1.2
pytest-cov==3.0.0
requests-mock==1.9.3
<|file_sep|>current/requirements-test.txt
coverage==6.4.3
freezegun==1.2.2
pytest==7.1.2
pytest-cov==3.0.0
requests-mock==1.9.3
<|file_sep|>updated/requirements-test.txt | coverage==6.4.4
freezegun==1.2.2
pytest==7.1.2
pytest-cov==3.0.0
requests-mock==1.9.3 | <|file_sep|>original/requirements-test.txt
coverage==6.4.3
freezegun==1.2.2
pytest==7.1.2
pytest-cov==3.0.0
requests-mock==1.9.3
<|file_sep|>current/requirements-test.txt
coverage==6.4.3
freezegun==1.2.2
pytest==7.1.2
pytest-cov==3.0.0
requests-mock==1.9.3
<|file_sep|>updated/requirements-test.txt
coverage==6.4.4
freezegun==1.2.2
pytest==7.1.2
pytest-cov==3.0.0
requests-mock==1.9.3 | 2916540ec71fa37399f777c76a66f523203332bb | requirements-test.txt | requirements-test.txt | Text |
<|file_sep|>original/test/Interpreter/SDK/objc_nil.swift
<|file_sep|>current/test/Interpreter/SDK/objc_nil.swift
<|file_sep|>updated/test/Interpreter/SDK/objc_nil.swift | // RUN: rm -rf %t/clang-module-cache
// RUN: %swift -module-cache-path=%t/clang-module-cache -sdk=%sdk -i %s | FileCheck %s
// REQUIRES: sdk
import Foundation
var str : NSString = nil
var url : NSURL = nil
println("\(str == nil) \(nil == url) \(str == url)")
// CHECK: true true true
str = "abc"
url = NSURL(initWithString: "file:///")
println("\(str == nil) \(nil == url)")
// CHECK: false false | <|file_sep|>original/test/Interpreter/SDK/objc_nil.swift
<|file_sep|>current/test/Interpreter/SDK/objc_nil.swift
<|file_sep|>updated/test/Interpreter/SDK/objc_nil.swift
// RUN: rm -rf %t/clang-module-cache
// RUN: %swift -module-cache-path=%t/clang-module-cache -sdk=%sdk -i %s | FileCheck %s
// REQUIRES: sdk
import Foundation
var str : NSString = nil
var url : NSURL = nil
println("\(str == nil) \(nil == url) \(str == url)")
// CHECK: true true true
str = "abc"
url = NSURL(initWithString: "file:///")
println("\(str == nil) \(nil == url)")
// CHECK: false false | 3b35451bbee8d7dcf4d703ed82289d83b8511890 | test/Interpreter/SDK/objc_nil.swift | test/Interpreter/SDK/objc_nil.swift | Swift |
<|file_sep|>pom.xml.diff
original:
updated:
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
</properties>
<|file_sep|>original/pom.xml
<packaging>pom</packaging>
<modules>
<module>btrfs-io</module>
<module>btrfs-structures</module>
</modules>
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<version>3.6.1</version>
<configuration>
<source>1.8</source>
<target>1.8</target>
</configuration>
</plugin>
</plugins>
</build>
</project>
<|file_sep|>current/pom.xml
<packaging>pom</packaging>
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
</properties>
<modules>
<module>btrfs-io</module>
<module>btrfs-structures</module>
</modules>
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<version>3.6.1</version>
<configuration>
<source>1.8</source>
<target>1.8</target>
</configuration>
<|file_sep|>updated/pom.xml | <packaging>pom</packaging>
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
</properties>
<modules>
<module>btrfs-io</module>
<module>btrfs-structures</module>
</modules>
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<version>3.6.1</version>
<configuration>
<source>1.8</source>
<target>1.8</target>
</configuration> | <|file_sep|>pom.xml.diff
original:
updated:
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
</properties>
<|file_sep|>original/pom.xml
<packaging>pom</packaging>
<modules>
<module>btrfs-io</module>
<module>btrfs-structures</module>
</modules>
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<version>3.6.1</version>
<configuration>
<source>1.8</source>
<target>1.8</target>
</configuration>
</plugin>
</plugins>
</build>
</project>
<|file_sep|>current/pom.xml
<packaging>pom</packaging>
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
</properties>
<modules>
<module>btrfs-io</module>
<module>btrfs-structures</module>
</modules>
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<version>3.6.1</version>
<configuration>
<source>1.8</source>
<target>1.8</target>
</configuration>
<|file_sep|>updated/pom.xml
<packaging>pom</packaging>
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
</properties>
<modules>
<module>btrfs-io</module>
<module>btrfs-structures</module>
</modules>
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<version>3.6.1</version>
<configuration>
<source>1.8</source>
<target>1.8</target>
</configuration> | 3fd74e374b27ce752b03d121ffd91da3dc819b3a | pom.xml | pom.xml | XML |
<|file_sep|>original/.travis.yml
language: node_js
node_js:
- 'latest'
# Use containers.
# http://docs.travis-ci.com/user/workers/container-based-infrastructure/
sudo: false
<|file_sep|>current/.travis.yml
language: node_js
node_js:
- 'latest'
# Use containers.
# http://docs.travis-ci.com/user/workers/container-based-infrastructure/
sudo: false
<|file_sep|>updated/.travis.yml | language: node_js
node_js:
- 'stable'
# Use containers.
# http://docs.travis-ci.com/user/workers/container-based-infrastructure/
sudo: false | <|file_sep|>original/.travis.yml
language: node_js
node_js:
- 'latest'
# Use containers.
# http://docs.travis-ci.com/user/workers/container-based-infrastructure/
sudo: false
<|file_sep|>current/.travis.yml
language: node_js
node_js:
- 'latest'
# Use containers.
# http://docs.travis-ci.com/user/workers/container-based-infrastructure/
sudo: false
<|file_sep|>updated/.travis.yml
language: node_js
node_js:
- 'stable'
# Use containers.
# http://docs.travis-ci.com/user/workers/container-based-infrastructure/
sudo: false | f555c8190d0be1c355fdc44753a11a629ccedbb9 | .travis.yml | .travis.yml | YAML |
<|file_sep|>package.json.diff
original:
"fh-js-sdk": "^2.17.1"
},
"devDependencies": {
updated:
"fh-js-sdk": "^2.17.1",
<|file_sep|>original/package.json
},
"devDependencies": {
"browserify": "^13.0.1",
"browserify-shim": "^3.8.12",
"grunt": "^0.4.4",
"grunt-browserify": "^5.0.0",
"grunt-cli": "^1.2.0",
"grunt-contrib-clean": "~0.5.0",
"grunt-contrib-connect": "^0.7.1",
"grunt-contrib-watch": "^0.6.1",
"grunt-newer": "~0.5.4",
"grunt-rev": "~0.1.0",
"load-grunt-tasks": "~0.2.0",
"time-grunt": "~0.2.1"
},
"engines": {
"node": ">=0.8.0"
},
"scripts": {
"install": "./node_modules/grunt-cli/bin/grunt browserify"
},
<|file_sep|>current/package.json
"browserify": "^13.0.1",
"browserify-shim": "^3.8.12",
"grunt": "^0.4.4",
"grunt-browserify": "^5.0.0",
"grunt-cli": "^1.2.0",
"grunt-contrib-clean": "~0.5.0",
"grunt-contrib-connect": "^0.7.1",
"grunt-contrib-watch": "^0.6.1",
"grunt-newer": "~0.5.4",
"grunt-rev": "~0.1.0",
"load-grunt-tasks": "~0.2.0",
"time-grunt": "~0.2.1"
},
"engines": {
"node": ">=0.8.0"
},
"scripts": {
"install": "./node_modules/grunt-cli/bin/grunt browserify"
},
"license": "mit"
}
<|file_sep|>updated/package.json | "grunt-browserify": "^5.0.0",
"grunt-cli": "^1.2.0",
"grunt-contrib-clean": "~0.5.0",
"grunt-contrib-connect": "^0.7.1",
"grunt-contrib-watch": "^0.6.1",
"grunt-newer": "~0.5.4",
"grunt-rev": "~0.1.0",
"load-grunt-tasks": "~0.2.0",
"time-grunt": "~0.2.1"
},
"devDependencies": {
},
"engines": {
"node": ">=0.8.0"
},
"scripts": {
"install": "./node_modules/grunt-cli/bin/grunt browserify"
},
"license": "mit"
} | <|file_sep|>package.json.diff
original:
"fh-js-sdk": "^2.17.1"
},
"devDependencies": {
updated:
"fh-js-sdk": "^2.17.1",
<|file_sep|>original/package.json
},
"devDependencies": {
"browserify": "^13.0.1",
"browserify-shim": "^3.8.12",
"grunt": "^0.4.4",
"grunt-browserify": "^5.0.0",
"grunt-cli": "^1.2.0",
"grunt-contrib-clean": "~0.5.0",
"grunt-contrib-connect": "^0.7.1",
"grunt-contrib-watch": "^0.6.1",
"grunt-newer": "~0.5.4",
"grunt-rev": "~0.1.0",
"load-grunt-tasks": "~0.2.0",
"time-grunt": "~0.2.1"
},
"engines": {
"node": ">=0.8.0"
},
"scripts": {
"install": "./node_modules/grunt-cli/bin/grunt browserify"
},
<|file_sep|>current/package.json
"browserify": "^13.0.1",
"browserify-shim": "^3.8.12",
"grunt": "^0.4.4",
"grunt-browserify": "^5.0.0",
"grunt-cli": "^1.2.0",
"grunt-contrib-clean": "~0.5.0",
"grunt-contrib-connect": "^0.7.1",
"grunt-contrib-watch": "^0.6.1",
"grunt-newer": "~0.5.4",
"grunt-rev": "~0.1.0",
"load-grunt-tasks": "~0.2.0",
"time-grunt": "~0.2.1"
},
"engines": {
"node": ">=0.8.0"
},
"scripts": {
"install": "./node_modules/grunt-cli/bin/grunt browserify"
},
"license": "mit"
}
<|file_sep|>updated/package.json
"grunt-browserify": "^5.0.0",
"grunt-cli": "^1.2.0",
"grunt-contrib-clean": "~0.5.0",
"grunt-contrib-connect": "^0.7.1",
"grunt-contrib-watch": "^0.6.1",
"grunt-newer": "~0.5.4",
"grunt-rev": "~0.1.0",
"load-grunt-tasks": "~0.2.0",
"time-grunt": "~0.2.1"
},
"devDependencies": {
},
"engines": {
"node": ">=0.8.0"
},
"scripts": {
"install": "./node_modules/grunt-cli/bin/grunt browserify"
},
"license": "mit"
} | b88b7bc96da86bd32fbdcb6a949fc3caba9e7edb | package.json | package.json | JSON |
<|file_sep|>slides/layout.rb.diff
original:
head = headline 'Flows'
@image = fully_shown_image 'images/flow.png', head.height + OFFSET_FOR_EMPTY_LINE
updated:
headline 'Flows'
@image = fully_shown_image 'images/flow.png', 150
<|file_sep|>original/slides/layout.rb
class Flows < Slide
def content
head = headline 'Flows'
@image = fully_shown_image 'images/flow.png', head.height + OFFSET_FOR_EMPTY_LINE
add_effect do
@image.remove
code 'code/flow.rb', true
end
end
end
class Stacks < Slide
def content
head = headline 'Stacks'
@image = fully_shown_image 'images/stack.png', head.height + OFFSET_FOR_EMPTY_LINE
add_effect do
@image.remove
code 'code/stack.rb', true
end
end
end
<|file_sep|>current/slides/layout.rb
class Flows < Slide
def content
headline 'Flows'
@image = fully_shown_image 'images/flow.png', 150
add_effect do
@image.remove
code 'code/flow.rb', true
end
end
end
class Stacks < Slide
def content
head = headline 'Stacks'
@image = fully_shown_image 'images/stack.png', head.height + OFFSET_FOR_EMPTY_LINE
add_effect do
@image.remove
code 'code/stack.rb', true
end
end
end
<|file_sep|>updated/slides/layout.rb | class Flows < Slide
def content
headline 'Flows'
@image = fully_shown_image 'images/flow.png', 150
add_effect do
@image.remove
code 'code/flow.rb', true
end
end
end
class Stacks < Slide
def content
headline 'Stacks'
@image = fully_shown_image 'images/stack.png', 150
add_effect do
@image.remove
code 'code/stack.rb', true
end
end
end | <|file_sep|>slides/layout.rb.diff
original:
head = headline 'Flows'
@image = fully_shown_image 'images/flow.png', head.height + OFFSET_FOR_EMPTY_LINE
updated:
headline 'Flows'
@image = fully_shown_image 'images/flow.png', 150
<|file_sep|>original/slides/layout.rb
class Flows < Slide
def content
head = headline 'Flows'
@image = fully_shown_image 'images/flow.png', head.height + OFFSET_FOR_EMPTY_LINE
add_effect do
@image.remove
code 'code/flow.rb', true
end
end
end
class Stacks < Slide
def content
head = headline 'Stacks'
@image = fully_shown_image 'images/stack.png', head.height + OFFSET_FOR_EMPTY_LINE
add_effect do
@image.remove
code 'code/stack.rb', true
end
end
end
<|file_sep|>current/slides/layout.rb
class Flows < Slide
def content
headline 'Flows'
@image = fully_shown_image 'images/flow.png', 150
add_effect do
@image.remove
code 'code/flow.rb', true
end
end
end
class Stacks < Slide
def content
head = headline 'Stacks'
@image = fully_shown_image 'images/stack.png', head.height + OFFSET_FOR_EMPTY_LINE
add_effect do
@image.remove
code 'code/stack.rb', true
end
end
end
<|file_sep|>updated/slides/layout.rb
class Flows < Slide
def content
headline 'Flows'
@image = fully_shown_image 'images/flow.png', 150
add_effect do
@image.remove
code 'code/flow.rb', true
end
end
end
class Stacks < Slide
def content
headline 'Stacks'
@image = fully_shown_image 'images/stack.png', 150
add_effect do
@image.remove
code 'code/stack.rb', true
end
end
end | 867718ede864bb0cd01b2ca655b3ab2196a18770 | slides/layout.rb | slides/layout.rb | Ruby |
<|file_sep|>native/chat/compose-thread-button.react.js.diff
original:
import Icon from 'react-native-vector-icons/Ionicons';
updated:
import Icon from 'react-native-vector-icons/MaterialCommunityIcons';
<|file_sep|>native/chat/compose-thread-button.react.js.diff
original:
name="ios-create-outline"
size={30}
updated:
name="pencil-plus-outline"
size={26}
<|file_sep|>original/native/chat/compose-thread-button.react.js
navigate: PropTypes.func.isRequired,
};
render() {
return (
<Button onPress={this.onPress} androidBorderlessRipple={true}>
<Icon
name="ios-create-outline"
size={30}
style={styles.composeButton}
color="#036AFF"
/>
</Button>
);
}
onPress = () => {
this.props.navigate({
routeName: ComposeThreadRouteName,
params: {},
});
<|file_sep|>current/native/chat/compose-thread-button.react.js
navigate: PropTypes.func.isRequired,
};
render() {
return (
<Button onPress={this.onPress} androidBorderlessRipple={true}>
<Icon
name="pencil-plus-outline"
size={26}
style={styles.composeButton}
color="#036AFF"
/>
</Button>
);
}
onPress = () => {
this.props.navigate({
routeName: ComposeThreadRouteName,
params: {},
});
<|file_sep|>updated/native/chat/compose-thread-button.react.js | navigate: PropTypes.func.isRequired,
};
render() {
return (
<Button onPress={this.onPress} androidBorderlessRipple={true}>
<Icon
name="pencil-plus-outline"
size={26}
style={styles.composeButton}
color="#036ABB"
/>
</Button>
);
}
onPress = () => {
this.props.navigate({
routeName: ComposeThreadRouteName,
params: {},
}); | <|file_sep|>native/chat/compose-thread-button.react.js.diff
original:
import Icon from 'react-native-vector-icons/Ionicons';
updated:
import Icon from 'react-native-vector-icons/MaterialCommunityIcons';
<|file_sep|>native/chat/compose-thread-button.react.js.diff
original:
name="ios-create-outline"
size={30}
updated:
name="pencil-plus-outline"
size={26}
<|file_sep|>original/native/chat/compose-thread-button.react.js
navigate: PropTypes.func.isRequired,
};
render() {
return (
<Button onPress={this.onPress} androidBorderlessRipple={true}>
<Icon
name="ios-create-outline"
size={30}
style={styles.composeButton}
color="#036AFF"
/>
</Button>
);
}
onPress = () => {
this.props.navigate({
routeName: ComposeThreadRouteName,
params: {},
});
<|file_sep|>current/native/chat/compose-thread-button.react.js
navigate: PropTypes.func.isRequired,
};
render() {
return (
<Button onPress={this.onPress} androidBorderlessRipple={true}>
<Icon
name="pencil-plus-outline"
size={26}
style={styles.composeButton}
color="#036AFF"
/>
</Button>
);
}
onPress = () => {
this.props.navigate({
routeName: ComposeThreadRouteName,
params: {},
});
<|file_sep|>updated/native/chat/compose-thread-button.react.js
navigate: PropTypes.func.isRequired,
};
render() {
return (
<Button onPress={this.onPress} androidBorderlessRipple={true}>
<Icon
name="pencil-plus-outline"
size={26}
style={styles.composeButton}
color="#036ABB"
/>
</Button>
);
}
onPress = () => {
this.props.navigate({
routeName: ComposeThreadRouteName,
params: {},
}); | bdbdc4e3503e5967e2c810b5d61443aaaa2e4941 | native/chat/compose-thread-button.react.js | native/chat/compose-thread-button.react.js | JavaScript |
<|file_sep|>original/.travis.yml
language: node_js
node_js:
- 0.4
- 0.6
<|file_sep|>current/.travis.yml
language: node_js
node_js:
- 0.4
- 0.6
<|file_sep|>updated/.travis.yml | language: node_js
node_js:
- 0.8
- 0.10
- 0.11 | <|file_sep|>original/.travis.yml
language: node_js
node_js:
- 0.4
- 0.6
<|file_sep|>current/.travis.yml
language: node_js
node_js:
- 0.4
- 0.6
<|file_sep|>updated/.travis.yml
language: node_js
node_js:
- 0.8
- 0.10
- 0.11 | 79edbea1166a21a4260373e960b02948fc6be874 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/Libraries/TurboModule/TurboModuleRegistry.js
'use strict';
const NativeModules = require('../BatchedBridge/NativeModules');
import type {TurboModule} from './RCTExport';
import invariant from 'invariant';
const turboModuleProxy = global.__turboModuleProxy;
export function get<T: TurboModule>(name: string): ?T {
// Backward compatibility layer during migration.
const legacyModule = NativeModules[name];
if (legacyModule != null) {
return ((legacyModule: any): T);
}
if (turboModuleProxy != null) {
const module: ?T = turboModuleProxy(name);
return module;
}
return null;
<|file_sep|>current/Libraries/TurboModule/TurboModuleRegistry.js
'use strict';
const NativeModules = require('../BatchedBridge/NativeModules');
import type {TurboModule} from './RCTExport';
import invariant from 'invariant';
const turboModuleProxy = global.__turboModuleProxy;
export function get<T: TurboModule>(name: string): ?T {
// Backward compatibility layer during migration.
const legacyModule = NativeModules[name];
if (legacyModule != null) {
return ((legacyModule: any): T);
}
if (turboModuleProxy != null) {
const module: ?T = turboModuleProxy(name);
return module;
}
return null;
<|file_sep|>updated/Libraries/TurboModule/TurboModuleRegistry.js | 'use strict';
const NativeModules = require('../BatchedBridge/NativeModules');
import type {TurboModule} from './RCTExport';
import invariant from 'invariant';
const turboModuleProxy = global.__turboModuleProxy;
export function get<T: TurboModule>(name: string): ?T {
// Bridgeless mode requires TurboModules
if (!global.RN$Bridgeless) {
// Backward compatibility layer during migration.
const legacyModule = NativeModules[name];
if (legacyModule != null) {
return ((legacyModule: any): T);
}
}
if (turboModuleProxy != null) {
const module: ?T = turboModuleProxy(name);
return module; | <|file_sep|>original/Libraries/TurboModule/TurboModuleRegistry.js
'use strict';
const NativeModules = require('../BatchedBridge/NativeModules');
import type {TurboModule} from './RCTExport';
import invariant from 'invariant';
const turboModuleProxy = global.__turboModuleProxy;
export function get<T: TurboModule>(name: string): ?T {
// Backward compatibility layer during migration.
const legacyModule = NativeModules[name];
if (legacyModule != null) {
return ((legacyModule: any): T);
}
if (turboModuleProxy != null) {
const module: ?T = turboModuleProxy(name);
return module;
}
return null;
<|file_sep|>current/Libraries/TurboModule/TurboModuleRegistry.js
'use strict';
const NativeModules = require('../BatchedBridge/NativeModules');
import type {TurboModule} from './RCTExport';
import invariant from 'invariant';
const turboModuleProxy = global.__turboModuleProxy;
export function get<T: TurboModule>(name: string): ?T {
// Backward compatibility layer during migration.
const legacyModule = NativeModules[name];
if (legacyModule != null) {
return ((legacyModule: any): T);
}
if (turboModuleProxy != null) {
const module: ?T = turboModuleProxy(name);
return module;
}
return null;
<|file_sep|>updated/Libraries/TurboModule/TurboModuleRegistry.js
'use strict';
const NativeModules = require('../BatchedBridge/NativeModules');
import type {TurboModule} from './RCTExport';
import invariant from 'invariant';
const turboModuleProxy = global.__turboModuleProxy;
export function get<T: TurboModule>(name: string): ?T {
// Bridgeless mode requires TurboModules
if (!global.RN$Bridgeless) {
// Backward compatibility layer during migration.
const legacyModule = NativeModules[name];
if (legacyModule != null) {
return ((legacyModule: any): T);
}
}
if (turboModuleProxy != null) {
const module: ?T = turboModuleProxy(name);
return module; | c19c6cef3fc64113e8c0795454798a15c5741383 | Libraries/TurboModule/TurboModuleRegistry.js | Libraries/TurboModule/TurboModuleRegistry.js | JavaScript |
<|file_sep|>resources/rubygems/operating_system.rb.diff
original:
have_tools = %w{gcc make sh}.all? do |t|
system("#{t} --version > NUL 2>&1")
end
updated:
begin
load 'devkit'
rescue LoadError
have_tools = %w{gcc make sh}.all? do |t|
system("#{t} --version > NUL 2>&1")
end
<|file_sep|>resources/rubygems/operating_system.rb.diff
original:
unless have_tools
raise Gem::InstallError,<<-EOT
updated:
unless have_tools
raise Gem::InstallError,<<-EOT
<|file_sep|>original/resources/rubygems/operating_system.rb
# :DK-BEG: missing DevKit/build tool convenience notice
Gem.pre_install do |gem_installer|
unless gem_installer.spec.extensions.empty?
have_tools = %w{gcc make sh}.all? do |t|
system("#{t} --version > NUL 2>&1")
end
unless have_tools
raise Gem::InstallError,<<-EOT
The '#{gem_installer.spec.name}' native gem requires installed build tools.
Please update your PATH to include build tools or download the DevKit
from 'http://rubyinstaller.org/downloads' and follow the instructions
at 'http://github.com/oneclick/rubyinstaller/wiki/Development-Kit'
EOT
end
end
end
# :DK-END:
<|file_sep|>current/resources/rubygems/operating_system.rb
Gem.pre_install do |gem_installer|
unless gem_installer.spec.extensions.empty?
begin
load 'devkit'
rescue LoadError
have_tools = %w{gcc make sh}.all? do |t|
system("#{t} --version > NUL 2>&1")
end
unless have_tools
raise Gem::InstallError,<<-EOT
The '#{gem_installer.spec.name}' native gem requires installed build tools.
Please update your PATH to include build tools or download the DevKit
from 'http://rubyinstaller.org/downloads' and follow the instructions
at 'http://github.com/oneclick/rubyinstaller/wiki/Development-Kit'
EOT
end
end
end
# :DK-END:
<|file_sep|>updated/resources/rubygems/operating_system.rb | unless gem_installer.spec.extensions.empty?
begin
load 'devkit'
rescue LoadError
have_tools = %w{gcc make sh}.all? do |t|
system("#{t} --version > NUL 2>&1")
end
unless have_tools
raise Gem::InstallError,<<-EOT
The '#{gem_installer.spec.name}' native gem requires installed build tools.
Please update your PATH to include build tools or download the DevKit
from 'http://rubyinstaller.org/downloads' and follow the instructions
at 'http://github.com/oneclick/rubyinstaller/wiki/Development-Kit'
EOT
end
end
end
end
# :DK-END: | <|file_sep|>resources/rubygems/operating_system.rb.diff
original:
have_tools = %w{gcc make sh}.all? do |t|
system("#{t} --version > NUL 2>&1")
end
updated:
begin
load 'devkit'
rescue LoadError
have_tools = %w{gcc make sh}.all? do |t|
system("#{t} --version > NUL 2>&1")
end
<|file_sep|>resources/rubygems/operating_system.rb.diff
original:
unless have_tools
raise Gem::InstallError,<<-EOT
updated:
unless have_tools
raise Gem::InstallError,<<-EOT
<|file_sep|>original/resources/rubygems/operating_system.rb
# :DK-BEG: missing DevKit/build tool convenience notice
Gem.pre_install do |gem_installer|
unless gem_installer.spec.extensions.empty?
have_tools = %w{gcc make sh}.all? do |t|
system("#{t} --version > NUL 2>&1")
end
unless have_tools
raise Gem::InstallError,<<-EOT
The '#{gem_installer.spec.name}' native gem requires installed build tools.
Please update your PATH to include build tools or download the DevKit
from 'http://rubyinstaller.org/downloads' and follow the instructions
at 'http://github.com/oneclick/rubyinstaller/wiki/Development-Kit'
EOT
end
end
end
# :DK-END:
<|file_sep|>current/resources/rubygems/operating_system.rb
Gem.pre_install do |gem_installer|
unless gem_installer.spec.extensions.empty?
begin
load 'devkit'
rescue LoadError
have_tools = %w{gcc make sh}.all? do |t|
system("#{t} --version > NUL 2>&1")
end
unless have_tools
raise Gem::InstallError,<<-EOT
The '#{gem_installer.spec.name}' native gem requires installed build tools.
Please update your PATH to include build tools or download the DevKit
from 'http://rubyinstaller.org/downloads' and follow the instructions
at 'http://github.com/oneclick/rubyinstaller/wiki/Development-Kit'
EOT
end
end
end
# :DK-END:
<|file_sep|>updated/resources/rubygems/operating_system.rb
unless gem_installer.spec.extensions.empty?
begin
load 'devkit'
rescue LoadError
have_tools = %w{gcc make sh}.all? do |t|
system("#{t} --version > NUL 2>&1")
end
unless have_tools
raise Gem::InstallError,<<-EOT
The '#{gem_installer.spec.name}' native gem requires installed build tools.
Please update your PATH to include build tools or download the DevKit
from 'http://rubyinstaller.org/downloads' and follow the instructions
at 'http://github.com/oneclick/rubyinstaller/wiki/Development-Kit'
EOT
end
end
end
end
# :DK-END: | 688b690bdbd65dd875e08a83e4ae743948bc4a42 | resources/rubygems/operating_system.rb | resources/rubygems/operating_system.rb | Ruby |
<|file_sep|>test/phantomBase.js.diff
original:
updated:
var async = require('async');
<|file_sep|>test/phantomBase.js.diff
original:
updated:
<|file_sep|>test/phantomBase.js.diff
original:
describe('Base PhantomJS functions', function () {
this.timeout(30000);
updated:
describe('Testing base PhantomJS functions', function () {
this.timeout(50000);
<|file_sep|>original/test/phantomBase.js
/**
* Created by JiaHao on 18/8/15.
*/
var phantomBase = require('./../lib/phantomBase');
const testUrls = ['http://techcrunch.com/', 'https://www.facebook.com/'];
describe('Base PhantomJS functions', function () {
this.timeout(30000);
it('Can open a page', function (done) {
phantomBase.openPage(testUrls[1], function (error, page, ph) {
ph.exit();
done(error);
})
})
});
<|file_sep|>current/test/phantomBase.js
*/
var async = require('async');
var phantomBase = require('./../lib/phantomBase');
const testUrls = ['http://techcrunch.com/', 'https://www.facebook.com/'];
describe('Testing base PhantomJS functions', function () {
this.timeout(50000);
it('Can open a page', function (done) {
phantomBase.openPage(testUrls[1], function (error, page, ph) {
ph.exit();
done(error);
})
})
});
<|file_sep|>updated/test/phantomBase.js | var phantomBase = require('./../lib/phantomBase');
const testUrls = ['http://techcrunch.com/', 'https://www.facebook.com/'];
describe('Testing base PhantomJS functions', function () {
this.timeout(50000);
it('Can open a page', function (done) {
async.each(testUrls, function (testUrl, callback) {
phantomBase.openPage(testUrl, function (error, page, ph) {
ph.exit();
callback(error);
});
}, function (error) {
done(error);
});
})
| <|file_sep|>test/phantomBase.js.diff
original:
updated:
var async = require('async');
<|file_sep|>test/phantomBase.js.diff
original:
updated:
<|file_sep|>test/phantomBase.js.diff
original:
describe('Base PhantomJS functions', function () {
this.timeout(30000);
updated:
describe('Testing base PhantomJS functions', function () {
this.timeout(50000);
<|file_sep|>original/test/phantomBase.js
/**
* Created by JiaHao on 18/8/15.
*/
var phantomBase = require('./../lib/phantomBase');
const testUrls = ['http://techcrunch.com/', 'https://www.facebook.com/'];
describe('Base PhantomJS functions', function () {
this.timeout(30000);
it('Can open a page', function (done) {
phantomBase.openPage(testUrls[1], function (error, page, ph) {
ph.exit();
done(error);
})
})
});
<|file_sep|>current/test/phantomBase.js
*/
var async = require('async');
var phantomBase = require('./../lib/phantomBase');
const testUrls = ['http://techcrunch.com/', 'https://www.facebook.com/'];
describe('Testing base PhantomJS functions', function () {
this.timeout(50000);
it('Can open a page', function (done) {
phantomBase.openPage(testUrls[1], function (error, page, ph) {
ph.exit();
done(error);
})
})
});
<|file_sep|>updated/test/phantomBase.js
var phantomBase = require('./../lib/phantomBase');
const testUrls = ['http://techcrunch.com/', 'https://www.facebook.com/'];
describe('Testing base PhantomJS functions', function () {
this.timeout(50000);
it('Can open a page', function (done) {
async.each(testUrls, function (testUrl, callback) {
phantomBase.openPage(testUrl, function (error, page, ph) {
ph.exit();
callback(error);
});
}, function (error) {
done(error);
});
})
| ccbfcbf7f3e304b3ba6c84152ce91daa48613ea6 | test/phantomBase.js | test/phantomBase.js | JavaScript |
<|file_sep|>tools/make/lib/ls/pkgs/namespaces.mk.diff
original:
updated:
# Define the directory from which to search for namespaces:
LIST_PACKAGE_NAMESPACES_DIR ?= $(SRC_DIR)
<|file_sep|>tools/make/lib/ls/pkgs/namespaces.mk.diff
original:
# TARGETS #
updated:
<|file_sep|>tools/make/lib/ls/pkgs/namespaces.mk.diff
original:
# List all package namespaces.
updated:
# RULES #
#/
# Prints a list of all package namespaces.
<|file_sep|>tools/make/lib/ls/pkgs/namespaces.mk.diff
original:
# This target prints a list of all package namespaces.
updated:
# @example
# make list-pkgs-namespaces
#
# @example
# make list-pkgs-namespaces LIST_PACKAGE_NAMESPACES_DIR=$PWD/lib/node_modules/\@stdlib/utils
#/
<|file_sep|>original/tools/make/lib/ls/pkgs/namespaces.mk
#/
# VARIABLES #
# Define the path of the executable:
LIST_PACKAGE_NAMESPACES ?= $(TOOLS_PKGS_DIR)/pkgs/namespaces/bin/cli
# Define the command flags:
LIST_PACKAGE_NAMESPACES_FLAGS ?=
# TARGETS #
# List all package namespaces.
#
# This target prints a list of all package namespaces.
list-pkgs-namespaces: $(LIST_PACKAGE_NAMESPACES) $(NODE_MODULES)
$(QUIET) NODE_PATH="$(NODE_PATH)" $(NODE) $(LIST_PACKAGE_NAMESPACES) $(LIST_PACKAGE_NAMESPACES_FLAGS) $(SRC_DIR)
.PHONY: list-pkgs-namespaces
<|file_sep|>current/tools/make/lib/ls/pkgs/namespaces.mk
LIST_PACKAGE_NAMESPACES_FLAGS ?=
# Define the directory from which to search for namespaces:
LIST_PACKAGE_NAMESPACES_DIR ?= $(SRC_DIR)
# RULES #
#/
# Prints a list of all package namespaces.
#
# @example
# make list-pkgs-namespaces
#
# @example
# make list-pkgs-namespaces LIST_PACKAGE_NAMESPACES_DIR=$PWD/lib/node_modules/\@stdlib/utils
#/
list-pkgs-namespaces: $(LIST_PACKAGE_NAMESPACES) $(NODE_MODULES)
$(QUIET) NODE_PATH="$(NODE_PATH)" $(NODE) $(LIST_PACKAGE_NAMESPACES) $(LIST_PACKAGE_NAMESPACES_FLAGS) $(SRC_DIR)
.PHONY: list-pkgs-namespaces
<|file_sep|>updated/tools/make/lib/ls/pkgs/namespaces.mk | LIST_PACKAGE_NAMESPACES_FLAGS ?=
# Define the directory from which to search for namespaces:
LIST_PACKAGE_NAMESPACES_DIR ?= $(SRC_DIR)
# RULES #
#/
# Prints a list of all package namespaces.
#
# @example
# make list-pkgs-namespaces
#
# @example
# make list-pkgs-namespaces LIST_PACKAGE_NAMESPACES_DIR=$PWD/lib/node_modules/\@stdlib/utils
#/
list-pkgs-namespaces: $(LIST_PACKAGE_NAMESPACES) $(NODE_MODULES)
$(QUIET) NODE_PATH="$(NODE_PATH)" $(NODE) $(LIST_PACKAGE_NAMESPACES) $(LIST_PACKAGE_NAMESPACES_FLAGS) $(LIST_PACKAGE_NAMESPACES_DIR)
.PHONY: list-pkgs-namespaces | <|file_sep|>tools/make/lib/ls/pkgs/namespaces.mk.diff
original:
updated:
# Define the directory from which to search for namespaces:
LIST_PACKAGE_NAMESPACES_DIR ?= $(SRC_DIR)
<|file_sep|>tools/make/lib/ls/pkgs/namespaces.mk.diff
original:
# TARGETS #
updated:
<|file_sep|>tools/make/lib/ls/pkgs/namespaces.mk.diff
original:
# List all package namespaces.
updated:
# RULES #
#/
# Prints a list of all package namespaces.
<|file_sep|>tools/make/lib/ls/pkgs/namespaces.mk.diff
original:
# This target prints a list of all package namespaces.
updated:
# @example
# make list-pkgs-namespaces
#
# @example
# make list-pkgs-namespaces LIST_PACKAGE_NAMESPACES_DIR=$PWD/lib/node_modules/\@stdlib/utils
#/
<|file_sep|>original/tools/make/lib/ls/pkgs/namespaces.mk
#/
# VARIABLES #
# Define the path of the executable:
LIST_PACKAGE_NAMESPACES ?= $(TOOLS_PKGS_DIR)/pkgs/namespaces/bin/cli
# Define the command flags:
LIST_PACKAGE_NAMESPACES_FLAGS ?=
# TARGETS #
# List all package namespaces.
#
# This target prints a list of all package namespaces.
list-pkgs-namespaces: $(LIST_PACKAGE_NAMESPACES) $(NODE_MODULES)
$(QUIET) NODE_PATH="$(NODE_PATH)" $(NODE) $(LIST_PACKAGE_NAMESPACES) $(LIST_PACKAGE_NAMESPACES_FLAGS) $(SRC_DIR)
.PHONY: list-pkgs-namespaces
<|file_sep|>current/tools/make/lib/ls/pkgs/namespaces.mk
LIST_PACKAGE_NAMESPACES_FLAGS ?=
# Define the directory from which to search for namespaces:
LIST_PACKAGE_NAMESPACES_DIR ?= $(SRC_DIR)
# RULES #
#/
# Prints a list of all package namespaces.
#
# @example
# make list-pkgs-namespaces
#
# @example
# make list-pkgs-namespaces LIST_PACKAGE_NAMESPACES_DIR=$PWD/lib/node_modules/\@stdlib/utils
#/
list-pkgs-namespaces: $(LIST_PACKAGE_NAMESPACES) $(NODE_MODULES)
$(QUIET) NODE_PATH="$(NODE_PATH)" $(NODE) $(LIST_PACKAGE_NAMESPACES) $(LIST_PACKAGE_NAMESPACES_FLAGS) $(SRC_DIR)
.PHONY: list-pkgs-namespaces
<|file_sep|>updated/tools/make/lib/ls/pkgs/namespaces.mk
LIST_PACKAGE_NAMESPACES_FLAGS ?=
# Define the directory from which to search for namespaces:
LIST_PACKAGE_NAMESPACES_DIR ?= $(SRC_DIR)
# RULES #
#/
# Prints a list of all package namespaces.
#
# @example
# make list-pkgs-namespaces
#
# @example
# make list-pkgs-namespaces LIST_PACKAGE_NAMESPACES_DIR=$PWD/lib/node_modules/\@stdlib/utils
#/
list-pkgs-namespaces: $(LIST_PACKAGE_NAMESPACES) $(NODE_MODULES)
$(QUIET) NODE_PATH="$(NODE_PATH)" $(NODE) $(LIST_PACKAGE_NAMESPACES) $(LIST_PACKAGE_NAMESPACES_FLAGS) $(LIST_PACKAGE_NAMESPACES_DIR)
.PHONY: list-pkgs-namespaces | 92948b3c5fe1c438906c19126c48caa8b469d435 | tools/make/lib/ls/pkgs/namespaces.mk | tools/make/lib/ls/pkgs/namespaces.mk | Makefile |
<|file_sep|>original/test/integration/users_login_test.rb
<|file_sep|>current/test/integration/users_login_test.rb
<|file_sep|>updated/test/integration/users_login_test.rb | require 'test_helper'
class UsersLoginTest < ActionDispatch::IntegrationTest
test 'login with invalid information' do
get login_path
assert_template 'sessions/new'
post login_path, session: {email: '', password: ''}
assert_template 'sessions/new'
assert_not flash.empty?
# ensure flash doesn't persist for an extra pageload
get login_path
assert flash.empty?
end
end | <|file_sep|>original/test/integration/users_login_test.rb
<|file_sep|>current/test/integration/users_login_test.rb
<|file_sep|>updated/test/integration/users_login_test.rb
require 'test_helper'
class UsersLoginTest < ActionDispatch::IntegrationTest
test 'login with invalid information' do
get login_path
assert_template 'sessions/new'
post login_path, session: {email: '', password: ''}
assert_template 'sessions/new'
assert_not flash.empty?
# ensure flash doesn't persist for an extra pageload
get login_path
assert flash.empty?
end
end | 0c8300b466d32c25dd7c063bd966c59733949bbd | test/integration/users_login_test.rb | test/integration/users_login_test.rb | Ruby |
<|file_sep|>original/metadata/itkach.aard2.txt
commit=0.26
init=rm libs/android-support-v13.jar
gradle=yes
srclibs=slobber@0.9,slobj@0.7
prebuild=cp -fR $$slobber$$ ../ && \
cp -fR $$slobj$$ ../ && \
sed -i -e '/support-v13/acompile "com.android.support:support-v13:22.2.0"' -e 's|compile files|//compile files|g' build.gradle
target=android-19
Maintainer Notes:
* slobj has a jar file
* throws error (java8) on slobj
* settings.gradle uses includeflat, which points to ../projectname (see cp -fR).
.
Auto Update Mode:None
#Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.26
Current Version Code:26
<|file_sep|>current/metadata/itkach.aard2.txt
commit=0.26
init=rm libs/android-support-v13.jar
gradle=yes
srclibs=slobber@0.9,slobj@0.7
prebuild=cp -fR $$slobber$$ ../ && \
cp -fR $$slobj$$ ../ && \
sed -i -e '/support-v13/acompile "com.android.support:support-v13:22.2.0"' -e 's|compile files|//compile files|g' build.gradle
target=android-19
Maintainer Notes:
* slobj has a jar file
* throws error (java8) on slobj
* settings.gradle uses includeflat, which points to ../projectname (see cp -fR).
.
Auto Update Mode:None
#Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.26
Current Version Code:26
<|file_sep|>updated/metadata/itkach.aard2.txt | commit=0.26
init=rm libs/android-support-v13.jar
gradle=yes
srclibs=slobber@0.9,slobj@0.7
prebuild=cp -fR $$slobber$$ ../ && \
cp -fR $$slobj$$ ../ && \
sed -i -e '/support-v13/acompile "com.android.support:support-v13:22.2.0"' -e 's|compile files|//compile files|g' build.gradle
target=android-19
Maintainer Notes:
* slobj has a jar file
* throws error (java8) on slobj
* settings.gradle uses includeflat, which points to ../projectname (see cp -fR).
.
Auto Update Mode:None
#Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.27
Current Version Code:27
| <|file_sep|>original/metadata/itkach.aard2.txt
commit=0.26
init=rm libs/android-support-v13.jar
gradle=yes
srclibs=slobber@0.9,slobj@0.7
prebuild=cp -fR $$slobber$$ ../ && \
cp -fR $$slobj$$ ../ && \
sed -i -e '/support-v13/acompile "com.android.support:support-v13:22.2.0"' -e 's|compile files|//compile files|g' build.gradle
target=android-19
Maintainer Notes:
* slobj has a jar file
* throws error (java8) on slobj
* settings.gradle uses includeflat, which points to ../projectname (see cp -fR).
.
Auto Update Mode:None
#Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.26
Current Version Code:26
<|file_sep|>current/metadata/itkach.aard2.txt
commit=0.26
init=rm libs/android-support-v13.jar
gradle=yes
srclibs=slobber@0.9,slobj@0.7
prebuild=cp -fR $$slobber$$ ../ && \
cp -fR $$slobj$$ ../ && \
sed -i -e '/support-v13/acompile "com.android.support:support-v13:22.2.0"' -e 's|compile files|//compile files|g' build.gradle
target=android-19
Maintainer Notes:
* slobj has a jar file
* throws error (java8) on slobj
* settings.gradle uses includeflat, which points to ../projectname (see cp -fR).
.
Auto Update Mode:None
#Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.26
Current Version Code:26
<|file_sep|>updated/metadata/itkach.aard2.txt
commit=0.26
init=rm libs/android-support-v13.jar
gradle=yes
srclibs=slobber@0.9,slobj@0.7
prebuild=cp -fR $$slobber$$ ../ && \
cp -fR $$slobj$$ ../ && \
sed -i -e '/support-v13/acompile "com.android.support:support-v13:22.2.0"' -e 's|compile files|//compile files|g' build.gradle
target=android-19
Maintainer Notes:
* slobj has a jar file
* throws error (java8) on slobj
* settings.gradle uses includeflat, which points to ../projectname (see cp -fR).
.
Auto Update Mode:None
#Auto Update Mode:Version %v
Update Check Mode:Tags
Current Version:0.27
Current Version Code:27
| 21d3bd659a5d68709f13dbc29a37506b0ebfd88f | metadata/itkach.aard2.txt | metadata/itkach.aard2.txt | Text |
<|file_sep|>original/weatherdownloaderlibrary/src/main/AndroidManifest.xml
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="studios.codelight.weatherdownloaderlibrary">
<uses-permission android:name="android.permission.INTERNET"/>
<application
android:allowBackup="false"
android:label="@string/app_name"
android:supportsRtl="true">
</application>
</manifest>
<|file_sep|>current/weatherdownloaderlibrary/src/main/AndroidManifest.xml
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="studios.codelight.weatherdownloaderlibrary">
<uses-permission android:name="android.permission.INTERNET"/>
<application
android:allowBackup="false"
android:label="@string/app_name"
android:supportsRtl="true">
</application>
</manifest>
<|file_sep|>updated/weatherdownloaderlibrary/src/main/AndroidManifest.xml | <manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="studios.codelight.weatherdownloaderlibrary">
<uses-permission android:name="android.permission.INTERNET"/>
<application
android:label="@string/app_name">
</application>
</manifest> | <|file_sep|>original/weatherdownloaderlibrary/src/main/AndroidManifest.xml
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="studios.codelight.weatherdownloaderlibrary">
<uses-permission android:name="android.permission.INTERNET"/>
<application
android:allowBackup="false"
android:label="@string/app_name"
android:supportsRtl="true">
</application>
</manifest>
<|file_sep|>current/weatherdownloaderlibrary/src/main/AndroidManifest.xml
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="studios.codelight.weatherdownloaderlibrary">
<uses-permission android:name="android.permission.INTERNET"/>
<application
android:allowBackup="false"
android:label="@string/app_name"
android:supportsRtl="true">
</application>
</manifest>
<|file_sep|>updated/weatherdownloaderlibrary/src/main/AndroidManifest.xml
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
package="studios.codelight.weatherdownloaderlibrary">
<uses-permission android:name="android.permission.INTERNET"/>
<application
android:label="@string/app_name">
</application>
</manifest> | ecf9fd1ea72b2af00d3d6023b9cecb65a39bce6b | weatherdownloaderlibrary/src/main/AndroidManifest.xml | weatherdownloaderlibrary/src/main/AndroidManifest.xml | XML |
<|file_sep|>original/front_end/models/issues_manager/descriptions/CoepCorpNotSameSite.md
# Specify a more permissive Cross-Origin Resource Policy to prevent a resource from being blocked
Your site tries to access an external resource that only allows same-site usage.
This behavior prevents a document from loading any non-same-site resources which don’t explicitly grant permission to be loaded.
To solve this, add the following to the resource’s HTML response header: `Cross-Origin-Embedder-Policy: cross-origin`
⚠️If you set this header, any website can embed this resource.
<|file_sep|>current/front_end/models/issues_manager/descriptions/CoepCorpNotSameSite.md
# Specify a more permissive Cross-Origin Resource Policy to prevent a resource from being blocked
Your site tries to access an external resource that only allows same-site usage.
This behavior prevents a document from loading any non-same-site resources which don’t explicitly grant permission to be loaded.
To solve this, add the following to the resource’s HTML response header: `Cross-Origin-Embedder-Policy: cross-origin`
⚠️If you set this header, any website can embed this resource.
<|file_sep|>updated/front_end/models/issues_manager/descriptions/CoepCorpNotSameSite.md | # Specify a more permissive Cross-Origin Resource Policy to prevent a resource from being blocked
Your site tries to access an external resource that only allows same-site usage.
This behavior prevents a document from loading any non-same-site resources which don’t explicitly grant permission to be loaded.
To solve this, add the following to the resource’s HTML response header: `Cross-Origin-Resource-Policy: cross-origin`
⚠️If you set this header, any website can embed this resource. | <|file_sep|>original/front_end/models/issues_manager/descriptions/CoepCorpNotSameSite.md
# Specify a more permissive Cross-Origin Resource Policy to prevent a resource from being blocked
Your site tries to access an external resource that only allows same-site usage.
This behavior prevents a document from loading any non-same-site resources which don’t explicitly grant permission to be loaded.
To solve this, add the following to the resource’s HTML response header: `Cross-Origin-Embedder-Policy: cross-origin`
⚠️If you set this header, any website can embed this resource.
<|file_sep|>current/front_end/models/issues_manager/descriptions/CoepCorpNotSameSite.md
# Specify a more permissive Cross-Origin Resource Policy to prevent a resource from being blocked
Your site tries to access an external resource that only allows same-site usage.
This behavior prevents a document from loading any non-same-site resources which don’t explicitly grant permission to be loaded.
To solve this, add the following to the resource’s HTML response header: `Cross-Origin-Embedder-Policy: cross-origin`
⚠️If you set this header, any website can embed this resource.
<|file_sep|>updated/front_end/models/issues_manager/descriptions/CoepCorpNotSameSite.md
# Specify a more permissive Cross-Origin Resource Policy to prevent a resource from being blocked
Your site tries to access an external resource that only allows same-site usage.
This behavior prevents a document from loading any non-same-site resources which don’t explicitly grant permission to be loaded.
To solve this, add the following to the resource’s HTML response header: `Cross-Origin-Resource-Policy: cross-origin`
⚠️If you set this header, any website can embed this resource. | c13c29dc98e39bc5c96f6b58be4856b8eda4fee6 | front_end/models/issues_manager/descriptions/CoepCorpNotSameSite.md | front_end/models/issues_manager/descriptions/CoepCorpNotSameSite.md | Markdown |
<|file_sep|>original/templates/pam_sudo_deb.erb
#%PAM-1.0
# This file is managed by Puppet
auth required pam_env.so readenv=1 user_readenv=0
auth required pam_env.so readenv=1 envfile=/etc/default/locale user_readenv=0
<%- if @pam_radius_enforce_real == 'strict'-%>
# strict - don't allow failback to local password
auth [success=1 default=ignore] pam_listfile.so onerr=success item=user sense=allow file=<%=@pam_radius_users_file%>
auth [success=2 default=ignore] pam_succeed_if.so user notingroup <%=@pam_radius_admins_group%> quiet
auth [success=done new_authtok_reqd=done ignore=ignore default=die] pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok
<% end -%>
<%- if @pam_radius_enforce_real == 'permissive'-%>
# permissive - allow failback to local password
auth sufficient pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok use_first_pass
<% end -%>
@include common-auth
@include common-account
@include common-session-noninteractive
<|file_sep|>current/templates/pam_sudo_deb.erb
#%PAM-1.0
# This file is managed by Puppet
auth required pam_env.so readenv=1 user_readenv=0
auth required pam_env.so readenv=1 envfile=/etc/default/locale user_readenv=0
<%- if @pam_radius_enforce_real == 'strict'-%>
# strict - don't allow failback to local password
auth [success=1 default=ignore] pam_listfile.so onerr=success item=user sense=allow file=<%=@pam_radius_users_file%>
auth [success=2 default=ignore] pam_succeed_if.so user notingroup <%=@pam_radius_admins_group%> quiet
auth [success=done new_authtok_reqd=done ignore=ignore default=die] pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok
<% end -%>
<%- if @pam_radius_enforce_real == 'permissive'-%>
# permissive - allow failback to local password
auth sufficient pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok use_first_pass
<% end -%>
@include common-auth
@include common-account
@include common-session-noninteractive
<|file_sep|>updated/templates/pam_sudo_deb.erb | #%PAM-1.0
# This file is managed by Puppet
auth required pam_env.so readenv=1 user_readenv=0
auth required pam_env.so readenv=1 envfile=/etc/default/locale user_readenv=0
<%- if @pam_radius_enforce_real == 'strict'-%>
# strict - don't allow failback to local password
auth [success=1 default=ignore] pam_listfile.so onerr=success item=user sense=allow file=/etc/<%=@pam_radius_users_file%>
auth [success=2 default=ignore] pam_succeed_if.so user notingroup <%=@pam_radius_admins_group%> quiet
auth [success=done new_authtok_reqd=done ignore=ignore default=die] pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok
<% end -%>
<%- if @pam_radius_enforce_real == 'permissive'-%>
# permissive - allow failback to local password
auth sufficient pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok use_first_pass
<% end -%>
@include common-auth
@include common-account
@include common-session-noninteractive | <|file_sep|>original/templates/pam_sudo_deb.erb
#%PAM-1.0
# This file is managed by Puppet
auth required pam_env.so readenv=1 user_readenv=0
auth required pam_env.so readenv=1 envfile=/etc/default/locale user_readenv=0
<%- if @pam_radius_enforce_real == 'strict'-%>
# strict - don't allow failback to local password
auth [success=1 default=ignore] pam_listfile.so onerr=success item=user sense=allow file=<%=@pam_radius_users_file%>
auth [success=2 default=ignore] pam_succeed_if.so user notingroup <%=@pam_radius_admins_group%> quiet
auth [success=done new_authtok_reqd=done ignore=ignore default=die] pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok
<% end -%>
<%- if @pam_radius_enforce_real == 'permissive'-%>
# permissive - allow failback to local password
auth sufficient pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok use_first_pass
<% end -%>
@include common-auth
@include common-account
@include common-session-noninteractive
<|file_sep|>current/templates/pam_sudo_deb.erb
#%PAM-1.0
# This file is managed by Puppet
auth required pam_env.so readenv=1 user_readenv=0
auth required pam_env.so readenv=1 envfile=/etc/default/locale user_readenv=0
<%- if @pam_radius_enforce_real == 'strict'-%>
# strict - don't allow failback to local password
auth [success=1 default=ignore] pam_listfile.so onerr=success item=user sense=allow file=<%=@pam_radius_users_file%>
auth [success=2 default=ignore] pam_succeed_if.so user notingroup <%=@pam_radius_admins_group%> quiet
auth [success=done new_authtok_reqd=done ignore=ignore default=die] pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok
<% end -%>
<%- if @pam_radius_enforce_real == 'permissive'-%>
# permissive - allow failback to local password
auth sufficient pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok use_first_pass
<% end -%>
@include common-auth
@include common-account
@include common-session-noninteractive
<|file_sep|>updated/templates/pam_sudo_deb.erb
#%PAM-1.0
# This file is managed by Puppet
auth required pam_env.so readenv=1 user_readenv=0
auth required pam_env.so readenv=1 envfile=/etc/default/locale user_readenv=0
<%- if @pam_radius_enforce_real == 'strict'-%>
# strict - don't allow failback to local password
auth [success=1 default=ignore] pam_listfile.so onerr=success item=user sense=allow file=/etc/<%=@pam_radius_users_file%>
auth [success=2 default=ignore] pam_succeed_if.so user notingroup <%=@pam_radius_admins_group%> quiet
auth [success=done new_authtok_reqd=done ignore=ignore default=die] pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok
<% end -%>
<%- if @pam_radius_enforce_real == 'permissive'-%>
# permissive - allow failback to local password
auth sufficient pam_radius_auth.so localifdown
auth sufficient pam_unix.so nullok use_first_pass
<% end -%>
@include common-auth
@include common-account
@include common-session-noninteractive | dc33efd1f17213ed3d18d5d1965ee2a1496133b7 | templates/pam_sudo_deb.erb | templates/pam_sudo_deb.erb | HTML+ERB |
<|file_sep|>original/idea.contracts/com/google/common/base/annotations.xml
~
~ Licensed under the Apache License, Version 2.0 (the "License");
~ you may not use this file except in compliance with the License.
~ You may obtain a copy of the License at
~
~ http://www.apache.org/licenses/LICENSE-2.0
~
~ Unless required by applicable law or agreed to in writing, software
~ distributed under the License is distributed on an "AS IS" BASIS,
~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
~ See the License for the specific language governing permissions and
~ limitations under the License.
-->
<root>
<item name='com.google.common.base.Preconditions void checkState(boolean, java.lang.String, java.lang.Object...)'>
<annotation name='org.jetbrains.annotations.Contract'>
<val val=""false, _, _ -> fail""/>
</annotation>
</item>
</root>
<|file_sep|>current/idea.contracts/com/google/common/base/annotations.xml
~
~ Licensed under the Apache License, Version 2.0 (the "License");
~ you may not use this file except in compliance with the License.
~ You may obtain a copy of the License at
~
~ http://www.apache.org/licenses/LICENSE-2.0
~
~ Unless required by applicable law or agreed to in writing, software
~ distributed under the License is distributed on an "AS IS" BASIS,
~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
~ See the License for the specific language governing permissions and
~ limitations under the License.
-->
<root>
<item name='com.google.common.base.Preconditions void checkState(boolean, java.lang.String, java.lang.Object...)'>
<annotation name='org.jetbrains.annotations.Contract'>
<val val=""false, _, _ -> fail""/>
</annotation>
</item>
</root>
<|file_sep|>updated/idea.contracts/com/google/common/base/annotations.xml | ~ http://www.apache.org/licenses/LICENSE-2.0
~
~ Unless required by applicable law or agreed to in writing, software
~ distributed under the License is distributed on an "AS IS" BASIS,
~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
~ See the License for the specific language governing permissions and
~ limitations under the License.
-->
<root>
<item name='com.google.common.base.Preconditions void checkArgument(boolean, java.lang.String, java.lang.Object...)'>
<annotation name='org.jetbrains.annotations.Contract'>
<val val=""false, _, _ -> fail""/>
</annotation>
</item>
<item name='com.google.common.base.Preconditions void checkState(boolean, java.lang.String, java.lang.Object...)'>
<annotation name='org.jetbrains.annotations.Contract'>
<val val=""false, _, _ -> fail""/>
</annotation>
</item>
</root> | <|file_sep|>original/idea.contracts/com/google/common/base/annotations.xml
~
~ Licensed under the Apache License, Version 2.0 (the "License");
~ you may not use this file except in compliance with the License.
~ You may obtain a copy of the License at
~
~ http://www.apache.org/licenses/LICENSE-2.0
~
~ Unless required by applicable law or agreed to in writing, software
~ distributed under the License is distributed on an "AS IS" BASIS,
~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
~ See the License for the specific language governing permissions and
~ limitations under the License.
-->
<root>
<item name='com.google.common.base.Preconditions void checkState(boolean, java.lang.String, java.lang.Object...)'>
<annotation name='org.jetbrains.annotations.Contract'>
<val val=""false, _, _ -> fail""/>
</annotation>
</item>
</root>
<|file_sep|>current/idea.contracts/com/google/common/base/annotations.xml
~
~ Licensed under the Apache License, Version 2.0 (the "License");
~ you may not use this file except in compliance with the License.
~ You may obtain a copy of the License at
~
~ http://www.apache.org/licenses/LICENSE-2.0
~
~ Unless required by applicable law or agreed to in writing, software
~ distributed under the License is distributed on an "AS IS" BASIS,
~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
~ See the License for the specific language governing permissions and
~ limitations under the License.
-->
<root>
<item name='com.google.common.base.Preconditions void checkState(boolean, java.lang.String, java.lang.Object...)'>
<annotation name='org.jetbrains.annotations.Contract'>
<val val=""false, _, _ -> fail""/>
</annotation>
</item>
</root>
<|file_sep|>updated/idea.contracts/com/google/common/base/annotations.xml
~ http://www.apache.org/licenses/LICENSE-2.0
~
~ Unless required by applicable law or agreed to in writing, software
~ distributed under the License is distributed on an "AS IS" BASIS,
~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
~ See the License for the specific language governing permissions and
~ limitations under the License.
-->
<root>
<item name='com.google.common.base.Preconditions void checkArgument(boolean, java.lang.String, java.lang.Object...)'>
<annotation name='org.jetbrains.annotations.Contract'>
<val val=""false, _, _ -> fail""/>
</annotation>
</item>
<item name='com.google.common.base.Preconditions void checkState(boolean, java.lang.String, java.lang.Object...)'>
<annotation name='org.jetbrains.annotations.Contract'>
<val val=""false, _, _ -> fail""/>
</annotation>
</item>
</root> | c30bc8d0a3939eaf8c16169fa9c6d1ee27efbdf2 | idea.contracts/com/google/common/base/annotations.xml | idea.contracts/com/google/common/base/annotations.xml | XML |
<|file_sep|>original/src/control/Meta.js
import Observable from '../util/Observable';
export default class MetaControl extends Observable {
constructor() {
super();
this._registry = [];
this._controlEvents = ['engage', 'disengage'];
}
_createEventProxy(event) {
return (...args) => this.fireEvent.call(this, event, ...args);
}
addControl(control) {
this._registry.push(control);
const listeners = {};
this._controlEvents.forEach(
event => listeners[event] = this._createEventProxy(event)
);
<|file_sep|>current/src/control/Meta.js
import Observable from '../util/Observable';
export default class MetaControl extends Observable {
constructor() {
super();
this._registry = [];
this._controlEvents = ['engage', 'disengage'];
}
_createEventProxy(event) {
return (...args) => this.fireEvent.call(this, event, ...args);
}
addControl(control) {
this._registry.push(control);
const listeners = {};
this._controlEvents.forEach(
event => listeners[event] = this._createEventProxy(event)
);
<|file_sep|>updated/src/control/Meta.js | import Observable from '../util/Observable';
import * as _ from 'lodash';
export default class MetaControl extends Observable {
constructor() {
super();
this._registry = [];
this._controlEvents = ['engage', 'disengage'];
}
_createEventProxy(event) {
return (...args) => this.fireEvent.call(this, event, ...args);
}
addControl(control) {
this._registry.push(control);
const listeners = {};
this._controlEvents.forEach(
event => listeners[event] = this._createEventProxy(event) | <|file_sep|>original/src/control/Meta.js
import Observable from '../util/Observable';
export default class MetaControl extends Observable {
constructor() {
super();
this._registry = [];
this._controlEvents = ['engage', 'disengage'];
}
_createEventProxy(event) {
return (...args) => this.fireEvent.call(this, event, ...args);
}
addControl(control) {
this._registry.push(control);
const listeners = {};
this._controlEvents.forEach(
event => listeners[event] = this._createEventProxy(event)
);
<|file_sep|>current/src/control/Meta.js
import Observable from '../util/Observable';
export default class MetaControl extends Observable {
constructor() {
super();
this._registry = [];
this._controlEvents = ['engage', 'disengage'];
}
_createEventProxy(event) {
return (...args) => this.fireEvent.call(this, event, ...args);
}
addControl(control) {
this._registry.push(control);
const listeners = {};
this._controlEvents.forEach(
event => listeners[event] = this._createEventProxy(event)
);
<|file_sep|>updated/src/control/Meta.js
import Observable from '../util/Observable';
import * as _ from 'lodash';
export default class MetaControl extends Observable {
constructor() {
super();
this._registry = [];
this._controlEvents = ['engage', 'disengage'];
}
_createEventProxy(event) {
return (...args) => this.fireEvent.call(this, event, ...args);
}
addControl(control) {
this._registry.push(control);
const listeners = {};
this._controlEvents.forEach(
event => listeners[event] = this._createEventProxy(event) | d426a54b238d0dae6143c076191bf80d4d9b70ee | src/control/Meta.js | src/control/Meta.js | JavaScript |
<|file_sep|>original/p3/templates/p3/my_schedule.html
{% extends "p3/schedule_list.html" %}
{% block CONTENT_TITLE %}My Europython schedule <a style="border: 0" href="{% url "p3-schedule-my-schedule-ics" conference=conference %}"><img src="{{ STATIC_URL }}p6/i/icon-ical.png" /></a>{% endblock %}
<|file_sep|>current/p3/templates/p3/my_schedule.html
{% extends "p3/schedule_list.html" %}
{% block CONTENT_TITLE %}My Europython schedule <a style="border: 0" href="{% url "p3-schedule-my-schedule-ics" conference=conference %}"><img src="{{ STATIC_URL }}p6/i/icon-ical.png" /></a>{% endblock %}
<|file_sep|>updated/p3/templates/p3/my_schedule.html | {% extends "p3/schedule_list.html" %}
{% block CONTENT_TITLE %}My Europython schedule <a style="border: 0" href="{% url "p3-schedule-my-schedule-ics" conference=conference %}"><img src="{{ STATIC_URL }}p6/images/icon-ics.png" /></a>{% endblock %} | <|file_sep|>original/p3/templates/p3/my_schedule.html
{% extends "p3/schedule_list.html" %}
{% block CONTENT_TITLE %}My Europython schedule <a style="border: 0" href="{% url "p3-schedule-my-schedule-ics" conference=conference %}"><img src="{{ STATIC_URL }}p6/i/icon-ical.png" /></a>{% endblock %}
<|file_sep|>current/p3/templates/p3/my_schedule.html
{% extends "p3/schedule_list.html" %}
{% block CONTENT_TITLE %}My Europython schedule <a style="border: 0" href="{% url "p3-schedule-my-schedule-ics" conference=conference %}"><img src="{{ STATIC_URL }}p6/i/icon-ical.png" /></a>{% endblock %}
<|file_sep|>updated/p3/templates/p3/my_schedule.html
{% extends "p3/schedule_list.html" %}
{% block CONTENT_TITLE %}My Europython schedule <a style="border: 0" href="{% url "p3-schedule-my-schedule-ics" conference=conference %}"><img src="{{ STATIC_URL }}p6/images/icon-ics.png" /></a>{% endblock %} | e193684d52cc60bfa4ee72ffb12188d229210a70 | p3/templates/p3/my_schedule.html | p3/templates/p3/my_schedule.html | HTML |
<|file_sep|>original/package.json
"react",
"slider",
"component"
],
"author": "Michał Powaga <michalpowaga13@gmail.com>",
"contributors": [{
"name": "Florian Klampfer",
"email": "f.klampfer@gmail.com"
}],
"license": "MIT",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "https://github.com/mpowaga/react-slider.git"
},
"peerDependencies": {
"react": ">=0.14"
}
}
<|file_sep|>current/package.json
"react",
"slider",
"component"
],
"author": "Michał Powaga <michalpowaga13@gmail.com>",
"contributors": [{
"name": "Florian Klampfer",
"email": "f.klampfer@gmail.com"
}],
"license": "MIT",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "https://github.com/mpowaga/react-slider.git"
},
"peerDependencies": {
"react": ">=0.14"
}
}
<|file_sep|>updated/package.json | "react",
"slider",
"component"
],
"author": "Michał Powaga <michalpowaga13@gmail.com>",
"contributors": [{
"name": "Florian Klampfer",
"email": "f.klampfer@gmail.com"
}],
"license": "MIT",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "https://github.com/mpowaga/react-slider.git"
},
"peerDependencies": {
"react": ">=0.14 || ^15.0.0"
}
} | <|file_sep|>original/package.json
"react",
"slider",
"component"
],
"author": "Michał Powaga <michalpowaga13@gmail.com>",
"contributors": [{
"name": "Florian Klampfer",
"email": "f.klampfer@gmail.com"
}],
"license": "MIT",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "https://github.com/mpowaga/react-slider.git"
},
"peerDependencies": {
"react": ">=0.14"
}
}
<|file_sep|>current/package.json
"react",
"slider",
"component"
],
"author": "Michał Powaga <michalpowaga13@gmail.com>",
"contributors": [{
"name": "Florian Klampfer",
"email": "f.klampfer@gmail.com"
}],
"license": "MIT",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "https://github.com/mpowaga/react-slider.git"
},
"peerDependencies": {
"react": ">=0.14"
}
}
<|file_sep|>updated/package.json
"react",
"slider",
"component"
],
"author": "Michał Powaga <michalpowaga13@gmail.com>",
"contributors": [{
"name": "Florian Klampfer",
"email": "f.klampfer@gmail.com"
}],
"license": "MIT",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "https://github.com/mpowaga/react-slider.git"
},
"peerDependencies": {
"react": ">=0.14 || ^15.0.0"
}
} | e1d0df43598833ffd19d13956258e03cf73a1cfd | package.json | package.json | JSON |
<|file_sep|>original/Resources/config/orm/doctrine-entity/RefreshToken.orm.xml
<?xml version="1.0" encoding="UTF-8"?>
<doctrine-mapping xmlns="http://doctrine-project.org/schemas/orm/doctrine-mapping"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://doctrine-project.org/schemas/orm/doctrine-mapping
http://doctrine-project.org/schemas/orm/doctrine-mapping.xsd">
<entity name="Gesdinet\JWTRefreshTokenBundle\Entity\RefreshToken" table="refresh_tokens">
<id name="id" type="integer">
<generator strategy="AUTO"/>
</id>
</entity>
</doctrine-mapping>
<|file_sep|>current/Resources/config/orm/doctrine-entity/RefreshToken.orm.xml
<?xml version="1.0" encoding="UTF-8"?>
<doctrine-mapping xmlns="http://doctrine-project.org/schemas/orm/doctrine-mapping"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://doctrine-project.org/schemas/orm/doctrine-mapping
http://doctrine-project.org/schemas/orm/doctrine-mapping.xsd">
<entity name="Gesdinet\JWTRefreshTokenBundle\Entity\RefreshToken" table="refresh_tokens">
<id name="id" type="integer">
<generator strategy="AUTO"/>
</id>
</entity>
</doctrine-mapping>
<|file_sep|>updated/Resources/config/orm/doctrine-entity/RefreshToken.orm.xml | <?xml version="1.0" encoding="UTF-8"?>
<doctrine-mapping xmlns="http://doctrine-project.org/schemas/orm/doctrine-mapping"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://doctrine-project.org/schemas/orm/doctrine-mapping
http://doctrine-project.org/schemas/orm/doctrine-mapping.xsd">
<entity name="Gesdinet\JWTRefreshTokenBundle\Entity\RefreshToken" table="refresh_tokens" repository-class="Gesdinet\JWTRefreshTokenBundle\Entity\RefreshTokenRepository">
<id name="id" type="integer">
<generator strategy="AUTO"/>
</id>
</entity>
</doctrine-mapping> | <|file_sep|>original/Resources/config/orm/doctrine-entity/RefreshToken.orm.xml
<?xml version="1.0" encoding="UTF-8"?>
<doctrine-mapping xmlns="http://doctrine-project.org/schemas/orm/doctrine-mapping"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://doctrine-project.org/schemas/orm/doctrine-mapping
http://doctrine-project.org/schemas/orm/doctrine-mapping.xsd">
<entity name="Gesdinet\JWTRefreshTokenBundle\Entity\RefreshToken" table="refresh_tokens">
<id name="id" type="integer">
<generator strategy="AUTO"/>
</id>
</entity>
</doctrine-mapping>
<|file_sep|>current/Resources/config/orm/doctrine-entity/RefreshToken.orm.xml
<?xml version="1.0" encoding="UTF-8"?>
<doctrine-mapping xmlns="http://doctrine-project.org/schemas/orm/doctrine-mapping"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://doctrine-project.org/schemas/orm/doctrine-mapping
http://doctrine-project.org/schemas/orm/doctrine-mapping.xsd">
<entity name="Gesdinet\JWTRefreshTokenBundle\Entity\RefreshToken" table="refresh_tokens">
<id name="id" type="integer">
<generator strategy="AUTO"/>
</id>
</entity>
</doctrine-mapping>
<|file_sep|>updated/Resources/config/orm/doctrine-entity/RefreshToken.orm.xml
<?xml version="1.0" encoding="UTF-8"?>
<doctrine-mapping xmlns="http://doctrine-project.org/schemas/orm/doctrine-mapping"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://doctrine-project.org/schemas/orm/doctrine-mapping
http://doctrine-project.org/schemas/orm/doctrine-mapping.xsd">
<entity name="Gesdinet\JWTRefreshTokenBundle\Entity\RefreshToken" table="refresh_tokens" repository-class="Gesdinet\JWTRefreshTokenBundle\Entity\RefreshTokenRepository">
<id name="id" type="integer">
<generator strategy="AUTO"/>
</id>
</entity>
</doctrine-mapping> | f5472176b7599f0ea39c1b868d542cda722c33b3 | Resources/config/orm/doctrine-entity/RefreshToken.orm.xml | Resources/config/orm/doctrine-entity/RefreshToken.orm.xml | XML |
<|file_sep|>original/_posts/2020-01-15-Another_Internship.md
c---
title: "We are looking for on more intern !"
categories:
- Bordeaux
tags:
- robocup
- SSPL
- Pepper
---
We are opening an intern position within the team on the themes of Robot Navigation & Mapping !
## Main goal:
Enhance robot capabilities in terms of navigation strategies specific situations (crossing doors, going backwards, ...)
and perception (exploring the use of the RTABMAP and OCTOMAP frameworks).
The robot you will focus your experiments on will mainly be Pepper,
but being able to apply this on our Open Platform (Palbator) would be great !
If this piqued your interest, don't hesitate to take a look at the [full proposition here](/assets/pdf/PFE\ Robotique\ -\ Robocup\ 2020.pdf) and send us a message !
<|file_sep|>current/_posts/2020-01-15-Another_Internship.md
c---
title: "We are looking for on more intern !"
categories:
- Bordeaux
tags:
- robocup
- SSPL
- Pepper
---
We are opening an intern position within the team on the themes of Robot Navigation & Mapping !
## Main goal:
Enhance robot capabilities in terms of navigation strategies specific situations (crossing doors, going backwards, ...)
and perception (exploring the use of the RTABMAP and OCTOMAP frameworks).
The robot you will focus your experiments on will mainly be Pepper,
but being able to apply this on our Open Platform (Palbator) would be great !
If this piqued your interest, don't hesitate to take a look at the [full proposition here](/assets/pdf/PFE\ Robotique\ -\ Robocup\ 2020.pdf) and send us a message !
<|file_sep|>updated/_posts/2020-01-15-Another_Internship.md | ---
title: "We are looking for one more intern !"
categories:
- Bordeaux
tags:
- robocup
- SSPL
- Pepper
---
We are opening an intern position within the team on the themes of Robot Navigation & Mapping !
## Main goal:
Enhance robot capabilities in terms of navigation strategies specific situations (crossing doors, going backwards, ...)
and perception (exploring the use of the RTABMAP and OCTOMAP frameworks).
The robot you will focus your experiments on will mainly be Pepper,
but being able to apply this on our Open Platform (Palbator) would be great !
If this piqued your interest, don't hesitate to take a look at the [full proposition here](/assets/pdf/PFE\ Robotique\ -\ Robocup\ 2020.pdf) and send us a message ! | <|file_sep|>original/_posts/2020-01-15-Another_Internship.md
c---
title: "We are looking for on more intern !"
categories:
- Bordeaux
tags:
- robocup
- SSPL
- Pepper
---
We are opening an intern position within the team on the themes of Robot Navigation & Mapping !
## Main goal:
Enhance robot capabilities in terms of navigation strategies specific situations (crossing doors, going backwards, ...)
and perception (exploring the use of the RTABMAP and OCTOMAP frameworks).
The robot you will focus your experiments on will mainly be Pepper,
but being able to apply this on our Open Platform (Palbator) would be great !
If this piqued your interest, don't hesitate to take a look at the [full proposition here](/assets/pdf/PFE\ Robotique\ -\ Robocup\ 2020.pdf) and send us a message !
<|file_sep|>current/_posts/2020-01-15-Another_Internship.md
c---
title: "We are looking for on more intern !"
categories:
- Bordeaux
tags:
- robocup
- SSPL
- Pepper
---
We are opening an intern position within the team on the themes of Robot Navigation & Mapping !
## Main goal:
Enhance robot capabilities in terms of navigation strategies specific situations (crossing doors, going backwards, ...)
and perception (exploring the use of the RTABMAP and OCTOMAP frameworks).
The robot you will focus your experiments on will mainly be Pepper,
but being able to apply this on our Open Platform (Palbator) would be great !
If this piqued your interest, don't hesitate to take a look at the [full proposition here](/assets/pdf/PFE\ Robotique\ -\ Robocup\ 2020.pdf) and send us a message !
<|file_sep|>updated/_posts/2020-01-15-Another_Internship.md
---
title: "We are looking for one more intern !"
categories:
- Bordeaux
tags:
- robocup
- SSPL
- Pepper
---
We are opening an intern position within the team on the themes of Robot Navigation & Mapping !
## Main goal:
Enhance robot capabilities in terms of navigation strategies specific situations (crossing doors, going backwards, ...)
and perception (exploring the use of the RTABMAP and OCTOMAP frameworks).
The robot you will focus your experiments on will mainly be Pepper,
but being able to apply this on our Open Platform (Palbator) would be great !
If this piqued your interest, don't hesitate to take a look at the [full proposition here](/assets/pdf/PFE\ Robotique\ -\ Robocup\ 2020.pdf) and send us a message ! | cc8e365dc27172d8fb452b1615ea39fdfec42a60 | _posts/2020-01-15-Another_Internship.md | _posts/2020-01-15-Another_Internship.md | Markdown |
<|file_sep|>original/src/main/java/com/ctci/treesandgraphs/ValidateBST.java
<|file_sep|>current/src/main/java/com/ctci/treesandgraphs/ValidateBST.java
<|file_sep|>updated/src/main/java/com/ctci/treesandgraphs/ValidateBST.java | package com.ctci.treesandgraphs;
import java.util.ArrayList;
import java.util.List;
/**
* @author rampatra
* @since 2019-02-17
*/
public class ValidateBST {
private static boolean isBST(TreeNode node) {
return isBST(node, new ArrayList<>());
}
private static boolean isBST(TreeNode node, List<Integer> values) {
if (node == null) return true;
isBST(node.left, values);
if (values.isEmpty() || node.val > values.get(values.size() - 1)) {
values.add(node.val); | <|file_sep|>original/src/main/java/com/ctci/treesandgraphs/ValidateBST.java
<|file_sep|>current/src/main/java/com/ctci/treesandgraphs/ValidateBST.java
<|file_sep|>updated/src/main/java/com/ctci/treesandgraphs/ValidateBST.java
package com.ctci.treesandgraphs;
import java.util.ArrayList;
import java.util.List;
/**
* @author rampatra
* @since 2019-02-17
*/
public class ValidateBST {
private static boolean isBST(TreeNode node) {
return isBST(node, new ArrayList<>());
}
private static boolean isBST(TreeNode node, List<Integer> values) {
if (node == null) return true;
isBST(node.left, values);
if (values.isEmpty() || node.val > values.get(values.size() - 1)) {
values.add(node.val); | 6d0c5c4c8ac537577ad6b5f184b35d8659bb69fe | src/main/java/com/ctci/treesandgraphs/ValidateBST.java | src/main/java/com/ctci/treesandgraphs/ValidateBST.java | Java |
<|file_sep|>original/src/application.bat
### application.bat - 2015 - Merel ###
echo "Copyright Benoit MEREL - 2015"
version = "2"
build_hour = "11H00"
echo "Build ${build_hour}"
echo "version ${version}"
<|file_sep|>current/src/application.bat
### application.bat - 2015 - Merel ###
echo "Copyright Benoit MEREL - 2015"
version = "2"
build_hour = "11H00"
echo "Build ${build_hour}"
echo "version ${version}"
<|file_sep|>updated/src/application.bat | ### application.bat - 2015 - Guillet ###
echo "Copyright Thomas Guillet - 2015"
version = "2"
build_hour = "11H00"
echo "Build ${build_hour}"
echo "version ${version}"
| <|file_sep|>original/src/application.bat
### application.bat - 2015 - Merel ###
echo "Copyright Benoit MEREL - 2015"
version = "2"
build_hour = "11H00"
echo "Build ${build_hour}"
echo "version ${version}"
<|file_sep|>current/src/application.bat
### application.bat - 2015 - Merel ###
echo "Copyright Benoit MEREL - 2015"
version = "2"
build_hour = "11H00"
echo "Build ${build_hour}"
echo "version ${version}"
<|file_sep|>updated/src/application.bat
### application.bat - 2015 - Guillet ###
echo "Copyright Thomas Guillet - 2015"
version = "2"
build_hour = "11H00"
echo "Build ${build_hour}"
echo "version ${version}"
| eafdae19b5d7bac54835cdca95383013eb356643 | src/application.bat | src/application.bat | Batchfile |
<|file_sep|>original/.travis.yml
compiler: gcc
addons:
apt:
sources: ['ubuntu-toolchain-r-test']
packages: ['g++-5']
env: COMPILER=g++-5
- os: linux
compiler: clang
addons:
apt:
sources: ['ubuntu-toolchain-r-test', 'llvm-toolchain-precise-3.8']
packages: ['clang-3.8']
env: COMPILER=clang++-3.8
before_script:
git clone https://github.com/siquel/kaluketju
script:
- if [ "$CXX" == "clang++" ]; then make CXX=$COMPILER linux-clang && .build/linux64_clang/bin/unit_testDebug; fi
- if [ "$CXX" == "g++" ]; then make CXX=$COMPILER linux-gcc && .build/linux64_gcc/bin/unit_testDebug; fi
<|file_sep|>current/.travis.yml
compiler: gcc
addons:
apt:
sources: ['ubuntu-toolchain-r-test']
packages: ['g++-5']
env: COMPILER=g++-5
- os: linux
compiler: clang
addons:
apt:
sources: ['ubuntu-toolchain-r-test', 'llvm-toolchain-precise-3.8']
packages: ['clang-3.8']
env: COMPILER=clang++-3.8
before_script:
git clone https://github.com/siquel/kaluketju
script:
- if [ "$CXX" == "clang++" ]; then make CXX=$COMPILER linux-clang && .build/linux64_clang/bin/unit_testDebug; fi
- if [ "$CXX" == "g++" ]; then make CXX=$COMPILER linux-gcc && .build/linux64_gcc/bin/unit_testDebug; fi
<|file_sep|>updated/.travis.yml | compiler: gcc
addons:
apt:
sources: ['ubuntu-toolchain-r-test']
packages: ['g++-5']
env: COMPILER=g++-5
# - os: linux
# compiler: clang
# addons:
# apt:
# sources: ['ubuntu-toolchain-r-test', 'llvm-toolchain-precise-3.8']
# packages: ['clang-3.8']
# env: COMPILER=clang++-3.8
before_script:
git clone https://github.com/siquel/kaluketju
script:
- if [ "$CXX" == "clang++" ]; then make CXX=$COMPILER linux-clang && .build/linux64_clang/bin/unit_testDebug; fi
- if [ "$CXX" == "g++" ]; then make CXX=$COMPILER linux-gcc && .build/linux64_gcc/bin/unit_testDebug; fi | <|file_sep|>original/.travis.yml
compiler: gcc
addons:
apt:
sources: ['ubuntu-toolchain-r-test']
packages: ['g++-5']
env: COMPILER=g++-5
- os: linux
compiler: clang
addons:
apt:
sources: ['ubuntu-toolchain-r-test', 'llvm-toolchain-precise-3.8']
packages: ['clang-3.8']
env: COMPILER=clang++-3.8
before_script:
git clone https://github.com/siquel/kaluketju
script:
- if [ "$CXX" == "clang++" ]; then make CXX=$COMPILER linux-clang && .build/linux64_clang/bin/unit_testDebug; fi
- if [ "$CXX" == "g++" ]; then make CXX=$COMPILER linux-gcc && .build/linux64_gcc/bin/unit_testDebug; fi
<|file_sep|>current/.travis.yml
compiler: gcc
addons:
apt:
sources: ['ubuntu-toolchain-r-test']
packages: ['g++-5']
env: COMPILER=g++-5
- os: linux
compiler: clang
addons:
apt:
sources: ['ubuntu-toolchain-r-test', 'llvm-toolchain-precise-3.8']
packages: ['clang-3.8']
env: COMPILER=clang++-3.8
before_script:
git clone https://github.com/siquel/kaluketju
script:
- if [ "$CXX" == "clang++" ]; then make CXX=$COMPILER linux-clang && .build/linux64_clang/bin/unit_testDebug; fi
- if [ "$CXX" == "g++" ]; then make CXX=$COMPILER linux-gcc && .build/linux64_gcc/bin/unit_testDebug; fi
<|file_sep|>updated/.travis.yml
compiler: gcc
addons:
apt:
sources: ['ubuntu-toolchain-r-test']
packages: ['g++-5']
env: COMPILER=g++-5
# - os: linux
# compiler: clang
# addons:
# apt:
# sources: ['ubuntu-toolchain-r-test', 'llvm-toolchain-precise-3.8']
# packages: ['clang-3.8']
# env: COMPILER=clang++-3.8
before_script:
git clone https://github.com/siquel/kaluketju
script:
- if [ "$CXX" == "clang++" ]; then make CXX=$COMPILER linux-clang && .build/linux64_clang/bin/unit_testDebug; fi
- if [ "$CXX" == "g++" ]; then make CXX=$COMPILER linux-gcc && .build/linux64_gcc/bin/unit_testDebug; fi | f5881cca371eeafe672a9f6163644a25b323b71a | .travis.yml | .travis.yml | YAML |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.