commit stringlengths 40 40 | old_file stringlengths 4 184 | new_file stringlengths 4 184 | old_contents stringlengths 1 3.6k | new_contents stringlengths 5 3.38k | subject stringlengths 15 778 | message stringlengths 16 6.74k | lang stringclasses 201 values | license stringclasses 13 values | repos stringlengths 6 116k | config stringclasses 201 values | content stringlengths 137 7.24k | diff stringlengths 26 5.55k | diff_length int64 1 123 | relative_diff_length float64 0.01 89 | n_lines_added int64 0 108 | n_lines_deleted int64 0 106 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
f33dccc38905998303676e34a066efccd839678f | src/main/webapp/SpecRunner.html | src/main/webapp/SpecRunner.html | <!DOCTYPE html>
<html>
<head>
<meta charset="utf-8">
<title>Jasmine Spec Runner v3.5.0</title>
<link rel="shortcut icon" type="image/png" href="webapptests/lib/jasmine-3.5.0/jasmine_favicon.png">
<link rel="stylesheet" href="webapptests/lib/jasmine-3.5.0/jasmine.css">
<script src="webapptests/lib/jasmine-3.5.0/jasmine.js"></script>
<script src="webapptests/lib/jasmine-3.5.0/jasmine-html.js"></script>
<script src="webapptests/lib/jasmine-3.5.0/boot.js"></script>
<!-- include source files here... -->
<script src="search.js"></script>
<script src="availability.js"></script>
<!-- include spec files here... -->
<script src="webapptests/spec/SpecSearch.js"></script>
<script src="webapptests/spec/SpecAvailability.js"></script>
</head>
<body>
</body>
</html>
| <!DOCTYPE html>
<html>
<head>
<meta charset="utf-8">
<title>Jasmine Spec Runner v3.5.0</title>
<link rel="shortcut icon" type="image/png" href="webapptests/lib/jasmine-3.5.0/jasmine_favicon.png">
<link rel="stylesheet" href="webapptests/lib/jasmine-3.5.0/jasmine.css">
<script src="webapptests/lib/jasmine-3.5.0/jasmine.js"></script>
<script src="webapptests/lib/jasmine-3.5.0/jasmine-html.js"></script>
<script src="webapptests/lib/jasmine-3.5.0/boot.js"></script>
<!-- include source files here... -->
<script src="search.js"></script>
<script src="availability.js"></script>
<script src="scheduling.js"></script>
<script src="confirmation.js"></script>
<!-- include spec files here... -->
<script src="webapptests/spec/SpecSearch.js"></script>
<script src="webapptests/spec/SpecAvailability.js"></script>
<script src="webapptests/spec/SpecScheduling.js"></script>
<script src="webapptests/spec/SpecConfirmation.js"></script>
</head>
<body>
</body>
</html>
| Add jasmine test files to the spec runner. | Add jasmine test files to the spec runner.
| HTML | apache-2.0 | googleinterns/step132-2020,googleinterns/step132-2020,googleinterns/step132-2020 | html | ## Code Before:
<!DOCTYPE html>
<html>
<head>
<meta charset="utf-8">
<title>Jasmine Spec Runner v3.5.0</title>
<link rel="shortcut icon" type="image/png" href="webapptests/lib/jasmine-3.5.0/jasmine_favicon.png">
<link rel="stylesheet" href="webapptests/lib/jasmine-3.5.0/jasmine.css">
<script src="webapptests/lib/jasmine-3.5.0/jasmine.js"></script>
<script src="webapptests/lib/jasmine-3.5.0/jasmine-html.js"></script>
<script src="webapptests/lib/jasmine-3.5.0/boot.js"></script>
<!-- include source files here... -->
<script src="search.js"></script>
<script src="availability.js"></script>
<!-- include spec files here... -->
<script src="webapptests/spec/SpecSearch.js"></script>
<script src="webapptests/spec/SpecAvailability.js"></script>
</head>
<body>
</body>
</html>
## Instruction:
Add jasmine test files to the spec runner.
## Code After:
<!DOCTYPE html>
<html>
<head>
<meta charset="utf-8">
<title>Jasmine Spec Runner v3.5.0</title>
<link rel="shortcut icon" type="image/png" href="webapptests/lib/jasmine-3.5.0/jasmine_favicon.png">
<link rel="stylesheet" href="webapptests/lib/jasmine-3.5.0/jasmine.css">
<script src="webapptests/lib/jasmine-3.5.0/jasmine.js"></script>
<script src="webapptests/lib/jasmine-3.5.0/jasmine-html.js"></script>
<script src="webapptests/lib/jasmine-3.5.0/boot.js"></script>
<!-- include source files here... -->
<script src="search.js"></script>
<script src="availability.js"></script>
<script src="scheduling.js"></script>
<script src="confirmation.js"></script>
<!-- include spec files here... -->
<script src="webapptests/spec/SpecSearch.js"></script>
<script src="webapptests/spec/SpecAvailability.js"></script>
<script src="webapptests/spec/SpecScheduling.js"></script>
<script src="webapptests/spec/SpecConfirmation.js"></script>
</head>
<body>
</body>
</html>
| <!DOCTYPE html>
<html>
<head>
<meta charset="utf-8">
<title>Jasmine Spec Runner v3.5.0</title>
<link rel="shortcut icon" type="image/png" href="webapptests/lib/jasmine-3.5.0/jasmine_favicon.png">
<link rel="stylesheet" href="webapptests/lib/jasmine-3.5.0/jasmine.css">
<script src="webapptests/lib/jasmine-3.5.0/jasmine.js"></script>
<script src="webapptests/lib/jasmine-3.5.0/jasmine-html.js"></script>
<script src="webapptests/lib/jasmine-3.5.0/boot.js"></script>
<!-- include source files here... -->
<script src="search.js"></script>
<script src="availability.js"></script>
+ <script src="scheduling.js"></script>
+ <script src="confirmation.js"></script>
<!-- include spec files here... -->
<script src="webapptests/spec/SpecSearch.js"></script>
<script src="webapptests/spec/SpecAvailability.js"></script>
+ <script src="webapptests/spec/SpecScheduling.js"></script>
+ <script src="webapptests/spec/SpecConfirmation.js"></script>
</head>
<body>
</body>
</html> | 4 | 0.153846 | 4 | 0 |
6846085cf9c53e260fc32f923736e71ea987ed04 | app/views/pages/tos.html.erb | app/views/pages/tos.html.erb | <% title 'Terms and Conditions' %>
<%= render 'shared/navbar' %>
<div class="page page-flat">
<div class="container">
<div class="row">
<div class="col-md-8 col-xs-8 col-md-offset-2 col-xs-offset-2">
<div class="page-header">
<h2 class="page-header-title">Terms and Conditions</h2>
<p>
Last updated on <%= markdown_mtime('app/views/pages/tos_content').to_date.to_s(:long) %>
(<a class="small" href="https://github.com/assemblymade/legal" rel="nofollow">Previous versions</a>)
</p>
</div>
<p class="lead">
Welcome to Assembly, a web site that lets Members submit App Ideas for feedback from and development by the Assembly community of users (the “Community”). The web site is first and foremost designed to facilitate the creation and evolution of viable software products. However, selected App Ideas will further be developed by the Community and may ultimately be commercialized, produced and licensed or sold by Assembly. The assembly.com web site and its associated services (the “Services”) may be found at the domain and its related subdomains found at assembly.com (the “Site”).
</p>
<%= markdown_content 'app/views/pages/tos_content' %>
</div>
</div>
</div>
</div>
<%= render 'shared/footer' %>
| <% title 'Terms and Conditions' %>
<%= render 'shared/navbar' %>
<div class="page page-flat">
<div class="container">
<div class="row">
<div class="col-md-8 col-xs-8 col-md-offset-2 col-xs-offset-2">
<div class="page-header">
<h2 class="page-header-title">Terms and Conditions</h2>
<p>
Last updated on <%= markdown_mtime('app/views/pages/tos_content').to_date.to_s(:long) %>
(<a class="small" href="https://github.com/assemblymade/legal" rel="nofollow">Previous versions</a>)
</p>
</div>
<p class="lead">
Welcome to Assembly, a web site that lets Members submit App Ideas for feedback from and development by the Assembly community of users (the “Community”). The web site is first and foremost designed to facilitate the creation and evolution of viable software products. However, selected App Ideas will further be developed by the Community and may ultimately be commercialized, produced and licensed or sold by Assembly (on behalf of the community, as described in this Agreement). The assembly.com web site and its associated services (the “Services”) may be found at the domain and its related subdomains found at assembly.com (the “Site”).
</p>
<%= markdown_content 'app/views/pages/tos_content' %>
</div>
</div>
</div>
</div>
<%= render 'shared/footer' %>
| Update terms to make the selling of products more clear | Update terms to make the selling of products more clear
| HTML+ERB | agpl-3.0 | lachlanjc/meta,assemblymade/meta,assemblymade/meta,lachlanjc/meta,lachlanjc/meta,assemblymade/meta,assemblymade/meta,lachlanjc/meta | html+erb | ## Code Before:
<% title 'Terms and Conditions' %>
<%= render 'shared/navbar' %>
<div class="page page-flat">
<div class="container">
<div class="row">
<div class="col-md-8 col-xs-8 col-md-offset-2 col-xs-offset-2">
<div class="page-header">
<h2 class="page-header-title">Terms and Conditions</h2>
<p>
Last updated on <%= markdown_mtime('app/views/pages/tos_content').to_date.to_s(:long) %>
(<a class="small" href="https://github.com/assemblymade/legal" rel="nofollow">Previous versions</a>)
</p>
</div>
<p class="lead">
Welcome to Assembly, a web site that lets Members submit App Ideas for feedback from and development by the Assembly community of users (the “Community”). The web site is first and foremost designed to facilitate the creation and evolution of viable software products. However, selected App Ideas will further be developed by the Community and may ultimately be commercialized, produced and licensed or sold by Assembly. The assembly.com web site and its associated services (the “Services”) may be found at the domain and its related subdomains found at assembly.com (the “Site”).
</p>
<%= markdown_content 'app/views/pages/tos_content' %>
</div>
</div>
</div>
</div>
<%= render 'shared/footer' %>
## Instruction:
Update terms to make the selling of products more clear
## Code After:
<% title 'Terms and Conditions' %>
<%= render 'shared/navbar' %>
<div class="page page-flat">
<div class="container">
<div class="row">
<div class="col-md-8 col-xs-8 col-md-offset-2 col-xs-offset-2">
<div class="page-header">
<h2 class="page-header-title">Terms and Conditions</h2>
<p>
Last updated on <%= markdown_mtime('app/views/pages/tos_content').to_date.to_s(:long) %>
(<a class="small" href="https://github.com/assemblymade/legal" rel="nofollow">Previous versions</a>)
</p>
</div>
<p class="lead">
Welcome to Assembly, a web site that lets Members submit App Ideas for feedback from and development by the Assembly community of users (the “Community”). The web site is first and foremost designed to facilitate the creation and evolution of viable software products. However, selected App Ideas will further be developed by the Community and may ultimately be commercialized, produced and licensed or sold by Assembly (on behalf of the community, as described in this Agreement). The assembly.com web site and its associated services (the “Services”) may be found at the domain and its related subdomains found at assembly.com (the “Site”).
</p>
<%= markdown_content 'app/views/pages/tos_content' %>
</div>
</div>
</div>
</div>
<%= render 'shared/footer' %>
| <% title 'Terms and Conditions' %>
<%= render 'shared/navbar' %>
<div class="page page-flat">
<div class="container">
<div class="row">
<div class="col-md-8 col-xs-8 col-md-offset-2 col-xs-offset-2">
<div class="page-header">
<h2 class="page-header-title">Terms and Conditions</h2>
<p>
Last updated on <%= markdown_mtime('app/views/pages/tos_content').to_date.to_s(:long) %>
(<a class="small" href="https://github.com/assemblymade/legal" rel="nofollow">Previous versions</a>)
</p>
</div>
<p class="lead">
- Welcome to Assembly, a web site that lets Members submit App Ideas for feedback from and development by the Assembly community of users (the “Community”). The web site is first and foremost designed to facilitate the creation and evolution of viable software products. However, selected App Ideas will further be developed by the Community and may ultimately be commercialized, produced and licensed or sold by Assembly. The assembly.com web site and its associated services (the “Services”) may be found at the domain and its related subdomains found at assembly.com (the “Site”).
+ Welcome to Assembly, a web site that lets Members submit App Ideas for feedback from and development by the Assembly community of users (the “Community”). The web site is first and foremost designed to facilitate the creation and evolution of viable software products. However, selected App Ideas will further be developed by the Community and may ultimately be commercialized, produced and licensed or sold by Assembly (on behalf of the community, as described in this Agreement). The assembly.com web site and its associated services (the “Services”) may be found at the domain and its related subdomains found at assembly.com (the “Site”).
? +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
</p>
<%= markdown_content 'app/views/pages/tos_content' %>
</div>
</div>
</div>
</div>
<%= render 'shared/footer' %> | 2 | 0.076923 | 1 | 1 |
d40ddc3d3bcc16440e285c008645d02c19bca530 | build/version_plugin.js | build/version_plugin.js | const commit = require('git-rev-sync').short();
const pkg = require('../package.json');
const version = JSON.stringify({
commit,
source: pkg.homepage,
version: process.env.CIRCLE_TAG || `v${pkg.version}`
});
class VersionPlugin {
apply(compiler) {
compiler.hooks.emit.tap('VersionPlugin', compilation => {
compilation.assets['version.json'] = {
source() {
return version;
},
size() {
return version.length;
}
};
});
}
}
module.exports = VersionPlugin;
| const gitRevSync = require('git-rev-sync');
const pkg = require('../package.json');
let commit = 'unknown';
try {
commit = gitRevSync.short();
} catch (e) {
console.warn('Error fetching current git commit: ' + e);
}
const version = JSON.stringify({
commit,
source: pkg.homepage,
version: process.env.CIRCLE_TAG || `v${pkg.version}`
});
class VersionPlugin {
apply(compiler) {
compiler.hooks.emit.tap('VersionPlugin', compilation => {
compilation.assets['version.json'] = {
source() {
return version;
},
size() {
return version.length;
}
};
});
}
}
module.exports = VersionPlugin;
| Fix npm build with missing .git | Fix npm build with missing .git
| JavaScript | mpl-2.0 | mozilla/send,mozilla/send,mozilla/send,mozilla/send,mozilla/send | javascript | ## Code Before:
const commit = require('git-rev-sync').short();
const pkg = require('../package.json');
const version = JSON.stringify({
commit,
source: pkg.homepage,
version: process.env.CIRCLE_TAG || `v${pkg.version}`
});
class VersionPlugin {
apply(compiler) {
compiler.hooks.emit.tap('VersionPlugin', compilation => {
compilation.assets['version.json'] = {
source() {
return version;
},
size() {
return version.length;
}
};
});
}
}
module.exports = VersionPlugin;
## Instruction:
Fix npm build with missing .git
## Code After:
const gitRevSync = require('git-rev-sync');
const pkg = require('../package.json');
let commit = 'unknown';
try {
commit = gitRevSync.short();
} catch (e) {
console.warn('Error fetching current git commit: ' + e);
}
const version = JSON.stringify({
commit,
source: pkg.homepage,
version: process.env.CIRCLE_TAG || `v${pkg.version}`
});
class VersionPlugin {
apply(compiler) {
compiler.hooks.emit.tap('VersionPlugin', compilation => {
compilation.assets['version.json'] = {
source() {
return version;
},
size() {
return version.length;
}
};
});
}
}
module.exports = VersionPlugin;
| - const commit = require('git-rev-sync').short();
? ^^^^ --------
+ const gitRevSync = require('git-rev-sync');
? ^ +++++++
const pkg = require('../package.json');
+
+ let commit = 'unknown';
+
+ try {
+ commit = gitRevSync.short();
+ } catch (e) {
+ console.warn('Error fetching current git commit: ' + e);
+ }
const version = JSON.stringify({
commit,
source: pkg.homepage,
version: process.env.CIRCLE_TAG || `v${pkg.version}`
});
class VersionPlugin {
apply(compiler) {
compiler.hooks.emit.tap('VersionPlugin', compilation => {
compilation.assets['version.json'] = {
source() {
return version;
},
size() {
return version.length;
}
};
});
}
}
module.exports = VersionPlugin; | 10 | 0.4 | 9 | 1 |
9f556a5b7f7c34c73ab6c753cf97a8a1630251d3 | packages/components/containers/apps/ProtonMailBridgeSection.js | packages/components/containers/apps/ProtonMailBridgeSection.js | import React from 'react';
import { c } from 'ttag';
import { SubTitle, Alert } from 'react-components';
const ProtonMailBridgeSection = () => {
return (
<>
<SubTitle>ProtonMail Bridge</SubTitle>
<Alert learnMore="https://protonmail.com/bridge/">{c('Info')
.t`ProtonMail Supports IMAP/SMTP via the ProtonMail Bridge application. Thunderbird, Microsoft Outlook, and Apple Mail are officially supported on both Windows and MacOS.`}</Alert>
</>
);
};
export default ProtonMailBridgeSection;
| import React from 'react';
import { c } from 'ttag';
import { SubTitle, Alert, PrimaryButton, useUser, useNotifications } from 'react-components';
const ProtonMailBridgeSection = () => {
const [{ isPaidMail }] = useUser();
const { createNotification } = useNotifications();
const handleClick = () =>
createNotification({
type: 'info',
text: c('Info').t`This feature is only available for paid users.`
});
return (
<>
<SubTitle>ProtonMail Bridge</SubTitle>
<Alert learnMore="https://protonmail.com/bridge/">{c('Info')
.t`ProtonMail Supports IMAP/SMTP via the ProtonMail Bridge application. Thunderbird, Microsoft Outlook, and Apple Mail are officially supported on both Windows and MacOS.`}</Alert>
{isPaidMail ? null : <PrimaryButton onClick={handleClick}>{c('Action').t`Activate`}</PrimaryButton>}
</>
);
};
export default ProtonMailBridgeSection;
| Add action to Activate bridge | Add action to Activate bridge
| JavaScript | mit | ProtonMail/WebClient,ProtonMail/WebClient,ProtonMail/WebClient | javascript | ## Code Before:
import React from 'react';
import { c } from 'ttag';
import { SubTitle, Alert } from 'react-components';
const ProtonMailBridgeSection = () => {
return (
<>
<SubTitle>ProtonMail Bridge</SubTitle>
<Alert learnMore="https://protonmail.com/bridge/">{c('Info')
.t`ProtonMail Supports IMAP/SMTP via the ProtonMail Bridge application. Thunderbird, Microsoft Outlook, and Apple Mail are officially supported on both Windows and MacOS.`}</Alert>
</>
);
};
export default ProtonMailBridgeSection;
## Instruction:
Add action to Activate bridge
## Code After:
import React from 'react';
import { c } from 'ttag';
import { SubTitle, Alert, PrimaryButton, useUser, useNotifications } from 'react-components';
const ProtonMailBridgeSection = () => {
const [{ isPaidMail }] = useUser();
const { createNotification } = useNotifications();
const handleClick = () =>
createNotification({
type: 'info',
text: c('Info').t`This feature is only available for paid users.`
});
return (
<>
<SubTitle>ProtonMail Bridge</SubTitle>
<Alert learnMore="https://protonmail.com/bridge/">{c('Info')
.t`ProtonMail Supports IMAP/SMTP via the ProtonMail Bridge application. Thunderbird, Microsoft Outlook, and Apple Mail are officially supported on both Windows and MacOS.`}</Alert>
{isPaidMail ? null : <PrimaryButton onClick={handleClick}>{c('Action').t`Activate`}</PrimaryButton>}
</>
);
};
export default ProtonMailBridgeSection;
| import React from 'react';
import { c } from 'ttag';
- import { SubTitle, Alert } from 'react-components';
+ import { SubTitle, Alert, PrimaryButton, useUser, useNotifications } from 'react-components';
const ProtonMailBridgeSection = () => {
+ const [{ isPaidMail }] = useUser();
+ const { createNotification } = useNotifications();
+ const handleClick = () =>
+ createNotification({
+ type: 'info',
+ text: c('Info').t`This feature is only available for paid users.`
+ });
return (
<>
<SubTitle>ProtonMail Bridge</SubTitle>
<Alert learnMore="https://protonmail.com/bridge/">{c('Info')
.t`ProtonMail Supports IMAP/SMTP via the ProtonMail Bridge application. Thunderbird, Microsoft Outlook, and Apple Mail are officially supported on both Windows and MacOS.`}</Alert>
+ {isPaidMail ? null : <PrimaryButton onClick={handleClick}>{c('Action').t`Activate`}</PrimaryButton>}
</>
);
};
export default ProtonMailBridgeSection; | 10 | 0.666667 | 9 | 1 |
4682dba3e6257a48d91fa5f35c1df3b16b030ab1 | src/main/java/modtweaker2/ModProps.java | src/main/java/modtweaker2/ModProps.java | package modtweaker2;
public class ModProps {
public static final String NAME = "Mod Tweaker 2", name = NAME;
public static final String MODID = "modtweaker2", modid = MODID;
public static final String VERSION = "0.9.4", version = VERSION;
public static final String DEPENDENCIES = "required-after:MineTweaker3", dependencies = DEPENDENCIES;
}
| package modtweaker2;
public class ModProps {
public static final String NAME = "Mod Tweaker 2", name = NAME;
public static final String MODID = "modtweaker2", modid = MODID;
public static final String VERSION = "0.9.4", version = VERSION;
public static final String DEPENDENCIES = "required-after:MineTweaker3;after:Forestry@[4.0.3,);", dependencies = DEPENDENCIES;
}
| Update mod dependency info for Forestry 4, thanks @mezz | Update mod dependency info for Forestry 4, thanks @mezz
| Java | mit | elucent/ModTweaker,jaredlll08/ModTweaker2,yrsegal/ModTweaker2,jaredlll08/ModTweaker | java | ## Code Before:
package modtweaker2;
public class ModProps {
public static final String NAME = "Mod Tweaker 2", name = NAME;
public static final String MODID = "modtweaker2", modid = MODID;
public static final String VERSION = "0.9.4", version = VERSION;
public static final String DEPENDENCIES = "required-after:MineTweaker3", dependencies = DEPENDENCIES;
}
## Instruction:
Update mod dependency info for Forestry 4, thanks @mezz
## Code After:
package modtweaker2;
public class ModProps {
public static final String NAME = "Mod Tweaker 2", name = NAME;
public static final String MODID = "modtweaker2", modid = MODID;
public static final String VERSION = "0.9.4", version = VERSION;
public static final String DEPENDENCIES = "required-after:MineTweaker3;after:Forestry@[4.0.3,);", dependencies = DEPENDENCIES;
}
| package modtweaker2;
public class ModProps {
public static final String NAME = "Mod Tweaker 2", name = NAME;
public static final String MODID = "modtweaker2", modid = MODID;
public static final String VERSION = "0.9.4", version = VERSION;
- public static final String DEPENDENCIES = "required-after:MineTweaker3", dependencies = DEPENDENCIES;
+ public static final String DEPENDENCIES = "required-after:MineTweaker3;after:Forestry@[4.0.3,);", dependencies = DEPENDENCIES;
? +++++++++++++++++++++++++
} | 2 | 0.222222 | 1 | 1 |
67f4b60b3a2a2fcaedebfe028453f6592af6c509 | tools/python.mk | tools/python.mk | .ONESHELL:
SHELL = bash
CPU_CORES = $(shell nproc)
# PIP_INSTALL = --editable .
# PYLINT_ARG =
# MYPY_ARG =
# PYTEST_ARG =
SHARED_DEV_REQUIREMENTS = \
black \
isort \
mypy \
pylint \
pytest \
pytest-cov \
pytest-xdist
VENV = venv
$(VENV):
python3 -m venv $(VENV)
source $(VENV)/bin/activate
pip install --upgrade pip setuptools wheel
pip install $(SHARED_DEV_REQUIREMENTS) $(PIP_INSTALL)
.PHONY: .format
.format: $(VENV)
source $(VENV)/bin/activate
black .
isort --profile black .
.PHONY: .pylint
.pylint: $(VENV)
source $(VENV)/bin/activate
pylint --output-format=colorized $(PYLINT_ARG) || true
.PHONY: .mypy
.mypy: $(VENV)
source $(VENV)/bin/activate
mypy $(MYPY_ARG) || true
.PHONY: .pytest
.pytest: $(VENV)
source venv/bin/activate
pytest -n $(CPU_CORES) --color=yes -v $(PYTEST_ARG)
.PHONY: .clean
.clean:
rm -Rf $(VENV)
| .ONESHELL:
SHELL = bash
CPU_CORES = $(shell nproc)
# PIP_INSTALL = --editable .
# PYLINT_ARG =
# MYPY_ARG =
# PYTEST_ARG =
SHARED_DEV_REQUIREMENTS = \
black \
isort \
mypy \
pylint \
pytest \
pytest-cov \
pytest-xdist
VENV = venv
$(VENV):
python3 -m venv $(VENV)
source $(VENV)/bin/activate
$(MAKE) install
install: venv
source $(VENV)/bin/activate
pip install --upgrade pip setuptools wheel
pip install $(SHARED_DEV_REQUIREMENTS) $(PIP_INSTALL)
.PHONY: .format
.format: $(VENV)
source $(VENV)/bin/activate
black .
isort --profile black .
.PHONY: .pylint
.pylint: $(VENV)
source $(VENV)/bin/activate
pylint --output-format=colorized $(PYLINT_ARG) || true
.PHONY: .mypy
.mypy: $(VENV)
source $(VENV)/bin/activate
mypy $(MYPY_ARG) || true
.PHONY: .pytest
.pytest: $(VENV)
source venv/bin/activate
pytest -n $(CPU_CORES) --color=yes -v $(PYTEST_ARG)
.PHONY: .clean
.clean:
rm -Rf $(VENV)
| Split venv setup and package install targets | Split venv setup and package install targets
This allow to reinstall the packages without
rebuilding the venv in case of dependencies change.
| Makefile | agpl-3.0 | LibreTime/libretime,LibreTime/libretime,LibreTime/libretime,LibreTime/libretime,LibreTime/libretime,LibreTime/libretime | makefile | ## Code Before:
.ONESHELL:
SHELL = bash
CPU_CORES = $(shell nproc)
# PIP_INSTALL = --editable .
# PYLINT_ARG =
# MYPY_ARG =
# PYTEST_ARG =
SHARED_DEV_REQUIREMENTS = \
black \
isort \
mypy \
pylint \
pytest \
pytest-cov \
pytest-xdist
VENV = venv
$(VENV):
python3 -m venv $(VENV)
source $(VENV)/bin/activate
pip install --upgrade pip setuptools wheel
pip install $(SHARED_DEV_REQUIREMENTS) $(PIP_INSTALL)
.PHONY: .format
.format: $(VENV)
source $(VENV)/bin/activate
black .
isort --profile black .
.PHONY: .pylint
.pylint: $(VENV)
source $(VENV)/bin/activate
pylint --output-format=colorized $(PYLINT_ARG) || true
.PHONY: .mypy
.mypy: $(VENV)
source $(VENV)/bin/activate
mypy $(MYPY_ARG) || true
.PHONY: .pytest
.pytest: $(VENV)
source venv/bin/activate
pytest -n $(CPU_CORES) --color=yes -v $(PYTEST_ARG)
.PHONY: .clean
.clean:
rm -Rf $(VENV)
## Instruction:
Split venv setup and package install targets
This allow to reinstall the packages without
rebuilding the venv in case of dependencies change.
## Code After:
.ONESHELL:
SHELL = bash
CPU_CORES = $(shell nproc)
# PIP_INSTALL = --editable .
# PYLINT_ARG =
# MYPY_ARG =
# PYTEST_ARG =
SHARED_DEV_REQUIREMENTS = \
black \
isort \
mypy \
pylint \
pytest \
pytest-cov \
pytest-xdist
VENV = venv
$(VENV):
python3 -m venv $(VENV)
source $(VENV)/bin/activate
$(MAKE) install
install: venv
source $(VENV)/bin/activate
pip install --upgrade pip setuptools wheel
pip install $(SHARED_DEV_REQUIREMENTS) $(PIP_INSTALL)
.PHONY: .format
.format: $(VENV)
source $(VENV)/bin/activate
black .
isort --profile black .
.PHONY: .pylint
.pylint: $(VENV)
source $(VENV)/bin/activate
pylint --output-format=colorized $(PYLINT_ARG) || true
.PHONY: .mypy
.mypy: $(VENV)
source $(VENV)/bin/activate
mypy $(MYPY_ARG) || true
.PHONY: .pytest
.pytest: $(VENV)
source venv/bin/activate
pytest -n $(CPU_CORES) --color=yes -v $(PYTEST_ARG)
.PHONY: .clean
.clean:
rm -Rf $(VENV)
| .ONESHELL:
SHELL = bash
CPU_CORES = $(shell nproc)
# PIP_INSTALL = --editable .
# PYLINT_ARG =
# MYPY_ARG =
# PYTEST_ARG =
SHARED_DEV_REQUIREMENTS = \
black \
isort \
mypy \
pylint \
pytest \
pytest-cov \
pytest-xdist
VENV = venv
$(VENV):
python3 -m venv $(VENV)
+ source $(VENV)/bin/activate
+ $(MAKE) install
+
+ install: venv
source $(VENV)/bin/activate
pip install --upgrade pip setuptools wheel
pip install $(SHARED_DEV_REQUIREMENTS) $(PIP_INSTALL)
.PHONY: .format
.format: $(VENV)
source $(VENV)/bin/activate
black .
isort --profile black .
.PHONY: .pylint
.pylint: $(VENV)
source $(VENV)/bin/activate
pylint --output-format=colorized $(PYLINT_ARG) || true
.PHONY: .mypy
.mypy: $(VENV)
source $(VENV)/bin/activate
mypy $(MYPY_ARG) || true
.PHONY: .pytest
.pytest: $(VENV)
source venv/bin/activate
pytest -n $(CPU_CORES) --color=yes -v $(PYTEST_ARG)
.PHONY: .clean
.clean:
rm -Rf $(VENV) | 4 | 0.08 | 4 | 0 |
a2388b5ddfc496b8aee0d41fc3648765d2613dd5 | .circleci/config.yml | .circleci/config.yml | version: 2
jobs:
build:
docker:
- image: circleci/python:3.6.1
environment:
FLASK_CONFIG: testing
TEST_DATABASE_URL: postgresql://ubuntu@localhost/circle_test?sslmode=disable
- image: circleci/postgres:9.6.5-alpine-ram
environment:
POSTGRES_USER: ubuntu
POSTGRES_DB: circle_test
POSTGRES_PASSWORD: ""
steps:
- checkout
- restore_cache:
key: deps1-{{ .Branch }}-{{ checksum "requirements.txt" }}
- run:
name: Install python libs
command: |
python3 -m venv venv
. venv/bin/activate
pip install -r requirements.txt
- save_cache:
key: deps1-{{ .Branch }}-{{ checksum "requirements.txt" }}
paths:
- "venv"
- run:
name: Run tox
command: |
. venv/bin/activate
tox
| version: 2
jobs:
build:
docker:
- image: circleci/python:3.6.1
environment:
SQLALCHEMY_DATABASE_URI: postgresql://ubuntu@localhost/circle_test
- image: circleci/postgres:9.6.5-alpine-ram
environment:
POSTGRES_USER: ubuntu
POSTGRES_DB: circle_test
POSTGRES_PASSWORD: ""
steps:
- checkout
- restore_cache:
key: deps1-{{ .Branch }}-{{ checksum "requirements.txt" }}
- run:
name: Install python libs
command: |
python3 -m venv venv
. venv/bin/activate
pip install -r requirements.txt
- save_cache:
key: deps1-{{ .Branch }}-{{ checksum "requirements.txt" }}
paths:
- "venv"
- run:
name: Run tox
command: |
. venv/bin/activate
tox
| Fix circle mistyped env varible | Fix circle mistyped env varible
| YAML | mit | danielSbastos/gistified,danielSbastos/gistified,danielSbastos/gistified | yaml | ## Code Before:
version: 2
jobs:
build:
docker:
- image: circleci/python:3.6.1
environment:
FLASK_CONFIG: testing
TEST_DATABASE_URL: postgresql://ubuntu@localhost/circle_test?sslmode=disable
- image: circleci/postgres:9.6.5-alpine-ram
environment:
POSTGRES_USER: ubuntu
POSTGRES_DB: circle_test
POSTGRES_PASSWORD: ""
steps:
- checkout
- restore_cache:
key: deps1-{{ .Branch }}-{{ checksum "requirements.txt" }}
- run:
name: Install python libs
command: |
python3 -m venv venv
. venv/bin/activate
pip install -r requirements.txt
- save_cache:
key: deps1-{{ .Branch }}-{{ checksum "requirements.txt" }}
paths:
- "venv"
- run:
name: Run tox
command: |
. venv/bin/activate
tox
## Instruction:
Fix circle mistyped env varible
## Code After:
version: 2
jobs:
build:
docker:
- image: circleci/python:3.6.1
environment:
SQLALCHEMY_DATABASE_URI: postgresql://ubuntu@localhost/circle_test
- image: circleci/postgres:9.6.5-alpine-ram
environment:
POSTGRES_USER: ubuntu
POSTGRES_DB: circle_test
POSTGRES_PASSWORD: ""
steps:
- checkout
- restore_cache:
key: deps1-{{ .Branch }}-{{ checksum "requirements.txt" }}
- run:
name: Install python libs
command: |
python3 -m venv venv
. venv/bin/activate
pip install -r requirements.txt
- save_cache:
key: deps1-{{ .Branch }}-{{ checksum "requirements.txt" }}
paths:
- "venv"
- run:
name: Run tox
command: |
. venv/bin/activate
tox
| version: 2
jobs:
build:
docker:
- image: circleci/python:3.6.1
environment:
- FLASK_CONFIG: testing
- TEST_DATABASE_URL: postgresql://ubuntu@localhost/circle_test?sslmode=disable
? ^ ^^ ^ ----------------
+ SQLALCHEMY_DATABASE_URI: postgresql://ubuntu@localhost/circle_test
? ^^^^^^^ ^^ ^
- image: circleci/postgres:9.6.5-alpine-ram
environment:
POSTGRES_USER: ubuntu
POSTGRES_DB: circle_test
POSTGRES_PASSWORD: ""
steps:
- checkout
- restore_cache:
key: deps1-{{ .Branch }}-{{ checksum "requirements.txt" }}
- run:
name: Install python libs
command: |
python3 -m venv venv
. venv/bin/activate
pip install -r requirements.txt
- save_cache:
key: deps1-{{ .Branch }}-{{ checksum "requirements.txt" }}
paths:
- "venv"
- run:
name: Run tox
command: |
. venv/bin/activate
tox | 3 | 0.09375 | 1 | 2 |
9a5ad19120f7aa610ca88ef2d0bc77d89b0efd83 | app/scripts/app/utils/data-store.js | app/scripts/app/utils/data-store.js | var DataStore = Ember.Object.extend({
init: function () {
this._super();
this.set('store', {});
},
addType: function (type) {
this.store[type] = [];
},
findById: function (type, id) {
var type = this.store[type],
model;
if (!type) {
throw new Error('There is no model of that type in the datastore!');
}
model = type.filter(function (model) {
return model.id && model.id === id;
});
return model;
},
findByKey: function (type, keyName, val) {
var type = this.store[type],
model;
if (!type) {
throw new Error('There is no model of that type in the datastore!');
}
model = type.filter(function (model) {
return model[key] && model[key] === val;
});
return model;
},
});
export default DataStore;
| import parseFields from 'client/utils/parse-fields-metadata';
var DataStore = Ember.Object.extend({
init: function () {
this._super();
this.set('store', Ember.Object.create());
},
addType: function (type) {
this.store[type] = Ember.ArrayProxy.create({ content: Ember.A() });
},
load: function (type, payload) {
if (!this.store[type]) {
throw new Error('There is no model of type ' + type + ' in the datastore!');
}
if (typeof payload !== 'object') {
throw new Error('Payload for type ' + type + ' was not an object!', payload);
}
// turn `fields` metadata field from a string like "[]" into a JSON object
parseFields(payload);
if (!this.store[type].length) {
this.store[type].pushObjects(payload);
return;
}
if (!Array.isArray(payload) && !this.store[type].findBy('id', payload.id)) {
this.store[type].pushObject(payload);
return;
}
// now we know both payload and the modelType in the store are non-zero-length arrays
// we need to check for collisions and update those that exist, and insert those that don't
},
binarySearch: function (sortedArray, options) {
},
all: function (type, key, val) {
var modelType = this.store[type];
if (!modelType) {
throw new Error('There is no model of type ' + type + ' in the datastore!');
}
if (!key || !val) {
return modelType;
}
return modelType.filterBy(key, val);
},
findById: function (type, id) {
var modelType = this.store[type];
if (!modelType) {
throw new Error('There is no model of type ' + type + ' in the datastore!');
}
return modelType.findBy('id', id);
},
findByKey: function (type, keyName, val) {
var modelType = this.store[type];
if (!modelType) {
throw new Error('There is no model of type ' + type + ' in the datastore!');
}
return modelType.filterBy(keyName, val);
},
});
export default DataStore;
| Build up how DataStore works with and transforms its data types | Build up how DataStore works with and transforms its data types
| JavaScript | mit | darvelo/wishlist,darvelo/wishlist | javascript | ## Code Before:
var DataStore = Ember.Object.extend({
init: function () {
this._super();
this.set('store', {});
},
addType: function (type) {
this.store[type] = [];
},
findById: function (type, id) {
var type = this.store[type],
model;
if (!type) {
throw new Error('There is no model of that type in the datastore!');
}
model = type.filter(function (model) {
return model.id && model.id === id;
});
return model;
},
findByKey: function (type, keyName, val) {
var type = this.store[type],
model;
if (!type) {
throw new Error('There is no model of that type in the datastore!');
}
model = type.filter(function (model) {
return model[key] && model[key] === val;
});
return model;
},
});
export default DataStore;
## Instruction:
Build up how DataStore works with and transforms its data types
## Code After:
import parseFields from 'client/utils/parse-fields-metadata';
var DataStore = Ember.Object.extend({
init: function () {
this._super();
this.set('store', Ember.Object.create());
},
addType: function (type) {
this.store[type] = Ember.ArrayProxy.create({ content: Ember.A() });
},
load: function (type, payload) {
if (!this.store[type]) {
throw new Error('There is no model of type ' + type + ' in the datastore!');
}
if (typeof payload !== 'object') {
throw new Error('Payload for type ' + type + ' was not an object!', payload);
}
// turn `fields` metadata field from a string like "[]" into a JSON object
parseFields(payload);
if (!this.store[type].length) {
this.store[type].pushObjects(payload);
return;
}
if (!Array.isArray(payload) && !this.store[type].findBy('id', payload.id)) {
this.store[type].pushObject(payload);
return;
}
// now we know both payload and the modelType in the store are non-zero-length arrays
// we need to check for collisions and update those that exist, and insert those that don't
},
binarySearch: function (sortedArray, options) {
},
all: function (type, key, val) {
var modelType = this.store[type];
if (!modelType) {
throw new Error('There is no model of type ' + type + ' in the datastore!');
}
if (!key || !val) {
return modelType;
}
return modelType.filterBy(key, val);
},
findById: function (type, id) {
var modelType = this.store[type];
if (!modelType) {
throw new Error('There is no model of type ' + type + ' in the datastore!');
}
return modelType.findBy('id', id);
},
findByKey: function (type, keyName, val) {
var modelType = this.store[type];
if (!modelType) {
throw new Error('There is no model of type ' + type + ' in the datastore!');
}
return modelType.filterBy(keyName, val);
},
});
export default DataStore;
| + import parseFields from 'client/utils/parse-fields-metadata';
+
var DataStore = Ember.Object.extend({
init: function () {
this._super();
- this.set('store', {});
+ this.set('store', Ember.Object.create());
},
addType: function (type) {
+ this.store[type] = Ember.ArrayProxy.create({ content: Ember.A() });
+ },
+
+ load: function (type, payload) {
- this.store[type] = [];
? ^^^^^
+ if (!this.store[type]) {
? +++++ + ^
+ throw new Error('There is no model of type ' + type + ' in the datastore!');
+ }
+
+ if (typeof payload !== 'object') {
+ throw new Error('Payload for type ' + type + ' was not an object!', payload);
+ }
+
+ // turn `fields` metadata field from a string like "[]" into a JSON object
+ parseFields(payload);
+
+ if (!this.store[type].length) {
+ this.store[type].pushObjects(payload);
+ return;
+ }
+
+ if (!Array.isArray(payload) && !this.store[type].findBy('id', payload.id)) {
+ this.store[type].pushObject(payload);
+ return;
+ }
+
+ // now we know both payload and the modelType in the store are non-zero-length arrays
+ // we need to check for collisions and update those that exist, and insert those that don't
+
+ },
+
+ binarySearch: function (sortedArray, options) {
+
+ },
+
+ all: function (type, key, val) {
+ var modelType = this.store[type];
+
+ if (!modelType) {
+ throw new Error('There is no model of type ' + type + ' in the datastore!');
+ }
+
+ if (!key || !val) {
+ return modelType;
+ }
+
+ return modelType.filterBy(key, val);
},
findById: function (type, id) {
- var type = this.store[type],
? ^ ^
+ var modelType = this.store[type];
? ^^^^^^ ^
- model;
- if (!type) {
? ^
+ if (!modelType) {
? ^^^^^^
- throw new Error('There is no model of that type in the datastore!');
? ^^^
+ throw new Error('There is no model of type ' + type + ' in the datastore!');
? ^^^^^^^ ++++
}
+ return modelType.findBy('id', id);
- model = type.filter(function (model) {
- return model.id && model.id === id;
- });
-
- return model;
},
findByKey: function (type, keyName, val) {
- var type = this.store[type],
? ^ ^
+ var modelType = this.store[type];
? ^^^^^^ ^
- model;
- if (!type) {
? ^
+ if (!modelType) {
? ^^^^^^
- throw new Error('There is no model of that type in the datastore!');
? ^^^
+ throw new Error('There is no model of type ' + type + ' in the datastore!');
? ^^^^^^^ ++++
}
+ return modelType.filterBy(keyName, val);
- model = type.filter(function (model) {
- return model[key] && model[key] === val;
- });
-
- return model;
},
});
export default DataStore; | 77 | 1.833333 | 57 | 20 |
b6e22554721824e601bc667ff3221f978cc87b1a | qml/common/AsyncImage.qml | qml/common/AsyncImage.qml | // Copyright (c) 2015 Piotr Tworek. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE.YTPlayer file.
import QtQuick 2.0
import Sailfish.Silica 1.0
Image {
fillMode: Image.PreserveAspectCrop
property alias indicatorSize: indicator.size
BusyIndicator {
id: indicator
size: BusyIndicatorSize.Small
anchors.centerIn: parent
running: parent.status === Image.Loading
}
Rectangle {
anchors.fill: parent
color: Theme.secondaryHighlightColor
visible: status === Image.Error
Label {
anchors.centerIn: parent
font.pixelSize: Theme.fontSizeExtraSmall
//: Label shown where video thumbnail is not valid, can't be loaded.
//: Should be very short 8-10 characters max.
//% "No image"
text: qsTrId("ytplayer-label-broken-image")
color: Theme.primaryColor
}
}
}
| // Copyright (c) 2015 Piotr Tworek. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE.YTPlayer file.
import QtQuick 2.0
import Sailfish.Silica 1.0
Image {
fillMode: Image.PreserveAspectCrop
property alias indicatorSize: indicator.size
BusyIndicator {
id: indicator
size: BusyIndicatorSize.Small
anchors.centerIn: parent
running: parent.status === Image.Loading
}
Rectangle {
anchors.fill: parent
color: Theme.secondaryHighlightColor
visible: status === Image.Error
Label {
anchors.centerIn: parent
width: parent.width
maximumLineCount: 1
font.pixelSize: Theme.fontSizeExtraSmall
horizontalAlignment: Text.AlignHCenter
//: Label shown where video thumbnail is not valid, can't be loaded.
//: Should be very short 8-10 characters max.
//% "No image"
text: qsTrId("ytplayer-label-broken-image")
color: Theme.primaryColor
}
}
}
| Add maximum width to no-image placeholder. | Add maximum width to no-image placeholder.
| QML | unlicense | tworaz/sailfish-ytplayer,tworaz/sailfish-ytplayer,tworaz/sailfish-ytplayer,tworaz/sailfish-ytplayer,tworaz/sailfish-ytplayer | qml | ## Code Before:
// Copyright (c) 2015 Piotr Tworek. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE.YTPlayer file.
import QtQuick 2.0
import Sailfish.Silica 1.0
Image {
fillMode: Image.PreserveAspectCrop
property alias indicatorSize: indicator.size
BusyIndicator {
id: indicator
size: BusyIndicatorSize.Small
anchors.centerIn: parent
running: parent.status === Image.Loading
}
Rectangle {
anchors.fill: parent
color: Theme.secondaryHighlightColor
visible: status === Image.Error
Label {
anchors.centerIn: parent
font.pixelSize: Theme.fontSizeExtraSmall
//: Label shown where video thumbnail is not valid, can't be loaded.
//: Should be very short 8-10 characters max.
//% "No image"
text: qsTrId("ytplayer-label-broken-image")
color: Theme.primaryColor
}
}
}
## Instruction:
Add maximum width to no-image placeholder.
## Code After:
// Copyright (c) 2015 Piotr Tworek. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE.YTPlayer file.
import QtQuick 2.0
import Sailfish.Silica 1.0
Image {
fillMode: Image.PreserveAspectCrop
property alias indicatorSize: indicator.size
BusyIndicator {
id: indicator
size: BusyIndicatorSize.Small
anchors.centerIn: parent
running: parent.status === Image.Loading
}
Rectangle {
anchors.fill: parent
color: Theme.secondaryHighlightColor
visible: status === Image.Error
Label {
anchors.centerIn: parent
width: parent.width
maximumLineCount: 1
font.pixelSize: Theme.fontSizeExtraSmall
horizontalAlignment: Text.AlignHCenter
//: Label shown where video thumbnail is not valid, can't be loaded.
//: Should be very short 8-10 characters max.
//% "No image"
text: qsTrId("ytplayer-label-broken-image")
color: Theme.primaryColor
}
}
}
| // Copyright (c) 2015 Piotr Tworek. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE.YTPlayer file.
import QtQuick 2.0
import Sailfish.Silica 1.0
Image {
fillMode: Image.PreserveAspectCrop
property alias indicatorSize: indicator.size
BusyIndicator {
id: indicator
size: BusyIndicatorSize.Small
anchors.centerIn: parent
running: parent.status === Image.Loading
}
Rectangle {
anchors.fill: parent
color: Theme.secondaryHighlightColor
visible: status === Image.Error
Label {
anchors.centerIn: parent
+ width: parent.width
+ maximumLineCount: 1
font.pixelSize: Theme.fontSizeExtraSmall
+ horizontalAlignment: Text.AlignHCenter
//: Label shown where video thumbnail is not valid, can't be loaded.
//: Should be very short 8-10 characters max.
//% "No image"
text: qsTrId("ytplayer-label-broken-image")
color: Theme.primaryColor
}
}
} | 3 | 0.085714 | 3 | 0 |
284f4eefe9245aac42675aec588460abd37073f6 | test/resque_failure_redis_test.rb | test/resque_failure_redis_test.rb | require 'test_helper'
require 'resque/failure/redis'
describe "Resque::Failure::Redis" do
before do
@bad_string = [39, 250, 141, 168, 138, 191, 52, 211, 159, 86, 93, 95, 39].map { |c| c.chr }.join
exception = StandardError.exception(@bad_string)
worker = Resque::Worker.new(:test)
queue = "queue"
payload = { "class" => Object, "args" => 3 }
@redis_backend = Resque::Failure::Redis.new(exception, worker, queue, payload)
end
it 'cleans up bad strings before saving the failure, in order to prevent errors on the resque UI' do
# test assumption: the bad string should not be able to round trip though JSON
assert_raises(MultiJson::DecodeError, JSON::GeneratorError) {
MultiJson.decode(MultiJson.encode(@bad_string))
}
@redis_backend.save
Resque::Failure::Redis.all # should not raise an error
end
end
| require 'test_helper'
require 'resque/failure/redis'
require 'json'
describe "Resque::Failure::Redis" do
before do
@bad_string = [39, 250, 141, 168, 138, 191, 52, 211, 159, 86, 93, 95, 39].map { |c| c.chr }.join
exception = StandardError.exception(@bad_string)
worker = Resque::Worker.new(:test)
queue = "queue"
payload = { "class" => Object, "args" => 3 }
@redis_backend = Resque::Failure::Redis.new(exception, worker, queue, payload)
end
it 'cleans up bad strings before saving the failure, in order to prevent errors on the resque UI' do
# test assumption: the bad string should not be able to round trip though JSON
assert_raises(MultiJson::DecodeError, JSON::GeneratorError) {
MultiJson.decode(MultiJson.encode(@bad_string))
}
@redis_backend.save
Resque::Failure::Redis.all # should not raise an error
end
end
| Fix the build on MRI. | Fix the build on MRI.
My fix for JRuby broke MRI :(
| Ruby | mit | stitchfix/resque,timo-p/resque,begriffs/resque,ktheory/resque,begriffs/resque,ychaim/resque,sideci-sample/sideci-sample-resque,stitchfix/resque,kangkot/resque,tylerdooling/resque,lileeyao/resque,partap/resque,ducthanh/resque,lukeasrodgers/resque,m0dd3r/resque,vladimirich/resque,igorjancic/resque,bhighley/resque,fmitech/resque,delectable/resque,fmitech/resque,lileeyao/resque,delectable/resque,ktheory/resque,lukeasrodgers/resque,hoffmanc/resque,m0dd3r/resque,vladimirich/resque,AlexVPopov/resque,rutikasb/resque,timo-p/resque,l4u/resque,paran0ids0ul/resque,bonekost/resque,ryan-drake-stoker/resque,hoffmanc/resque,redhotpenguin/resque,eileencodes/resque,ngpestelos/resque,sideci-sample/sideci-sample-resque,paran0ids0ul/resque,l4u/resque,jonatas/resque,jacknguyen/resque,sufery/resque,vietnt/resque,sufery/resque,igorjancic/resque,jacknguyen/resque,tylerdooling/resque | ruby | ## Code Before:
require 'test_helper'
require 'resque/failure/redis'
describe "Resque::Failure::Redis" do
before do
@bad_string = [39, 250, 141, 168, 138, 191, 52, 211, 159, 86, 93, 95, 39].map { |c| c.chr }.join
exception = StandardError.exception(@bad_string)
worker = Resque::Worker.new(:test)
queue = "queue"
payload = { "class" => Object, "args" => 3 }
@redis_backend = Resque::Failure::Redis.new(exception, worker, queue, payload)
end
it 'cleans up bad strings before saving the failure, in order to prevent errors on the resque UI' do
# test assumption: the bad string should not be able to round trip though JSON
assert_raises(MultiJson::DecodeError, JSON::GeneratorError) {
MultiJson.decode(MultiJson.encode(@bad_string))
}
@redis_backend.save
Resque::Failure::Redis.all # should not raise an error
end
end
## Instruction:
Fix the build on MRI.
My fix for JRuby broke MRI :(
## Code After:
require 'test_helper'
require 'resque/failure/redis'
require 'json'
describe "Resque::Failure::Redis" do
before do
@bad_string = [39, 250, 141, 168, 138, 191, 52, 211, 159, 86, 93, 95, 39].map { |c| c.chr }.join
exception = StandardError.exception(@bad_string)
worker = Resque::Worker.new(:test)
queue = "queue"
payload = { "class" => Object, "args" => 3 }
@redis_backend = Resque::Failure::Redis.new(exception, worker, queue, payload)
end
it 'cleans up bad strings before saving the failure, in order to prevent errors on the resque UI' do
# test assumption: the bad string should not be able to round trip though JSON
assert_raises(MultiJson::DecodeError, JSON::GeneratorError) {
MultiJson.decode(MultiJson.encode(@bad_string))
}
@redis_backend.save
Resque::Failure::Redis.all # should not raise an error
end
end
| require 'test_helper'
require 'resque/failure/redis'
+
+ require 'json'
describe "Resque::Failure::Redis" do
before do
@bad_string = [39, 250, 141, 168, 138, 191, 52, 211, 159, 86, 93, 95, 39].map { |c| c.chr }.join
exception = StandardError.exception(@bad_string)
worker = Resque::Worker.new(:test)
queue = "queue"
payload = { "class" => Object, "args" => 3 }
@redis_backend = Resque::Failure::Redis.new(exception, worker, queue, payload)
end
it 'cleans up bad strings before saving the failure, in order to prevent errors on the resque UI' do
# test assumption: the bad string should not be able to round trip though JSON
assert_raises(MultiJson::DecodeError, JSON::GeneratorError) {
MultiJson.decode(MultiJson.encode(@bad_string))
}
@redis_backend.save
Resque::Failure::Redis.all # should not raise an error
end
end | 2 | 0.086957 | 2 | 0 |
ae4426acb1dd2301ffe6c63a790fb07a9d8be63f | project/src/styl/new-game.styl | project/src/styl/new-game.styl | .game_form_popup
background bglight
width 85%
left 7.5%
.game_form
padding-top 10px
fieldset
padding 0px 20px
button[type=submit]
width 100%
border-radius 0 0 5px 5px
margin 30px 0 0
background gold
color #FFF
font-size 0.85em
text-transform uppercase
#new_game_form
.nice-radio
input.radio.computer:empty + label
text-align right
border-radius 0 25px 25px 0
input.radio.human:empty + label
border-radius 30px 0 0 30px
input.radio.human:empty + label:after
left auto
right 0
.rating_range
margin-top 10px
.select_input
margin-top 5px
select
padding-left 60px
| .game_form_popup
background bglight
width 85%
left 7.5%
.game_form
padding-top 10px
fieldset
padding 0px 20px
button[type=submit]
width 100%
border-radius 0 0 5px 5px
margin 30px 0 0
background gold
color #FFF
font-size 0.85em
text-transform uppercase
&.active
background lighten(gold, 10%)
#new_game_form
.nice-radio
input.radio.computer:empty + label
text-align right
border-radius 0 25px 25px 0
input.radio.human:empty + label
border-radius 30px 0 0 30px
input.radio.human:empty + label:after
left auto
right 0
.rating_range
margin-top 10px
.select_input
margin-top 5px
select
padding-left 60px
| Add active style to new game button | Add active style to new game button
| Stylus | mit | garawaa/lichobile,btrent/lichobile,btrent/lichobile,garawaa/lichobile,btrent/lichobile,btrent/lichobile,garawaa/lichobile | stylus | ## Code Before:
.game_form_popup
background bglight
width 85%
left 7.5%
.game_form
padding-top 10px
fieldset
padding 0px 20px
button[type=submit]
width 100%
border-radius 0 0 5px 5px
margin 30px 0 0
background gold
color #FFF
font-size 0.85em
text-transform uppercase
#new_game_form
.nice-radio
input.radio.computer:empty + label
text-align right
border-radius 0 25px 25px 0
input.radio.human:empty + label
border-radius 30px 0 0 30px
input.radio.human:empty + label:after
left auto
right 0
.rating_range
margin-top 10px
.select_input
margin-top 5px
select
padding-left 60px
## Instruction:
Add active style to new game button
## Code After:
.game_form_popup
background bglight
width 85%
left 7.5%
.game_form
padding-top 10px
fieldset
padding 0px 20px
button[type=submit]
width 100%
border-radius 0 0 5px 5px
margin 30px 0 0
background gold
color #FFF
font-size 0.85em
text-transform uppercase
&.active
background lighten(gold, 10%)
#new_game_form
.nice-radio
input.radio.computer:empty + label
text-align right
border-radius 0 25px 25px 0
input.radio.human:empty + label
border-radius 30px 0 0 30px
input.radio.human:empty + label:after
left auto
right 0
.rating_range
margin-top 10px
.select_input
margin-top 5px
select
padding-left 60px
| .game_form_popup
background bglight
width 85%
left 7.5%
.game_form
padding-top 10px
fieldset
padding 0px 20px
button[type=submit]
width 100%
border-radius 0 0 5px 5px
margin 30px 0 0
background gold
color #FFF
font-size 0.85em
text-transform uppercase
+ &.active
+ background lighten(gold, 10%)
#new_game_form
.nice-radio
input.radio.computer:empty + label
text-align right
border-radius 0 25px 25px 0
input.radio.human:empty + label
border-radius 30px 0 0 30px
input.radio.human:empty + label:after
left auto
right 0
.rating_range
margin-top 10px
.select_input
margin-top 5px
select
padding-left 60px
| 2 | 0.054054 | 2 | 0 |
ff489b1541f896025a0c630be6abe2d23843ec36 | examples/05_alternative_language.py | examples/05_alternative_language.py |
from pyhmsa.datafile import DataFile
from pyhmsa.type.language import langstr
datafile = DataFile()
author = langstr('Fyodor Dostoyevsky', {'ru': u'Фёдор Миха́йлович Достое́вский'})
datafile.header.author = author
print(datafile.header.author.alternatives['ru']) # Returns ... |
from pyhmsa.datafile import DataFile
from pyhmsa.type.language import langstr
datafile = DataFile()
author = langstr('Wilhelm Conrad Roentgen', {'de': u'Wilhelm Conrad Röntgen'})
datafile.header.author = author
print(datafile.header.author.alternatives['de']) # Returns ... | Replace name in alternative language to prevent compilation problems with LaTeX | Replace name in alternative language to prevent compilation problems
with LaTeX | Python | mit | pyhmsa/pyhmsa | python | ## Code Before:
from pyhmsa.datafile import DataFile
from pyhmsa.type.language import langstr
datafile = DataFile()
author = langstr('Fyodor Dostoyevsky', {'ru': u'Фёдор Миха́йлович Достое́вский'})
datafile.header.author = author
print(datafile.header.author.alternatives['ru']) # Returns ...
## Instruction:
Replace name in alternative language to prevent compilation problems
with LaTeX
## Code After:
from pyhmsa.datafile import DataFile
from pyhmsa.type.language import langstr
datafile = DataFile()
author = langstr('Wilhelm Conrad Roentgen', {'de': u'Wilhelm Conrad Röntgen'})
datafile.header.author = author
print(datafile.header.author.alternatives['de']) # Returns ... |
from pyhmsa.datafile import DataFile
from pyhmsa.type.language import langstr
datafile = DataFile()
- author = langstr('Fyodor Dostoyevsky', {'ru': u'Фёдор Миха́йлович Достое́вский'})
+ author = langstr('Wilhelm Conrad Roentgen', {'de': u'Wilhelm Conrad Röntgen'})
datafile.header.author = author
- print(datafile.header.author.alternatives['ru']) # Returns ...
? ^^
+ print(datafile.header.author.alternatives['de']) # Returns ...
? ^^
| 4 | 0.444444 | 2 | 2 |
dd06c6459ab9b6a7536e9de8c1b0daeefbd888ef | azure-pipelines-release.yml | azure-pipelines-release.yml |
trigger:
tags:
include:
- v*.*.*
pool:
vmImage: 'Ubuntu-16.04'
steps:
- task: NodeTool@0
inputs:
versionSpec: '10.x'
displayName: 'install Node.js'
- script: |
npm install
npm run build_prod
npm run testsingle
npm run testcafe_ci
displayName: 'npm install and build'
- script: |
npm run doc_gen
displayName: 'generate docs'
- script: |
chmod +x ./deploy_npm.sh && ./deploy_npm.sh
displayName: 'npm publish'
- script: |
npm run doc_update
displayName: 'update docs'
|
trigger:
tags:
include:
- v*.*.*
pool:
vmImage: 'Ubuntu-16.04'
steps:
- task: NodeTool@0
inputs:
versionSpec: '10.x'
displayName: 'install Node.js'
- script: |
npm install
npm run build_prod
npm run testsingle
npm run testcafe_ci
displayName: 'npm install and build'
- script: |
npm run doc_gen
displayName: 'generate docs'
- task: Npm@1
inputs:
command: 'publish'
workingDir: 'packages/survey-angular'
publishEndpoint: 'NPM Registry'
verbose: true
- task: Npm@2
inputs:
command: 'publish'
workingDir: 'packages/survey-react'
publishEndpoint: 'NPM Registry'
verbose: true
- task: Npm@3
inputs:
command: 'publish'
workingDir: 'packages/survey-vue'
publishEndpoint: 'NPM Registry'
verbose: true
- task: Npm@4
inputs:
command: 'publish'
workingDir: 'packages/survey-knockout'
publishEndpoint: 'NPM Registry'
verbose: true
- task: Npm@5
inputs:
command: 'publish'
workingDir: 'packages/survey-jquery'
publishEndpoint: 'NPM Registry'
verbose: true
- task: Npm@6
inputs:
command: 'publish'
workingDir: 'packages/survey-core'
publishEndpoint: 'NPM Registry'
verbose: true
- script: |
npm run doc_update
displayName: 'update docs'
| Add npm tasks in azure pipeline release | Add npm tasks in azure pipeline release
| YAML | mit | andrewtelnov/surveyjs,andrewtelnov/surveyjs,surveyjs/surveyjs,surveyjs/surveyjs,surveyjs/surveyjs,andrewtelnov/surveyjs,surveyjs/surveyjs | yaml | ## Code Before:
trigger:
tags:
include:
- v*.*.*
pool:
vmImage: 'Ubuntu-16.04'
steps:
- task: NodeTool@0
inputs:
versionSpec: '10.x'
displayName: 'install Node.js'
- script: |
npm install
npm run build_prod
npm run testsingle
npm run testcafe_ci
displayName: 'npm install and build'
- script: |
npm run doc_gen
displayName: 'generate docs'
- script: |
chmod +x ./deploy_npm.sh && ./deploy_npm.sh
displayName: 'npm publish'
- script: |
npm run doc_update
displayName: 'update docs'
## Instruction:
Add npm tasks in azure pipeline release
## Code After:
trigger:
tags:
include:
- v*.*.*
pool:
vmImage: 'Ubuntu-16.04'
steps:
- task: NodeTool@0
inputs:
versionSpec: '10.x'
displayName: 'install Node.js'
- script: |
npm install
npm run build_prod
npm run testsingle
npm run testcafe_ci
displayName: 'npm install and build'
- script: |
npm run doc_gen
displayName: 'generate docs'
- task: Npm@1
inputs:
command: 'publish'
workingDir: 'packages/survey-angular'
publishEndpoint: 'NPM Registry'
verbose: true
- task: Npm@2
inputs:
command: 'publish'
workingDir: 'packages/survey-react'
publishEndpoint: 'NPM Registry'
verbose: true
- task: Npm@3
inputs:
command: 'publish'
workingDir: 'packages/survey-vue'
publishEndpoint: 'NPM Registry'
verbose: true
- task: Npm@4
inputs:
command: 'publish'
workingDir: 'packages/survey-knockout'
publishEndpoint: 'NPM Registry'
verbose: true
- task: Npm@5
inputs:
command: 'publish'
workingDir: 'packages/survey-jquery'
publishEndpoint: 'NPM Registry'
verbose: true
- task: Npm@6
inputs:
command: 'publish'
workingDir: 'packages/survey-core'
publishEndpoint: 'NPM Registry'
verbose: true
- script: |
npm run doc_update
displayName: 'update docs'
|
trigger:
tags:
include:
- v*.*.*
pool:
vmImage: 'Ubuntu-16.04'
steps:
- task: NodeTool@0
inputs:
versionSpec: '10.x'
displayName: 'install Node.js'
- script: |
npm install
npm run build_prod
npm run testsingle
npm run testcafe_ci
displayName: 'npm install and build'
- script: |
npm run doc_gen
displayName: 'generate docs'
- - script: |
- chmod +x ./deploy_npm.sh && ./deploy_npm.sh
- displayName: 'npm publish'
+ - task: Npm@1
+ inputs:
+ command: 'publish'
+ workingDir: 'packages/survey-angular'
+ publishEndpoint: 'NPM Registry'
+ verbose: true
+
+ - task: Npm@2
+ inputs:
+ command: 'publish'
+ workingDir: 'packages/survey-react'
+ publishEndpoint: 'NPM Registry'
+ verbose: true
+
+ - task: Npm@3
+ inputs:
+ command: 'publish'
+ workingDir: 'packages/survey-vue'
+ publishEndpoint: 'NPM Registry'
+ verbose: true
+
+ - task: Npm@4
+ inputs:
+ command: 'publish'
+ workingDir: 'packages/survey-knockout'
+ publishEndpoint: 'NPM Registry'
+ verbose: true
+
+ - task: Npm@5
+ inputs:
+ command: 'publish'
+ workingDir: 'packages/survey-jquery'
+ publishEndpoint: 'NPM Registry'
+ verbose: true
+
+ - task: Npm@6
+ inputs:
+ command: 'publish'
+ workingDir: 'packages/survey-core'
+ publishEndpoint: 'NPM Registry'
+ verbose: true
+
- script: |
npm run doc_update
displayName: 'update docs' | 45 | 1.363636 | 42 | 3 |
38a7a113b1e9dec6756a90cb80d770e37b321d28 | css/jquery.fileupload-ui.css | css/jquery.fileupload-ui.css | @charset 'UTF-8';
/*
* jQuery File Upload UI Plugin CSS 6.2
* https://github.com/blueimp/jQuery-File-Upload
*
* Copyright 2010, Sebastian Tschan
* https://blueimp.net
*
* Licensed under the MIT license:
* http://www.opensource.org/licenses/MIT
*/
.fileinput-button {
position: relative;
overflow: hidden;
float: left;
margin-right: 4px;
}
.fileinput-button input {
position: absolute;
top: 0;
right: 0;
margin: 0;
border: solid transparent;
border-width: 0 0 100px 200px;
opacity: 0;
filter: alpha(opacity=0);
-moz-transform: translate(-300px, 0) scale(4);
direction: ltr;
cursor: pointer;
}
.fileupload-buttonbar button {
margin-bottom: 5px;
}
.files .progress {
width: 200px;
}
.progress-animated .bar {
background: url(../img/progressbar.gif);
filter: none;
}
| @charset 'UTF-8';
/*
* jQuery File Upload UI Plugin CSS 6.2
* https://github.com/blueimp/jQuery-File-Upload
*
* Copyright 2010, Sebastian Tschan
* https://blueimp.net
*
* Licensed under the MIT license:
* http://www.opensource.org/licenses/MIT
*/
.fileinput-button {
position: relative;
overflow: hidden;
float: left;
margin-right: 4px;
}
.fileinput-button input {
position: absolute;
top: 0;
right: 0;
margin: 0;
border: solid transparent;
border-width: 0 0 100px 200px;
opacity: 0;
filter: alpha(opacity=0);
-moz-transform: translate(-300px, 0) scale(4);
direction: ltr;
cursor: pointer;
}
.fileupload-buttonbar .btn,
.fileupload-buttonbar .toggle {
margin-bottom: 5px;
}
.files .progress {
width: 200px;
}
.progress-animated .bar {
background: url(../img/progressbar.gif);
filter: none;
}
| Add bottom margin to both buttons and inputs. | Add bottom margin to both buttons and inputs.
| CSS | apache-2.0 | Verde1705/verde1705.github.io,GrizliK1988/PhotoWidget,blueimp/jQuery-File-Upload,blueimp/jQuery-File-Upload,blueimp/jQuery-File-Upload,blueimp/jQuery-File-Upload,Verde1705/verde1705.github.io,blueimp/jQuery-File-Upload,blueimp/jQuery-File-Upload | css | ## Code Before:
@charset 'UTF-8';
/*
* jQuery File Upload UI Plugin CSS 6.2
* https://github.com/blueimp/jQuery-File-Upload
*
* Copyright 2010, Sebastian Tschan
* https://blueimp.net
*
* Licensed under the MIT license:
* http://www.opensource.org/licenses/MIT
*/
.fileinput-button {
position: relative;
overflow: hidden;
float: left;
margin-right: 4px;
}
.fileinput-button input {
position: absolute;
top: 0;
right: 0;
margin: 0;
border: solid transparent;
border-width: 0 0 100px 200px;
opacity: 0;
filter: alpha(opacity=0);
-moz-transform: translate(-300px, 0) scale(4);
direction: ltr;
cursor: pointer;
}
.fileupload-buttonbar button {
margin-bottom: 5px;
}
.files .progress {
width: 200px;
}
.progress-animated .bar {
background: url(../img/progressbar.gif);
filter: none;
}
## Instruction:
Add bottom margin to both buttons and inputs.
## Code After:
@charset 'UTF-8';
/*
* jQuery File Upload UI Plugin CSS 6.2
* https://github.com/blueimp/jQuery-File-Upload
*
* Copyright 2010, Sebastian Tschan
* https://blueimp.net
*
* Licensed under the MIT license:
* http://www.opensource.org/licenses/MIT
*/
.fileinput-button {
position: relative;
overflow: hidden;
float: left;
margin-right: 4px;
}
.fileinput-button input {
position: absolute;
top: 0;
right: 0;
margin: 0;
border: solid transparent;
border-width: 0 0 100px 200px;
opacity: 0;
filter: alpha(opacity=0);
-moz-transform: translate(-300px, 0) scale(4);
direction: ltr;
cursor: pointer;
}
.fileupload-buttonbar .btn,
.fileupload-buttonbar .toggle {
margin-bottom: 5px;
}
.files .progress {
width: 200px;
}
.progress-animated .bar {
background: url(../img/progressbar.gif);
filter: none;
}
| @charset 'UTF-8';
/*
* jQuery File Upload UI Plugin CSS 6.2
* https://github.com/blueimp/jQuery-File-Upload
*
* Copyright 2010, Sebastian Tschan
* https://blueimp.net
*
* Licensed under the MIT license:
* http://www.opensource.org/licenses/MIT
*/
.fileinput-button {
position: relative;
overflow: hidden;
float: left;
margin-right: 4px;
}
.fileinput-button input {
position: absolute;
top: 0;
right: 0;
margin: 0;
border: solid transparent;
border-width: 0 0 100px 200px;
opacity: 0;
filter: alpha(opacity=0);
-moz-transform: translate(-300px, 0) scale(4);
direction: ltr;
cursor: pointer;
}
- .fileupload-buttonbar button {
? - -- ^^
+ .fileupload-buttonbar .btn,
? + ^
+ .fileupload-buttonbar .toggle {
margin-bottom: 5px;
}
.files .progress {
width: 200px;
}
.progress-animated .bar {
background: url(../img/progressbar.gif);
filter: none;
} | 3 | 0.073171 | 2 | 1 |
053fe2a5b1f0b3e46a2d0b144e1403074970e819 | src/main/java/org/jtrfp/trcl/Renderable.java | src/main/java/org/jtrfp/trcl/Renderable.java | /*******************************************************************************
* This file is part of TERMINAL RECALL
* Copyright (c) 2012-2014 Chuck Ritola
* Part of the jTRFP.org project
* All rights reserved. This program and the accompanying materials
* are made available under the terms of the GNU Public License v3.0
* which accompanies this distribution, and is available at
* http://www.gnu.org/licenses/gpl.html
*
* Contributors:
* chuck - initial API and implementation
******************************************************************************/
package org.jtrfp.trcl;
import java.nio.ByteBuffer;
public interface Renderable {
ByteBuffer getOpaqueObjectDefinitionAddresses();
ByteBuffer getTransparentObjectDefinitionAddresses();
void updateStateToGPU();
}
| /*******************************************************************************
* This file is part of TERMINAL RECALL
* Copyright (c) 2012-2014 Chuck Ritola
* Part of the jTRFP.org project
* All rights reserved. This program and the accompanying materials
* are made available under the terms of the GNU Public License v3.0
* which accompanies this distribution, and is available at
* http://www.gnu.org/licenses/gpl.html
*
* Contributors:
* chuck - initial API and implementation
******************************************************************************/
package org.jtrfp.trcl;
import java.util.Collection;
import org.jtrfp.trcl.mem.VEC4Address;
public interface Renderable {
Collection<VEC4Address> getOpaqueObjectDefinitionAddresses();
Collection<VEC4Address> getTransparentObjectDefinitionAddresses();
void updateStateToGPU();
}
| Append to VEC4Address object defs commit (fix broken build). | Append to VEC4Address object defs commit (fix broken build). | Java | epl-1.0 | jtrfp/terminal-recall,jtrfp/terminal-recall,jtrfp/terminal-recall | java | ## Code Before:
/*******************************************************************************
* This file is part of TERMINAL RECALL
* Copyright (c) 2012-2014 Chuck Ritola
* Part of the jTRFP.org project
* All rights reserved. This program and the accompanying materials
* are made available under the terms of the GNU Public License v3.0
* which accompanies this distribution, and is available at
* http://www.gnu.org/licenses/gpl.html
*
* Contributors:
* chuck - initial API and implementation
******************************************************************************/
package org.jtrfp.trcl;
import java.nio.ByteBuffer;
public interface Renderable {
ByteBuffer getOpaqueObjectDefinitionAddresses();
ByteBuffer getTransparentObjectDefinitionAddresses();
void updateStateToGPU();
}
## Instruction:
Append to VEC4Address object defs commit (fix broken build).
## Code After:
/*******************************************************************************
* This file is part of TERMINAL RECALL
* Copyright (c) 2012-2014 Chuck Ritola
* Part of the jTRFP.org project
* All rights reserved. This program and the accompanying materials
* are made available under the terms of the GNU Public License v3.0
* which accompanies this distribution, and is available at
* http://www.gnu.org/licenses/gpl.html
*
* Contributors:
* chuck - initial API and implementation
******************************************************************************/
package org.jtrfp.trcl;
import java.util.Collection;
import org.jtrfp.trcl.mem.VEC4Address;
public interface Renderable {
Collection<VEC4Address> getOpaqueObjectDefinitionAddresses();
Collection<VEC4Address> getTransparentObjectDefinitionAddresses();
void updateStateToGPU();
}
| /*******************************************************************************
* This file is part of TERMINAL RECALL
* Copyright (c) 2012-2014 Chuck Ritola
* Part of the jTRFP.org project
* All rights reserved. This program and the accompanying materials
* are made available under the terms of the GNU Public License v3.0
* which accompanies this distribution, and is available at
* http://www.gnu.org/licenses/gpl.html
*
* Contributors:
* chuck - initial API and implementation
******************************************************************************/
package org.jtrfp.trcl;
- import java.nio.ByteBuffer;
+ import java.util.Collection;
+
+ import org.jtrfp.trcl.mem.VEC4Address;
public interface Renderable {
- ByteBuffer getOpaqueObjectDefinitionAddresses();
? ^^ ^^^^^^
+ Collection<VEC4Address> getOpaqueObjectDefinitionAddresses();
? ^^^^^^ ++++++++++++ ^^^
- ByteBuffer getTransparentObjectDefinitionAddresses();
? ^^ ^^^^^^
+ Collection<VEC4Address> getTransparentObjectDefinitionAddresses();
? ^^^^^^ ++++++++++++ ^^^
void updateStateToGPU();
} | 8 | 0.347826 | 5 | 3 |
ae7ed85528dfa069f4e47554cf61068569a0f227 | etl/s3-to-hdfs.sh | etl/s3-to-hdfs.sh | if [ $# -eq 0 ]; then
echo 'Target is missing.'
exit
fi
DIR=$(dirname $1)/$(basename $1)
for f in `aws s3 ls s3://t-ip$DIR/ | awk '{print $4}'`;
do
if [ ! -f /hdfs$DIR/$f ]; then
echo $f
aws s3 cp s3://t-ip$DIR/$f /tmp/
cp /tmp/$f /hdfs$DIR/
rm /tmp/$f
fi
done
| if [ $# -eq 0 ]; then
echo 'Target is missing.'
exit
fi
DIR=$(dirname $1)/$(basename $1)
for f in `/usr/local/bin/aws s3 ls s3://t-ip$DIR/ | awk '{print $4}'`;
do
if [ ! -f /hdfs$DIR/$f ]; then
echo $f
/usr/local/bin/aws s3 cp s3://t-ip$DIR/$f /tmp/
cp /tmp/$f /hdfs$DIR/
rm /tmp/$f
fi
done
| Change into absolute path for CRON job | Change into absolute path for CRON job
| Shell | apache-2.0 | dongjoon-hyun/tools,dongjoon-hyun/tools,dongjoon-hyun/tools,dongjoon-hyun/tools,dongjoon-hyun/tools | shell | ## Code Before:
if [ $# -eq 0 ]; then
echo 'Target is missing.'
exit
fi
DIR=$(dirname $1)/$(basename $1)
for f in `aws s3 ls s3://t-ip$DIR/ | awk '{print $4}'`;
do
if [ ! -f /hdfs$DIR/$f ]; then
echo $f
aws s3 cp s3://t-ip$DIR/$f /tmp/
cp /tmp/$f /hdfs$DIR/
rm /tmp/$f
fi
done
## Instruction:
Change into absolute path for CRON job
## Code After:
if [ $# -eq 0 ]; then
echo 'Target is missing.'
exit
fi
DIR=$(dirname $1)/$(basename $1)
for f in `/usr/local/bin/aws s3 ls s3://t-ip$DIR/ | awk '{print $4}'`;
do
if [ ! -f /hdfs$DIR/$f ]; then
echo $f
/usr/local/bin/aws s3 cp s3://t-ip$DIR/$f /tmp/
cp /tmp/$f /hdfs$DIR/
rm /tmp/$f
fi
done
| if [ $# -eq 0 ]; then
echo 'Target is missing.'
exit
fi
DIR=$(dirname $1)/$(basename $1)
- for f in `aws s3 ls s3://t-ip$DIR/ | awk '{print $4}'`;
+ for f in `/usr/local/bin/aws s3 ls s3://t-ip$DIR/ | awk '{print $4}'`;
? +++++++++++++++
do
if [ ! -f /hdfs$DIR/$f ]; then
echo $f
- aws s3 cp s3://t-ip$DIR/$f /tmp/
+ /usr/local/bin/aws s3 cp s3://t-ip$DIR/$f /tmp/
? +++++++++++++++
cp /tmp/$f /hdfs$DIR/
rm /tmp/$f
fi
done | 4 | 0.285714 | 2 | 2 |
15e199acbc8580a147d0d579c1e808f68d5daa8b | src/data/quiz.js | src/data/quiz.js | import Random from '../util/random';
import countries from '../data/countries';
export const QUIZ_OPTIONS_COUNT = 5;
export const generateQuizOptions = (num = QUIZ_OPTIONS_COUNT) => {
const options = Random.getRandomRange(countries, num);
const answer = Random.getRandomRange(options, 1)[0];
return {
options,
answer
}
}
| import Random from '../util/random';
import countries from '../data/countries';
export const QUIZ_OPTIONS_COUNT = 5;
let queue = [];
export const generateQuizOptions = (num = QUIZ_OPTIONS_COUNT) => {
if (queue.length > countries.length * .2) {
queue = queue.slice(1);
}
const options = Random.getRandomRange(countries.filter(country => queue.indexOf(country.iso3) === -1), num);
const answer = Random.getRandomRange(options, 1)[0];
queue.push(answer.iso3);
return {
options,
answer
}
}
| Add answers to a cooldown queue | Add answers to a cooldown queue
| JavaScript | isc | mrkiffie/flag-quiz,mrkiffie/flag-quiz | javascript | ## Code Before:
import Random from '../util/random';
import countries from '../data/countries';
export const QUIZ_OPTIONS_COUNT = 5;
export const generateQuizOptions = (num = QUIZ_OPTIONS_COUNT) => {
const options = Random.getRandomRange(countries, num);
const answer = Random.getRandomRange(options, 1)[0];
return {
options,
answer
}
}
## Instruction:
Add answers to a cooldown queue
## Code After:
import Random from '../util/random';
import countries from '../data/countries';
export const QUIZ_OPTIONS_COUNT = 5;
let queue = [];
export const generateQuizOptions = (num = QUIZ_OPTIONS_COUNT) => {
if (queue.length > countries.length * .2) {
queue = queue.slice(1);
}
const options = Random.getRandomRange(countries.filter(country => queue.indexOf(country.iso3) === -1), num);
const answer = Random.getRandomRange(options, 1)[0];
queue.push(answer.iso3);
return {
options,
answer
}
}
| import Random from '../util/random';
import countries from '../data/countries';
export const QUIZ_OPTIONS_COUNT = 5;
+ let queue = [];
+
export const generateQuizOptions = (num = QUIZ_OPTIONS_COUNT) => {
- const options = Random.getRandomRange(countries, num);
+
+ if (queue.length > countries.length * .2) {
+ queue = queue.slice(1);
+ }
+
+ const options = Random.getRandomRange(countries.filter(country => queue.indexOf(country.iso3) === -1), num);
const answer = Random.getRandomRange(options, 1)[0];
+
+ queue.push(answer.iso3);
+
return {
options,
answer
}
} | 12 | 0.923077 | 11 | 1 |
04e01981e8dd3b523bde6db52797ad23d3f50488 | app/controllers/users/recapitulatif_controller.rb | app/controllers/users/recapitulatif_controller.rb | class Users::RecapitulatifController < UsersController
before_action only: [:show] do
authorized_routes? self.class
end
def show
create_dossier_facade
end
def initiate
create_dossier_facade
@facade.dossier.next_step! 'user', 'initiate'
flash.notice = 'Dossier soumis avec succès.'
render 'show'
end
def submit
create_dossier_facade
@facade.dossier.submit!
flash.notice = 'Dossier déposé avec succès.'
render 'show'
end
def self.route_authorization
{
states: [:initiated, :replied, :updated, :validated, :received, :submitted, :without_continuation, :closed]
}
end
private
def create_dossier_facade
@facade = DossierFacades.new current_user_dossier.id, current_user.email
rescue ActiveRecord::RecordNotFound
flash.alert = t('errors.messages.dossier_not_found')
redirect_to url_for(root_path)
end
end
| class Users::RecapitulatifController < UsersController
before_action only: [:show] do
authorized_routes? self.class
end
def show
create_dossier_facade
end
def initiate
create_dossier_facade
@facade.dossier.next_step! 'user', 'initiate'
flash.notice = 'Dossier soumis avec succès.'
render 'show'
end
def submit
create_dossier_facade
@facade.dossier.submit!
flash.notice = 'Dossier déposé avec succès.'
render 'show'
end
def self.route_authorization
{
states: [:initiated, :replied, :updated, :validated, :received, :submitted, :without_continuation, :closed, :refused]
}
end
private
def create_dossier_facade
@facade = DossierFacades.new current_user_dossier.id, current_user.email
rescue ActiveRecord::RecordNotFound
flash.alert = t('errors.messages.dossier_not_found')
redirect_to url_for(root_path)
end
end
| Fix authorized routes on show dossier | Fix authorized routes on show dossier
| Ruby | agpl-3.0 | sgmap/tps,sgmap/tps,sgmap/tps | ruby | ## Code Before:
class Users::RecapitulatifController < UsersController
before_action only: [:show] do
authorized_routes? self.class
end
def show
create_dossier_facade
end
def initiate
create_dossier_facade
@facade.dossier.next_step! 'user', 'initiate'
flash.notice = 'Dossier soumis avec succès.'
render 'show'
end
def submit
create_dossier_facade
@facade.dossier.submit!
flash.notice = 'Dossier déposé avec succès.'
render 'show'
end
def self.route_authorization
{
states: [:initiated, :replied, :updated, :validated, :received, :submitted, :without_continuation, :closed]
}
end
private
def create_dossier_facade
@facade = DossierFacades.new current_user_dossier.id, current_user.email
rescue ActiveRecord::RecordNotFound
flash.alert = t('errors.messages.dossier_not_found')
redirect_to url_for(root_path)
end
end
## Instruction:
Fix authorized routes on show dossier
## Code After:
class Users::RecapitulatifController < UsersController
before_action only: [:show] do
authorized_routes? self.class
end
def show
create_dossier_facade
end
def initiate
create_dossier_facade
@facade.dossier.next_step! 'user', 'initiate'
flash.notice = 'Dossier soumis avec succès.'
render 'show'
end
def submit
create_dossier_facade
@facade.dossier.submit!
flash.notice = 'Dossier déposé avec succès.'
render 'show'
end
def self.route_authorization
{
states: [:initiated, :replied, :updated, :validated, :received, :submitted, :without_continuation, :closed, :refused]
}
end
private
def create_dossier_facade
@facade = DossierFacades.new current_user_dossier.id, current_user.email
rescue ActiveRecord::RecordNotFound
flash.alert = t('errors.messages.dossier_not_found')
redirect_to url_for(root_path)
end
end
| class Users::RecapitulatifController < UsersController
before_action only: [:show] do
authorized_routes? self.class
end
def show
create_dossier_facade
end
def initiate
create_dossier_facade
@facade.dossier.next_step! 'user', 'initiate'
flash.notice = 'Dossier soumis avec succès.'
render 'show'
end
def submit
create_dossier_facade
@facade.dossier.submit!
flash.notice = 'Dossier déposé avec succès.'
render 'show'
end
def self.route_authorization
{
- states: [:initiated, :replied, :updated, :validated, :received, :submitted, :without_continuation, :closed]
+ states: [:initiated, :replied, :updated, :validated, :received, :submitted, :without_continuation, :closed, :refused]
? ++++++++++
}
end
private
def create_dossier_facade
@facade = DossierFacades.new current_user_dossier.id, current_user.email
rescue ActiveRecord::RecordNotFound
flash.alert = t('errors.messages.dossier_not_found')
redirect_to url_for(root_path)
end
end | 2 | 0.045455 | 1 | 1 |
12d01ff0fbf733b909002b6cab14c27ece26909c | .travis.yml | .travis.yml | language: ruby
rvm:
- 1.9.3
services:
- postgresql
- elasticsearch
notifications:
webhooks: http://hooks.holderdeord.no/deploy
recipients:
- jari@holderdeord.no
irc:
channels:
- "irc.freenode.net#holderdeord"
on_success: change
on_failure: change
template:
- "%{author} | %{commit} | %{message} | %{build_url}"
script: "RAILS_ENV=test SPEC_OPTS=--backtrace bundle exec rake travis"
before_script:
- mv config/database.yml.travis config/database.yml
- sh -e /etc/init.d/xvfb start
- 'npm install -g buster autolint || (sleep 5 && npm install -g buster autolint)'
env:
- DISPLAY=:99.0 COVERAGE_THRESHOLD=83
| language: ruby
rvm:
- 1.9.3
services:
- postgresql
- elasticsearch
notifications:
webhooks: http://hooks.holderdeord.no/deploy
recipients:
- jari@holderdeord.no
irc:
channels:
- "irc.freenode.net#holderdeord"
on_success: change
on_failure: change
template:
- "%{author} | %{commit} | %{message} | %{build_url}"
script: "RAILS_ENV=test SPEC_OPTS=--backtrace bundle exec rake travis"
before_script:
- dpkg -s elasticsearch | grep Version
- mv config/database.yml.travis config/database.yml
- sh -e /etc/init.d/xvfb start
- 'npm install -g buster autolint || (sleep 5 && npm install -g buster autolint)'
env:
- DISPLAY=:99.0 COVERAGE_THRESHOLD=83
| Print the elasticsearch version used on Travis. | Print the elasticsearch version used on Travis.
| YAML | bsd-3-clause | holderdeord/hdo-site,holderdeord/hdo-site,holderdeord/hdo-site,holderdeord/hdo-site | yaml | ## Code Before:
language: ruby
rvm:
- 1.9.3
services:
- postgresql
- elasticsearch
notifications:
webhooks: http://hooks.holderdeord.no/deploy
recipients:
- jari@holderdeord.no
irc:
channels:
- "irc.freenode.net#holderdeord"
on_success: change
on_failure: change
template:
- "%{author} | %{commit} | %{message} | %{build_url}"
script: "RAILS_ENV=test SPEC_OPTS=--backtrace bundle exec rake travis"
before_script:
- mv config/database.yml.travis config/database.yml
- sh -e /etc/init.d/xvfb start
- 'npm install -g buster autolint || (sleep 5 && npm install -g buster autolint)'
env:
- DISPLAY=:99.0 COVERAGE_THRESHOLD=83
## Instruction:
Print the elasticsearch version used on Travis.
## Code After:
language: ruby
rvm:
- 1.9.3
services:
- postgresql
- elasticsearch
notifications:
webhooks: http://hooks.holderdeord.no/deploy
recipients:
- jari@holderdeord.no
irc:
channels:
- "irc.freenode.net#holderdeord"
on_success: change
on_failure: change
template:
- "%{author} | %{commit} | %{message} | %{build_url}"
script: "RAILS_ENV=test SPEC_OPTS=--backtrace bundle exec rake travis"
before_script:
- dpkg -s elasticsearch | grep Version
- mv config/database.yml.travis config/database.yml
- sh -e /etc/init.d/xvfb start
- 'npm install -g buster autolint || (sleep 5 && npm install -g buster autolint)'
env:
- DISPLAY=:99.0 COVERAGE_THRESHOLD=83
| language: ruby
rvm:
- 1.9.3
services:
- postgresql
- elasticsearch
notifications:
webhooks: http://hooks.holderdeord.no/deploy
recipients:
- jari@holderdeord.no
irc:
channels:
- "irc.freenode.net#holderdeord"
on_success: change
on_failure: change
template:
- "%{author} | %{commit} | %{message} | %{build_url}"
script: "RAILS_ENV=test SPEC_OPTS=--backtrace bundle exec rake travis"
before_script:
+ - dpkg -s elasticsearch | grep Version
- mv config/database.yml.travis config/database.yml
- sh -e /etc/init.d/xvfb start
- 'npm install -g buster autolint || (sleep 5 && npm install -g buster autolint)'
env:
- DISPLAY=:99.0 COVERAGE_THRESHOLD=83 | 1 | 0.041667 | 1 | 0 |
7388f0742135aada695bf4750cfd588d96f716c8 | app/views/opinio/comments/create.js.erb | app/views/opinio/comments/create.js.erb | $('#no_comments').fadeOut()
<% if @reply %>
if($('#comment_<%= @comment.commentable_id %> ul').length == 0)
$('#comment_<%= @comment.commentable_id %>').append('<ul id="comment_<%= @comment.commentable_id %>_replies" class="replies"></ul>')
$('#comment_<%= @comment.commentable_id %>_replies').append("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
$('#commentable_id').val('<%= @comment.commentable.commentable_id %>')
$('#commentable_type').val('<%= @comment.commentable.commentable_type %>')
<% else %>
$('#comments').prepend("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
<% end %>
| $('#no_comments').fadeOut()
<% if @comment.valid? %>
<% if @reply %>
if($('#comment_<%= @comment.commentable_id %> ul').length == 0)
$('#comment_<%= @comment.commentable_id %>').append('<ul id="comment_<%= @comment.commentable_id %>_replies" class="replies"></ul>')
$('#comment_<%= @comment.commentable_id %>_replies').append("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
$('#commentable_id').val('<%= @comment.commentable.commentable_id %>')
$('#commentable_type').val('<%= @comment.commentable.commentable_type %>')
<% else %>
$('#comments').prepend("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
<% end %>
<% else %>
jQuery('#comments').prepend("<%= escape_javascript(flash[:notice]) %>");
jQuery('#comments').prepend("<%= escape_javascript(flash[:error]) %>");
<% end %>
| Return error to user if couldn't save the comment | Return error to user if couldn't save the comment
| HTML+ERB | mit | Draiken/opinio,Draiken/opinio | html+erb | ## Code Before:
$('#no_comments').fadeOut()
<% if @reply %>
if($('#comment_<%= @comment.commentable_id %> ul').length == 0)
$('#comment_<%= @comment.commentable_id %>').append('<ul id="comment_<%= @comment.commentable_id %>_replies" class="replies"></ul>')
$('#comment_<%= @comment.commentable_id %>_replies').append("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
$('#commentable_id').val('<%= @comment.commentable.commentable_id %>')
$('#commentable_type').val('<%= @comment.commentable.commentable_type %>')
<% else %>
$('#comments').prepend("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
<% end %>
## Instruction:
Return error to user if couldn't save the comment
## Code After:
$('#no_comments').fadeOut()
<% if @comment.valid? %>
<% if @reply %>
if($('#comment_<%= @comment.commentable_id %> ul').length == 0)
$('#comment_<%= @comment.commentable_id %>').append('<ul id="comment_<%= @comment.commentable_id %>_replies" class="replies"></ul>')
$('#comment_<%= @comment.commentable_id %>_replies').append("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
$('#commentable_id').val('<%= @comment.commentable.commentable_id %>')
$('#commentable_type').val('<%= @comment.commentable.commentable_type %>')
<% else %>
$('#comments').prepend("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
<% end %>
<% else %>
jQuery('#comments').prepend("<%= escape_javascript(flash[:notice]) %>");
jQuery('#comments').prepend("<%= escape_javascript(flash[:error]) %>");
<% end %>
| $('#no_comments').fadeOut()
+ <% if @comment.valid? %>
- <% if @reply %>
+ <% if @reply %>
? ++
- if($('#comment_<%= @comment.commentable_id %> ul').length == 0)
+ if($('#comment_<%= @comment.commentable_id %> ul').length == 0)
? ++
- $('#comment_<%= @comment.commentable_id %>').append('<ul id="comment_<%= @comment.commentable_id %>_replies" class="replies"></ul>')
+ $('#comment_<%= @comment.commentable_id %>').append('<ul id="comment_<%= @comment.commentable_id %>_replies" class="replies"></ul>')
? ++
- $('#comment_<%= @comment.commentable_id %>_replies').append("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
+ $('#comment_<%= @comment.commentable_id %>_replies').append("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
? ++
- $('#commentable_id').val('<%= @comment.commentable.commentable_id %>')
+ $('#commentable_id').val('<%= @comment.commentable.commentable_id %>')
? ++
- $('#commentable_type').val('<%= @comment.commentable.commentable_type %>')
+ $('#commentable_type').val('<%= @comment.commentable.commentable_type %>')
? ++
+ <% else %>
+ $('#comments').prepend("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
+ <% end %>
<% else %>
- $('#comments').prepend("<%= escape_javascript( render @comment, :locals => {:reply => @reply} ) %>")
+ jQuery('#comments').prepend("<%= escape_javascript(flash[:notice]) %>");
+ jQuery('#comments').prepend("<%= escape_javascript(flash[:error]) %>");
<% end %>
- | 20 | 1.818182 | 12 | 8 |
57b74f8ce3282db91bbdaecb7db4d19712a9e40d | config/CMakeLists.txt | config/CMakeLists.txt | INSTALL(
FILES
csync_log.conf
DESTINATION
${DATA_INSTALL_PREFIX}
)
| macro_copy_file(${CMAKE_CURRENT_SOURCE_DIR}/csync.conf ${CMAKE_CURRENT_BINARY_DIR}/csync.conf)
macro_copy_file(${CMAKE_CURRENT_SOURCE_DIR}/csync_log.conf ${CMAKE_CURRENT_BINARY_DIR}/csync_log.conf)
install(
FILES
csync.conf
csync_log.conf
DESTINATION
${SYSCONF_INSTALL_DIR}/csync
)
| Copy config to the build directory. | Copy config to the build directory.
| Text | lgpl-2.1 | gco/csync,meeh420/csync,gco/csync,gco/csync,meeh420/csync,gco/csync,meeh420/csync | text | ## Code Before:
INSTALL(
FILES
csync_log.conf
DESTINATION
${DATA_INSTALL_PREFIX}
)
## Instruction:
Copy config to the build directory.
## Code After:
macro_copy_file(${CMAKE_CURRENT_SOURCE_DIR}/csync.conf ${CMAKE_CURRENT_BINARY_DIR}/csync.conf)
macro_copy_file(${CMAKE_CURRENT_SOURCE_DIR}/csync_log.conf ${CMAKE_CURRENT_BINARY_DIR}/csync_log.conf)
install(
FILES
csync.conf
csync_log.conf
DESTINATION
${SYSCONF_INSTALL_DIR}/csync
)
| - INSTALL(
+ macro_copy_file(${CMAKE_CURRENT_SOURCE_DIR}/csync.conf ${CMAKE_CURRENT_BINARY_DIR}/csync.conf)
+ macro_copy_file(${CMAKE_CURRENT_SOURCE_DIR}/csync_log.conf ${CMAKE_CURRENT_BINARY_DIR}/csync_log.conf)
+
+ install(
FILES
+ csync.conf
csync_log.conf
DESTINATION
- ${DATA_INSTALL_PREFIX}
+ ${SYSCONF_INSTALL_DIR}/csync
) | 8 | 1.333333 | 6 | 2 |
67ff617ab87f248c62d6ce422ebdb581802e4229 | recipes-debian/gperf/gperf_debian.bb | recipes-debian/gperf/gperf_debian.bb | require recipes-extended/gperf/gperf_3.0.4.bb
inherit debian-package
DEBIAN_SECTION = "devel"
LICENSE = "GPLv3"
LIC_FILES_CHKSUM = "file://COPYING;md5=d32239bcb673463ab874e80d47fae504"
#
# Debian Native Test
#
inherit debian-test
SRC_URI_DEBIAN_TEST = "\
file://gperf-native-test/run_native_test_gperf \
file://gperf-native-test/run_with_option-L \
file://gperf-native-test/run_with_option-N \
file://gperf-native-test/run_with_option-H \
file://gperf-native-test/run_with_option-Z \
file://gperf-native-test/command_line_options.gperf \
"
DEBIAN_NATIVE_TESTS = "run_native_test_gperf"
TEST_DIR = "${B}/native-test"
|
PR = "r0"
inherit debian-package
LICENSE = "GPLv3"
LIC_FILES_CHKSUM = "file://COPYING;md5=d32239bcb673463ab874e80d47fae504"
inherit autotools
# autoreconf couldn't find acinclude.m4 when stepping into subdirectory. Instead of
# duplicating acinclude.m4 in every subdirectory, use absolute include path to aclocal
EXTRA_AUTORECONF += " -I ${S}"
do_configure_prepend() {
if [ ! -e ${S}/acinclude.m4 ]; then
cat ${S}/aclocal.m4 > ${S}/acinclude.m4
fi
}
BBCLASSEXTEND = "native"
#
# Debian Native Test
#
inherit debian-test
SRC_URI_DEBIAN_TEST = "\
file://gperf-native-test/run_native_test_gperf \
file://gperf-native-test/run_with_option-L \
file://gperf-native-test/run_with_option-N \
file://gperf-native-test/run_with_option-H \
file://gperf-native-test/run_with_option-Z \
file://gperf-native-test/command_line_options.gperf \
"
DEBIAN_NATIVE_TESTS = "run_native_test_gperf"
TEST_DIR = "${B}/native-test"
| Update recipe for gperf package | Update recipe for gperf package
Signed-off-by: Tran Thi Van Dung <4602a2150458a8fc79ca562d1ec5fffebd9f9638@toshiba-tsdv.com>
| BitBake | mit | nghiaht-tsdv/meta-debian,rofehr/meta-debian,nghiaht-tsdv/meta-debian,rofehr/meta-debian,congnt-tsdv/meta-debian,nghiaht-tsdv/meta-debian,meta-debian/meta-debian,congnt-tsdv/meta-debian,tienlee/meta-debian,rofehr/meta-debian,meta-debian/meta-debian,nghiaht-tsdv/meta-debian,tienlee/meta-debian,congnt-tsdv/meta-debian,meta-debian/meta-debian,nghiaht-tsdv/meta-debian,meta-debian/meta-debian,meta-debian/meta-debian,rofehr/meta-debian,nghiaht-tsdv/meta-debian,tienlee/meta-debian,tienlee/meta-debian,rofehr/meta-debian,congnt-tsdv/meta-debian,tienlee/meta-debian,congnt-tsdv/meta-debian,tienlee/meta-debian,meta-debian/meta-debian,rofehr/meta-debian,congnt-tsdv/meta-debian | bitbake | ## Code Before:
require recipes-extended/gperf/gperf_3.0.4.bb
inherit debian-package
DEBIAN_SECTION = "devel"
LICENSE = "GPLv3"
LIC_FILES_CHKSUM = "file://COPYING;md5=d32239bcb673463ab874e80d47fae504"
#
# Debian Native Test
#
inherit debian-test
SRC_URI_DEBIAN_TEST = "\
file://gperf-native-test/run_native_test_gperf \
file://gperf-native-test/run_with_option-L \
file://gperf-native-test/run_with_option-N \
file://gperf-native-test/run_with_option-H \
file://gperf-native-test/run_with_option-Z \
file://gperf-native-test/command_line_options.gperf \
"
DEBIAN_NATIVE_TESTS = "run_native_test_gperf"
TEST_DIR = "${B}/native-test"
## Instruction:
Update recipe for gperf package
Signed-off-by: Tran Thi Van Dung <4602a2150458a8fc79ca562d1ec5fffebd9f9638@toshiba-tsdv.com>
## Code After:
PR = "r0"
inherit debian-package
LICENSE = "GPLv3"
LIC_FILES_CHKSUM = "file://COPYING;md5=d32239bcb673463ab874e80d47fae504"
inherit autotools
# autoreconf couldn't find acinclude.m4 when stepping into subdirectory. Instead of
# duplicating acinclude.m4 in every subdirectory, use absolute include path to aclocal
EXTRA_AUTORECONF += " -I ${S}"
do_configure_prepend() {
if [ ! -e ${S}/acinclude.m4 ]; then
cat ${S}/aclocal.m4 > ${S}/acinclude.m4
fi
}
BBCLASSEXTEND = "native"
#
# Debian Native Test
#
inherit debian-test
SRC_URI_DEBIAN_TEST = "\
file://gperf-native-test/run_native_test_gperf \
file://gperf-native-test/run_with_option-L \
file://gperf-native-test/run_with_option-N \
file://gperf-native-test/run_with_option-H \
file://gperf-native-test/run_with_option-Z \
file://gperf-native-test/command_line_options.gperf \
"
DEBIAN_NATIVE_TESTS = "run_native_test_gperf"
TEST_DIR = "${B}/native-test"
| - require recipes-extended/gperf/gperf_3.0.4.bb
+
+ PR = "r0"
inherit debian-package
- DEBIAN_SECTION = "devel"
LICENSE = "GPLv3"
LIC_FILES_CHKSUM = "file://COPYING;md5=d32239bcb673463ab874e80d47fae504"
+
+ inherit autotools
+
+ # autoreconf couldn't find acinclude.m4 when stepping into subdirectory. Instead of
+ # duplicating acinclude.m4 in every subdirectory, use absolute include path to aclocal
+ EXTRA_AUTORECONF += " -I ${S}"
+
+ do_configure_prepend() {
+ if [ ! -e ${S}/acinclude.m4 ]; then
+ cat ${S}/aclocal.m4 > ${S}/acinclude.m4
+ fi
+ }
+
+ BBCLASSEXTEND = "native"
#
# Debian Native Test
#
inherit debian-test
SRC_URI_DEBIAN_TEST = "\
file://gperf-native-test/run_native_test_gperf \
file://gperf-native-test/run_with_option-L \
file://gperf-native-test/run_with_option-N \
file://gperf-native-test/run_with_option-H \
file://gperf-native-test/run_with_option-Z \
file://gperf-native-test/command_line_options.gperf \
"
DEBIAN_NATIVE_TESTS = "run_native_test_gperf"
TEST_DIR = "${B}/native-test" | 18 | 0.75 | 16 | 2 |
f5fc61950e20bba6f8b8cc3decc8a4c223700660 | app/src/main/res/menu/menu_main.xml | app/src/main/res/menu/menu_main.xml | <?xml version="1.0" encoding="utf-8"?>
<menu xmlns:android="http://schemas.android.com/apk/res/android"
xmlns:app="http://schemas.android.com/apk/res-auto"
xmlns:tools="http://schemas.android.com/tools"
tools:context=".ui.activities.MainActivity">
<item
android:id="@+id/action_library"
android:title="@string/action_library"
app:showAsAction="never">
<menu>
<group android:checkableBehavior="single">
<item
android:id="@+id/action_library_jpop"
android:title="@string/jpop" />
<item
android:id="@+id/action_library_kpop"
android:title="@string/kpop" />
</group>
</menu>
</item>
<item
android:id="@+id/action_logout"
android:title="@string/logout"
app:showAsAction="never" />
<item
android:id="@+id/action_settings"
android:title="@string/settings"
app:showAsAction="never" />
<item
android:id="@+id/action_about"
android:title="@string/about"
app:showAsAction="never" />
<item
android:id="@+id/action_sleep_timer"
android:title="@string/sleep_timer"
app:showAsAction="never" />
</menu>
| <?xml version="1.0" encoding="utf-8"?>
<menu xmlns:android="http://schemas.android.com/apk/res/android"
xmlns:app="http://schemas.android.com/apk/res-auto"
xmlns:tools="http://schemas.android.com/tools"
tools:context=".ui.activities.MainActivity">
<item
android:id="@+id/action_library"
android:title="@string/action_library"
app:showAsAction="never">
<menu>
<group android:checkableBehavior="single">
<item
android:id="@+id/action_library_jpop"
android:title="@string/jpop" />
<item
android:id="@+id/action_library_kpop"
android:title="@string/kpop" />
</group>
</menu>
</item>
<item
android:id="@+id/action_settings"
android:title="@string/settings"
app:showAsAction="never" />
<item
android:id="@+id/action_about"
android:title="@string/about"
app:showAsAction="never" />
<item
android:id="@+id/action_sleep_timer"
android:title="@string/sleep_timer"
app:showAsAction="never" />
<item
android:id="@+id/action_logout"
android:title="@string/logout"
app:showAsAction="never" />
</menu>
| Move logout to bottom of menu | Move logout to bottom of menu
| XML | mit | LISTEN-moe/android-app,arkon/LISTEN.moe-Unofficial-Android-App | xml | ## Code Before:
<?xml version="1.0" encoding="utf-8"?>
<menu xmlns:android="http://schemas.android.com/apk/res/android"
xmlns:app="http://schemas.android.com/apk/res-auto"
xmlns:tools="http://schemas.android.com/tools"
tools:context=".ui.activities.MainActivity">
<item
android:id="@+id/action_library"
android:title="@string/action_library"
app:showAsAction="never">
<menu>
<group android:checkableBehavior="single">
<item
android:id="@+id/action_library_jpop"
android:title="@string/jpop" />
<item
android:id="@+id/action_library_kpop"
android:title="@string/kpop" />
</group>
</menu>
</item>
<item
android:id="@+id/action_logout"
android:title="@string/logout"
app:showAsAction="never" />
<item
android:id="@+id/action_settings"
android:title="@string/settings"
app:showAsAction="never" />
<item
android:id="@+id/action_about"
android:title="@string/about"
app:showAsAction="never" />
<item
android:id="@+id/action_sleep_timer"
android:title="@string/sleep_timer"
app:showAsAction="never" />
</menu>
## Instruction:
Move logout to bottom of menu
## Code After:
<?xml version="1.0" encoding="utf-8"?>
<menu xmlns:android="http://schemas.android.com/apk/res/android"
xmlns:app="http://schemas.android.com/apk/res-auto"
xmlns:tools="http://schemas.android.com/tools"
tools:context=".ui.activities.MainActivity">
<item
android:id="@+id/action_library"
android:title="@string/action_library"
app:showAsAction="never">
<menu>
<group android:checkableBehavior="single">
<item
android:id="@+id/action_library_jpop"
android:title="@string/jpop" />
<item
android:id="@+id/action_library_kpop"
android:title="@string/kpop" />
</group>
</menu>
</item>
<item
android:id="@+id/action_settings"
android:title="@string/settings"
app:showAsAction="never" />
<item
android:id="@+id/action_about"
android:title="@string/about"
app:showAsAction="never" />
<item
android:id="@+id/action_sleep_timer"
android:title="@string/sleep_timer"
app:showAsAction="never" />
<item
android:id="@+id/action_logout"
android:title="@string/logout"
app:showAsAction="never" />
</menu>
| <?xml version="1.0" encoding="utf-8"?>
<menu xmlns:android="http://schemas.android.com/apk/res/android"
xmlns:app="http://schemas.android.com/apk/res-auto"
xmlns:tools="http://schemas.android.com/tools"
tools:context=".ui.activities.MainActivity">
<item
android:id="@+id/action_library"
android:title="@string/action_library"
app:showAsAction="never">
<menu>
<group android:checkableBehavior="single">
<item
android:id="@+id/action_library_jpop"
android:title="@string/jpop" />
<item
android:id="@+id/action_library_kpop"
android:title="@string/kpop" />
</group>
</menu>
</item>
<item
- android:id="@+id/action_logout"
- android:title="@string/logout"
- app:showAsAction="never" />
-
- <item
android:id="@+id/action_settings"
android:title="@string/settings"
app:showAsAction="never" />
<item
android:id="@+id/action_about"
android:title="@string/about"
app:showAsAction="never" />
<item
android:id="@+id/action_sleep_timer"
android:title="@string/sleep_timer"
app:showAsAction="never" />
+ <item
+ android:id="@+id/action_logout"
+ android:title="@string/logout"
+ app:showAsAction="never" />
+
</menu> | 10 | 0.217391 | 5 | 5 |
a79804ff154d17011f6bd95568812cf13828c3ee | app/partials/actions/list.html | app/partials/actions/list.html | <div class="row">
<label for="presenters" style="text-transform: capitalize;">{{resourceName}}</label>
<div
multi-select
input-model="models"
output-model="selectedModels"
button-label="name"
item-label="name"
selection-mode="{{selectionMode}}"
tick-property="ticked"
></div>
<!-- <button type="button" class="btn btn-default btn-block" ng-click="addPresenter()">+ New Presenter</button> -->
</div> | <div class="row">
<label style="text-transform: capitalize;">{{resourceName}}</label>
<div
multi-select
input-model="models"
output-model="selectedModels"
button-label="name"
item-label="name"
selection-mode="{{selectionMode}}"
tick-property="ticked"
></div>
<!-- <button type="button" class="btn btn-default btn-block" ng-click="addPresenter()">+ New Presenter</button> -->
</div> | Remove unnecessary presenters reference in label | Remove unnecessary presenters reference in label
| HTML | mit | TechAtNYU/intranet,TechAtNYU/intranet | html | ## Code Before:
<div class="row">
<label for="presenters" style="text-transform: capitalize;">{{resourceName}}</label>
<div
multi-select
input-model="models"
output-model="selectedModels"
button-label="name"
item-label="name"
selection-mode="{{selectionMode}}"
tick-property="ticked"
></div>
<!-- <button type="button" class="btn btn-default btn-block" ng-click="addPresenter()">+ New Presenter</button> -->
</div>
## Instruction:
Remove unnecessary presenters reference in label
## Code After:
<div class="row">
<label style="text-transform: capitalize;">{{resourceName}}</label>
<div
multi-select
input-model="models"
output-model="selectedModels"
button-label="name"
item-label="name"
selection-mode="{{selectionMode}}"
tick-property="ticked"
></div>
<!-- <button type="button" class="btn btn-default btn-block" ng-click="addPresenter()">+ New Presenter</button> -->
</div> | <div class="row">
- <label for="presenters" style="text-transform: capitalize;">{{resourceName}}</label>
? -----------------
+ <label style="text-transform: capitalize;">{{resourceName}}</label>
<div
multi-select
input-model="models"
output-model="selectedModels"
button-label="name"
item-label="name"
selection-mode="{{selectionMode}}"
tick-property="ticked"
></div>
<!-- <button type="button" class="btn btn-default btn-block" ng-click="addPresenter()">+ New Presenter</button> -->
</div> | 2 | 0.153846 | 1 | 1 |
64a0fae01e96a0864bb08766307c9dd8fe5a98ff | test/acceptance/features/create-a-contact.feature | test/acceptance/features/create-a-contact.feature | @contact-create
Feature: Create New Contact
As an existing user
I would like to create a new contact for a company
Scenario: Verify Add new Contact option
Given I am an authenticated user on Data Hub website
When I navigate to Contacts page of any company
Then I verify an option to add a new contact
Scenario: Add a new Primary Contact
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
Scenario: Add a new Primary Contact with new Company address
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact with a new company address
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
Scenario: Add a new non Primary Contact
Given I am an authenticated user on Data Hub website
When I add a new non Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
Scenario: Verify newly added contact under search landing page
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact under search landing page
| @contact-create
Feature: Create New Contact
As an existing user
I would like to create a new contact for a company
@contact-create-option
Scenario: Verify Add new Contact option
Given I am an authenticated user on Data Hub website
When I navigate to Contacts page of any company
Then I verify an option to add a new contact
@contact-create-primary
Scenario: Add a new Primary Contact
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
@contact-create-primary-new-company-address
Scenario: Add a new Primary Contact with new Company address
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact with a new company address
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
@contact-create-non-primary
Scenario: Add a new non-Primary Contact
Given I am an authenticated user on Data Hub website
When I add a new non Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
@contact-create-verify
Scenario: Verify newly added contact under search landing page
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact under search landing page
| Add tags to interaction and contact scenarios | E2E: Add tags to interaction and contact scenarios
| Cucumber | mit | uktrade/data-hub-frontend,uktrade/data-hub-frontend,uktrade/data-hub-frontend,uktrade/data-hub-fe-beta2,uktrade/data-hub-fe-beta2 | cucumber | ## Code Before:
@contact-create
Feature: Create New Contact
As an existing user
I would like to create a new contact for a company
Scenario: Verify Add new Contact option
Given I am an authenticated user on Data Hub website
When I navigate to Contacts page of any company
Then I verify an option to add a new contact
Scenario: Add a new Primary Contact
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
Scenario: Add a new Primary Contact with new Company address
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact with a new company address
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
Scenario: Add a new non Primary Contact
Given I am an authenticated user on Data Hub website
When I add a new non Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
Scenario: Verify newly added contact under search landing page
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact under search landing page
## Instruction:
E2E: Add tags to interaction and contact scenarios
## Code After:
@contact-create
Feature: Create New Contact
As an existing user
I would like to create a new contact for a company
@contact-create-option
Scenario: Verify Add new Contact option
Given I am an authenticated user on Data Hub website
When I navigate to Contacts page of any company
Then I verify an option to add a new contact
@contact-create-primary
Scenario: Add a new Primary Contact
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
@contact-create-primary-new-company-address
Scenario: Add a new Primary Contact with new Company address
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact with a new company address
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
@contact-create-non-primary
Scenario: Add a new non-Primary Contact
Given I am an authenticated user on Data Hub website
When I add a new non Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
@contact-create-verify
Scenario: Verify newly added contact under search landing page
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact under search landing page
| @contact-create
Feature: Create New Contact
As an existing user
I would like to create a new contact for a company
+ @contact-create-option
Scenario: Verify Add new Contact option
Given I am an authenticated user on Data Hub website
When I navigate to Contacts page of any company
Then I verify an option to add a new contact
+ @contact-create-primary
Scenario: Add a new Primary Contact
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
+ @contact-create-primary-new-company-address
Scenario: Add a new Primary Contact with new Company address
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact with a new company address
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
+ @contact-create-non-primary
- Scenario: Add a new non Primary Contact
? ^
+ Scenario: Add a new non-Primary Contact
? ^
Given I am an authenticated user on Data Hub website
When I add a new non Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact in company profile
+ @contact-create-verify
Scenario: Verify newly added contact under search landing page
Given I am an authenticated user on Data Hub website
When I add a new Primary Contact
Then I see the contact creation confirmation message
And I verify my newly added contact under search landing page
| 7 | 0.205882 | 6 | 1 |
e5cfdd5aa08199882030d05e08758873be4cfec7 | README.md | README.md | Demonstration of developing, testing, and delivering production software with Docker
| Demonstration of developing, testing, and delivering production software with Docker.
## Overview
The MEAN stack application code is used to demonstrate how to:
- Containerize a multi-tier application
- Create development and test environments with Docker Compose
- Create production environments using Docker Compose override files
- Release a new feature from development, through test, and into production
## Getting Started
An Azure Resource Manager template is provided to create an environment with three virtual machines in a virtual network with Docker installed on each. Certificates are shared and setup for TLS secured communication between the three Docker hosts. The created environment resembles the following:
<img src="https://user-images.githubusercontent.com/3911650/27996794-58870aa8-64a7-11e7-8227-e1f137603dba.png" alt="Azure environment">
### One-Click Deploy
<a href="https://portal.azure.com/#create/Microsoft.Template/uri/https%3A%2F%2Fraw.githubusercontent.com%2Flrakai%2Fdocker-software-delivery%2Fmaster%2Finfrastructure%2Farm-template.json">
<img src="https://camo.githubusercontent.com/9285dd3998997a0835869065bb15e5d500475034/687474703a2f2f617a7572656465706c6f792e6e65742f6465706c6f79627574746f6e2e706e67" data-canonical-src="http://azuredeploy.net/deploybutton.png" style="max-width:100%;">
</a>
### Using PowerShell
The following command sequence can prepare the environment in the West US 2 Azure region:
```ps1
Login-AzureRmAccount
New-AzureRmResourceGroup -Name docker -Location westus2
New-AzureRmResourceGroupDeployment -ResourceGroupName docker -TemplateFile .\infrastructure\arm-template.json -Name dsd
```
When finished, the following can tear down the environment:
```ps1
Remove-AzureRmResourceGroup -Name docker
``` | Include overview and getting started sections | Include overview and getting started sections
| Markdown | mit | lrakai/docker-software-delivery,lrakai/docker-software-delivery,lrakai/docker-software-delivery | markdown | ## Code Before:
Demonstration of developing, testing, and delivering production software with Docker
## Instruction:
Include overview and getting started sections
## Code After:
Demonstration of developing, testing, and delivering production software with Docker.
## Overview
The MEAN stack application code is used to demonstrate how to:
- Containerize a multi-tier application
- Create development and test environments with Docker Compose
- Create production environments using Docker Compose override files
- Release a new feature from development, through test, and into production
## Getting Started
An Azure Resource Manager template is provided to create an environment with three virtual machines in a virtual network with Docker installed on each. Certificates are shared and setup for TLS secured communication between the three Docker hosts. The created environment resembles the following:
<img src="https://user-images.githubusercontent.com/3911650/27996794-58870aa8-64a7-11e7-8227-e1f137603dba.png" alt="Azure environment">
### One-Click Deploy
<a href="https://portal.azure.com/#create/Microsoft.Template/uri/https%3A%2F%2Fraw.githubusercontent.com%2Flrakai%2Fdocker-software-delivery%2Fmaster%2Finfrastructure%2Farm-template.json">
<img src="https://camo.githubusercontent.com/9285dd3998997a0835869065bb15e5d500475034/687474703a2f2f617a7572656465706c6f792e6e65742f6465706c6f79627574746f6e2e706e67" data-canonical-src="http://azuredeploy.net/deploybutton.png" style="max-width:100%;">
</a>
### Using PowerShell
The following command sequence can prepare the environment in the West US 2 Azure region:
```ps1
Login-AzureRmAccount
New-AzureRmResourceGroup -Name docker -Location westus2
New-AzureRmResourceGroupDeployment -ResourceGroupName docker -TemplateFile .\infrastructure\arm-template.json -Name dsd
```
When finished, the following can tear down the environment:
```ps1
Remove-AzureRmResourceGroup -Name docker
``` | - Demonstration of developing, testing, and delivering production software with Docker
+ Demonstration of developing, testing, and delivering production software with Docker.
? +
+
+ ## Overview
+ The MEAN stack application code is used to demonstrate how to:
+ - Containerize a multi-tier application
+ - Create development and test environments with Docker Compose
+ - Create production environments using Docker Compose override files
+ - Release a new feature from development, through test, and into production
+
+ ## Getting Started
+ An Azure Resource Manager template is provided to create an environment with three virtual machines in a virtual network with Docker installed on each. Certificates are shared and setup for TLS secured communication between the three Docker hosts. The created environment resembles the following:
+ <img src="https://user-images.githubusercontent.com/3911650/27996794-58870aa8-64a7-11e7-8227-e1f137603dba.png" alt="Azure environment">
+
+ ### One-Click Deploy
+ <a href="https://portal.azure.com/#create/Microsoft.Template/uri/https%3A%2F%2Fraw.githubusercontent.com%2Flrakai%2Fdocker-software-delivery%2Fmaster%2Finfrastructure%2Farm-template.json">
+ <img src="https://camo.githubusercontent.com/9285dd3998997a0835869065bb15e5d500475034/687474703a2f2f617a7572656465706c6f792e6e65742f6465706c6f79627574746f6e2e706e67" data-canonical-src="http://azuredeploy.net/deploybutton.png" style="max-width:100%;">
+ </a>
+
+ ### Using PowerShell
+ The following command sequence can prepare the environment in the West US 2 Azure region:
+ ```ps1
+ Login-AzureRmAccount
+ New-AzureRmResourceGroup -Name docker -Location westus2
+ New-AzureRmResourceGroupDeployment -ResourceGroupName docker -TemplateFile .\infrastructure\arm-template.json -Name dsd
+ ```
+ When finished, the following can tear down the environment:
+ ```ps1
+ Remove-AzureRmResourceGroup -Name docker
+ ``` | 30 | 30 | 29 | 1 |
a9a5ad493bdaa1ae83731becdb301c58bde93cdb | db/migrate/20140113000001_create_simple_slug_history_slug.rb | db/migrate/20140113000001_create_simple_slug_history_slug.rb | class CreateSimpleSlugHistorySlug < ActiveRecord::Migration
def change
create_table :simple_slug_history_slugs do |t|
t.string :slug, null: false, limit: 191
t.integer :sluggable_id, null: false
t.string :sluggable_type, limit: 50, null: false
t.datetime :created_at
end
add_index :simple_slug_history_slugs, :slug
add_index :simple_slug_history_slugs, [:sluggable_type, :sluggable_id], name: 'simple_slug_history_slugs_on_sluggable_type_and_sluggable_id'
end
end
| class CreateSimpleSlugHistorySlug < ActiveRecord::Migration
def change
create_table :simple_slug_history_slugs do |t|
t.string :slug, null: false, limit: 191
t.string :locale, limit: 10
t.integer :sluggable_id, null: false
t.string :sluggable_type, limit: 50, null: false
t.timestamps
end
add_index :simple_slug_history_slugs, :slug
add_index :simple_slug_history_slugs, [:sluggable_type, :sluggable_id], name: 'simple_slug_history_slugs_on_sluggable_type_and_sluggable_id'
end
end
| Add locale to history slug migration | Add locale to history slug migration
| Ruby | mit | leschenko/simple_slug | ruby | ## Code Before:
class CreateSimpleSlugHistorySlug < ActiveRecord::Migration
def change
create_table :simple_slug_history_slugs do |t|
t.string :slug, null: false, limit: 191
t.integer :sluggable_id, null: false
t.string :sluggable_type, limit: 50, null: false
t.datetime :created_at
end
add_index :simple_slug_history_slugs, :slug
add_index :simple_slug_history_slugs, [:sluggable_type, :sluggable_id], name: 'simple_slug_history_slugs_on_sluggable_type_and_sluggable_id'
end
end
## Instruction:
Add locale to history slug migration
## Code After:
class CreateSimpleSlugHistorySlug < ActiveRecord::Migration
def change
create_table :simple_slug_history_slugs do |t|
t.string :slug, null: false, limit: 191
t.string :locale, limit: 10
t.integer :sluggable_id, null: false
t.string :sluggable_type, limit: 50, null: false
t.timestamps
end
add_index :simple_slug_history_slugs, :slug
add_index :simple_slug_history_slugs, [:sluggable_type, :sluggable_id], name: 'simple_slug_history_slugs_on_sluggable_type_and_sluggable_id'
end
end
| class CreateSimpleSlugHistorySlug < ActiveRecord::Migration
def change
create_table :simple_slug_history_slugs do |t|
t.string :slug, null: false, limit: 191
+ t.string :locale, limit: 10
t.integer :sluggable_id, null: false
t.string :sluggable_type, limit: 50, null: false
- t.datetime :created_at
+ t.timestamps
end
add_index :simple_slug_history_slugs, :slug
add_index :simple_slug_history_slugs, [:sluggable_type, :sluggable_id], name: 'simple_slug_history_slugs_on_sluggable_type_and_sluggable_id'
end
end | 3 | 0.230769 | 2 | 1 |
9fe73989da42cf45a9a317af465b04498b456e60 | INSTALL.md | INSTALL.md | ```bash
Android 5.0.1 or greater (>= 4.2 support coming soon)
An internet connection
```
## Installing a Nightly or Release Build
To install either the nightly or release build, simply navigate to the APK package below using either Firefox or Chrome on your Android device. If using Firefox, navigate to the "Downloads" menu and tap on the APK. If using Chrome, use the "Downloads" application and tap on the APK to begin the install process. If you have previously installed a version of the app from source, you may need to delete the app from your device prior to installing a nightly or release build.
#### Release – http://mzl.la/webmaker-release
#### Nightly – http://mzl.la/webmaker-nightly
## Installing From Source
Complete instructions for how to setup a local development environment and build using Android Studio can be found [here](https://github.com/mozilla/webmaker-android/blob/develop/README.md).
## Installing From Google Play
*Coming soon!*
## Contact Us
IRC: `#webmaker` on `irc.mozilla.org`
Forum: [https://groups.google.com/forum/#!forum/mozilla.webmaker](https://groups.google.com/forum/#!forum/mozilla.webmaker)
| ```bash
Android 5.0.1 or greater (>= 4.2 support coming soon)
An internet connection
```
## Enable Untrusted Sources
- Open the Android "Settings" application
- Navigate to "Applications"
- If the check box for "Unknown Sources" is unchecked, check it
## Installing a Nightly or Release Build
To install either the nightly or release build, simply navigate to the APK package below using either Firefox or Chrome on your Android device. If using Firefox, navigate to the "Downloads" menu and tap on the APK. If using Chrome, use the "Downloads" application and tap on the APK to begin the install process. If you have previously installed a version of the app from source, you may need to delete the app from your device prior to installing a nightly or release build.
#### Release – http://mzl.la/webmaker-release
#### Nightly – http://mzl.la/webmaker-nightly
## Installing From Source
Complete instructions for how to setup a local development environment and build using Android Studio can be found [here](https://github.com/mozilla/webmaker-android/blob/develop/README.md).
## Installing From Google Play
*Coming soon!*
## Contact Us
IRC: `#webmaker` on `irc.mozilla.org`
Forum: [https://groups.google.com/forum/#!forum/mozilla.webmaker](https://groups.google.com/forum/#!forum/mozilla.webmaker)
| Add 'untrusted sources' to install documentation | Add 'untrusted sources' to install documentation
| Markdown | mpl-2.0 | gvn/webmaker-android,k88hudson/webmaker-android,k88hudson/webmaker-android,rodmoreno/webmaker-android,bolaram/webmaker-android,j796160836/webmaker-android,alanmoo/webmaker-android,bolaram/webmaker-android,alicoding/webmaker-android,rodmoreno/webmaker-android,mozilla/webmaker-android,j796160836/webmaker-android,adamlofting/webmaker-android,alicoding/webmaker-android,gvn/webmaker-android,mozilla/webmaker-android,alanmoo/webmaker-android | markdown | ## Code Before:
```bash
Android 5.0.1 or greater (>= 4.2 support coming soon)
An internet connection
```
## Installing a Nightly or Release Build
To install either the nightly or release build, simply navigate to the APK package below using either Firefox or Chrome on your Android device. If using Firefox, navigate to the "Downloads" menu and tap on the APK. If using Chrome, use the "Downloads" application and tap on the APK to begin the install process. If you have previously installed a version of the app from source, you may need to delete the app from your device prior to installing a nightly or release build.
#### Release – http://mzl.la/webmaker-release
#### Nightly – http://mzl.la/webmaker-nightly
## Installing From Source
Complete instructions for how to setup a local development environment and build using Android Studio can be found [here](https://github.com/mozilla/webmaker-android/blob/develop/README.md).
## Installing From Google Play
*Coming soon!*
## Contact Us
IRC: `#webmaker` on `irc.mozilla.org`
Forum: [https://groups.google.com/forum/#!forum/mozilla.webmaker](https://groups.google.com/forum/#!forum/mozilla.webmaker)
## Instruction:
Add 'untrusted sources' to install documentation
## Code After:
```bash
Android 5.0.1 or greater (>= 4.2 support coming soon)
An internet connection
```
## Enable Untrusted Sources
- Open the Android "Settings" application
- Navigate to "Applications"
- If the check box for "Unknown Sources" is unchecked, check it
## Installing a Nightly or Release Build
To install either the nightly or release build, simply navigate to the APK package below using either Firefox or Chrome on your Android device. If using Firefox, navigate to the "Downloads" menu and tap on the APK. If using Chrome, use the "Downloads" application and tap on the APK to begin the install process. If you have previously installed a version of the app from source, you may need to delete the app from your device prior to installing a nightly or release build.
#### Release – http://mzl.la/webmaker-release
#### Nightly – http://mzl.la/webmaker-nightly
## Installing From Source
Complete instructions for how to setup a local development environment and build using Android Studio can be found [here](https://github.com/mozilla/webmaker-android/blob/develop/README.md).
## Installing From Google Play
*Coming soon!*
## Contact Us
IRC: `#webmaker` on `irc.mozilla.org`
Forum: [https://groups.google.com/forum/#!forum/mozilla.webmaker](https://groups.google.com/forum/#!forum/mozilla.webmaker)
| ```bash
Android 5.0.1 or greater (>= 4.2 support coming soon)
An internet connection
```
+
+ ## Enable Untrusted Sources
+ - Open the Android "Settings" application
+ - Navigate to "Applications"
+ - If the check box for "Unknown Sources" is unchecked, check it
## Installing a Nightly or Release Build
To install either the nightly or release build, simply navigate to the APK package below using either Firefox or Chrome on your Android device. If using Firefox, navigate to the "Downloads" menu and tap on the APK. If using Chrome, use the "Downloads" application and tap on the APK to begin the install process. If you have previously installed a version of the app from source, you may need to delete the app from your device prior to installing a nightly or release build.
#### Release – http://mzl.la/webmaker-release
#### Nightly – http://mzl.la/webmaker-nightly
## Installing From Source
Complete instructions for how to setup a local development environment and build using Android Studio can be found [here](https://github.com/mozilla/webmaker-android/blob/develop/README.md).
## Installing From Google Play
*Coming soon!*
## Contact Us
IRC: `#webmaker` on `irc.mozilla.org`
Forum: [https://groups.google.com/forum/#!forum/mozilla.webmaker](https://groups.google.com/forum/#!forum/mozilla.webmaker) | 5 | 0.238095 | 5 | 0 |
ff4e769102295280b9e5ad703c5b676f399df894 | test/test_basic.py | test/test_basic.py | import unittest
class MyTestCase(unittest.TestCase):
def test_something(self):
self.assertEqual(True, False)
if __name__ == '__main__':
unittest.main()
| import unittest
import openfigi
class MyTestCase(unittest.TestCase):
def test_wkn_ticker_anonymous(self):
"""Get an ETF by WKN and check if response makes sense"""
ofg = openfigi.OpenFigi()
ofg.enqueue_request(id_type='ID_WERTPAPIER', id_value='A0YEDG')
response = ofg.fetch_response()
self.assertTrue(type(response) is list)
self.assertTrue(len(response) > 0)
self.assertTrue(type(response[0]) is dict)
self.assertTrue('data' in response[0].keys())
self.assertTrue(len(response[0]['data']) > 0)
if __name__ == '__main__':
unittest.main()
| Add a basic unit test | Add a basic unit test
| Python | mit | jwergieluk/openfigi,jwergieluk/openfigi | python | ## Code Before:
import unittest
class MyTestCase(unittest.TestCase):
def test_something(self):
self.assertEqual(True, False)
if __name__ == '__main__':
unittest.main()
## Instruction:
Add a basic unit test
## Code After:
import unittest
import openfigi
class MyTestCase(unittest.TestCase):
def test_wkn_ticker_anonymous(self):
"""Get an ETF by WKN and check if response makes sense"""
ofg = openfigi.OpenFigi()
ofg.enqueue_request(id_type='ID_WERTPAPIER', id_value='A0YEDG')
response = ofg.fetch_response()
self.assertTrue(type(response) is list)
self.assertTrue(len(response) > 0)
self.assertTrue(type(response[0]) is dict)
self.assertTrue('data' in response[0].keys())
self.assertTrue(len(response[0]['data']) > 0)
if __name__ == '__main__':
unittest.main()
| import unittest
+ import openfigi
class MyTestCase(unittest.TestCase):
- def test_something(self):
- self.assertEqual(True, False)
+ def test_wkn_ticker_anonymous(self):
+ """Get an ETF by WKN and check if response makes sense"""
+ ofg = openfigi.OpenFigi()
+ ofg.enqueue_request(id_type='ID_WERTPAPIER', id_value='A0YEDG')
+ response = ofg.fetch_response()
+ self.assertTrue(type(response) is list)
+ self.assertTrue(len(response) > 0)
+ self.assertTrue(type(response[0]) is dict)
+ self.assertTrue('data' in response[0].keys())
+ self.assertTrue(len(response[0]['data']) > 0)
if __name__ == '__main__':
unittest.main() | 13 | 1.3 | 11 | 2 |
36ff97bd836fc1e4b8ca63ec3d3cb168c8bbc5ee | src/two-columns.js | src/two-columns.js | export const layoutWrapper = {
display: "flex",
justifyContent: "space-between",
};
export const sideBar = {
marginLeft: "35px",
marginTop: "25px",
order: "1",
flex: "1 6 25%",
};
export const mainContent = {
order: "2",
flex: "3 1 75%",
}; | export const layoutWrapper = {
display: "flex",
justifyContent: "space-between",
};
export const sideBar = {
marginLeft: "35px",
marginTop: "25px",
order: "1",
flex: "1 6 35%",
};
export const mainContent = {
marginTop: "25px",
order: "2",
flex: "3 1 65%",
}; | Make sidebar wider in 2-column layout | Make sidebar wider in 2-column layout
| JavaScript | mit | orouvinen/mytype-frontend,orouvinen/mytype-frontend | javascript | ## Code Before:
export const layoutWrapper = {
display: "flex",
justifyContent: "space-between",
};
export const sideBar = {
marginLeft: "35px",
marginTop: "25px",
order: "1",
flex: "1 6 25%",
};
export const mainContent = {
order: "2",
flex: "3 1 75%",
};
## Instruction:
Make sidebar wider in 2-column layout
## Code After:
export const layoutWrapper = {
display: "flex",
justifyContent: "space-between",
};
export const sideBar = {
marginLeft: "35px",
marginTop: "25px",
order: "1",
flex: "1 6 35%",
};
export const mainContent = {
marginTop: "25px",
order: "2",
flex: "3 1 65%",
}; | export const layoutWrapper = {
display: "flex",
justifyContent: "space-between",
};
export const sideBar = {
marginLeft: "35px",
marginTop: "25px",
order: "1",
- flex: "1 6 25%",
? ^
+ flex: "1 6 35%",
? ^
};
export const mainContent = {
+ marginTop: "25px",
order: "2",
- flex: "3 1 75%",
? ^
+ flex: "3 1 65%",
? ^
}; | 5 | 0.3125 | 3 | 2 |
86d2fda9905f719370271f30934ee39bcb5cec62 | genautomake.sh | genautomake.sh |
aclocal
libtoolize=libtoolize
if which glibtoolize > /dev/null 2>&1; then
libtoolize=glibtoolize
fi
$libtoolize --copy --force
autoheader
automake --add-missing --copy --force-missing
autoconf
# hack to nuke rpath:
# http://wiki.debian.org/RpathIssue
# http://fedoraproject.org/wiki/RPath_Packaging_Draft
ed libtool > /dev/null <<EOF
,s|^hardcode_libdir_flag_spec=.*|hardcode_libdir_flag_spec=""|g
,s|^runpath_var=LD_RUN_PATH|runpath_var=DIE_RPATH_DIE|g
wq
EOF
|
aclocal
libtoolize=libtoolize
if which glibtoolize > /dev/null 2>&1; then
libtoolize=glibtoolize
fi
$libtoolize --copy --force
autoheader
automake --add-missing --copy --force-missing
autoconf
| Revert "avoid using rpath when linking" | Revert "avoid using rpath when linking"
This reverts commit 27641444173e3a21b531d0384ad612a76837b90c.
| Shell | mit | rclasen/srmio,rclasen/srmio,rclasen/srmio,rclasen/srmio | shell | ## Code Before:
aclocal
libtoolize=libtoolize
if which glibtoolize > /dev/null 2>&1; then
libtoolize=glibtoolize
fi
$libtoolize --copy --force
autoheader
automake --add-missing --copy --force-missing
autoconf
# hack to nuke rpath:
# http://wiki.debian.org/RpathIssue
# http://fedoraproject.org/wiki/RPath_Packaging_Draft
ed libtool > /dev/null <<EOF
,s|^hardcode_libdir_flag_spec=.*|hardcode_libdir_flag_spec=""|g
,s|^runpath_var=LD_RUN_PATH|runpath_var=DIE_RPATH_DIE|g
wq
EOF
## Instruction:
Revert "avoid using rpath when linking"
This reverts commit 27641444173e3a21b531d0384ad612a76837b90c.
## Code After:
aclocal
libtoolize=libtoolize
if which glibtoolize > /dev/null 2>&1; then
libtoolize=glibtoolize
fi
$libtoolize --copy --force
autoheader
automake --add-missing --copy --force-missing
autoconf
|
aclocal
libtoolize=libtoolize
if which glibtoolize > /dev/null 2>&1; then
libtoolize=glibtoolize
fi
$libtoolize --copy --force
autoheader
automake --add-missing --copy --force-missing
autoconf
- # hack to nuke rpath:
- # http://wiki.debian.org/RpathIssue
- # http://fedoraproject.org/wiki/RPath_Packaging_Draft
- ed libtool > /dev/null <<EOF
- ,s|^hardcode_libdir_flag_spec=.*|hardcode_libdir_flag_spec=""|g
- ,s|^runpath_var=LD_RUN_PATH|runpath_var=DIE_RPATH_DIE|g
- wq
- EOF
- | 9 | 0.428571 | 0 | 9 |
6cd3a9056373b2decea6472a77823fbfbee54a1a | tests/unit/DrupalTestUserTest.php | tests/unit/DrupalTestUserTest.php | <?php
use \Codeception\Module\Drupal\UserRegistry\DrupalTestUser;
use \Codeception\Util\Fixtures;
/**
* Unit tests for DrushTestUserManager class.
*/
class DrupalTestUserTest extends \Codeception\TestCase\Test
{
/**
* @var \UnitTester
* Store the Actor object being used to test.
*/
protected $tester;
/**
* Objects of this class should be instantiable.
*
* @test
*/
public function shouldBeInstantiatable()
{
$user = Fixtures::get("drupalTestUser");
$this->assertInstanceOf(
'\Codeception\Module\Drupal\UserRegistry\DrupalTestUser',
new DrupalTestUser($user->name, $user->pass)
);
}
}
| <?php
use \Codeception\Module\Drupal\UserRegistry\DrupalTestUser;
use \Codeception\Util\Fixtures;
/**
* Unit tests for DrushTestUserManager class.
*/
class DrupalTestUserTest extends \Codeception\TestCase\Test
{
/**
* @var \UnitTester
* Store the Actor object being used to test.
*/
protected $tester;
/**
* Objects of this class should be instantiable.
*
* @test
*/
public function shouldBeInstantiatable()
{
$user = Fixtures::get("drupalTestUser");
$this->assertInstanceOf(
'\Codeception\Module\Drupal\UserRegistry\DrupalTestUser',
new DrupalTestUser($user->name, $user->pass)
);
}
/**
* Test the class __toString() method.
*
* @test
*/
public function testToString()
{
$user = new DrupalTestUser("Test Username", "password");
$this->assertEquals("Test Username", $user->__toString());
$this->assertEquals("Test Username", $user . "");
}
}
| Test the __toString() method of DrupalTestUser class. | Test the __toString() method of DrupalTestUser class.
| PHP | mit | pfaocle/codeception-module-drupal-user-registry,chriscohen/codeception-module-drupal-user-registry,ixis/codeception-module-drupal-user-registry | php | ## Code Before:
<?php
use \Codeception\Module\Drupal\UserRegistry\DrupalTestUser;
use \Codeception\Util\Fixtures;
/**
* Unit tests for DrushTestUserManager class.
*/
class DrupalTestUserTest extends \Codeception\TestCase\Test
{
/**
* @var \UnitTester
* Store the Actor object being used to test.
*/
protected $tester;
/**
* Objects of this class should be instantiable.
*
* @test
*/
public function shouldBeInstantiatable()
{
$user = Fixtures::get("drupalTestUser");
$this->assertInstanceOf(
'\Codeception\Module\Drupal\UserRegistry\DrupalTestUser',
new DrupalTestUser($user->name, $user->pass)
);
}
}
## Instruction:
Test the __toString() method of DrupalTestUser class.
## Code After:
<?php
use \Codeception\Module\Drupal\UserRegistry\DrupalTestUser;
use \Codeception\Util\Fixtures;
/**
* Unit tests for DrushTestUserManager class.
*/
class DrupalTestUserTest extends \Codeception\TestCase\Test
{
/**
* @var \UnitTester
* Store the Actor object being used to test.
*/
protected $tester;
/**
* Objects of this class should be instantiable.
*
* @test
*/
public function shouldBeInstantiatable()
{
$user = Fixtures::get("drupalTestUser");
$this->assertInstanceOf(
'\Codeception\Module\Drupal\UserRegistry\DrupalTestUser',
new DrupalTestUser($user->name, $user->pass)
);
}
/**
* Test the class __toString() method.
*
* @test
*/
public function testToString()
{
$user = new DrupalTestUser("Test Username", "password");
$this->assertEquals("Test Username", $user->__toString());
$this->assertEquals("Test Username", $user . "");
}
}
| <?php
use \Codeception\Module\Drupal\UserRegistry\DrupalTestUser;
use \Codeception\Util\Fixtures;
/**
* Unit tests for DrushTestUserManager class.
*/
class DrupalTestUserTest extends \Codeception\TestCase\Test
{
/**
* @var \UnitTester
* Store the Actor object being used to test.
*/
protected $tester;
/**
* Objects of this class should be instantiable.
*
* @test
*/
public function shouldBeInstantiatable()
{
$user = Fixtures::get("drupalTestUser");
$this->assertInstanceOf(
'\Codeception\Module\Drupal\UserRegistry\DrupalTestUser',
new DrupalTestUser($user->name, $user->pass)
);
}
+
+ /**
+ * Test the class __toString() method.
+ *
+ * @test
+ */
+ public function testToString()
+ {
+ $user = new DrupalTestUser("Test Username", "password");
+ $this->assertEquals("Test Username", $user->__toString());
+ $this->assertEquals("Test Username", $user . "");
+ }
} | 12 | 0.4 | 12 | 0 |
ff87177f52e577de48cb10ffdfa1470a2279ba93 | create-widget.sh | create-widget.sh |
if [ -e ./proximityreminder.wgt ]; then
rm -v proximityreminder.wgt
fi
# Zip all the html, javascript, CSS, images and other information.
zip proximityreminder.wgt *.html ./js/*.js ./css/*.css ./css/images/* config.xml remind_me.png
if [ -e ./proximityreminder-nowebinosjs.wgt ]; then
rm -v proximityreminder-nowebinosjs.wgt
fi
# Do it again, but this time without webinos.js
zip proximityreminder-nowebinosjs.wgt *.html ./js/*.js ./css/*.css ./css/images/* config.xml remind_me.png -x ./js/webinos.js
|
if [ -e ./proximityreminder.wgt ]; then
rm -v proximityreminder.wgt
fi
# Zip all the html, javascript, CSS, images and other information.
zip -r proximityreminder.wgt *.html ./js/*.js ./css/* config.xml remind_me.png -x *~ -x */*~
if [ -e ./proximityreminder-nowebinosjs.wgt ]; then
rm -v proximityreminder-nowebinosjs.wgt
fi
# Do it again, but this time without webinos.js
zip -r proximityreminder-nowebinosjs.wgt *.html ./js/*.js ./css/* config.xml remind_me.png -x ./js/webinos.js -x *~ -x */*~
| Fix the widget creation script to respect directory structure | Fix the widget creation script to respect directory structure
The widget creation script now uses the -r flag and removes
all temporary files
Jira issue: APPPR-17
| Shell | apache-2.0 | webinos-apps/app-proximity-reminders,webinos-apps/app-proximity-reminders | shell | ## Code Before:
if [ -e ./proximityreminder.wgt ]; then
rm -v proximityreminder.wgt
fi
# Zip all the html, javascript, CSS, images and other information.
zip proximityreminder.wgt *.html ./js/*.js ./css/*.css ./css/images/* config.xml remind_me.png
if [ -e ./proximityreminder-nowebinosjs.wgt ]; then
rm -v proximityreminder-nowebinosjs.wgt
fi
# Do it again, but this time without webinos.js
zip proximityreminder-nowebinosjs.wgt *.html ./js/*.js ./css/*.css ./css/images/* config.xml remind_me.png -x ./js/webinos.js
## Instruction:
Fix the widget creation script to respect directory structure
The widget creation script now uses the -r flag and removes
all temporary files
Jira issue: APPPR-17
## Code After:
if [ -e ./proximityreminder.wgt ]; then
rm -v proximityreminder.wgt
fi
# Zip all the html, javascript, CSS, images and other information.
zip -r proximityreminder.wgt *.html ./js/*.js ./css/* config.xml remind_me.png -x *~ -x */*~
if [ -e ./proximityreminder-nowebinosjs.wgt ]; then
rm -v proximityreminder-nowebinosjs.wgt
fi
# Do it again, but this time without webinos.js
zip -r proximityreminder-nowebinosjs.wgt *.html ./js/*.js ./css/* config.xml remind_me.png -x ./js/webinos.js -x *~ -x */*~
|
if [ -e ./proximityreminder.wgt ]; then
rm -v proximityreminder.wgt
fi
# Zip all the html, javascript, CSS, images and other information.
- zip proximityreminder.wgt *.html ./js/*.js ./css/*.css ./css/images/* config.xml remind_me.png
? -------------------
+ zip -r proximityreminder.wgt *.html ./js/*.js ./css/* config.xml remind_me.png -x *~ -x */*~
? +++ ++++++++++++++
if [ -e ./proximityreminder-nowebinosjs.wgt ]; then
rm -v proximityreminder-nowebinosjs.wgt
fi
# Do it again, but this time without webinos.js
- zip proximityreminder-nowebinosjs.wgt *.html ./js/*.js ./css/*.css ./css/images/* config.xml remind_me.png -x ./js/webinos.js
? -------------------
+ zip -r proximityreminder-nowebinosjs.wgt *.html ./js/*.js ./css/* config.xml remind_me.png -x ./js/webinos.js -x *~ -x */*~
? +++ ++++++++++++++
| 4 | 0.285714 | 2 | 2 |
463eac252b2e0ba16db908fe366fe372c4853517 | docker-plugin/src/main/java/com/nirima/jenkins/plugins/docker/utils/LogUtils.java | docker-plugin/src/main/java/com/nirima/jenkins/plugins/docker/utils/LogUtils.java | package com.nirima.jenkins.plugins.docker.utils;
import com.github.dockerjava.api.model.ResponseItem;
import hudson.model.BuildListener;
/**
* @author Kanstantsin Shautsou
*/
public class LogUtils {
private LogUtils() {
}
public static void printResponseItemToListener(BuildListener listener, ResponseItem item) {
if (item.getStatus() != null) {
if (item.getError() != null) {
listener.error(item.getError());
}
final StringBuilder stringBuffer = new StringBuilder();
if (item.getId() != null) {
stringBuffer.append(item.getId()).append(": "); // Doesn't exist before "Digest"
}
stringBuffer.append(item.getStatus());
if (item.getProgress() != null) {
stringBuffer.append(" ").append(item.getProgress());
}
listener.getLogger().println(stringBuffer.toString());
}
}
}
| package com.nirima.jenkins.plugins.docker.utils;
import com.github.dockerjava.api.model.ResponseItem;
import hudson.model.BuildListener;
import hudson.model.TaskListener;
/**
* @author Kanstantsin Shautsou
*/
public class LogUtils {
private LogUtils() {
}
public static void printResponseItemToListener(TaskListener listener, ResponseItem item) {
if (item.getStatus() != null) {
if (item.getError() != null) {
listener.error(item.getError());
}
final StringBuilder stringBuffer = new StringBuilder();
if (item.getId() != null) {
stringBuffer.append(item.getId()).append(": "); // Doesn't exist before "Digest"
}
stringBuffer.append(item.getStatus());
if (item.getProgress() != null) {
stringBuffer.append(" ").append(item.getProgress());
}
listener.getLogger().println(stringBuffer.toString());
}
}
}
| Make logging take superclass so not tied to Builds only. | Make logging take superclass so not tied to Builds only.
Signed-off-by: Nigel Magnay <0c422baef524090650d5dc166beb1e8269e8fa23@gmail.com>
| Java | mit | eaglerainbow/docker-plugin,jenkinsci/docker-plugin,jenkinsci/docker-plugin,jenkinsci/docker-plugin,eaglerainbow/docker-plugin,eaglerainbow/docker-plugin | java | ## Code Before:
package com.nirima.jenkins.plugins.docker.utils;
import com.github.dockerjava.api.model.ResponseItem;
import hudson.model.BuildListener;
/**
* @author Kanstantsin Shautsou
*/
public class LogUtils {
private LogUtils() {
}
public static void printResponseItemToListener(BuildListener listener, ResponseItem item) {
if (item.getStatus() != null) {
if (item.getError() != null) {
listener.error(item.getError());
}
final StringBuilder stringBuffer = new StringBuilder();
if (item.getId() != null) {
stringBuffer.append(item.getId()).append(": "); // Doesn't exist before "Digest"
}
stringBuffer.append(item.getStatus());
if (item.getProgress() != null) {
stringBuffer.append(" ").append(item.getProgress());
}
listener.getLogger().println(stringBuffer.toString());
}
}
}
## Instruction:
Make logging take superclass so not tied to Builds only.
Signed-off-by: Nigel Magnay <0c422baef524090650d5dc166beb1e8269e8fa23@gmail.com>
## Code After:
package com.nirima.jenkins.plugins.docker.utils;
import com.github.dockerjava.api.model.ResponseItem;
import hudson.model.BuildListener;
import hudson.model.TaskListener;
/**
* @author Kanstantsin Shautsou
*/
public class LogUtils {
private LogUtils() {
}
public static void printResponseItemToListener(TaskListener listener, ResponseItem item) {
if (item.getStatus() != null) {
if (item.getError() != null) {
listener.error(item.getError());
}
final StringBuilder stringBuffer = new StringBuilder();
if (item.getId() != null) {
stringBuffer.append(item.getId()).append(": "); // Doesn't exist before "Digest"
}
stringBuffer.append(item.getStatus());
if (item.getProgress() != null) {
stringBuffer.append(" ").append(item.getProgress());
}
listener.getLogger().println(stringBuffer.toString());
}
}
}
| package com.nirima.jenkins.plugins.docker.utils;
import com.github.dockerjava.api.model.ResponseItem;
import hudson.model.BuildListener;
+ import hudson.model.TaskListener;
/**
* @author Kanstantsin Shautsou
*/
public class LogUtils {
private LogUtils() {
}
- public static void printResponseItemToListener(BuildListener listener, ResponseItem item) {
? ^^^^^
+ public static void printResponseItemToListener(TaskListener listener, ResponseItem item) {
? ^^^^
if (item.getStatus() != null) {
if (item.getError() != null) {
listener.error(item.getError());
}
final StringBuilder stringBuffer = new StringBuilder();
if (item.getId() != null) {
stringBuffer.append(item.getId()).append(": "); // Doesn't exist before "Digest"
}
stringBuffer.append(item.getStatus());
if (item.getProgress() != null) {
stringBuffer.append(" ").append(item.getProgress());
}
listener.getLogger().println(stringBuffer.toString());
}
}
} | 3 | 0.088235 | 2 | 1 |
3fbf6fb486b80c1b73542a26a2c0974c33089bfa | .gitlab-ci.yml | .gitlab-ci.yml | image: node:8.1
cache:
paths:
- node_modules/
- public/content
- ~/.yarn-cache
stages:
- test
- deploy
test:
stage: test
script:
- rm -rf node_modules
- yarn install
- npm run test-with-coverage
deploy_production:
stage: deploy
script:
- rm -rf node_modules
- yarn install
- npm run sync-content
- npm run build
- npm run build:minify
- npm run deploy:production
only:
- master
| image: node:8.1
cache:
paths:
- node_modules/
- public/content
- ~/.yarn-cache
stages:
- test
- deploy
test:
stage: test
script:
- yarn install
- npm run test-with-coverage
deploy_production:
stage: deploy
script:
- yarn install
- npm run sync-content
- npm run build
- npm run build:minify
- npm run deploy:production
only:
- master
| Remove node_modules removal in CI | Remove node_modules removal in CI
Was introduced to clear the caches in gitlab ci (no manual way yet)
| YAML | mit | reality-scheveningen/reality-website,reality-scheveningen/reality-website | yaml | ## Code Before:
image: node:8.1
cache:
paths:
- node_modules/
- public/content
- ~/.yarn-cache
stages:
- test
- deploy
test:
stage: test
script:
- rm -rf node_modules
- yarn install
- npm run test-with-coverage
deploy_production:
stage: deploy
script:
- rm -rf node_modules
- yarn install
- npm run sync-content
- npm run build
- npm run build:minify
- npm run deploy:production
only:
- master
## Instruction:
Remove node_modules removal in CI
Was introduced to clear the caches in gitlab ci (no manual way yet)
## Code After:
image: node:8.1
cache:
paths:
- node_modules/
- public/content
- ~/.yarn-cache
stages:
- test
- deploy
test:
stage: test
script:
- yarn install
- npm run test-with-coverage
deploy_production:
stage: deploy
script:
- yarn install
- npm run sync-content
- npm run build
- npm run build:minify
- npm run deploy:production
only:
- master
| image: node:8.1
cache:
paths:
- node_modules/
- public/content
- ~/.yarn-cache
stages:
- test
- deploy
test:
stage: test
script:
- - rm -rf node_modules
- yarn install
- npm run test-with-coverage
deploy_production:
stage: deploy
script:
- - rm -rf node_modules
- yarn install
- npm run sync-content
- npm run build
- npm run build:minify
- npm run deploy:production
only:
- master | 2 | 0.066667 | 0 | 2 |
ac6f9a9e666229a7571839bb1c2dd06ac4c3dc24 | app/controllers/backend/application_controller.rb | app/controllers/backend/application_controller.rb | module Backend
class ApplicationController < InheritedResources::Base
helper_method :attributes
before_filter :add_breadcrumbs
def attributes
resource_class.attribute_names - %w(id created_at updated_at)
end
protected
def collection
@q = end_of_association_chain.search(params[:q])
get_collection_ivar || set_collection_ivar(@q.result(distinct: true).paginate(:page => params[:page]))
end
def paths
request.path.split("/")[1..-1]
end
def add_breadcrumbs
url = ""
paths.each do |path|
url = url + "/" + path
add_breadcrumb path, url
end
end
end
end
| module Backend
class ApplicationController < InheritedResources::Base
helper_method :attributes
before_filter :add_breadcrumbs
def attributes
resource_class.attribute_names - %w(id created_at updated_at)
end
protected
def collection
@q = end_of_association_chain.search(params[:q])
get_collection_ivar || set_collection_ivar(@q.result(distinct: true).paginate(:page => params[:page]))
end
def paths
request.path.split("/")[1..-1]
end
def add_breadcrumbs
if paths
url = ""
paths.each do |path|
url = url + "/" + path
add_breadcrumb path, url
end
end
end
end
end
| Fix breadcrumbs when there is no path | Fix breadcrumbs when there is no path | Ruby | agpl-3.0 | bluelemons/rails_backend,bluelemons/rails_backend | ruby | ## Code Before:
module Backend
class ApplicationController < InheritedResources::Base
helper_method :attributes
before_filter :add_breadcrumbs
def attributes
resource_class.attribute_names - %w(id created_at updated_at)
end
protected
def collection
@q = end_of_association_chain.search(params[:q])
get_collection_ivar || set_collection_ivar(@q.result(distinct: true).paginate(:page => params[:page]))
end
def paths
request.path.split("/")[1..-1]
end
def add_breadcrumbs
url = ""
paths.each do |path|
url = url + "/" + path
add_breadcrumb path, url
end
end
end
end
## Instruction:
Fix breadcrumbs when there is no path
## Code After:
module Backend
class ApplicationController < InheritedResources::Base
helper_method :attributes
before_filter :add_breadcrumbs
def attributes
resource_class.attribute_names - %w(id created_at updated_at)
end
protected
def collection
@q = end_of_association_chain.search(params[:q])
get_collection_ivar || set_collection_ivar(@q.result(distinct: true).paginate(:page => params[:page]))
end
def paths
request.path.split("/")[1..-1]
end
def add_breadcrumbs
if paths
url = ""
paths.each do |path|
url = url + "/" + path
add_breadcrumb path, url
end
end
end
end
end
| module Backend
class ApplicationController < InheritedResources::Base
helper_method :attributes
before_filter :add_breadcrumbs
def attributes
resource_class.attribute_names - %w(id created_at updated_at)
end
protected
def collection
@q = end_of_association_chain.search(params[:q])
get_collection_ivar || set_collection_ivar(@q.result(distinct: true).paginate(:page => params[:page]))
end
def paths
request.path.split("/")[1..-1]
end
def add_breadcrumbs
+ if paths
- url = ""
+ url = ""
? ++
- paths.each do |path|
+ paths.each do |path|
? ++
- url = url + "/" + path
+ url = url + "/" + path
? ++
- add_breadcrumb path, url
+ add_breadcrumb path, url
? ++
+ end
end
end
end
end | 10 | 0.333333 | 6 | 4 |
af31d9c6a3484aaf356c4bf00a6b985fc0c96e08 | lib/skroutz_api/paginated_collection.rb | lib/skroutz_api/paginated_collection.rb | class SkroutzApi::PaginatedCollection < Array
include SkroutzApi::Parsing
delegate :model_name, :client, to: :context
attr_reader :response, :context
def initialize(context, response)
@context = context
@response = response
super(parse_array(response, context.resource_prefix))
end
def is_at_first?
!first?
end
def is_at_last?
!last?
end
%w[first last next previous].each do |meth|
define_method "#{meth}?" do
link_header(response).key? meth.to_sym
end
define_method "#{meth}_page" do
return [] if !self.send("#{meth}?")
target_query = URI.parse(link_header(response)[meth.to_sym]).query
target_uri = URI.parse(context.resource_prefix)
target_uri.query = target_query
response = context.client.get(target_uri)
return parse(response) unless block_given?
yield response
end
end
end
| class SkroutzApi::PaginatedCollection < Array
include SkroutzApi::Parsing
delegate :model_name, :client, to: :context
attr_reader :response, :context
def initialize(context, response, collection)
@context = context
@response = response
super(collection)
end
def is_at_first?
!first?
end
def is_at_last?
!last?
end
%w[first last next previous].each do |meth|
define_method "#{meth}?" do
link_header(response).key? meth.to_sym
end
define_method "#{meth}_page" do
return [] if !self.send("#{meth}?")
target_query = URI.parse(link_header(response)[meth.to_sym]).query
target_uri = URI.parse(context.resource_prefix)
target_uri.query = target_query
response = context.client.get(target_uri)
return parse(response) unless block_given?
yield response
end
end
end
| Initialize SkroutzApi::PaginatedCollection with parsed collection | Initialize SkroutzApi::PaginatedCollection with parsed collection
| Ruby | mit | skroutz/skroutz.rb | ruby | ## Code Before:
class SkroutzApi::PaginatedCollection < Array
include SkroutzApi::Parsing
delegate :model_name, :client, to: :context
attr_reader :response, :context
def initialize(context, response)
@context = context
@response = response
super(parse_array(response, context.resource_prefix))
end
def is_at_first?
!first?
end
def is_at_last?
!last?
end
%w[first last next previous].each do |meth|
define_method "#{meth}?" do
link_header(response).key? meth.to_sym
end
define_method "#{meth}_page" do
return [] if !self.send("#{meth}?")
target_query = URI.parse(link_header(response)[meth.to_sym]).query
target_uri = URI.parse(context.resource_prefix)
target_uri.query = target_query
response = context.client.get(target_uri)
return parse(response) unless block_given?
yield response
end
end
end
## Instruction:
Initialize SkroutzApi::PaginatedCollection with parsed collection
## Code After:
class SkroutzApi::PaginatedCollection < Array
include SkroutzApi::Parsing
delegate :model_name, :client, to: :context
attr_reader :response, :context
def initialize(context, response, collection)
@context = context
@response = response
super(collection)
end
def is_at_first?
!first?
end
def is_at_last?
!last?
end
%w[first last next previous].each do |meth|
define_method "#{meth}?" do
link_header(response).key? meth.to_sym
end
define_method "#{meth}_page" do
return [] if !self.send("#{meth}?")
target_query = URI.parse(link_header(response)[meth.to_sym]).query
target_uri = URI.parse(context.resource_prefix)
target_uri.query = target_query
response = context.client.get(target_uri)
return parse(response) unless block_given?
yield response
end
end
end
| class SkroutzApi::PaginatedCollection < Array
include SkroutzApi::Parsing
delegate :model_name, :client, to: :context
attr_reader :response, :context
- def initialize(context, response)
+ def initialize(context, response, collection)
? ++++++++++++
@context = context
@response = response
- super(parse_array(response, context.resource_prefix))
+
+ super(collection)
end
def is_at_first?
!first?
end
def is_at_last?
!last?
end
%w[first last next previous].each do |meth|
define_method "#{meth}?" do
link_header(response).key? meth.to_sym
end
define_method "#{meth}_page" do
return [] if !self.send("#{meth}?")
target_query = URI.parse(link_header(response)[meth.to_sym]).query
target_uri = URI.parse(context.resource_prefix)
target_uri.query = target_query
response = context.client.get(target_uri)
return parse(response) unless block_given?
yield response
end
end
end | 5 | 0.125 | 3 | 2 |
da1046c5a303edf8964882dfdb7eb2807b4adb40 | lib/daimon_skycrawlers/processor/default.rb | lib/daimon_skycrawlers/processor/default.rb | require "daimon_skycrawlers/storage/rdb"
require "daimon_skycrawlers/processor/base"
module DaimonSkycrawlers
module Processor
class Default < Base
def call(message)
url = message[:url]
page = storage.find(url)
headers = JSON.parse(page.headers)
headers_string = headers.map {|key, value| " #{key}: #{value}" }.join("\n")
message = <<~LOG
URL: #{page.url}
Body: #{page.body.bytesize} bytes
Headers:
#{headers_string}
LOG
log.info(message)
end
end
end
end
| require "daimon_skycrawlers/storage/rdb"
require "daimon_skycrawlers/processor/base"
module DaimonSkycrawlers
module Processor
class Default < Base
def call(message)
url = message[:url]
page = storage.find(url)
headers = JSON.parse(page.headers)
headers_string = headers.map {|key, value| " #{key}: #{value}" }.join("\n")
message = <<LOG
URL: #{page.url}
Body: #{page.body.bytesize} bytes
Headers:
#{headers_string}
LOG
log.info(message)
end
end
end
end
| Use plain here document to support Ruby 2.2 | Use plain here document to support Ruby 2.2
| Ruby | mit | bm-sms/daimon-skycrawlers,bm-sms/daimon-skycrawlers,bm-sms/daimon_skycrawlers,yucao24hours/daimon-skycrawlers,bm-sms/daimon_skycrawlers,bm-sms/daimon_skycrawlers,yucao24hours/daimon-skycrawlers,yucao24hours/daimon-skycrawlers | ruby | ## Code Before:
require "daimon_skycrawlers/storage/rdb"
require "daimon_skycrawlers/processor/base"
module DaimonSkycrawlers
module Processor
class Default < Base
def call(message)
url = message[:url]
page = storage.find(url)
headers = JSON.parse(page.headers)
headers_string = headers.map {|key, value| " #{key}: #{value}" }.join("\n")
message = <<~LOG
URL: #{page.url}
Body: #{page.body.bytesize} bytes
Headers:
#{headers_string}
LOG
log.info(message)
end
end
end
end
## Instruction:
Use plain here document to support Ruby 2.2
## Code After:
require "daimon_skycrawlers/storage/rdb"
require "daimon_skycrawlers/processor/base"
module DaimonSkycrawlers
module Processor
class Default < Base
def call(message)
url = message[:url]
page = storage.find(url)
headers = JSON.parse(page.headers)
headers_string = headers.map {|key, value| " #{key}: #{value}" }.join("\n")
message = <<LOG
URL: #{page.url}
Body: #{page.body.bytesize} bytes
Headers:
#{headers_string}
LOG
log.info(message)
end
end
end
end
| require "daimon_skycrawlers/storage/rdb"
require "daimon_skycrawlers/processor/base"
module DaimonSkycrawlers
module Processor
class Default < Base
def call(message)
url = message[:url]
page = storage.find(url)
headers = JSON.parse(page.headers)
headers_string = headers.map {|key, value| " #{key}: #{value}" }.join("\n")
- message = <<~LOG
? -
+ message = <<LOG
- URL: #{page.url}
? ----------
+ URL: #{page.url}
- Body: #{page.body.bytesize} bytes
? ----------
+ Body: #{page.body.bytesize} bytes
- Headers:
+ Headers:
- #{headers_string}
? ----------
+ #{headers_string}
- LOG
+ LOG
log.info(message)
end
end
end
end | 12 | 0.545455 | 6 | 6 |
2b1052a59ca3bc5941bd399a20c6d12bfa4ca696 | config-management/roles/harrow.lxd-host/tasks/tools.yml | config-management/roles/harrow.lxd-host/tasks/tools.yml | ---
- apt:
name: "{{ item }}"
update_cache: true
with_items:
- python3-requests
- python3-dateutil
- name: copy lxd-kill-old script
template:
src: usr/local/bin/lxd-kill-old.py.j2
dest: /usr/local/bin/lxd-kill-old
owner: root
group: root
mode: 0700
tags:
- lxd
- name: copy report-container-info-to-influxdb script
template:
src: usr/local/bin/report-container-info-to-influxdb.py.j2
dest: /usr/local/bin/report-container-info-to-influxdb
owner: root
group: root
mode: 0700
tags:
- lxd
- name: set MAILTO to nothing in crontab
cron:
name: MAILTO
minute: MAILTO=""
hour: ""
day: ""
month: ""
weekday: ""
job: ""
- name: periodically kill old containers
cron:
name: "kill too-old containers"
minute: "*/5"
job: "/usr/local/bin/lxd-kill-old kill"
tags:
- lxd
- name: report container info to influxdb
cron:
name: "report-container-info-to-influxdb"
minute: "*/5"
job: "/usr/local/bin/report-container-info-to-influxdb"
tags:
- lxd
| ---
- apt:
name: "{{ item }}"
update_cache: true
with_items:
- python3-requests
- python3-dateutil
- name: copy lxd-kill-old script
template:
src: usr/local/bin/lxd-kill-old.py.j2
dest: /usr/local/bin/lxd-kill-old
owner: root
group: root
mode: 0700
tags:
- lxd
- name: copy report-container-info-to-influxdb script
template:
src: usr/local/bin/report-container-info-to-influxdb.py.j2
dest: /usr/local/bin/report-container-info-to-influxdb
owner: root
group: root
mode: 0700
tags:
- lxd
- name: periodically kill old containers
cron:
name: "kill too-old containers"
minute: "*/5"
job: "/usr/local/bin/lxd-kill-old kill"
tags:
- lxd
- name: report container info to influxdb
cron:
name: "report-container-info-to-influxdb"
minute: "*/5"
job: "/usr/local/bin/report-container-info-to-influxdb"
tags:
- lxd
| Fix spacing and indentation in task name | Fix spacing and indentation in task name
| YAML | agpl-3.0 | harrowio/harrow,harrowio/harrow,harrowio/harrow,harrowio/harrow,harrowio/harrow,harrowio/harrow | yaml | ## Code Before:
---
- apt:
name: "{{ item }}"
update_cache: true
with_items:
- python3-requests
- python3-dateutil
- name: copy lxd-kill-old script
template:
src: usr/local/bin/lxd-kill-old.py.j2
dest: /usr/local/bin/lxd-kill-old
owner: root
group: root
mode: 0700
tags:
- lxd
- name: copy report-container-info-to-influxdb script
template:
src: usr/local/bin/report-container-info-to-influxdb.py.j2
dest: /usr/local/bin/report-container-info-to-influxdb
owner: root
group: root
mode: 0700
tags:
- lxd
- name: set MAILTO to nothing in crontab
cron:
name: MAILTO
minute: MAILTO=""
hour: ""
day: ""
month: ""
weekday: ""
job: ""
- name: periodically kill old containers
cron:
name: "kill too-old containers"
minute: "*/5"
job: "/usr/local/bin/lxd-kill-old kill"
tags:
- lxd
- name: report container info to influxdb
cron:
name: "report-container-info-to-influxdb"
minute: "*/5"
job: "/usr/local/bin/report-container-info-to-influxdb"
tags:
- lxd
## Instruction:
Fix spacing and indentation in task name
## Code After:
---
- apt:
name: "{{ item }}"
update_cache: true
with_items:
- python3-requests
- python3-dateutil
- name: copy lxd-kill-old script
template:
src: usr/local/bin/lxd-kill-old.py.j2
dest: /usr/local/bin/lxd-kill-old
owner: root
group: root
mode: 0700
tags:
- lxd
- name: copy report-container-info-to-influxdb script
template:
src: usr/local/bin/report-container-info-to-influxdb.py.j2
dest: /usr/local/bin/report-container-info-to-influxdb
owner: root
group: root
mode: 0700
tags:
- lxd
- name: periodically kill old containers
cron:
name: "kill too-old containers"
minute: "*/5"
job: "/usr/local/bin/lxd-kill-old kill"
tags:
- lxd
- name: report container info to influxdb
cron:
name: "report-container-info-to-influxdb"
minute: "*/5"
job: "/usr/local/bin/report-container-info-to-influxdb"
tags:
- lxd
| ---
- apt:
name: "{{ item }}"
update_cache: true
with_items:
- python3-requests
- python3-dateutil
- name: copy lxd-kill-old script
template:
src: usr/local/bin/lxd-kill-old.py.j2
dest: /usr/local/bin/lxd-kill-old
owner: root
group: root
mode: 0700
tags:
- lxd
- name: copy report-container-info-to-influxdb script
template:
src: usr/local/bin/report-container-info-to-influxdb.py.j2
dest: /usr/local/bin/report-container-info-to-influxdb
owner: root
group: root
mode: 0700
tags:
- lxd
- - name: set MAILTO to nothing in crontab
- cron:
- name: MAILTO
- minute: MAILTO=""
- hour: ""
- day: ""
- month: ""
- weekday: ""
- job: ""
-
- name: periodically kill old containers
cron:
name: "kill too-old containers"
minute: "*/5"
job: "/usr/local/bin/lxd-kill-old kill"
tags:
- lxd
- - name: report container info to influxdb
? --
+ - name: report container info to influxdb
cron:
name: "report-container-info-to-influxdb"
minute: "*/5"
job: "/usr/local/bin/report-container-info-to-influxdb"
tags:
- lxd
-
-
- | 15 | 0.267857 | 1 | 14 |
44fa96b817da339e27e60b30f4b8354f952f0d74 | src/boxmuller.cpp | src/boxmuller.cpp |
template <class T>
void boxmuller(T* data, size_t count) {
assert(count % 2 == 0);
static const T twopi = T(2.0 * 3.14159265358979323846);
LCG r;
for (size_t i = 0; i < count; i += 2) {
T u1, u2;
do {
u1 = r();
u2 = r();
} while (u1 <= std::numeric_limits<T>::min());
T radius = std::sqrt(-2 * std::log(u1));
T theta = twopi * u2;
data[i ] = radius * std::cos(theta);
data[i + 1] = radius * std::sin(theta);
}
}
static void normaldistf_boxmuller(float* data, size_t count) {
boxmuller(data, count);
}
static void normaldist_boxmuller(double* data, size_t count) {
boxmuller(data, count);
}
REGISTER_TEST(boxmuller);
|
template <class T>
void boxmuller(T* data, size_t count) {
assert(count % 2 == 0);
static const T twopi = T(2.0 * 3.14159265358979323846);
LCG r;
for (size_t i = 0; i < count; i += 2) {
T u1, u2;
u1 = r();
u2 = r();
u1 = std::max(u1, std::numeric_limits<T>::min());
T radius = std::sqrt(-2 * std::log(u1));
T theta = twopi * u2;
data[i ] = radius * std::cos(theta);
data[i + 1] = radius * std::sin(theta);
}
}
static void normaldistf_boxmuller(float* data, size_t count) {
boxmuller(data, count);
}
static void normaldist_boxmuller(double* data, size_t count) {
boxmuller(data, count);
}
REGISTER_TEST(boxmuller);
| Use minimum instead of do-while in Box-Muller | Use minimum instead of do-while in Box-Muller
| C++ | mit | miloyip/normaldist-benchmark,miloyip/normaldist-benchmark,miloyip/normaldist-benchmark,miloyip/normaldist-benchmark,miloyip/normaldist-benchmark | c++ | ## Code Before:
template <class T>
void boxmuller(T* data, size_t count) {
assert(count % 2 == 0);
static const T twopi = T(2.0 * 3.14159265358979323846);
LCG r;
for (size_t i = 0; i < count; i += 2) {
T u1, u2;
do {
u1 = r();
u2 = r();
} while (u1 <= std::numeric_limits<T>::min());
T radius = std::sqrt(-2 * std::log(u1));
T theta = twopi * u2;
data[i ] = radius * std::cos(theta);
data[i + 1] = radius * std::sin(theta);
}
}
static void normaldistf_boxmuller(float* data, size_t count) {
boxmuller(data, count);
}
static void normaldist_boxmuller(double* data, size_t count) {
boxmuller(data, count);
}
REGISTER_TEST(boxmuller);
## Instruction:
Use minimum instead of do-while in Box-Muller
## Code After:
template <class T>
void boxmuller(T* data, size_t count) {
assert(count % 2 == 0);
static const T twopi = T(2.0 * 3.14159265358979323846);
LCG r;
for (size_t i = 0; i < count; i += 2) {
T u1, u2;
u1 = r();
u2 = r();
u1 = std::max(u1, std::numeric_limits<T>::min());
T radius = std::sqrt(-2 * std::log(u1));
T theta = twopi * u2;
data[i ] = radius * std::cos(theta);
data[i + 1] = radius * std::sin(theta);
}
}
static void normaldistf_boxmuller(float* data, size_t count) {
boxmuller(data, count);
}
static void normaldist_boxmuller(double* data, size_t count) {
boxmuller(data, count);
}
REGISTER_TEST(boxmuller);
|
template <class T>
void boxmuller(T* data, size_t count) {
assert(count % 2 == 0);
static const T twopi = T(2.0 * 3.14159265358979323846);
LCG r;
for (size_t i = 0; i < count; i += 2) {
T u1, u2;
- do {
- u1 = r();
? ----
+ u1 = r();
- u2 = r();
? ----
+ u2 = r();
- } while (u1 <= std::numeric_limits<T>::min());
? ^ ----- ^^^
+ u1 = std::max(u1, std::numeric_limits<T>::min());
? ^^ ++++++++++ ^
T radius = std::sqrt(-2 * std::log(u1));
T theta = twopi * u2;
data[i ] = radius * std::cos(theta);
data[i + 1] = radius * std::sin(theta);
}
}
static void normaldistf_boxmuller(float* data, size_t count) {
boxmuller(data, count);
}
static void normaldist_boxmuller(double* data, size_t count) {
boxmuller(data, count);
}
REGISTER_TEST(boxmuller); | 7 | 0.233333 | 3 | 4 |
95b167955c246946859fa3c58bd0805671c7e954 | README.md | README.md | 
# [daniel.ga/llegos](https://daniel.ga/llegos)
This is my portfolio website. There are many like it, but this one is mine.
It's a constant work in progress.
## things used:
* Jekyll (Markdown :heart:)
* GitHub Pages (Free hosting, w00t!)
* HTML5/CSS3/JS (and more buzzwords)
* [SMACCS](https://smacss.com/) for SASS layout
* [BEM](http://getbem.com/) for CSS structure
## code license:
[MIT](/LICENSE.md)
## content:
:copyright: Daniel A. Gallegos 2018
| 
# [daniel.ga/llegos](https://daniel.ga/llegos)
This is my portfolio website. There are many like it, but this one is mine.
It's a constant work in progress.
## things used:
* Jekyll (Markdown :heart:)
* GitHub Pages (Free hosting, w00t!)
* HTML5/CSS3/JS (and more buzzwords)
* [SMACCS](https://smacss.com/) for SASS layout
* [BEM](http://getbem.com/) for CSS structure
## code license:
[MIT](/LICENSE.md)
## content:
:copyright: Daniel A. Gallegos 2018
| Fix image to refer to master branch | Fix image to refer to master branch
| Markdown | mit | thattacoguy/thattacoguy.github.io | markdown | ## Code Before:

# [daniel.ga/llegos](https://daniel.ga/llegos)
This is my portfolio website. There are many like it, but this one is mine.
It's a constant work in progress.
## things used:
* Jekyll (Markdown :heart:)
* GitHub Pages (Free hosting, w00t!)
* HTML5/CSS3/JS (and more buzzwords)
* [SMACCS](https://smacss.com/) for SASS layout
* [BEM](http://getbem.com/) for CSS structure
## code license:
[MIT](/LICENSE.md)
## content:
:copyright: Daniel A. Gallegos 2018
## Instruction:
Fix image to refer to master branch
## Code After:

# [daniel.ga/llegos](https://daniel.ga/llegos)
This is my portfolio website. There are many like it, but this one is mine.
It's a constant work in progress.
## things used:
* Jekyll (Markdown :heart:)
* GitHub Pages (Free hosting, w00t!)
* HTML5/CSS3/JS (and more buzzwords)
* [SMACCS](https://smacss.com/) for SASS layout
* [BEM](http://getbem.com/) for CSS structure
## code license:
[MIT](/LICENSE.md)
## content:
:copyright: Daniel A. Gallegos 2018
| - 
? ---- ----------- ^^ ^^^
+ 
? ++ ^^^^^ ^^ +++++++++
# [daniel.ga/llegos](https://daniel.ga/llegos)
This is my portfolio website. There are many like it, but this one is mine.
It's a constant work in progress.
## things used:
* Jekyll (Markdown :heart:)
* GitHub Pages (Free hosting, w00t!)
* HTML5/CSS3/JS (and more buzzwords)
* [SMACCS](https://smacss.com/) for SASS layout
* [BEM](http://getbem.com/) for CSS structure
## code license:
[MIT](/LICENSE.md)
## content:
:copyright: Daniel A. Gallegos 2018 | 2 | 0.1 | 1 | 1 |
3fcd9a4e44e2f8851a54d47558d814c512d76c19 | app/CakeInput.tsx | app/CakeInput.tsx | import * as React from "react";
class CakeInput extends React.Component<any, void> {
constructor(props: React.Props<any>) {
super(props);
// Required because of the ES6 class syntax decifiency
this.onSubmit = this.onSubmit.bind(this);
}
refs: {
[string: string]: any;
cakeName: any;
}
onSubmit(e: Event) {
e.preventDefault();
const newCake = {
name: this.refs.cakeName.value,
description: "",
date: new Date()
};
// Propagate to external handler
this.props.onSubmit(newCake);
}
render() {
return (
<form onSubmit={this.onSubmit}>
<h3>New cake</h3>
<input ref="cakeName" type="text" defaultValue="" placeholder="Cake name" />
<button type="submit">Add</button>
</form>
);
}
}
export default CakeInput;
| import * as React from "react";
interface CakeInputState {
value: string;
}
class CakeInput extends React.Component<any, CakeInputState> {
constructor(props: React.Props<any>) {
super(props);
// Required because of the ES6 class syntax decifiency
this.onSubmit = this.onSubmit.bind(this);
this.state = { value: "" };
}
refs: {
[string: string]: any;
cakeName: any;
}
onSubmit(e: Event) {
e.preventDefault();
const newCake = {
name: this.refs.cakeName.value,
description: "",
date: new Date()
};
// Propagate to external handler
this.props.onSubmit(newCake);
// State didn't for some reason propagate correctly to the input's default value,
// so clearing the input field directly for now
this.refs.cakeName.value = "";
}
render() {
return (
<form onSubmit={this.onSubmit}>
<h3>New cake</h3>
<input ref="cakeName" type="text" placeholder="Cake name" />
<button type="submit">Add</button>
</form>
);
}
}
export default CakeInput;
| Clear text field on input | Clear text field on input
| TypeScript | mit | mieky/we-love-cake,mieky/we-love-cake | typescript | ## Code Before:
import * as React from "react";
class CakeInput extends React.Component<any, void> {
constructor(props: React.Props<any>) {
super(props);
// Required because of the ES6 class syntax decifiency
this.onSubmit = this.onSubmit.bind(this);
}
refs: {
[string: string]: any;
cakeName: any;
}
onSubmit(e: Event) {
e.preventDefault();
const newCake = {
name: this.refs.cakeName.value,
description: "",
date: new Date()
};
// Propagate to external handler
this.props.onSubmit(newCake);
}
render() {
return (
<form onSubmit={this.onSubmit}>
<h3>New cake</h3>
<input ref="cakeName" type="text" defaultValue="" placeholder="Cake name" />
<button type="submit">Add</button>
</form>
);
}
}
export default CakeInput;
## Instruction:
Clear text field on input
## Code After:
import * as React from "react";
interface CakeInputState {
value: string;
}
class CakeInput extends React.Component<any, CakeInputState> {
constructor(props: React.Props<any>) {
super(props);
// Required because of the ES6 class syntax decifiency
this.onSubmit = this.onSubmit.bind(this);
this.state = { value: "" };
}
refs: {
[string: string]: any;
cakeName: any;
}
onSubmit(e: Event) {
e.preventDefault();
const newCake = {
name: this.refs.cakeName.value,
description: "",
date: new Date()
};
// Propagate to external handler
this.props.onSubmit(newCake);
// State didn't for some reason propagate correctly to the input's default value,
// so clearing the input field directly for now
this.refs.cakeName.value = "";
}
render() {
return (
<form onSubmit={this.onSubmit}>
<h3>New cake</h3>
<input ref="cakeName" type="text" placeholder="Cake name" />
<button type="submit">Add</button>
</form>
);
}
}
export default CakeInput;
| import * as React from "react";
+ interface CakeInputState {
+ value: string;
+ }
+
- class CakeInput extends React.Component<any, void> {
? ^^^^
+ class CakeInput extends React.Component<any, CakeInputState> {
? ^^^^^^^^^^^^^^
constructor(props: React.Props<any>) {
super(props);
// Required because of the ES6 class syntax decifiency
this.onSubmit = this.onSubmit.bind(this);
+ this.state = { value: "" };
- }
? -
+ }
refs: {
[string: string]: any;
cakeName: any;
}
onSubmit(e: Event) {
e.preventDefault();
const newCake = {
name: this.refs.cakeName.value,
description: "",
date: new Date()
};
// Propagate to external handler
this.props.onSubmit(newCake);
+
+ // State didn't for some reason propagate correctly to the input's default value,
+ // so clearing the input field directly for now
+ this.refs.cakeName.value = "";
}
render() {
return (
<form onSubmit={this.onSubmit}>
<h3>New cake</h3>
- <input ref="cakeName" type="text" defaultValue="" placeholder="Cake name" />
? ----------------
+ <input ref="cakeName" type="text" placeholder="Cake name" />
<button type="submit">Add</button>
</form>
);
}
}
export default CakeInput; | 15 | 0.375 | 12 | 3 |
cbde377559c912cf28458234e0bcbd3b7b0bc7c4 | src/main/java/water/api/TypeaheadFileRequest.java | src/main/java/water/api/TypeaheadFileRequest.java |
package water.api;
import java.io.File;
import com.google.gson.*;
public class TypeaheadFileRequest extends TypeaheadRequest {
public TypeaheadFileRequest() {
super("Provides a simple JSON array of filtered local files.","");
}
@Override
protected JsonArray serve(String filter, int limit) {
File base = null;
String filterPrefix = "";
if( !filter.isEmpty() ) {
File file = new File(filter);
if( file.isDirectory() ) {
base = file;
} else {
base = file.getParentFile();
filterPrefix = file.getName().toLowerCase();
}
}
if( base == null ) base = new File(".");
JsonArray array = new JsonArray();
for( File file : base.listFiles() ) {
if( file.isHidden() ) continue;
if( file.getName().toLowerCase().startsWith(filterPrefix) ) {
String s = file.getPath();
array.add(new JsonPrimitive(s));
}
if( array.size() == limit) break;
}
return array;
}
}
|
package water.api;
import java.io.File;
import com.google.gson.*;
public class TypeaheadFileRequest extends TypeaheadRequest {
public TypeaheadFileRequest() {
super("Provides a simple JSON array of filtered local files.","");
}
@Override
protected JsonArray serve(String filter, int limit) {
File base = null;
String filterPrefix = "";
if( !filter.isEmpty() ) {
File file = new File(filter);
if( file.isDirectory() ) {
base = file;
} else {
base = file.getParentFile();
filterPrefix = file.getName().toLowerCase();
}
}
if( base == null ) base = new File(".");
JsonArray array = new JsonArray();
File[] files = base.listFiles();
if( files == null ) return array;
for( File file : files ) {
if( file.isHidden() ) continue;
if( file.getName().toLowerCase().startsWith(filterPrefix) )
array.add(new JsonPrimitive(file.getPath()));
if( array.size() == limit) break;
}
return array;
}
}
| Fix NPE for mispelled/missing directories in type-ahead file lookup | Fix NPE for mispelled/missing directories in type-ahead file lookup
| Java | apache-2.0 | 100star/h2o,elkingtonmcb/h2o-2,h2oai/h2o,rowhit/h2o-2,eg-zhang/h2o-2,100star/h2o,h2oai/h2o-2,elkingtonmcb/h2o-2,100star/h2o,h2oai/h2o,calvingit21/h2o-2,eg-zhang/h2o-2,h2oai/h2o-2,elkingtonmcb/h2o-2,rowhit/h2o-2,rowhit/h2o-2,eg-zhang/h2o-2,rowhit/h2o-2,calvingit21/h2o-2,eg-zhang/h2o-2,h2oai/h2o-2,vbelakov/h2o,111t8e/h2o-2,calvingit21/h2o-2,h2oai/h2o,rowhit/h2o-2,111t8e/h2o-2,h2oai/h2o-2,111t8e/h2o-2,vbelakov/h2o,eg-zhang/h2o-2,111t8e/h2o-2,111t8e/h2o-2,calvingit21/h2o-2,rowhit/h2o-2,vbelakov/h2o,eg-zhang/h2o-2,calvingit21/h2o-2,h2oai/h2o-2,elkingtonmcb/h2o-2,100star/h2o,vbelakov/h2o,h2oai/h2o,calvingit21/h2o-2,h2oai/h2o,rowhit/h2o-2,111t8e/h2o-2,vbelakov/h2o,h2oai/h2o-2,rowhit/h2o-2,elkingtonmcb/h2o-2,rowhit/h2o-2,111t8e/h2o-2,vbelakov/h2o,calvingit21/h2o-2,111t8e/h2o-2,h2oai/h2o,rowhit/h2o-2,vbelakov/h2o,eg-zhang/h2o-2,h2oai/h2o-2,100star/h2o,100star/h2o,100star/h2o,elkingtonmcb/h2o-2,h2oai/h2o-2,vbelakov/h2o,h2oai/h2o,111t8e/h2o-2,elkingtonmcb/h2o-2,111t8e/h2o-2,elkingtonmcb/h2o-2,elkingtonmcb/h2o-2,h2oai/h2o,h2oai/h2o,eg-zhang/h2o-2,calvingit21/h2o-2,eg-zhang/h2o-2,vbelakov/h2o,100star/h2o,h2oai/h2o,calvingit21/h2o-2,calvingit21/h2o-2,elkingtonmcb/h2o-2,h2oai/h2o-2,eg-zhang/h2o-2,vbelakov/h2o,h2oai/h2o-2,100star/h2o | java | ## Code Before:
package water.api;
import java.io.File;
import com.google.gson.*;
public class TypeaheadFileRequest extends TypeaheadRequest {
public TypeaheadFileRequest() {
super("Provides a simple JSON array of filtered local files.","");
}
@Override
protected JsonArray serve(String filter, int limit) {
File base = null;
String filterPrefix = "";
if( !filter.isEmpty() ) {
File file = new File(filter);
if( file.isDirectory() ) {
base = file;
} else {
base = file.getParentFile();
filterPrefix = file.getName().toLowerCase();
}
}
if( base == null ) base = new File(".");
JsonArray array = new JsonArray();
for( File file : base.listFiles() ) {
if( file.isHidden() ) continue;
if( file.getName().toLowerCase().startsWith(filterPrefix) ) {
String s = file.getPath();
array.add(new JsonPrimitive(s));
}
if( array.size() == limit) break;
}
return array;
}
}
## Instruction:
Fix NPE for mispelled/missing directories in type-ahead file lookup
## Code After:
package water.api;
import java.io.File;
import com.google.gson.*;
public class TypeaheadFileRequest extends TypeaheadRequest {
public TypeaheadFileRequest() {
super("Provides a simple JSON array of filtered local files.","");
}
@Override
protected JsonArray serve(String filter, int limit) {
File base = null;
String filterPrefix = "";
if( !filter.isEmpty() ) {
File file = new File(filter);
if( file.isDirectory() ) {
base = file;
} else {
base = file.getParentFile();
filterPrefix = file.getName().toLowerCase();
}
}
if( base == null ) base = new File(".");
JsonArray array = new JsonArray();
File[] files = base.listFiles();
if( files == null ) return array;
for( File file : files ) {
if( file.isHidden() ) continue;
if( file.getName().toLowerCase().startsWith(filterPrefix) )
array.add(new JsonPrimitive(file.getPath()));
if( array.size() == limit) break;
}
return array;
}
}
|
package water.api;
import java.io.File;
import com.google.gson.*;
public class TypeaheadFileRequest extends TypeaheadRequest {
public TypeaheadFileRequest() {
super("Provides a simple JSON array of filtered local files.","");
}
@Override
protected JsonArray serve(String filter, int limit) {
File base = null;
String filterPrefix = "";
if( !filter.isEmpty() ) {
File file = new File(filter);
if( file.isDirectory() ) {
base = file;
} else {
base = file.getParentFile();
filterPrefix = file.getName().toLowerCase();
}
}
if( base == null ) base = new File(".");
JsonArray array = new JsonArray();
+ File[] files = base.listFiles();
+ if( files == null ) return array;
- for( File file : base.listFiles() ) {
? ^^^^^^^^^^ --
+ for( File file : files ) {
? ^
if( file.isHidden() ) continue;
- if( file.getName().toLowerCase().startsWith(filterPrefix) ) {
? --
+ if( file.getName().toLowerCase().startsWith(filterPrefix) )
- String s = file.getPath();
- array.add(new JsonPrimitive(s));
? ^
+ array.add(new JsonPrimitive(file.getPath()));
? ^^^^^^^^^^^^^^
- }
if( array.size() == limit) break;
}
return array;
}
} | 10 | 0.25 | 5 | 5 |
63901d2a06eb1a955caf262db08c8c27cd98a599 | src/Eloquent/Model.php | src/Eloquent/Model.php | <?php namespace duxet\Rethinkdb\Eloquent;
use Carbon\Carbon;
use duxet\Rethinkdb\Query\Builder as QueryBuilder;
class Model extends \Illuminate\Database\Eloquent\Model {
/**
* Get the format for database stored dates.
*
* @return string
*/
protected function getDateFormat()
{
return 'Y-m-d H:i:s';
}
/**
* Return a timestamp as DateTime object.
*
* @param mixed $value
* @return \Carbon\Carbon
*/
protected function asDateTime($value)
{
return Carbon::instance($value);
}
/**
* Get a new query builder instance for the connection.
*
* @return Builder
*/
protected function newBaseQueryBuilder()
{
$connection = $this->getConnection();
// Check the connection type
if ($connection instanceof \duxet\Rethinkdb\Connection)
{
return new QueryBuilder($connection);
}
return parent::newBaseQueryBuilder();
}
/**
* Create a new Eloquent query builder for the model.
*
* @param \duxet\Rethinkdb\Query\Builder $query
* @return \duxet\Rethinkdb\Eloquent\Builder|static
*/
public function newEloquentBuilder($query)
{
return new Builder($query);
}
} | <?php namespace duxet\Rethinkdb\Eloquent;
use Carbon\Carbon;
use duxet\Rethinkdb\Query\Builder as QueryBuilder;
class Model extends \Illuminate\Database\Eloquent\Model {
/**
* Get the format for database stored dates.
*
* @return string
*/
protected function getDateFormat()
{
return 'Y-m-d H:i:s';
}
/**
* Return DateTime object as Carbon instance.
*
* @param DateTime $value
* @return \Carbon\Carbon
*/
protected function asDateTime($value)
{
return Carbon::instance($value);
}
/**
* Get a new query builder instance for the connection.
*
* @return Builder
*/
protected function newBaseQueryBuilder()
{
$connection = $this->getConnection();
// Check the connection type
if ($connection instanceof \duxet\Rethinkdb\Connection)
{
return new QueryBuilder($connection);
}
return parent::newBaseQueryBuilder();
}
/**
* Create a new Eloquent query builder for the model.
*
* @param \duxet\Rethinkdb\Query\Builder $query
* @return \duxet\Rethinkdb\Eloquent\Builder|static
*/
public function newEloquentBuilder($query)
{
return new Builder($query);
}
} | Update phpdoc for last commit | Update phpdoc for last commit
| PHP | mit | duxet/laravel-rethinkdb,dkuzmenchuk/laravel-rethinkdb | php | ## Code Before:
<?php namespace duxet\Rethinkdb\Eloquent;
use Carbon\Carbon;
use duxet\Rethinkdb\Query\Builder as QueryBuilder;
class Model extends \Illuminate\Database\Eloquent\Model {
/**
* Get the format for database stored dates.
*
* @return string
*/
protected function getDateFormat()
{
return 'Y-m-d H:i:s';
}
/**
* Return a timestamp as DateTime object.
*
* @param mixed $value
* @return \Carbon\Carbon
*/
protected function asDateTime($value)
{
return Carbon::instance($value);
}
/**
* Get a new query builder instance for the connection.
*
* @return Builder
*/
protected function newBaseQueryBuilder()
{
$connection = $this->getConnection();
// Check the connection type
if ($connection instanceof \duxet\Rethinkdb\Connection)
{
return new QueryBuilder($connection);
}
return parent::newBaseQueryBuilder();
}
/**
* Create a new Eloquent query builder for the model.
*
* @param \duxet\Rethinkdb\Query\Builder $query
* @return \duxet\Rethinkdb\Eloquent\Builder|static
*/
public function newEloquentBuilder($query)
{
return new Builder($query);
}
}
## Instruction:
Update phpdoc for last commit
## Code After:
<?php namespace duxet\Rethinkdb\Eloquent;
use Carbon\Carbon;
use duxet\Rethinkdb\Query\Builder as QueryBuilder;
class Model extends \Illuminate\Database\Eloquent\Model {
/**
* Get the format for database stored dates.
*
* @return string
*/
protected function getDateFormat()
{
return 'Y-m-d H:i:s';
}
/**
* Return DateTime object as Carbon instance.
*
* @param DateTime $value
* @return \Carbon\Carbon
*/
protected function asDateTime($value)
{
return Carbon::instance($value);
}
/**
* Get a new query builder instance for the connection.
*
* @return Builder
*/
protected function newBaseQueryBuilder()
{
$connection = $this->getConnection();
// Check the connection type
if ($connection instanceof \duxet\Rethinkdb\Connection)
{
return new QueryBuilder($connection);
}
return parent::newBaseQueryBuilder();
}
/**
* Create a new Eloquent query builder for the model.
*
* @param \duxet\Rethinkdb\Query\Builder $query
* @return \duxet\Rethinkdb\Eloquent\Builder|static
*/
public function newEloquentBuilder($query)
{
return new Builder($query);
}
} | <?php namespace duxet\Rethinkdb\Eloquent;
use Carbon\Carbon;
use duxet\Rethinkdb\Query\Builder as QueryBuilder;
class Model extends \Illuminate\Database\Eloquent\Model {
/**
* Get the format for database stored dates.
*
* @return string
*/
protected function getDateFormat()
{
return 'Y-m-d H:i:s';
}
/**
- * Return a timestamp as DateTime object.
+ * Return DateTime object as Carbon instance.
*
- * @param mixed $value
? -- -
+ * @param DateTime $value
? ++++++
* @return \Carbon\Carbon
*/
protected function asDateTime($value)
{
return Carbon::instance($value);
}
/**
* Get a new query builder instance for the connection.
*
* @return Builder
*/
protected function newBaseQueryBuilder()
{
$connection = $this->getConnection();
// Check the connection type
if ($connection instanceof \duxet\Rethinkdb\Connection)
{
return new QueryBuilder($connection);
}
return parent::newBaseQueryBuilder();
}
/**
* Create a new Eloquent query builder for the model.
*
* @param \duxet\Rethinkdb\Query\Builder $query
* @return \duxet\Rethinkdb\Eloquent\Builder|static
*/
public function newEloquentBuilder($query)
{
return new Builder($query);
}
} | 4 | 0.071429 | 2 | 2 |
1ede74e4f0972bc2fdf85adbc074fd147189ccba | spec/spec_helper.rb | spec/spec_helper.rb | $LOAD_PATH.unshift File.expand_path('../../lib', __FILE__)
require 'lstash'
require 'rspec/its'
require 'timecop'
ENV['ES_URL'] = nil
RSpec.configure do |config|
config.order = 'random'
end
require 'stringio'
def capture_stdout(&blk)
old = $stdout
$stdout = fake = StringIO.new
blk.call
fake.string
ensure
$stdout = old
end
def capture_stderr(&blk)
old = $stderr
$stderr = fake = StringIO.new
blk.call
fake.string
ensure
$stderr = old
end | $LOAD_PATH.unshift File.expand_path('../../lib', __FILE__)
require 'lstash'
require 'rspec/its'
require 'timecop'
ENV['ES_URL'] = nil
ENV['TZ'] = 'Europe/Amsterdam' # Test in a specific timezone.
RSpec.configure do |config|
config.order = 'random'
end
require 'stringio'
def capture_stdout(&blk)
old = $stdout
$stdout = fake = StringIO.new
blk.call
fake.string
ensure
$stdout = old
end
def capture_stderr(&blk)
old = $stderr
$stderr = fake = StringIO.new
blk.call
fake.string
ensure
$stderr = old
end | Fix the timezone for the specs. We test in a specific timezone. | Fix the timezone for the specs. We test in a specific timezone.
| Ruby | mit | kjwierenga/lstash | ruby | ## Code Before:
$LOAD_PATH.unshift File.expand_path('../../lib', __FILE__)
require 'lstash'
require 'rspec/its'
require 'timecop'
ENV['ES_URL'] = nil
RSpec.configure do |config|
config.order = 'random'
end
require 'stringio'
def capture_stdout(&blk)
old = $stdout
$stdout = fake = StringIO.new
blk.call
fake.string
ensure
$stdout = old
end
def capture_stderr(&blk)
old = $stderr
$stderr = fake = StringIO.new
blk.call
fake.string
ensure
$stderr = old
end
## Instruction:
Fix the timezone for the specs. We test in a specific timezone.
## Code After:
$LOAD_PATH.unshift File.expand_path('../../lib', __FILE__)
require 'lstash'
require 'rspec/its'
require 'timecop'
ENV['ES_URL'] = nil
ENV['TZ'] = 'Europe/Amsterdam' # Test in a specific timezone.
RSpec.configure do |config|
config.order = 'random'
end
require 'stringio'
def capture_stdout(&blk)
old = $stdout
$stdout = fake = StringIO.new
blk.call
fake.string
ensure
$stdout = old
end
def capture_stderr(&blk)
old = $stderr
$stderr = fake = StringIO.new
blk.call
fake.string
ensure
$stderr = old
end | $LOAD_PATH.unshift File.expand_path('../../lib', __FILE__)
require 'lstash'
require 'rspec/its'
require 'timecop'
ENV['ES_URL'] = nil
+ ENV['TZ'] = 'Europe/Amsterdam' # Test in a specific timezone.
RSpec.configure do |config|
config.order = 'random'
end
require 'stringio'
def capture_stdout(&blk)
old = $stdout
$stdout = fake = StringIO.new
blk.call
fake.string
ensure
$stdout = old
end
def capture_stderr(&blk)
old = $stderr
$stderr = fake = StringIO.new
blk.call
fake.string
ensure
$stderr = old
end | 1 | 0.03125 | 1 | 0 |
58c9d62af1054753acd468d61aa903269a1839a3 | Resources/views/User/edit_content.twig | Resources/views/User/edit_content.twig | {% form_theme form form.getOption('theme') %}
<form action="{% path 'doctrine_user_user_update' %}" {{ form|render_enctype }} method="POST" class="doctrine_user_user_edit">
{{ form|render }}
<div>
<input type="submit" value="{% trans 'Update user' from 'DoctrineUserBundle' %}" />
</div>
</form>
| {% form_theme form form.getOption('theme') %}
<form action="{% path 'doctrine_user_user_update' with ['username': username] %}" {{ form|render_enctype }} method="POST" class="doctrine_user_user_edit">
{{ form|render }}
<div>
<input type="submit" value="{% trans 'Update user' from 'DoctrineUserBundle' %}" />
</div>
</form>
| Add missing username parameter in the user edit twig view | Add missing username parameter in the user edit twig view
| Twig | mit | XWB/FOSUserBundle,dunglas/DoctrineUserBundle,XWB/FOSUserBundle,FriendsOfSymfony/FOSUserBundle | twig | ## Code Before:
{% form_theme form form.getOption('theme') %}
<form action="{% path 'doctrine_user_user_update' %}" {{ form|render_enctype }} method="POST" class="doctrine_user_user_edit">
{{ form|render }}
<div>
<input type="submit" value="{% trans 'Update user' from 'DoctrineUserBundle' %}" />
</div>
</form>
## Instruction:
Add missing username parameter in the user edit twig view
## Code After:
{% form_theme form form.getOption('theme') %}
<form action="{% path 'doctrine_user_user_update' with ['username': username] %}" {{ form|render_enctype }} method="POST" class="doctrine_user_user_edit">
{{ form|render }}
<div>
<input type="submit" value="{% trans 'Update user' from 'DoctrineUserBundle' %}" />
</div>
</form>
| {% form_theme form form.getOption('theme') %}
- <form action="{% path 'doctrine_user_user_update' %}" {{ form|render_enctype }} method="POST" class="doctrine_user_user_edit">
+ <form action="{% path 'doctrine_user_user_update' with ['username': username] %}" {{ form|render_enctype }} method="POST" class="doctrine_user_user_edit">
? ++++++++++++++++++++++++++++
{{ form|render }}
<div>
<input type="submit" value="{% trans 'Update user' from 'DoctrineUserBundle' %}" />
</div>
</form> | 2 | 0.285714 | 1 | 1 |
30a3b8ab3cd24328fbf0576d4143f6349fa27b5b | src/boss_cache.erl | src/boss_cache.erl | -module(boss_cache).
-export([start/0, start/1]).
-export([stop/0]).
-export([get/2, set/4, delete/2]).
-define(POOLNAME, boss_cache_pool).
start() ->
Adapter = boss_cache_adapter_memcached_bin,
start([{adapter, Adapter}, {cache_servers, [{"127.0.0.1", 11211, 1}]}]).
start(Options) ->
Adapter = proplists:get_value(adapter, Options, boss_cache_adapter_memcached_bin),
Adapter:start(Options),
boss_cache_sup:start_link(Options).
stop() ->
ok.
set(Prefix, Key, Val, TTL) ->
boss_pool:call(?POOLNAME, {set, Prefix, Key, Val, TTL}).
get(Prefix, Key) ->
boss_pool:call(?POOLNAME, {get, Prefix, Key}).
delete(Prefix, Key) ->
boss_pool:call(?POOLNAME, {delete, Prefix, Key}).
| -module(boss_cache).
-export([start/0, start/1]).
-export([stop/0]).
-export([get/2, set/4, delete/2]).
-define(POOLNAME, boss_cache_pool).
start() ->
Adapter = boss_cache_adapter_memcached_bin,
start([{adapter, Adapter}, {cache_servers, [{"127.0.0.1", 11211, 1}]}]).
start(Options) ->
AdapterName = proplists:get_value(adapter, Options, memcached_bin),
Adapter = list_to_atom(lists:concat(["boss_cache_adapter_", AdapterName])),
Adapter:start(Options),
boss_cache_sup:start_link(Options).
stop() ->
ok.
set(Prefix, Key, Val, TTL) ->
boss_pool:call(?POOLNAME, {set, Prefix, Key, Val, TTL}).
get(Prefix, Key) ->
boss_pool:call(?POOLNAME, {get, Prefix, Key}).
delete(Prefix, Key) ->
boss_pool:call(?POOLNAME, {delete, Prefix, Key}).
| Fix cache adapter name bug | Fix cache adapter name bug
| Erlang | mit | badubizzle/boss_db,QBkira/boss_db,mihawk/boss_db | erlang | ## Code Before:
-module(boss_cache).
-export([start/0, start/1]).
-export([stop/0]).
-export([get/2, set/4, delete/2]).
-define(POOLNAME, boss_cache_pool).
start() ->
Adapter = boss_cache_adapter_memcached_bin,
start([{adapter, Adapter}, {cache_servers, [{"127.0.0.1", 11211, 1}]}]).
start(Options) ->
Adapter = proplists:get_value(adapter, Options, boss_cache_adapter_memcached_bin),
Adapter:start(Options),
boss_cache_sup:start_link(Options).
stop() ->
ok.
set(Prefix, Key, Val, TTL) ->
boss_pool:call(?POOLNAME, {set, Prefix, Key, Val, TTL}).
get(Prefix, Key) ->
boss_pool:call(?POOLNAME, {get, Prefix, Key}).
delete(Prefix, Key) ->
boss_pool:call(?POOLNAME, {delete, Prefix, Key}).
## Instruction:
Fix cache adapter name bug
## Code After:
-module(boss_cache).
-export([start/0, start/1]).
-export([stop/0]).
-export([get/2, set/4, delete/2]).
-define(POOLNAME, boss_cache_pool).
start() ->
Adapter = boss_cache_adapter_memcached_bin,
start([{adapter, Adapter}, {cache_servers, [{"127.0.0.1", 11211, 1}]}]).
start(Options) ->
AdapterName = proplists:get_value(adapter, Options, memcached_bin),
Adapter = list_to_atom(lists:concat(["boss_cache_adapter_", AdapterName])),
Adapter:start(Options),
boss_cache_sup:start_link(Options).
stop() ->
ok.
set(Prefix, Key, Val, TTL) ->
boss_pool:call(?POOLNAME, {set, Prefix, Key, Val, TTL}).
get(Prefix, Key) ->
boss_pool:call(?POOLNAME, {get, Prefix, Key}).
delete(Prefix, Key) ->
boss_pool:call(?POOLNAME, {delete, Prefix, Key}).
| -module(boss_cache).
-export([start/0, start/1]).
-export([stop/0]).
-export([get/2, set/4, delete/2]).
-define(POOLNAME, boss_cache_pool).
start() ->
Adapter = boss_cache_adapter_memcached_bin,
start([{adapter, Adapter}, {cache_servers, [{"127.0.0.1", 11211, 1}]}]).
start(Options) ->
- Adapter = proplists:get_value(adapter, Options, boss_cache_adapter_memcached_bin),
? -------------------
+ AdapterName = proplists:get_value(adapter, Options, memcached_bin),
? ++++
+ Adapter = list_to_atom(lists:concat(["boss_cache_adapter_", AdapterName])),
Adapter:start(Options),
boss_cache_sup:start_link(Options).
stop() ->
ok.
set(Prefix, Key, Val, TTL) ->
boss_pool:call(?POOLNAME, {set, Prefix, Key, Val, TTL}).
get(Prefix, Key) ->
boss_pool:call(?POOLNAME, {get, Prefix, Key}).
delete(Prefix, Key) ->
boss_pool:call(?POOLNAME, {delete, Prefix, Key}). | 3 | 0.111111 | 2 | 1 |
518ec5090aa225ac36daf4180d15d19f4d0d39ad | src/libaktualizr/logging/default_log_sink.cc | src/libaktualizr/logging/default_log_sink.cc |
static void color_fmt(boost::log::record_view const& rec, boost::log::formatting_ostream& strm) {
auto severity = rec[boost::log::trivial::severity];
bool color = false;
if (severity) {
switch (severity.get()) {
case boost::log::trivial::warning:
strm << "\033[33m";
color = true;
break;
case boost::log::trivial::error:
case boost::log::trivial::fatal:
strm << "\033[31m";
color = true;
break;
default:
break;
}
}
// setw(7) = the longest log level. eg "warning"
strm << std::setw(7) << std::setfill(' ') << severity << ": " << rec[boost::log::expressions::smessage];
if (color) {
strm << "\033[0m";
}
}
void logger_init_sink(bool use_colors = false) {
auto sink = boost::log::add_console_log(std::cout, boost::log::keywords::format = "%Message%",
boost::log::keywords::auto_flush = true);
if (use_colors) {
sink->set_formatter(&color_fmt);
}
}
|
static void color_fmt(boost::log::record_view const& rec, boost::log::formatting_ostream& strm) {
auto severity = rec[boost::log::trivial::severity];
bool color = false;
if (severity) {
switch (severity.get()) {
case boost::log::trivial::warning:
strm << "\033[33m";
color = true;
break;
case boost::log::trivial::error:
case boost::log::trivial::fatal:
strm << "\033[31m";
color = true;
break;
default:
break;
}
}
// setw(7) = the longest log level. eg "warning"
strm << std::setw(7) << std::setfill(' ') << severity << ": " << rec[boost::log::expressions::smessage];
if (color) {
strm << "\033[0m";
}
}
void logger_init_sink(bool use_colors = false) {
auto stream = &std::cerr;
if (getenv("LOG_STDERR") == nullptr) {
stream = &std::cout;
}
auto sink = boost::log::add_console_log(*stream, boost::log::keywords::format = "%Message%",
boost::log::keywords::auto_flush = true);
if (use_colors) {
sink->set_formatter(&color_fmt);
}
}
| Allow logger to use STDERR | logging: Allow logger to use STDERR
This allows libaktualizr to log to STDERR if LOG_STDERR is defined
in the environment. A handy way this can be used is for something
like aktualizr-get:
LOG_STDERR=1 aktualizr-get ....
This allows you to understand what is logging (stderr's fd) and what
is the output (stdout's fd).
Signed-off-by: Andy Doan <e3579b1e47f273529f0f929453e939a68ede9fd1@foundries.io>
| C++ | mpl-2.0 | advancedtelematic/aktualizr,advancedtelematic/sota_client_cpp,advancedtelematic/aktualizr,advancedtelematic/sota_client_cpp,advancedtelematic/aktualizr,advancedtelematic/aktualizr | c++ | ## Code Before:
static void color_fmt(boost::log::record_view const& rec, boost::log::formatting_ostream& strm) {
auto severity = rec[boost::log::trivial::severity];
bool color = false;
if (severity) {
switch (severity.get()) {
case boost::log::trivial::warning:
strm << "\033[33m";
color = true;
break;
case boost::log::trivial::error:
case boost::log::trivial::fatal:
strm << "\033[31m";
color = true;
break;
default:
break;
}
}
// setw(7) = the longest log level. eg "warning"
strm << std::setw(7) << std::setfill(' ') << severity << ": " << rec[boost::log::expressions::smessage];
if (color) {
strm << "\033[0m";
}
}
void logger_init_sink(bool use_colors = false) {
auto sink = boost::log::add_console_log(std::cout, boost::log::keywords::format = "%Message%",
boost::log::keywords::auto_flush = true);
if (use_colors) {
sink->set_formatter(&color_fmt);
}
}
## Instruction:
logging: Allow logger to use STDERR
This allows libaktualizr to log to STDERR if LOG_STDERR is defined
in the environment. A handy way this can be used is for something
like aktualizr-get:
LOG_STDERR=1 aktualizr-get ....
This allows you to understand what is logging (stderr's fd) and what
is the output (stdout's fd).
Signed-off-by: Andy Doan <e3579b1e47f273529f0f929453e939a68ede9fd1@foundries.io>
## Code After:
static void color_fmt(boost::log::record_view const& rec, boost::log::formatting_ostream& strm) {
auto severity = rec[boost::log::trivial::severity];
bool color = false;
if (severity) {
switch (severity.get()) {
case boost::log::trivial::warning:
strm << "\033[33m";
color = true;
break;
case boost::log::trivial::error:
case boost::log::trivial::fatal:
strm << "\033[31m";
color = true;
break;
default:
break;
}
}
// setw(7) = the longest log level. eg "warning"
strm << std::setw(7) << std::setfill(' ') << severity << ": " << rec[boost::log::expressions::smessage];
if (color) {
strm << "\033[0m";
}
}
void logger_init_sink(bool use_colors = false) {
auto stream = &std::cerr;
if (getenv("LOG_STDERR") == nullptr) {
stream = &std::cout;
}
auto sink = boost::log::add_console_log(*stream, boost::log::keywords::format = "%Message%",
boost::log::keywords::auto_flush = true);
if (use_colors) {
sink->set_formatter(&color_fmt);
}
}
|
static void color_fmt(boost::log::record_view const& rec, boost::log::formatting_ostream& strm) {
auto severity = rec[boost::log::trivial::severity];
bool color = false;
if (severity) {
switch (severity.get()) {
case boost::log::trivial::warning:
strm << "\033[33m";
color = true;
break;
case boost::log::trivial::error:
case boost::log::trivial::fatal:
strm << "\033[31m";
color = true;
break;
default:
break;
}
}
// setw(7) = the longest log level. eg "warning"
strm << std::setw(7) << std::setfill(' ') << severity << ": " << rec[boost::log::expressions::smessage];
if (color) {
strm << "\033[0m";
}
}
void logger_init_sink(bool use_colors = false) {
+ auto stream = &std::cerr;
+ if (getenv("LOG_STDERR") == nullptr) {
+ stream = &std::cout;
+ }
- auto sink = boost::log::add_console_log(std::cout, boost::log::keywords::format = "%Message%",
? ^^^^^^^
+ auto sink = boost::log::add_console_log(*stream, boost::log::keywords::format = "%Message%",
? + ^^^^
boost::log::keywords::auto_flush = true);
if (use_colors) {
sink->set_formatter(&color_fmt);
}
} | 6 | 0.181818 | 5 | 1 |
6717365393dbc078e3ae4d5374e354eefe43c7ce | README.md | README.md | Helper classes that are useful for Unity projects
I'll add helper classes that I often use in Unity projects
## Description
## Mass Set Tags Tool
This tool simply sets the tags of all selected objects and their children.
The class is `Scripts/EditorTools/Editor/MassSetTags.cs` and it can be accessed via `UnityHelpers -> Mass Set Tags` | Helper classes that are useful for Unity projects
I'll add helper classes that I often use in Unity projects
## Description
## Mass Set Tags Tool
This tool simply sets the tags of all selected objects and their children.
The class is `Scripts/EditorTools/Editor/MassSetTags.cs` and it can be accessed via `UnityHelpers -> Mass Set Tags`
## Singleton
A singleton pattern implemenation is found under `Scripts/Singleton/Singleton.cs` the full class name is `UnityHelpers.Singleton.Singleton<T>`.
## Component extensions
An extension for getting or adding a component to a gameobject is located in `Scripts/Components/ComponentExtensions`. This includes extension methods for `GameObject` and `Component`. | Add readme for singleton and component extension | Add readme for singleton and component extension
| Markdown | mit | petereichinger/unity-helpers | markdown | ## Code Before:
Helper classes that are useful for Unity projects
I'll add helper classes that I often use in Unity projects
## Description
## Mass Set Tags Tool
This tool simply sets the tags of all selected objects and their children.
The class is `Scripts/EditorTools/Editor/MassSetTags.cs` and it can be accessed via `UnityHelpers -> Mass Set Tags`
## Instruction:
Add readme for singleton and component extension
## Code After:
Helper classes that are useful for Unity projects
I'll add helper classes that I often use in Unity projects
## Description
## Mass Set Tags Tool
This tool simply sets the tags of all selected objects and their children.
The class is `Scripts/EditorTools/Editor/MassSetTags.cs` and it can be accessed via `UnityHelpers -> Mass Set Tags`
## Singleton
A singleton pattern implemenation is found under `Scripts/Singleton/Singleton.cs` the full class name is `UnityHelpers.Singleton.Singleton<T>`.
## Component extensions
An extension for getting or adding a component to a gameobject is located in `Scripts/Components/ComponentExtensions`. This includes extension methods for `GameObject` and `Component`. | Helper classes that are useful for Unity projects
I'll add helper classes that I often use in Unity projects
## Description
## Mass Set Tags Tool
This tool simply sets the tags of all selected objects and their children.
The class is `Scripts/EditorTools/Editor/MassSetTags.cs` and it can be accessed via `UnityHelpers -> Mass Set Tags`
+
+ ## Singleton
+
+ A singleton pattern implemenation is found under `Scripts/Singleton/Singleton.cs` the full class name is `UnityHelpers.Singleton.Singleton<T>`.
+
+ ## Component extensions
+
+ An extension for getting or adding a component to a gameobject is located in `Scripts/Components/ComponentExtensions`. This includes extension methods for `GameObject` and `Component`. | 8 | 0.8 | 8 | 0 |
5a5a69fc8535cc4cabf524f435ac31b71360601d | metadata/com.vanderbie.heart_rate_monitor.txt | metadata/com.vanderbie.heart_rate_monitor.txt | Categories:Science & Education
License:Apache2
Web Site:https://github.com/joeyvanderbie/android-heart-rate-monitor/blob/master/README.md
Source Code:https://github.com/joeyvanderbie/android-heart-rate-monitor
Issue Tracker:
Summary:Heart Rate Monitor
Description:
Uses the camera and flash to determine the users heart rate in beats per
minute. It can send captured data via UDP.
A fork of [[com.jwetherell.heart_rate_monitor]].
.
Repo Type:git
Repo:https://github.com/joeyvanderbie/android-heart-rate-monitor.git
Build:1.0,1
commit=45b8967c4
Auto Update Mode:None
Update Check Mode:RepoManifest
Current Version:1.0
Current Version Code:1
| Categories:Science & Education
License:Apache2
Web Site:https://github.com/joeyvanderbie/android-heart-rate-monitor/blob/master/README.md
Source Code:https://github.com/joeyvanderbie/android-heart-rate-monitor
Issue Tracker:
Auto Name:Heart rate
Summary:Heart Rate Monitor
Description:
Uses the camera and flash to determine the users heart rate in beats per
minute. It can send captured data via UDP.
A fork of [[com.jwetherell.heart_rate_monitor]].
.
Repo Type:git
Repo:https://github.com/joeyvanderbie/android-heart-rate-monitor.git
Build:1.0,1
commit=45b8967c4
Auto Update Mode:None
Update Check Mode:RepoManifest
Current Version:1.0
Current Version Code:1
| Set autoname of Heart rate | Set autoname of Heart rate
| Text | agpl-3.0 | f-droid/fdroiddata,f-droid/fdroiddata,f-droid/fdroid-data | text | ## Code Before:
Categories:Science & Education
License:Apache2
Web Site:https://github.com/joeyvanderbie/android-heart-rate-monitor/blob/master/README.md
Source Code:https://github.com/joeyvanderbie/android-heart-rate-monitor
Issue Tracker:
Summary:Heart Rate Monitor
Description:
Uses the camera and flash to determine the users heart rate in beats per
minute. It can send captured data via UDP.
A fork of [[com.jwetherell.heart_rate_monitor]].
.
Repo Type:git
Repo:https://github.com/joeyvanderbie/android-heart-rate-monitor.git
Build:1.0,1
commit=45b8967c4
Auto Update Mode:None
Update Check Mode:RepoManifest
Current Version:1.0
Current Version Code:1
## Instruction:
Set autoname of Heart rate
## Code After:
Categories:Science & Education
License:Apache2
Web Site:https://github.com/joeyvanderbie/android-heart-rate-monitor/blob/master/README.md
Source Code:https://github.com/joeyvanderbie/android-heart-rate-monitor
Issue Tracker:
Auto Name:Heart rate
Summary:Heart Rate Monitor
Description:
Uses the camera and flash to determine the users heart rate in beats per
minute. It can send captured data via UDP.
A fork of [[com.jwetherell.heart_rate_monitor]].
.
Repo Type:git
Repo:https://github.com/joeyvanderbie/android-heart-rate-monitor.git
Build:1.0,1
commit=45b8967c4
Auto Update Mode:None
Update Check Mode:RepoManifest
Current Version:1.0
Current Version Code:1
| Categories:Science & Education
License:Apache2
Web Site:https://github.com/joeyvanderbie/android-heart-rate-monitor/blob/master/README.md
Source Code:https://github.com/joeyvanderbie/android-heart-rate-monitor
Issue Tracker:
+ Auto Name:Heart rate
Summary:Heart Rate Monitor
Description:
Uses the camera and flash to determine the users heart rate in beats per
minute. It can send captured data via UDP.
A fork of [[com.jwetherell.heart_rate_monitor]].
.
Repo Type:git
Repo:https://github.com/joeyvanderbie/android-heart-rate-monitor.git
Build:1.0,1
commit=45b8967c4
Auto Update Mode:None
Update Check Mode:RepoManifest
Current Version:1.0
Current Version Code:1
| 1 | 0.04 | 1 | 0 |
9223c6b0e257556172a580c29e1140405c9f6fdc | docs/source/local_setup.rst | docs/source/local_setup.rst | Local setup
===========
Available hosts
---------------
::
localhost:8080
platform.genestack.org
internal-dev.genestack.com
Setup users for local installation
----------------------------------
To use local system installation add users via `genestack-user-setup.py`::
root
root@genestack.com
localhost:8080
pwdRoot
public
public@genestack.com
localhost:8080
pwdPublic
tester
tester@genestack.com
localhost:8080
pwdTester123
| Local setup
===========
Available hosts
---------------
::
localhost:8080
platform.genestack.org
internal-dev.genestack.com
Setup users for local installation
----------------------------------
To use local system installation add users via `genestack-user-setup.py`::
root
root@genestack.com
localhost:8080
pwdRoot
public
public@genestack.com
localhost:8080
pwdPublic
tester
tester@genestack.com
localhost:8080
pwdTester123
Make tester default user. `genestack-user-setup.py default tester`
| Add mention about tester default user for localhost. | Add mention about tester default user for localhost.
| reStructuredText | mit | genestack/python-client | restructuredtext | ## Code Before:
Local setup
===========
Available hosts
---------------
::
localhost:8080
platform.genestack.org
internal-dev.genestack.com
Setup users for local installation
----------------------------------
To use local system installation add users via `genestack-user-setup.py`::
root
root@genestack.com
localhost:8080
pwdRoot
public
public@genestack.com
localhost:8080
pwdPublic
tester
tester@genestack.com
localhost:8080
pwdTester123
## Instruction:
Add mention about tester default user for localhost.
## Code After:
Local setup
===========
Available hosts
---------------
::
localhost:8080
platform.genestack.org
internal-dev.genestack.com
Setup users for local installation
----------------------------------
To use local system installation add users via `genestack-user-setup.py`::
root
root@genestack.com
localhost:8080
pwdRoot
public
public@genestack.com
localhost:8080
pwdPublic
tester
tester@genestack.com
localhost:8080
pwdTester123
Make tester default user. `genestack-user-setup.py default tester`
| Local setup
===========
Available hosts
---------------
::
localhost:8080
platform.genestack.org
internal-dev.genestack.com
Setup users for local installation
----------------------------------
To use local system installation add users via `genestack-user-setup.py`::
root
root@genestack.com
localhost:8080
pwdRoot
public
public@genestack.com
localhost:8080
pwdPublic
tester
tester@genestack.com
localhost:8080
pwdTester123
+ Make tester default user. `genestack-user-setup.py default tester` | 1 | 0.03125 | 1 | 0 |
237a3f87f0d456d9b32cde42e4d7c783955f5a45 | spec/rails_helper.rb | spec/rails_helper.rb | require "spec_helper"
ENV["RAILS_ENV"] ||= "test"
require "simplecov"
SimpleCov.start "rails"
require File.expand_path("../config/environment", __dir__)
# Prevent database truncation if the environment is production
abort("The Rails environment is running in production mode!") if Rails.env.production?
require "rspec/rails"
Dir[Rails.root.join("spec/support/**/*.rb")].sort.each { |f| require f }
GovukTest.configure
require "webmock/rspec"
WebMock.disable_net_connect!(allow_localhost: true)
RSpec.configure do |config|
require "gds_api/test_helpers/content_store"
include GdsApi::TestHelpers::ContentStore
include ClickGovukStartButton
config.use_active_record = false # Remove this line to enable support for ActiveRecord
config.infer_spec_type_from_file_location!
# Filter lines from Rails gems in backtraces.
config.filter_rails_from_backtrace!
end
# Without this GovukContentSchemaTestHelpers configuration, get `NoMethodError: undefined method `<<' for nil:NilClass`
require "slimmer/test"
require "govuk-content-schema-test-helpers/test_unit"
GovukContentSchemaTestHelpers.configure do |config|
config.schema_type = "publisher_v2"
config.project_root = Rails.root
end
| require "spec_helper"
ENV["RAILS_ENV"] ||= "test"
require "simplecov"
SimpleCov.start "rails" do
add_group "Smart Answer Flows", "lib/smart_answer_flows"
end
require File.expand_path("../config/environment", __dir__)
# Prevent database truncation if the environment is production
abort("The Rails environment is running in production mode!") if Rails.env.production?
require "rspec/rails"
Dir[Rails.root.join("spec/support/**/*.rb")].sort.each { |f| require f }
GovukTest.configure
require "webmock/rspec"
WebMock.disable_net_connect!(allow_localhost: true)
RSpec.configure do |config|
require "gds_api/test_helpers/content_store"
include GdsApi::TestHelpers::ContentStore
include ClickGovukStartButton
config.use_active_record = false # Remove this line to enable support for ActiveRecord
config.infer_spec_type_from_file_location!
# Filter lines from Rails gems in backtraces.
config.filter_rails_from_backtrace!
end
# Without this GovukContentSchemaTestHelpers configuration, get `NoMethodError: undefined method `<<' for nil:NilClass`
require "slimmer/test"
require "govuk-content-schema-test-helpers/test_unit"
GovukContentSchemaTestHelpers.configure do |config|
config.schema_type = "publisher_v2"
config.project_root = Rails.root
end
| Add grouping to SimpleCov report | Add grouping to SimpleCov report
What
Add a report grouping (tab) specifically for flows to the SimpleCov report output for RSpec tests.
Why
Makes finding the flow tests easier to find.
| Ruby | mit | alphagov/smart-answers,alphagov/smart-answers,alphagov/smart-answers,alphagov/smart-answers | ruby | ## Code Before:
require "spec_helper"
ENV["RAILS_ENV"] ||= "test"
require "simplecov"
SimpleCov.start "rails"
require File.expand_path("../config/environment", __dir__)
# Prevent database truncation if the environment is production
abort("The Rails environment is running in production mode!") if Rails.env.production?
require "rspec/rails"
Dir[Rails.root.join("spec/support/**/*.rb")].sort.each { |f| require f }
GovukTest.configure
require "webmock/rspec"
WebMock.disable_net_connect!(allow_localhost: true)
RSpec.configure do |config|
require "gds_api/test_helpers/content_store"
include GdsApi::TestHelpers::ContentStore
include ClickGovukStartButton
config.use_active_record = false # Remove this line to enable support for ActiveRecord
config.infer_spec_type_from_file_location!
# Filter lines from Rails gems in backtraces.
config.filter_rails_from_backtrace!
end
# Without this GovukContentSchemaTestHelpers configuration, get `NoMethodError: undefined method `<<' for nil:NilClass`
require "slimmer/test"
require "govuk-content-schema-test-helpers/test_unit"
GovukContentSchemaTestHelpers.configure do |config|
config.schema_type = "publisher_v2"
config.project_root = Rails.root
end
## Instruction:
Add grouping to SimpleCov report
What
Add a report grouping (tab) specifically for flows to the SimpleCov report output for RSpec tests.
Why
Makes finding the flow tests easier to find.
## Code After:
require "spec_helper"
ENV["RAILS_ENV"] ||= "test"
require "simplecov"
SimpleCov.start "rails" do
add_group "Smart Answer Flows", "lib/smart_answer_flows"
end
require File.expand_path("../config/environment", __dir__)
# Prevent database truncation if the environment is production
abort("The Rails environment is running in production mode!") if Rails.env.production?
require "rspec/rails"
Dir[Rails.root.join("spec/support/**/*.rb")].sort.each { |f| require f }
GovukTest.configure
require "webmock/rspec"
WebMock.disable_net_connect!(allow_localhost: true)
RSpec.configure do |config|
require "gds_api/test_helpers/content_store"
include GdsApi::TestHelpers::ContentStore
include ClickGovukStartButton
config.use_active_record = false # Remove this line to enable support for ActiveRecord
config.infer_spec_type_from_file_location!
# Filter lines from Rails gems in backtraces.
config.filter_rails_from_backtrace!
end
# Without this GovukContentSchemaTestHelpers configuration, get `NoMethodError: undefined method `<<' for nil:NilClass`
require "slimmer/test"
require "govuk-content-schema-test-helpers/test_unit"
GovukContentSchemaTestHelpers.configure do |config|
config.schema_type = "publisher_v2"
config.project_root = Rails.root
end
| require "spec_helper"
ENV["RAILS_ENV"] ||= "test"
require "simplecov"
+
- SimpleCov.start "rails"
+ SimpleCov.start "rails" do
? +++
+ add_group "Smart Answer Flows", "lib/smart_answer_flows"
+ end
require File.expand_path("../config/environment", __dir__)
# Prevent database truncation if the environment is production
abort("The Rails environment is running in production mode!") if Rails.env.production?
require "rspec/rails"
Dir[Rails.root.join("spec/support/**/*.rb")].sort.each { |f| require f }
GovukTest.configure
require "webmock/rspec"
WebMock.disable_net_connect!(allow_localhost: true)
RSpec.configure do |config|
require "gds_api/test_helpers/content_store"
include GdsApi::TestHelpers::ContentStore
include ClickGovukStartButton
config.use_active_record = false # Remove this line to enable support for ActiveRecord
config.infer_spec_type_from_file_location!
# Filter lines from Rails gems in backtraces.
config.filter_rails_from_backtrace!
end
# Without this GovukContentSchemaTestHelpers configuration, get `NoMethodError: undefined method `<<' for nil:NilClass`
require "slimmer/test"
require "govuk-content-schema-test-helpers/test_unit"
GovukContentSchemaTestHelpers.configure do |config|
config.schema_type = "publisher_v2"
config.project_root = Rails.root
end | 5 | 0.131579 | 4 | 1 |
07c81824d684d3fac422330bb6506ec8a2afc10e | template/.travis.yml | template/.travis.yml | language: java
sudo: required
jdk:
- oraclejdk8
env:
matrix:
- ENGINE=lucee@4.5
- ENGINE=lucee@5
- ENGINE=adobe@2016
- ENGINE=adobe@11
- ENGINE=adobe@10
before_install:
- sudo apt-key adv --keyserver keys.gnupg.net --recv 6DA70622
- sudo echo "deb http://downloads.ortussolutions.com/debs/noarch /" | sudo tee -a
/etc/apt/sources.list.d/commandbox.list
install:
- sudo apt-get update && sudo apt-get --assume-yes install commandbox
- box install
before_script:
- box server start cfengine=$ENGINE port=8500
script: >
testResults="$(box testbox run runner='http://127.0.0.1:8500/tests/runner.cfm' )";
echo "$testResults";
if grep -i "\[(Failures|Errors): [1-9][0-9]\?[0-9]\?\]\|\[Errors: [1-9][0-9]\?[0-9]\?\]\|<t[^>]*>\|<b[^>]*>" <<< $testResults; then exit 1; fi
notifications:
email: false | language: java
sudo: required
jdk:
- oraclejdk8
env:
matrix:
- ENGINE=lucee@4.5
- ENGINE=lucee@5
- ENGINE=adobe@2016
- ENGINE=adobe@11
- ENGINE=adobe@10
before_install:
- sudo apt-key adv --keyserver keys.gnupg.net --recv 6DA70622
- sudo echo "deb http://downloads.ortussolutions.com/debs/noarch /" | sudo tee -a
/etc/apt/sources.list.d/commandbox.list
install:
- sudo apt-get update && sudo apt-get --assume-yes install commandbox
- box install
before_script:
- box server start cfengine=$ENGINE port=8500
script: >
testResults="$(box testbox run runner='http://127.0.0.1:8500/tests/runner.cfm' )";
echo "$testResults";
if grep -i "\[Failures: [1-9][0-9]\?\]\|<t[^>]*>\|<b[^>]*>" <<< $testResults; then exit 1; fi
notifications:
email: false
| Fix regex to fail tests | Fix regex to fail tests | YAML | mit | elpete/cb-module-template | yaml | ## Code Before:
language: java
sudo: required
jdk:
- oraclejdk8
env:
matrix:
- ENGINE=lucee@4.5
- ENGINE=lucee@5
- ENGINE=adobe@2016
- ENGINE=adobe@11
- ENGINE=adobe@10
before_install:
- sudo apt-key adv --keyserver keys.gnupg.net --recv 6DA70622
- sudo echo "deb http://downloads.ortussolutions.com/debs/noarch /" | sudo tee -a
/etc/apt/sources.list.d/commandbox.list
install:
- sudo apt-get update && sudo apt-get --assume-yes install commandbox
- box install
before_script:
- box server start cfengine=$ENGINE port=8500
script: >
testResults="$(box testbox run runner='http://127.0.0.1:8500/tests/runner.cfm' )";
echo "$testResults";
if grep -i "\[(Failures|Errors): [1-9][0-9]\?[0-9]\?\]\|\[Errors: [1-9][0-9]\?[0-9]\?\]\|<t[^>]*>\|<b[^>]*>" <<< $testResults; then exit 1; fi
notifications:
email: false
## Instruction:
Fix regex to fail tests
## Code After:
language: java
sudo: required
jdk:
- oraclejdk8
env:
matrix:
- ENGINE=lucee@4.5
- ENGINE=lucee@5
- ENGINE=adobe@2016
- ENGINE=adobe@11
- ENGINE=adobe@10
before_install:
- sudo apt-key adv --keyserver keys.gnupg.net --recv 6DA70622
- sudo echo "deb http://downloads.ortussolutions.com/debs/noarch /" | sudo tee -a
/etc/apt/sources.list.d/commandbox.list
install:
- sudo apt-get update && sudo apt-get --assume-yes install commandbox
- box install
before_script:
- box server start cfengine=$ENGINE port=8500
script: >
testResults="$(box testbox run runner='http://127.0.0.1:8500/tests/runner.cfm' )";
echo "$testResults";
if grep -i "\[Failures: [1-9][0-9]\?\]\|<t[^>]*>\|<b[^>]*>" <<< $testResults; then exit 1; fi
notifications:
email: false
| language: java
sudo: required
jdk:
- oraclejdk8
env:
matrix:
- ENGINE=lucee@4.5
- ENGINE=lucee@5
- ENGINE=adobe@2016
- ENGINE=adobe@11
- ENGINE=adobe@10
before_install:
- sudo apt-key adv --keyserver keys.gnupg.net --recv 6DA70622
- sudo echo "deb http://downloads.ortussolutions.com/debs/noarch /" | sudo tee -a
/etc/apt/sources.list.d/commandbox.list
install:
- sudo apt-get update && sudo apt-get --assume-yes install commandbox
- box install
before_script:
- box server start cfengine=$ENGINE port=8500
script: >
testResults="$(box testbox run runner='http://127.0.0.1:8500/tests/runner.cfm' )";
echo "$testResults";
- if grep -i "\[(Failures|Errors): [1-9][0-9]\?[0-9]\?\]\|\[Errors: [1-9][0-9]\?[0-9]\?\]\|<t[^>]*>\|<b[^>]*>" <<< $testResults; then exit 1; fi
? - -------- ----------------------------------------
+ if grep -i "\[Failures: [1-9][0-9]\?\]\|<t[^>]*>\|<b[^>]*>" <<< $testResults; then exit 1; fi
? +
notifications:
email: false | 2 | 0.076923 | 1 | 1 |
00cac9ff0e4c6aeba168ab4f524a6fbe56ae39ee | packages/cm/cmark-sections.yaml | packages/cm/cmark-sections.yaml | homepage: http://github.com/aelve/cmark-sections
changelog-type: markdown
hash: 5ff7821b64fd89f94ed4453b4719cf5ce50a0eeb4ae71b25c1ba283cb6317db8
test-bench-deps:
base-prelude: ==1.*
base: ! '>=4.7 && <5'
hspec: ! '>=2.2 && <2.4'
text: -any
cmark: -any
cmark-sections: -any
containers: -any
QuickCheck: ! '>=2.8 && <3'
maintainer: yom@artyom.me
synopsis: Represent cmark-parsed Markdown as a tree of sections
changelog: ! '# 0.1.0.1
* Bumped hspec version.
# 0.1.0.0
First release.
'
basic-deps:
base-prelude: ==1.*
split: ==0.2.*
base: ! '>=4.7 && <5'
text: -any
cmark: ! '>=0.5 && <0.5.4'
containers: -any
microlens: ==0.4.*
all-versions:
- '0.1.0.0'
- '0.1.0.1'
author: Artyom
latest: '0.1.0.1'
description-type: haddock
description: Represent cmark-parsed Markdown as a tree of sections
license-name: BSD3
| homepage: http://github.com/aelve/cmark-sections
changelog-type: markdown
hash: b08cd73ff6103e39e763f5809d757887399059505fe8ac468ef377dc616962ef
test-bench-deps:
base-prelude: ==1.*
base: ! '>=4.7 && <5'
hspec: ! '>=2.2 && <2.4'
text: -any
cmark: -any
cmark-sections: -any
containers: -any
QuickCheck: ! '>=2.8 && <3'
maintainer: yom@artyom.me
synopsis: Represent cmark-parsed Markdown as a tree of sections
changelog: ! '# 0.1.0.2
* Bumped cmark version.
# 0.1.0.1
* Bumped hspec version.
# 0.1.0.0
First release.
'
basic-deps:
base-prelude: ==1.*
split: ==0.2.*
base: ! '>=4.7 && <5'
text: -any
cmark: ! '>=0.5 && <0.5.5'
containers: -any
microlens: ==0.4.*
all-versions:
- '0.1.0.0'
- '0.1.0.1'
- '0.1.0.2'
author: Artyom
latest: '0.1.0.2'
description-type: haddock
description: Represent cmark-parsed Markdown as a tree of sections
license-name: BSD3
| Update from Hackage at 2016-11-19T11:22:51Z | Update from Hackage at 2016-11-19T11:22:51Z | YAML | mit | commercialhaskell/all-cabal-metadata | yaml | ## Code Before:
homepage: http://github.com/aelve/cmark-sections
changelog-type: markdown
hash: 5ff7821b64fd89f94ed4453b4719cf5ce50a0eeb4ae71b25c1ba283cb6317db8
test-bench-deps:
base-prelude: ==1.*
base: ! '>=4.7 && <5'
hspec: ! '>=2.2 && <2.4'
text: -any
cmark: -any
cmark-sections: -any
containers: -any
QuickCheck: ! '>=2.8 && <3'
maintainer: yom@artyom.me
synopsis: Represent cmark-parsed Markdown as a tree of sections
changelog: ! '# 0.1.0.1
* Bumped hspec version.
# 0.1.0.0
First release.
'
basic-deps:
base-prelude: ==1.*
split: ==0.2.*
base: ! '>=4.7 && <5'
text: -any
cmark: ! '>=0.5 && <0.5.4'
containers: -any
microlens: ==0.4.*
all-versions:
- '0.1.0.0'
- '0.1.0.1'
author: Artyom
latest: '0.1.0.1'
description-type: haddock
description: Represent cmark-parsed Markdown as a tree of sections
license-name: BSD3
## Instruction:
Update from Hackage at 2016-11-19T11:22:51Z
## Code After:
homepage: http://github.com/aelve/cmark-sections
changelog-type: markdown
hash: b08cd73ff6103e39e763f5809d757887399059505fe8ac468ef377dc616962ef
test-bench-deps:
base-prelude: ==1.*
base: ! '>=4.7 && <5'
hspec: ! '>=2.2 && <2.4'
text: -any
cmark: -any
cmark-sections: -any
containers: -any
QuickCheck: ! '>=2.8 && <3'
maintainer: yom@artyom.me
synopsis: Represent cmark-parsed Markdown as a tree of sections
changelog: ! '# 0.1.0.2
* Bumped cmark version.
# 0.1.0.1
* Bumped hspec version.
# 0.1.0.0
First release.
'
basic-deps:
base-prelude: ==1.*
split: ==0.2.*
base: ! '>=4.7 && <5'
text: -any
cmark: ! '>=0.5 && <0.5.5'
containers: -any
microlens: ==0.4.*
all-versions:
- '0.1.0.0'
- '0.1.0.1'
- '0.1.0.2'
author: Artyom
latest: '0.1.0.2'
description-type: haddock
description: Represent cmark-parsed Markdown as a tree of sections
license-name: BSD3
| homepage: http://github.com/aelve/cmark-sections
changelog-type: markdown
- hash: 5ff7821b64fd89f94ed4453b4719cf5ce50a0eeb4ae71b25c1ba283cb6317db8
+ hash: b08cd73ff6103e39e763f5809d757887399059505fe8ac468ef377dc616962ef
test-bench-deps:
base-prelude: ==1.*
base: ! '>=4.7 && <5'
hspec: ! '>=2.2 && <2.4'
text: -any
cmark: -any
cmark-sections: -any
containers: -any
QuickCheck: ! '>=2.8 && <3'
maintainer: yom@artyom.me
synopsis: Represent cmark-parsed Markdown as a tree of sections
- changelog: ! '# 0.1.0.1
? ^
+ changelog: ! '# 0.1.0.2
? ^
+
+
+ * Bumped cmark version.
+
+
+ # 0.1.0.1
* Bumped hspec version.
# 0.1.0.0
First release.
'
basic-deps:
base-prelude: ==1.*
split: ==0.2.*
base: ! '>=4.7 && <5'
text: -any
- cmark: ! '>=0.5 && <0.5.4'
? ^
+ cmark: ! '>=0.5 && <0.5.5'
? ^
containers: -any
microlens: ==0.4.*
all-versions:
- '0.1.0.0'
- '0.1.0.1'
+ - '0.1.0.2'
author: Artyom
- latest: '0.1.0.1'
? ^
+ latest: '0.1.0.2'
? ^
description-type: haddock
description: Represent cmark-parsed Markdown as a tree of sections
license-name: BSD3 | 15 | 0.357143 | 11 | 4 |
af09689eaab30c8f378fbfe52104a2da836ce014 | README.md | README.md | actionpack-xml\_parser
======================
A XML parameters parser for Action Pack (removed from core in Rails 4.0)
Installation
------------
Include this gem into your Gemfile:
```ruby
gem 'actionpack-xml_parser'
```
Then, add `ActionDispatch::XmlParamsParser` middleware after `ActionDispatch::ParamsParser`
in `config/application.rb`:
```ruby
config.middleware.insert_after ActionDispatch::ParamsParser, ActionDispatch::XmlParamsParser
```
You may need to require the `ActionDispatch::XmlParamsParser` manually. Add the following
in your `config/application.rb`:
```ruby
require 'action_dispatch/xml_params_parser'
```
| actionpack-xml\_parser
======================
A XML parameters parser for Action Pack (removed from core in Rails 4.0)
Installation
------------
Include this gem into your Gemfile:
```ruby
gem 'actionpack-xml_parser'
```
Then, add `ActionDispatch::XmlParamsParser` middleware after `ActionDispatch::ParamsParser`
in `config/application.rb`:
```ruby
config.middleware.insert_after ActionDispatch::ParamsParser, ActionDispatch::XmlParamsParser
```
You may need to require the `ActionDispatch::XmlParamsParser` manually. Add the following
in your `config/application.rb`:
```ruby
require 'action_dispatch/xml_params_parser'
```
Parameters parsing rules
------------------------
The parameters parsing is handled by `ActiveSupport::XMLConverter` so there may
be specific features and subtle differences depending on the chosen XML backend.
### Hashes
Basically, each node represents a key. With the following XML:
```xml
<person><name>David</name></person>
```
The resulting parameters will be:
```ruby
{"person" => {"name" => "David"}}
```
### File attachment
You can specify the `type` attribute of a node to attach files:
```xml
<person>
<avatar type="file" name="me.jpg" content_type="image/jpg"><!-- File content --></avatar>
</person>
```
The resulting parameters will include a `StringIO` object with the given content,
name and content type set accordingly:
```ruby
{"person" => {"avatar" => #<StringIO:...>}}
```
### Arrays
There are several ways to pass an array. You can either specify multiple nodes
with the same name:
```xml
<person>
<address city="Chicago"/>
<address city="Ottawa"/>
</person>
```
The resulting parameters will be:
```ruby
{"person" => {"address" => [{"city" => "Chicago"}, {"city" => "Ottawa"}]}}
```
You can also specify the `type` attribute of a node and nest child nodes inside:
```xml
<person>
<addresses type="array">
<address city="Melbourne"/>
<address city="Paris"/>
</addresses>
</person>
```
will result in:
```ruby
{"person" => {"addresses" => [{"city" => "Melbourne"}, {"city" => "Paris"}]}}
```
| Add examples of parameters parsing | Add examples of parameters parsing [ci skip]
| Markdown | mit | rails/actionpack-xml_parser,eileencodes/actionpack-xml_parser | markdown | ## Code Before:
actionpack-xml\_parser
======================
A XML parameters parser for Action Pack (removed from core in Rails 4.0)
Installation
------------
Include this gem into your Gemfile:
```ruby
gem 'actionpack-xml_parser'
```
Then, add `ActionDispatch::XmlParamsParser` middleware after `ActionDispatch::ParamsParser`
in `config/application.rb`:
```ruby
config.middleware.insert_after ActionDispatch::ParamsParser, ActionDispatch::XmlParamsParser
```
You may need to require the `ActionDispatch::XmlParamsParser` manually. Add the following
in your `config/application.rb`:
```ruby
require 'action_dispatch/xml_params_parser'
```
## Instruction:
Add examples of parameters parsing [ci skip]
## Code After:
actionpack-xml\_parser
======================
A XML parameters parser for Action Pack (removed from core in Rails 4.0)
Installation
------------
Include this gem into your Gemfile:
```ruby
gem 'actionpack-xml_parser'
```
Then, add `ActionDispatch::XmlParamsParser` middleware after `ActionDispatch::ParamsParser`
in `config/application.rb`:
```ruby
config.middleware.insert_after ActionDispatch::ParamsParser, ActionDispatch::XmlParamsParser
```
You may need to require the `ActionDispatch::XmlParamsParser` manually. Add the following
in your `config/application.rb`:
```ruby
require 'action_dispatch/xml_params_parser'
```
Parameters parsing rules
------------------------
The parameters parsing is handled by `ActiveSupport::XMLConverter` so there may
be specific features and subtle differences depending on the chosen XML backend.
### Hashes
Basically, each node represents a key. With the following XML:
```xml
<person><name>David</name></person>
```
The resulting parameters will be:
```ruby
{"person" => {"name" => "David"}}
```
### File attachment
You can specify the `type` attribute of a node to attach files:
```xml
<person>
<avatar type="file" name="me.jpg" content_type="image/jpg"><!-- File content --></avatar>
</person>
```
The resulting parameters will include a `StringIO` object with the given content,
name and content type set accordingly:
```ruby
{"person" => {"avatar" => #<StringIO:...>}}
```
### Arrays
There are several ways to pass an array. You can either specify multiple nodes
with the same name:
```xml
<person>
<address city="Chicago"/>
<address city="Ottawa"/>
</person>
```
The resulting parameters will be:
```ruby
{"person" => {"address" => [{"city" => "Chicago"}, {"city" => "Ottawa"}]}}
```
You can also specify the `type` attribute of a node and nest child nodes inside:
```xml
<person>
<addresses type="array">
<address city="Melbourne"/>
<address city="Paris"/>
</addresses>
</person>
```
will result in:
```ruby
{"person" => {"addresses" => [{"city" => "Melbourne"}, {"city" => "Paris"}]}}
```
| actionpack-xml\_parser
======================
A XML parameters parser for Action Pack (removed from core in Rails 4.0)
Installation
------------
Include this gem into your Gemfile:
```ruby
gem 'actionpack-xml_parser'
```
Then, add `ActionDispatch::XmlParamsParser` middleware after `ActionDispatch::ParamsParser`
in `config/application.rb`:
```ruby
config.middleware.insert_after ActionDispatch::ParamsParser, ActionDispatch::XmlParamsParser
```
- You may need to require the `ActionDispatch::XmlParamsParser` manually. Add the following
? -
+ You may need to require the `ActionDispatch::XmlParamsParser` manually. Add the following
in your `config/application.rb`:
```ruby
require 'action_dispatch/xml_params_parser'
```
+
+ Parameters parsing rules
+ ------------------------
+
+ The parameters parsing is handled by `ActiveSupport::XMLConverter` so there may
+ be specific features and subtle differences depending on the chosen XML backend.
+
+ ### Hashes
+
+ Basically, each node represents a key. With the following XML:
+
+ ```xml
+ <person><name>David</name></person>
+ ```
+
+ The resulting parameters will be:
+
+ ```ruby
+ {"person" => {"name" => "David"}}
+ ```
+
+ ### File attachment
+
+ You can specify the `type` attribute of a node to attach files:
+
+ ```xml
+ <person>
+ <avatar type="file" name="me.jpg" content_type="image/jpg"><!-- File content --></avatar>
+ </person>
+ ```
+
+ The resulting parameters will include a `StringIO` object with the given content,
+ name and content type set accordingly:
+
+ ```ruby
+ {"person" => {"avatar" => #<StringIO:...>}}
+ ```
+
+ ### Arrays
+
+ There are several ways to pass an array. You can either specify multiple nodes
+ with the same name:
+
+ ```xml
+ <person>
+ <address city="Chicago"/>
+ <address city="Ottawa"/>
+ </person>
+ ```
+
+ The resulting parameters will be:
+
+ ```ruby
+ {"person" => {"address" => [{"city" => "Chicago"}, {"city" => "Ottawa"}]}}
+ ```
+
+ You can also specify the `type` attribute of a node and nest child nodes inside:
+
+ ```xml
+ <person>
+ <addresses type="array">
+ <address city="Melbourne"/>
+ <address city="Paris"/>
+ </addresses>
+ </person>
+ ```
+
+ will result in:
+
+ ```ruby
+ {"person" => {"addresses" => [{"city" => "Melbourne"}, {"city" => "Paris"}]}}
+ ``` | 74 | 2.740741 | 73 | 1 |
93687c1491e8df144829804e7d7c142ca70d9f32 | lib/Slic3r/GUI/Notifier.pm | lib/Slic3r/GUI/Notifier.pm | package Slic3r::GUI::Notifier;
use Moo;
has 'growler' => (is => 'rw');
my $icon = "$Slic3r::var/Slic3r.png";
sub BUILD {
my ($self) = @_;
if (eval 'use Growl::GNTP; 1') {
# register with growl
eval {
$self->growler(Growl::GNTP->new(AppName => 'Slic3r', AppIcon => $icon));
$self->growler->register([{Name => 'SKEIN_DONE', DisplayName => 'Slicing Done'}]);
};
}
}
sub notify {
my ($self, $message) = @_;
my $title = 'Slicing Done!';
eval {
$self->growler->notify(Event => 'SKEIN_DONE', Title => $title, Message => $message)
if $self->growler;
};
# Net::DBus is broken in multithreaded environment
if (0 && eval 'use Net::DBus; 1') {
eval {
my $session = Net::DBus->session;
my $serv = $session->get_service('org.freedesktop.Notifications');
my $notifier = $serv->get_object('/org/freedesktop/Notifications',
'org.freedesktop.Notifications');
$notifier->Notify('Slic3r', 0, $icon, $title, $message, [], {}, -1);
undef $Net::DBus::bus_session;
};
}
}
1;
| package Slic3r::GUI::Notifier;
use Moo;
has 'growler' => (is => 'rw');
my $icon = "$Slic3r::var/Slic3r.png";
sub BUILD {
my ($self) = @_;
if (eval 'use Growl::GNTP; 1') {
# register with growl
eval {
$self->growler(Growl::GNTP->new(AppName => 'Slic3r', AppIcon => $icon));
$self->growler->register([{Name => 'SKEIN_DONE', DisplayName => 'Slicing Done'}]);
};
# if register() fails (for example because of a timeout), disable growler at all
$self->growler(undef) if $@;
}
}
sub notify {
my ($self, $message) = @_;
my $title = 'Slicing Done!';
eval {
$self->growler->notify(Event => 'SKEIN_DONE', Title => $title, Message => $message)
if $self->growler;
};
# Net::DBus is broken in multithreaded environment
if (0 && eval 'use Net::DBus; 1') {
eval {
my $session = Net::DBus->session;
my $serv = $session->get_service('org.freedesktop.Notifications');
my $notifier = $serv->get_object('/org/freedesktop/Notifications',
'org.freedesktop.Notifications');
$notifier->Notify('Slic3r', 0, $icon, $title, $message, [], {}, -1);
undef $Net::DBus::bus_session;
};
}
}
1;
| Disable Growl notifications if register() fails - this prevents a deadlock on growler->notify() | Disable Growl notifications if register() fails - this prevents a deadlock on growler->notify()
| Perl | agpl-3.0 | platsch/Slic3r,lordofhyphens/Slic3r,platsch/Slic3r,tmotl/Slic3r,alexrj/Slic3r,alexrj/Slic3r,xoan/Slic3r,lordofhyphens/Slic3r,alexrj/Slic3r,xoan/Slic3r,pieis2pi/Slic3r,prusa3d/Slic3r,prusa3d/Slic3r,prusa3d/Slic3r,mcilhargey/Slic3r,curieos/Slic3r,mcilhargey/Slic3r,lordofhyphens/Slic3r,platsch/Slic3r,xoan/Slic3r,curieos/Slic3r,tmotl/Slic3r,curieos/Slic3r,pieis2pi/Slic3r,curieos/Slic3r,lordofhyphens/Slic3r,alexrj/Slic3r,pieis2pi/Slic3r,prusa3d/Slic3r,tmotl/Slic3r,alexrj/Slic3r,prusa3d/Slic3r,platsch/Slic3r,mcilhargey/Slic3r,tmotl/Slic3r,prusa3d/Slic3r,xoan/Slic3r,prusa3d/Slic3r,platsch/Slic3r,curieos/Slic3r,pieis2pi/Slic3r,lordofhyphens/Slic3r,curieos/Slic3r,platsch/Slic3r,pieis2pi/Slic3r,mcilhargey/Slic3r,mcilhargey/Slic3r,pieis2pi/Slic3r,xoan/Slic3r,mcilhargey/Slic3r,xoan/Slic3r,mcilhargey/Slic3r,alexrj/Slic3r,tmotl/Slic3r,lordofhyphens/Slic3r,tmotl/Slic3r | perl | ## Code Before:
package Slic3r::GUI::Notifier;
use Moo;
has 'growler' => (is => 'rw');
my $icon = "$Slic3r::var/Slic3r.png";
sub BUILD {
my ($self) = @_;
if (eval 'use Growl::GNTP; 1') {
# register with growl
eval {
$self->growler(Growl::GNTP->new(AppName => 'Slic3r', AppIcon => $icon));
$self->growler->register([{Name => 'SKEIN_DONE', DisplayName => 'Slicing Done'}]);
};
}
}
sub notify {
my ($self, $message) = @_;
my $title = 'Slicing Done!';
eval {
$self->growler->notify(Event => 'SKEIN_DONE', Title => $title, Message => $message)
if $self->growler;
};
# Net::DBus is broken in multithreaded environment
if (0 && eval 'use Net::DBus; 1') {
eval {
my $session = Net::DBus->session;
my $serv = $session->get_service('org.freedesktop.Notifications');
my $notifier = $serv->get_object('/org/freedesktop/Notifications',
'org.freedesktop.Notifications');
$notifier->Notify('Slic3r', 0, $icon, $title, $message, [], {}, -1);
undef $Net::DBus::bus_session;
};
}
}
1;
## Instruction:
Disable Growl notifications if register() fails - this prevents a deadlock on growler->notify()
## Code After:
package Slic3r::GUI::Notifier;
use Moo;
has 'growler' => (is => 'rw');
my $icon = "$Slic3r::var/Slic3r.png";
sub BUILD {
my ($self) = @_;
if (eval 'use Growl::GNTP; 1') {
# register with growl
eval {
$self->growler(Growl::GNTP->new(AppName => 'Slic3r', AppIcon => $icon));
$self->growler->register([{Name => 'SKEIN_DONE', DisplayName => 'Slicing Done'}]);
};
# if register() fails (for example because of a timeout), disable growler at all
$self->growler(undef) if $@;
}
}
sub notify {
my ($self, $message) = @_;
my $title = 'Slicing Done!';
eval {
$self->growler->notify(Event => 'SKEIN_DONE', Title => $title, Message => $message)
if $self->growler;
};
# Net::DBus is broken in multithreaded environment
if (0 && eval 'use Net::DBus; 1') {
eval {
my $session = Net::DBus->session;
my $serv = $session->get_service('org.freedesktop.Notifications');
my $notifier = $serv->get_object('/org/freedesktop/Notifications',
'org.freedesktop.Notifications');
$notifier->Notify('Slic3r', 0, $icon, $title, $message, [], {}, -1);
undef $Net::DBus::bus_session;
};
}
}
1;
| package Slic3r::GUI::Notifier;
use Moo;
has 'growler' => (is => 'rw');
my $icon = "$Slic3r::var/Slic3r.png";
sub BUILD {
my ($self) = @_;
if (eval 'use Growl::GNTP; 1') {
# register with growl
eval {
$self->growler(Growl::GNTP->new(AppName => 'Slic3r', AppIcon => $icon));
$self->growler->register([{Name => 'SKEIN_DONE', DisplayName => 'Slicing Done'}]);
};
+ # if register() fails (for example because of a timeout), disable growler at all
+ $self->growler(undef) if $@;
}
}
sub notify {
my ($self, $message) = @_;
my $title = 'Slicing Done!';
eval {
$self->growler->notify(Event => 'SKEIN_DONE', Title => $title, Message => $message)
if $self->growler;
};
# Net::DBus is broken in multithreaded environment
if (0 && eval 'use Net::DBus; 1') {
eval {
my $session = Net::DBus->session;
my $serv = $session->get_service('org.freedesktop.Notifications');
my $notifier = $serv->get_object('/org/freedesktop/Notifications',
'org.freedesktop.Notifications');
$notifier->Notify('Slic3r', 0, $icon, $title, $message, [], {}, -1);
undef $Net::DBus::bus_session;
};
}
}
1; | 2 | 0.04878 | 2 | 0 |
09cc981721efb28717d58a7e5654270cf0a029bb | services/web/src/main/ember/app/serializers/application.js | services/web/src/main/ember/app/serializers/application.js | import Ember from 'ember';
import DS from 'ember-data';
// Adapted from http://springember.blogspot.com.au/2014/08/using-ember-data-restadapter-with.html
export default DS.RESTSerializer.extend({
serializeIntoHash: function(hash, type, record, options) {
var serialized = this.serialize(record, options);
//Include the id in the payload
//Jackson was complaining when it received a null id ...
serialized.id = record.id ? record.id : null;
//remove the root element
Ember.merge(hash, serialized);
}
});
| import Ember from 'ember';
import DS from 'ember-data';
// Adapted from http://springember.blogspot.com.au/2014/08/using-ember-data-restadapter-with.html
export default DS.RESTSerializer.extend({
serializeIntoHash: function(hash, type, record, options) {
var serialized = this.serialize(record, options);
//Remove id from the payload for new records
//Jackson was complaining when it received a null id ...
if (record.id == null) {
delete serialized.id;
}
//remove the root element
Ember.merge(hash, serialized);
}
});
| Fix Jackson failure when id:null was present in the json request | Fix Jackson failure when id:null was present in the json request
| JavaScript | agpl-3.0 | MarSik/shelves,MarSik/shelves,MarSik/shelves,MarSik/shelves | javascript | ## Code Before:
import Ember from 'ember';
import DS from 'ember-data';
// Adapted from http://springember.blogspot.com.au/2014/08/using-ember-data-restadapter-with.html
export default DS.RESTSerializer.extend({
serializeIntoHash: function(hash, type, record, options) {
var serialized = this.serialize(record, options);
//Include the id in the payload
//Jackson was complaining when it received a null id ...
serialized.id = record.id ? record.id : null;
//remove the root element
Ember.merge(hash, serialized);
}
});
## Instruction:
Fix Jackson failure when id:null was present in the json request
## Code After:
import Ember from 'ember';
import DS from 'ember-data';
// Adapted from http://springember.blogspot.com.au/2014/08/using-ember-data-restadapter-with.html
export default DS.RESTSerializer.extend({
serializeIntoHash: function(hash, type, record, options) {
var serialized = this.serialize(record, options);
//Remove id from the payload for new records
//Jackson was complaining when it received a null id ...
if (record.id == null) {
delete serialized.id;
}
//remove the root element
Ember.merge(hash, serialized);
}
});
| import Ember from 'ember';
import DS from 'ember-data';
// Adapted from http://springember.blogspot.com.au/2014/08/using-ember-data-restadapter-with.html
export default DS.RESTSerializer.extend({
serializeIntoHash: function(hash, type, record, options) {
var serialized = this.serialize(record, options);
- //Include the id in the payload
+ //Remove id from the payload for new records
//Jackson was complaining when it received a null id ...
- serialized.id = record.id ? record.id : null;
+ if (record.id == null) {
+ delete serialized.id;
+ }
//remove the root element
Ember.merge(hash, serialized);
}
}); | 6 | 0.375 | 4 | 2 |
d526038503d59481965f6614c05ec6acb534b986 | contrib/zfs/README.md | contrib/zfs/README.md |
This is a placeholder to declare the presence and status of ZFS storage driver
for containers.
The current development is done in Gurjeet Singh's fork of Docker, under the
branch named [zfs_driver].
[zfs_driver]: https://github.com/gurjeet/docker/tree/zfs_driver
# Status
Pre-alpha
The code is under development. Contributions in the form of suggestions,
code-reviews, and patches are welcome.
Please send the communication to gurjeet@singh.im and CC at least one Docker
mailing list.
|
This is a placeholder to declare the presence and status of ZFS storage driver
for containers.
The current development is done in Gurjeet Singh's fork of Docker, under the
branch named [zfs_driver].
[zfs_driver]: https://github.com/gurjeet/docker/tree/zfs_driver
# Status
Alpha: The code is now capable of creating, running and destroying containers
and images.
The code is under development. Contributions in the form of suggestions,
code-reviews, and patches are welcome.
Please send the communication to gurjeet@singh.im and CC at least one Docker
mailing list.
| Update readme to mark ZFS driver as Alpha quality. | Update readme to mark ZFS driver as Alpha quality.
| Markdown | apache-2.0 | mrjana/docker,Ye-Yong-Chi/docker,pwaller/docker,mavenugo/docker,devmeyster/docker,shaded-enmity/docker,Gemini-Cain/docker,lyft/docker,diogomonica/docker,reem/docker,chatenilesh/docker,noqcks/docker,flythecircle/gotojourney,menglingwei/denverdino.github.io,coolsvap/docker,manchoz/docker,pramodhkp/moby,kkirsche/docker,Zenithar/docker,alena1108/docker,MarginC/docker,remh/docker,airandfingers/docker,josegonzalez/docker,bleuchtang/docker,rbarlow/docker,gdevillele/docker.github.io,sreejithr/docker,rhatdan/docker,twaugh/docker,vishh/docker,draghuram/docker,nawawi/docker,opreaadrian/docker,kojiromike/docker,hyperhq/hypercli,Ramzec/docker,yongtang/docker,beni55/docker,1HLtd/docker,pwaller/docker,dave-tucker/docker,xiaozi0lei/docker,glennblock/docker,rhatdan/storage,jvanz/docker,vincentbernat/docker,tutumcloud/docker,newrelic-forks/docker,elancom/docker,jc-m/docker,hacfi/docker,crawford/docker,lmesz/docker,ponsfrilus/docker,DoraALin/docker,youprofit/docker,ferbe/docker,l0rd/docker,nghiant2710/docker,mruse/docker,twistlock/docker,Mashimiao/docker,1HLtd/docker,scapewang/docker,marimokkori/docker,mstanleyjones/docker,dyema/docker,diogomonica/docker,alecbenson/docker,foobar2github/docker,lizf-os/docker,rhuss/docker,qzio/docker,markprokoudine/docker,ibuildthecloud/docker,ColinHebert/docker,boucher/docker,Test-Betta-Inc/psychic-barnacle,prertik/docker,andrewnguyen/docker,jvgogh/docker,mieciu/docker,ggerman/docker,uestcfyl/docker,nalind/docker,resin-io/docker,shot/docker,flowlo/docker,stainboy/docker,alex-aizman/docker,kolyshkin/docker,ctmnz/docker,gzjim/docker,airandfingers/docker,sidzan/docker,duguhaotian/docker,LoHChina/docker,dhiltgen/docker,jecarey/docker,cnbin/docker,gzjim/docker,Distrotech/docker,ffoysal/docker,raja-sami-10p/moby,xiekeyang/docker,sallyom/docker,sanscontext/docker.github.io,trebonian/docker,Kaffa-MY/docker,oguzcanoguz/docker,dmcgowan/docker,q384566678/docker,mohmost12/docker,zelahi/docker,zhangjianfnst/docker,haveatry/docker,ponsfrilus/docker,denverdino/denverdino.github.io,willmtemple/docker,jimjag/docker,beatle/docker,ClusterHQ/docker-plugins,Shopify/docker,tiw/docker,trebonian/docker,icaoweiwei/docker,Collinux/docker,robort/docker,rajurs/docker,bbox-kula/docker,arun-gupta/docker,hyperhq/hypercli,dit4c/docker,rillig/docker.github.io,mtrmac/docker,porjo/docker,yoanisgil/docker,mohmost12/docker,thaJeztah/docker,davinashreddy/docker,slok/docker,hemanthkumarsa13/test,slok/docker,thaJeztah/docker,lifeeth/docker-arm,crquan/docker,alecbenson/docker,bajh/docker,Akasurde/docker,gjaskiewicz/moby,yasiramin10p/moby,yongtang/docker,diogomonica/docker,uestcfyl/docker,affo/docker,KingEmet/docker,nabeken/docker,yummypeng/docker,ZenikaOuest/docker,nicolasmiller/docker,Yhgenomics/docker,BWITS/docker,ctrombley/docker,lsxia/docker,thieman/docker,mc2014/docker,dongjoon-hyun/docker,kiranmeduri/docker,jonahzheng/docker,scotttlin/docker,shiroyuki/docker,anchal-agrawal/docker,eskaaren/docker,cezarsa/docker,supasate/docker,rghv/docker,duguhaotian/docker,StevenLudwig/docker,seblu/docker,contiv/docker,xiaozi0lei/docker,sfsmithcha/docker,mathewkoshy493/docker,beatle/docker,darrenstahlmsft/docker,gdi2290/docker,moypray/docker,jonahzheng/docker,ideaar/docker,proppy/docker,brianfoshee/docker,nalind/docker,shin-/docker.github.io,csfrancis/docker,krisdages/docker,srust/docker,q384566678/docker,Shinzu/docker,carlanton/docker,qudongfang/docker,pugnascotia/docker,James0tang/docker,huxingbang/docker,HideoYamauchi/docker,cflannagan/docker,crquan/docker,Zenithar/docker,chavezom/docker,heartlock/docker,sidzan/docker,pugnascotia/docker,fsoppelsa/docker,JohnyDays/docker,dongjoon-hyun/docker,paulczar/docker,gdevillele/docker.github.io,matthiasr/docker,vmware/docker,0xfoo/docker,johnstep/docker.github.io,marineam/docker,hewei10/docker,phiroict/docker,rremer/moby,vlajos/docker,franklyhuan/docker-1,NERSC/docker,noironetworks/docker,liu-rui/docker,ColinHebert/docker,pragyasardana/Docker-cr-combined,rsampaio/docker,joaofnfernandes/docker.github.io,slok/docker,ai-traders/docker,wzzrd/docker,moypray/docker,VincentYu/docker,aaronlehmann/docker,gabrielhartmann/docker,tsuna/docker,krishnasrinivas/docker,inatatsu/docker,haveatry/docker,rghv/docker,dongjiaqiang/docker,coreos/docker,trebonian/docker,kimh/docker,camallen/docker,asbjornenge/docker,shot/docker,Jimmy-Xu/hypercli,Yhgenomics/docker,vdemeester/docker,skatsuta/docker,Gobella/docker,sanimej/docker,ashmo123/docker,bertyuan/docker,gs11/docker,bdwill/docker.github.io,endophage/docker,KumaranKamalanathan/docker,aidanhs/docker,timwraight/docker,spinolacastro/docker,chrisseto/docker,NIWAHideyuki/docker,bq-xiao/docker,Ninir/docker,xnox/docker,pragyasardana/Docker-cr-combined,j-stew/git_sandbox,mkumatag/docker,TrumanLing/docker,tiw/docker,pkdevbox/docker,jrjang/docker,Gobella/docker,buddhamagnet/docker,lauly/dr,joeuo/docker.github.io,tophj-ibm/docker,yosifkit/docker,wackyapples/docker,vanloswang/docker,gnawux/docker,yang1992/docker,seblu/docker,hacpai/docker,armbuild/docker,johnstep/moby-moby,chrisseto/docker,jloehel/docker,azweb76/docker,sunyuan3/docker,kingland/docker,wdxxs2z/docker,mieciu/docker,catinthesky/docker,NERSC/docker,jlebon/docker,mheon/docker,anusha-ragunathan/docker,crawford/docker,noqcks/docker,linjk/docker,mauidev/docker,mudverma/docker,YuPengZTE/docker,bkeyoumarsi/docker,vegasbrianc/docker,psquickitjayant/docker,Akasurde/docker,duglin/docker,AkihiroSuda/docker,wcwxyz/docker,ydhydhjanson/docker,johnstep/docker.github.io,d23/docker,seanmcdaniel/docker,abiosoft/docker,ZJUshuaizhou/docker,13W/docker,imeoer/hypercli,ChanderG/docker,pdevine/docker,roth1002/docker,keerthanabs1/docker,glycerine/docker,anshulr/docker,ajaymdesai/docker,roxyboy/docker,Jarema/docker,dmcgowan/docker,HackToday/docker,wagzhi/docker,mimoralea/docker,bobrik/docker,gaowenbin/docker,mbanikazemi/docker,ottok/docker,sunxiang0918/docker,chatenilesh/docker,cyphar/docker,xiaods/docker,jonathan-beard/docker,quaintcow/docker,stannie42/docker,cyli/docker,twosigma/docker-old,mattrobenolt/docker,sharifmamun/docker,kkirsche/docker,RichardScothern/docker,lixiaobing10051267/docker,DBCDK/docker,elipiwas/DockerBirthday,raychaser/docker,hypriot/docker,insequent/docker,paralin/docker,alexisbellido/docker.github.io,menglingwei/denverdino.github.io,iskradelta/docker,thaJeztah/docker.github.io,NunoEdgarGub1/docker,lizf-os/docker,boite/docker,cjlesiw/docker,phiroict/docker,gdi2290/docker,mhrivnak/docker,cezarsa/docker,x1022as/docker,shubheksha/docker.github.io,xiekeyang/docker,raja-sami-10p/moby,crawford/docker,shubheksha/docker.github.io,abudulemusa/docker,tklauser/docker,vanloswang/docker,seanmcdaniel/docker,hustcat/docker,AkihiroSuda/docker,docker-zh/docker.github.io,yasker/docker,gnailuy/docker,fancyisbest/docker,scotttlin/docker,anshulr/docker,jimmyyan/docker,beni55/docker,ygravrand/docker,hemanthkumarsa13/test,xiaojingchen/docker,himanshu0503/docker,jefby/docker,laijs/moby,jvanz/docker,mike1729/docker,kencochrane/docker,jameszhan/docker,matthewvalimaki/docker,Kaffa-MY/docker,endophage/docker,sharidas/docker,jjn2009/docker,nakuljavali/docker,haugene/docker,sanscontext/docker.github.io,paulczar/docker,mgireesh05/docker,paralin/docker,jzwlqx/denverdino.github.io,Gemini-Cain/docker,alibaba/docker,vlajos/docker,mixja/docker,mruse/docker,songinfo/docker,aneshas/docker,SFDMumbaiDockerHackathon/docker,jwuwork/docker,mc2014/docker,fmoliveira/docker,seblu/docker,msagheer/docker,anujbahuguna/docker,dongjiaqiang/docker,bbox-kula/docker,solganik/docker,sdurrheimer/docker,keerthanabs1/docker,msagheer/docker,runcom/docker,MabinGo/docker,robhaswell/docker,Collinux/docker,shliujing/docker,slavau/docker,JohnyDays/docker,ItsVeryWindy/docker,ottok/docker,gdevillele/docker.github.io,mahak/docker,wesseljt/docker,nnordrum/docker,kyle-wang/docker,sanjeo/docker,sirxun/docker,James0tang/docker,gonkulator/docker,jloehel/docker,naveed-jamil-tenpearls/moby,sanketsaurav/docker,KarthikNayak/docker,BWITS/docker,ashahab-altiscale/docker,twistlock/docker,inversion/docker,rsmitty/docker,WPH95/docker,crquan/docker,elipiwas/DockerBirthday,sanjeo/docker,vijaykilari/docker,JimGalasyn/docker.github.io,PatrickAuld/docker,YanFeng-Adam/docker,shuhuai007/docker,gabrielhartmann/docker,KimTaehee/docker,kostickm/docker,basvdlei/docker,matthiasr/docker,docker/docker.github.io,HackToday/docker,sheimi/docker,Evalle/docker,makinacorpus/docker,jwuwork/docker,d23/docker,madhanrm/docker,dcbw/docker,anortef/docker,tt/docker,hongbinz/docker,kostickm/docker,matthewvalimaki/docker,nghiant2710/test-docker,afein/docker,mikedanese/docker,rhatdan/storage,lizf-os/docker,shhsu/docker,zaxbbun/docker,dashengSun/docker,tombee/docker,itsnuwan/docker,solganik/docker,alfred-landrum/docker,hemanthkumarsa13/test,laijs/docker,swasd/baleno,squaremo/docker,justincormack/docker,snailwalker/docker,scapewang/docker,sc4599/docker,majst01/docker,ClusterHQ/docker,kirankbs/docker,raja-sami-10p/moby,nghiant2710/test-docker,kirankbs/docker,Chhunlong/docker,alfred-landrum/docker,tt/docker,aarondav/docker,seblu/docker,SvenDowideit/docker,gbarboza/docker,mhrivnak/docker,sivabizruntime/dockerfile,wdxxs2z/docker,chrismckinnel/docker,ponsfrilus/docker,justanshulsharma/docker,nicolasmiller/docker,shakamunyi/docker,vmware/docker,ferbe/docker,ffoysal/docker,SaiedKazemi/docker,zelahi/docker,erikstmartin/docker,wescravens/docker,jjn2009/docker,ktraghavendra/docker,tbronchain/docker-1,trebonian/docker,campoy/docker,vishh/docker,menglingwei/denverdino.github.io,vincentbernat/docker,gschaetz/moby,containers/storage,yuanzougaofei/docker,q384566678/docker,anusha-ragunathan/docker,erikstmartin/docker,Ramzec/docker,izaakschroeder/docker,RepmujNetsik/docker,ntrvic/docker,tombee/docker,nghiant2710/docker,beealone/docker,mukgupta/docker,stainboy/docker,wesseljt/docker,sfsmithcha/docker,ankushagarwal/docker,mountkin/docker,abiosoft/docker,jingjidejuren/docker,dwcramer/docker,YuanZhewei/docker,peterlei/docker,LaynePeng/docker,rajdeepd/docker,Jarema/docker,JohnyDays/docker,dnephin/docker,rremer/moby,gschaetz/moby,kickinz1/docker,kojiromike/docker,mountkin/docker,SophiaGitHub/hello,balajibr/docker,PatrickLang/moby,DoraALin/docker,mchasal/docker,sharmasushant/docker,maaquib/docker,dezon/docker,denverdino/docker.github.io,campoy/docker,rajdeepd/docker,endocode/docker,mariusGundersen/docker,jssnlw/docker,psquickitjayant/docker,songinfo/docker,guard163/docker,xiaojingchen/docker,robort/docker,PatrickLang/moby,manoj0077/docker,glycerine/docker,achanda/docker,balajibr/docker,moby/moby,johnstep/docker.github.io,gschaetz/moby,jagdeesh109/docker,jengeb/docker,hongbinz/docker,moby/moby,srevereault/docker,DarknessBeforeDawn/docker,BrianHicks/docker,mheon/docker,zhuguihua/docker,mimoralea/docker,waitingkuo/docker,dims/docker,omni360/docker,twistlock/docker,cezarsa/docker,KimTaehee/docker,carlanton/docker,xiaoshaozi52/docker,ctmnz/docker,HideoYamauchi/docker,marimokkori/docker,ken-saka/docker,justyntemme/docker,rogaha/docker,toophy/docker,Ramzec/docker,mgireesh05/docker,vyshaliOrg/docker,ibuildthecloud/docker,londoncalling/docker.github.io,nf/docker,POLJM242/docker,camallen/docker,hariharan16s/docker_vlan,binario200/docker,13W/docker,lloydde/docker,mieciu/docker,sheavner/docker,ncdc/docker,Microsoft/docker,cr7pt0gr4ph7/docker,zhuguihua/docker,thtanaka/docker,EricPai/docker,POLJM242/docker,rsmoorthy/docker-1,michelvocks/docker,bru7us/docker,DDCCOE/docker,michelvocks/docker,SophiaGitHub/newGitHub,shin-/docker.github.io,0xfoo/docker,endophage/docker,labkode/docker,csfrancis/docker,aneshas/docker,affo/docker,toli/docker,draghuram/docker,kwk/docker,sergeyevstifeev/docker,kvchampa/Docker,caofangkun/docker-docker,adouang/docker,aboch/docker,nicholaskh/docker,shakamunyi/docker,ajneu/docker,misterbisson/docker,jonathanperret/docker,andrewnguyen/docker,imeoer/hypercli,rancher/docker,ch3lo/docker,angryxlt/docker,hbdrawn/docker,azurezk/docker,moul/docker,clearlinux/docker,gcuisinier/docker,remh/docker,bbox-kula/docker,madhanrm/docker,dhiltgen/docker,jongiddy/docker,hariharan16s/docker_vlan,susan51531/docker,rghv/docker,justsml/docker,hewei10/docker,Mokolea/docker,vikstrous/docker,afein/docker,euank/docker,softprops/docker,moypray/docker,liu-rui/docker,morninj/docker,xyliuke/docker,xnox/docker,asridharan/docker,janghwan/docker,srust/docker,klshxsh/docker,mbanikazemi/docker,kostickm/docker,rwincewicz/docker,twosigma/docker-old,ksasi/docker,adouang/docker,alfred-landrum/docker,cc272309126/docker,AlligatorSinensis/docker,SophiaGitHub/newGitHub,darrenstahlmsft/moby,prertik/docker,alanmpitts/docker,rvesse/docker,goodwinos/docker,londoncalling/docker.github.io,iaintshine/docker,lauly/dr,allencloud/docker,aditirajagopal/docker,buddhamagnet/docker,toophy/docker,jmtd/docker,eskaaren/docker,ctrombley/docker,matthiasr/docker,alekhas/docker,gorcz/docker,fsoppelsa/docker,nicolasmiller/docker,mariusGundersen/docker,jamesodhunt/docker,gbarboza/docker,divya88ssn/docker,KingEmet/docker,jonathanperret/docker,jhorwit2/docker,dims/docker,tonymwinters/docker,dangwy/docker,dsau/docker,michael-holzheu/docker,icaroseara/docker,maxamillion/docker,TheBigBear/docker,mattrobenolt/docker,pradipd/moby,ajneu/docker,pyotr777/docker,izarov/docker,kolyshkin/moby,hyperhq/hypercli,mapuri/docker,yasker/docker,ZJU-SEL/docker,xiaoshaozi52/docker,Chhunlong/docker,tpounds/docker,timwraight/docker,maximkulkin/docker,dianping/docker,crosbymichael/docker,cjlesiw/docker,jongiddy/docker,lmesz/docker,cpuguy83/docker,1uptalent/docker,Test-Betta-Inc/psychic-barnacle,hallyn/docker,azweb76/docker,londoncalling/docker.github.io,roth1002/docker,lauly/dr,mikedougherty/docker,gorcz/docker,tagomoris/docker,sheavner/docker,anchal-agrawal/docker,OrenKishon/docker,jameseggers/docker,vmware/docker,Collinux/docker,keerthanabs1/docker,l0rd/docker,anthonydahanne/docker,lyndaoleary/docker,dccurtis/docker,wangcan2014/docker,kvasdopil/docker,nakuljavali/docker,lukaszraczylo/docker,14rcole/docker,cmehay/docker-1,304471720/docker,tombee/docker,hristozov/docker,srust/docker,mtrmac/docker,caofangkun/docker-docker,v47/docker,AlligatorSinensis/docker,wangcy6/docker,bertyuan/docker,huxingbang/docker,FollowMyDev/docker,lixiaobing1/docker,mattrobenolt/docker,mike1729/docker,jc-m/docker-medallia,matthewvalimaki/docker,hristozov/docker,ngpestelos/docker,kenken0926/docker,lsxia/docker,rentongzhang/docker,jasonamyers/docker,jwhonce/docker,smw/docker,ivansaracino/docker,portworx/docker_v1.9.0,sivabizruntime/dockerfile,alex-aizman/docker,zaxbbun/docker,mc2014/docker,manoj0077/docker,linjk/docker,ABaldwinHunter/docker,anshulr/docker,miminar/docker,quaintcow/docker,JMesser81/docker,alecbenson/docker,foobar2github/docker,whuwxl/docker,mahak/docker,sidharthgoel/docker,hemanthkumarsa13/test,gdi2290/docker,tomzhang/docker,bpradipt/docker,porjo/docker,luxas/docker,thomasfrancis1/docker,MjAbuz/docker,nickandrew/docker,sanimej/docker,rickcuri/docker,chenrui2014/docker,ndeloof/docker,youprofit/docker,JeffBellegarde/docker,MaxCy330/docker,sanimej/docker,cloudflare/docker,vishh/docker,Morsicus/docker,terminalcloud/docker,robhaswell/docker,rickyzhang82/docker,ethernetdan/docker,ngorskig/docker,icaoweiwei/docker,sterchelen/docker,boucher/docker,goodwinos/docker,tklauser/docker,BSWANG/denverdino.github.io,charleszheng44/moby,yummypeng/docker,ggerman/docker,squaremo/docker,rsmoorthy/docker-1,JeffBellegarde/docker,v47/docker,estesp/docker,AkihiroSuda/docker,coolljt0725/docker,cc272309126/docker,lyndaoleary/docker,chiting/docker,Shinzu/docker,DDCCOE/docker,mtesselH/docker,splunk/docker,kyle-wang/docker,willmtemple/docker,yummypeng/docker,kolyshkin/moby,renrujue/docker,tuxknight/docker,AdamOssenford/docker-pi,jonathan-beard/docker,jzwlqx/denverdino.github.io,rkazak/docker,moxiegirl/docker,alibaba/docker,jwuwork/docker,nathanleclaire/docker,yoshiotu/docker,torfuzx/docker,hypriot/docker,jonathan-beard/docker,Tenk42/docker,marineam/docker,dnephin/docker,DBCDK/docker,hacpai/docker,wlan0/docker,CrazyJvm/docker,BrianBland/docker,azurezk/docker,vegasbrianc/docker,mavenugo/docker,gluwaile/docker,bexelbie/docker,Lanzafame/docker,dperny/docker,ashahab-altiscale/docker,sallyom/docker,VishvajitP/docker,liaoqingwei/docker,liu-rui/docker,kerneltime/docker,andersjanmyr/docker,sallyom/docker,wesseljt/docker,bobrik/docker,achanda/docker,kenken0926/docker,afein/docker,morninj/docker,bogdangrigg/docker,tophj-ibm/docker,cr7pt0gr4ph7/docker,tklauser/docker,JimGalasyn/docker.github.io,gunjan5/docker,davinashreddy/docker,rocknrollMarc/docker,vijaykilari/docker,makinacorpus/docker,JimGalasyn/docker.github.io,dnephin/docker,stevvooe/docker,mavenugo/docker,errordeveloper/docker,sanscontext/docker.github.io,harche/docker,mieciu/docker,cyphar/docker,l0rd/docker,MaxCy330/docker,markprokoudine/docker,pgmcd/docker,jzwlqx/denverdino.github.io,PatrickLang/moby,olleolleolle/docker,ClusterHQ/docker,zoomis/docker,gjaskiewicz/moby,ivansaracino/docker,ndeloof/docker,aarondav/docker,bluebreezecf/docker,yangdongsheng/docker,HideoYamauchi/docker,NathanMcCauley/docker,NOX73/docker,CrazyJvm/docker,wenchma/docker,AdamOssenford/docker-pi,hewei10/docker,majst01/docker,thaJeztah/docker,abhinavdahiya/docker,BrianHicks/docker,dwcramer/docker,wzzrd/docker,AtwoodCSB/Docker,kunalkushwaha/docker,bobrik/docker,bru7us/docker,bonczj/moby-json-file-logger,konstruktoid/docker-1,vincentbernat/docker,erikstmartin/docker,londoncalling/docker.github.io,outcoldman/docker,gbarboza/docker,1dot75cm/docker,TrumanLing/docker,aboch/docker,asbjornenge/docker,dyema/docker,huxingbang/docker,dianping/docker,minidoll/docker,micahhausler/docker,jhorwit2/docker,lioryanko/docker,giorrrgio/docker,hypriot/docker,phiroict/docker,crosbymichael/docker,DarknessBeforeDawn/docker,fmoliveira/docker,1yvT0s/docker,slavau/docker,zhaoshengbo/docker,ailispaw/docker,Aegeaner/docker,alexisbellido/docker.github.io,geekylucas/docker,Distrotech/docker,alexisbellido/docker.github.io,bru7us/docker,aboch/docker,coolljt0725/docker,KumaranKamalanathan/docker,dgageot/docker,krisdages/docker,sarnautov/moby,sergeyevstifeev/docker,proppy/docker,spf13/docker,mike1729/docker,errordeveloper/docker,sevki/docker,wcwxyz/docker,stefanschneider/docker,tbonza/docker,NERSC/docker,RichardScothern/docker,opreaadrian/docker,WPH95/docker,CodyGuo/docker,darrenstahlmsft/moby,YHM07/docker,qilong0/docker,ntrvic/docker,pradipd/moby,cnbin/docker,paralin/docker,rentongzhang/docker,rickcuri/docker,ktraghavendra/docker,mnuessler/docker,charleszheng44/moby,icecrime/docker,justanshulsharma/docker,mohmost12/docker,glennblock/docker,tonistiigi/docker,dims/docker,beealone/docker,zoomis/docker,sharmasushant/docker,ClusterHQ/docker-plugins,KingEmet/docker,wangcy6/docker,dnascimento/docker,rickcuri/docker,toli/docker,ckeyer/docker,OnePaaS/docker,Morsicus/docker,asridharan/docker,AtwoodCSB/Docker,zhangg/docker,andyxator/docker,danyalKhaliq10p/moby,VishvajitP/docker,ivinpolosony/docker,runcom/docker,carolfh/docker,inversion/docker,jrslv/docker,rsmitty/docker,imikushin/docker,unodba/docker-1,sean-jc/docker,lixiaobing10051267/docker,ygravrand/docker,willmtemple/docker,clintonskitson/docker,ashahab-altiscale/docker,jengeb/docker,outcoldman/docker,paralin/docker,beatle/docker,nabeken/docker,Morsicus/docker,mssola/docker,pmalmgren/docker,thaJeztah/docker,naveed-jamil-tenpearls/moby,erikh/docker,micahyoung/docker-archlinuxarm,mssola/docker,vbatts/docker,shliujing/docker,aishraj/docker,erikh/docker,thockin/docker,msagheer/docker,stefanberger/docker,nawawi/docker,basvdlei/docker,Morgy93/docker,xiaods/docker,jefby/docker,VincentYu/docker,justsml/docker,unclejack/docker,thockin/docker,jingjidejuren/docker,docker/docker.github.io,rremer/moby,rogaha/docker,nguyentm83/docker,crquan/docker,srevereault/docker,OnePaaS/docker,raychaser/docker,ibuildthecloud/docker,itsNikolay/docker,mittalms/docker,danix800/docker.github.io,StevenLudwig/docker,MabinGo/docker,scapewang/docker,icaoweiwei/docker,dsheets/moby,SeMeKh/docker,rogaha/docker,resin-io/docker,dit4c/docker,bearstech/moby,dmizelle/docker,cmehay/docker-1,msabansal/docker,brianbirke/docker,wulonghui/docker,shin-/docker.github.io,vbatts/docker,basvdlei/docker,darrenstahlmsft/docker,brianbirke/docker,tpounds/docker,tangkun75/docker,dmizelle/docker,Ninir/docker,vishh/docker,wcwxyz/docker,yyljlyy/docker,misterbisson/docker,hacfi/docker,aditirajagopal/docker,deliangyang/docker,dongjiaqiang/docker,mtesselH/docker,ibuildthecloud/docker,pgmcd/docker,wenchma/docker,nf/docker,lsm5/docker,imkin/moby,jamesmarva/docker,ygravrand/docker,myPublicGit/docker,metalivedev/dockerclone,lixiaobing10051267/docker,dsau/docker,projectatomic/docker,maaquib/docker,tutumcloud/docker,flythecircle/gotojourney,troy0820/docker.github.io,cflannagan/docker,cvallance/docker,pkdevbox/docker,psquickitjayant/docker,qilong0/docker,itsnuwan/docker,ivinpolosony/docker,HackToday/docker,cnaize/docker,wdxxs2z/docker,kdomanski/docker,guard163/docker,justone/docker,allencloud/docker,azweb76/docker,thieman/docker,s7v7nislands/docker,mbanikazemi/docker,lyft/docker,renrujue/docker,dongjoon-hyun/docker,duguhaotian/docker,veggiemonk/docker,yosifkit/docker,errodr/docker,mahak/docker,TrumanLing/docker,izaakschroeder/docker,euank/docker,leroy-chen/docker,jagdeesh109/docker,arun-gupta/docker,pradipd/moby,anthonydahanne/docker,Ajunboys/docker,srevereault/docker,JeffBellegarde/docker,joaofnfernandes/docker.github.io,menglingwei/denverdino.github.io,axsh/docker,nakuljavali/docker,yoshiotu/docker,alena1108/docker,fcrisciani/docker,lifeeth/docker-arm,tomzhang/docker,cyli/docker,rancher/docker,clearlinux/docker,denverdino/denverdino.github.io,phiroict/docker,liaoqingwei/docker,solanolabs/docker,danix800/docker.github.io,AdamOssenford/docker-pi,zhangjianfnst/docker,noxiouz/docker,cwandrews/docker,arnib/docker,tangkun75/docker,haveatry/docker,kingland/docker,xuxinkun/docker,BenHall/docker,BSWANG/denverdino.github.io,mittalms/docker,danyalKhaliq10p/moby,mhrivnak/docker,rootfs/docker,SeMeKh/docker,shhsu/docker,cyli/docker,sterchelen/docker,mnuessler/docker,chenrui2014/docker,shenxiaochen/docker,Chhunlong/docker,Altiscale/docker,deliangyang/docker,macat/docker,kkirsche/docker,LalatenduMohanty/docker,devx/docker,mj340522/docker,lukaszraczylo/docker,tkopczynski/docker,Kevin2030/docker,rkazak/docker,AtwoodCSB/Docker,lyft/docker,caofangkun/docker-docker,zhaoshengbo/docker,j-mracek/docker,EricPai/docker,originye/docker,YujiOshima/docker,mukgupta/docker,carlosvquezada/docker,miminar/docker,luxas/docker,Lanzafame/docker,chenrui2014/docker,j-bennet/docker,xiaojingchen/docker,troy0820/docker.github.io,remh/docker,shiroyuki/docker,yang1992/docker,aarondav/docker,spf13/docker,orih/docker,sevki/docker,LuisBosquez/docker.github.io,shaded-enmity/docker,jlhawn/docker,yasiramin10p/moby,brahmaroutu/docker,sterchelen/docker,sidharthgoel/docker,gluwaile/docker,lordzero0000/docker,medallia/docker,ItsVeryWindy/docker,mchasal/docker,reem/docker,chavezom/docker,xyliuke/docker,hustcat/docker,mudverma/docker,Carrotzpc/docker,kiranmeduri/docker,docker-zh/docker.github.io,ngorskig/docker,michael-donat/docker,AlligatorSinensis/docker,nalind/docker,duglin/docker,davinashreddy/docker,cloudflare/docker,sunyuan3/docker,sirxun/docker,ottok/docker,gigidwan/docker,uestcfyl/docker,imkin/moby,makinacorpus/docker,maxwell92/docker,femibyte/docker,pugnascotia/docker,charleswhchan/docker,umerazad/docker,sirxun/docker,kencochrane/docker,CodyGuo/docker,cpswan/docker,mikedougherty/docker,boite/docker,abhinavdahiya/docker,medallia/docker,splunk/docker,rohitkadam19/docker,ydhydhjanson/docker,gigidwan/docker,samuelkarp/docker,rsampaio/docker,zhangg/docker,konstruktoid/docker-1,wagzhi/docker,anhzhi/docker,Zenithar/docker,Shopify/docker,ndeloof/docker,isubuz/docker,dbonev/docker,dcbw/docker,tagomoris/docker,nf/docker,cyphar/docker,merwok/docker,carlosvquezada/docker,yosifkit/docker,aishraj/docker,ABaldwinHunter/docker,HuKeping/docker,pkdevbox/docker,dyema/docker,brianfoshee/docker,rabbitcount/docker,rentongzhang/docker,Mashimiao/docker,aduermael/docker.github.io,hustcat/docker,lyndaoleary/docker,flavio/docker,justyntemme/docker,Test-Betta-Inc/psychic-barnacle,chiting/docker,stefanschneider/docker,x1022as/docker,dmizelle/docker,alibaba/docker,thtanaka/docker,raychaser/docker,rsampaio/docker,roth1002/docker,anweiss/docker.github.io,JimGalasyn/docker.github.io,stormltf/docker,ethernetdan/docker,rgl/docker,andersjanmyr/docker,jvanz/docker,shakamunyi/docker,dit4c/docker,mj340522/docker,nguyentm83/docker,Carrotzpc/docker,stainboy/docker,kawamuray/docker,cwandrews/docker,ewindisch/docker,Kevin2030/docker,rajurs/docker,SophiaGitHub/hello,bearstech/moby,denverdino/denverdino.github.io,bleuchtang/docker,1dot75cm/docker,alekhas/docker,gospo/docker,jthelin/docker,mathewkoshy493/docker,gigidwan/docker,carolfh/docker,Syntaxide/docker,ewindisch/docker,devmeyster/docker,beealone/docker,tuxknight/docker,iaintshine/docker,pramodhkp/moby,michael-donat/docker,moxiegirl/docker,orih/docker,wzzrd/docker,insequent/docker,NOX73/docker,inatatsu/docker,Ye-Yong-Chi/docker,1yvT0s/docker,amatsus/docker,flavio/docker,hvnsweeting/docker,gs11/docker,dinhxuanvu/docker,LoeAple/docker,YHM07/docker,lloydde/docker,tmp6154/docker,Tenk42/docker,konstruktoid/docker-1,abudulemusa/docker,alexisbellido/docker.github.io,pmorie/docker,angryxlt/docker,michaelhuettermann/docker,dave-tucker/docker,newrelic-forks/docker,NathanMcCauley/docker,BenHall/docker,noxiouz/docker,deletoe/docker,booyaa/docker,seanmcdaniel/docker,csfrancis/docker,cr7pt0gr4ph7/docker,mkumatag/docker,devonharvey/docker,abhinavdahiya/docker,gnailuy/docker,kvchampa/Docker,bfirsh/docker,lucasjo/docker,keyan/docker,ideaar/docker,chiting/docker,BrianBland/docker,michael-holzheu/docker,cc272309126/docker,Jarema/docker,qilong0/docker,arneluehrs/docker,labkode/docker,binario200/docker,troy0820/docker.github.io,myPublicGit/docker,dnascimento/docker,orih/docker,pdevine/docker,harche/docker,imeoer/hypercli,jacobxk/docker,newrelic-forks/docker,dperny/docker,Ninir/docker,misterbisson/docker,susan51531/docker,menglingwei/denverdino.github.io,dinhxuanvu/docker,NunoEdgarGub1/docker,draghuram/docker,1yvT0s/docker,beni55/docker,zhouxiaoxiaoxujian/docker,AtwoodCSB/Docker,LuisBosquez/docker.github.io,morninj/docker,spinolacastro/docker,dave-tucker/docker,insequent/docker,thaJeztah/docker.github.io,laijs/moby,jmtd/docker,Gemini-Cain/docker,j-bennet/docker,terminalcloud/docker,LoHChina/docker,ycaihua/docker,ngpestelos/docker,dgageot/docker,Syntaxide/docker,tangkun75/docker,1uptalent/docker,solanolabs/docker,yyljlyy/docker,whuwxl/docker,1HLtd/docker,mauidev/docker,mariusGundersen/docker,portworx/docker,nalind/docker,bluebreezecf/docker,noironetworks/docker,tonymwinters/docker,gmanfunky/docker,ashahab-altiscale/docker,bertyuan/docker,kvchampa/Docker,lsm5/docker,swasd/baleno,tonymwinters/docker,toophy/docker,joaofnfernandes/docker.github.io,VishvajitP/docker,wangcan2014/docker,Jimmy-Xu/hypercli,sanjeo/docker,michaelhuettermann/docker,danix800/docker.github.io,nicholaskh/docker,ankushagarwal/docker,bdwill/docker.github.io,gmanfunky/docker,sc4599/docker,timwraight/docker,mrjana/docker,xiekeyang/docker,abudulemusa/docker,ivinpolosony/docker,originye/docker,dianping/docker,pramodhkp/moby,hqhq/moby,vyshaliOrg/docker,E312232595/docker,xiaoshaozi52/docker,mpatlasov/docker,rogaha/docker,jvgogh/docker,coreos/docker,portworx/docker,jlhawn/docker,supasate/docker,carlye566/docker_eric,ken-saka/docker,aneshas/docker,anoshf/docker,bearstech/moby,LoeAple/docker,carlanton/docker,dcbw/docker,tkopczynski/docker,danix800/docker.github.io,pecameron/docker,affo/docker,franklyhuan/docker-1,kunalkushwaha/docker,msabansal/docker,yuchangchun1/docker,flavio/docker,lsm5/docker,SaiedKazemi/docker,SvenDowideit/docker,jhorwit2/docker,iskradelta/docker,ekuric/docker,harche/docker,madhanrm/docker,wtschreiter/docker,dyema/docker,docteurklein/docker,alekhas/docker,peterlei/docker,jongiddy/docker,gzjim/docker,4is2cbetter/docker,springning/docker,mathewkoshy493/docker,circleci/docker,jzwlqx/denverdino.github.io,bfirsh/docker,OrenKishon/docker,xuxinkun/docker,dwcramer/docker,yuchangchun1/docker,deletoe/docker,NathanMcCauley/docker,yasker/docker,jonahzheng/docker,mpatlasov/docker,wenchma/docker,ABaldwinHunter/docker,justincormack/docker,labkode/docker,foobar2github/docker,noqcks/docker,maxamillion/docker,sdurrheimer/docker,itsNikolay/docker,denverdino/docker.github.io,axsh/docker,mgood/docker,nawawi/docker,Evalle/docker,sunxiang0918/docker,terminalcloud/docker,jingjidejuren/docker,ycaihua/docker,devonharvey/docker,hbdrawn/docker,matrix-stone/docker,mgoelzer/docker,jrjang/docker,Mokolea/docker,macat/docker,BWITS/docker,ktraghavendra/docker,estesp/docker,shubheksha/docker.github.io,j-stew/git_sandbox,moby/moby,pmalmgren/docker,shin-/docker.github.io,Shopify/docker,jameseggers/docker,kerneltime/docker,micahhausler/docker,docteurklein/docker,lloydde/docker,sfsmithcha/docker,burakcakirel/docker,iaintshine/docker,anortef/docker,summershrimp/docker,mukgupta/docker,gorcz/docker,johnstep/moby-moby,ankushagarwal/docker,ChanderG/docker,mikebrow/docker,dianping/docker,airandfingers/docker,tbronchain/docker-1,lioryanko/docker,chatenilesh/docker,arnib/docker,booyaa/docker,oguzcanoguz/docker,bsmr-docker/docker,upmio/docker,chrisseto/docker,olleolleolle/docker,TheBigBear/docker,arnib/docker,carlye566/docker_eric,rocknrollMarc/docker,resin-io/docker,armbuild/docker,Tenk42/docker,tsuna/docker,ClusterHQ/docker-plugins,franklyhuan/docker-1,michael-holzheu/docker,cvallance/docker,Ye-Yong-Chi/docker,johnstep/docker.github.io,stormltf/docker,josemonsalve2/docker,LoeAple/docker,alexisbellido/docker.github.io,carlosvquezada/docker,yoanisgil/docker,cjlesiw/docker,giorrrgio/docker,gao-feng/docker,awanke/docker,tomzhang/docker,sevki/docker,kickinz1/docker,krishnasrinivas/docker,MjAbuz/docker,awanke/docker,justone/docker,moul/docker,Jimmy-Xu/hypercli,sarnautov/moby,gospo/docker,jc-m/docker,lastgeniusleft2004/docker,squaremo/docker,maaquib/docker,pyotr777/docker,unclejack/docker,awanke/docker,scotttlin/docker,vyshaliOrg/docker,solganik/docker,sarnautov/moby,burakcakirel/docker,omni360/docker,pragyasardana/Docker-cr-combined,toolchainX/docker,justincormack/docker,wagzhi/docker,geekylucas/docker,tpounds/docker,Shinzu/docker,songinfo/docker,mtesselH/docker,softprops/docker,wulonghui/docker,vlajos/docker,x1022as/docker,ashahab-altiscale/docker,jimjag/docker,charleswhchan/docker,adusia/docker,tophj-ibm/moby,mapuri/docker,brianfoshee/docker,rhardih/docker,gs11/docker,torfuzx/docker,aidanhs/docker,twaugh/docker,docker/docker,josegonzalez/docker,jameszhan/docker,euank/docker,izarov/docker,tsuna/docker,docker/docker,yang1992/docker,heartlock/docker,jordimassaguerpla/docker,catinthesky/docker,moul/docker,NIWAHideyuki/docker,rickyzhang82/docker,samsabed/docker,metalivedev/dockerclone,snailwalker/docker,1dot75cm/docker,waitingkuo/docker,jloehel/docker,thomasfrancis1/docker,itsNikolay/docker,vincentwoo/docker,xiaods/docker,danyalKhaliq10p/moby,HuKeping/docker,shhsu/docker,torfuzx/docker,burakcakirel/docker,ckeyer/docker,johnstep/moby-moby,BenHall/docker,yasiramin10p/moby,flowlo/docker,yyljlyy/docker,originye/docker,anujbahuguna/docker,RichardScothern/docker,ailispaw/docker,runcom/docker,konstruktoid/docker-upstream,docker/docker,selvik/docker,imikushin/docker,mikedougherty/docker,kyle-wang/docker,Aegeaner/docker,coreos/docker,samsabed/docker,maxamillion/docker,rhuss/docker,oguzcanoguz/docker,DBCDK/docker,thtanaka/docker,jlebon/docker,noironetworks/docker,ChanderG/docker,luxas/docker,lsm5/docker,maximkulkin/docker,docker-zh/docker.github.io,fmoliveira/docker,mchasal/docker,sanscontext/docker.github.io,metalivedev/dockerclone,kojiromike/docker,s7v7nislands/docker,Gobella/docker,LuisBosquez/docker.github.io,sanketsaurav/docker,mpatlasov/docker,chavezom/docker,mtanlee/docker,unclejack/docker,raharsha/docker,hemanthkumarsa13/test,shishir-a412ed/docker,Ajunboys/docker,micahhausler/docker,cpuguy83/docker,jimmyyan/docker,dsheets/moby,amatsus/docker,anweiss/docker.github.io,vdemeester/docker,pgmcd/docker,booyaa/docker,darrenstahlmsft/moby,yangdongsheng/docker,SvenDowideit/docker,rhatdan/docker,kwk/docker,binario200/docker,splunk/docker,wackyapples/docker,jlebon/docker,minidoll/docker,scollier/docker,Syntaxide/docker,kunalkushwaha/docker,ycaihua/docker,docker/docker.github.io,BrianBland/docker,krishnasrinivas/docker,teotihuacanada/docker,xuxinkun/docker,gonkulator/docker,aduermael/docker.github.io,bleuchtang/docker,yosifkit/docker,nigelpoulton/docker,gmanfunky/docker,josemonsalve2/docker,ajaymdesai/docker,Evalle/docker,deliangyang/docker,mixja/docker,skatsuta/docker,cmehay/docker-1,abudulemusa/docker,klshxsh/docker,zoomis/docker,mountkin/docker,qzio/docker,rohitkadam19/docker,bajh/docker,campoy/docker,shishir-a412ed/docker,sunxiang0918/docker,mtanlee/docker,rancher/docker,wtschreiter/docker,prertik/docker,dnascimento/docker,jwhonce/docker,keyan/docker,joaofnfernandes/docker.github.io,mstanleyjones/docker,majst01/docker,camallen/docker,ken-saka/docker,josemonsalve2/docker,dccurtis/docker,hbdrawn/docker,cjlesiw/docker,vijaykilari/docker,vincentwoo/docker,bdwill/docker.github.io,jwhonce/docker,Microsoft/docker,SaiedKazemi/docker,fcrisciani/docker,chenchun/docker,kerneltime/docker,micahyoung/docker-archlinuxarm,dyema/docker,cezarsa/docker,stannie42/docker,rvesse/docker,estesp/docker,ClusterHQ/docker,adouang/docker,vanloswang/docker,nigelpoulton/docker,kimh/docker,chenchun/docker,smw/docker,rootfs/docker,klshxsh/docker,aditirajagopal/docker,mrjana/docker,BrianHicks/docker,ctrombley/docker,outcoldman/docker,docker/docker.github.io,wescravens/docker,janghwan/docker,zhaoshengbo/docker,shin-/docker.github.io,lukaszraczylo/docker,charleswhchan/docker,KumaranKamalanathan/docker,Carrotzpc/docker,jordimassaguerpla/docker,ypid/docker,sharifmamun/docker,ffoysal/docker,thomasfrancis1/docker,d23/docker,erikh/docker,bsmr-docker/docker,JMesser81/docker,bexelbie/docker,kawamuray/docker,contiv/docker,hallyn/docker,lioryanko/docker,yuanzougaofei/docker,micahyoung/docker-archlinuxarm,laijs/moby,nigelpoulton/docker,whuwxl/docker,MarginC/docker,jecarey/docker,zaxbbun/docker,WPH95/docker,shenxiaochen/docker,EricPai/docker,lucasjo/docker,unodba/docker-1,rocknrollMarc/docker,summershrimp/docker,jimjag/docker,Samurais/docker,devx/docker,gdi2290/docker,sunyuan3/docker,keyan/docker,jrjang/docker,ajneu/docker,rgl/docker,hvnsweeting/docker,imikushin/docker,asbjornenge/docker,ggerman/docker,ideaar/docker,bpradipt/docker,KimTaehee/docker,thaJeztah/docker.github.io,sheavner/docker,skatsuta/docker,gospo/docker,boite/docker,femibyte/docker,containers/storage,tmp6154/docker,roxyboy/docker,veggiemonk/docker,raharsha/docker,vincentwoo/docker,hongbinz/docker,errordeveloper/docker,dhiltgen/docker,heartlock/docker,gnawux/docker,dharmit/docker,hallyn/docker,konstruktoid/docker-upstream,kiranmeduri/docker,YujiOshima/docker,thieman/docker,manchoz/docker,susan51531/docker,LuisBosquez/docker.github.io,ripcurld0/moby,ngpestelos/docker,kickinz1/docker,jthelin/docker,rohitkadam19/docker,tophj-ibm/moby,mheon/docker,VincentYu/docker,docker-zh/docker.github.io,jpopelka/docker,LaynePeng/docker,pwaller/docker,sidharthgoel/docker,michelvocks/docker,buddhamagnet/docker,samsabed/docker,kathir2903/docker,hqhq/moby,mgood/docker,bq-xiao/docker,euank/docker,boucher/docker,myPublicGit/docker,denverdino/denverdino.github.io,leroy-chen/docker,jc-m/docker,aidanhs/docker,RepmujNetsik/docker,ClusterHQ/docker,OrenKishon/docker,lastgeniusleft2004/docker,ksasi/docker,axsh/docker,bluebreezecf/docker,pecameron/docker,gcuisinier/docker,miminar/docker,jrslv/docker,dashengSun/docker,ZenikaOuest/docker,tbronchain/docker-1,Yhgenomics/docker,NERSC/docker,charleszheng44/moby,jpopelka/docker,thockin/docker,wackyapples/docker,dharmit/docker,selvik/docker,justsml/docker,304471720/docker,anweiss/docker.github.io,cnbin/docker,dsau/docker,carlye566/docker_eric,abiosoft/docker,cvallance/docker,icaroseara/docker,aaronlehmann/docker,shaded-enmity/docker,porjo/docker,rabbitcount/docker,dezon/docker,j-stew/git_sandbox,unodba/docker-1,sheimi/docker,chrismckinnel/docker,denverdino/denverdino.github.io,fancyisbest/docker,lbthomsen/docker,alanmpitts/docker,kvasdopil/docker,cnaize/docker,rajdeepd/docker,raghavtan/docker,vegasbrianc/docker,ZJUshuaizhou/docker,icecrime/docker,mtanlee/docker,SFDMumbaiDockerHackathon/docker,rhatdan/docker,j-bennet/docker,AnselZhangGit/docker,mrfuxi/docker,slavau/docker,gnawux/docker,jlhawn/docker,jameszhan/docker,stefanschneider/docker,clearlinux/docker,ch3lo/docker,docker-zh/docker.github.io,michaelhuettermann/docker,tonistiigi/docker,maxwell92/docker,devmeyster/docker,mgoelzer/docker,rillig/docker.github.io,wangcan2014/docker,wescravens/docker,SFDMumbaiDockerHackathon/docker,olleolleolle/docker,James0tang/docker,tiw/docker,Altiscale/docker,ch3lo/docker,yp-engineering/docker,mj340522/docker,geekylucas/docker,josegonzalez/docker,gaobrian/docker,4is2cbetter/docker,bobrik/docker,mruse/docker,dmcgowan/docker,MabinGo/docker,clintonskitson/docker,Microsoft/docker,justyntemme/docker,laijs/docker,0xfoo/docker,andyxator/docker,djs55/docker,bonczj/moby-json-file-logger,mixja/docker,mahak/docker,mapuri/docker,sheimi/docker,nickandrew/docker,13W/docker,sequenceiq/docker,Mokolea/docker,zhangg/docker,kvasdopil/docker,anoshf/docker,ethernetdan/docker,devonharvey/docker,13W/docker,mrfuxi/docker,TheBigBear/docker,imkin/moby,veggiemonk/docker,AtwoodCSB/Docker,DoraALin/docker,Mashimiao/docker,lifeeth/docker-arm,isubuz/docker,AnselZhangGit/docker,achanda/docker,ashmo123/docker,FollowMyDev/docker,gaobrian/docker,yp-engineering/docker,gjaskiewicz/moby,mtrmac/docker,marineam/docker,sean-jc/docker,Kevin2030/docker,YHM07/docker,wulonghui/docker,sharidas/docker,toli/docker,ai-traders/docker,ycaihua/docker,dashengSun/docker,vanpire110/docker,pyotr777/docker,ajaymdesai/docker,SophiaGitHub/hello,jssnlw/docker,adusia/docker,shakamunyi/docker,mikebrow/docker,fancyisbest/docker,merwok/docker,yp-engineering/docker,jacobxk/docker,raghavtan/docker,haugene/docker,jasonamyers/docker,tonistiigi/docker,toolchainX/docker,zmarouf/docker,LoeAple/docker,ClusterHQ/docker,ypid/docker,NunoEdgarGub1/docker,icecrime/docker,askb/docker,wlan0/docker,hqhq/moby,nicholaskh/docker,anhzhi/docker,michael-donat/docker,lsxia/docker,vdemeester/docker,bexelbie/docker,rillig/docker.github.io,icaroseara/docker,mittalms/docker,medallia/docker,dangwy/docker,linjk/docker,merwok/docker,hariharan16s/docker_vlan,gaowenbin/docker,xiaozi0lei/docker,rhatdan/storage,kolyshkin/moby,LalatenduMohanty/docker,sivabizruntime/dockerfile,LuisBosquez/docker.github.io,tiw/docker,yoshiotu/docker,vishh/docker,gdi2290/docker,lastgeniusleft2004/docker,lordzero0000/docker,jimjag/docker,E312232595/docker,OnePaaS/docker,JMesser81/docker,fangyuan930917/docker,gaowenbin/docker,marimokkori/docker,cezarsa/docker,portworx/docker_v1.9.0,rbarlow/docker,ColinHebert/docker,cjlesiw/docker,hvnsweeting/docker,sequenceiq/docker,omni360/docker,gabrielhartmann/docker,seblu/docker,KarthikNayak/docker,mgood/docker,micahyoung/docker-archlinuxarm,joaofnfernandes/docker.github.io,yuanzougaofei/docker,YuPengZTE/docker,akira-baruah/docker,noxiouz/docker,mssola/docker,softprops/docker,coolsvap/docker,scollier/docker,zmarouf/docker,inatatsu/docker,pmorie/docker,clintonskitson/docker,sharifmamun/docker,deletoe/docker,adusia/docker,circleci/docker,xnox/docker,brianbirke/docker,rwincewicz/docker,jordimassaguerpla/docker,umerazad/docker,hristozov/docker,lsm5/docker,twaugh/docker,stevvooe/docker,amatsus/docker,kdomanski/docker,shuhuai007/docker,ibuildthecloud/docker,cstroe/docker,naveed-jamil-tenpearls/moby,shliujing/docker,joeuo/docker.github.io,tkopczynski/docker,denverdino/docker.github.io,jvgogh/docker,arneluehrs/docker,iskradelta/docker,jzwlqx/denverdino.github.io,BSWANG/denverdino.github.io,errodr/docker,rootfs/docker,proppy/docker,zhouxiaoxiaoxujian/docker,Neverous/docker,shot/docker,anweiss/docker.github.io,bpradipt/docker,nghiant2710/docker,ZJUshuaizhou/docker,YanFeng-Adam/docker,troy0820/docker.github.io,divya88ssn/docker,arun-gupta/docker,phiroict/docker,jc-m/docker-medallia,dccurtis/docker,crquan/docker,majst01/docker,jamesmarva/docker,Morgy93/docker,dezon/docker,roxyboy/docker,jjn2009/docker,femibyte/docker,vanpire110/docker,djs55/docker,docteurklein/docker,ripcurld0/moby,gunjan5/docker,LaynePeng/docker,bobrik/docker,janghwan/docker,errodr/docker,ncdc/docker,maxwell92/docker,mikedanese/docker,tutumcloud/docker,rsmoorthy/docker-1,liaoqingwei/docker,anchal-agrawal/docker,bfirsh/docker,zhuohual/docker,vmware/docker,NOX73/docker,johnstep/docker.github.io,tagomoris/docker,jpopelka/docker,supasate/docker,andersjanmyr/docker,kawamuray/docker,inversion/docker,YanFeng-Adam/docker,rillig/docker.github.io,denverdino/docker.github.io,ckeyer/docker,hacfi/docker,allencloud/docker,kdomanski/docker,macat/docker,stormltf/docker,vmware/docker,londoncalling/docker.github.io,darrenstahlmsft/docker,Samurais/docker,sreejithr/docker,bdwill/docker.github.io,nawawi/docker,opreaadrian/docker,isubuz/docker,tt/docker,raghavtan/docker,anortef/docker,Ramzec/docker,fangyuan930917/docker,CrazyJvm/docker,kirankbs/docker,andyxator/docker,gnailuy/docker,bonczj/moby-json-file-logger,ekuric/docker,mikebrow/docker,MarginC/docker,gluwaile/docker,mstanleyjones/docker,aaronlehmann/docker,1uptalent/docker,zhuguihua/docker,Kaffa-MY/docker,circleci/docker,gunjan5/docker,ZJU-SEL/docker,reem/docker,akira-baruah/docker,ai-traders/docker,sanscontext/docker.github.io,mkumatag/docker,lixiaobing1/docker,dharmit/docker,tophj-ibm/docker,pdevine/docker,sequenceiq/docker,krisdages/docker,gdevillele/docker.github.io,andrewnguyen/docker,mimoralea/docker,crosbymichael/docker,porjo/docker,BSWANG/denverdino.github.io,ctmnz/docker,FollowMyDev/docker,alanmpitts/docker,LoeAple/docker,thaJeztah/docker.github.io,robort/docker,portworx/docker,msabansal/docker,flowlo/docker,projectatomic/docker,gao-feng/docker,aduermael/docker.github.io,izarov/docker,brahmaroutu/docker,himanshu0503/docker,nnordrum/docker,Neverous/docker,qudongfang/docker,justanshulsharma/docker,StevenLudwig/docker,ItsVeryWindy/docker,ewindisch/docker,kencochrane/docker,coolljt0725/docker,dianping/docker,Mokolea/docker,carolfh/docker,shishir-a412ed/docker,ycaihua/docker,giorrrgio/docker,euank/docker,sequenceiq/docker,rabbitcount/docker,kimh/docker,itsnuwan/docker,wlan0/docker,bogdangrigg/docker,lixiaobing1/docker,nnordrum/docker,cloudflare/docker,markprokoudine/docker,ZenikaOuest/docker,SeMeKh/docker,jmtd/docker,HuKeping/docker,dinhxuanvu/docker,thaJeztah/docker.github.io,flavio/docker,contiv/docker,brahmaroutu/docker,jimmyyan/docker,13W/docker,lucasjo/docker,tianon/docker,gao-feng/docker,jacobxk/docker,konstruktoid/docker-upstream,joeuo/docker.github.io,kenken0926/docker,JimGalasyn/docker.github.io,alena1108/docker,yuchangchun1/docker,Ajunboys/docker,xyliuke/docker,aduermael/docker.github.io,chenchun/docker,E312232595/docker,sergeyevstifeev/docker,tmp6154/docker,jrslv/docker,BSWANG/denverdino.github.io,smw/docker,teotihuacanada/docker,summershrimp/docker,nghiant2710/test-docker,manoj0077/docker,fsoppelsa/docker,sc4599/docker,rvesse/docker,j-mracek/docker,MjAbuz/docker,minidoll/docker,Altiscale/docker,pecameron/docker,kolyshkin/docker,stevvooe/docker,springning/docker,sdurrheimer/docker,stannie42/docker,solanolabs/docker,spinolacastro/docker,majst01/docker,yoanisgil/docker,anweiss/docker.github.io,4is2cbetter/docker,LoHChina/docker,moxiegirl/docker,endocode/docker,ncdc/docker,toolchainX/docker,devx/docker,dsheets/moby,Aegeaner/docker,YuPengZTE/docker,manchoz/docker,izaakschroeder/docker,Morgy93/docker,PatrickAuld/docker,tophj-ibm/moby,hacpai/docker,samuelkarp/docker,himanshu0503/docker,matrix-stone/docker,dangwy/docker,gdevillele/docker.github.io,gcuisinier/docker,DDCCOE/docker,ferbe/docker,laijs/docker,vikstrous/docker,proppy/docker,RepmujNetsik/docker,denverdino/docker.github.io,fangyuan930917/docker,304471720/docker,gonkulator/docker,divya88ssn/docker,KarthikNayak/docker,swasd/baleno,goodwinos/docker,Akasurde/docker,nabeken/docker,jagdeesh109/docker,djs55/docker,ntrvic/docker,wangcy6/docker,shubheksha/docker.github.io,raharsha/docker,joeuo/docker.github.io,qudongfang/docker,YujiOshima/docker,sharmasushant/docker,ydhydhjanson/docker,youprofit/docker,dgageot/docker,ypid/docker,sanketsaurav/docker,1uptalent/docker,askb/docker,chrismckinnel/docker,justone/docker,anhzhi/docker,asridharan/docker,duglin/docker,cwandrews/docker,rkazak/docker,NERSC/docker,kathir2903/docker,elancom/docker,konstruktoid/docker-upstream,jssnlw/docker,cnaize/docker,rbarlow/docker,ripcurld0/moby,tbonza/docker,yongtang/docker,shenxiaochen/docker,Samurais/docker,lbthomsen/docker,cflannagan/docker,sharidas/docker,cstroe/docker,moby/moby,lordzero0000/docker,mnuessler/docker,upmio/docker,robhaswell/docker,CodyGuo/docker,anoshf/docker,endocode/docker,cstroe/docker,micahyoung/docker-archlinuxarm,jonathanperret/docker,balajibr/docker,spf13/docker,elipiwas/DockerBirthday,AkihiroSuda/docker,guard163/docker,tianon/docker,coolsvap/docker,azurezk/docker,elancom/docker,renrujue/docker,peterlei/docker,arneluehrs/docker,glennblock/docker,maximkulkin/docker,jamesodhunt/docker,lifeeth/docker-arm,projectatomic/docker,shubheksha/docker.github.io,leroy-chen/docker,umerazad/docker,rajurs/docker,SophiaGitHub/newGitHub,sidzan/docker,catinthesky/docker,YuanZhewei/docker,containers/storage,wtschreiter/docker,AnselZhangGit/docker,porjo/docker,teotihuacanada/docker,springning/docker,snailwalker/docker,twosigma/docker-old,selvik/docker,sean-jc/docker,proppy/docker,eskaaren/docker,mieciu/docker,docker/docker.github.io,sequenceiq/docker,bkeyoumarsi/docker,14rcole/docker,cpuguy83/docker,rsmitty/docker,glycerine/docker,NIWAHideyuki/docker,nickandrew/docker,kingland/docker,matrix-stone/docker,ailispaw/docker,waitingkuo/docker,bajh/docker,cpswan/docker,rhardih/docker,cpuguy83/docker,jamesodhunt/docker,jamesmarva/docker,upmio/docker,tutumcloud/docker,medallia/docker,dbonev/docker,alex-aizman/docker,caofangkun/docker-docker,zmarouf/docker,fcrisciani/docker,lbthomsen/docker,quaintcow/docker,zhangjianfnst/docker,DarknessBeforeDawn/docker,MaxCy330/docker,kathir2903/docker,dperny/docker,mgoelzer/docker,moul/docker,bkeyoumarsi/docker,caofangkun/docker-docker,bdwill/docker.github.io,pmorie/docker,anujbahuguna/docker,mudverma/docker,abudulemusa/docker,tuxknight/docker,ivansaracino/docker,bearstech/moby,jengeb/docker,mgireesh05/docker,paulczar/docker,v47/docker,zhuohual/docker,vbatts/docker,gnailuy/docker,bq-xiao/docker,nathanleclaire/docker,stefanberger/docker,Lanzafame/docker,rwincewicz/docker,joeuo/docker.github.io,samuelkarp/docker | markdown | ## Code Before:
This is a placeholder to declare the presence and status of ZFS storage driver
for containers.
The current development is done in Gurjeet Singh's fork of Docker, under the
branch named [zfs_driver].
[zfs_driver]: https://github.com/gurjeet/docker/tree/zfs_driver
# Status
Pre-alpha
The code is under development. Contributions in the form of suggestions,
code-reviews, and patches are welcome.
Please send the communication to gurjeet@singh.im and CC at least one Docker
mailing list.
## Instruction:
Update readme to mark ZFS driver as Alpha quality.
## Code After:
This is a placeholder to declare the presence and status of ZFS storage driver
for containers.
The current development is done in Gurjeet Singh's fork of Docker, under the
branch named [zfs_driver].
[zfs_driver]: https://github.com/gurjeet/docker/tree/zfs_driver
# Status
Alpha: The code is now capable of creating, running and destroying containers
and images.
The code is under development. Contributions in the form of suggestions,
code-reviews, and patches are welcome.
Please send the communication to gurjeet@singh.im and CC at least one Docker
mailing list.
|
This is a placeholder to declare the presence and status of ZFS storage driver
for containers.
The current development is done in Gurjeet Singh's fork of Docker, under the
branch named [zfs_driver].
[zfs_driver]: https://github.com/gurjeet/docker/tree/zfs_driver
# Status
- Pre-alpha
+ Alpha: The code is now capable of creating, running and destroying containers
+ and images.
The code is under development. Contributions in the form of suggestions,
code-reviews, and patches are welcome.
Please send the communication to gurjeet@singh.im and CC at least one Docker
mailing list.
| 3 | 0.142857 | 2 | 1 |
46bee322a1ee07f33d8e75c766a029814c42747f | lib/lol/tournament_code.rb | lib/lol/tournament_code.rb | require 'lol/model'
module Lol
class TournamentCode < Lol::Model
attr_reader :id
attr_reader :provider_id
# @!attribute [r] tournament_id
# @return [Fixnum] Tournament ID
attr_reader :tournament_id
# @!attribute [r] code
# @return [String] Code
attr_reader :code
# @!attribute [r] region
# @return [String] Region
attr_reader :region
# @!attribute [r] map
# @return [String] Map
attr_reader :map
# @!attribute [r] team_size
# @return [Fixnum] Team Size
attr_reader :team_size
# @!attribute [r] spectators
# @return [String] Spectators Type
attr_reader :spectators
# @!attribute [r] pick_type
# @return [String] Pick Type
attr_reader :pick_type
# @!attribute [r] lobby_name
# @return [String] Lobby Name
attr_reader :lobby_name
# @!attribute [r] participants
# @return [Array] Summoner IDS of Allowed participants
attr_reader :participants
attr_reader :password
private
attr_writer :id, :provider_id, :tournament_id, :code, :region, :map, :team_size,
:spectators, :pick_type, :lobby_name, :password, :participants
end
end
| require 'lol/model'
module Lol
class TournamentCode < Lol::Model
attr_reader :id
attr_reader :provider_id
# @!attribute [r] tournament_id
# @return [Fixnum] Tournament ID
attr_reader :tournament_id
# @!attribute [r] code
# @return [String] Code
attr_reader :code
# @!attribute [r] region
# @return [String] Region
attr_reader :region
# @!attribute [r] map
# @return [String] Map
attr_reader :map
# @!attribute [r] team_size
# @return [Fixnum] Team Size
attr_reader :team_size
# @!attribute [r] spectators
# @return [String] Spectators Type
attr_reader :spectators
# @!attribute [r] pick_type
# @return [String] Pick Type
attr_reader :pick_type
# @!attribute [r] lobby_name
# @return [String] Lobby Name
attr_reader :lobby_name
# @!attribute [r] participants
# @return [Array] Summoner IDS of Allowed participants
attr_reader :participants
attr_reader :password
private
attr_writer :id, :provider_id, :tournament_id, :code, :region, :map, :team_size,
:spectators, :pick_type, :lobby_name, :password, :participants
def initialize options = {}
super
@participants = [] if @participants.nil?
end
end
end
| Add participants default value to TournamentCode | Add participants default value to TournamentCode
| Ruby | mit | mikamai/ruby-lol,rhuberdeau/ruby-lol | ruby | ## Code Before:
require 'lol/model'
module Lol
class TournamentCode < Lol::Model
attr_reader :id
attr_reader :provider_id
# @!attribute [r] tournament_id
# @return [Fixnum] Tournament ID
attr_reader :tournament_id
# @!attribute [r] code
# @return [String] Code
attr_reader :code
# @!attribute [r] region
# @return [String] Region
attr_reader :region
# @!attribute [r] map
# @return [String] Map
attr_reader :map
# @!attribute [r] team_size
# @return [Fixnum] Team Size
attr_reader :team_size
# @!attribute [r] spectators
# @return [String] Spectators Type
attr_reader :spectators
# @!attribute [r] pick_type
# @return [String] Pick Type
attr_reader :pick_type
# @!attribute [r] lobby_name
# @return [String] Lobby Name
attr_reader :lobby_name
# @!attribute [r] participants
# @return [Array] Summoner IDS of Allowed participants
attr_reader :participants
attr_reader :password
private
attr_writer :id, :provider_id, :tournament_id, :code, :region, :map, :team_size,
:spectators, :pick_type, :lobby_name, :password, :participants
end
end
## Instruction:
Add participants default value to TournamentCode
## Code After:
require 'lol/model'
module Lol
class TournamentCode < Lol::Model
attr_reader :id
attr_reader :provider_id
# @!attribute [r] tournament_id
# @return [Fixnum] Tournament ID
attr_reader :tournament_id
# @!attribute [r] code
# @return [String] Code
attr_reader :code
# @!attribute [r] region
# @return [String] Region
attr_reader :region
# @!attribute [r] map
# @return [String] Map
attr_reader :map
# @!attribute [r] team_size
# @return [Fixnum] Team Size
attr_reader :team_size
# @!attribute [r] spectators
# @return [String] Spectators Type
attr_reader :spectators
# @!attribute [r] pick_type
# @return [String] Pick Type
attr_reader :pick_type
# @!attribute [r] lobby_name
# @return [String] Lobby Name
attr_reader :lobby_name
# @!attribute [r] participants
# @return [Array] Summoner IDS of Allowed participants
attr_reader :participants
attr_reader :password
private
attr_writer :id, :provider_id, :tournament_id, :code, :region, :map, :team_size,
:spectators, :pick_type, :lobby_name, :password, :participants
def initialize options = {}
super
@participants = [] if @participants.nil?
end
end
end
| require 'lol/model'
module Lol
class TournamentCode < Lol::Model
attr_reader :id
attr_reader :provider_id
# @!attribute [r] tournament_id
# @return [Fixnum] Tournament ID
attr_reader :tournament_id
# @!attribute [r] code
# @return [String] Code
attr_reader :code
# @!attribute [r] region
# @return [String] Region
attr_reader :region
# @!attribute [r] map
# @return [String] Map
attr_reader :map
# @!attribute [r] team_size
# @return [Fixnum] Team Size
attr_reader :team_size
# @!attribute [r] spectators
# @return [String] Spectators Type
attr_reader :spectators
# @!attribute [r] pick_type
# @return [String] Pick Type
attr_reader :pick_type
# @!attribute [r] lobby_name
# @return [String] Lobby Name
attr_reader :lobby_name
# @!attribute [r] participants
# @return [Array] Summoner IDS of Allowed participants
attr_reader :participants
attr_reader :password
private
attr_writer :id, :provider_id, :tournament_id, :code, :region, :map, :team_size,
:spectators, :pick_type, :lobby_name, :password, :participants
+
+ def initialize options = {}
+ super
+ @participants = [] if @participants.nil?
+ end
+
end
end | 6 | 0.115385 | 6 | 0 |
1d51f80c3fce45363300e330f9c38dd6d50d70f8 | src/scss/custom.scss | src/scss/custom.scss | // Your custom CSS
@font-face {
font-family: 'Courier New';
src: url('CourierNew.ttf') format('ttf');
}
body {
font-family: 'Courier New', sans-serif !important;
}
.copyright {
color: #999;
font-size: 80%;
font-weight: normal;
}
.poster {
}
.bodycopy {
color: #666;
}
.video-title {
color: #777;
font-weight: normal;
}
// Responsive Typography
html {
font-size: 0.6rem;
}
@include media-breakpoint-up(sm) {
html {
font-size: 0.8rem;
}
}
@include media-breakpoint-up(md) {
html {
font-size: 0.9rem;
}
}
@include media-breakpoint-up(lg) {
html {
font-size: 1.0rem;
}
}
@include media-breakpoint-up(xl) {
html {
font-size: 1.1rem;
}
}
| // Your custom CSS
@font-face {
font-family: 'Courier New';
src: url('CourierNew.ttf') format('ttf');
}
body {
font-family: 'Courier New', sans-serif !important;
}
.nav-link {
font-weight: bold;
}
.copyright {
color: #000000;
font-size: 80%;
font-weight: normal;
}
.poster {
}
.bodycopy {
color: #000000;
}
.video-title {
color: #000000;
}
// Responsive Typography
html {
font-size: 0.6rem;
}
@include media-breakpoint-up(sm) {
html {
font-size: 0.8rem;
}
}
@include media-breakpoint-up(md) {
html {
font-size: 0.9rem;
}
}
@include media-breakpoint-up(lg) {
html {
font-size: 1.0rem;
}
}
@include media-breakpoint-up(xl) {
html {
font-size: 1.1rem;
}
}
| Update font weight and color. | Update font weight and color.
| SCSS | mit | quantumtom/marcflennert,quantumtom/marcflennert | scss | ## Code Before:
// Your custom CSS
@font-face {
font-family: 'Courier New';
src: url('CourierNew.ttf') format('ttf');
}
body {
font-family: 'Courier New', sans-serif !important;
}
.copyright {
color: #999;
font-size: 80%;
font-weight: normal;
}
.poster {
}
.bodycopy {
color: #666;
}
.video-title {
color: #777;
font-weight: normal;
}
// Responsive Typography
html {
font-size: 0.6rem;
}
@include media-breakpoint-up(sm) {
html {
font-size: 0.8rem;
}
}
@include media-breakpoint-up(md) {
html {
font-size: 0.9rem;
}
}
@include media-breakpoint-up(lg) {
html {
font-size: 1.0rem;
}
}
@include media-breakpoint-up(xl) {
html {
font-size: 1.1rem;
}
}
## Instruction:
Update font weight and color.
## Code After:
// Your custom CSS
@font-face {
font-family: 'Courier New';
src: url('CourierNew.ttf') format('ttf');
}
body {
font-family: 'Courier New', sans-serif !important;
}
.nav-link {
font-weight: bold;
}
.copyright {
color: #000000;
font-size: 80%;
font-weight: normal;
}
.poster {
}
.bodycopy {
color: #000000;
}
.video-title {
color: #000000;
}
// Responsive Typography
html {
font-size: 0.6rem;
}
@include media-breakpoint-up(sm) {
html {
font-size: 0.8rem;
}
}
@include media-breakpoint-up(md) {
html {
font-size: 0.9rem;
}
}
@include media-breakpoint-up(lg) {
html {
font-size: 1.0rem;
}
}
@include media-breakpoint-up(xl) {
html {
font-size: 1.1rem;
}
}
| // Your custom CSS
@font-face {
font-family: 'Courier New';
src: url('CourierNew.ttf') format('ttf');
}
body {
font-family: 'Courier New', sans-serif !important;
}
+ .nav-link {
+ font-weight: bold;
+ }
+
.copyright {
- color: #999;
+ color: #000000;
font-size: 80%;
font-weight: normal;
}
.poster {
}
.bodycopy {
- color: #666;
+ color: #000000;
}
.video-title {
+ color: #000000;
- color: #777;
- font-weight: normal;
}
// Responsive Typography
html {
font-size: 0.6rem;
}
@include media-breakpoint-up(sm) {
html {
font-size: 0.8rem;
}
}
@include media-breakpoint-up(md) {
html {
font-size: 0.9rem;
}
}
@include media-breakpoint-up(lg) {
html {
font-size: 1.0rem;
}
}
@include media-breakpoint-up(xl) {
html {
font-size: 1.1rem;
}
} | 11 | 0.189655 | 7 | 4 |
dbb65cdbc5022033f1de040cbef831a27434b74c | pillars/grains/debian/system.sls | pillars/grains/debian/system.sls | node:
source: path
| node:
archive: https://nodejs.org/dist/v5.1.0/node-v5.1.0-linux-armv7l.tar.gz
checksum: c7bbb8c51bdaf9b46daffbebdc889ce561c8f7b6cdd3c804c5e8cb7314a4d40d
| Use ARM archive for raspberrypi. | [NodeJS] Use ARM archive for raspberrypi.
| SaltStack | bsd-3-clause | stefano-pogliani/salt-states,stefano-pogliani/salt-states,stefano-pogliani/salt-states,stefano-pogliani/salt-states | saltstack | ## Code Before:
node:
source: path
## Instruction:
[NodeJS] Use ARM archive for raspberrypi.
## Code After:
node:
archive: https://nodejs.org/dist/v5.1.0/node-v5.1.0-linux-armv7l.tar.gz
checksum: c7bbb8c51bdaf9b46daffbebdc889ce561c8f7b6cdd3c804c5e8cb7314a4d40d
| node:
- source: path
+ archive: https://nodejs.org/dist/v5.1.0/node-v5.1.0-linux-armv7l.tar.gz
+ checksum: c7bbb8c51bdaf9b46daffbebdc889ce561c8f7b6cdd3c804c5e8cb7314a4d40d | 3 | 1.5 | 2 | 1 |
357449a0bdaff9e494e2a1befb252431e105a01c | build.boot | build.boot | (def project 'radicalzephyr/webfx)
(def version "0.1.0-SNAPSHOT")
(set-env! :resource-paths #{"src"}
:dependencies [])
(task-options!
pom {:project project
:version version
:description "A configurable shim for running a web application in a local JavaFX Webkit web view."
:url "https://github.com/RadicalZephyr/webfx"
:scm {:url "https://github.com/RadicalZephyr/webfx"}
:license {"Eclipse Public License"
"http://www.eclipse.org/legal/epl-v10.html"}})
(deftask build
"Build and install the project locally."
[]
(comp (javac) (pom) (jar) (install)))
| (def project 'radicalzephyr/webfx)
(def version "0.1.0-SNAPSHOT")
(set-env! :resource-paths #{"src"}
:dependencies [])
(task-options!
pom {:project project
:version version
:description "A configurable shim for running a web application in a local JavaFX Webkit web view."
:url "https://github.com/RadicalZephyr/webfx"
:scm {:url "https://github.com/RadicalZephyr/webfx"}
:license {"Eclipse Public License"
"http://www.eclipse.org/legal/epl-v10.html"}}
jar {:main 'radicalzephyr.WebFX})
(deftask build
"Build and install the project locally."
[]
(comp (javac) (pom) (jar) (install)))
| Add WebFX as the main class for the jar manifest | Add WebFX as the main class for the jar manifest
That makes it into a standalone command-line "app-ifier" that can be
configured via command line arguments.
| Clojure | epl-1.0 | RadicalZephyr/webfx | clojure | ## Code Before:
(def project 'radicalzephyr/webfx)
(def version "0.1.0-SNAPSHOT")
(set-env! :resource-paths #{"src"}
:dependencies [])
(task-options!
pom {:project project
:version version
:description "A configurable shim for running a web application in a local JavaFX Webkit web view."
:url "https://github.com/RadicalZephyr/webfx"
:scm {:url "https://github.com/RadicalZephyr/webfx"}
:license {"Eclipse Public License"
"http://www.eclipse.org/legal/epl-v10.html"}})
(deftask build
"Build and install the project locally."
[]
(comp (javac) (pom) (jar) (install)))
## Instruction:
Add WebFX as the main class for the jar manifest
That makes it into a standalone command-line "app-ifier" that can be
configured via command line arguments.
## Code After:
(def project 'radicalzephyr/webfx)
(def version "0.1.0-SNAPSHOT")
(set-env! :resource-paths #{"src"}
:dependencies [])
(task-options!
pom {:project project
:version version
:description "A configurable shim for running a web application in a local JavaFX Webkit web view."
:url "https://github.com/RadicalZephyr/webfx"
:scm {:url "https://github.com/RadicalZephyr/webfx"}
:license {"Eclipse Public License"
"http://www.eclipse.org/legal/epl-v10.html"}}
jar {:main 'radicalzephyr.WebFX})
(deftask build
"Build and install the project locally."
[]
(comp (javac) (pom) (jar) (install)))
| (def project 'radicalzephyr/webfx)
(def version "0.1.0-SNAPSHOT")
(set-env! :resource-paths #{"src"}
:dependencies [])
(task-options!
pom {:project project
:version version
:description "A configurable shim for running a web application in a local JavaFX Webkit web view."
:url "https://github.com/RadicalZephyr/webfx"
:scm {:url "https://github.com/RadicalZephyr/webfx"}
:license {"Eclipse Public License"
- "http://www.eclipse.org/legal/epl-v10.html"}})
? -
+ "http://www.eclipse.org/legal/epl-v10.html"}}
+ jar {:main 'radicalzephyr.WebFX})
(deftask build
"Build and install the project locally."
[]
(comp (javac) (pom) (jar) (install))) | 3 | 0.157895 | 2 | 1 |
881cae43c4b8643d3acdfe0bf07749939b0fb326 | tests/jsstyle.cfg | tests/jsstyle.cfg | {
"preset": "crockford",
"disallowDanglingUnderscores": null,
"requireMultipleVarDecl": null,
"requireCamelCaseOrUpperCaseIdentifiers": null,
"requireVarDeclFirst": null
}
| {
"preset": "crockford",
"disallowDanglingUnderscores": null,
"requireMultipleVarDecl": null,
"requireCamelCaseOrUpperCaseIdentifiers": null
}
| Revert added rule for jscs. | Revert added rule for jscs.
Apparently I can't make the work with both jscs 2.x and jscs 1.6, since we fail on unknown rules.
| INI | apache-2.0 | girder/girder,salamb/girder,msmolens/girder,Kitware/girder,data-exp-lab/girder,Xarthisius/girder,opadron/girder,essamjoubori/girder,RafaelPalomar/girder,essamjoubori/girder,kotfic/girder,RafaelPalomar/girder,sutartmelson/girder,jbeezley/girder,msmolens/girder,girder/girder,msmolens/girder,RafaelPalomar/girder,jbeezley/girder,adsorensen/girder,opadron/girder,girder/girder,data-exp-lab/girder,Kitware/girder,girder/girder,RafaelPalomar/girder,sutartmelson/girder,Xarthisius/girder,adsorensen/girder,opadron/girder,sutartmelson/girder,jbeezley/girder,manthey/girder,manthey/girder,manthey/girder,essamjoubori/girder,kotfic/girder,Xarthisius/girder,manthey/girder,kotfic/girder,adsorensen/girder,adsorensen/girder,msmolens/girder,salamb/girder,data-exp-lab/girder,opadron/girder,RafaelPalomar/girder,kotfic/girder,sutartmelson/girder,Kitware/girder,data-exp-lab/girder,Kitware/girder,adsorensen/girder,essamjoubori/girder,data-exp-lab/girder,Xarthisius/girder,msmolens/girder,salamb/girder,kotfic/girder,sutartmelson/girder,opadron/girder,Xarthisius/girder,jbeezley/girder,salamb/girder,essamjoubori/girder,salamb/girder | ini | ## Code Before:
{
"preset": "crockford",
"disallowDanglingUnderscores": null,
"requireMultipleVarDecl": null,
"requireCamelCaseOrUpperCaseIdentifiers": null,
"requireVarDeclFirst": null
}
## Instruction:
Revert added rule for jscs.
Apparently I can't make the work with both jscs 2.x and jscs 1.6, since we fail on unknown rules.
## Code After:
{
"preset": "crockford",
"disallowDanglingUnderscores": null,
"requireMultipleVarDecl": null,
"requireCamelCaseOrUpperCaseIdentifiers": null
}
| {
"preset": "crockford",
"disallowDanglingUnderscores": null,
"requireMultipleVarDecl": null,
- "requireCamelCaseOrUpperCaseIdentifiers": null,
? -
+ "requireCamelCaseOrUpperCaseIdentifiers": null
- "requireVarDeclFirst": null
} | 3 | 0.428571 | 1 | 2 |
dc1635c9da885491e0be5246429758554c2775f3 | gulpfile.coffee | gulpfile.coffee | gulp = require 'gulp'
gutil = require 'gulp-util'
clean = require('gulp-clean')
coffee = require 'gulp-coffee'
coffeelint = require('gulp-coffeelint')
istanbul = require 'gulp-istanbul'
mocha = require 'gulp-mocha'
gulp.task 'clean', ->
gulp.src(['lib/**/*.js'], read: false).pipe(clean())
gulp.task 'coffee', ->
gulp.src ['src/**/*.coffee']
.pipe coffee({bare: true}).on('error', gutil.log)
.pipe gulp.dest 'lib'
gulp.task 'lint', ->
gulp.src 'src/**/*.coffee'
.pipe coffeelint()
.pipe coffeelint.reporter()
gulp.task 'test', ['coffee', 'lint'], ->
gulp.src ['lib/**/*.js']
.pipe istanbul() # Covering files
.on 'finish', ->
gulp.src ['spec/**/*.spec.coffee']
.pipe mocha reporter: 'spec', compilers: 'coffee:coffee-script'
.pipe istanbul.writeReports() # Creating the reports after tests run
gulp.task 'default', ['clean', 'coffee']
gulp.task 'build', ['lint', 'clean', 'coffee', 'test']
| gulp = require 'gulp'
gutil = require 'gulp-util'
clean = require('gulp-clean')
coffee = require 'gulp-coffee'
coffeelint = require('gulp-coffeelint')
istanbul = require 'gulp-istanbul'
mocha = require 'gulp-mocha'
gulp.task 'clean', ->
gulp.src(['lib/**/*.js'], read: false).pipe(clean())
gulp.task 'coffee', ->
gulp.src ['src/**/*.coffee']
.pipe coffee({bare: true}).on('error', gutil.log)
.pipe gulp.dest 'lib'
gulp.task 'lint', ->
gulp.src 'src/**/*.coffee'
.pipe coffeelint()
.pipe coffeelint.reporter()
istanbul = require 'gulp-istanbul'
mocha = require 'gulp-mocha'
gulp.task 'test', ['coffee', 'lint'], ->
gulp.src ['lib/**/*.js']
.pipe istanbul({includeUntested: true}) # Covering files
.on 'finish', ->
gulp.src ['spec/**/*.spec.coffee']
.pipe mocha reporter: 'spec', compilers: 'coffee:coffee-script'
.pipe istanbul.writeReports() # Creating the reports after tests run
gulp.task 'default', ['clean', 'coffee']
gulp.task 'build', ['lint', 'clean', 'coffee', 'test']
| Add include untested option to istanbul | Add include untested option to istanbul
| CoffeeScript | mit | duereg/fit-parser | coffeescript | ## Code Before:
gulp = require 'gulp'
gutil = require 'gulp-util'
clean = require('gulp-clean')
coffee = require 'gulp-coffee'
coffeelint = require('gulp-coffeelint')
istanbul = require 'gulp-istanbul'
mocha = require 'gulp-mocha'
gulp.task 'clean', ->
gulp.src(['lib/**/*.js'], read: false).pipe(clean())
gulp.task 'coffee', ->
gulp.src ['src/**/*.coffee']
.pipe coffee({bare: true}).on('error', gutil.log)
.pipe gulp.dest 'lib'
gulp.task 'lint', ->
gulp.src 'src/**/*.coffee'
.pipe coffeelint()
.pipe coffeelint.reporter()
gulp.task 'test', ['coffee', 'lint'], ->
gulp.src ['lib/**/*.js']
.pipe istanbul() # Covering files
.on 'finish', ->
gulp.src ['spec/**/*.spec.coffee']
.pipe mocha reporter: 'spec', compilers: 'coffee:coffee-script'
.pipe istanbul.writeReports() # Creating the reports after tests run
gulp.task 'default', ['clean', 'coffee']
gulp.task 'build', ['lint', 'clean', 'coffee', 'test']
## Instruction:
Add include untested option to istanbul
## Code After:
gulp = require 'gulp'
gutil = require 'gulp-util'
clean = require('gulp-clean')
coffee = require 'gulp-coffee'
coffeelint = require('gulp-coffeelint')
istanbul = require 'gulp-istanbul'
mocha = require 'gulp-mocha'
gulp.task 'clean', ->
gulp.src(['lib/**/*.js'], read: false).pipe(clean())
gulp.task 'coffee', ->
gulp.src ['src/**/*.coffee']
.pipe coffee({bare: true}).on('error', gutil.log)
.pipe gulp.dest 'lib'
gulp.task 'lint', ->
gulp.src 'src/**/*.coffee'
.pipe coffeelint()
.pipe coffeelint.reporter()
istanbul = require 'gulp-istanbul'
mocha = require 'gulp-mocha'
gulp.task 'test', ['coffee', 'lint'], ->
gulp.src ['lib/**/*.js']
.pipe istanbul({includeUntested: true}) # Covering files
.on 'finish', ->
gulp.src ['spec/**/*.spec.coffee']
.pipe mocha reporter: 'spec', compilers: 'coffee:coffee-script'
.pipe istanbul.writeReports() # Creating the reports after tests run
gulp.task 'default', ['clean', 'coffee']
gulp.task 'build', ['lint', 'clean', 'coffee', 'test']
| gulp = require 'gulp'
gutil = require 'gulp-util'
clean = require('gulp-clean')
coffee = require 'gulp-coffee'
coffeelint = require('gulp-coffeelint')
istanbul = require 'gulp-istanbul'
mocha = require 'gulp-mocha'
gulp.task 'clean', ->
gulp.src(['lib/**/*.js'], read: false).pipe(clean())
gulp.task 'coffee', ->
gulp.src ['src/**/*.coffee']
.pipe coffee({bare: true}).on('error', gutil.log)
.pipe gulp.dest 'lib'
gulp.task 'lint', ->
gulp.src 'src/**/*.coffee'
.pipe coffeelint()
.pipe coffeelint.reporter()
+ istanbul = require 'gulp-istanbul'
+ mocha = require 'gulp-mocha'
+
gulp.task 'test', ['coffee', 'lint'], ->
gulp.src ['lib/**/*.js']
- .pipe istanbul() # Covering files
+ .pipe istanbul({includeUntested: true}) # Covering files
? +++++++++++++++++++++++
.on 'finish', ->
gulp.src ['spec/**/*.spec.coffee']
.pipe mocha reporter: 'spec', compilers: 'coffee:coffee-script'
.pipe istanbul.writeReports() # Creating the reports after tests run
gulp.task 'default', ['clean', 'coffee']
gulp.task 'build', ['lint', 'clean', 'coffee', 'test'] | 5 | 0.15625 | 4 | 1 |
49ee8782e838aad4c945bd5cf69726cc544de81c | README.md | README.md | SystemJS script tag loading plugin
===
Install
---
```
jspm install npm:system-script
```
Basic Use
---
```javascript
import 'www.google-analytics.com/analytics.js!system-script'
```
The example above loads google analytics via a script tag instead of a XHR request which avoid CORS issues.
### License
MIT
| SystemJS script tag loading plugin
===
Install
---
```
jspm install npm:system-script
```
Basic Use
---
```javascript
import 'www.google-analytics.com/analytics.js!system-script'
```
The example above loads google analytics via a script tag instead of a XHR request which avoid CORS issues.
SystemJS 0.17+
---
Note that as of SystemJS 0.17 the functionality in this plugin can be achieved via the following config:
```javascript
System.config({
...
meta: {
"ga": {
"scriptLoad": true,
"format": "global",
"exports": "ga"
}
},
map: {
"ga": "//www.google-analytics.com/analytics.js"
}
});
```
### License
MIT
| Add note about SystemJS 0.17 | Add note about SystemJS 0.17
| Markdown | mit | circlingthesun/systemjs-script-plugin,circlingthesun/systemjs-script-plugin | markdown | ## Code Before:
SystemJS script tag loading plugin
===
Install
---
```
jspm install npm:system-script
```
Basic Use
---
```javascript
import 'www.google-analytics.com/analytics.js!system-script'
```
The example above loads google analytics via a script tag instead of a XHR request which avoid CORS issues.
### License
MIT
## Instruction:
Add note about SystemJS 0.17
## Code After:
SystemJS script tag loading plugin
===
Install
---
```
jspm install npm:system-script
```
Basic Use
---
```javascript
import 'www.google-analytics.com/analytics.js!system-script'
```
The example above loads google analytics via a script tag instead of a XHR request which avoid CORS issues.
SystemJS 0.17+
---
Note that as of SystemJS 0.17 the functionality in this plugin can be achieved via the following config:
```javascript
System.config({
...
meta: {
"ga": {
"scriptLoad": true,
"format": "global",
"exports": "ga"
}
},
map: {
"ga": "//www.google-analytics.com/analytics.js"
}
});
```
### License
MIT
| SystemJS script tag loading plugin
===
Install
---
```
jspm install npm:system-script
```
Basic Use
---
```javascript
import 'www.google-analytics.com/analytics.js!system-script'
```
The example above loads google analytics via a script tag instead of a XHR request which avoid CORS issues.
+ SystemJS 0.17+
+ ---
+ Note that as of SystemJS 0.17 the functionality in this plugin can be achieved via the following config:
+
+ ```javascript
+ System.config({
+ ...
+ meta: {
+ "ga": {
+ "scriptLoad": true,
+ "format": "global",
+ "exports": "ga"
+ }
+ },
+ map: {
+ "ga": "//www.google-analytics.com/analytics.js"
+ }
+ });
+
+ ```
+
### License
MIT | 21 | 0.954545 | 21 | 0 |
d37555f71d61aa2f40b6d959833d7dd08bc269d4 | tmserver/jtui/__init__.py | tmserver/jtui/__init__.py | import logging
from flask import Blueprint, current_app, jsonify
from tmserver.error import register_http_error_classes
jtui = Blueprint('jtui', __name__)
logger = logging.getLogger(__name__)
def register_error(cls):
"""Decorator to register exception classes as errors that can be
serialized to JSON"""
@jtui.errorhandler(cls)
def handle_invalid_usage(error):
current_app.logger.error(error)
response = jsonify(error=error)
response.status_code = error.status_code
return response
return cls
register_http_error_classes(jtui.errorhandler)
import tmserver.jtui.api
| import logging
from flask import Blueprint
jtui = Blueprint('jtui', __name__)
logger = logging.getLogger(__name__)
import tmserver.jtui.api
| Remove jtui blueprint specific error handler | Remove jtui blueprint specific error handler
| Python | agpl-3.0 | TissueMAPS/TmServer | python | ## Code Before:
import logging
from flask import Blueprint, current_app, jsonify
from tmserver.error import register_http_error_classes
jtui = Blueprint('jtui', __name__)
logger = logging.getLogger(__name__)
def register_error(cls):
"""Decorator to register exception classes as errors that can be
serialized to JSON"""
@jtui.errorhandler(cls)
def handle_invalid_usage(error):
current_app.logger.error(error)
response = jsonify(error=error)
response.status_code = error.status_code
return response
return cls
register_http_error_classes(jtui.errorhandler)
import tmserver.jtui.api
## Instruction:
Remove jtui blueprint specific error handler
## Code After:
import logging
from flask import Blueprint
jtui = Blueprint('jtui', __name__)
logger = logging.getLogger(__name__)
import tmserver.jtui.api
| import logging
- from flask import Blueprint, current_app, jsonify
+ from flask import Blueprint
- from tmserver.error import register_http_error_classes
jtui = Blueprint('jtui', __name__)
logger = logging.getLogger(__name__)
-
- def register_error(cls):
- """Decorator to register exception classes as errors that can be
- serialized to JSON"""
- @jtui.errorhandler(cls)
- def handle_invalid_usage(error):
- current_app.logger.error(error)
- response = jsonify(error=error)
- response.status_code = error.status_code
- return response
- return cls
-
-
- register_http_error_classes(jtui.errorhandler)
-
-
import tmserver.jtui.api
| 19 | 0.678571 | 1 | 18 |
665f39450c87bc43b6b86365a74c9531476849fb | app/Http/Controllers/All/AuthController.php | app/Http/Controllers/All/AuthController.php | <?php
namespace App\Http\Controllers\All;
use App\Http\Controllers\Controller;
use App\Models\User;
use EtuUTT;
use Request;
use Config;
use View;
use Crypt;
use Redirect;
use Auth;
use Response;
class AuthController extends Controller
{
/**
* Show the authentication for newcomer page
*
* @return Response
*/
public function login()
{
return View::make('newcomer.login');
}
/**
* Authenticate the newcomer
*
* @return Response
*/
public function loginSubmit()
{
$user = User::where('login', Request::get('login'))->get()->first();
if ($user && !empty($user->password)) {
if (password_verify(Request::get('password'), $user->password)) {
Auth::login($user, true);
if ($user->isNewcomer()) {
return Redirect::route('newcomer.home')->withSuccess('Vous êtes maintenant connecté.');
}
else {
return Redirect::route('menu')->withSuccess('Vous êtes maintenant connecté.');
}
}
}
return $this->error('Identifiant ou mot de passe incorrect. Contactez integration@utt.fr si vous n\'arrivez pas à vous connecter.');
}
/**
* log out the newcomer
*
* @return Response
*/
public function logout()
{
Auth::logout();
return Redirect::route('index')->withSuccess('Vous êtes maintenant déconnecté.');
}
}
| <?php
namespace App\Http\Controllers\All;
use App\Http\Controllers\Controller;
use App\Models\User;
use EtuUTT;
use Request;
use Config;
use View;
use Crypt;
use Redirect;
use Auth;
use Response;
class AuthController extends Controller
{
/**
* Show the authentication for newcomer page
*
* @return Response
*/
public function login()
{
return View::make('newcomer.login');
}
/**
* Authenticate the newcomer
*
* @return Response
*/
public function loginSubmit()
{
$user = User::where('login', Request::get('login'))->get()->first();
if ($user && !empty($user->password)) {
if (password_verify(Request::get('password'), $user->password)) {
$user->last_login = new \DateTime();
$user->save();
Auth::login($user, true);
if ($user->isNewcomer()) {
return Redirect::route('newcomer.home')->withSuccess('Vous êtes maintenant connecté.');
}
else {
return Redirect::route('menu')->withSuccess('Vous êtes maintenant connecté.');
}
}
}
return $this->error('Identifiant ou mot de passe incorrect. Contactez integration@utt.fr si vous n\'arrivez pas à vous connecter.');
}
/**
* log out the newcomer
*
* @return Response
*/
public function logout()
{
Auth::logout();
return Redirect::route('index')->withSuccess('Vous êtes maintenant déconnecté.');
}
}
| Update last login time on newcomer auth | Update last login time on newcomer auth
| PHP | mit | ungdev/integration-UTT,nashe/integration-UTT,nashe/integration-UTT,ungdev/integration-UTT,ungdev/integration-UTT | php | ## Code Before:
<?php
namespace App\Http\Controllers\All;
use App\Http\Controllers\Controller;
use App\Models\User;
use EtuUTT;
use Request;
use Config;
use View;
use Crypt;
use Redirect;
use Auth;
use Response;
class AuthController extends Controller
{
/**
* Show the authentication for newcomer page
*
* @return Response
*/
public function login()
{
return View::make('newcomer.login');
}
/**
* Authenticate the newcomer
*
* @return Response
*/
public function loginSubmit()
{
$user = User::where('login', Request::get('login'))->get()->first();
if ($user && !empty($user->password)) {
if (password_verify(Request::get('password'), $user->password)) {
Auth::login($user, true);
if ($user->isNewcomer()) {
return Redirect::route('newcomer.home')->withSuccess('Vous êtes maintenant connecté.');
}
else {
return Redirect::route('menu')->withSuccess('Vous êtes maintenant connecté.');
}
}
}
return $this->error('Identifiant ou mot de passe incorrect. Contactez integration@utt.fr si vous n\'arrivez pas à vous connecter.');
}
/**
* log out the newcomer
*
* @return Response
*/
public function logout()
{
Auth::logout();
return Redirect::route('index')->withSuccess('Vous êtes maintenant déconnecté.');
}
}
## Instruction:
Update last login time on newcomer auth
## Code After:
<?php
namespace App\Http\Controllers\All;
use App\Http\Controllers\Controller;
use App\Models\User;
use EtuUTT;
use Request;
use Config;
use View;
use Crypt;
use Redirect;
use Auth;
use Response;
class AuthController extends Controller
{
/**
* Show the authentication for newcomer page
*
* @return Response
*/
public function login()
{
return View::make('newcomer.login');
}
/**
* Authenticate the newcomer
*
* @return Response
*/
public function loginSubmit()
{
$user = User::where('login', Request::get('login'))->get()->first();
if ($user && !empty($user->password)) {
if (password_verify(Request::get('password'), $user->password)) {
$user->last_login = new \DateTime();
$user->save();
Auth::login($user, true);
if ($user->isNewcomer()) {
return Redirect::route('newcomer.home')->withSuccess('Vous êtes maintenant connecté.');
}
else {
return Redirect::route('menu')->withSuccess('Vous êtes maintenant connecté.');
}
}
}
return $this->error('Identifiant ou mot de passe incorrect. Contactez integration@utt.fr si vous n\'arrivez pas à vous connecter.');
}
/**
* log out the newcomer
*
* @return Response
*/
public function logout()
{
Auth::logout();
return Redirect::route('index')->withSuccess('Vous êtes maintenant déconnecté.');
}
}
| <?php
namespace App\Http\Controllers\All;
use App\Http\Controllers\Controller;
use App\Models\User;
use EtuUTT;
use Request;
use Config;
use View;
use Crypt;
use Redirect;
use Auth;
use Response;
class AuthController extends Controller
{
/**
* Show the authentication for newcomer page
*
* @return Response
*/
public function login()
{
return View::make('newcomer.login');
}
/**
* Authenticate the newcomer
*
* @return Response
*/
public function loginSubmit()
{
$user = User::where('login', Request::get('login'))->get()->first();
if ($user && !empty($user->password)) {
if (password_verify(Request::get('password'), $user->password)) {
+ $user->last_login = new \DateTime();
+ $user->save();
Auth::login($user, true);
if ($user->isNewcomer()) {
return Redirect::route('newcomer.home')->withSuccess('Vous êtes maintenant connecté.');
}
else {
return Redirect::route('menu')->withSuccess('Vous êtes maintenant connecté.');
}
}
}
return $this->error('Identifiant ou mot de passe incorrect. Contactez integration@utt.fr si vous n\'arrivez pas à vous connecter.');
}
/**
* log out the newcomer
*
* @return Response
*/
public function logout()
{
Auth::logout();
return Redirect::route('index')->withSuccess('Vous êtes maintenant déconnecté.');
}
} | 2 | 0.033333 | 2 | 0 |
f2f52fb4bd11ca3f06071b4b22c377a95b5fd876 | .travis.yml | .travis.yml | language: ruby
rvm:
- 2.2
env:
- GRAPE_VERSION=0.8.0
- GRAPE_VERSION=0.9.0
- GRAPE_VERSION=0.10.0
- GRAPE_VERSION=0.10.1
- GRAPE_VERSION=0.11.0
- GRAPE_VERSION=0.12.0
- GRAPE_VERSION=0.13.0
- GRAPE_VERSION=0.14.0
- GRAPE_VERSION=HEAD
| language: ruby
rvm:
- 2.2
- 2.0
- rbx-2.5.8
- jruby-19mode
env:
- GRAPE_VERSION=0.8.0
- GRAPE_VERSION=0.9.0
- GRAPE_VERSION=0.10.0
- GRAPE_VERSION=0.10.1
- GRAPE_VERSION=0.11.0
- GRAPE_VERSION=0.12.0
- GRAPE_VERSION=0.13.0
- GRAPE_VERSION=0.14.0
- GRAPE_VERSION=HEAD
| Expand Ruby version test matrix. | Expand Ruby version test matrix.
| YAML | mit | xinminlabs/newrelic-grape | yaml | ## Code Before:
language: ruby
rvm:
- 2.2
env:
- GRAPE_VERSION=0.8.0
- GRAPE_VERSION=0.9.0
- GRAPE_VERSION=0.10.0
- GRAPE_VERSION=0.10.1
- GRAPE_VERSION=0.11.0
- GRAPE_VERSION=0.12.0
- GRAPE_VERSION=0.13.0
- GRAPE_VERSION=0.14.0
- GRAPE_VERSION=HEAD
## Instruction:
Expand Ruby version test matrix.
## Code After:
language: ruby
rvm:
- 2.2
- 2.0
- rbx-2.5.8
- jruby-19mode
env:
- GRAPE_VERSION=0.8.0
- GRAPE_VERSION=0.9.0
- GRAPE_VERSION=0.10.0
- GRAPE_VERSION=0.10.1
- GRAPE_VERSION=0.11.0
- GRAPE_VERSION=0.12.0
- GRAPE_VERSION=0.13.0
- GRAPE_VERSION=0.14.0
- GRAPE_VERSION=HEAD
| language: ruby
rvm:
- 2.2
+ - 2.0
+ - rbx-2.5.8
+ - jruby-19mode
env:
- GRAPE_VERSION=0.8.0
- GRAPE_VERSION=0.9.0
- GRAPE_VERSION=0.10.0
- GRAPE_VERSION=0.10.1
- GRAPE_VERSION=0.11.0
- GRAPE_VERSION=0.12.0
- GRAPE_VERSION=0.13.0
- GRAPE_VERSION=0.14.0
- GRAPE_VERSION=HEAD | 3 | 0.2 | 3 | 0 |
48ec7e84f1182782e762dd359558b28c22244650 | .before_travis.sh | .before_travis.sh | if [ "$ADMIN" == true ]; then
cd backend
bundle exec bin/rails s -p 3002 &
pid=$!
export RAILSPID=$pid
echo "pid - $pid"
echo "rails pid - $RAILSPID"
cd ..
sleep 5
echo "Started rails server for backend";
else
echo "Did not start backend server";
fi
if [ "$ADMIN" == true ]; then
cd admin
echo "cd into admin";
else
echo "not cd into admin";
fi
if [ "$BACKEND" == true ]; then
cd backend
echo "cd into backend";
else
echo "not cd into backend";
fi
| if [ "$ADMIN" == true ]; then
cd backend
rake db:migrate
bundle exec bin/rails s -p 3002 &
pid=$!
export RAILSPID=$pid
echo "pid - $pid"
echo "rails pid - $RAILSPID"
cd ..
sleep 5
echo "Started rails server for backend";
else
echo "Did not start backend server";
fi
if [ "$ADMIN" == true ]; then
cd admin
echo "cd into admin";
else
echo "not cd into admin";
fi
if [ "$BACKEND" == true ]; then
cd backend
echo "cd into backend";
else
echo "not cd into backend";
fi
| Migrate backend before starting it | Migrate backend before starting it
| Shell | apache-2.0 | Habitats/bitcoupon,bitcoupon/bitcoupon-bitcoin-library,bitcoupon/bitcoupon-on-rails,Habitats/bitcoupon,bitcoupon/bitcoupon-on-rails,bitcoupon/bitcoupon-on-rails,Habitats/bitcoupon,bitcoupon/bitcoupon-android-client,bitcoupon/bitcoupon-on-rails,Habitats/bitcoupon,bitcoupon/bitcoupon-bitcoin-library | shell | ## Code Before:
if [ "$ADMIN" == true ]; then
cd backend
bundle exec bin/rails s -p 3002 &
pid=$!
export RAILSPID=$pid
echo "pid - $pid"
echo "rails pid - $RAILSPID"
cd ..
sleep 5
echo "Started rails server for backend";
else
echo "Did not start backend server";
fi
if [ "$ADMIN" == true ]; then
cd admin
echo "cd into admin";
else
echo "not cd into admin";
fi
if [ "$BACKEND" == true ]; then
cd backend
echo "cd into backend";
else
echo "not cd into backend";
fi
## Instruction:
Migrate backend before starting it
## Code After:
if [ "$ADMIN" == true ]; then
cd backend
rake db:migrate
bundle exec bin/rails s -p 3002 &
pid=$!
export RAILSPID=$pid
echo "pid - $pid"
echo "rails pid - $RAILSPID"
cd ..
sleep 5
echo "Started rails server for backend";
else
echo "Did not start backend server";
fi
if [ "$ADMIN" == true ]; then
cd admin
echo "cd into admin";
else
echo "not cd into admin";
fi
if [ "$BACKEND" == true ]; then
cd backend
echo "cd into backend";
else
echo "not cd into backend";
fi
| if [ "$ADMIN" == true ]; then
cd backend
+ rake db:migrate
bundle exec bin/rails s -p 3002 &
pid=$!
export RAILSPID=$pid
echo "pid - $pid"
echo "rails pid - $RAILSPID"
cd ..
sleep 5
echo "Started rails server for backend";
else
echo "Did not start backend server";
fi
if [ "$ADMIN" == true ]; then
cd admin
echo "cd into admin";
else
echo "not cd into admin";
fi
if [ "$BACKEND" == true ]; then
cd backend
echo "cd into backend";
else
echo "not cd into backend";
fi | 1 | 0.037037 | 1 | 0 |
7c2ffa5789d676157b509b577e4a2492b7c3261b | module/Application/src/Application/Service/UserService.php | module/Application/src/Application/Service/UserService.php | <?php
namespace Application\Service;
class UserService implements \Zend\ServiceManager\ServiceLocatorAwareInterface {
use \Zend\ServiceManager\ServiceLocatorAwareTrait;
public function getCurrentUserEntity() {
$uid = $this->getServiceLocator()->get('zfcuser_auth_service')->getIdentity()->getId();
return $this->getServiceLocator()->get('UserRepository')->find((int) $uid);
}
}
| <?php
namespace Application\Service;
class UserService implements \Zend\ServiceManager\ServiceLocatorAwareInterface {
use \Zend\ServiceManager\ServiceLocatorAwareTrait;
public function getCurrentUserEntity() {
if($user = $this->getServiceLocator()->get('zfcuser_auth_service')->getIdentity()) {
return $this->getServiceLocator()->get('UserRepository')->find((int) $user->getId());
}
return null;
}
}
| Fix User Service when guest | Fix User Service when guest | PHP | mit | ee-projet-univ/zf2-blog-app,ee-projet-univ/zf2-blog-app | php | ## Code Before:
<?php
namespace Application\Service;
class UserService implements \Zend\ServiceManager\ServiceLocatorAwareInterface {
use \Zend\ServiceManager\ServiceLocatorAwareTrait;
public function getCurrentUserEntity() {
$uid = $this->getServiceLocator()->get('zfcuser_auth_service')->getIdentity()->getId();
return $this->getServiceLocator()->get('UserRepository')->find((int) $uid);
}
}
## Instruction:
Fix User Service when guest
## Code After:
<?php
namespace Application\Service;
class UserService implements \Zend\ServiceManager\ServiceLocatorAwareInterface {
use \Zend\ServiceManager\ServiceLocatorAwareTrait;
public function getCurrentUserEntity() {
if($user = $this->getServiceLocator()->get('zfcuser_auth_service')->getIdentity()) {
return $this->getServiceLocator()->get('UserRepository')->find((int) $user->getId());
}
return null;
}
}
| <?php
namespace Application\Service;
class UserService implements \Zend\ServiceManager\ServiceLocatorAwareInterface {
use \Zend\ServiceManager\ServiceLocatorAwareTrait;
public function getCurrentUserEntity() {
- $uid = $this->getServiceLocator()->get('zfcuser_auth_service')->getIdentity()->getId();
? ^^ -------- ^
+ if($user = $this->getServiceLocator()->get('zfcuser_auth_service')->getIdentity()) {
? +++ ^^^ ^^
- return $this->getServiceLocator()->get('UserRepository')->find((int) $uid);
? ^
+ return $this->getServiceLocator()->get('UserRepository')->find((int) $user->getId());
? ++++ ^^^^^^^^^ ++
+ }
+ return null;
}
} | 6 | 0.461538 | 4 | 2 |
f32898ebdf0b4e37cc169cd515e422c6cc6b051e | README.md | README.md |
A tool which can diff C# code semantically, rather than simply by the program's text.
This project aims to make developers aware of when they can do the following:
1. Cause a merge conflict when their code doesn't change any logic in the program
2. Merge in code which was written under different semantics
The goal of this project is to ultimately save significant time for large projects.
## Contribution Policy
1. Create an Issue describing the work you want to do.
2. Fork the repo, create a branch, and work on that branch.
3. Make a pull request describing what you did and reference your issue from (1)
4. Discuss the code and have it merged in!
|
A tool which diffs C# code semantically, rather than simply by the program's text in order to predict possible merge conflicts.
This project aims to make developers aware of relevant changes made in other pull requests, such as:
* The base class of an edited method being changed.
* A method being moved and changed both locally and in a pull request
The goal of this project is to ultimately save significant time for large projects.
## Requirements
* Only C# projects are supported
* Requires a Visual Studio Version that supports Roslyn Diagnostic Analyzers
* SemDiff was developed and tested on Visual Studio 2015
* Git/GitHub must be used for source control and project hosting
* The GitHub repo that contains pull requests must be the first in the config file.
* For example, if you fork the repository before you begin working. Add an "upstream" remote before the "origin". Otherwise SemDiff will search the forked repository for pull requests.
## Installation
SemDiff is available on the [NuGet Gallery](https://www.nuget.org/packages/SemDiff)
```
Install-Package SemDiff
```
Consult the [Configuration Guide](https://github.com/semdiffdotnet/semdiff/wiki/Configuration) to learn how to add GitHub credentials to the config file.
## Contribution Policy
1. Create an Issue describing the work you want to do.
2. Fork the repo, create a branch, and work on that branch.
3. Make a pull request describing what you did and reference your issue from (1)
4. Discuss the code and have it merged in!
| Add Requirements and Installation Sections | Add Requirements and Installation Sections | Markdown | mit | semdiffdotnet/semdiff | markdown | ## Code Before:
A tool which can diff C# code semantically, rather than simply by the program's text.
This project aims to make developers aware of when they can do the following:
1. Cause a merge conflict when their code doesn't change any logic in the program
2. Merge in code which was written under different semantics
The goal of this project is to ultimately save significant time for large projects.
## Contribution Policy
1. Create an Issue describing the work you want to do.
2. Fork the repo, create a branch, and work on that branch.
3. Make a pull request describing what you did and reference your issue from (1)
4. Discuss the code and have it merged in!
## Instruction:
Add Requirements and Installation Sections
## Code After:
A tool which diffs C# code semantically, rather than simply by the program's text in order to predict possible merge conflicts.
This project aims to make developers aware of relevant changes made in other pull requests, such as:
* The base class of an edited method being changed.
* A method being moved and changed both locally and in a pull request
The goal of this project is to ultimately save significant time for large projects.
## Requirements
* Only C# projects are supported
* Requires a Visual Studio Version that supports Roslyn Diagnostic Analyzers
* SemDiff was developed and tested on Visual Studio 2015
* Git/GitHub must be used for source control and project hosting
* The GitHub repo that contains pull requests must be the first in the config file.
* For example, if you fork the repository before you begin working. Add an "upstream" remote before the "origin". Otherwise SemDiff will search the forked repository for pull requests.
## Installation
SemDiff is available on the [NuGet Gallery](https://www.nuget.org/packages/SemDiff)
```
Install-Package SemDiff
```
Consult the [Configuration Guide](https://github.com/semdiffdotnet/semdiff/wiki/Configuration) to learn how to add GitHub credentials to the config file.
## Contribution Policy
1. Create an Issue describing the work you want to do.
2. Fork the repo, create a branch, and work on that branch.
3. Make a pull request describing what you did and reference your issue from (1)
4. Discuss the code and have it merged in!
|
- A tool which can diff C# code semantically, rather than simply by the program's text.
? ----
+ A tool which diffs C# code semantically, rather than simply by the program's text in order to predict possible merge conflicts.
? + +++++++++++++++++++++++++++++++++++++++++++++
- This project aims to make developers aware of when they can do the following:
+ This project aims to make developers aware of relevant changes made in other pull requests, such as:
- 1. Cause a merge conflict when their code doesn't change any logic in the program
- 2. Merge in code which was written under different semantics
+ * The base class of an edited method being changed.
+ * A method being moved and changed both locally and in a pull request
The goal of this project is to ultimately save significant time for large projects.
+
+ ## Requirements
+
+ * Only C# projects are supported
+ * Requires a Visual Studio Version that supports Roslyn Diagnostic Analyzers
+ * SemDiff was developed and tested on Visual Studio 2015
+ * Git/GitHub must be used for source control and project hosting
+ * The GitHub repo that contains pull requests must be the first in the config file.
+ * For example, if you fork the repository before you begin working. Add an "upstream" remote before the "origin". Otherwise SemDiff will search the forked repository for pull requests.
+
+ ## Installation
+
+ SemDiff is available on the [NuGet Gallery](https://www.nuget.org/packages/SemDiff)
+
+ ```
+ Install-Package SemDiff
+ ```
+
+ Consult the [Configuration Guide](https://github.com/semdiffdotnet/semdiff/wiki/Configuration) to learn how to add GitHub credentials to the config file.
+
## Contribution Policy
1. Create an Issue describing the work you want to do.
2. Fork the repo, create a branch, and work on that branch.
3. Make a pull request describing what you did and reference your issue from (1)
4. Discuss the code and have it merged in! | 28 | 1.75 | 24 | 4 |
8ce509a2a3df1fcc38442e35e8733bf00bd788ad | harvest/decorators.py | harvest/decorators.py | import os
from functools import wraps
from argparse import ArgumentParser
from fabric.context_managers import prefix
def cached_property(func):
cach_attr = '_{0}'.format(func.__name__)
@property
def wrap(self):
if not hasattr(self, cach_attr):
value = func(self)
if value is not None:
setattr(self, cach_attr, value)
return getattr(self, cach_attr, None)
return wrap
def cli(*args, **kwargs):
def decorator(func):
class Parser(ArgumentParser):
def handle(self, *args, **kwargs):
try:
func(*args, **kwargs)
except Exception, e:
self.error(e.message)
# No catching of exceptions
def handle_raw(self, *args, **kwargs):
func(*args, **kwargs)
return Parser(*args, **kwargs)
return decorator
def virtualenv(path):
"Wraps a function and prefixes the call with the virtualenv active."
if path is None:
activate = None
else:
activate = os.path.join(path, 'bin/activate')
def decorator(func):
@wraps(func)
def inner(*args, **kwargs):
if path is not None:
with prefix('source {0}'.format(activate)):
func(*args, **kwargs)
else:
func(*args, **kwargs)
return inner
return decorator
| import os
from functools import wraps
from argparse import ArgumentParser
from fabric.context_managers import prefix
def cached_property(func):
cach_attr = '_{0}'.format(func.__name__)
@property
def wrap(self):
if not hasattr(self, cach_attr):
value = func(self)
if value is not None:
setattr(self, cach_attr, value)
return getattr(self, cach_attr, None)
return wrap
def cli(*args, **kwargs):
def decorator(func):
class Parser(ArgumentParser):
def handle(self, *args, **kwargs):
try:
func(*args, **kwargs)
except Exception, e:
self.error(e.message)
# No catching of exceptions
def handle_raw(self, *args, **kwargs):
func(*args, **kwargs)
return Parser(*args, **kwargs)
return decorator
def virtualenv(path):
"Wraps a function and prefixes the call with the virtualenv active."
if path is None:
activate = None
else:
activate = os.path.join(path, 'bin/activate')
def decorator(func):
@wraps(func)
def inner(*args, **kwargs):
if path is not None:
with prefix('source {0}'.format(activate)):
return func(*args, **kwargs)
else:
return func(*args, **kwargs)
return inner
return decorator
| Return resulting object from decorated function if any | Return resulting object from decorated function if any
| Python | bsd-2-clause | chop-dbhi/harvest,tjrivera/harvest | python | ## Code Before:
import os
from functools import wraps
from argparse import ArgumentParser
from fabric.context_managers import prefix
def cached_property(func):
cach_attr = '_{0}'.format(func.__name__)
@property
def wrap(self):
if not hasattr(self, cach_attr):
value = func(self)
if value is not None:
setattr(self, cach_attr, value)
return getattr(self, cach_attr, None)
return wrap
def cli(*args, **kwargs):
def decorator(func):
class Parser(ArgumentParser):
def handle(self, *args, **kwargs):
try:
func(*args, **kwargs)
except Exception, e:
self.error(e.message)
# No catching of exceptions
def handle_raw(self, *args, **kwargs):
func(*args, **kwargs)
return Parser(*args, **kwargs)
return decorator
def virtualenv(path):
"Wraps a function and prefixes the call with the virtualenv active."
if path is None:
activate = None
else:
activate = os.path.join(path, 'bin/activate')
def decorator(func):
@wraps(func)
def inner(*args, **kwargs):
if path is not None:
with prefix('source {0}'.format(activate)):
func(*args, **kwargs)
else:
func(*args, **kwargs)
return inner
return decorator
## Instruction:
Return resulting object from decorated function if any
## Code After:
import os
from functools import wraps
from argparse import ArgumentParser
from fabric.context_managers import prefix
def cached_property(func):
cach_attr = '_{0}'.format(func.__name__)
@property
def wrap(self):
if not hasattr(self, cach_attr):
value = func(self)
if value is not None:
setattr(self, cach_attr, value)
return getattr(self, cach_attr, None)
return wrap
def cli(*args, **kwargs):
def decorator(func):
class Parser(ArgumentParser):
def handle(self, *args, **kwargs):
try:
func(*args, **kwargs)
except Exception, e:
self.error(e.message)
# No catching of exceptions
def handle_raw(self, *args, **kwargs):
func(*args, **kwargs)
return Parser(*args, **kwargs)
return decorator
def virtualenv(path):
"Wraps a function and prefixes the call with the virtualenv active."
if path is None:
activate = None
else:
activate = os.path.join(path, 'bin/activate')
def decorator(func):
@wraps(func)
def inner(*args, **kwargs):
if path is not None:
with prefix('source {0}'.format(activate)):
return func(*args, **kwargs)
else:
return func(*args, **kwargs)
return inner
return decorator
| import os
from functools import wraps
from argparse import ArgumentParser
from fabric.context_managers import prefix
def cached_property(func):
cach_attr = '_{0}'.format(func.__name__)
@property
def wrap(self):
if not hasattr(self, cach_attr):
value = func(self)
if value is not None:
setattr(self, cach_attr, value)
return getattr(self, cach_attr, None)
return wrap
def cli(*args, **kwargs):
def decorator(func):
class Parser(ArgumentParser):
def handle(self, *args, **kwargs):
try:
func(*args, **kwargs)
except Exception, e:
self.error(e.message)
# No catching of exceptions
def handle_raw(self, *args, **kwargs):
func(*args, **kwargs)
return Parser(*args, **kwargs)
return decorator
def virtualenv(path):
"Wraps a function and prefixes the call with the virtualenv active."
if path is None:
activate = None
else:
activate = os.path.join(path, 'bin/activate')
def decorator(func):
@wraps(func)
def inner(*args, **kwargs):
if path is not None:
with prefix('source {0}'.format(activate)):
- func(*args, **kwargs)
+ return func(*args, **kwargs)
? +++++++
else:
- func(*args, **kwargs)
+ return func(*args, **kwargs)
? +++++++
return inner
return decorator | 4 | 0.075472 | 2 | 2 |
72604e3e6d1fcb08b090275f594298b2efa3126e | lib/shaper.rb | lib/shaper.rb | require 'active_support/all'
require 'shaper/base'
require 'shaper/property_shaper'
require 'shaper/association_shaper'
require 'shaper/data_visitor'
| require 'active_support/all'
require 'shaper/base'
require 'shaper/property_shaper'
require 'shaper/association_shaper'
require 'shaper/data_visitor'
require 'shaper/renderers'
module Shaper
extend ActiveSupport::Concern
include Shaper::Base
include Shaper::Renderers
end
| Make it easier to include. | Make it easier to include.
| Ruby | mit | robincurry/shape | ruby | ## Code Before:
require 'active_support/all'
require 'shaper/base'
require 'shaper/property_shaper'
require 'shaper/association_shaper'
require 'shaper/data_visitor'
## Instruction:
Make it easier to include.
## Code After:
require 'active_support/all'
require 'shaper/base'
require 'shaper/property_shaper'
require 'shaper/association_shaper'
require 'shaper/data_visitor'
require 'shaper/renderers'
module Shaper
extend ActiveSupport::Concern
include Shaper::Base
include Shaper::Renderers
end
| require 'active_support/all'
require 'shaper/base'
require 'shaper/property_shaper'
require 'shaper/association_shaper'
require 'shaper/data_visitor'
+ require 'shaper/renderers'
+
+ module Shaper
+ extend ActiveSupport::Concern
+ include Shaper::Base
+ include Shaper::Renderers
+ end | 7 | 1.4 | 7 | 0 |
d655b0b9e6af1267e726ec659404dadc23cdcac6 | .eslintrc.js | .eslintrc.js | module.exports = {
env: {
browser: true,
es6: true,
},
extends: [
'airbnb',
'plugin:import/errors',
'plugin:import/warnings'
],
globals: {
Atomics: 'readonly',
SharedArrayBuffer: 'readonly',
PRODUCTION: 'readonly',
SSR: 'readonly',
jest: 'readonly',
describe: 'readonly',
beforeEach: 'readonly',
it: 'readonly',
expect: 'readonly',
},
parserOptions: {
ecmaFeatures: {
jsx: true,
},
ecmaVersion: 2018,
sourceType: 'module',
},
plugins: [
'react',
],
rules: {
'class-methods-use-this': 'off',
'import/namespace': ['on', { allowComputed: true }],
'import/no-duplicates': 'off',
'import/prefer-default-export': 'off',
'eact/jsx-filename-extension': 'off',
'no-underscore-dangle': 'off',
'react/jsx-filename-extension': 'off',
'react/prefer-stateless-function': 'off',
'no-use-before-define': 'off',
},
settings: {
'import/resolver': {
webpack: {
config: './config/webpack/webpack.config.client.babel.js'
}
}
}
};
| module.exports = {
env: {
browser: true,
es6: true,
},
extends: [
'airbnb',
'plugin:import/errors',
'plugin:import/warnings'
],
globals: {
Atomics: 'readonly',
SharedArrayBuffer: 'readonly',
PRODUCTION: 'readonly',
SSR: 'readonly',
jest: 'readonly',
describe: 'readonly',
beforeEach: 'readonly',
it: 'readonly',
expect: 'readonly',
process: 'readonly',
},
parser: 'babel-eslint',
parserOptions: {
ecmaFeatures: {
jsx: true,
},
ecmaVersion: 2018,
sourceType: 'module',
},
plugins: [
'react',
'babel',
],
rules: {
'class-methods-use-this': 'off',
'import/namespace': ['on', { allowComputed: true }],
'import/no-duplicates': 'off',
'import/prefer-default-export': 'off',
'eact/jsx-filename-extension': 'off',
'no-underscore-dangle': 'off',
'react/jsx-filename-extension': 'off',
'react/prefer-stateless-function': 'off',
'no-use-before-define': 'off',
},
settings: {
'import/resolver': {
webpack: {
config: './config/webpack/webpack.config.client.babel.js'
}
}
}
};
| Update eslint with babel-eslint parser | Update eslint with babel-eslint parser
| JavaScript | mit | dashukin/react-redux-template,dashukin/react-redux-template | javascript | ## Code Before:
module.exports = {
env: {
browser: true,
es6: true,
},
extends: [
'airbnb',
'plugin:import/errors',
'plugin:import/warnings'
],
globals: {
Atomics: 'readonly',
SharedArrayBuffer: 'readonly',
PRODUCTION: 'readonly',
SSR: 'readonly',
jest: 'readonly',
describe: 'readonly',
beforeEach: 'readonly',
it: 'readonly',
expect: 'readonly',
},
parserOptions: {
ecmaFeatures: {
jsx: true,
},
ecmaVersion: 2018,
sourceType: 'module',
},
plugins: [
'react',
],
rules: {
'class-methods-use-this': 'off',
'import/namespace': ['on', { allowComputed: true }],
'import/no-duplicates': 'off',
'import/prefer-default-export': 'off',
'eact/jsx-filename-extension': 'off',
'no-underscore-dangle': 'off',
'react/jsx-filename-extension': 'off',
'react/prefer-stateless-function': 'off',
'no-use-before-define': 'off',
},
settings: {
'import/resolver': {
webpack: {
config: './config/webpack/webpack.config.client.babel.js'
}
}
}
};
## Instruction:
Update eslint with babel-eslint parser
## Code After:
module.exports = {
env: {
browser: true,
es6: true,
},
extends: [
'airbnb',
'plugin:import/errors',
'plugin:import/warnings'
],
globals: {
Atomics: 'readonly',
SharedArrayBuffer: 'readonly',
PRODUCTION: 'readonly',
SSR: 'readonly',
jest: 'readonly',
describe: 'readonly',
beforeEach: 'readonly',
it: 'readonly',
expect: 'readonly',
process: 'readonly',
},
parser: 'babel-eslint',
parserOptions: {
ecmaFeatures: {
jsx: true,
},
ecmaVersion: 2018,
sourceType: 'module',
},
plugins: [
'react',
'babel',
],
rules: {
'class-methods-use-this': 'off',
'import/namespace': ['on', { allowComputed: true }],
'import/no-duplicates': 'off',
'import/prefer-default-export': 'off',
'eact/jsx-filename-extension': 'off',
'no-underscore-dangle': 'off',
'react/jsx-filename-extension': 'off',
'react/prefer-stateless-function': 'off',
'no-use-before-define': 'off',
},
settings: {
'import/resolver': {
webpack: {
config: './config/webpack/webpack.config.client.babel.js'
}
}
}
};
| module.exports = {
env: {
browser: true,
es6: true,
},
extends: [
'airbnb',
'plugin:import/errors',
'plugin:import/warnings'
],
globals: {
Atomics: 'readonly',
SharedArrayBuffer: 'readonly',
PRODUCTION: 'readonly',
SSR: 'readonly',
jest: 'readonly',
describe: 'readonly',
beforeEach: 'readonly',
it: 'readonly',
expect: 'readonly',
+ process: 'readonly',
},
+ parser: 'babel-eslint',
parserOptions: {
ecmaFeatures: {
jsx: true,
},
ecmaVersion: 2018,
sourceType: 'module',
},
plugins: [
'react',
+ 'babel',
],
rules: {
'class-methods-use-this': 'off',
'import/namespace': ['on', { allowComputed: true }],
'import/no-duplicates': 'off',
'import/prefer-default-export': 'off',
'eact/jsx-filename-extension': 'off',
'no-underscore-dangle': 'off',
'react/jsx-filename-extension': 'off',
'react/prefer-stateless-function': 'off',
'no-use-before-define': 'off',
},
settings: {
'import/resolver': {
webpack: {
config: './config/webpack/webpack.config.client.babel.js'
}
}
}
}; | 3 | 0.06 | 3 | 0 |
f4f9bfeeeeca960347481fd2a8ac038147da68d8 | app/views/permit_steps/_screener_question.html.erb | app/views/permit_steps/_screener_question.html.erb | <% if question %>
<label for="test-label"><%= question %></label>
<% end %>
<% if @permit.errors[attribute].length >= 1 %>
<!-- create an errors div to style error message and make input red -->
<div class="field_with_errors">
<!-- Print out the error message (from validator) -->
<% permit.errors[attribute].each do | msg | %>
<p><%= msg %></p>
<% end %>
</div>
<% end %>
<% option.each do | op | %>
<div class="radio">
<%= f.label attribute, op[:value], value: op[:value] do %>
<%= f.radio_button attribute, op[:value], id: op[:value] %>
<%= op[:label] %>
<% end %>
</div>
<% end %>
| <% if question %>
<label for="test-label"><%= question %></label>
<% end %>
<% if @permit.errors[attribute].length >= 1 %>
<!-- create an errors div to style error message and make input red -->
<div class="field_with_errors">
<!-- Print out the error message (from validator) -->
<% permit.errors[attribute].each do | msg | %>
<p><%= msg %></p>
<% end %>
</div>
<% end %>
<% option.each do | op | %>
<div class="radio">
<%= f.radio_button attribute, op[:value] %>
<%= f.label attribute, op[:label], value: op[:value] %>
</div>
<% end %>
| Fix radio label issues, but screwed up the position of it | Fix radio label issues, but screwed up the position of it
| HTML+ERB | isc | schlos/homebase,schlos/homebase,randy-r-masters/homebase,codeforamerica/homebase,randy-r-masters/homebase,randy-r-masters/homebase,codeforamerica/homebase,codeforamerica/homebase,schlos/homebase | html+erb | ## Code Before:
<% if question %>
<label for="test-label"><%= question %></label>
<% end %>
<% if @permit.errors[attribute].length >= 1 %>
<!-- create an errors div to style error message and make input red -->
<div class="field_with_errors">
<!-- Print out the error message (from validator) -->
<% permit.errors[attribute].each do | msg | %>
<p><%= msg %></p>
<% end %>
</div>
<% end %>
<% option.each do | op | %>
<div class="radio">
<%= f.label attribute, op[:value], value: op[:value] do %>
<%= f.radio_button attribute, op[:value], id: op[:value] %>
<%= op[:label] %>
<% end %>
</div>
<% end %>
## Instruction:
Fix radio label issues, but screwed up the position of it
## Code After:
<% if question %>
<label for="test-label"><%= question %></label>
<% end %>
<% if @permit.errors[attribute].length >= 1 %>
<!-- create an errors div to style error message and make input red -->
<div class="field_with_errors">
<!-- Print out the error message (from validator) -->
<% permit.errors[attribute].each do | msg | %>
<p><%= msg %></p>
<% end %>
</div>
<% end %>
<% option.each do | op | %>
<div class="radio">
<%= f.radio_button attribute, op[:value] %>
<%= f.label attribute, op[:label], value: op[:value] %>
</div>
<% end %>
| <% if question %>
<label for="test-label"><%= question %></label>
<% end %>
<% if @permit.errors[attribute].length >= 1 %>
<!-- create an errors div to style error message and make input red -->
<div class="field_with_errors">
<!-- Print out the error message (from validator) -->
<% permit.errors[attribute].each do | msg | %>
<p><%= msg %></p>
<% end %>
</div>
<% end %>
<% option.each do | op | %>
<div class="radio">
+ <%= f.radio_button attribute, op[:value] %>
- <%= f.label attribute, op[:value], value: op[:value] do %>
? ^ -- ---
+ <%= f.label attribute, op[:label], value: op[:value] %>
? ^ ++
- <%= f.radio_button attribute, op[:value], id: op[:value] %>
- <%= op[:label] %>
- <% end %>
</div>
<% end %> | 6 | 0.26087 | 2 | 4 |
4cc6a1d63d35aa247d7def6dbf883a6a2c4d7044 | Examples/README.md | Examples/README.md | The apps in this folder are designed to provide examples of the basic workings of the Google Client Library. They should not be used in production.
Eg you should not ask for a Client key as it should be embedded within the codebase of your app.
## Installation instructons
In order to use the example projects in this folder, you will need to ensure that you have downloaded the below two Git Submodules:
- gtm-session-fetcher
- AppAuth
- GTMAppAuth
## How to download Git Submodules
You can download Git Submodules using the following command:
[Credit: Stackoverflow Question](http://stackoverflow.com/questions/8090761/pull-using-git-including-submodule)
```
git clone git://url
cd repo
git submodule init
git submodule update
```
Then, add another step after the git pull.
```
git pull ...
git submodule update --recursive
```
## Requirements to use
These example apps need an oAuth2 Client key, which you can obtain from the [Google Developer Console](https://console.developers.google.com/).
### To obtain an oAuth2 client Key:
1. Visit the [Google Developer Console](https://console.developers.google.com/)
2. Click Credentials on the left hand navigation pane
3. Click the `Create Credentials Button` and select `oAuth Client id` from the resulting dropdown menu.
4. Select the `IOS` radio button
5. Enter a memorable name for this key to help you identify it easily in the future.
6. Copy the Bundle Identifier from your IOS App. This is found in your main project target under the `General` tab
7. Click the `Create` Button and your are complete.
## Note
These sample apps should automatically build and copy over the GTLR.framework as part of the build-and-run process.
| The apps in this folder are designed to provide examples of the basic workings of the Google Client Library. They should not be used in production.
Eg you should not ask for a Client key as it should be embedded within the codebase of your app.
## Installation instructons
In order to use the example projects in this folder, you will need to use CocoaPods to install their dependencies.
For help with installing CocoaPods, see [their getting started guide](https://guides.cocoapods.org/using/getting-started.html#getting-started)
To fetch/update the dependencies:
```
cd Examples/[sample_to_build]
pod install
```
## Requirements to use
These example apps need an oAuth2 Client key, which you can obtain from the [Google Developer Console](https://console.developers.google.com/).
### To obtain an oAuth2 client Key:
1. Visit the [Google Developer Console](https://console.developers.google.com/)
2. Click Credentials on the left hand navigation pane
3. Click the `Create Credentials Button` and select `oAuth Client id` from the resulting dropdown menu.
4. Select the `IOS` radio button
5. Enter a memorable name for this key to help you identify it easily in the future.
6. Copy the Bundle Identifier from your IOS App. This is found in your main project target under the `General` tab
7. Click the `Create` Button and your are complete.
| Update example notes about using pods to get the dependencies. | Update example notes about using pods to get the dependencies.
Fixes #260
| Markdown | apache-2.0 | google/google-api-objectivec-client-for-rest,google/google-api-objectivec-client-for-rest,google/google-api-objectivec-client-for-rest,google/google-api-objectivec-client-for-rest,google/google-api-objectivec-client-for-rest | markdown | ## Code Before:
The apps in this folder are designed to provide examples of the basic workings of the Google Client Library. They should not be used in production.
Eg you should not ask for a Client key as it should be embedded within the codebase of your app.
## Installation instructons
In order to use the example projects in this folder, you will need to ensure that you have downloaded the below two Git Submodules:
- gtm-session-fetcher
- AppAuth
- GTMAppAuth
## How to download Git Submodules
You can download Git Submodules using the following command:
[Credit: Stackoverflow Question](http://stackoverflow.com/questions/8090761/pull-using-git-including-submodule)
```
git clone git://url
cd repo
git submodule init
git submodule update
```
Then, add another step after the git pull.
```
git pull ...
git submodule update --recursive
```
## Requirements to use
These example apps need an oAuth2 Client key, which you can obtain from the [Google Developer Console](https://console.developers.google.com/).
### To obtain an oAuth2 client Key:
1. Visit the [Google Developer Console](https://console.developers.google.com/)
2. Click Credentials on the left hand navigation pane
3. Click the `Create Credentials Button` and select `oAuth Client id` from the resulting dropdown menu.
4. Select the `IOS` radio button
5. Enter a memorable name for this key to help you identify it easily in the future.
6. Copy the Bundle Identifier from your IOS App. This is found in your main project target under the `General` tab
7. Click the `Create` Button and your are complete.
## Note
These sample apps should automatically build and copy over the GTLR.framework as part of the build-and-run process.
## Instruction:
Update example notes about using pods to get the dependencies.
Fixes #260
## Code After:
The apps in this folder are designed to provide examples of the basic workings of the Google Client Library. They should not be used in production.
Eg you should not ask for a Client key as it should be embedded within the codebase of your app.
## Installation instructons
In order to use the example projects in this folder, you will need to use CocoaPods to install their dependencies.
For help with installing CocoaPods, see [their getting started guide](https://guides.cocoapods.org/using/getting-started.html#getting-started)
To fetch/update the dependencies:
```
cd Examples/[sample_to_build]
pod install
```
## Requirements to use
These example apps need an oAuth2 Client key, which you can obtain from the [Google Developer Console](https://console.developers.google.com/).
### To obtain an oAuth2 client Key:
1. Visit the [Google Developer Console](https://console.developers.google.com/)
2. Click Credentials on the left hand navigation pane
3. Click the `Create Credentials Button` and select `oAuth Client id` from the resulting dropdown menu.
4. Select the `IOS` radio button
5. Enter a memorable name for this key to help you identify it easily in the future.
6. Copy the Bundle Identifier from your IOS App. This is found in your main project target under the `General` tab
7. Click the `Create` Button and your are complete.
| The apps in this folder are designed to provide examples of the basic workings of the Google Client Library. They should not be used in production.
Eg you should not ask for a Client key as it should be embedded within the codebase of your app.
## Installation instructons
+ In order to use the example projects in this folder, you will need to use CocoaPods to install their dependencies.
+ For help with installing CocoaPods, see [their getting started guide](https://guides.cocoapods.org/using/getting-started.html#getting-started)
- In order to use the example projects in this folder, you will need to ensure that you have downloaded the below two Git Submodules:
- - gtm-session-fetcher
- - AppAuth
- - GTMAppAuth
+ To fetch/update the dependencies:
-
- ## How to download Git Submodules
- You can download Git Submodules using the following command:
- [Credit: Stackoverflow Question](http://stackoverflow.com/questions/8090761/pull-using-git-including-submodule)
-
-
```
+ cd Examples/[sample_to_build]
+ pod install
- git clone git://url
- cd repo
- git submodule init
- git submodule update
- ```
-
- Then, add another step after the git pull.
-
- ```
- git pull ...
- git submodule update --recursive
```
## Requirements to use
These example apps need an oAuth2 Client key, which you can obtain from the [Google Developer Console](https://console.developers.google.com/).
### To obtain an oAuth2 client Key:
1. Visit the [Google Developer Console](https://console.developers.google.com/)
2. Click Credentials on the left hand navigation pane
3. Click the `Create Credentials Button` and select `oAuth Client id` from the resulting dropdown menu.
4. Select the `IOS` radio button
5. Enter a memorable name for this key to help you identify it easily in the future.
6. Copy the Bundle Identifier from your IOS App. This is found in your main project target under the `General` tab
7. Click the `Create` Button and your are complete.
-
-
- ## Note
- These sample apps should automatically build and copy over the GTLR.framework as part of the build-and-run process. | 30 | 0.681818 | 5 | 25 |
3af0f0e1fbd0f7b10b74f58157d5b3fe3418bd8f | README.md | README.md | imuco
=====
Inertial Measurement Unit Coprocessor
| IMUco
=====
Inertial Measurement Unit Coprocessor
Hardware
========
The current IMUco implementation supports nine degrees of freedom (9 DoF) IMU sensor break-out boards equipped with the following sensors:
* InvenSense [ITG3200/ITG3205](http://invensense.com/mems/gyro/itg3200.html) 3-axis gyroscope
* Analog Devices [ADXL345](http://www.analog.com/en/mems-sensors/mems-inertial-sensors/adxl345/products/product.html) 3-axis accelerometer
* Honeywell [HMC5883L](http://www51.honeywell.com/aero/common/documents/myaerospacecatalog-documents/Defense_Brochures-documents/HMC5883L_3-Axis_Digital_Compass_IC.pdf) 3-axis magnetometer
These break-out boards are widely available online (e.g. on [eBay](http://www.ebay.co.uk/itm/130988986740)).
| Add initial description of the supported hardware | Add initial description of the supported hardware | Markdown | bsd-2-clause | henrikbrixandersen/imuco | markdown | ## Code Before:
imuco
=====
Inertial Measurement Unit Coprocessor
## Instruction:
Add initial description of the supported hardware
## Code After:
IMUco
=====
Inertial Measurement Unit Coprocessor
Hardware
========
The current IMUco implementation supports nine degrees of freedom (9 DoF) IMU sensor break-out boards equipped with the following sensors:
* InvenSense [ITG3200/ITG3205](http://invensense.com/mems/gyro/itg3200.html) 3-axis gyroscope
* Analog Devices [ADXL345](http://www.analog.com/en/mems-sensors/mems-inertial-sensors/adxl345/products/product.html) 3-axis accelerometer
* Honeywell [HMC5883L](http://www51.honeywell.com/aero/common/documents/myaerospacecatalog-documents/Defense_Brochures-documents/HMC5883L_3-Axis_Digital_Compass_IC.pdf) 3-axis magnetometer
These break-out boards are widely available online (e.g. on [eBay](http://www.ebay.co.uk/itm/130988986740)).
| - imuco
+ IMUco
=====
+ Inertial Measurement Unit Coprocessor
- Inertial Measurement Unit Coprocessor
+
+ Hardware
+ ========
+ The current IMUco implementation supports nine degrees of freedom (9 DoF) IMU sensor break-out boards equipped with the following sensors:
+ * InvenSense [ITG3200/ITG3205](http://invensense.com/mems/gyro/itg3200.html) 3-axis gyroscope
+ * Analog Devices [ADXL345](http://www.analog.com/en/mems-sensors/mems-inertial-sensors/adxl345/products/product.html) 3-axis accelerometer
+ * Honeywell [HMC5883L](http://www51.honeywell.com/aero/common/documents/myaerospacecatalog-documents/Defense_Brochures-documents/HMC5883L_3-Axis_Digital_Compass_IC.pdf) 3-axis magnetometer
+
+ These break-out boards are widely available online (e.g. on [eBay](http://www.ebay.co.uk/itm/130988986740)). | 13 | 3.25 | 11 | 2 |
6435b1daa004c046d34954a7269f0a2b179a4681 | tests/test_mem_leaks.sh | tests/test_mem_leaks.sh |
UNIT_TEST=./check_mem_leaks
VALGRIND_LOG_FILE=${UNIT_TEST}.valgrind
LEAK_MESSAGE="are definitely lost"
# This test runs valgrind against the check_mem_leaks unit test
# program, looking for memory leaks. If any are found, "exit 1"
# is invoked, and one must look through the resulting valgrind log
# file for details on the leak.
rm -f ${VALGRIND_LOG_FILE}
libtool --mode=execute valgrind --leak-check=full ${UNIT_TEST} 2>&1 | tee ${VALGRIND_LOG_FILE}
NUM_LEAKS=$(grep "${LEAK_MESSAGE}" ${VALGRIND_LOG_FILE} | wc -l)
if test ${NUM_LEAKS} -gt 0; then
echo "ERROR: ${NUM_LEAKS} memory leaks were detected by valgrind."
echo " Look through ${VALGRIND_LOG_FILE} for details,"
echo " searching for \"${LEAK_MESSAGE}\"."
exit 1
else
echo "No memory leaks found"
exit 0
fi
|
SCRIPT_PATH="$(dirname "$(readlink -f "$0")")"
if test -z "${1}"; then
UNIT_TEST="${SCRIPT_PATH}/check_mem_leaks"
else
UNIT_TEST="${1}"
fi
VALGRIND_LOG_FILE=${UNIT_TEST}.valgrind
LEAK_MESSAGE="are definitely lost"
# This test runs valgrind against the check_mem_leaks unit test
# program, looking for memory leaks. If any are found, "exit 1"
# is invoked, and one must look through the resulting valgrind log
# file for details on the leak.
rm -f ${VALGRIND_LOG_FILE}
libtool --mode=execute valgrind --leak-check=full ${UNIT_TEST} 2>&1 | tee ${VALGRIND_LOG_FILE}
NUM_LEAKS=$(grep "${LEAK_MESSAGE}" ${VALGRIND_LOG_FILE} | wc -l)
if test ${NUM_LEAKS} -gt 0; then
echo "ERROR: ${NUM_LEAKS} memory leaks were detected by valgrind."
echo " Look through ${VALGRIND_LOG_FILE} for details,"
echo " searching for \"${LEAK_MESSAGE}\"."
exit 1
else
echo "No memory leaks found"
exit 0
fi
| Fix script to work from any path and to take argument | Fix script to work from any path and to take argument
Signed-off-by: Mikko Koivunalho <b1070ee6d40473e0bf54e1be38982cfcedc3436a@iki.fi>
| Shell | lgpl-2.1 | libcheck/check,libcheck/check,libcheck/check,libcheck/check,libcheck/check | shell | ## Code Before:
UNIT_TEST=./check_mem_leaks
VALGRIND_LOG_FILE=${UNIT_TEST}.valgrind
LEAK_MESSAGE="are definitely lost"
# This test runs valgrind against the check_mem_leaks unit test
# program, looking for memory leaks. If any are found, "exit 1"
# is invoked, and one must look through the resulting valgrind log
# file for details on the leak.
rm -f ${VALGRIND_LOG_FILE}
libtool --mode=execute valgrind --leak-check=full ${UNIT_TEST} 2>&1 | tee ${VALGRIND_LOG_FILE}
NUM_LEAKS=$(grep "${LEAK_MESSAGE}" ${VALGRIND_LOG_FILE} | wc -l)
if test ${NUM_LEAKS} -gt 0; then
echo "ERROR: ${NUM_LEAKS} memory leaks were detected by valgrind."
echo " Look through ${VALGRIND_LOG_FILE} for details,"
echo " searching for \"${LEAK_MESSAGE}\"."
exit 1
else
echo "No memory leaks found"
exit 0
fi
## Instruction:
Fix script to work from any path and to take argument
Signed-off-by: Mikko Koivunalho <b1070ee6d40473e0bf54e1be38982cfcedc3436a@iki.fi>
## Code After:
SCRIPT_PATH="$(dirname "$(readlink -f "$0")")"
if test -z "${1}"; then
UNIT_TEST="${SCRIPT_PATH}/check_mem_leaks"
else
UNIT_TEST="${1}"
fi
VALGRIND_LOG_FILE=${UNIT_TEST}.valgrind
LEAK_MESSAGE="are definitely lost"
# This test runs valgrind against the check_mem_leaks unit test
# program, looking for memory leaks. If any are found, "exit 1"
# is invoked, and one must look through the resulting valgrind log
# file for details on the leak.
rm -f ${VALGRIND_LOG_FILE}
libtool --mode=execute valgrind --leak-check=full ${UNIT_TEST} 2>&1 | tee ${VALGRIND_LOG_FILE}
NUM_LEAKS=$(grep "${LEAK_MESSAGE}" ${VALGRIND_LOG_FILE} | wc -l)
if test ${NUM_LEAKS} -gt 0; then
echo "ERROR: ${NUM_LEAKS} memory leaks were detected by valgrind."
echo " Look through ${VALGRIND_LOG_FILE} for details,"
echo " searching for \"${LEAK_MESSAGE}\"."
exit 1
else
echo "No memory leaks found"
exit 0
fi
|
- UNIT_TEST=./check_mem_leaks
+ SCRIPT_PATH="$(dirname "$(readlink -f "$0")")"
+ if test -z "${1}"; then
+ UNIT_TEST="${SCRIPT_PATH}/check_mem_leaks"
+ else
+ UNIT_TEST="${1}"
+ fi
VALGRIND_LOG_FILE=${UNIT_TEST}.valgrind
LEAK_MESSAGE="are definitely lost"
# This test runs valgrind against the check_mem_leaks unit test
# program, looking for memory leaks. If any are found, "exit 1"
# is invoked, and one must look through the resulting valgrind log
# file for details on the leak.
rm -f ${VALGRIND_LOG_FILE}
libtool --mode=execute valgrind --leak-check=full ${UNIT_TEST} 2>&1 | tee ${VALGRIND_LOG_FILE}
NUM_LEAKS=$(grep "${LEAK_MESSAGE}" ${VALGRIND_LOG_FILE} | wc -l)
if test ${NUM_LEAKS} -gt 0; then
echo "ERROR: ${NUM_LEAKS} memory leaks were detected by valgrind."
echo " Look through ${VALGRIND_LOG_FILE} for details,"
echo " searching for \"${LEAK_MESSAGE}\"."
exit 1
else
echo "No memory leaks found"
exit 0
fi | 7 | 0.291667 | 6 | 1 |
fbb8f67f838f8fe820d6446fca2d702b172b1472 | src/common/Label.js | src/common/Label.js | /* @flow */
import React, { PureComponent } from 'react';
import { Text } from 'react-native';
import { FormattedMessage } from 'react-intl';
import type { LocalizableText, Style } from '../types';
type Props = {
text: LocalizableText,
style?: Style,
};
export default class Label extends PureComponent<Props> {
static contextTypes = {
styles: () => null,
};
props: Props;
render() {
const { text, style, ...restProps } = this.props;
const message = text.text || text;
const { styles } = this.context;
return (
<Text style={[styles.label, style]} {...restProps}>
<FormattedMessage id={message} defaultMessage={message} values={text.values} />
</Text>
);
}
}
| /* @flow */
import React, { PureComponent } from 'react';
import { Text } from 'react-native';
import TranslatedText from './TranslatedText';
import type { LocalizableText, Style } from '../types';
type Props = {
text: LocalizableText,
style?: Style,
};
export default class Label extends PureComponent<Props> {
static contextTypes = {
styles: () => null,
};
props: Props;
render() {
const { styles } = this.context;
const { text, style, ...restProps } = this.props;
return (
<Text style={[styles.label, style]} {...restProps}>
<TranslatedText text={text} />
</Text>
);
}
}
| Use TranslatedText component to get translated text. | label: Use TranslatedText component to get translated text.
| JavaScript | apache-2.0 | vishwesh3/zulip-mobile,vishwesh3/zulip-mobile,vishwesh3/zulip-mobile,vishwesh3/zulip-mobile | javascript | ## Code Before:
/* @flow */
import React, { PureComponent } from 'react';
import { Text } from 'react-native';
import { FormattedMessage } from 'react-intl';
import type { LocalizableText, Style } from '../types';
type Props = {
text: LocalizableText,
style?: Style,
};
export default class Label extends PureComponent<Props> {
static contextTypes = {
styles: () => null,
};
props: Props;
render() {
const { text, style, ...restProps } = this.props;
const message = text.text || text;
const { styles } = this.context;
return (
<Text style={[styles.label, style]} {...restProps}>
<FormattedMessage id={message} defaultMessage={message} values={text.values} />
</Text>
);
}
}
## Instruction:
label: Use TranslatedText component to get translated text.
## Code After:
/* @flow */
import React, { PureComponent } from 'react';
import { Text } from 'react-native';
import TranslatedText from './TranslatedText';
import type { LocalizableText, Style } from '../types';
type Props = {
text: LocalizableText,
style?: Style,
};
export default class Label extends PureComponent<Props> {
static contextTypes = {
styles: () => null,
};
props: Props;
render() {
const { styles } = this.context;
const { text, style, ...restProps } = this.props;
return (
<Text style={[styles.label, style]} {...restProps}>
<TranslatedText text={text} />
</Text>
);
}
}
| /* @flow */
import React, { PureComponent } from 'react';
import { Text } from 'react-native';
- import { FormattedMessage } from 'react-intl';
+ import TranslatedText from './TranslatedText';
import type { LocalizableText, Style } from '../types';
type Props = {
text: LocalizableText,
style?: Style,
};
export default class Label extends PureComponent<Props> {
static contextTypes = {
styles: () => null,
};
props: Props;
render() {
+ const { styles } = this.context;
const { text, style, ...restProps } = this.props;
- const message = text.text || text;
- const { styles } = this.context;
return (
<Text style={[styles.label, style]} {...restProps}>
- <FormattedMessage id={message} defaultMessage={message} values={text.values} />
+ <TranslatedText text={text} />
</Text>
);
}
} | 7 | 0.225806 | 3 | 4 |
582811074db86be964648dc9457855db3549a2b5 | data_structures/Disjoint_Set_Union/Python/dsu.py | data_structures/Disjoint_Set_Union/Python/dsu.py |
parent=[]
size=[]
def initialize(n):
for i in range(0,n+1):
parent.append(i)
size.append(1)
def find(x):
if parent[x] == x:
return x
else:
return find(parent[x])
def join(a,b):
p_a = find(a)
p_b = find(b)
if p_a != p_b:
if size[p_a] < size[p_b]:
parent[p_a] = p_b
size[p_b] += size[p_a]
else:
parent[p_b] = p_a
size[p_a] += size[p_b]
''' Main Program Starts Here '''
n=5
initialize(n)
join(1,2)
join(2,3)
join(4,5)
print(find(3))
|
parent=[]
size=[]
def initialize(n):
for i in range(0,n+1):
parent.append(i)
size.append(1)
def find(x):
if parent[x] == x:
return x
else:
return find(parent[x])
def join(a,b):
p_a = find(a)
p_b = find(b)
if p_a != p_b:
if size[p_a] < size[p_b]:
parent[p_a] = p_b
size[p_b] += size[p_a]
else:
parent[p_b] = p_a
size[p_a] += size[p_b]
''' Main Program Starts Here '''
def main():
n=5
initialize(n)
join(1,2)
assert(find(2) == 1)
assert(find(3) == 3)
join(2,3)
assert(find(3) == 1)
assert(find(5) == 5)
join(4,5)
assert(find(5) == 4)
join(3,4)
assert(find(5) == 1)
if __name__ == '__main__':
main()
| Test for DSU on Python | Test for DSU on Python
| Python | cc0-1.0 | ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms | python | ## Code Before:
parent=[]
size=[]
def initialize(n):
for i in range(0,n+1):
parent.append(i)
size.append(1)
def find(x):
if parent[x] == x:
return x
else:
return find(parent[x])
def join(a,b):
p_a = find(a)
p_b = find(b)
if p_a != p_b:
if size[p_a] < size[p_b]:
parent[p_a] = p_b
size[p_b] += size[p_a]
else:
parent[p_b] = p_a
size[p_a] += size[p_b]
''' Main Program Starts Here '''
n=5
initialize(n)
join(1,2)
join(2,3)
join(4,5)
print(find(3))
## Instruction:
Test for DSU on Python
## Code After:
parent=[]
size=[]
def initialize(n):
for i in range(0,n+1):
parent.append(i)
size.append(1)
def find(x):
if parent[x] == x:
return x
else:
return find(parent[x])
def join(a,b):
p_a = find(a)
p_b = find(b)
if p_a != p_b:
if size[p_a] < size[p_b]:
parent[p_a] = p_b
size[p_b] += size[p_a]
else:
parent[p_b] = p_a
size[p_a] += size[p_b]
''' Main Program Starts Here '''
def main():
n=5
initialize(n)
join(1,2)
assert(find(2) == 1)
assert(find(3) == 3)
join(2,3)
assert(find(3) == 1)
assert(find(5) == 5)
join(4,5)
assert(find(5) == 4)
join(3,4)
assert(find(5) == 1)
if __name__ == '__main__':
main()
|
parent=[]
size=[]
def initialize(n):
for i in range(0,n+1):
parent.append(i)
size.append(1)
def find(x):
if parent[x] == x:
return x
else:
return find(parent[x])
def join(a,b):
p_a = find(a)
p_b = find(b)
if p_a != p_b:
if size[p_a] < size[p_b]:
parent[p_a] = p_b
size[p_b] += size[p_a]
else:
parent[p_b] = p_a
size[p_a] += size[p_b]
''' Main Program Starts Here '''
+ def main():
+ n=5
+ initialize(n)
+ join(1,2)
+ assert(find(2) == 1)
+ assert(find(3) == 3)
+ join(2,3)
+ assert(find(3) == 1)
+ assert(find(5) == 5)
+ join(4,5)
+ assert(find(5) == 4)
+ join(3,4)
+ assert(find(5) == 1)
+ if __name__ == '__main__':
+ main()
- n=5
- initialize(n)
- join(1,2)
- join(2,3)
- join(4,5)
-
- print(find(3))
| 22 | 0.564103 | 15 | 7 |
b337576b958d6280933a7a44eac466f32885fa47 | lib/phil_columns/command/generate/seed.rb | lib/phil_columns/command/generate/seed.rb | require 'fileutils'
require 'ostruct'
require 'pathname'
module PhilColumns
module Command
module Generate
class Seed < Generator
def execute
write "Generating seed #{seed_filepath} ... "
erb_template_to_file( template_filepath, seed_filepath, class_name: seed_class_name )
say_ok
end
protected
def template_filepath
'templates/seed_class.erb'
end
end
end
end
end
| require 'fileutils'
require 'ostruct'
require 'pathname'
module PhilColumns
module Command
module Generate
class Seed < Generator
def execute
write "Generating seed #{seed_filepath} ... "
erb_template_to_file( template_filepath, seed_filepath, class_name: seed_class_name )
say_ok
end
protected
def template_filepath
File.expand_path( '../../../../../templates/seed_class.erb', __FILE__ )
end
end
end
end
end
| Fix issue with template path in generator. | Fix issue with template path in generator.
| Ruby | mit | midas/phil_columns | ruby | ## Code Before:
require 'fileutils'
require 'ostruct'
require 'pathname'
module PhilColumns
module Command
module Generate
class Seed < Generator
def execute
write "Generating seed #{seed_filepath} ... "
erb_template_to_file( template_filepath, seed_filepath, class_name: seed_class_name )
say_ok
end
protected
def template_filepath
'templates/seed_class.erb'
end
end
end
end
end
## Instruction:
Fix issue with template path in generator.
## Code After:
require 'fileutils'
require 'ostruct'
require 'pathname'
module PhilColumns
module Command
module Generate
class Seed < Generator
def execute
write "Generating seed #{seed_filepath} ... "
erb_template_to_file( template_filepath, seed_filepath, class_name: seed_class_name )
say_ok
end
protected
def template_filepath
File.expand_path( '../../../../../templates/seed_class.erb', __FILE__ )
end
end
end
end
end
| require 'fileutils'
require 'ostruct'
require 'pathname'
module PhilColumns
module Command
module Generate
class Seed < Generator
def execute
write "Generating seed #{seed_filepath} ... "
erb_template_to_file( template_filepath, seed_filepath, class_name: seed_class_name )
say_ok
end
protected
def template_filepath
- 'templates/seed_class.erb'
+ File.expand_path( '../../../../../templates/seed_class.erb', __FILE__ )
end
end
end
end
end | 2 | 0.08 | 1 | 1 |
8bcfe832ac30da108c0df6a6a7623ced9738722d | code/controllers/subsystem/server_maintenance.dm | code/controllers/subsystem/server_maintenance.dm | var/datum/subsystem/server_maint/SSserver
/datum/subsystem/server_maint
name = "Server Tasks"
wait = 6000
flags = SS_NO_TICK_CHECK
/datum/subsystem/server_maint/New()
NEW_SS_GLOBAL(SSserver)
/datum/subsystem/server_maint/Initialize(timeofday)
if (config.hub)
message_admins("This server is on the SS13 hub.")
world.visibility = 1
..()
/datum/subsystem/server_maint/fire()
//handle kicking inactive players
if(config.kick_inactive > 0)
for(var/client/C in clients)
if(C.is_afk(INACTIVITY_KICK))
if(!istype(C.mob, /mob/dead))
log_access("AFK: [key_name(C)]")
C << "<span class='danger'>You have been inactive for more than 10 minutes and have been disconnected.</span>"
qdel(C)
if(config.sql_enabled)
sql_poll_players()
sql_poll_admins()
| var/datum/subsystem/server_maint/SSserver
/datum/subsystem/server_maint
name = "Server Tasks"
wait = 6000
flags = SS_NO_TICK_CHECK
/datum/subsystem/server_maint/New()
NEW_SS_GLOBAL(SSserver)
/datum/subsystem/server_maint/Initialize(timeofday)
if (config.hub)
world.log << "This server has hub broadcast enabled."
world.visibility = 1
else
world.log << "This server has hub broadcast disabled."
..()
/datum/subsystem/server_maint/fire()
//handle kicking inactive players
if(config.kick_inactive > 0)
for(var/client/C in clients)
if(C.is_afk(INACTIVITY_KICK))
if(!istype(C.mob, /mob/dead))
log_access("AFK: [key_name(C)]")
C << "<span class='danger'>You have been inactive for more than 10 minutes and have been disconnected.</span>"
qdel(C)
if(config.sql_enabled)
sql_poll_players()
sql_poll_admins()
| Add more hub debugging code. | Add more hub debugging code.
| DM | agpl-3.0 | Memendia/HippieStation13,HippieStationCode/HippieStation13,Memendia/HippieStation13,HippieStationCode/HippieStation13,Memendia/HippieStation13,HippieStationCode/HippieStation13,HippieStationCode/HippieStation13,Memendia/HippieStation13 | dm | ## Code Before:
var/datum/subsystem/server_maint/SSserver
/datum/subsystem/server_maint
name = "Server Tasks"
wait = 6000
flags = SS_NO_TICK_CHECK
/datum/subsystem/server_maint/New()
NEW_SS_GLOBAL(SSserver)
/datum/subsystem/server_maint/Initialize(timeofday)
if (config.hub)
message_admins("This server is on the SS13 hub.")
world.visibility = 1
..()
/datum/subsystem/server_maint/fire()
//handle kicking inactive players
if(config.kick_inactive > 0)
for(var/client/C in clients)
if(C.is_afk(INACTIVITY_KICK))
if(!istype(C.mob, /mob/dead))
log_access("AFK: [key_name(C)]")
C << "<span class='danger'>You have been inactive for more than 10 minutes and have been disconnected.</span>"
qdel(C)
if(config.sql_enabled)
sql_poll_players()
sql_poll_admins()
## Instruction:
Add more hub debugging code.
## Code After:
var/datum/subsystem/server_maint/SSserver
/datum/subsystem/server_maint
name = "Server Tasks"
wait = 6000
flags = SS_NO_TICK_CHECK
/datum/subsystem/server_maint/New()
NEW_SS_GLOBAL(SSserver)
/datum/subsystem/server_maint/Initialize(timeofday)
if (config.hub)
world.log << "This server has hub broadcast enabled."
world.visibility = 1
else
world.log << "This server has hub broadcast disabled."
..()
/datum/subsystem/server_maint/fire()
//handle kicking inactive players
if(config.kick_inactive > 0)
for(var/client/C in clients)
if(C.is_afk(INACTIVITY_KICK))
if(!istype(C.mob, /mob/dead))
log_access("AFK: [key_name(C)]")
C << "<span class='danger'>You have been inactive for more than 10 minutes and have been disconnected.</span>"
qdel(C)
if(config.sql_enabled)
sql_poll_players()
sql_poll_admins()
| var/datum/subsystem/server_maint/SSserver
/datum/subsystem/server_maint
name = "Server Tasks"
wait = 6000
flags = SS_NO_TICK_CHECK
/datum/subsystem/server_maint/New()
NEW_SS_GLOBAL(SSserver)
/datum/subsystem/server_maint/Initialize(timeofday)
if (config.hub)
- message_admins("This server is on the SS13 hub.")
+ world.log << "This server has hub broadcast enabled."
world.visibility = 1
+ else
+ world.log << "This server has hub broadcast disabled."
..()
/datum/subsystem/server_maint/fire()
//handle kicking inactive players
if(config.kick_inactive > 0)
for(var/client/C in clients)
if(C.is_afk(INACTIVITY_KICK))
if(!istype(C.mob, /mob/dead))
log_access("AFK: [key_name(C)]")
C << "<span class='danger'>You have been inactive for more than 10 minutes and have been disconnected.</span>"
qdel(C)
if(config.sql_enabled)
sql_poll_players()
sql_poll_admins() | 4 | 0.137931 | 3 | 1 |
b48c9f339895a40b02c86fd4a662f9ce7b278eed | recipes/kraken/meta.yaml | recipes/kraken/meta.yaml | package:
name: kraken
version: "0.10.6_eaf8fb68"
source:
fn: kraken-eaf8fb68.tar.gz
url: https://github.com/DerrickWood/kraken/archive/eaf8fb68.tar.gz
build:
number: 2
skip: True # [osx]
has_prefix_files:
- libexec/kraken
- libexec/kraken-build
- libexec/kraken-filter
- libexec/kraken-mpa-report
- libexec/kraken-report
- libexec/kraken-translate
requirements:
build:
run:
- perl-threaded
test:
commands:
- kraken --version 2>&1 > /dev/null
about:
home: http://ccb.jhu.edu/software/kraken/
license: GPLv3
summary: Kraken is a system for assigning taxonomic labels to short DNA sequences, usually obtained through metagenomic studies.
| package:
name: kraken
version: "0.10.6_eaf8fb68"
source:
fn: kraken-eaf8fb68.tar.gz
url: https://github.com/DerrickWood/kraken/archive/eaf8fb68.tar.gz
sha256: 0425fb88ce1d66ca5187cf348970d1dd2cec61d69bd53c466ca43f957de71ffd
build:
number: 3
skip: True # [osx]
has_prefix_files:
- libexec/kraken
- libexec/kraken-build
- libexec/kraken-filter
- libexec/kraken-mpa-report
- libexec/kraken-report
- libexec/kraken-translate
requirements:
build:
- perl
run:
- jellyfish 1.*
- perl
test:
commands:
- kraken --version 2>&1 > /dev/null
- kraken-build --version
about:
home: http://ccb.jhu.edu/software/kraken/
license: GPLv3
summary: Kraken is a system for assigning taxonomic labels to short DNA sequences, usually obtained through metagenomic studies.
| Add jellyfish 1 as run requirement for kraken | Add jellyfish 1 as run requirement for kraken
It is required for kraken-build.
Also add perl as build requirement.
Change perl-threaded to perl.
| YAML | mit | xguse/bioconda-recipes,dkoppstein/recipes,rob-p/bioconda-recipes,omicsnut/bioconda-recipes,bebatut/bioconda-recipes,lpantano/recipes,bow/bioconda-recipes,mdehollander/bioconda-recipes,abims-sbr/bioconda-recipes,acaprez/recipes,cokelaer/bioconda-recipes,matthdsm/bioconda-recipes,ivirshup/bioconda-recipes,jasper1918/bioconda-recipes,xguse/bioconda-recipes,bioconda/recipes,bebatut/bioconda-recipes,saketkc/bioconda-recipes,npavlovikj/bioconda-recipes,joachimwolff/bioconda-recipes,hardingnj/bioconda-recipes,cokelaer/bioconda-recipes,jasper1918/bioconda-recipes,colinbrislawn/bioconda-recipes,blankenberg/bioconda-recipes,gregvonkuster/bioconda-recipes,daler/bioconda-recipes,saketkc/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,daler/bioconda-recipes,lpantano/recipes,blankenberg/bioconda-recipes,peterjc/bioconda-recipes,martin-mann/bioconda-recipes,rvalieris/bioconda-recipes,martin-mann/bioconda-recipes,bow/bioconda-recipes,npavlovikj/bioconda-recipes,hardingnj/bioconda-recipes,HassanAmr/bioconda-recipes,jfallmann/bioconda-recipes,gregvonkuster/bioconda-recipes,ostrokach/bioconda-recipes,rvalieris/bioconda-recipes,dmaticzka/bioconda-recipes,ostrokach/bioconda-recipes,ivirshup/bioconda-recipes,CGATOxford/bioconda-recipes,rob-p/bioconda-recipes,martin-mann/bioconda-recipes,dmaticzka/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,rob-p/bioconda-recipes,abims-sbr/bioconda-recipes,peterjc/bioconda-recipes,bow/bioconda-recipes,acaprez/recipes,daler/bioconda-recipes,acaprez/recipes,instituteofpathologyheidelberg/bioconda-recipes,HassanAmr/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,ostrokach/bioconda-recipes,phac-nml/bioconda-recipes,CGATOxford/bioconda-recipes,gregvonkuster/bioconda-recipes,phac-nml/bioconda-recipes,omicsnut/bioconda-recipes,matthdsm/bioconda-recipes,abims-sbr/bioconda-recipes,lpantano/recipes,cokelaer/bioconda-recipes,saketkc/bioconda-recipes,bow/bioconda-recipes,lpantano/recipes,bioconda/bioconda-recipes,jasper1918/bioconda-recipes,dmaticzka/bioconda-recipes,oena/bioconda-recipes,bioconda/recipes,zachcp/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,HassanAmr/bioconda-recipes,peterjc/bioconda-recipes,xguse/bioconda-recipes,abims-sbr/bioconda-recipes,rvalieris/bioconda-recipes,oena/bioconda-recipes,daler/bioconda-recipes,colinbrislawn/bioconda-recipes,jfallmann/bioconda-recipes,mcornwell1957/bioconda-recipes,jfallmann/bioconda-recipes,dmaticzka/bioconda-recipes,colinbrislawn/bioconda-recipes,ivirshup/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,CGATOxford/bioconda-recipes,saketkc/bioconda-recipes,omicsnut/bioconda-recipes,colinbrislawn/bioconda-recipes,roryk/recipes,bioconda/bioconda-recipes,matthdsm/bioconda-recipes,jasper1918/bioconda-recipes,CGATOxford/bioconda-recipes,dmaticzka/bioconda-recipes,peterjc/bioconda-recipes,gvlproject/bioconda-recipes,xguse/bioconda-recipes,bow/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,gregvonkuster/bioconda-recipes,hardingnj/bioconda-recipes,joachimwolff/bioconda-recipes,peterjc/bioconda-recipes,keuv-grvl/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,xguse/bioconda-recipes,ostrokach/bioconda-recipes,mcornwell1957/bioconda-recipes,martin-mann/bioconda-recipes,matthdsm/bioconda-recipes,abims-sbr/bioconda-recipes,saketkc/bioconda-recipes,keuv-grvl/bioconda-recipes,CGATOxford/bioconda-recipes,rvalieris/bioconda-recipes,daler/bioconda-recipes,bebatut/bioconda-recipes,hardingnj/bioconda-recipes,omicsnut/bioconda-recipes,HassanAmr/bioconda-recipes,mdehollander/bioconda-recipes,keuv-grvl/bioconda-recipes,chapmanb/bioconda-recipes,zachcp/bioconda-recipes,joachimwolff/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,dkoppstein/recipes,dkoppstein/recipes,ivirshup/bioconda-recipes,jfallmann/bioconda-recipes,joachimwolff/bioconda-recipes,bebatut/bioconda-recipes,keuv-grvl/bioconda-recipes,joachimwolff/bioconda-recipes,phac-nml/bioconda-recipes,bioconda/bioconda-recipes,zachcp/bioconda-recipes,blankenberg/bioconda-recipes,shenwei356/bioconda-recipes,HassanAmr/bioconda-recipes,HassanAmr/bioconda-recipes,zachcp/bioconda-recipes,ostrokach/bioconda-recipes,abims-sbr/bioconda-recipes,keuv-grvl/bioconda-recipes,rvalieris/bioconda-recipes,martin-mann/bioconda-recipes,acaprez/recipes,mdehollander/bioconda-recipes,rvalieris/bioconda-recipes,shenwei356/bioconda-recipes,keuv-grvl/bioconda-recipes,Luobiny/bioconda-recipes,chapmanb/bioconda-recipes,bioconda/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,rob-p/bioconda-recipes,shenwei356/bioconda-recipes,matthdsm/bioconda-recipes,dmaticzka/bioconda-recipes,mcornwell1957/bioconda-recipes,cokelaer/bioconda-recipes,gvlproject/bioconda-recipes,gvlproject/bioconda-recipes,colinbrislawn/bioconda-recipes,peterjc/bioconda-recipes,phac-nml/bioconda-recipes,npavlovikj/bioconda-recipes,npavlovikj/bioconda-recipes,ivirshup/bioconda-recipes,oena/bioconda-recipes,Luobiny/bioconda-recipes,bow/bioconda-recipes,colinbrislawn/bioconda-recipes,mcornwell1957/bioconda-recipes,Luobiny/bioconda-recipes,omicsnut/bioconda-recipes,roryk/recipes,phac-nml/bioconda-recipes,Luobiny/bioconda-recipes,oena/bioconda-recipes,chapmanb/bioconda-recipes,gvlproject/bioconda-recipes,bioconda/recipes,mdehollander/bioconda-recipes,chapmanb/bioconda-recipes,mdehollander/bioconda-recipes,hardingnj/bioconda-recipes,CGATOxford/bioconda-recipes,blankenberg/bioconda-recipes,gvlproject/bioconda-recipes,matthdsm/bioconda-recipes,daler/bioconda-recipes,chapmanb/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,gvlproject/bioconda-recipes,ivirshup/bioconda-recipes,shenwei356/bioconda-recipes,mcornwell1957/bioconda-recipes,joachimwolff/bioconda-recipes,oena/bioconda-recipes,saketkc/bioconda-recipes,mdehollander/bioconda-recipes,ostrokach/bioconda-recipes,roryk/recipes,jasper1918/bioconda-recipes | yaml | ## Code Before:
package:
name: kraken
version: "0.10.6_eaf8fb68"
source:
fn: kraken-eaf8fb68.tar.gz
url: https://github.com/DerrickWood/kraken/archive/eaf8fb68.tar.gz
build:
number: 2
skip: True # [osx]
has_prefix_files:
- libexec/kraken
- libexec/kraken-build
- libexec/kraken-filter
- libexec/kraken-mpa-report
- libexec/kraken-report
- libexec/kraken-translate
requirements:
build:
run:
- perl-threaded
test:
commands:
- kraken --version 2>&1 > /dev/null
about:
home: http://ccb.jhu.edu/software/kraken/
license: GPLv3
summary: Kraken is a system for assigning taxonomic labels to short DNA sequences, usually obtained through metagenomic studies.
## Instruction:
Add jellyfish 1 as run requirement for kraken
It is required for kraken-build.
Also add perl as build requirement.
Change perl-threaded to perl.
## Code After:
package:
name: kraken
version: "0.10.6_eaf8fb68"
source:
fn: kraken-eaf8fb68.tar.gz
url: https://github.com/DerrickWood/kraken/archive/eaf8fb68.tar.gz
sha256: 0425fb88ce1d66ca5187cf348970d1dd2cec61d69bd53c466ca43f957de71ffd
build:
number: 3
skip: True # [osx]
has_prefix_files:
- libexec/kraken
- libexec/kraken-build
- libexec/kraken-filter
- libexec/kraken-mpa-report
- libexec/kraken-report
- libexec/kraken-translate
requirements:
build:
- perl
run:
- jellyfish 1.*
- perl
test:
commands:
- kraken --version 2>&1 > /dev/null
- kraken-build --version
about:
home: http://ccb.jhu.edu/software/kraken/
license: GPLv3
summary: Kraken is a system for assigning taxonomic labels to short DNA sequences, usually obtained through metagenomic studies.
| package:
name: kraken
version: "0.10.6_eaf8fb68"
source:
fn: kraken-eaf8fb68.tar.gz
url: https://github.com/DerrickWood/kraken/archive/eaf8fb68.tar.gz
+ sha256: 0425fb88ce1d66ca5187cf348970d1dd2cec61d69bd53c466ca43f957de71ffd
build:
- number: 2
? ^
+ number: 3
? ^
skip: True # [osx]
has_prefix_files:
- libexec/kraken
- libexec/kraken-build
- libexec/kraken-filter
- libexec/kraken-mpa-report
- libexec/kraken-report
- libexec/kraken-translate
requirements:
build:
+ - perl
run:
- - perl-threaded
+ - jellyfish 1.*
+ - perl
test:
commands:
- kraken --version 2>&1 > /dev/null
+ - kraken-build --version
about:
home: http://ccb.jhu.edu/software/kraken/
license: GPLv3
summary: Kraken is a system for assigning taxonomic labels to short DNA sequences, usually obtained through metagenomic studies. | 8 | 0.258065 | 6 | 2 |
deaab8d14229aacb2c4730a6c6f201ff4f1f971b | README.md | README.md | ember-cli-data-factory-guy
===========
##About
ember-cli-data-factory-guy is a wrapper around ember-data-factory-guy for ember-cli.
##Install
To install, run:
- `npm install ember-cli-data-factory-guy`
- `bower install ember-data-factory-guy` or `ember g ember-cli-data-factory-guy`
| ember-cli-data-factory-guy
===========
##About
ember-cli-data-factory-guy is a wrapper around ember-data-factory-guy for
ember-cli.
##Install
To install, run:
- `npm install ember-cli-data-factory-guy`
- `ember g ember-cli-data-factory-guy` to install ember-data-factory-guy and add
its predefined variables to tests/.jshintrc
| Add FactoryGuy and FactoryGuyTestMixin to the tests jsHint | Add FactoryGuy and FactoryGuyTestMixin to the tests jsHint
| Markdown | mit | cristinawithout/ember-cli-data-factory-guy | markdown | ## Code Before:
ember-cli-data-factory-guy
===========
##About
ember-cli-data-factory-guy is a wrapper around ember-data-factory-guy for ember-cli.
##Install
To install, run:
- `npm install ember-cli-data-factory-guy`
- `bower install ember-data-factory-guy` or `ember g ember-cli-data-factory-guy`
## Instruction:
Add FactoryGuy and FactoryGuyTestMixin to the tests jsHint
## Code After:
ember-cli-data-factory-guy
===========
##About
ember-cli-data-factory-guy is a wrapper around ember-data-factory-guy for
ember-cli.
##Install
To install, run:
- `npm install ember-cli-data-factory-guy`
- `ember g ember-cli-data-factory-guy` to install ember-data-factory-guy and add
its predefined variables to tests/.jshintrc
| ember-cli-data-factory-guy
===========
##About
- ember-cli-data-factory-guy is a wrapper around ember-data-factory-guy for ember-cli.
? ----------
+ ember-cli-data-factory-guy is a wrapper around ember-data-factory-guy for
+ ember-cli.
##Install
To install, run:
- `npm install ember-cli-data-factory-guy`
- - `bower install ember-data-factory-guy` or `ember g ember-cli-data-factory-guy`
+ - `ember g ember-cli-data-factory-guy` to install ember-data-factory-guy and add
+ its predefined variables to tests/.jshintrc | 6 | 0.461538 | 4 | 2 |
e4f9953362d367f89e9a830c60df900c24b4d2e7 | etc/standalone/dragonflow.ini | etc/standalone/dragonflow.ini | [DEFAULT]
[df]
pub_sub_driver = etcd_pubsub_driver
enable_selective_topology_distribution = False
enable_df_pub_sub = True
apps_list = portbinding,l2,l3_proactive,dhcp,dnat,sg,portqos,classifier,tunneling,provider,metadata_service,active_port_detection,migration
integration_bridge = br-int
tunnel_types = vxlan,geneve,gre
local_ip = LOCAL_IP
management_ip = MANAGEMENT_IP
enable_dpdk = False
enable_neutron_notifier = False
remote_db_hosts = DB_SERVER_IP
nb_db_class = etcd_nb_db_driver
auto_detect_port_behind_port = False
publisher_rate_limit_count = 1
publisher_rate_limit_timeout = 180
monitor_table_poll_time = 30
datapath_layout_path=/etc/dragonflow/dragonflow_datapath_layout.yaml
[df_l2_app]
l2_responder = True
[df_loadbalancer]
auto_enable_vip_ports = True
[df_metadata]
metadata_interface = tap-metadata
port = 18080
ip = 169.254.169.254
| [DEFAULT]
[df]
pub_sub_driver = etcd_pubsub_driver
enable_selective_topology_distribution = False
enable_df_pub_sub = True
apps_list = portbinding,l2,l3_proactive,dhcp,dnat,sg,portqos,classifier,tunneling,provider,metadata_service,active_port_detection,migration
integration_bridge = br-int
tunnel_types = vxlan,geneve,gre
local_ip = LOCAL_IP
management_ip = MANAGEMENT_IP
enable_dpdk = False
enable_neutron_notifier = False
remote_db_hosts = DB_SERVER_IP
nb_db_class = etcd_nb_db_driver
auto_detect_port_behind_port = False
publisher_rate_limit_count = 1
publisher_rate_limit_timeout = 180
monitor_table_poll_time = 30
datapath_layout_path=/etc/dragonflow/dragonflow_datapath_layout.yaml
[df_l2_app]
l2_responder = True
[df_loadbalancer]
auto_enable_vip_ports = True
[df_metadata]
metadata_interface = tap-metadata
port = 18080
ip = 169.254.169.254
[df_ryu]
of_listen_address = LOCAL_IP
| Add OpenFlow listen address to standalone configuration | Add OpenFlow listen address to standalone configuration
In standalone configuration, the local controller needs to specify
on which address it is listening.
One scenario is in case the local controller is containerised.
Dragonflow informs OVS where the controller is listening, and
then OVS knows where to connect to the controller.
Related-Bug: #1778007
Change-Id: I4b4cdae3aa848aa02a8148a9d47f6262f707e0ce
| INI | apache-2.0 | openstack/dragonflow,openstack/dragonflow,openstack/dragonflow | ini | ## Code Before:
[DEFAULT]
[df]
pub_sub_driver = etcd_pubsub_driver
enable_selective_topology_distribution = False
enable_df_pub_sub = True
apps_list = portbinding,l2,l3_proactive,dhcp,dnat,sg,portqos,classifier,tunneling,provider,metadata_service,active_port_detection,migration
integration_bridge = br-int
tunnel_types = vxlan,geneve,gre
local_ip = LOCAL_IP
management_ip = MANAGEMENT_IP
enable_dpdk = False
enable_neutron_notifier = False
remote_db_hosts = DB_SERVER_IP
nb_db_class = etcd_nb_db_driver
auto_detect_port_behind_port = False
publisher_rate_limit_count = 1
publisher_rate_limit_timeout = 180
monitor_table_poll_time = 30
datapath_layout_path=/etc/dragonflow/dragonflow_datapath_layout.yaml
[df_l2_app]
l2_responder = True
[df_loadbalancer]
auto_enable_vip_ports = True
[df_metadata]
metadata_interface = tap-metadata
port = 18080
ip = 169.254.169.254
## Instruction:
Add OpenFlow listen address to standalone configuration
In standalone configuration, the local controller needs to specify
on which address it is listening.
One scenario is in case the local controller is containerised.
Dragonflow informs OVS where the controller is listening, and
then OVS knows where to connect to the controller.
Related-Bug: #1778007
Change-Id: I4b4cdae3aa848aa02a8148a9d47f6262f707e0ce
## Code After:
[DEFAULT]
[df]
pub_sub_driver = etcd_pubsub_driver
enable_selective_topology_distribution = False
enable_df_pub_sub = True
apps_list = portbinding,l2,l3_proactive,dhcp,dnat,sg,portqos,classifier,tunneling,provider,metadata_service,active_port_detection,migration
integration_bridge = br-int
tunnel_types = vxlan,geneve,gre
local_ip = LOCAL_IP
management_ip = MANAGEMENT_IP
enable_dpdk = False
enable_neutron_notifier = False
remote_db_hosts = DB_SERVER_IP
nb_db_class = etcd_nb_db_driver
auto_detect_port_behind_port = False
publisher_rate_limit_count = 1
publisher_rate_limit_timeout = 180
monitor_table_poll_time = 30
datapath_layout_path=/etc/dragonflow/dragonflow_datapath_layout.yaml
[df_l2_app]
l2_responder = True
[df_loadbalancer]
auto_enable_vip_ports = True
[df_metadata]
metadata_interface = tap-metadata
port = 18080
ip = 169.254.169.254
[df_ryu]
of_listen_address = LOCAL_IP
| [DEFAULT]
[df]
pub_sub_driver = etcd_pubsub_driver
enable_selective_topology_distribution = False
enable_df_pub_sub = True
apps_list = portbinding,l2,l3_proactive,dhcp,dnat,sg,portqos,classifier,tunneling,provider,metadata_service,active_port_detection,migration
integration_bridge = br-int
tunnel_types = vxlan,geneve,gre
local_ip = LOCAL_IP
management_ip = MANAGEMENT_IP
enable_dpdk = False
enable_neutron_notifier = False
remote_db_hosts = DB_SERVER_IP
nb_db_class = etcd_nb_db_driver
auto_detect_port_behind_port = False
publisher_rate_limit_count = 1
publisher_rate_limit_timeout = 180
monitor_table_poll_time = 30
datapath_layout_path=/etc/dragonflow/dragonflow_datapath_layout.yaml
[df_l2_app]
l2_responder = True
[df_loadbalancer]
auto_enable_vip_ports = True
[df_metadata]
metadata_interface = tap-metadata
port = 18080
ip = 169.254.169.254
+ [df_ryu]
+ of_listen_address = LOCAL_IP | 2 | 0.060606 | 2 | 0 |
65cb681a26530e038c7f45ac26dbd2009d6c0e3b | components/media-bar/media-bar.html | components/media-bar/media-bar.html | <app-scroll-scroller
overlay
horizontal
class="section fill-darker media-bar"
:class="{
'-is-loading': !mediaItems || !mediaItems.length,
}"
>
<div class="-items">
<div class="-loading-container" v-if="!mediaItems || !mediaItems.length">
<app-loading centered no-color stationary hide-label />
</div>
<template v-if="mediaItems && mediaItems.length">
<app-media-bar-item
v-for="item of mediaItems"
:key="item.id"
:item="item"
/>
</template>
</div>
</app-scroll-scroller>
| <app-scroll-scroller overlay horizontal>
<div
class="section fill-darker media-bar"
:class="{
'-is-loading': !mediaItems || !mediaItems.length,
}"
>
<div class="-items">
<div class="-loading-container" v-if="!mediaItems || !mediaItems.length">
<app-loading centered no-color stationary hide-label />
</div>
<template v-if="mediaItems && mediaItems.length">
<app-media-bar-item
v-for="item of mediaItems"
:key="item.id"
:item="item"
/>
</template>
</div>
</div>
</app-scroll-scroller>
| Fix media bar with overlay scroll on chrome | Fix media bar with overlay scroll on chrome
| HTML | mit | gamejolt/frontend-lib,gamejolt/frontend-lib,gamejolt/frontend-lib,gamejolt/frontend-lib | html | ## Code Before:
<app-scroll-scroller
overlay
horizontal
class="section fill-darker media-bar"
:class="{
'-is-loading': !mediaItems || !mediaItems.length,
}"
>
<div class="-items">
<div class="-loading-container" v-if="!mediaItems || !mediaItems.length">
<app-loading centered no-color stationary hide-label />
</div>
<template v-if="mediaItems && mediaItems.length">
<app-media-bar-item
v-for="item of mediaItems"
:key="item.id"
:item="item"
/>
</template>
</div>
</app-scroll-scroller>
## Instruction:
Fix media bar with overlay scroll on chrome
## Code After:
<app-scroll-scroller overlay horizontal>
<div
class="section fill-darker media-bar"
:class="{
'-is-loading': !mediaItems || !mediaItems.length,
}"
>
<div class="-items">
<div class="-loading-container" v-if="!mediaItems || !mediaItems.length">
<app-loading centered no-color stationary hide-label />
</div>
<template v-if="mediaItems && mediaItems.length">
<app-media-bar-item
v-for="item of mediaItems"
:key="item.id"
:item="item"
/>
</template>
</div>
</div>
</app-scroll-scroller>
| + <app-scroll-scroller overlay horizontal>
+ <div
- <app-scroll-scroller
- overlay
- horizontal
- class="section fill-darker media-bar"
+ class="section fill-darker media-bar"
? +
- :class="{
+ :class="{
? +
- '-is-loading': !mediaItems || !mediaItems.length,
+ '-is-loading': !mediaItems || !mediaItems.length,
? +
- }"
+ }"
? +
- >
+ >
? +
- <div class="-items">
+ <div class="-items">
? +
- <div class="-loading-container" v-if="!mediaItems || !mediaItems.length">
+ <div class="-loading-container" v-if="!mediaItems || !mediaItems.length">
? +
- <app-loading centered no-color stationary hide-label />
+ <app-loading centered no-color stationary hide-label />
? +
+ </div>
+
+ <template v-if="mediaItems && mediaItems.length">
+ <app-media-bar-item
+ v-for="item of mediaItems"
+ :key="item.id"
+ :item="item"
+ />
+ </template>
+
</div>
-
- <template v-if="mediaItems && mediaItems.length">
- <app-media-bar-item
- v-for="item of mediaItems"
- :key="item.id"
- :item="item"
- />
- </template>
-
</div>
</app-scroll-scroller> | 40 | 1.666667 | 20 | 20 |
fe78f638842930e2d8b9271e2565b80c8d5e85f3 | lib/functions.zsh | lib/functions.zsh | function zsh_stats() {
history | awk '{CMD[$2]++;count++;}END { for (a in CMD)print CMD[a] " " CMD[a]/count*100 "% " a;}' | grep -v "./" | column -c3 -s " " -t | sort -nr | nl | head -n20
}
function uninstall_oh_my_zsh() {
/usr/bin/env ZSH=$ZSH /bin/sh $ZSH/tools/uninstall.sh
}
function upgrade_oh_my_zsh() {
/usr/bin/env ZSH=$ZSH /bin/sh $ZSH/tools/upgrade.sh
}
function take() {
mkdir -p $1
cd $1
}
| function zsh_stats() {
history | awk '{CMD[$2]++;count++;}END { for (a in CMD)print CMD[a] " " CMD[a]/count*100 "% " a;}' | grep -v "./" | column -c3 -s " " -t | sort -nr | nl | head -n20
}
function uninstall_oh_my_zsh() {
/usr/bin/env ZSH=$ZSH /bin/sh $ZSH/tools/uninstall.sh
}
function upgrade_oh_my_zsh() {
/usr/bin/env ZSH=$ZSH /bin/sh $ZSH/tools/upgrade.sh
}
function take() {
mkdir -p $1
cd $1
}
function swap () {
[ $# -ne 2 ] && echo "swap: 2 arguments needed" && return 1
[ ! -e $1 ] && echo "swap: $1 does not exit" && return 1
[ ! -e $2 ] && echo "swap: $2 does not exit" && return 1
local TMPNAME=tmp.$$
mv "$1" TMPNAME;
mv "$2" "$1"
mv TMPNAME "$2"
}
| Add a handy function to swap files | Add a handy function to swap files
| Shell | mit | toctan/oh-my-zsh,toctan/oh-my-zsh,toctan/oh-my-zsh | shell | ## Code Before:
function zsh_stats() {
history | awk '{CMD[$2]++;count++;}END { for (a in CMD)print CMD[a] " " CMD[a]/count*100 "% " a;}' | grep -v "./" | column -c3 -s " " -t | sort -nr | nl | head -n20
}
function uninstall_oh_my_zsh() {
/usr/bin/env ZSH=$ZSH /bin/sh $ZSH/tools/uninstall.sh
}
function upgrade_oh_my_zsh() {
/usr/bin/env ZSH=$ZSH /bin/sh $ZSH/tools/upgrade.sh
}
function take() {
mkdir -p $1
cd $1
}
## Instruction:
Add a handy function to swap files
## Code After:
function zsh_stats() {
history | awk '{CMD[$2]++;count++;}END { for (a in CMD)print CMD[a] " " CMD[a]/count*100 "% " a;}' | grep -v "./" | column -c3 -s " " -t | sort -nr | nl | head -n20
}
function uninstall_oh_my_zsh() {
/usr/bin/env ZSH=$ZSH /bin/sh $ZSH/tools/uninstall.sh
}
function upgrade_oh_my_zsh() {
/usr/bin/env ZSH=$ZSH /bin/sh $ZSH/tools/upgrade.sh
}
function take() {
mkdir -p $1
cd $1
}
function swap () {
[ $# -ne 2 ] && echo "swap: 2 arguments needed" && return 1
[ ! -e $1 ] && echo "swap: $1 does not exit" && return 1
[ ! -e $2 ] && echo "swap: $2 does not exit" && return 1
local TMPNAME=tmp.$$
mv "$1" TMPNAME;
mv "$2" "$1"
mv TMPNAME "$2"
}
| function zsh_stats() {
history | awk '{CMD[$2]++;count++;}END { for (a in CMD)print CMD[a] " " CMD[a]/count*100 "% " a;}' | grep -v "./" | column -c3 -s " " -t | sort -nr | nl | head -n20
}
function uninstall_oh_my_zsh() {
/usr/bin/env ZSH=$ZSH /bin/sh $ZSH/tools/uninstall.sh
}
function upgrade_oh_my_zsh() {
/usr/bin/env ZSH=$ZSH /bin/sh $ZSH/tools/upgrade.sh
}
function take() {
mkdir -p $1
cd $1
}
+ function swap () {
+ [ $# -ne 2 ] && echo "swap: 2 arguments needed" && return 1
+ [ ! -e $1 ] && echo "swap: $1 does not exit" && return 1
+ [ ! -e $2 ] && echo "swap: $2 does not exit" && return 1
+
+ local TMPNAME=tmp.$$
+ mv "$1" TMPNAME;
+ mv "$2" "$1"
+ mv TMPNAME "$2"
+ } | 10 | 0.588235 | 10 | 0 |
38ae02399bf4d6300dcb8f8661de26a46945999b | .travis.yml | .travis.yml | sudo: required
services:
- docker
#git:
# submodules: false
# need to replace SSH submodule URLs with HTTPS
#before_install:
# - sed -i 's/git@github.com:/https:\/\/github.com\//' .gitmodules
# - sed -i 's/git@github.com:/https:\/\/github.com\//' catkin_ws/.gitmodules
install: true
script:
- docker pull openhumanoids/oh-distro-base:1.0
- docker run openhumanoids/oh-distro-base "./build-oh-distro.sh"
| sudo: required
services:
- docker
#git:
# submodules: false
# need to replace SSH submodule URLs with HTTPS
#before_install:
# - sed -i 's/git@github.com:/https:\/\/github.com\//' .gitmodules
# - sed -i 's/git@github.com:/https:\/\/github.com\//' catkin_ws/.gitmodules
install: true
script:
- docker pull openhumanoids/oh-distro-base:latest
- docker run openhumanoids/oh-distro-base /bin/bash -c "git clone https://github.com/oh-dev/oh-distro.git && cd oh-distro && git checkout wolfgang-make-public-compilable && git config --global url.https://github.com/.insteadOf git@github.com: && echo '' >> ~/.ssh/known_hosts && ssh-keyscan -H github.com >> ~/.ssh/known_hosts && ssh-keyscan -H bitbucket.org >> ~/.ssh/known_hosts && sed -i 's/git@github.com:/https:\/\/github.com\//g' .gitmodules && sed -i 's/git@github.com:/https:\/\/github.com\//g' catkin_ws/.gitmodules && sed -i 's/ssh:\/\/git@github.com\//https:\/\/github.com\//g' software/externals/cmake/externals.cmake && git config --global user.email "buildserver@example" && git config --global user.name "build server" && git commit -am "module changes" && git submodule deinit catkin_ws/src/exotica-dev && git rm catkin_ws/src/exotica-dev && git submodule update --init --recursive && cd software && source ~/oh-distro/software/config/drc_environment.sh && cd externals && mkdir pod-build && cd pod-build && cmake .. -DBUILD_PRIVATE_EXTERNALS:BOOL=OFF && cd .. && make -j8 && cd .. && make -j8"
| Update docker run command and change to latest base image | Update docker run command and change to latest base image
| YAML | bsd-3-clause | openhumanoids/oh-distro,openhumanoids/oh-distro,openhumanoids/oh-distro,openhumanoids/oh-distro,openhumanoids/oh-distro,openhumanoids/oh-distro,openhumanoids/oh-distro,openhumanoids/oh-distro | yaml | ## Code Before:
sudo: required
services:
- docker
#git:
# submodules: false
# need to replace SSH submodule URLs with HTTPS
#before_install:
# - sed -i 's/git@github.com:/https:\/\/github.com\//' .gitmodules
# - sed -i 's/git@github.com:/https:\/\/github.com\//' catkin_ws/.gitmodules
install: true
script:
- docker pull openhumanoids/oh-distro-base:1.0
- docker run openhumanoids/oh-distro-base "./build-oh-distro.sh"
## Instruction:
Update docker run command and change to latest base image
## Code After:
sudo: required
services:
- docker
#git:
# submodules: false
# need to replace SSH submodule URLs with HTTPS
#before_install:
# - sed -i 's/git@github.com:/https:\/\/github.com\//' .gitmodules
# - sed -i 's/git@github.com:/https:\/\/github.com\//' catkin_ws/.gitmodules
install: true
script:
- docker pull openhumanoids/oh-distro-base:latest
- docker run openhumanoids/oh-distro-base /bin/bash -c "git clone https://github.com/oh-dev/oh-distro.git && cd oh-distro && git checkout wolfgang-make-public-compilable && git config --global url.https://github.com/.insteadOf git@github.com: && echo '' >> ~/.ssh/known_hosts && ssh-keyscan -H github.com >> ~/.ssh/known_hosts && ssh-keyscan -H bitbucket.org >> ~/.ssh/known_hosts && sed -i 's/git@github.com:/https:\/\/github.com\//g' .gitmodules && sed -i 's/git@github.com:/https:\/\/github.com\//g' catkin_ws/.gitmodules && sed -i 's/ssh:\/\/git@github.com\//https:\/\/github.com\//g' software/externals/cmake/externals.cmake && git config --global user.email "buildserver@example" && git config --global user.name "build server" && git commit -am "module changes" && git submodule deinit catkin_ws/src/exotica-dev && git rm catkin_ws/src/exotica-dev && git submodule update --init --recursive && cd software && source ~/oh-distro/software/config/drc_environment.sh && cd externals && mkdir pod-build && cd pod-build && cmake .. -DBUILD_PRIVATE_EXTERNALS:BOOL=OFF && cd .. && make -j8 && cd .. && make -j8"
| sudo: required
services:
- docker
#git:
# submodules: false
# need to replace SSH submodule URLs with HTTPS
#before_install:
# - sed -i 's/git@github.com:/https:\/\/github.com\//' .gitmodules
# - sed -i 's/git@github.com:/https:\/\/github.com\//' catkin_ws/.gitmodules
install: true
script:
- - docker pull openhumanoids/oh-distro-base:1.0
? ^^^
+ - docker pull openhumanoids/oh-distro-base:latest
? ^^^^^^
- - docker run openhumanoids/oh-distro-base "./build-oh-distro.sh"
+ - docker run openhumanoids/oh-distro-base /bin/bash -c "git clone https://github.com/oh-dev/oh-distro.git && cd oh-distro && git checkout wolfgang-make-public-compilable && git config --global url.https://github.com/.insteadOf git@github.com: && echo '' >> ~/.ssh/known_hosts && ssh-keyscan -H github.com >> ~/.ssh/known_hosts && ssh-keyscan -H bitbucket.org >> ~/.ssh/known_hosts && sed -i 's/git@github.com:/https:\/\/github.com\//g' .gitmodules && sed -i 's/git@github.com:/https:\/\/github.com\//g' catkin_ws/.gitmodules && sed -i 's/ssh:\/\/git@github.com\//https:\/\/github.com\//g' software/externals/cmake/externals.cmake && git config --global user.email "buildserver@example" && git config --global user.name "build server" && git commit -am "module changes" && git submodule deinit catkin_ws/src/exotica-dev && git rm catkin_ws/src/exotica-dev && git submodule update --init --recursive && cd software && source ~/oh-distro/software/config/drc_environment.sh && cd externals && mkdir pod-build && cd pod-build && cmake .. -DBUILD_PRIVATE_EXTERNALS:BOOL=OFF && cd .. && make -j8 && cd .. && make -j8" | 4 | 0.222222 | 2 | 2 |
30ed04ece86cd5b5505da7a792e06f8fd51a9a62 | src/oldie.js | src/oldie.js | // this code is from developer.mozilla.org
// adds the forEach array method if it doesn't exist
if ( !Array.prototype.forEach )
{
Array.prototype.forEach = function( fn, scope )
{
"use strict";
for(var i = 0, len = this.length; i < len; ++i) {
fn.call(scope, this[i], i, this);
}
}
}
// adds the filter array method if it doesn't exist
if (!Array.prototype.filter)
{
Array.prototype.filter = function(fun /*, thisp*/)
{
"use strict";
if (this == null)
throw new TypeError();
var t = Object(this);
var len = t.length >>> 0;
if (typeof fun != "function")
throw new TypeError();
var res = [];
var thisp = arguments[1];
for (var i = 0; i < len; i++)
{
if (i in t)
{
var val = t[i]; // in case fun mutates this
if (fun.call(thisp, val, i, t))
res.push(val);
}
}
return res;
};
}
| // this code is from developer.mozilla.org
// adds the forEach array method if it doesn't exist
if ( !Array.prototype.forEach )
{
Array.prototype.forEach = function( fn, scope )
{
"use strict";
for(var i = 0, len = this.length; i < len; ++i) {
fn.call(scope, this[i], i, this);
}
}
}
// adds the filter array method if it doesn't exist
if (!Array.prototype.filter)
{
Array.prototype.filter = function(fun /*, thisp*/)
{
"use strict";
if (this == null)
throw new TypeError();
var t = Object(this);
var len = t.length >>> 0;
if (typeof fun != "function")
throw new TypeError();
var res = [];
var thisp = arguments[1];
for (var i = 0; i < len; i++)
{
if (i in t)
{
var val = t[i]; // in case fun mutates this
if (fun.call(thisp, val, i, t))
res.push(val);
}
}
return res;
};
}
if( typeof console === 'undefined' )
{
console = { log: function( string ){ }, assert: function( condition, message ) { } };
}
| Create a dummy console for browsers without one | Create a dummy console for browsers without one
Fixes issue with IE8 under Win7
| JavaScript | mit | steelbreeze/state.js,steelbreeze/state.js | javascript | ## Code Before:
// this code is from developer.mozilla.org
// adds the forEach array method if it doesn't exist
if ( !Array.prototype.forEach )
{
Array.prototype.forEach = function( fn, scope )
{
"use strict";
for(var i = 0, len = this.length; i < len; ++i) {
fn.call(scope, this[i], i, this);
}
}
}
// adds the filter array method if it doesn't exist
if (!Array.prototype.filter)
{
Array.prototype.filter = function(fun /*, thisp*/)
{
"use strict";
if (this == null)
throw new TypeError();
var t = Object(this);
var len = t.length >>> 0;
if (typeof fun != "function")
throw new TypeError();
var res = [];
var thisp = arguments[1];
for (var i = 0; i < len; i++)
{
if (i in t)
{
var val = t[i]; // in case fun mutates this
if (fun.call(thisp, val, i, t))
res.push(val);
}
}
return res;
};
}
## Instruction:
Create a dummy console for browsers without one
Fixes issue with IE8 under Win7
## Code After:
// this code is from developer.mozilla.org
// adds the forEach array method if it doesn't exist
if ( !Array.prototype.forEach )
{
Array.prototype.forEach = function( fn, scope )
{
"use strict";
for(var i = 0, len = this.length; i < len; ++i) {
fn.call(scope, this[i], i, this);
}
}
}
// adds the filter array method if it doesn't exist
if (!Array.prototype.filter)
{
Array.prototype.filter = function(fun /*, thisp*/)
{
"use strict";
if (this == null)
throw new TypeError();
var t = Object(this);
var len = t.length >>> 0;
if (typeof fun != "function")
throw new TypeError();
var res = [];
var thisp = arguments[1];
for (var i = 0; i < len; i++)
{
if (i in t)
{
var val = t[i]; // in case fun mutates this
if (fun.call(thisp, val, i, t))
res.push(val);
}
}
return res;
};
}
if( typeof console === 'undefined' )
{
console = { log: function( string ){ }, assert: function( condition, message ) { } };
}
| // this code is from developer.mozilla.org
// adds the forEach array method if it doesn't exist
if ( !Array.prototype.forEach )
{
Array.prototype.forEach = function( fn, scope )
{
"use strict";
for(var i = 0, len = this.length; i < len; ++i) {
fn.call(scope, this[i], i, this);
}
}
}
// adds the filter array method if it doesn't exist
if (!Array.prototype.filter)
{
Array.prototype.filter = function(fun /*, thisp*/)
{
"use strict";
if (this == null)
throw new TypeError();
var t = Object(this);
var len = t.length >>> 0;
if (typeof fun != "function")
throw new TypeError();
var res = [];
var thisp = arguments[1];
for (var i = 0; i < len; i++)
{
if (i in t)
{
var val = t[i]; // in case fun mutates this
if (fun.call(thisp, val, i, t))
res.push(val);
}
}
return res;
};
}
+
+ if( typeof console === 'undefined' )
+ {
+ console = { log: function( string ){ }, assert: function( condition, message ) { } };
+ } | 5 | 0.111111 | 5 | 0 |
f68919331a8d88beabf3debc36bf8058afb3e17f | src/test/java/code/donbonifacio/saft/ResultTest.java | src/test/java/code/donbonifacio/saft/ResultTest.java | package code.donbonifacio.saft;
import junit.framework.TestCase;
/**
* Tests for the Result class
*/
public class ResultTest extends TestCase {
/**
* Tests the Result.success singleton
*/
public void testSuccessSingleton() {
Result r1 = Result.success();
assertTrue(r1.isSucceeded());
assertFalse(r1.isFailed());
Result r2 = Result.success();
assertTrue(r2.isSucceeded());
assertFalse(r2.isFailed());
assertTrue(r1 == r2);
}
/**
* Tests a failed result via exception
*/
public void testResultException() {
Exception ex = new RuntimeException();
Result result = Result.exception(ex);
assertFalse(result.isSucceeded());
assertTrue(result.isException());
}
}
| package code.donbonifacio.saft;
import junit.framework.TestCase;
import java.util.Arrays;
/**
* Tests for the Result class
*/
public class ResultTest extends TestCase {
/**
* Tests the Result.success singleton
*/
public void testSuccessSingleton() {
Result r1 = Result.success();
assertTrue(r1.isSucceeded());
assertFalse(r1.isFailed());
Result r2 = Result.success();
assertTrue(r2.isSucceeded());
assertFalse(r2.isFailed());
assertTrue(r1 == r2);
}
/**
* Tests a failed result via exception
*/
public void testResultException() {
Exception ex = new RuntimeException();
Result result = Result.exception(ex);
assertFalse(result.isSucceeded());
assertTrue(result.isException());
}
/**
* Test toString generation
*/
public void testToString() {
Result r1 = Result.failure("Test1");
Result r2 = Result.failure("Test2");
Result r3 = Result.fromResults(Arrays.asList(r1, r2));
assertEquals("Result{succeeded=false, reason='From list of results', results=[Result{succeeded=false, reason='Test1'}, Result{succeeded=false, reason='Test2'}, ]}", r3.toString());
}
/**
* If fromResults gets a collection with only one result, return
* that result
*/
public void testFromResultsAsOne() {
Result r1 = Result.success();
Result r2 = Result.fromResults(r1.asList());
assertTrue(r1 == r2);
}
}
| Add more tests to Result | Add more tests to Result
| Java | mit | donbonifacio/saft-utils | java | ## Code Before:
package code.donbonifacio.saft;
import junit.framework.TestCase;
/**
* Tests for the Result class
*/
public class ResultTest extends TestCase {
/**
* Tests the Result.success singleton
*/
public void testSuccessSingleton() {
Result r1 = Result.success();
assertTrue(r1.isSucceeded());
assertFalse(r1.isFailed());
Result r2 = Result.success();
assertTrue(r2.isSucceeded());
assertFalse(r2.isFailed());
assertTrue(r1 == r2);
}
/**
* Tests a failed result via exception
*/
public void testResultException() {
Exception ex = new RuntimeException();
Result result = Result.exception(ex);
assertFalse(result.isSucceeded());
assertTrue(result.isException());
}
}
## Instruction:
Add more tests to Result
## Code After:
package code.donbonifacio.saft;
import junit.framework.TestCase;
import java.util.Arrays;
/**
* Tests for the Result class
*/
public class ResultTest extends TestCase {
/**
* Tests the Result.success singleton
*/
public void testSuccessSingleton() {
Result r1 = Result.success();
assertTrue(r1.isSucceeded());
assertFalse(r1.isFailed());
Result r2 = Result.success();
assertTrue(r2.isSucceeded());
assertFalse(r2.isFailed());
assertTrue(r1 == r2);
}
/**
* Tests a failed result via exception
*/
public void testResultException() {
Exception ex = new RuntimeException();
Result result = Result.exception(ex);
assertFalse(result.isSucceeded());
assertTrue(result.isException());
}
/**
* Test toString generation
*/
public void testToString() {
Result r1 = Result.failure("Test1");
Result r2 = Result.failure("Test2");
Result r3 = Result.fromResults(Arrays.asList(r1, r2));
assertEquals("Result{succeeded=false, reason='From list of results', results=[Result{succeeded=false, reason='Test1'}, Result{succeeded=false, reason='Test2'}, ]}", r3.toString());
}
/**
* If fromResults gets a collection with only one result, return
* that result
*/
public void testFromResultsAsOne() {
Result r1 = Result.success();
Result r2 = Result.fromResults(r1.asList());
assertTrue(r1 == r2);
}
}
| package code.donbonifacio.saft;
import junit.framework.TestCase;
+
+ import java.util.Arrays;
/**
* Tests for the Result class
*/
public class ResultTest extends TestCase {
/**
* Tests the Result.success singleton
*/
public void testSuccessSingleton() {
Result r1 = Result.success();
assertTrue(r1.isSucceeded());
assertFalse(r1.isFailed());
Result r2 = Result.success();
assertTrue(r2.isSucceeded());
assertFalse(r2.isFailed());
assertTrue(r1 == r2);
}
/**
* Tests a failed result via exception
*/
public void testResultException() {
Exception ex = new RuntimeException();
Result result = Result.exception(ex);
assertFalse(result.isSucceeded());
assertTrue(result.isException());
}
+
+ /**
+ * Test toString generation
+ */
+ public void testToString() {
+ Result r1 = Result.failure("Test1");
+ Result r2 = Result.failure("Test2");
+ Result r3 = Result.fromResults(Arrays.asList(r1, r2));
+
+ assertEquals("Result{succeeded=false, reason='From list of results', results=[Result{succeeded=false, reason='Test1'}, Result{succeeded=false, reason='Test2'}, ]}", r3.toString());
+ }
+
+ /**
+ * If fromResults gets a collection with only one result, return
+ * that result
+ */
+ public void testFromResultsAsOne() {
+ Result r1 = Result.success();
+ Result r2 = Result.fromResults(r1.asList());
+
+ assertTrue(r1 == r2);
+
+ }
} | 25 | 0.735294 | 25 | 0 |
3c5aa921ab8d218f13d8479cef35c2010bbca7c9 | composer.json | composer.json | {
"name": "petercoles/countries",
"type": "library",
"description": "Lookup Objects for ISO 3166 Country Names and Codes",
"keywords": ["country lists", "country lookup", "country codes", "ISO 3166", "laravel"],
"homepage": "https://github.com/petercoles/countries",
"license": "MIT",
"authors": [
{
"name": "Peter Coles",
"email": "peterdcoles@gmail.com",
"homepage": "http://petercoles.com"
}
],
"support": {
"issues": "https://github.com/petercoles/countries/issues"
},
"repositories": [
{
"type": "vcs",
"url": "https://github.com/umpirsky/country-list.git"
}
],
"require": {
"php": ">=5.6.0",
"illuminate/support": "~5",
"umpirsky/country-list": "@dev"
},
"require-dev": {
"phpunit/phpunit": "~5.0",
"mockery/mockery": "^0.9.4"
},
"autoload": {
"psr-4": {
"PeterColes\\Countries\\": "src/"
}
}
}
| {
"name": "petercoles/multilingual-country-list",
"type": "library",
"description": "Lookup Objects for ISO 3166 Country Names and Codes",
"keywords": ["country lists", "country lookup", "country codes", "ISO 3166", "laravel"],
"homepage": "https://github.com/petercoles/countries",
"license": "MIT",
"authors": [
{
"name": "Peter Coles",
"email": "peterdcoles@gmail.com",
"homepage": "http://petercoles.com"
}
],
"support": {
"issues": "https://github.com/petercoles/multilingual-country-list/issues"
},
"repositories": [
{
"type": "vcs",
"url": "https://github.com/petercoles/multilingual-country-list.git"
}
],
"require": {
"php": ">=5.6.0",
"illuminate/support": "~5",
"umpirsky/country-list": "@dev"
},
"require-dev": {
"phpunit/phpunit": "~5.0",
"mockery/mockery": "^0.9.4"
},
"autoload": {
"psr-4": {
"PeterColes\\Countries\\": "src/"
}
}
}
| Align internal package references with the repo name | Align internal package references with the repo name
| JSON | mit | petercoles/Multilingual-Country-List | json | ## Code Before:
{
"name": "petercoles/countries",
"type": "library",
"description": "Lookup Objects for ISO 3166 Country Names and Codes",
"keywords": ["country lists", "country lookup", "country codes", "ISO 3166", "laravel"],
"homepage": "https://github.com/petercoles/countries",
"license": "MIT",
"authors": [
{
"name": "Peter Coles",
"email": "peterdcoles@gmail.com",
"homepage": "http://petercoles.com"
}
],
"support": {
"issues": "https://github.com/petercoles/countries/issues"
},
"repositories": [
{
"type": "vcs",
"url": "https://github.com/umpirsky/country-list.git"
}
],
"require": {
"php": ">=5.6.0",
"illuminate/support": "~5",
"umpirsky/country-list": "@dev"
},
"require-dev": {
"phpunit/phpunit": "~5.0",
"mockery/mockery": "^0.9.4"
},
"autoload": {
"psr-4": {
"PeterColes\\Countries\\": "src/"
}
}
}
## Instruction:
Align internal package references with the repo name
## Code After:
{
"name": "petercoles/multilingual-country-list",
"type": "library",
"description": "Lookup Objects for ISO 3166 Country Names and Codes",
"keywords": ["country lists", "country lookup", "country codes", "ISO 3166", "laravel"],
"homepage": "https://github.com/petercoles/countries",
"license": "MIT",
"authors": [
{
"name": "Peter Coles",
"email": "peterdcoles@gmail.com",
"homepage": "http://petercoles.com"
}
],
"support": {
"issues": "https://github.com/petercoles/multilingual-country-list/issues"
},
"repositories": [
{
"type": "vcs",
"url": "https://github.com/petercoles/multilingual-country-list.git"
}
],
"require": {
"php": ">=5.6.0",
"illuminate/support": "~5",
"umpirsky/country-list": "@dev"
},
"require-dev": {
"phpunit/phpunit": "~5.0",
"mockery/mockery": "^0.9.4"
},
"autoload": {
"psr-4": {
"PeterColes\\Countries\\": "src/"
}
}
}
| {
- "name": "petercoles/countries",
? -
+ "name": "petercoles/multilingual-country-list",
? +++++++++++++ +++ +
"type": "library",
"description": "Lookup Objects for ISO 3166 Country Names and Codes",
"keywords": ["country lists", "country lookup", "country codes", "ISO 3166", "laravel"],
"homepage": "https://github.com/petercoles/countries",
"license": "MIT",
"authors": [
{
"name": "Peter Coles",
"email": "peterdcoles@gmail.com",
"homepage": "http://petercoles.com"
}
],
"support": {
- "issues": "https://github.com/petercoles/countries/issues"
? -
+ "issues": "https://github.com/petercoles/multilingual-country-list/issues"
? +++++++++++++ +++ +
},
"repositories": [
{
"type": "vcs",
- "url": "https://github.com/umpirsky/country-list.git"
? ^^ ^^^^^
+ "url": "https://github.com/petercoles/multilingual-country-list.git"
? ++++++++++++ ^^ ^^^^^^^^
}
],
"require": {
"php": ">=5.6.0",
"illuminate/support": "~5",
"umpirsky/country-list": "@dev"
},
"require-dev": {
"phpunit/phpunit": "~5.0",
"mockery/mockery": "^0.9.4"
},
"autoload": {
"psr-4": {
"PeterColes\\Countries\\": "src/"
}
}
} | 6 | 0.157895 | 3 | 3 |
789e04879e362ced05ee7b4a1ab534004e7291f2 | requirements-common.txt | requirements-common.txt | Django==1.5.5
Markdown==2.4
Pygments==1.6
South==0.8.4
django-generic-aggregation==0.3.2
django-haystack==2.1.0
django-pagination==1.0.7
django-registration==1.0
django-taggit==0.11.2
-e git+https://github.com/django-de/cab#egg=cab
-e git+https://github.com/jezdez/django-simple-ratings#egg=django-simple-ratings
| Django==1.5.5
Markdown==2.4
Pygments==1.6
South==0.8.4
django-generic-aggregation==0.3.2
django-haystack==2.1.0
django-pagination==1.0.7
django-registration==1.0
django-taggit==0.11.2
django-cab==0.3
django-simple-ratings==0.3.3
| Use released versions of own packages. | Use released versions of own packages.
| Text | bsd-3-clause | django/djangosnippets.org,django-de/djangosnippets.org,django-de/djangosnippets.org,django/djangosnippets.org,django-de/djangosnippets.org,django/djangosnippets.org,django-de/djangosnippets.org,django/djangosnippets.org,django/djangosnippets.org | text | ## Code Before:
Django==1.5.5
Markdown==2.4
Pygments==1.6
South==0.8.4
django-generic-aggregation==0.3.2
django-haystack==2.1.0
django-pagination==1.0.7
django-registration==1.0
django-taggit==0.11.2
-e git+https://github.com/django-de/cab#egg=cab
-e git+https://github.com/jezdez/django-simple-ratings#egg=django-simple-ratings
## Instruction:
Use released versions of own packages.
## Code After:
Django==1.5.5
Markdown==2.4
Pygments==1.6
South==0.8.4
django-generic-aggregation==0.3.2
django-haystack==2.1.0
django-pagination==1.0.7
django-registration==1.0
django-taggit==0.11.2
django-cab==0.3
django-simple-ratings==0.3.3
| Django==1.5.5
Markdown==2.4
Pygments==1.6
South==0.8.4
django-generic-aggregation==0.3.2
django-haystack==2.1.0
django-pagination==1.0.7
django-registration==1.0
django-taggit==0.11.2
+ django-cab==0.3
+ django-simple-ratings==0.3.3
-
- -e git+https://github.com/django-de/cab#egg=cab
- -e git+https://github.com/jezdez/django-simple-ratings#egg=django-simple-ratings | 5 | 0.416667 | 2 | 3 |
2dd4a0bec817abeab78ed92f63d40f6fb5b1b46a | backend/app/views/spree/admin/promotion_categories/_form.html.erb | backend/app/views/spree/admin/promotion_categories/_form.html.erb | <%= render :partial => 'spree/shared/error_messages', :locals => { :target => @promotion } %>
<div class='row'>
<div class='alpha six columns'>
<%= f.field_container :name do %>
<%= f.label :name %>
<%= f.text_field :name, :class => 'fullwidth' %>
<% end %>
<%= f.field_container :code do %>
<%= f.label :code %>
<%= f.text_field :code, :class => 'fullwidth' %>
<% end %>
</div>
</div>
| <%= render :partial => 'spree/shared/error_messages', :locals => { :target => @promotion } %>
<div class='row'>
<div class='col-xs-4'>
<%= f.field_container :name do %>
<%= f.label :name %>
<%= f.text_field :name, :class => 'fullwidth' %>
<% end %>
<%= f.field_container :code do %>
<%= f.label :code %>
<%= f.text_field :code, :class => 'fullwidth' %>
<% end %>
</div>
</div>
| Change skeleton to bootstrap in promotion cat. view | Change skeleton to bootstrap in promotion cat. view
| HTML+ERB | bsd-3-clause | Arpsara/solidus,pervino/solidus,pervino/solidus,Arpsara/solidus,Arpsara/solidus,pervino/solidus,pervino/solidus,Arpsara/solidus | html+erb | ## Code Before:
<%= render :partial => 'spree/shared/error_messages', :locals => { :target => @promotion } %>
<div class='row'>
<div class='alpha six columns'>
<%= f.field_container :name do %>
<%= f.label :name %>
<%= f.text_field :name, :class => 'fullwidth' %>
<% end %>
<%= f.field_container :code do %>
<%= f.label :code %>
<%= f.text_field :code, :class => 'fullwidth' %>
<% end %>
</div>
</div>
## Instruction:
Change skeleton to bootstrap in promotion cat. view
## Code After:
<%= render :partial => 'spree/shared/error_messages', :locals => { :target => @promotion } %>
<div class='row'>
<div class='col-xs-4'>
<%= f.field_container :name do %>
<%= f.label :name %>
<%= f.text_field :name, :class => 'fullwidth' %>
<% end %>
<%= f.field_container :code do %>
<%= f.label :code %>
<%= f.text_field :code, :class => 'fullwidth' %>
<% end %>
</div>
</div>
| <%= render :partial => 'spree/shared/error_messages', :locals => { :target => @promotion } %>
<div class='row'>
- <div class='alpha six columns'>
+ <div class='col-xs-4'>
<%= f.field_container :name do %>
<%= f.label :name %>
<%= f.text_field :name, :class => 'fullwidth' %>
<% end %>
<%= f.field_container :code do %>
<%= f.label :code %>
<%= f.text_field :code, :class => 'fullwidth' %>
<% end %>
</div>
</div> | 2 | 0.142857 | 1 | 1 |
59e338feba942f6c094c67ba390e5210f81f274c | src/webvowl/js/elements/nodes/implementations/Hexagon.js | src/webvowl/js/elements/nodes/implementations/Hexagon.js | var HexagonNode = require("../HexagonNode");
module.exports = (function () {
var o = function (graph) {
HexagonNode.apply(this, arguments);
this.attributes(["hexagon"])
.type("Hexagon");
};
o.prototype = Object.create(HexagonNode.prototype);
o.prototype.constructor = o;
return o;
}());
| var HexagonNode = require("../HexagonNode");
module.exports = (function () {
var o = function (graph) {
HexagonNode.apply(this, arguments);
this.type("Rule");
};
o.prototype = Object.create(HexagonNode.prototype);
o.prototype.constructor = o;
return o;
}());
| Change hexagon type to rule | Change hexagon type to rule
| JavaScript | mit | leshek-pawlak/WebVOWL,leshek-pawlak/WebVOWL | javascript | ## Code Before:
var HexagonNode = require("../HexagonNode");
module.exports = (function () {
var o = function (graph) {
HexagonNode.apply(this, arguments);
this.attributes(["hexagon"])
.type("Hexagon");
};
o.prototype = Object.create(HexagonNode.prototype);
o.prototype.constructor = o;
return o;
}());
## Instruction:
Change hexagon type to rule
## Code After:
var HexagonNode = require("../HexagonNode");
module.exports = (function () {
var o = function (graph) {
HexagonNode.apply(this, arguments);
this.type("Rule");
};
o.prototype = Object.create(HexagonNode.prototype);
o.prototype.constructor = o;
return o;
}());
| var HexagonNode = require("../HexagonNode");
module.exports = (function () {
var o = function (graph) {
HexagonNode.apply(this, arguments);
+ this.type("Rule");
- this.attributes(["hexagon"])
- .type("Hexagon");
};
o.prototype = Object.create(HexagonNode.prototype);
o.prototype.constructor = o;
return o;
}()); | 3 | 0.2 | 1 | 2 |
be60e2408e8b4ccbfc33cd8c255182eb06af057f | src/main/java/org/codeswarm/lipsum/STGroupParagraphGenerator.java | src/main/java/org/codeswarm/lipsum/STGroupParagraphGenerator.java | package org.codeswarm.lipsum;
import org.stringtemplate.v4.ST;
import org.stringtemplate.v4.STGroup;
/**
* A {@link org.codeswarm.lipsum.Lipsum.ParagraphGenerator} implementation backed by an {@link STGroup}.
*/
class STGroupParagraphGenerator implements Lipsum.ParagraphGenerator {
interface TemplateNames {
int getMinIndex();
int getMaxIndex();
String getTemplateName(int index);
}
private final STGroup stg;
private final TemplateNames templateNames;
STGroupParagraphGenerator(STGroup stg, TemplateNames templateNames) {
this.stg = stg;
this.templateNames = templateNames;
}
@Override
public String paragraph(long key) {
int index = templateNames.getMinIndex() + Lipsum.remainder(
Math.abs(key - templateNames.getMinIndex()),
templateNames.getMaxIndex());
String tplName = templateNames.getTemplateName(index);
ST tpl = stg.getInstanceOf(tplName);
return tpl.render();
}
}
| package org.codeswarm.lipsum;
import org.stringtemplate.v4.ST;
import org.stringtemplate.v4.STGroup;
/**
* A {@link Lipsum.ParagraphGenerator} implementation backed by an {@link STGroup}.
*/
class STGroupParagraphGenerator implements Lipsum.ParagraphGenerator {
interface TemplateNames {
int getMinIndex();
int getMaxIndex();
String getTemplateName(int index);
}
private final STGroup stg;
private final TemplateNames templateNames;
STGroupParagraphGenerator(STGroup stg, TemplateNames templateNames) {
this.stg = stg;
this.templateNames = templateNames;
}
@Override
public String paragraph(long key) {
int index = templateNames.getMinIndex() + Lipsum.remainder(
Math.abs(key - templateNames.getMinIndex()),
templateNames.getMaxIndex());
String tplName = templateNames.getTemplateName(index);
ST tpl = stg.getInstanceOf(tplName);
return tpl.render();
}
}
| Remove unnecessary full class qualification | Remove unnecessary full class qualification
| Java | apache-2.0 | chris-martin/lipsum | java | ## Code Before:
package org.codeswarm.lipsum;
import org.stringtemplate.v4.ST;
import org.stringtemplate.v4.STGroup;
/**
* A {@link org.codeswarm.lipsum.Lipsum.ParagraphGenerator} implementation backed by an {@link STGroup}.
*/
class STGroupParagraphGenerator implements Lipsum.ParagraphGenerator {
interface TemplateNames {
int getMinIndex();
int getMaxIndex();
String getTemplateName(int index);
}
private final STGroup stg;
private final TemplateNames templateNames;
STGroupParagraphGenerator(STGroup stg, TemplateNames templateNames) {
this.stg = stg;
this.templateNames = templateNames;
}
@Override
public String paragraph(long key) {
int index = templateNames.getMinIndex() + Lipsum.remainder(
Math.abs(key - templateNames.getMinIndex()),
templateNames.getMaxIndex());
String tplName = templateNames.getTemplateName(index);
ST tpl = stg.getInstanceOf(tplName);
return tpl.render();
}
}
## Instruction:
Remove unnecessary full class qualification
## Code After:
package org.codeswarm.lipsum;
import org.stringtemplate.v4.ST;
import org.stringtemplate.v4.STGroup;
/**
* A {@link Lipsum.ParagraphGenerator} implementation backed by an {@link STGroup}.
*/
class STGroupParagraphGenerator implements Lipsum.ParagraphGenerator {
interface TemplateNames {
int getMinIndex();
int getMaxIndex();
String getTemplateName(int index);
}
private final STGroup stg;
private final TemplateNames templateNames;
STGroupParagraphGenerator(STGroup stg, TemplateNames templateNames) {
this.stg = stg;
this.templateNames = templateNames;
}
@Override
public String paragraph(long key) {
int index = templateNames.getMinIndex() + Lipsum.remainder(
Math.abs(key - templateNames.getMinIndex()),
templateNames.getMaxIndex());
String tplName = templateNames.getTemplateName(index);
ST tpl = stg.getInstanceOf(tplName);
return tpl.render();
}
}
| package org.codeswarm.lipsum;
import org.stringtemplate.v4.ST;
import org.stringtemplate.v4.STGroup;
/**
- * A {@link org.codeswarm.lipsum.Lipsum.ParagraphGenerator} implementation backed by an {@link STGroup}.
? ---------------------
+ * A {@link Lipsum.ParagraphGenerator} implementation backed by an {@link STGroup}.
*/
class STGroupParagraphGenerator implements Lipsum.ParagraphGenerator {
interface TemplateNames {
int getMinIndex();
int getMaxIndex();
String getTemplateName(int index);
}
private final STGroup stg;
private final TemplateNames templateNames;
STGroupParagraphGenerator(STGroup stg, TemplateNames templateNames) {
this.stg = stg;
this.templateNames = templateNames;
}
@Override
public String paragraph(long key) {
int index = templateNames.getMinIndex() + Lipsum.remainder(
Math.abs(key - templateNames.getMinIndex()),
templateNames.getMaxIndex());
String tplName = templateNames.getTemplateName(index);
ST tpl = stg.getInstanceOf(tplName);
return tpl.render();
}
} | 2 | 0.057143 | 1 | 1 |
2a528e4492e2284641e8c164dda5aba801f2153a | cmd/sqlite-squish/main.go | cmd/sqlite-squish/main.go | package main
import "fmt"
func main() {
fmt.Println("Hello, world!")
}
| package main
import (
"bufio"
"fmt"
"log"
"os"
"github.com/griffithsh/sql-squish/database"
)
func main() {
scanner := bufio.NewScanner(os.Stdin)
var concatted string
for scanner.Scan() {
concatted = concatted + scanner.Text()
}
d, err := database.FromString(concatted)
if err != nil {
log.Fatal(err)
}
for _, t := range d.Tables {
fmt.Println(t.String())
}
// files, err := db.AsSQL()
// if err != nil {
// log.Fatal(err)
// }
// for file := range files {
// fmt.Print(file)
// }
}
| Make sqlite-squish accept SQL on stdin | Make sqlite-squish accept SQL on stdin
| Go | mit | griffithsh/sqlite-squish | go | ## Code Before:
package main
import "fmt"
func main() {
fmt.Println("Hello, world!")
}
## Instruction:
Make sqlite-squish accept SQL on stdin
## Code After:
package main
import (
"bufio"
"fmt"
"log"
"os"
"github.com/griffithsh/sql-squish/database"
)
func main() {
scanner := bufio.NewScanner(os.Stdin)
var concatted string
for scanner.Scan() {
concatted = concatted + scanner.Text()
}
d, err := database.FromString(concatted)
if err != nil {
log.Fatal(err)
}
for _, t := range d.Tables {
fmt.Println(t.String())
}
// files, err := db.AsSQL()
// if err != nil {
// log.Fatal(err)
// }
// for file := range files {
// fmt.Print(file)
// }
}
| package main
- import "fmt"
+ import (
+ "bufio"
+ "fmt"
+ "log"
+ "os"
+
+ "github.com/griffithsh/sql-squish/database"
+ )
func main() {
- fmt.Println("Hello, world!")
+ scanner := bufio.NewScanner(os.Stdin)
+
+ var concatted string
+ for scanner.Scan() {
+ concatted = concatted + scanner.Text()
+ }
+ d, err := database.FromString(concatted)
+ if err != nil {
+ log.Fatal(err)
+ }
+ for _, t := range d.Tables {
+ fmt.Println(t.String())
+ }
+ // files, err := db.AsSQL()
+ // if err != nil {
+ // log.Fatal(err)
+ // }
+ // for file := range files {
+ // fmt.Print(file)
+ // }
} | 30 | 4.285714 | 28 | 2 |
67004518d28211c0111284bb32bb04eb4698d3b0 | app/controllers/experiments/fall2017_cmu_experiment_controller.rb | app/controllers/experiments/fall2017_cmu_experiment_controller.rb |
require "#{Rails.root}/lib/experiments/fall2017_cmu_experiment"
module Experiments
class Fall2017CmuExperimentController < ApplicationController
before_action :set_course_and_check_email_code, only: %i[opt_in opt_out]
before_action :require_admin_permissions, only: :course_list
def opt_in
Fall2017CmuExperiment.new(@course).opt_in
flash[:notice] = 'Thanks for opting in!'
redirect_to "/courses/#{@course.slug}"
end
def opt_out
Fall2017CmuExperiment.new(@course).opt_out
flash[:notice] = 'Okay, you have opted out.'
redirect_to "/courses/#{@course.slug}"
end
def course_list
send_data Fall2017CmuExperiment.course_list,
filename: "fall2017-cmu-experiment-courses-#{Time.zone.today}.csv"
end
private
def set_course_and_check_email_code
@course = Course.find(params[:course_id])
code = @course.flags[Fall2017CmuExperiment::EMAIL_CODE]
raise IncorrectEmailCodeError if code.nil?
raise IncorrectEmailCodeError unless code == params[:email_code]
end
end
class IncorrectEmailCodeError < StandardError; end
end
|
require "#{Rails.root}/lib/experiments/fall2017_cmu_experiment"
module Experiments
class Fall2017CmuExperimentController < ApplicationController
before_action :set_course_and_check_email_code, only: %i[opt_in opt_out]
before_action :require_admin_permissions, only: :course_list
def opt_in
Fall2017CmuExperiment.new(@course).opt_in
flash[:notice] = 'Thank you for opting in. We will add the video sessions to the relevant assignments on the timeline for your WikiEd dashboard and send you an email about next steps.'
redirect_to "/courses/#{@course.slug}"
end
def opt_out
Fall2017CmuExperiment.new(@course).opt_out
flash[:notice] = 'Okay, you have opted out.'
redirect_to "/courses/#{@course.slug}"
end
def course_list
send_data Fall2017CmuExperiment.course_list,
filename: "fall2017-cmu-experiment-courses-#{Time.zone.today}.csv"
end
private
def set_course_and_check_email_code
@course = Course.find(params[:course_id])
code = @course.flags[Fall2017CmuExperiment::EMAIL_CODE]
raise IncorrectEmailCodeError if code.nil?
raise IncorrectEmailCodeError unless code == params[:email_code]
end
end
class IncorrectEmailCodeError < StandardError; end
end
| Update opt-in flash message for CMU experiment | Update opt-in flash message for CMU experiment
| Ruby | mit | WikiEducationFoundation/WikiEduDashboard,majakomel/WikiEduDashboard,majakomel/WikiEduDashboard,KarmaHater/WikiEduDashboard,alpha721/WikiEduDashboard,majakomel/WikiEduDashboard,WikiEducationFoundation/WikiEduDashboard,sejalkhatri/WikiEduDashboard,WikiEducationFoundation/WikiEduDashboard,majakomel/WikiEduDashboard,sejalkhatri/WikiEduDashboard,KarmaHater/WikiEduDashboard,KarmaHater/WikiEduDashboard,alpha721/WikiEduDashboard,WikiEducationFoundation/WikiEduDashboard,alpha721/WikiEduDashboard,alpha721/WikiEduDashboard,sejalkhatri/WikiEduDashboard,KarmaHater/WikiEduDashboard,sejalkhatri/WikiEduDashboard,sejalkhatri/WikiEduDashboard,WikiEducationFoundation/WikiEduDashboard | ruby | ## Code Before:
require "#{Rails.root}/lib/experiments/fall2017_cmu_experiment"
module Experiments
class Fall2017CmuExperimentController < ApplicationController
before_action :set_course_and_check_email_code, only: %i[opt_in opt_out]
before_action :require_admin_permissions, only: :course_list
def opt_in
Fall2017CmuExperiment.new(@course).opt_in
flash[:notice] = 'Thanks for opting in!'
redirect_to "/courses/#{@course.slug}"
end
def opt_out
Fall2017CmuExperiment.new(@course).opt_out
flash[:notice] = 'Okay, you have opted out.'
redirect_to "/courses/#{@course.slug}"
end
def course_list
send_data Fall2017CmuExperiment.course_list,
filename: "fall2017-cmu-experiment-courses-#{Time.zone.today}.csv"
end
private
def set_course_and_check_email_code
@course = Course.find(params[:course_id])
code = @course.flags[Fall2017CmuExperiment::EMAIL_CODE]
raise IncorrectEmailCodeError if code.nil?
raise IncorrectEmailCodeError unless code == params[:email_code]
end
end
class IncorrectEmailCodeError < StandardError; end
end
## Instruction:
Update opt-in flash message for CMU experiment
## Code After:
require "#{Rails.root}/lib/experiments/fall2017_cmu_experiment"
module Experiments
class Fall2017CmuExperimentController < ApplicationController
before_action :set_course_and_check_email_code, only: %i[opt_in opt_out]
before_action :require_admin_permissions, only: :course_list
def opt_in
Fall2017CmuExperiment.new(@course).opt_in
flash[:notice] = 'Thank you for opting in. We will add the video sessions to the relevant assignments on the timeline for your WikiEd dashboard and send you an email about next steps.'
redirect_to "/courses/#{@course.slug}"
end
def opt_out
Fall2017CmuExperiment.new(@course).opt_out
flash[:notice] = 'Okay, you have opted out.'
redirect_to "/courses/#{@course.slug}"
end
def course_list
send_data Fall2017CmuExperiment.course_list,
filename: "fall2017-cmu-experiment-courses-#{Time.zone.today}.csv"
end
private
def set_course_and_check_email_code
@course = Course.find(params[:course_id])
code = @course.flags[Fall2017CmuExperiment::EMAIL_CODE]
raise IncorrectEmailCodeError if code.nil?
raise IncorrectEmailCodeError unless code == params[:email_code]
end
end
class IncorrectEmailCodeError < StandardError; end
end
|
require "#{Rails.root}/lib/experiments/fall2017_cmu_experiment"
module Experiments
class Fall2017CmuExperimentController < ApplicationController
before_action :set_course_and_check_email_code, only: %i[opt_in opt_out]
before_action :require_admin_permissions, only: :course_list
def opt_in
Fall2017CmuExperiment.new(@course).opt_in
- flash[:notice] = 'Thanks for opting in!'
+ flash[:notice] = 'Thank you for opting in. We will add the video sessions to the relevant assignments on the timeline for your WikiEd dashboard and send you an email about next steps.'
redirect_to "/courses/#{@course.slug}"
end
def opt_out
Fall2017CmuExperiment.new(@course).opt_out
flash[:notice] = 'Okay, you have opted out.'
redirect_to "/courses/#{@course.slug}"
end
def course_list
send_data Fall2017CmuExperiment.course_list,
filename: "fall2017-cmu-experiment-courses-#{Time.zone.today}.csv"
end
private
def set_course_and_check_email_code
@course = Course.find(params[:course_id])
code = @course.flags[Fall2017CmuExperiment::EMAIL_CODE]
raise IncorrectEmailCodeError if code.nil?
raise IncorrectEmailCodeError unless code == params[:email_code]
end
end
class IncorrectEmailCodeError < StandardError; end
end | 2 | 0.054054 | 1 | 1 |
06ab217f49e00bd52c9f8b632db904e1ebe7256d | pycroft/helpers/date.py | pycroft/helpers/date.py | from datetime import timedelta
from datetime import date
def diff_month(d1: date, d2: date) -> int:
"""Calculate the difference in months ignoring the days
If d1 > d2, the result is positive.
"""
return (d1.year - d2.year) * 12 + d1.month - d2.month
def last_day_of_month(d):
next_month = d.replace(day=28) + timedelta(4)
return d.replace(day=(next_month - timedelta(days=next_month.day)).day)
| from calendar import monthrange
from datetime import date
def diff_month(d1: date, d2: date) -> int:
"""Calculate the difference in months ignoring the days
If d1 > d2, the result is positive.
"""
return (d1.year - d2.year) * 12 + d1.month - d2.month
def last_day_of_month(d: date) -> date:
_, num_days = monthrange(d.year, d.month)
return d.replace(day=num_days)
| Use builtin function to find last month | Use builtin function to find last month
| Python | apache-2.0 | agdsn/pycroft,lukasjuhrich/pycroft,agdsn/pycroft,agdsn/pycroft,lukasjuhrich/pycroft,lukasjuhrich/pycroft,lukasjuhrich/pycroft,agdsn/pycroft,agdsn/pycroft | python | ## Code Before:
from datetime import timedelta
from datetime import date
def diff_month(d1: date, d2: date) -> int:
"""Calculate the difference in months ignoring the days
If d1 > d2, the result is positive.
"""
return (d1.year - d2.year) * 12 + d1.month - d2.month
def last_day_of_month(d):
next_month = d.replace(day=28) + timedelta(4)
return d.replace(day=(next_month - timedelta(days=next_month.day)).day)
## Instruction:
Use builtin function to find last month
## Code After:
from calendar import monthrange
from datetime import date
def diff_month(d1: date, d2: date) -> int:
"""Calculate the difference in months ignoring the days
If d1 > d2, the result is positive.
"""
return (d1.year - d2.year) * 12 + d1.month - d2.month
def last_day_of_month(d: date) -> date:
_, num_days = monthrange(d.year, d.month)
return d.replace(day=num_days)
| - from datetime import timedelta
+ from calendar import monthrange
from datetime import date
def diff_month(d1: date, d2: date) -> int:
"""Calculate the difference in months ignoring the days
If d1 > d2, the result is positive.
"""
return (d1.year - d2.year) * 12 + d1.month - d2.month
- def last_day_of_month(d):
+ def last_day_of_month(d: date) -> date:
? ++++++ ++++++++
- next_month = d.replace(day=28) + timedelta(4)
- return d.replace(day=(next_month - timedelta(days=next_month.day)).day)
+ _, num_days = monthrange(d.year, d.month)
+ return d.replace(day=num_days) | 8 | 0.533333 | 4 | 4 |
03af5f1806d6a9d7052a8836d07b2135277bc2d5 | .travis.yml | .travis.yml | sudo: required
language: java
before_install:
- mkdir ci-dependencies
- cd ci-dependencies
- curl -O https://openskywalking.github.io/skywalking-ci-assist/jars/ojdbc14-10.2.0.4.0.jar
- curl -O https://openskywalking.github.io/skywalking-ci-assist/jars/resin-4.0.41.jar
- cd ..
install:
- jdk_switcher use oraclejdk8
- mvn clean install -Ptravis-ci-submodule --quiet jacoco:report coveralls:report
- mvn clean javadoc:javadoc -Dmaven.test.skip=true --quiet
| sudo: required
cache:
directories:
- "$HOME/.m2"
language: java
before_install:
- mkdir ci-dependencies
- cd ci-dependencies
- curl -O https://openskywalking.github.io/skywalking-ci-assist/jars/ojdbc14-10.2.0.4.0.jar
- curl -O https://openskywalking.github.io/skywalking-ci-assist/jars/resin-4.0.41.jar
- cd ..
install:
- jdk_switcher use oraclejdk8
- mvn clean install -Ptravis-ci-submodule --quiet jacoco:report coveralls:report
- mvn clean javadoc:javadoc -Dmaven.test.skip=true --quiet
| Support m2 cache in CI. | Support m2 cache in CI.
| YAML | apache-2.0 | ascrutae/sky-walking,OpenSkywalking/skywalking,apache/skywalking,apache/skywalking,apache/skywalking,apache/skywalking,apache/skywalking,apache/skywalking,apache/skywalking,ascrutae/sky-walking,ascrutae/sky-walking,ascrutae/sky-walking,OpenSkywalking/skywalking | yaml | ## Code Before:
sudo: required
language: java
before_install:
- mkdir ci-dependencies
- cd ci-dependencies
- curl -O https://openskywalking.github.io/skywalking-ci-assist/jars/ojdbc14-10.2.0.4.0.jar
- curl -O https://openskywalking.github.io/skywalking-ci-assist/jars/resin-4.0.41.jar
- cd ..
install:
- jdk_switcher use oraclejdk8
- mvn clean install -Ptravis-ci-submodule --quiet jacoco:report coveralls:report
- mvn clean javadoc:javadoc -Dmaven.test.skip=true --quiet
## Instruction:
Support m2 cache in CI.
## Code After:
sudo: required
cache:
directories:
- "$HOME/.m2"
language: java
before_install:
- mkdir ci-dependencies
- cd ci-dependencies
- curl -O https://openskywalking.github.io/skywalking-ci-assist/jars/ojdbc14-10.2.0.4.0.jar
- curl -O https://openskywalking.github.io/skywalking-ci-assist/jars/resin-4.0.41.jar
- cd ..
install:
- jdk_switcher use oraclejdk8
- mvn clean install -Ptravis-ci-submodule --quiet jacoco:report coveralls:report
- mvn clean javadoc:javadoc -Dmaven.test.skip=true --quiet
| sudo: required
+
+ cache:
+ directories:
+ - "$HOME/.m2"
language: java
before_install:
- mkdir ci-dependencies
- cd ci-dependencies
- curl -O https://openskywalking.github.io/skywalking-ci-assist/jars/ojdbc14-10.2.0.4.0.jar
- curl -O https://openskywalking.github.io/skywalking-ci-assist/jars/resin-4.0.41.jar
- cd ..
install:
- jdk_switcher use oraclejdk8
- mvn clean install -Ptravis-ci-submodule --quiet jacoco:report coveralls:report
- mvn clean javadoc:javadoc -Dmaven.test.skip=true --quiet
| 4 | 0.266667 | 4 | 0 |
85ba78029a13ac78c9ee024526cc62a0b894553c | .travis.yml | .travis.yml | language: php
php:
- '5.4'
- '5.6'
- '7.0'
before_install:
- nvm install 4.1.1
- nvm use 4.1.1
install:
- npm install
script:
- npm run compile --silent
- npm test --silent
sudo: false
| language: php
php:
- '5.4'
- '5.6'
- '7.0'
before_install:
- nvm install 4.1.1
- nvm use 4.1.1
- npm install -g typings
install:
- npm install
- typings install
script:
- npm run compile --silent
- npm test --silent
sudo: false
| Add typings to Travis build | Add typings to Travis build
| YAML | mit | felixfbecker/vscode-php-debug,felixfbecker/vscode-php-debug,felixfbecker/vscode-php-debug | yaml | ## Code Before:
language: php
php:
- '5.4'
- '5.6'
- '7.0'
before_install:
- nvm install 4.1.1
- nvm use 4.1.1
install:
- npm install
script:
- npm run compile --silent
- npm test --silent
sudo: false
## Instruction:
Add typings to Travis build
## Code After:
language: php
php:
- '5.4'
- '5.6'
- '7.0'
before_install:
- nvm install 4.1.1
- nvm use 4.1.1
- npm install -g typings
install:
- npm install
- typings install
script:
- npm run compile --silent
- npm test --silent
sudo: false
| language: php
php:
- '5.4'
- '5.6'
- '7.0'
before_install:
- nvm install 4.1.1
- nvm use 4.1.1
+ - npm install -g typings
install:
- npm install
+ - typings install
script:
- npm run compile --silent
- npm test --silent
sudo: false | 2 | 0.142857 | 2 | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.