hash
stringlengths 40
40
| diff
stringlengths 172
2.63k
| message
stringlengths 12
593
| project
stringlengths 7
65
| split
stringclasses 1
value | diff_languages
stringclasses 54
values |
|---|---|---|---|---|---|
47bc2dcbf878f7a91d8804a4135cf9c75eb0e1d6
|
diff --git a/lib/utils/formize.js b/lib/utils/formize.js
index <HASH>..<HASH> 100644
--- a/lib/utils/formize.js
+++ b/lib/utils/formize.js
@@ -219,6 +219,11 @@ export default function formize({
checked !== undefined ? checked : undefined
)
+ if (isFile && file === undefined) {
+ entity['data-file'] = null
+ entity.value = ''
+ }
+
if (file !== undefined) {
entity['data-file'] = file
entity.value = ''
|
feat: handle file removal in setValue
|
eyedea-io_zefir
|
train
|
js
|
101218716da14b6000e2022486500d2317e08735
|
diff --git a/packages/letter/gulpfile.js b/packages/letter/gulpfile.js
index <HASH>..<HASH> 100644
--- a/packages/letter/gulpfile.js
+++ b/packages/letter/gulpfile.js
@@ -106,8 +106,7 @@ gulp.task("docs:index", () => {
return gulp
.src([
"dist/*.html",
- "dist/*.pdf",
- "dist/*.ico"
+ "dist/*.pdf"
])
.pipe(gulp.dest("."));
});
|
chore(letter): `docs` task doesn't need to copy `*.ico` anymore.
|
randytarampi_me
|
train
|
js
|
b2174c1ce8878e104c63664d1c8f652589fccd89
|
diff --git a/lib/instrumentation/generic-span.js b/lib/instrumentation/generic-span.js
index <HASH>..<HASH> 100644
--- a/lib/instrumentation/generic-span.js
+++ b/lib/instrumentation/generic-span.js
@@ -10,7 +10,7 @@ module.exports = GenericSpan
function GenericSpan (agent, type, opts) {
this._timer = new Timer(opts.timer, opts.startTime)
- this._context = TraceParent.startOrResume(opts.childOf, agent._conf)
+ this._context = TraceParent.startOrResume(opts.childOf, agent._conf) // _context is used by the OT bridge, and should unfortunately therefore be considered public
this._agent = agent
this._tags = null
|
chore: add code comment about not changing _context (#<I>)
|
elastic_apm-agent-nodejs
|
train
|
js
|
ab536d5a7772c61c6c74f8e5eb2c6fb8b2c97bb2
|
diff --git a/index.js b/index.js
index <HASH>..<HASH> 100755
--- a/index.js
+++ b/index.js
@@ -13,7 +13,7 @@
const path = require('path')
const Commands = require('./src/Commands')
const commandNames = []
-const needProviders = ['repl', 'route:list', 'install', 'run:instructions']
+const needProviders = ['repl', 'route:list', 'run:instructions']
const ace = require('./lib/ace')
diff --git a/src/Commands/Install/index.js b/src/Commands/Install/index.js
index <HASH>..<HASH> 100644
--- a/src/Commands/Install/index.js
+++ b/src/Commands/Install/index.js
@@ -17,20 +17,10 @@ Installation failed due to following error
=============================================`
class Install extends Command {
- constructor (Helpers) {
+ constructor () {
super()
- this.Helpers = Helpers
- }
-
- /**
- * Injecting dependencies
- *
- * @method inject
- *
- * @return {Array}
- */
- static get inject () {
- return ['Adonis/Src/Helpers']
+ const FakeHelpers = require('@adonisjs/ignitor/src/Helpers')
+ this.Helpers = new FakeHelpers(process.cwd())
}
/**
|
fix(install): do not require app ace file
application ace file should not required on install, since app errors should never stop installation
|
adonisjs_adonis-cli
|
train
|
js,js
|
1d84fb5195d8559ca74173f98019777a00206777
|
diff --git a/apps/sampler-swing/src/main/java/org/kordamp/ikonli/sampler/swing/Sampler.java b/apps/sampler-swing/src/main/java/org/kordamp/ikonli/sampler/swing/Sampler.java
index <HASH>..<HASH> 100644
--- a/apps/sampler-swing/src/main/java/org/kordamp/ikonli/sampler/swing/Sampler.java
+++ b/apps/sampler-swing/src/main/java/org/kordamp/ikonli/sampler/swing/Sampler.java
@@ -62,6 +62,7 @@ import javax.swing.JPanel;
import javax.swing.JScrollPane;
import javax.swing.JTabbedPane;
import javax.swing.SwingUtilities;
+import javax.swing.WindowConstants;
import java.awt.Dimension;
import java.awt.GridLayout;
import java.util.EnumSet;
@@ -122,6 +123,7 @@ public class Sampler {
JFrame frame = new JFrame("Ikonli Sampler");
frame.add(tabPane);
frame.setSize(new Dimension(1024, 1024));
+ frame.setDefaultCloseOperation(WindowConstants.EXIT_ON_CLOSE);
frame.setVisible(true);
}
|
fix: jvm not terminating after closing sampler-swing app
|
aalmiray_ikonli
|
train
|
java
|
f74e1698e46299e3c835e5c180a4d878502e40f4
|
diff --git a/lib/drop.js b/lib/drop.js
index <HASH>..<HASH> 100644
--- a/lib/drop.js
+++ b/lib/drop.js
@@ -31,5 +31,8 @@ function drop(schemas) {
*/
function dropSchema(result, schema) {
- return this.db.bookshelf.knex.schema.dropTableIfExists(schema.tableName);
+ return this.db.bookshelf.knex.schema.dropTableIfExists(schema.tableName)
+ .then(function () {
+ return result.concat([schema]);
+ });
}
\ No newline at end of file
diff --git a/lib/reset.js b/lib/reset.js
index <HASH>..<HASH> 100644
--- a/lib/reset.js
+++ b/lib/reset.js
@@ -35,6 +35,9 @@ function resetSchema(result, schema) {
return knex.schema.hasTable(schema.tableName)
.then(function (exists) {
if (! exists) return result;
- return knex(schema.tableName).del();
+ return knex(schema.tableName).del()
+ .then(function () {
+ return result.concat([schema]);
+ });
});
}
\ No newline at end of file
|
fix(manager): drop and reset promise resolution value
|
lemonde_knex-schema
|
train
|
js,js
|
9fd605c2007ffa321e1cc3e5649acb1dc63e337c
|
diff --git a/docs/src/components/DocApiEntry.js b/docs/src/components/DocApiEntry.js
index <HASH>..<HASH> 100644
--- a/docs/src/components/DocApiEntry.js
+++ b/docs/src/components/DocApiEntry.js
@@ -292,7 +292,7 @@ describe.events = events => {
}
else {
params.push(
- h('div', { class: 'text-italic q-pa-xs' }, '*None*')
+ h('div', { class: 'text-italic q-py-xs q-px-md' }, '*None*')
)
}
|
feat(docs): small tweak to API card design
|
quasarframework_quasar
|
train
|
js
|
ceed033b168f14c866d4dc8210562a561b975dca
|
diff --git a/gridsome/lib/develop.js b/gridsome/lib/develop.js
index <HASH>..<HASH> 100644
--- a/gridsome/lib/develop.js
+++ b/gridsome/lib/develop.js
@@ -126,8 +126,8 @@ module.exports = async (context, args) => {
config.entryPoints.store.forEach((entry, name) => {
config.entry(name)
- .prepend(`webpack-hot-middleware/client?name=${name}&reload=true&noInfo=true`)
- .prepend('webpack/hot/dev-server')
+ .prepend(`${require.resolve('webpack-hot-middleware/client')}?name=${name}&reload=true&noInfo=true`)
+ .prepend(require.resolve('webpack/hot/dev-server'))
})
return app.compiler.resolveWebpackConfig(false, config)
|
fix(gridsome): require.resolve develop entries
|
gridsome_gridsome
|
train
|
js
|
ddc5a76739e4f83eae66b2bae4659c80f58a0e6b
|
diff --git a/src/RollbarLogger.php b/src/RollbarLogger.php
index <HASH>..<HASH> 100644
--- a/src/RollbarLogger.php
+++ b/src/RollbarLogger.php
@@ -23,6 +23,14 @@ class RollbarLogger extends AbstractLogger
$this->truncation = new Truncation($this->config);
$this->queue = array();
}
+
+ /**
+ * @since 3.0
+ */
+ public function getConfig()
+ {
+ return $this->config;
+ }
public function enable()
{
diff --git a/tests/RollbarTest.php b/tests/RollbarTest.php
index <HASH>..<HASH> 100644
--- a/tests/RollbarTest.php
+++ b/tests/RollbarTest.php
@@ -30,7 +30,7 @@ class RollbarTest extends BaseRollbarTest
Rollbar::init(self::$simpleConfig);
$this->assertInstanceOf('Rollbar\RollbarLogger', Rollbar::logger());
- $this->assertAttributeEquals(new Config(self::$simpleConfig), 'config', Rollbar::logger());
+ $this->assertEquals(new Config(self::$simpleConfig), Rollbar::logger()->getConfig());
}
public function testInitWithLogger()
|
feat: add Rollbar\RollbarLogger::getConfig()
Necessitated by PHPUnit 9 removing assertions for values of
protected and private members, add a method to inspect the given
config.
|
rollbar_rollbar-php
|
train
|
php,php
|
35714414a2ab18e7833b39b7e006121fe567ddb9
|
diff --git a/core/frontier.js b/core/frontier.js
index <HASH>..<HASH> 100644
--- a/core/frontier.js
+++ b/core/frontier.js
@@ -119,11 +119,7 @@ exports.icFrontier = function (config) {
// get one level deeper for scalra system servers (monitor, entry...)
if (owner === 'scalra') {
owner = words[words.length-4];
-
- // also obtain proper project name if not for 'demo' project
- // TODO: hard-coded now, find a better approach?
- //if (project !== 'demo')
- // project = words[words.length-3];
+ project = words[words.length-3];
}
LOG.warn('extracting server info from path... \n[owner]: ' + owner + '\n[project]: ' + project + '\n[name]: ' + name, l_name);
|
fix: cannot find proper scalra path when deployed on dev server
|
imonology_scalra
|
train
|
js
|
45dfcefec9d3bff5cef7af0eb4bba07a516f4af8
|
diff --git a/scripts/build-playroom.js b/scripts/build-playroom.js
index <HASH>..<HASH> 100644
--- a/scripts/build-playroom.js
+++ b/scripts/build-playroom.js
@@ -1,3 +1,6 @@
+// First, ensure the build is running in production mode
+process.env.NODE_ENV = 'production';
+
const playroom = require('playroom/lib');
const { cwd } = require('../lib/cwd');
const makePlayroomConfig = require('../config/playroom/makePlayroomConfig');
|
fix(playroom): Ensure build is in production mode (#<I>)
|
seek-oss_sku
|
train
|
js
|
cdee09fa42d611c9ae95be1f87495ff232bd03d9
|
diff --git a/mamba/formatters.py b/mamba/formatters.py
index <HASH>..<HASH> 100644
--- a/mamba/formatters.py
+++ b/mamba/formatters.py
@@ -231,7 +231,7 @@ class JUnitFormatter(DocumentationFormatter):
self.suite.attrib['tests'] = str(example_count)
self.suite.attrib['skipped'] = str(pending_count)
self.suite.attrib['failures'] = str(failed_count)
- self.suite.attrib['time'] = str(duration)
+ self.suite.attrib['time'] = '%f' % (duration.total_seconds(), )
ElementTree.ElementTree(self.suite).write(sys.stdout, encoding='unicode')
def _dump_example(self, example, child=None):
@@ -239,7 +239,7 @@ class JUnitFormatter(DocumentationFormatter):
'classname': example.classname,
'name': self.format_full_example_name(example),
'file': example.file,
- 'time': str(example.elapsed_time.total_seconds())
+ 'time': '%f' % (example.elapsed_time.total_seconds(), )
})
if child is not None:
testcase.append(child)
|
fix: change junit 'time' field format, as it breaks some importers when parsing it
|
nestorsalceda_mamba
|
train
|
py
|
3a9c99556d45f01ab4bdbd1f24f246b7fff92a1e
|
diff --git a/tests/share_elements/test_examples.py b/tests/share_elements/test_examples.py
index <HASH>..<HASH> 100644
--- a/tests/share_elements/test_examples.py
+++ b/tests/share_elements/test_examples.py
@@ -105,7 +105,7 @@ def test_functionality_example(caplog):
rafcon.core.singleton.state_machine_execution_engine.join()
finally:
testing_utils.wait_for_gui() # to avoid execution and model notification clinches
- testing_utils.shutdown_environment(caplog=caplog, expected_warnings=1, expected_errors=4,
+ testing_utils.shutdown_environment(caplog=caplog, expected_warnings=0, expected_errors=4,
unpatch_threading=False)
|
test(examples): adapt number of expected warnings
no more warning as scripts are no longer compiled on state machine
loading
|
DLR-RM_RAFCON
|
train
|
py
|
7cba67620ebba8457863bdea473ee5df7c15e6b6
|
diff --git a/src/managers/BaseGuildEmojiManager.js b/src/managers/BaseGuildEmojiManager.js
index <HASH>..<HASH> 100644
--- a/src/managers/BaseGuildEmojiManager.js
+++ b/src/managers/BaseGuildEmojiManager.js
@@ -22,7 +22,7 @@ class BaseGuildEmojiManager extends CachedManager {
/**
* Data that can be resolved into a GuildEmoji object. This can be:
- * * A custom emoji identifier
+ * * A Snowflake
* * A GuildEmoji object
* * A ReactionEmoji object
* @typedef {Snowflake|GuildEmoji|ReactionEmoji} EmojiResolvable
|
docs(EmojiResolvable): update description of typedef to remove ambiguity (#<I>)
|
discordjs_discord.js
|
train
|
js
|
c491ee13da9ecd19d57435db992ad24deba3745c
|
diff --git a/__tests__/monorepo-deps.test.js b/__tests__/monorepo-deps.test.js
index <HASH>..<HASH> 100644
--- a/__tests__/monorepo-deps.test.js
+++ b/__tests__/monorepo-deps.test.js
@@ -53,7 +53,7 @@ expect.extend({ pkgToHaveDependenciesOn });
const boltPkgs = getPkgList();
describe('Bolt Components declare dependencies in package.json if used in Twig files', () => {
- const excludedPkgs = ['@bolt/generator-bolt'];
+ const excludedPkgs = ['@bolt/generator-bolt', '@bolt/components-editor'];
boltPkgs
.filter(boltPkg => !excludedPkgs.includes(boltPkg.name))
|
fix: temporarily exclude the editor from packages that have to get the dependency map checked
|
bolt-design-system_bolt
|
train
|
js
|
2464de55b8f850361a7866695c3cd71e8a537e30
|
diff --git a/tooling/util/proxies.js b/tooling/util/proxies.js
index <HASH>..<HASH> 100644
--- a/tooling/util/proxies.js
+++ b/tooling/util/proxies.js
@@ -8,19 +8,34 @@ const path = require('path');
const url = require('url');
const yakbak = require('yakbak');
+// NOTE: CONVERSATION_SERVICE does not end in URL
const services = [
{
+ defaultUrl: 'https://atlas-a.wbx2.com/admin/api/v1',
+ env: 'ATLAS_SERVICE_URL',
+ name: 'atlas',
+ port: 3010,
+ serviceUrl: process.env.ATLAS_SERVICE_URL
+ },
+ // {
+ // defaultUrl: 'https://conv-a.wbx2.com/conversation/api/v1',
+ // env: 'CONVERSATION_SERVICE',
+ // name: 'conversation',
+ // port: 3020,
+ // serviceUrl: process.env.CONVERSATION_SERVICE
+ // },
+ {
defaultUrl: 'https://api.ciscospark.com/v1',
env: 'HYDRA_SERVICE_URL',
name: 'hydra',
- port: 3010,
+ port: 3030,
serviceUrl: process.env.HYDRA_SERVICE_URL
},
{
defaultUrl: 'https://wdm-a.wbx2.com/wdm/api/v1',
env: 'WDM_SERVICE_URL',
name: 'wdm',
- port: 3020,
+ port: 3040,
serviceUrl: process.env.WDM_SERVICE_URL
}
];
|
chore(proxies): add services
|
webex_spark-js-sdk
|
train
|
js
|
a4bc5bae61ffe99fe3e6302db519510374941074
|
diff --git a/superset-frontend/temporary_superset_ui/superset-ui/plugins/legacy-plugin-chart-sankey-loop/src/SankeyLoop.js b/superset-frontend/temporary_superset_ui/superset-ui/plugins/legacy-plugin-chart-sankey-loop/src/SankeyLoop.js
index <HASH>..<HASH> 100644
--- a/superset-frontend/temporary_superset_ui/superset-ui/plugins/legacy-plugin-chart-sankey-loop/src/SankeyLoop.js
+++ b/superset-frontend/temporary_superset_ui/superset-ui/plugins/legacy-plugin-chart-sankey-loop/src/SankeyLoop.js
@@ -98,7 +98,10 @@ function SankeyLoop(element, props) {
)
.linkColor(d => color(d.source.name));
- const svg = select(element)
+ const div = select(element);
+ div.selectAll('*').remove();
+
+ const svg = div
.append('svg')
.classed('superset-legacy-chart-sankey-loop', true)
.style('width', width)
|
fix(legacy-plugin-chart-sankey-loop): clear element before redrawing (#<I>)
otherwise every (re)render adds a new svg to the element. this is essentially
the same as the fix that was applied to legacy-plugin-chart-sankey back in
apache/superset#<I>
|
apache_incubator-superset
|
train
|
js
|
da35b0fc5f4530617a8f2abc8db50d51b0a57c11
|
diff --git a/test/log.spec.js b/test/log.spec.js
index <HASH>..<HASH> 100644
--- a/test/log.spec.js
+++ b/test/log.spec.js
@@ -31,11 +31,16 @@ describe('.log', function () {
})
it('.log.tail', (done) => {
+ let i = setInterval(() => {
+ ipfs.files.add(Buffer.from('just adding some data to generate logs'))
+ }, 1000)
+
const req = ipfs.log.tail((err, res) => {
expect(err).to.not.exist()
expect(req).to.exist()
res.once('data', (obj) => {
+ clearInterval(i)
expect(obj).to.be.an('object')
done()
})
|
fix: fix log.tail by calling add after listening for events (#<I>)
|
ipfs_js-ipfs
|
train
|
js
|
daaec07bbe5ff216898504199f185fe033782dd5
|
diff --git a/packages/retail-ui/styleguide.config.js b/packages/retail-ui/styleguide.config.js
index <HASH>..<HASH> 100644
--- a/packages/retail-ui/styleguide.config.js
+++ b/packages/retail-ui/styleguide.config.js
@@ -179,7 +179,7 @@ const webpackConfig = {
]
},
{
- test: /\.less$/,
+ test: /\.(css|less)$/,
use: [
'style-loader',
'css-loader',
|
chore: update styleguidist webpack config
|
skbkontur_retail-ui
|
train
|
js
|
03a1441bbe98525412df754d3934141bc3b12168
|
diff --git a/pgjdbc/src/main/java/org/postgresql/jdbc/PgConnection.java b/pgjdbc/src/main/java/org/postgresql/jdbc/PgConnection.java
index <HASH>..<HASH> 100644
--- a/pgjdbc/src/main/java/org/postgresql/jdbc/PgConnection.java
+++ b/pgjdbc/src/main/java/org/postgresql/jdbc/PgConnection.java
@@ -655,6 +655,11 @@ public class PgConnection implements BaseConnection {
* {@inheritDoc}
*/
public void close() throws SQLException {
+ if (queryExecutor == null) {
+ // This might happen in case constructor throws an exception (e.g. host being not available).
+ // When that happens the connection is still registered in the finalizer queue, so it gets finalized
+ return;
+ }
releaseTimer();
queryExecutor.close();
openStackTrace = null;
|
fix: avoid NPE in PgConnection.finalize
Note: the NPE is not visible to the end users as it can be caught in debugger only.
fixes #<I>
|
pgjdbc_pgjdbc
|
train
|
java
|
5d27e0351c4c2e7f7fe8b85dafe0decf96cae466
|
diff --git a/samples/generateV4SignedPolicy.js b/samples/generateV4SignedPolicy.js
index <HASH>..<HASH> 100644
--- a/samples/generateV4SignedPolicy.js
+++ b/samples/generateV4SignedPolicy.js
@@ -21,7 +21,7 @@
*/
function main(bucketName = 'my-bucket', filename = 'test.txt') {
- // [START storage_generate_post_policy_url_v4]
+ // [START storage_generate_signed_post_policy_v4]
/**
* TODO(developer): Uncomment the following lines before running the sample.
*/
@@ -65,6 +65,6 @@ function main(bucketName = 'my-bucket', filename = 'test.txt') {
}
generateV4SignedPolicy().catch(console.error);
- // [END storage_generate_post_policy_url_v4]
+ // [END storage_generate_signed_post_policy_v4]
}
main(...process.argv.slice(2));
|
chore(samples): region tag for post policy sample (#<I>)
|
googleapis_nodejs-storage
|
train
|
js
|
7de8bb0e4c5aeacfcd0fd6c5f38531280c63489d
|
diff --git a/samcli/__init__.py b/samcli/__init__.py
index <HASH>..<HASH> 100644
--- a/samcli/__init__.py
+++ b/samcli/__init__.py
@@ -2,4 +2,4 @@
SAM CLI version
"""
-__version__ = "0.46.1"
+__version__ = "0.46.2"
|
chore: version bump <I> (#<I>)
Why is this change necessary?
* For releasing a new version of SAM CLI
|
awslabs_aws-sam-cli
|
train
|
py
|
e208c2617761b7543ecbe79fa992c10a5991ad91
|
diff --git a/lib/fetch/index.js b/lib/fetch/index.js
index <HASH>..<HASH> 100644
--- a/lib/fetch/index.js
+++ b/lib/fetch/index.js
@@ -32,6 +32,11 @@ const { ReadableStream } = require('stream/web')
const { performance } = require('perf_hooks')
const EE = require('events')
+// https://fetch.spec.whatwg.org/#garbage-collection
+const registry = new FinalizationRegistry(abort => {
+ abort()
+})
+
// https://fetch.spec.whatwg.org/#fetch-method
async function fetch (resource, init) {
const context = Object.assign(new EE(), {
@@ -1545,11 +1550,14 @@ function httpNetworkFetch (
pullAlgorithm = resume
+ const body = { stream }
+ registry.register(body, connection.abort)
+
response = makeResponse({
status,
statusText,
headersList: headers[kHeadersList],
- body: stream ? { stream } : null
+ body
})
context.on('terminated', onResponseAborted)
|
fix: terminate fetch on gc (#<I>)
* fix: terminate fetch on gc
If the user doesn't consume the response we end up leaking
a connection. When the stream gets GC:d and no further progress
can be made in the fetch, then terminate it.
Fixes: <URL>
|
mcollina_undici
|
train
|
js
|
2b2df1ae8e33dcf8c15e564b29f95476c8a6675c
|
diff --git a/test/forked_joins.js b/test/forked_joins.js
index <HASH>..<HASH> 100644
--- a/test/forked_joins.js
+++ b/test/forked_joins.js
@@ -123,7 +123,7 @@ test('fork way', function (t) {
test('delete a fork', function (t) {
t.plan(9)
// Delete first fork
- osm.del('F', {keys: [ways[1]]}, function (err, doc) {
+ osm.del('F', {links: [ways[1]]}, function (err, doc) {
ways[3] = doc.key
t.error(err)
osm.ready(ready)
|
test: use "links" instead of "keys" for desired effect
|
digidem_osm-p2p-db
|
train
|
js
|
dc612dc9b78150412be6260f573bda8cf0caf3f2
|
diff --git a/index.js b/index.js
index <HASH>..<HASH> 100644
--- a/index.js
+++ b/index.js
@@ -54,7 +54,7 @@ module.exports = {
treeForPublic: function(tree) {
if (this.project.targetIsCordova) {
- var platformAssets = getPlatformAssets();
+ var platformAssets = getPlatformAssets(this.project);
if (platformAssets.path === undefined) {
throw new Error('ember-cordova: Did not receive platform asset path, canot not build');
diff --git a/node-tests/unit/index.js b/node-tests/unit/index.js
index <HASH>..<HASH> 100644
--- a/node-tests/unit/index.js
+++ b/node-tests/unit/index.js
@@ -3,6 +3,7 @@
const td = require('testdouble');
var fs = require('fs');
const expect = require('../helpers/expect');
+const isObject = td.matchers.isA(Object);
const stubIndex = function() {
let stub = require('../../index');
@@ -90,7 +91,7 @@ describe('Index', () => {
projectIndex.treeForPublic()
}).to.throw(Error);
- td.verify(getAssetDouble());
+ td.verify(getAssetDouble(isObject));
});
it('throws an error if platformAssets.path is undefined', function() {
|
fix(get-assets-util): ensure project is passed
|
isleofcode_ember-cordova
|
train
|
js,js
|
d19913986f21d25065242976254c76bb3bad78d1
|
diff --git a/packages/plugin-logger/src/logger.js b/packages/plugin-logger/src/logger.js
index <HASH>..<HASH> 100644
--- a/packages/plugin-logger/src/logger.js
+++ b/packages/plugin-logger/src/logger.js
@@ -71,7 +71,7 @@ function walkAndFilter(object) {
const Logger = SparkPlugin.extend({
namespace: `Logger`,
- computed: {
+ derived: {
level: {
cache: false,
fn() {
|
fix(plugin-logger): use correct keyword for derived properties
|
webex_spark-js-sdk
|
train
|
js
|
a2457252c034655465ef395e621d1289fac1a67a
|
diff --git a/.eslintrc.js b/.eslintrc.js
index <HASH>..<HASH> 100644
--- a/.eslintrc.js
+++ b/.eslintrc.js
@@ -9,6 +9,7 @@ module.exports = {
rules: {
'arrow-parens': 0,
'generator-star-spacing': 0,
+ 'spaced-comment': 0,
'no-debugger': process.env.NODE_ENV === 'production' ? 2 : 0
}
}
|
chore: allow non-spaced comment
|
hjkcai_vue-modal-dialogs
|
train
|
js
|
1fc053141f574996ab57fd437df15a0a0f63415f
|
diff --git a/lib/plugins/load-plugin.js b/lib/plugins/load-plugin.js
index <HASH>..<HASH> 100644
--- a/lib/plugins/load-plugin.js
+++ b/lib/plugins/load-plugin.js
@@ -300,6 +300,7 @@ var initReq = function(req, res, isServer) {
oReq.headers = headers;
oReq.ruleValue = getValue(req, pluginOpts.RULE_VALUE_HEADER);
oReq.pipeValue = getValue(req, pluginOpts.PIPE_VALUE_HEADER);
+ oReq.hostValue = getValue(req, pluginOpts.HOST_VALUE_HEADER);
oReq.url = oReq.fullUrl = getValue(req, pluginOpts.FULL_URL_HEADER);
oReq.realUrl = getValue(req, pluginOpts.REAL_URL_HEADER);
oReq.method = getValue(req, pluginOpts.METHOD_HEADER) || 'GET';
|
refactor: add req.originalReq.hostValue to plugin
|
avwo_whistle
|
train
|
js
|
42ec004c1167e19a5167517355d79a167f54546d
|
diff --git a/protocol-designer/src/components/LabwareSelectionModal/LabwareSelectionModal.js b/protocol-designer/src/components/LabwareSelectionModal/LabwareSelectionModal.js
index <HASH>..<HASH> 100644
--- a/protocol-designer/src/components/LabwareSelectionModal/LabwareSelectionModal.js
+++ b/protocol-designer/src/components/LabwareSelectionModal/LabwareSelectionModal.js
@@ -72,7 +72,7 @@ const RECOMMENDED_LABWARE_BY_MODULE: { [ModuleType]: Array<string> } = {
'opentrons_24_aluminumblock_nest_2ml_snapcap',
'opentrons_24_aluminumblock_nest_0.5ml_screwcap',
],
- magdeck: ['biorad_96_wellplate_200ul_pcr'],
+ magdeck: ['nest_96_wellplate_100ul_pcr_full_skirt'],
thermocycler: ['nest_96_wellplate_100ul_pcr_full_skirt'],
}
|
fix(protocol-designer): Update magnetic module recommended labware (#<I>)
closes #<I>
|
Opentrons_opentrons
|
train
|
js
|
0802c138d3afc9e7b8a252938c8d65a831458a96
|
diff --git a/packages/@aws-cdk/aws-eks/lib/kubectl-handler/apply/__init__.py b/packages/@aws-cdk/aws-eks/lib/kubectl-handler/apply/__init__.py
index <HASH>..<HASH> 100644
--- a/packages/@aws-cdk/aws-eks/lib/kubectl-handler/apply/__init__.py
+++ b/packages/@aws-cdk/aws-eks/lib/kubectl-handler/apply/__init__.py
@@ -26,11 +26,13 @@ def apply_handler(event, context):
overwrite = props.get('Overwrite', False)
# "log in" to the cluster
- subprocess.check_call([ 'aws', 'eks', 'update-kubeconfig',
+ cmd = [ 'aws', 'eks', 'update-kubeconfig',
'--role-arn', role_arn,
'--name', cluster_name,
'--kubeconfig', kubeconfig
- ])
+ ]
+ logger.info(f'Running command: {cmd}')
+ subprocess.check_call(cmd)
# write resource manifests in sequence: { r1 }{ r2 }{ r3 } (this is how
# a stream of JSON objects can be included in a k8s manifest).
|
chore(eks): log the command before invoking aws eks (#<I>)
It seems it is possible to hit a condition where `aws eks update-kubeconfig`
command may indefinitely hang: never returns until lambda is killed
by timeout.
Add debug print so that it is possible to easier localize the issue.
----
*By submitting this pull request, I confirm that my contribution is made under the terms of the Apache-<I> license*
|
awslabs_aws-cdk
|
train
|
py
|
6e616edc696d76593f0f19094b5bc38d581ea3dd
|
diff --git a/webpack.config.js b/webpack.config.js
index <HASH>..<HASH> 100644
--- a/webpack.config.js
+++ b/webpack.config.js
@@ -49,12 +49,13 @@ module.exports = {
optimization: {
minimizer: [
new TerserJSPlugin({
+ cache: false,
terserOptions: {
mangle: { keep_classnames: true },
compress: { keep_classnames: true },
+ keep_classnames: true,
output: { comments: false },
},
- parallel: true,
}),
],
},
|
chore: keep_classnames literally everywhere
|
discordjs_discord.js
|
train
|
js
|
5f27e782c7bf698511c0c82f59dc91161d1d1c29
|
diff --git a/packages/runner/cypress/support/index.js b/packages/runner/cypress/support/index.js
index <HASH>..<HASH> 100644
--- a/packages/runner/cypress/support/index.js
+++ b/packages/runner/cypress/support/index.js
@@ -1,6 +1,6 @@
require('@packages/ui-components/cypress/support/customPercyCommand')({
elementOverrides: {
- '.stats .duration': ($el) => $el.text('XX.XX'),
+ '.runnable-header .duration': ($el) => $el.text('XX:XX'),
'.cy-tooltip': true,
},
})
|
fix(unify): Hide duration from Percy snapshots (#<I>)
* hide duration in Percy snapshots
* make it more realistic
|
cypress-io_cypress
|
train
|
js
|
02a9770a66bc8c8dc9c9cf20e30729743a1d18e3
|
diff --git a/components/serverless-files/serverless.js b/components/serverless-files/serverless.js
index <HASH>..<HASH> 100644
--- a/components/serverless-files/serverless.js
+++ b/components/serverless-files/serverless.js
@@ -20,6 +20,7 @@ class FilesComponent extends Component {
const {
region,
bucket,
+ storage = {},
functions: {
apolloService: apolloServiceInputs,
downloadFile: downloadFileInputs,
@@ -46,7 +47,7 @@ class FilesComponent extends Component {
// Create S3 bucket for storing files.
const s3 = await this.load("@serverless/aws-s3");
- const s3Output = await s3({ name: bucket, region });
+ const s3Output = await s3({ name: bucket, region, ...storage });
await configureS3Bucket({
component: this,
s3Output,
|
feat: add ability to configure "storage"
|
Webiny_webiny-js
|
train
|
js
|
6c51fc566d865f25d319b526f282d3e7553eb7d1
|
diff --git a/packages/razzle/config/env.js b/packages/razzle/config/env.js
index <HASH>..<HASH> 100644
--- a/packages/razzle/config/env.js
+++ b/packages/razzle/config/env.js
@@ -74,7 +74,7 @@ function getClientEnvironment(target, is_dev, options, paths) {
);
// Stringify all values so we can feed into Webpack DefinePlugin
const stringified = Object.keys(raw).reduce((env, key) => {
- if (!is_dev && !options.forceRuntimeEnvVars.includes(key)) {
+ if (!options.forceRuntimeEnvVars.includes(key) || is_dev) {
env[`process.env.${key}`] = JSON.stringify(raw[key]);
}
return env;
|
fix(razzle): fix a logic error with forceRuntimeEnvVars
|
jaredpalmer_razzle
|
train
|
js
|
37b5ae04781ef535c7800f3a822cf639e8ddeaf1
|
diff --git a/cloudvolume/mesh.py b/cloudvolume/mesh.py
index <HASH>..<HASH> 100644
--- a/cloudvolume/mesh.py
+++ b/cloudvolume/mesh.py
@@ -319,10 +319,12 @@ end_header
new_vertices = np.hstack((verts, np.arange(verts.shape[0])[:, np.newaxis]))
new_vertices[do_merge, 3] = -1
+ faces = faces.flatten()
+
# use unique to make the artificial vertex list unique and reindex faces
vertices, newfaces = np.unique(new_vertices[faces], return_inverse=True, axis=0)
- newfaces = newfaces.astype(np.uint32)
+ newfaces = newfaces.astype(np.uint32).reshape( (len(newfaces) // 3, 3) )
- return Mesh(vertices, newfaces, None, segid=self.segid,
+ return Mesh(vertices[:,0:3], newfaces, None, segid=self.segid,
encoding_type=self.encoding_type, encoding_options=self.encoding_options
)
|
fix: deduplication was corrupting faces
|
seung-lab_cloud-volume
|
train
|
py
|
292083a073606b22b974c0b1523ac0106464ba91
|
diff --git a/commitlint.config.js b/commitlint.config.js
index <HASH>..<HASH> 100644
--- a/commitlint.config.js
+++ b/commitlint.config.js
@@ -3,17 +3,21 @@
// so authorizing 4+5 = 9 characters more on master for the max header length should work
// until we reach PR #99999.
-let maxHeaderLength = 100;
+let maxLineLength = 100;
const prExtrasChars = 9;
const isPushEvent = process.env.GITHUB_EVENT_NAME === 'push';
if (isPushEvent) {
- maxHeaderLength += prExtrasChars;
+ maxLineLength += prExtrasChars;
}
module.exports = {
extends: ['@commitlint/config-conventional'],
- rules: { 'header-max-length': [2, 'always', maxHeaderLength] },
+ rules: {
+ 'header-max-length': [1, 'always', maxLineLength],
+ 'body-max-line-length': [1, 'always', maxLineLength],
+ 'footer-max-line-length': [1, 'always', maxLineLength],
+ },
};
|
ci: consider all linting errors about line length as warnings (#<I>)
|
ForestAdmin_forest-express-mongoose
|
train
|
js
|
8998c08e0b658ed0db7b97d3e086c46458b60860
|
diff --git a/packages/d3fc-element/src/css.js b/packages/d3fc-element/src/css.js
index <HASH>..<HASH> 100644
--- a/packages/d3fc-element/src/css.js
+++ b/packages/d3fc-element/src/css.js
@@ -1,5 +1,5 @@
// Adapted from https://github.com/substack/insert-css
-const css = `d3fc-canvas,d3fc-svg{position:relative}\
+const css = `d3fc-canvas,d3fc-svg{position:relative;display:block}\
d3fc-canvas>canvas,d3fc-svg>svg{position:absolute;top:0;right:0;left:0;bottom: 0}\
d3fc-svg>svg{overflow:visible}`;
|
fix: explicitly set display:block on d3fc-canvas/d3fc-svg (#<I>)
BREAKING CHANGE: `d3fc-canvas`/`d3fc-svg` are both now `display: block` (their previous implicit value was `inline`)
Closes #<I>
|
d3fc_d3fc
|
train
|
js
|
568379fca58744410b273cc6fdb40cec83700d08
|
diff --git a/packages-serverless/egg-layer/index.js b/packages-serverless/egg-layer/index.js
index <HASH>..<HASH> 100644
--- a/packages-serverless/egg-layer/index.js
+++ b/packages-serverless/egg-layer/index.js
@@ -45,6 +45,9 @@ module.exports = engine => {
async defaultInvokeHandler(context) {
return new Promise((resolve, reject) => {
delete context.headers['content-length'];
+ if (eggApp && eggApp.config && eggApp.config.proxy && !context.headers['X-Forwarded-For']) {
+ context.headers['X-Forwarded-For'] = context.ip;
+ }
request(
{
uri: `http://unix:${socketPath}:${context.path}`,
|
fix: egg layer x forwarded for (#<I>)
|
midwayjs_midway
|
train
|
js
|
9941cf8ce92146660e5a5849292a822c1c60167a
|
diff --git a/lib/watson_developer_cloud/websocket/speech_to_text_websocket_listener.rb b/lib/watson_developer_cloud/websocket/speech_to_text_websocket_listener.rb
index <HASH>..<HASH> 100644
--- a/lib/watson_developer_cloud/websocket/speech_to_text_websocket_listener.rb
+++ b/lib/watson_developer_cloud/websocket/speech_to_text_websocket_listener.rb
@@ -5,7 +5,7 @@ require("faye/websocket")
require("json")
ONE_KB = 1024
-TIMEOUT_PREFIX = "No speech detected for"
+TIMEOUT_PREFIX = "No speech detected for".freeze
CLOSE_SIGNAL = 1000
TEN_MILLISECONDS = 0.01
@@ -87,6 +87,17 @@ class WebSocketClient
end
end
+ def add_audio_chunk(chunk:)
+ @data_size += chunk.size
+ @queue << chunk
+ end
+
+ def stop_audio
+ @mic_running = false
+ end
+
+ private
+
def on_connect(_response)
@callback.on_connected
end
@@ -146,17 +157,6 @@ class WebSocketClient
transcripts
end
- def add_audio_chunk(chunk:)
- @data_size += chunk.size
- @queue << chunk
- end
-
- def stop_audio
- @mic_running = false
- end
-
- private
-
def send_chunk(chunk:, final: false)
return if chunk.nil?
@bytes_sent += chunk.size
|
refactor(websocket): Privatize methods not intended for the user
|
watson-developer-cloud_ruby-sdk
|
train
|
rb
|
d04aacfd735067653a8bad2ad3620bcbe069fd0c
|
diff --git a/packages/plugin-bibjson/src/json.js b/packages/plugin-bibjson/src/json.js
index <HASH>..<HASH> 100644
--- a/packages/plugin-bibjson/src/json.js
+++ b/packages/plugin-bibjson/src/json.js
@@ -63,15 +63,13 @@ const typeMap = {
incollection: 'chapter',
conference: 'paper-conference',
inproceedings: 'paper-conference',
+ manual: 'report',
+ misc: 'document',
online: 'website',
patent: 'patent',
phdthesis: 'thesis',
techreport: 'report',
- unpublished: 'manuscript',
-
- // no mapping, fallback
- manual: undefined,
- misc: undefined
+ unpublished: 'manuscript'
}
function quickscrapeSpecificProps () {
|
feat(plugin-bibjson): update mapping to CSL <I>
|
citation-js_citation-js
|
train
|
js
|
7e65d64fb1e1db0bd2b7648102c8a017177f5956
|
diff --git a/tests/test_staticjinja.py b/tests/test_staticjinja.py
index <HASH>..<HASH> 100644
--- a/tests/test_staticjinja.py
+++ b/tests/test_staticjinja.py
@@ -142,15 +142,16 @@ def test_build(site):
assert templates == list(site.templates)
-def test_with_reloader(reloader, site):
- reloader.watch_called = False
+def test_with_reloader(site):
+ watch_called = False
- def watch(self):
- reloader.watch_called = True
+ def fake_watch(self):
+ nonlocal watch_called
+ watch_called = True
- Reloader.watch = watch
+ Reloader.watch = fake_watch
site.render(use_reloader=True)
- assert reloader.watch_called
+ assert watch_called
def test_should_handle(reloader, template_path):
|
chore: Don't inject test variables into real object
Seems like bad news to mess with a real object.
|
Ceasar_staticjinja
|
train
|
py
|
05ef3bb93899de89b80deaa79ad90f7d72151649
|
diff --git a/cloudvolume/provenance.py b/cloudvolume/provenance.py
index <HASH>..<HASH> 100644
--- a/cloudvolume/provenance.py
+++ b/cloudvolume/provenance.py
@@ -5,7 +5,7 @@ import json5
__all__ = [ 'DatasetProvenance', 'DataLayerProvenance' ]
dataset_provenance_schema = {
- "$schema": "http://json-schema.org/draft-04/schema#",
+ "$schema": "http://json-schema.org/draft-04/schema",
"title": "Dataset Provenance",
"description": "Represents a dataset and its derived data layers.",
"required": [
@@ -46,7 +46,7 @@ classes = builder.build_classes()
DatasetProvenance = classes.DatasetProvenance
layer_provenance_schema = {
- "$schema": "http://json-schema.org/draft-04/schema#",
+ "$schema": "http://json-schema.org/draft-04/schema",
"title": "Data Layer Provenance",
"description": "Represents a data layer within a dataset. e.g. image, segmentation, etc",
"required": [
|
fix: warning from python-jsonschema-objects in newest version
|
seung-lab_cloud-volume
|
train
|
py
|
3783e68e6e7c03be7f0c75a1877e9c5c97adbfab
|
diff --git a/src/resolve.js b/src/resolve.js
index <HASH>..<HASH> 100644
--- a/src/resolve.js
+++ b/src/resolve.js
@@ -216,7 +216,7 @@ function $Resolve( $q, $injector) {
* propagated immediately. Once the `$resolve` promise has been rejected, no
* further invocables will be called.
*
- * Cyclic dependencies between invocables are not permitted and will caues `$resolve`
+ * Cyclic dependencies between invocables are not permitted and will cause `$resolve`
* to throw an error. As a special case, an injectable can depend on a parameter
* with the same name as the injectable, which will be fulfilled from the `parent`
* injectable of the same name. This allows inherited values to be decorated.
|
docs(resolve): fix a typo
caues → cause
|
ui-router_angular
|
train
|
js
|
840db11a927ed94e57668c453b7fd29a1edcfc09
|
diff --git a/script/release.js b/script/release.js
index <HASH>..<HASH> 100755
--- a/script/release.js
+++ b/script/release.js
@@ -141,6 +141,8 @@ function s3UrlsForVersion (version) {
}
function checkVersion () {
+ if (args.skipVersionCheck) return
+
console.log(`Verifying that app version matches package version ${pkgVersion}.`)
let startScript = path.join(__dirname, 'start.py')
let scriptArgs = ['--version']
|
chore: add flag to skip version check (#<I>)
|
electron_electron
|
train
|
js
|
0353cbe2dc2343f9d3f7cb3cca5118abea2d1658
|
diff --git a/src/pages/index.js b/src/pages/index.js
index <HASH>..<HASH> 100644
--- a/src/pages/index.js
+++ b/src/pages/index.js
@@ -15,6 +15,7 @@ const IndexPage = () => (
src="../images/gatsby-astronaut.png"
width={300}
quality={95}
+ formats={["AUTO", "WEBP", "AVIF"]}
alt="A Gatsby astronaut"
style={{ marginBottom: `1.45rem` }}
/>
|
chore(starters): Update hello-world and AVIF (#<I>)
|
gatsbyjs_gatsby-starter-default
|
train
|
js
|
66d26b948c222e61515647609b8002ee78d5bd96
|
diff --git a/test/test_spelling.py b/test/test_spelling.py
index <HASH>..<HASH> 100644
--- a/test/test_spelling.py
+++ b/test/test_spelling.py
@@ -430,6 +430,17 @@ class TestSpellcheckOnRegion(WordCacheTestCase):
self.assertThat(errors[0].suggestions,
Not(Contains("spl_lling")))
+ def test_spelling_errors_inside_triple_quotes(self):
+ """Catch spelling errors inside of triple quotes."""
+ contents = "#\n\"\"\"splelling\"\"\"".splitlines()
+ valid, technical = self.__class__.get_dictionaries(contents)
+
+ errors = [e for e in spelling.spellcheck_region(contents,
+ valid,
+ technical)]
+ self.assertThat(errors[0].suggestions,
+ Contains("spelling"))
+
class TestFilterNonspellcheckableTokens(TestCase):
"""Test cases for filtering out non-spellcheckable tokens."""
|
test: Add test that we can find things within triple quotes
|
polysquare_polysquare-generic-file-linter
|
train
|
py
|
9918981f373b51d13434b47b2ef41019ca85c305
|
diff --git a/lib/plugins/module-paths.js b/lib/plugins/module-paths.js
index <HASH>..<HASH> 100644
--- a/lib/plugins/module-paths.js
+++ b/lib/plugins/module-paths.js
@@ -1,7 +1,9 @@
var path = require('path');
var config = require('../config');
-var addon = config.addon || [];
+var addon = (config.addon || []).map(function(str) {
+ return path.resolve(str);
+});
addon = addon.concat(addon.map(formatPath));
config.systemPluginPath = formatPath(config.SYSTEM_PLUGIN_PATH);
|
refactor: w2 start -A relativePath
|
avwo_whistle
|
train
|
js
|
3a27f8fb5b5a172194bdacfea6f91beebe8eab67
|
diff --git a/routes/authorizations.js b/routes/authorizations.js
index <HASH>..<HASH> 100644
--- a/routes/authorizations.js
+++ b/routes/authorizations.js
@@ -38,7 +38,7 @@ module.exports = function (server) {
AccessToken.revoke(uid, cid, function (err, confirm) {
if (err) { return next(err); }
- res.status(204).send()
+ res.sendStatus(204);
});
});
|
fix(authorizations): Use correct function to send only status code
|
anvilresearch_connect
|
train
|
js
|
e8169599ca685a972a09c8898e3358cc3b5d4a10
|
diff --git a/lib/cli.js b/lib/cli.js
index <HASH>..<HASH> 100644
--- a/lib/cli.js
+++ b/lib/cli.js
@@ -72,8 +72,12 @@ var saveUserAsset = function(type, file) {
};
var rmUserAsset = function(type, name) {
- var assetPath = path.join(regDir, type, name + '.css');
- fs.unlinkSync(assetPath);
+ var assetPath = path.join(regDir, 'deck.js', 'themes', type, name + '.css');
+ if(fs.existsSync(assetPath)) {
+ fs.unlinkSync(assetPath);
+ } else {
+ console.log('Remove What? Path "' + assetPath + '" does not exist.');
+ }
};
var optsFromProgramArgs = function(program) {
@@ -129,11 +133,11 @@ module.exports = function(program) {
var themeName = program.rmTheme
, transName = program.rmTransition;
- if(themeName !== true) {
- rmUserAsset('theme', themeName);
+ if(themeName && themeName !== true) {
+ rmUserAsset('style', themeName);
}
- if(transName !== true) {
+ if(transName && transName !== true) {
rmUserAsset('transition', transName);
}
|
fix: Delete from correct user folder
|
jtrussell_bedecked
|
train
|
js
|
62f96a5d91e2949211602bcab191ba7310794eb3
|
diff --git a/components/grid/mediaQueryForStylus.js b/components/grid/mediaQueryForStylus.js
index <HASH>..<HASH> 100644
--- a/components/grid/mediaQueryForStylus.js
+++ b/components/grid/mediaQueryForStylus.js
@@ -5,9 +5,11 @@ var nodes = stylus.nodes;
module.exports = function() {
return function(style) {
for (var key in breakpoints) {
- style.define(key, function() {
- return new nodes.Unit(parseInt(breakpoints[key], 10), 'px')
- });
+ style.define(key, (function(key) {
+ return function() {
+ return new nodes.Unit(parseInt(breakpoints[key], 10), 'px')
+ }
+ })(key));
}
};
};
|
fix(Grid): breakpoints all are <I>px, close #<I>
|
ksc-fe_kpc
|
train
|
js
|
9b1711fad7989283d908c0cc6c09330ea242dca9
|
diff --git a/src/Config.php b/src/Config.php
index <HASH>..<HASH> 100644
--- a/src/Config.php
+++ b/src/Config.php
@@ -833,7 +833,7 @@ class Config
$this->verboseLogger()->info('Occurrence ignored due to custom check_ignore logic');
return true;
}
- } catch (\Exception $exception) {
+ } catch (Throwable $exception) {
$this->verboseLogger()->error(
'Exception occurred in the custom checkIgnore logic:' . $exception->getMessage()
);
|
feat: widen capture of error in custom check
We want to capture more than just an Exception: any failure of the
check ignore logic should interrupt its continued flow, be those
parse errors, div/0, etc.
|
rollbar_rollbar-php
|
train
|
php
|
9a8ef36da5df69a3d15e92f97772faa7f1939b23
|
diff --git a/lib/resolver.js b/lib/resolver.js
index <HASH>..<HASH> 100644
--- a/lib/resolver.js
+++ b/lib/resolver.js
@@ -21,7 +21,7 @@ var DependencyResolver = function() {
var resolveFile = function(filepath, files, alreadyResolvedMap) {
if (!fileMap[filepath]) {
- // console.log('IGORED', filepath);
+ // console.log('IGNORED', filepath);
files.push(filepath);
return;
}
@@ -41,7 +41,11 @@ var DependencyResolver = function() {
};
this.removeFile = function(filepath) {
- // TODO(vojta): handle if unknown file
+ if (!fileMap[filepath]) {
+ // console.log('Cannot remove unmapped file', filepath);
+ return;
+ }
+
fileMap[filepath].provides.forEach(function(dep) {
if (provideMap[dep] === filepath) {
provideMap[dep] = null;
|
fix: Don't attempt to remove an unmapped file.
I'm not really sure what the root cause is here, so this is a bit of a bandaid. When using IntelliJ I see it trying to remove files such as /Users/rob/.../tests/integration/TrackingIntegrationTests.js___jb_old___ even though I've configured just to map *.js.
|
karma-runner_karma-closure
|
train
|
js
|
b8f4c3f6a4b703d42791c9885a4a86cff5d12ac3
|
diff --git a/nodes/call-service/call-service.js b/nodes/call-service/call-service.js
index <HASH>..<HASH> 100644
--- a/nodes/call-service/call-service.js
+++ b/nodes/call-service/call-service.js
@@ -108,7 +108,7 @@ module.exports = function(RED) {
return this.nodeConfig.server.websocket
.callService(apiDomain, apiService, apiData)
.then(() => {
- this.setStatusSuccess(`${apiDomain}.${apiService}`);
+ this.setStatusSuccess(`${apiDomain}.${apiService} called`);
const contextKey = RED.util.parseContextStore(
this.nodeConfig.output_location
|
chore(call-service): Added back in missing work to status message
|
zachowj_node-red-contrib-home-assistant-websocket
|
train
|
js
|
120ff157446aaf2ccfdca82efcbf9cf75a3ad842
|
diff --git a/lib/canbus.js b/lib/canbus.js
index <HASH>..<HASH> 100644
--- a/lib/canbus.js
+++ b/lib/canbus.js
@@ -204,6 +204,11 @@ CanbusStream.prototype.sendPGN = function (msg) {
canid = getCanIdFromPGN(pgn)
buffer = pgn.data
}
+
+ if ( debug.enabled ) {
+ var str = toActisenseSerialFormat(msg.pgn, buffer, msg.dst, msg.src)
+ debug(str)
+ }
if ( buffer.length > 8 ) {
var pgns = getPlainPGNs(buffer)
|
chore: add debugging of toPgn conversion results (#<I>)
|
canboat_canboatjs
|
train
|
js
|
8fa475b2e4e2d85804a39f3f73d3a70b2b5e239b
|
diff --git a/easytrader/ht_clienttrader.py b/easytrader/ht_clienttrader.py
index <HASH>..<HASH> 100644
--- a/easytrader/ht_clienttrader.py
+++ b/easytrader/ht_clienttrader.py
@@ -55,7 +55,7 @@ class HTClientTrader(ClientTrader):
self._app.top_window().type_keys('%Y')
# detect login is success or not
- self._app.top_window().wait_not('exists', 2)
+ self._app.top_window().wait_not('exists', 10)
self._app = pywinauto.Application().connect(path=self._run_exe_path(exe_path), timeout=10)
self._close_prompt_windows()
diff --git a/easytrader/yh_clienttrader.py b/easytrader/yh_clienttrader.py
index <HASH>..<HASH> 100644
--- a/easytrader/yh_clienttrader.py
+++ b/easytrader/yh_clienttrader.py
@@ -58,7 +58,7 @@ class YHClientTrader(ClientTrader):
# detect login is success or not
try:
- self._app.top_window().wait_not('exists', 2)
+ self._app.top_window().wait_not('exists', 10)
break
except:
pass
|
fix: switch window to normal window after login
|
shidenggui_easytrader
|
train
|
py,py
|
28b1c02a179b659eb02d33f79c8aa3a90bf7494a
|
diff --git a/test/InjectorClient.js b/test/InjectorClient.js
index <HASH>..<HASH> 100644
--- a/test/InjectorClient.js
+++ b/test/InjectorClient.js
@@ -7,6 +7,9 @@ describe('InjectorClient', function() {
var injectorClient, debuggerClient, breakedObject;
function setupInjector(done) {
+ // increase the timeout for Travis CI
+ this.timeout(5000);
+
launcher.runPeriodicConsoleLog(false, function(childProcess, client) {
debuggerClient = client;
injectorClient = new InjectorClient({inject: false}, debuggerClient);
@@ -44,6 +47,9 @@ describe('InjectorClient', function() {
var injectorClient, debuggerClient, serverPort;
function setupInjector(done) {
+ // increase the timeout for Travis CI
+ this.timeout(5000);
+
launcher.runPeriodicConsoleLog(false, function(childProcess, client) {
debuggerClient = client;
injectorClient = new InjectorClient({}, debuggerClient);
|
test: increase test timeout in InjectorClient
The test setup takes more than 2 seconds in Travis CI, which caused the
build to fail. The timeout is 5 seconds now, which seems to be enough.
|
node-inspector_node-inspector
|
train
|
js
|
e6971457832ffed3e167d31d33943d8b7414d738
|
diff --git a/src/models/Transaction.js b/src/models/Transaction.js
index <HASH>..<HASH> 100644
--- a/src/models/Transaction.js
+++ b/src/models/Transaction.js
@@ -30,6 +30,7 @@ module.exports = function getSaleClass(config) {
attributes,
};
+ attributes.description = Transaction.descriptionFilter(attributes.description);
attributes.humanDescription = this.remapDescription(attributes.type);
const { error } = validator.validateSync('Transaction', data);
@@ -60,6 +61,12 @@ module.exports = function getSaleClass(config) {
}
}
+ static matchRedundantInfo = /(?:\s*(?:for|with)\s*info@cappasity.com)/gi;
+
+ static descriptionFilter(description) {
+ return description.replace(Transaction.matchRedundantInfo, '')
+ }
+
/**
* Serializes object into json:api output
* @param {Boolean} host - resource location
|
fix: rid of redundant information in descriptions of transactions
|
makeomatic_restify-utils
|
train
|
js
|
79eefceb09e211bd7b8c2dd84705199d0595de0a
|
diff --git a/test/test.js b/test/test.js
index <HASH>..<HASH> 100644
--- a/test/test.js
+++ b/test/test.js
@@ -52,11 +52,11 @@ describe('when the stream is null', function () {
return stream.send(key)
})
- it('should have saved the file', function (done) {
+ it('should not have saved the file', function (done) {
stream.client.getFile(key, function (err, res) {
if (err) return done(err)
- assert.equal(res.statusCode, 403)
+ assert(res.statusCode === 403 || res.statusCode === 404)
done()
})
})
|
test: fix test as s3 seems to return the correct status code now
|
collectors_s3-buffer-stream
|
train
|
js
|
2aa94e9c3b2801824a471c2ab6d594eda732a135
|
diff --git a/allauth/socialaccount/providers/stackexchange/views.py b/allauth/socialaccount/providers/stackexchange/views.py
index <HASH>..<HASH> 100644
--- a/allauth/socialaccount/providers/stackexchange/views.py
+++ b/allauth/socialaccount/providers/stackexchange/views.py
@@ -22,6 +22,7 @@ class StackExchangeOAuth2Adapter(OAuth2Adapter):
params={'access_token': token.token,
'key': app.key,
'site': site})
+ resp.raise_for_status()
extra_data = resp.json()['items'][0]
return self.get_provider().sociallogin_from_response(request,
extra_data)
|
fix(stackexchange): Check API response code
|
pennersr_django-allauth
|
train
|
py
|
09028dc17e2ee95f1ffc382bdef2e1c54a551d87
|
diff --git a/src/Task/Runner.php b/src/Task/Runner.php
index <HASH>..<HASH> 100644
--- a/src/Task/Runner.php
+++ b/src/Task/Runner.php
@@ -81,7 +81,8 @@ class Runner
7 => self::TASK_RETRY_ALWAYS, // CURLE_COULDNT_CONNECT
28 => self::TASK_RETRY_ALWAYS, // CURLE_OPERATION_TIMEOUTED
35 => self::TASK_RETRY_ALWAYS, // CURLE_SSL_CONNECT_ERROR
- 52 => self::TASK_RETRY_ALWAYS // CURLE_GOT_NOTHING
+ 52 => self::TASK_RETRY_ALWAYS, // CURLE_GOT_NOTHING
+ 'lighthouseError' => self::TASK_RETRY_NEVER
];
/**
|
fix: exclude PageSpeed Insights API errors from retries. (#<I>)
|
googleapis_google-api-php-client
|
train
|
php
|
2dbb8ff37c0f08356e6ee32f381bd50ce42af41a
|
diff --git a/packages/xod-fs/src/load.js b/packages/xod-fs/src/load.js
index <HASH>..<HASH> 100644
--- a/packages/xod-fs/src/load.js
+++ b/packages/xod-fs/src/load.js
@@ -138,7 +138,7 @@ export const getProjects = workspacePath => R.composeP(
// `filename` path relative to `dir`
// :: String -> String -> Promise (Pair String String)
const readImplFile = dir => filename =>
- fs.readFile(path.resolve(dir, filename)).then(content => [
+ fs.readFile(path.resolve(dir, filename), 'utf8').then(content => [
filename,
content,
]);
|
fix(xod-fs): fix bug on loading implementation files (define encoding for `fs.readFile` to ensure getting String, not a Buffer)
|
xodio_xod
|
train
|
js
|
9cb3848e51dc6d416e18f35d4cc3b1f1443e327f
|
diff --git a/cid.go b/cid.go
index <HASH>..<HASH> 100644
--- a/cid.go
+++ b/cid.go
@@ -337,8 +337,8 @@ func (c Cid) Type() uint64 {
}
// String returns the default string representation of a
-// Cid. Currently, Base58 is used as the encoding for the
-// multibase string.
+// Cid. Currently, Base32 is used for CIDV1 as the encoding for the
+// multibase string, Base58 is used for CIDV0.
func (c Cid) String() string {
switch c.Version() {
case 0:
|
chore: make comment reflect the code
During the change to default CIDV1 to Base<I>, the corresponding code
comment wasn't updated.
|
ipfs_go-cid
|
train
|
go
|
7bc4fc72c74cd92802e42595e57f942eb70cb51e
|
diff --git a/conf/rollup.config.js b/conf/rollup.config.js
index <HASH>..<HASH> 100644
--- a/conf/rollup.config.js
+++ b/conf/rollup.config.js
@@ -47,7 +47,7 @@ const common = {
|| id.indexOf('d3') === 0
|| id.indexOf('@nivo') === 0
|| id.indexOf('lodash') === 0
- || id.indexOf('recompose') === 0,
+ || id.indexOf('@react-spring') === 0,
}
const commonPlugins = [
|
fix(build): mark @react-spring packages as external (#<I>)
|
plouc_nivo
|
train
|
js
|
d1d0db43eaa50d62c10350a6ea62ca4554404ec5
|
diff --git a/src/providers/github.js b/src/providers/github.js
index <HASH>..<HASH> 100644
--- a/src/providers/github.js
+++ b/src/providers/github.js
@@ -6,7 +6,31 @@ export default function GitHub(options) {
type: "oauth",
authorization: "https://github.com/login/oauth/authorize?scope=read:user+user:email",
token: "https://github.com/login/oauth/access_token",
- userinfo: "https://api.github.com/user",
+ userinfo: {
+ url: "https://api.github.com/user",
+ async request({ client, tokens }) {
+ // Get base profile
+ const profile = await client.userinfo(tokens)
+
+ // If user has email hidden, get their primary email from the GitHub API
+ if (!profile.email) {
+ const emails = await (
+ await fetch("https://api.github.com/user/emails", {
+ headers: { Authorization: `token ${tokens.access_token}` },
+ })
+ ).json()
+
+ if (emails?.length > 0) {
+ // Get primary email
+ profile.email = emails.find(email => email.primary)?.email;
+ // And if for some reason it doesn't exist, just use the first
+ if (!profile.email) profile.email = emails[0].email;
+ }
+ }
+
+ return profile
+ },
+ },
profile(profile) {
return {
id: profile.id.toString(),
|
feat(providers): ensure GitHub provider always gives an email (#<I>)
* Ensure that GitHub provider always gives an email
* Update src/providers/github.js
|
iaincollins_next-auth
|
train
|
js
|
c2e0a6494a31da36f63bd9b43026c6617df065c6
|
diff --git a/src/Http/Resources/Json/ResourceMethodExtension.php b/src/Http/Resources/Json/ResourceMethodExtension.php
index <HASH>..<HASH> 100644
--- a/src/Http/Resources/Json/ResourceMethodExtension.php
+++ b/src/Http/Resources/Json/ResourceMethodExtension.php
@@ -13,13 +13,11 @@ declare(strict_types=1);
namespace NunoMaduro\LaravelCodeAnalyse\Http\Resources\Json;
+use ReflectionClass;
use Illuminate\Database\Eloquent\Model;
-use Mockery;
use PHPStan\Reflection\ClassReflection;
-use PHPStan\Reflection\MethodReflection;
use Illuminate\Http\Resources\Json\Resource;
use NunoMaduro\LaravelCodeAnalyse\AbstractExtension;
-use ReflectionClass;
final class ResourceMethodExtension extends AbstractExtension
{
|
refactor: removes unused classes
|
nunomaduro_larastan
|
train
|
php
|
0ec6a43816a9f8052cfd0c479919147f56f95644
|
diff --git a/lib/entities/entry.js b/lib/entities/entry.js
index <HASH>..<HASH> 100644
--- a/lib/entities/entry.js
+++ b/lib/entities/entry.js
@@ -31,6 +31,7 @@ export function wrapEntry (data) {
* @prop {number} limit
* @prop {Array<Entities.Entry>} items
* @prop {Array<Object>=} errors - Array of errors that might occur when retrieving entries.
+ * @prop {Object<Array>=} includes - Object with arrays of includes for Entries and Assets. This will be present if resolveLinks is on, and any linked entries or assets exist. Those links will be resolved in the Entries present in the items array, but they are also present here in raw form.
* @prop {function(): Object} toPlainObject() - Returns this Entry collection as a plain JS object
* @prop {function(?function=, space=): Object} stringifySafe(replacer,space) - Stringifies the entry collection, accounting for circular references. Circular references will be replaced with just a Link object, with a <code>circular</code> property set to <code>true</code>. See <a href="https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/JSON/stringify">MDN</a> and <a href="https://www.npmjs.com/package/json-stringify-safe">json-stringify-safe</a> for more details on the arguments this method can take.
*/
|
docs: Document includes property on getEntries response
Fixes #<I>
|
contentful_contentful.js
|
train
|
js
|
13879b354bc9a90a1a6864f74aeb76c002c3d5f0
|
diff --git a/src/Keboola/StorageApi/Client.php b/src/Keboola/StorageApi/Client.php
index <HASH>..<HASH> 100644
--- a/src/Keboola/StorageApi/Client.php
+++ b/src/Keboola/StorageApi/Client.php
@@ -1349,7 +1349,7 @@ class Client
{
$job = $jobCreatedResponse->json();
$maxEndTime = time() + $this->getTimeout();
- $maxWaitPeriod = 60;
+ $maxWaitPeriod = 20;
$retries = 0;
// poll for status
|
feat: handleAsyncTask wait max. <I> secs
|
keboola_storage-api-php-client
|
train
|
php
|
f40cfdda82f9f522d069bd92a21745db7065cc11
|
diff --git a/sync/options.go b/sync/options.go
index <HASH>..<HASH> 100644
--- a/sync/options.go
+++ b/sync/options.go
@@ -1,6 +1,7 @@
package sync
import (
+ "context"
"crypto/tls"
"time"
)
@@ -39,3 +40,10 @@ func WithTLS(t *tls.Config) Option {
o.TLSConfig = t
}
}
+
+// WithContext sets the syncs context, for any extra configuration
+func WithContext(c context.Context) Option {
+ return func(o *Options) {
+ o.Context = c
+ }
+}
diff --git a/sync/sync.go b/sync/sync.go
index <HASH>..<HASH> 100644
--- a/sync/sync.go
+++ b/sync/sync.go
@@ -2,6 +2,7 @@
package sync
import (
+ "context"
"crypto/tls"
"errors"
"time"
@@ -39,6 +40,7 @@ type Options struct {
Nodes []string
Prefix string
TLSConfig *tls.Config
+ Context context.Context
}
type Option func(o *Options)
|
feat(sync): Add additional context option to sync (#<I>)
|
micro_go-micro
|
train
|
go,go
|
9d17dc372f1a328142b5997e3b18ea74694d57a9
|
diff --git a/lib/uploads/index.js b/lib/uploads/index.js
index <HASH>..<HASH> 100644
--- a/lib/uploads/index.js
+++ b/lib/uploads/index.js
@@ -20,6 +20,8 @@ if (!fs.existsSync(UPLOADS_ROOT)) {
fs.mkdirSync(UPLOADS_ROOT);
}
+var foldersToWatch = [];
+
bundles.all().forEach(function (bundle) {
if (bundle.uploads && bundle.uploads.enabled) {
/* istanbul ignore next: Simple directory creation. */
@@ -32,10 +34,12 @@ bundles.all().forEach(function (bundle) {
defaultValue: [],
persistent: false
});
+
+ foldersToWatch.push(bundleUploadsPath + '/**/*');
}
});
-var watcher = chokidar.watch(path.join(UPLOADS_ROOT, '*/*'), {
+var watcher = chokidar.watch(foldersToWatch, {
ignored: /[\/\\]\./
});
|
fix(uploads): prevent crash related to uninstalled bundles
When a bundle is uninstalled but its uploads are still present, NodeCG would attempt to access an undefined Replicant and crash.
|
nodecg_nodecg
|
train
|
js
|
73c9d4148e3f14ef10bbc44da25085c058721550
|
diff --git a/data/preferences.js b/data/preferences.js
index <HASH>..<HASH> 100644
--- a/data/preferences.js
+++ b/data/preferences.js
@@ -64,7 +64,13 @@ exports.DEFAULT_FIREFOX_PREFS = {
// Point the url-classifier to a nonexistent local URL for fast failures.
"browser.safebrowsing.provider.0.gethashURL" : "http://localhost/safebrowsing-dummy/gethash",
"browser.safebrowsing.provider.0.keyURL" : "http://localhost/safebrowsing-dummy/newkey",
- "browser.safebrowsing.provider.0.updateURL" : "http://localhost/safebrowsing-dummy/update"
+ "browser.safebrowsing.provider.0.updateURL" : "http://localhost/safebrowsing-dummy/update",
+
+ // Disable self repair/SHIELD
+ "browser.selfsupport.url": "https://localhost/selfrepair",
+
+ // Disable Reader Mode UI tour
+ "browser.reader.detectedFirstArticle": true
};
// When launching a temporary new Thunderbird profile, use these preferences.
|
fix: Avoid JS errors from external file from normandy.cdn.mozilla.org (#<I>)
* fix: Avoid JS errors from external file from normandy.cdn.mozilla.org
* fix: Disable reader mode UI tour in temporary profiles
|
mozilla-jetpack_jpm
|
train
|
js
|
305302292d232710f7007adb33cab30e49f60da5
|
diff --git a/contextMenu.js b/contextMenu.js
index <HASH>..<HASH> 100644
--- a/contextMenu.js
+++ b/contextMenu.js
@@ -289,10 +289,8 @@ angular.module('ui.bootstrap.contextMenu', [])
function removeAllContextMenus(e) {
$(document.body).off('mousedown', removeOnOutsideClickEvent);
$(document).off('scroll', removeOnScrollEvent);
- if ( !$(event.currentTarget).hasClass('context') ) {
- $(event.currentTarget).removeClass('context');
- removeContextMenus();
- }
+ $(event.originalTarget).removeClass('context');
+ removeContextMenus();
}
if(level === 0) {
|
fix: remove all contextMenus regardless of context class
Remove the context menus even if the current object does not have the
class 'context' since it doesn't seem to be required.
Closes #<I>
|
Templarian_ui.bootstrap.contextMenu
|
train
|
js
|
35f7f20b3e57e4e2e542afa7956de6b5f55414cc
|
diff --git a/packages/eslint-config-peakfijn/index.js b/packages/eslint-config-peakfijn/index.js
index <HASH>..<HASH> 100644
--- a/packages/eslint-config-peakfijn/index.js
+++ b/packages/eslint-config-peakfijn/index.js
@@ -38,10 +38,9 @@ module.exports = {
int: 'number',
Number: 'number',
Object: 'Object',
- 'React.Component': 'ReactNode',
- 'React.ReactElement': 'ReactElement',
- 'React.ReactNode': 'ReactNode',
- 'react/element': 'ReactElement',
+ ReactComponent: 'React.Component',
+ ReactElement: 'React.Element',
+ ReactNode: 'React.Node',
String: 'string',
Undefined: 'undefined',
},
|
refactor(eslint): simplify jsdoc preferred types for react based on flow (#<I>)
|
Peakfijn_Conventions
|
train
|
js
|
b7f29dc67c872354c4b36e03c736e235972c7e5a
|
diff --git a/lib/env_pull_request/test_helper.rb b/lib/env_pull_request/test_helper.rb
index <HASH>..<HASH> 100644
--- a/lib/env_pull_request/test_helper.rb
+++ b/lib/env_pull_request/test_helper.rb
@@ -46,6 +46,9 @@ module EnvPullRequest
#
# @see Base#fetch_pull_request_id
def restore_env_pull_request
+ @original_travis_pull_request ||= nil
+ @original_circle_pr_number ||= nil
+ @original_ghprb_pull_id ||= nil
ENV['TRAVIS_PULL_REQUEST'] = @original_travis_pull_request
ENV['CIRCLE_PR_NUMBER'] = @original_circle_pr_number
ENV['ghprbPullId'] = @original_ghprb_pull_id
|
chore(helper): remove ruby warning
warning: instance variable @original_travis_pull_request not initialized
and simular warning.
|
packsaddle_ruby-env_pull_request
|
train
|
rb
|
73e35592877820875d334f33a915651ee0530191
|
diff --git a/core/corehttp/webui.go b/core/corehttp/webui.go
index <HASH>..<HASH> 100644
--- a/core/corehttp/webui.go
+++ b/core/corehttp/webui.go
@@ -1,7 +1,7 @@
package corehttp
// TODO: move to IPNS
-const WebUIPath = "/ipfs/bafybeifuexpvt6g4bkbmsrlb7rnudskfakn6vdrtoja4ml4zbxne2hu6bq" // v2.12.2
+const WebUIPath = "/ipfs/bafybeid26vjplsejg7t3nrh7mxmiaaxriebbm4xxrxxdunlk7o337m5sqq" // v2.12.3
// this is a list of all past webUI paths.
var WebUIPaths = []string{
|
fix: webui-<I>
This bumps webui to bugfix release based on feedback from <I>-rc1:
<URL>
|
ipfs_go-ipfs
|
train
|
go
|
5212d8846dce5369bc557e1bbaf29f997a71fe76
|
diff --git a/src/addons/Parallax.js b/src/addons/Parallax.js
index <HASH>..<HASH> 100644
--- a/src/addons/Parallax.js
+++ b/src/addons/Parallax.js
@@ -15,7 +15,7 @@ const START_TRANSLATE = 'translate(0px,0px)'
export class ParallaxLayer extends React.PureComponent {
static propTypes = {
- /** Size of a page, (1=100%, 1.5=1 and 1/5, ...) */
+ /** Size of a page, (1=100%, 1.5=1 and 1/2, ...) */
factor: PropTypes.number,
/** Determines where the layer will be at when scrolled to (0=start, 1=1st page, ...) */
offset: PropTypes.number,
|
fix: typo in parallax propstable
|
react-spring_react-spring
|
train
|
js
|
c5d1b9e69ae38f09cbb110965d6eb320eb45e65c
|
diff --git a/src/RemoteProcessor.php b/src/RemoteProcessor.php
index <HASH>..<HASH> 100644
--- a/src/RemoteProcessor.php
+++ b/src/RemoteProcessor.php
@@ -64,7 +64,7 @@ abstract class RemoteProcessor
.implode(PHP_EOL, $env).PHP_EOL
.'set -e'.PHP_EOL
.$task->script.PHP_EOL
- .$delimiter
+ .$delimiter,
]);
}
}
|
fix(Process): Resolve composer lint
|
laravel_envoy
|
train
|
php
|
01309f92f5613c6ceb47f62fa1a5284486404051
|
diff --git a/scripts/generate-base-colors.js b/scripts/generate-base-colors.js
index <HASH>..<HASH> 100644
--- a/scripts/generate-base-colors.js
+++ b/scripts/generate-base-colors.js
@@ -18,9 +18,9 @@ const variablePrefix = '--color';
const colors = [JSON.parse(file)];
-const declarationsBody = colors.reduce((acc, curr) => {
+const declarationsBody = colors.reduce((acc, colorGroup) => {
// Transforms the colors object into base colors
- Object.entries(curr).forEach(([colorName, variations]) => {
+ Object.entries(colorGroup).forEach(([colorName, variations]) => {
Object.entries(variations).forEach(
([variationNumber, lightnessPercentage]) => {
// defines the main colors variables
|
fix(generate-base-colors): changes curr var to color group
|
commercetools_ui-kit
|
train
|
js
|
92be90b0e7ccc385c6558e778f84f5cb04f5ff38
|
diff --git a/internal/service/batch/sweep.go b/internal/service/batch/sweep.go
index <HASH>..<HASH> 100644
--- a/internal/service/batch/sweep.go
+++ b/internal/service/batch/sweep.go
@@ -261,7 +261,7 @@ func sweepSchedulingPolicies(region string) error {
input := &batch.ListSchedulingPoliciesInput{}
var sweeperErrs *multierror.Error
- err = conn.ListSchedulingPolicies(input, func(page *batch.ListSchedulingPoliciesOutput, lastPage bool) bool {
+ err = conn.ListSchedulingPoliciesPages(input, func(page *batch.ListSchedulingPoliciesOutput, lastPage bool) bool {
if page == nil {
return !lastPage
}
|
fix(batch): use pages api for sweeper
|
terraform-providers_terraform-provider-aws
|
train
|
go
|
3404aac765334b89665ea61ecc54e0ef6d3432fc
|
diff --git a/jobs/reset.js b/jobs/reset.js
index <HASH>..<HASH> 100644
--- a/jobs/reset.js
+++ b/jobs/reset.js
@@ -88,7 +88,7 @@ module.exports = async function ({ repositoryFullName }) {
inclusive_end: true
})
const openIssues = issues.rows.filter(issue => issue.doc.state !== 'closed')
- log.info(`Started closing ${openIssues} open issues`)
+ log.info(`Started closing ${openIssues.length} open issues`, { issues: openIssues })
for (let issue of openIssues) {
try {
|
docs(reset): log the length of the issue array
|
greenkeeperio_greenkeeper
|
train
|
js
|
89910f6ef17be38b59bc066d754793cc65a84624
|
diff --git a/test/mysql.test.js b/test/mysql.test.js
index <HASH>..<HASH> 100644
--- a/test/mysql.test.js
+++ b/test/mysql.test.js
@@ -84,6 +84,15 @@ describe('test/mysql.test.js', () => {
assert(row.id === user.id);
});
+ it('should query one desc is NULL success', function* () {
+ const user = yield app.mysql.queryOne('select * from npm_auth where `desc` is NULL');
+ assert(user);
+ assert(typeof user.user_id === 'string' && user.user_id);
+
+ const row = yield app.mysql.get('npm_auth', { desc: null });
+ assert(row.id === user.id);
+ });
+
it('should query one not exists return null', function* () {
let user = yield app.mysql.queryOne('select * from npm_auth where id = -1');
assert(!user);
|
test: add null query test case (#<I>)
|
eggjs_egg-mysql
|
train
|
js
|
46feb215810255f2c84400e3b32ce9e7d7fdf052
|
diff --git a/src/utils/suggestions.js b/src/utils/suggestions.js
index <HASH>..<HASH> 100644
--- a/src/utils/suggestions.js
+++ b/src/utils/suggestions.js
@@ -71,7 +71,7 @@ const getPredictiveSuggestions = ({
if (regexExecution) {
const matchedString = parsedContent.slice(regexExecution.index, parsedContent.length);
- const suggestionPhrase = `${currentValueTrimmed}<mark class="highlight">${matchedString
+ const suggestionPhrase = `<em>${currentValueTrimmed}</em><mark class="highlight predictive-suggestions">${matchedString
.slice(currentValueTrimmed.length)
.split(' ')
.slice(0, wordsToShowAfterHighlight + 1)
@@ -234,7 +234,21 @@ const getSuggestions = ({
currentValue,
wordsToShowAfterHighlight,
});
- suggestionsList = predictiveSuggestions;
+
+ const finalSuggestions = [];
+
+ const predictiveValues = predictiveSuggestions.reduce((arr, suggestion) => {
+ finalSuggestions.push(suggestion);
+ arr.push(suggestion.value);
+ return arr;
+ }, []);
+
+ suggestionsList.forEach((suggestion) => {
+ if (!predictiveValues.includes(suggestion.value)) {
+ finalSuggestions.push(suggestion);
+ }
+ });
+ suggestionsList = finalSuggestions;
}
return suggestionsList;
};
|
fix: changed suggestions to return both predictive and normal suggestions when the former is enabled
|
appbaseio_reactivecore
|
train
|
js
|
a08cdfc0bde3a866a712999741dc729f023d2eb7
|
diff --git a/test/test_helper.rb b/test/test_helper.rb
index <HASH>..<HASH> 100644
--- a/test/test_helper.rb
+++ b/test/test_helper.rb
@@ -7,9 +7,7 @@ Bundler.require :test
require "minitest/autorun"
require "imgix"
require "webmock/minitest"
-
-require 'minitest/autorun'
-require 'minitest/reporters' # requires the gem
+require "minitest/reporters"
Minitest::Reporters.use! Minitest::Reporters::SpecReporter.new
|
test: remove comment and duplicated import
|
imgix_imgix-rb
|
train
|
rb
|
21cfd7e363309b5712a396f9d3d876010eb047bb
|
diff --git a/index.js b/index.js
index <HASH>..<HASH> 100644
--- a/index.js
+++ b/index.js
@@ -42,7 +42,7 @@ class Homefront {
/**
* Recursively merges given sources into data.
*
- * @param {{}[]} sources One or more, or array of, objects to merge into data (left to right).
+ * @param {...Object} sources One or more, or array of, objects to merge into data (left to right).
*
* @return {Homefront}
*/
@@ -68,6 +68,19 @@ class Homefront {
}
/**
+ * Static version of merge, allowing you to merge objects together.
+ *
+ * @param {...Object} sources One or more, or array of, objects to merge (left to right).
+ *
+ * @return {{}}
+ */
+ static merge(sources) {
+ sources = Array.isArray(sources) ? sources : Array.prototype.slice.call(arguments);
+
+ return extend.apply(extend, [true, {}].concat(sources));
+ }
+
+ /**
* Sets the mode.
*
* @param {String} [mode] Defaults to nested.
|
feat(merge): Added static merge to be used outside of statham instance
|
SpoonX_homefront
|
train
|
js
|
bf04e3d2f6cba980e19df38eda9118a7d2cedd93
|
diff --git a/pgjdbc/src/main/java/org/postgresql/core/v3/ConnectionFactoryImpl.java b/pgjdbc/src/main/java/org/postgresql/core/v3/ConnectionFactoryImpl.java
index <HASH>..<HASH> 100644
--- a/pgjdbc/src/main/java/org/postgresql/core/v3/ConnectionFactoryImpl.java
+++ b/pgjdbc/src/main/java/org/postgresql/core/v3/ConnectionFactoryImpl.java
@@ -239,10 +239,8 @@ public class ConnectionFactoryImpl extends ConnectionFactory {
tryConnect(user, database, info, socketFactory, hostSpec, SslMode.DISABLE,gssEncMode);
LOGGER.log(Level.FINE, "Downgraded to non-encrypted connection for host {0}",
hostSpec);
- } catch (SQLException ee) {
+ } catch (SQLException | IOException ee) {
ex = ee;
- } catch (IOException ee) {
- ex = ee; // Can't use multi-catch in Java 6 :(
}
if (ex != null) {
log(Level.FINE, "sslMode==PREFER, however non-SSL connection failed as well", ex);
|
refactor: Use multi-catch for exceptions in ConnectionFactoryImpl
|
pgjdbc_pgjdbc
|
train
|
java
|
9079d3e1e566c5ca32311addcf4dfb5fcf692c9a
|
diff --git a/packages/@vue/cli-ui/apollo-server/connectors/dependencies.js b/packages/@vue/cli-ui/apollo-server/connectors/dependencies.js
index <HASH>..<HASH> 100644
--- a/packages/@vue/cli-ui/apollo-server/connectors/dependencies.js
+++ b/packages/@vue/cli-ui/apollo-server/connectors/dependencies.js
@@ -60,7 +60,7 @@ function findDependencies (deps, type, file, context) {
id => ({
id,
versionRange: deps[id],
- installed: fs.existsSync(getPath({ id, file })),
+ installed: isInstalled({ id, file }),
website: getLink({ id, file }, context),
type,
baseFir: file
@@ -70,9 +70,15 @@ function findDependencies (deps, type, file, context) {
function getPath ({ id, file = cwd.get() }) {
const filePath = resolveModule(path.join(id, 'package.json'), file)
+ if (!filePath) return
return resolveModuleRoot(filePath, id)
}
+function isInstalled ({ id, file = cwd.get() }) {
+ const resolvedPath = getPath({ id, file })
+ return resolvedPath && fs.existsSync(resolvedPath)
+}
+
function readPackage ({ id, file }, context) {
try {
return folders.readPackage(getPath({ id, file }), context)
|
fix(deps): more robust isInstalled
|
vuejs_vue-cli
|
train
|
js
|
5cb838ce31a698a7a9ea354a5b7e488b474be02a
|
diff --git a/tooling/commands/list.js b/tooling/commands/list.js
index <HASH>..<HASH> 100644
--- a/tooling/commands/list.js
+++ b/tooling/commands/list.js
@@ -17,20 +17,25 @@ module.exports = {
type: `boolean`
}
},
+ // eslint-disable-next-line complexity
handler: wrapHandler(async ({fortests, forpipeline}) => {
let packages;
if (fortests) {
const changed = await updated({});
if (changed.includes(`tooling`)) {
packages = await list();
- packages.push(`legacy`);
}
else {
packages = await updated({dependents: true});
}
+ // We don't want to run legacy tests in validated merge
+ packages = packages.filter((p) => p !== `legacy`);
}
else {
packages = await list();
+ if (forpipeline) {
+ packages.push(`legacy`);
+ }
}
if (forpipeline || fortests) {
|
fix(tooling): do not run legacy tests during validated merge
Also, correctly run legacy tests in pipeline
|
webex_spark-js-sdk
|
train
|
js
|
812bff1cc3d1e814a37eb0683a0a7044bd3ac86b
|
diff --git a/test/ddescribe-iit.spec.js b/test/ddescribe-iit.spec.js
index <HASH>..<HASH> 100644
--- a/test/ddescribe-iit.spec.js
+++ b/test/ddescribe-iit.spec.js
@@ -513,7 +513,7 @@ describe('gulp-ddescribe-iit', function() {
var mockFile = new File({
path: 'mock-file.js',
contents: new Buffer(
- '\n\t describe \t.\n\n\t\t\tonly();\n//This should look good!')
+ '\n\t describe \t.\n\n\t\t\tonly(function() { \n});//This should look good!')
});
stream = ddescribeIit({ noColor: true });
var called = false;
@@ -526,9 +526,9 @@ describe('gulp-ddescribe-iit', function() {
" 2| describe .",
" | ^^^^^^^^^^^^^^^",
" 2| ",
- " 3| only();",
+ " 3| only(function() { ",
" | ^^^^",
- " 5| //This should look good!",
+ " 5| });//This should look good!",
""
].join('\n'));
}));
|
test(ddescribe-iit): make that last test slightly more convincing
|
caitp_gulp-ddescribe-iit
|
train
|
js
|
2c1fe1ed15b535ba458e607badc931c7f7bd71f8
|
diff --git a/tests/test_presenter.py b/tests/test_presenter.py
index <HASH>..<HASH> 100644
--- a/tests/test_presenter.py
+++ b/tests/test_presenter.py
@@ -185,12 +185,12 @@ def test_compare_default_match(compare_url, text, prefix, expected, markdown_pre
),
("## v3 \n## v2\n## v1", "([1-9])", "v", "## v3 \n## v2\n## v1"),
(
- "# v3 \n# v2\n# v1",
- "(?P<version>[1-9])",
- "v",
- "# [v3](https://github.com/LeMimit/auto-changelog/compare/v2...v3)"
- " \n# [v2](https://github.com/LeMimit/auto-changelog/compare/v1...v2)\n# v1",
- )
+ "# v3 \n# v2\n# v1",
+ "(?P<version>[1-9])",
+ "v",
+ "# [v3](https://github.com/LeMimit/auto-changelog/compare/v2...v3)"
+ " \n# [v2](https://github.com/LeMimit/auto-changelog/compare/v1...v2)\n# v1",
+ ),
],
)
def test_compare_custom_match(compare_url, text, tag_pattern, prefix, expected, markdown_presenter):
|
fix: take into account full specification of semver spec
|
Michael-F-Bryan_auto-changelog
|
train
|
py
|
c422764ab9050a74d8715a5d1487b50dc2856a7b
|
diff --git a/src/rest/delete.js b/src/rest/delete.js
index <HASH>..<HASH> 100644
--- a/src/rest/delete.js
+++ b/src/rest/delete.js
@@ -8,6 +8,6 @@ export default (req, MongooseModel) => new Promise((resolve, reject) => {
return reject({ status: 404 }, { text: 'Not Found' });
}
- return resolve();
+ return resolve({ status: 204 });
});
});
diff --git a/test/rest.delete.js b/test/rest.delete.js
index <HASH>..<HASH> 100644
--- a/test/rest.delete.js
+++ b/test/rest.delete.js
@@ -18,7 +18,8 @@ describe('rest.delete', function() {
it('should delete resource if it exist', async function() {
const res = await request(host).del(`/book(${data[0].id})`);
- res.status.should.be.equal(200);
+ console.log(res)
+ res.status.should.be.equal(204);
});
it('should be 404 if resource not exist', async function() {
const res = await request(host).del(`/book(not-exist-id)`);
|
fix(rest): fixed #<I>, DELETE operation should return <I>.
|
TossShinHwa_node-odata
|
train
|
js,js
|
429dd54ce5079ad00f63e09711475143aeba4f3f
|
diff --git a/webapps/client/scripts/tasklist/plugins/cam-tasklist-search-plugin.js b/webapps/client/scripts/tasklist/plugins/cam-tasklist-search-plugin.js
index <HASH>..<HASH> 100644
--- a/webapps/client/scripts/tasklist/plugins/cam-tasklist-search-plugin.js
+++ b/webapps/client/scripts/tasklist/plugins/cam-tasklist-search-plugin.js
@@ -45,8 +45,9 @@ define([
if(['Like', 'Before', 'After'].indexOf(operator) !== -1) {
out += operator;
}
- if(expressionsRegex.test(value)) {
- out += 'Expession';
+ if(expressionsRegex.test(value) &&
+ ['assignee', 'owner', 'candidateGroup', 'candidateUser', 'involvedUser'].indexOf(type) !== -1) {
+ out += 'Expression';
}
return out;
};
|
fix(search): use correct Expression suffix
related to CAM-<I>
|
camunda_camunda-bpm-platform
|
train
|
js
|
7926f91a1add374f176fbf2a53368e5d866496b2
|
diff --git a/manifest.php b/manifest.php
index <HASH>..<HASH> 100755
--- a/manifest.php
+++ b/manifest.php
@@ -47,7 +47,7 @@ return [
'author' => 'Open Assessment Technologies',
'requires' => [
'taoItems' => '>=10.19.0',
- 'tao' => '>=45.12.0',
+ 'tao' => '>=46.13.0',
'generis' => '>=12.17.0',
],
'models' => [
|
chore: update tao version in dependencies
|
oat-sa_extension-tao-itemqti
|
train
|
php
|
597105fc2e8048df71f745f27eac6f6eed8e1417
|
diff --git a/packages/card/src/css/index.js b/packages/card/src/css/index.js
index <HASH>..<HASH> 100644
--- a/packages/card/src/css/index.js
+++ b/packages/card/src/css/index.js
@@ -140,6 +140,11 @@ export default {
transition: `opacity ${core.motion.speedNormal}`,
pointerEvents: 'none',
opacity: 0,
+ zIndex: 10,
+
+ '> *': {
+ pointerEvents: 'all'
+ },
'&:focus-within, &[focus-within]': {
opacity: 1
@@ -154,7 +159,6 @@ export default {
// __action-bar__button
'.psds-card__action-bar__button': {
- pointerEvents: 'all',
fontSize: core.type.fontSizeXSmall,
padding: 0,
cursor: 'pointer',
|
fix(card): fix action-bar z-index & pointer events
- Action bar needs a z-index > 0 because it may contain an action menu that
overflows the card. The z-index will prevent the menu from going behind other
cards.
- Set pointer-events on ALL children of action-bar to be "all" so users
can interact with them. Allowed us to remove the pointer-events on
action-bar-buttons specifically.
|
pluralsight_design-system
|
train
|
js
|
fa27bea67b5dee64adb5281a987b2d7feb4d70ab
|
diff --git a/lib/sonos.js b/lib/sonos.js
index <HASH>..<HASH> 100644
--- a/lib/sonos.js
+++ b/lib/sonos.js
@@ -493,7 +493,7 @@ Sonos.prototype.getPlayMode = async function () {
})
}
-const playmodes = ['NORMAL', 'REPEAT_ALL', 'SHUFFLE', 'SHUFFLE_NOREPEAT']
+const playmodes = ['NORMAL', 'REPEAT_ONE', 'REPEAT_ALL', 'SHUFFLE', 'SHUFFLE_NOREPEAT', 'SHUFFLE_REPEAT_ONE']
/**
* Set Play Mode
* @param {String} playmode
|
feat: add all supported playmodes (#<I>)
|
bencevans_node-sonos
|
train
|
js
|
f5b73cac958d8d62af315f1b72656d61260bd18d
|
diff --git a/test/index.js b/test/index.js
index <HASH>..<HASH> 100644
--- a/test/index.js
+++ b/test/index.js
@@ -163,4 +163,34 @@ describe('hapi-sequelized', function() {
done();
});
});
-});
\ No newline at end of file
+
+ it('should connect to a database using the URI format',
+ function(done) {
+ options.uri = options.dialect + '://' + options.user + ':' + options.pass + '@' + options.host + ':' + options.port + '/' + options.database
+ delete options.host;
+ delete options.port;
+ delete options.dialect;
+ delete options.user;
+ delete options.pass;
+ delete options.database;
+
+ var register = {
+ register: require('..'),
+ options: options
+ };
+
+ server.register([register], function() {
+ var models = server.plugins['hapi-sequelized'].db.sequelize.models;
+
+ // check if all models were imported
+ expect(models.User).to.exist();
+ expect(models.Product).to.exist();
+ expect(models.Category).to.exist();
+ expect(models.BlogRoll).to.exist();
+ expect(models.Blog).to.exist();
+ expect(models.Post).to.exist();
+
+ done();
+ });
+ });
+});
|
test: Add test for connection URI.
Allow for the application of a database connection string whilst
maintaining options for Sequelize.
|
danecando_hapi-sequelize
|
train
|
js
|
bd8dab46c69ec397e204c7a7cb3689d1f692f580
|
diff --git a/tests/test_file.py b/tests/test_file.py
index <HASH>..<HASH> 100644
--- a/tests/test_file.py
+++ b/tests/test_file.py
@@ -60,6 +60,8 @@ def test_basic(): # type: () -> None
fn(cStringIO.StringIO(b))
fn(buffer(b)) # noqa: F821
+ if six.PY3:
+ fn(b)
def test_create_with_url(): # type: () -> None
f = File.create_with_url(
|
test: construct LCFile from bytes literal
|
leancloud_python-sdk
|
train
|
py
|
e79ce1489b05a0bab2eb16f39506536d167237cf
|
diff --git a/src/widgets/hierarchical-menu/hierarchical-menu.js b/src/widgets/hierarchical-menu/hierarchical-menu.js
index <HASH>..<HASH> 100644
--- a/src/widgets/hierarchical-menu/hierarchical-menu.js
+++ b/src/widgets/hierarchical-menu/hierarchical-menu.js
@@ -99,7 +99,7 @@ hierarchicalMenu({
* @property {number} [limit=10] How much facet values to get [*].
* @property {string} [separator=" > "] Separator used in the attributes to separate level values. [*].
* @property {string} [rootPath] Prefix path to use if the first level is not the root level.
- * @property {string} [showParentLevel=false] Show the parent level of the current refined value.
+ * @property {boolean} [showParentLevel=false] Show the parent level of the current refined value.
* @property {string[]|function} [sortBy=['name:asc']] How to sort refinements. Possible values: `count|isRefined|name:asc|name:desc`.
*
* You can also use a sort function that behaves like the standard Javascript [compareFunction](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Array/sort#Syntax).
|
chore(docs): hierarchicalMenu - showParentLevel is boolean (#<I>)
see #<I>
|
algolia_instantsearch.js
|
train
|
js
|
97eddbc3ba58f02f93a34f7be6761f87565ca3f0
|
diff --git a/controller/api/models.py b/controller/api/models.py
index <HASH>..<HASH> 100644
--- a/controller/api/models.py
+++ b/controller/api/models.py
@@ -618,7 +618,12 @@ def _etcd_purge_key(**kwargs):
def _etcd_purge_user(**kwargs):
username = kwargs['instance'].username
- _etcd_client.delete('/deis/builder/users/{}'.format(username), dir=True, recursive=True)
+ try:
+ _etcd_client.delete(
+ '/deis/builder/users/{}'.format(username), dir=True, recursive=True)
+ except KeyError:
+ # If _etcd_publish_key() wasn't called, there is no user dir to delete.
+ pass
def _etcd_publish_domains(**kwargs):
|
fix(controller): ignore KeyError when purging user from etcd
|
deis_deis
|
train
|
py
|
a44bd2c9c10fb80c41578ba1cc208ef258045f72
|
diff --git a/core/src/main/java/org/infinispan/cache/impl/AbstractDelegatingCache.java b/core/src/main/java/org/infinispan/cache/impl/AbstractDelegatingCache.java
index <HASH>..<HASH> 100644
--- a/core/src/main/java/org/infinispan/cache/impl/AbstractDelegatingCache.java
+++ b/core/src/main/java/org/infinispan/cache/impl/AbstractDelegatingCache.java
@@ -426,12 +426,12 @@ public abstract class AbstractDelegatingCache<K, V> implements Cache<K, V> {
@Override
public V computeIfAbsent(K key, Function<? super K, ? extends V> mappingFunction, long lifespan, TimeUnit lifespanUnit) {
- return computeIfAbsent(key, mappingFunction, lifespan, lifespanUnit);
+ return cache.computeIfAbsent(key, mappingFunction, lifespan, lifespanUnit);
}
@Override
public V computeIfAbsent(K key, Function<? super K, ? extends V> mappingFunction, long lifespan, TimeUnit lifespanUnit, long maxIdleTime, TimeUnit maxIdleTimeUnit) {
- return computeIfAbsent(key, mappingFunction, lifespan, lifespanUnit, maxIdleTime, maxIdleTimeUnit);
+ return cache.computeIfAbsent(key, mappingFunction, lifespan, lifespanUnit, maxIdleTime, maxIdleTimeUnit);
}
@Override
|
fix: avoid StackOverflowError caused by computeIfAbsent calling itself instead of the delegate cache
|
infinispan_infinispan
|
train
|
java
|
19a37e6ab85fe2dd9897e6f07666b32ae238492a
|
diff --git a/index.js b/index.js
index <HASH>..<HASH> 100644
--- a/index.js
+++ b/index.js
@@ -411,7 +411,6 @@ exports = module.exports = function () {
o[STACK_LABEL[9]] = PMEarly;
o[STACK_LABEL[10]] = PMOver;
o[STACK_LABEL[11]] = PMAfter;
- console.log((AMBefore + AMOver + AMRest + AMLate + AMWork + AMEarly) == (PMAfter + PMOver + PMRest + PMLate + PMWork + PMEarly));
result.push(o);
} else {
o[STACK_LABEL[0]] = HALF_MINUTE;
|
refactor:remove unless code
|
hankaibo_d3-stack-time
|
train
|
js
|
ab1b6aa8f6c435f7f00fa37a39b0a8826bf0eca4
|
diff --git a/src/gatsby-node.js b/src/gatsby-node.js
index <HASH>..<HASH> 100644
--- a/src/gatsby-node.js
+++ b/src/gatsby-node.js
@@ -2,6 +2,30 @@ const lunr = require("lunr");
const { enhanceLunr } = require("./common.js");
const fs = require("fs");
+
+exports.onCreateWebpackConfig = ({ actions, plugins }, { languages = [] }) => {
+ const languageNames = new Set(languages.map(language => language.name));
+
+ actions.setWebpackConfig({
+ plugins: [
+ plugins.ignore({
+ checkResource(resource, context) {
+ if (/lunr-languages$/.test(context)) {
+ const match = resource.match(/lunr\.(\w+)/);
+ if (match !== null) {
+ const name = match[1];
+ if (!languageNames.has(name)) {
+ // Skip the resource.
+ return true;
+ };
+ }
+ }
+ }
+ })
+ ]
+ });
+}
+
exports.onPostBootstrap = ({ getNodes, getNode }, pluginOptions) => {
const {
languages = [],
|
fix: do not bundle unneeded languages from lunr-languages (#<I>)
* fix: do not bundle unneeded languages from lunr-languages
* fix: wrong return values for webpack ignore plugin
|
humanseelabs_gatsby-plugin-lunr
|
train
|
js
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.