added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T04:34:50.053543
| 2017-01-19T07:52:22
|
201787317
|
{
"authors": [
"artm",
"janjon",
"mspae",
"thesolotraveller"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8849",
"repo": "mspae/react-wavesurfer",
"url": "https://github.com/mspae/react-wavesurfer/issues/28"
}
|
gharchive/issue
|
plugins/timeline Can not introduce
Introduction method
import Timeline from 'react-wavesurfer/lib/plugins/timeline';
error:
React.createElement: type should not be null, undefined, boolean, or number. It should be a string (for DOM elements) or a ReactClass (for composite components
Hello, can you post your code? Which transpilation library and config are you using? Can you maybe create a codepen which reproduces the issue. Thanks!
I'm using electron-compile for transpilation. I'm 95% sure it uses babel for the actual work. I get a different error when trying to import a plugin:
import Regions from "react-wavesurfer/lib/plugins/regions";
Uncaught ReferenceError: WaveSurfer is not defined
at WaveSurfer.Regions.init.wavesurfer.wavesurfer (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:288)
at Object.<anonymous> (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:290)
at __webpack_require__ (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:30)
at Object.module.exports (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:73)
at __webpack_require__ (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:30)
at Object.defineProperty.value (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:50)
at /home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:53
at webpackUniversalModuleDefinition (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:3)
at Object.<anonymous> (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:10)
at Object.<anonymous> (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:831)
ah, and apparently electron-compile uses webpack for which you're describing some extra confguration, I'll have a look into that.
I had a look and now I realize that electron-complile doesn't use webpack, the webpack in the stack trace comes form your code, sorry about the confusion. I don't understand the instructions though, how can I achieve the same effect if I'm using babel for transpilation?
Ok, I've figured it out. In my case (transpilation by babel via electron-compile) the following requires/imports section works:
window.WaveSurfer = require("wavesurfer.js");
let Regions = require("react-wavesurfer/lib/plugins/regions").default;
let Minimap = require("react-wavesurfer/lib/plugins/minimap").default;
import React from "react";
import Wavesurfer from "react-wavesurfer";
import Segmentator from "./segmentator";
Hi !
In my case regions is not loading and it is causing all other components to not load in main app.
Here is the error it shows on including -
import Regions from 'react-wavesurfer/lib/plugins/regions';
bundle.js:42575 Uncaught ReferenceError: WaveSurfer is not defined
at WaveSurfer.Regions.init.wavesurfer.wavesurfer (bundle.js:42575)
at Object. (bundle.js:42577)
at webpack_require (bundle.js:42317)
at Object.module.exports (bundle.js:42360)
at webpack_require (bundle.js:42317)
at Object.defineProperty.value (bundle.js:42337)
at bundle.js:42340
at webpackUniversalModuleDefinition (bundle.js:42290)
at Object. (bundle.js:42297)
at webpack_require (bundle.js:20)
Please refer to this: https://github.com/mspae/react-wavesurfer#prerequisites-and-common-pitfalls – If this doesn't fix it I'll reopen the issue.
|
2025-04-01T04:34:50.064760
| 2019-09-15T08:38:33
|
493711254
|
{
"authors": [
"mstksg",
"rnhmjoj"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8850",
"repo": "mstksg/uncertain",
"url": "https://github.com/mstksg/uncertain/issues/4"
}
|
gharchive/issue
|
Initialize values from a covariance matrix
Sometimes you have a set of correlated numbers (for example parameters estimate of a fitted curve) that are given by a vector and a covariance matrix.
I don't think it's possibile to handle these with the current interface
will look into this! :) thanks for the idea!
|
2025-04-01T04:34:50.123842
| 2021-04-20T16:03:45
|
862974064
|
{
"authors": [
"camchenry",
"kettanaito",
"marcosvega91"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8852",
"repo": "mswjs/data",
"url": "https://github.com/mswjs/data/issues/78"
}
|
gharchive/issue
|
Error when calling findFirst even though nothing has changed
First, thanks for the great library. This solves the exact problem I have been working on, and we're already using MSW so this is a natural fit.
I have been running into an issue where the relations seem to "disappear" from the database, even though nothing has changed? I created a CodeSandbox which consistently crashes:
https://codesandbox.io/s/mswjs-relation-undefined-error-wkcfp?file=/src/App.tsx
import { factory, oneOf, primaryKey } from "@mswjs/data";
import * as faker from "faker";
const db = factory({
user: {
id: primaryKey(faker.datatype.uuid),
firstName: String
},
userObject: {
id: primaryKey(faker.datatype.uuid),
data: String,
user: oneOf("user")
}
});
// Seed the database
const seededUser = db.user.create({
id: "ab4f631c-cca4-498f-a5aa-4828352a7c69",
firstName: "Test"
});
db.userObject.create({
user: seededUser,
data: "test data 1 - associated with user"
});
const queryObject = () => {
const object = db.userObject.findFirst({
where: {
user: {
id: {
equals: "ab4f631c-cca4-498f-a5aa-4828352a7c69"
}
}
}
});
console.log("query:", { object, user: object.user });
};
// This will work OK, user is not undefined
queryObject();
// This will NOT work ok
setTimeout(() => {
console.log("deferred query");
queryObject();
}, 1000);
I make the same queryObject call twice: once immediately after creation, and another 1 second after seeding the database. My expectation is that both invocations of the function should do exactly the same thing. However, what happens is that the first call works as expected, while the second invocation crashes with the error TypeError: can't access property "__type", actualValue is undefined.
Willing to help out with this, let me know what questions you have. Thanks in advance.
(I apologize for all of the issues, but I am desperately trying to figure out this issue, and I have had no luck. I'd love to use this project.)
Hi @camchenry thanks for raising the issue :).
I think that making the property enumerable should solve the issue. I have already done it in the same PR as for the other issue because I need it for deleting internal properties.
@marcosvega91 Sounds great! I will look forward to when that releases then.
Thanks for raising this, @camchenry. You've been immensely helpful in uncovering these behaviors. That's indeed an issue and we should fix it in the upcoming release.
@marcosvega91, I'll take a look at the internal properties removal pull request once more, would love to get it merged. It's a superb piece of work already, so huge thanks! I have one concern, but we'll figure it out.
The issue itself is addressed in #67, but we need to add a test for it. @camchenry, please, would you be interested in contributing the relevant test? We can support you throughout the process and have this not only fixed, but bullet-proof against any future regressions?
@kettanaito Sure! I'll try to get around to it today and put up a PR.
With the test added, is there anything left to address to mak this as done, @camchenry?
@kettanaito i think with the tests written and the manual testing I've done, I think the original issue should be fixed 🤞
@kettanaito Any chance this will make it into a new release sometime soon? This was somewhat of a blocker in my project, but I'd like to give it another go and see if I can uncover any more issues.
Released in 0.3.0. Please give it a try, your feedback is highly valuable!
|
2025-04-01T04:34:50.143928
| 2018-07-09T23:22:29
|
339640451
|
{
"authors": [
"deronnax",
"mszep",
"nitrocode"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8853",
"repo": "mszep/pandoc_resume",
"url": "https://github.com/mszep/pandoc_resume/issues/49"
}
|
gharchive/issue
|
Thank You
Thank you @mszep for starting this project because it helped get me create multiple resumes, cover letters, interviews, which lead to multiple job offers, and finally a job!
I enjoyed adding my own contributions to it and hope it helps others. :smile:
Great to hear @nitrocode, and congrats on your new job! I had hoped this project might help someone in this way, but you're the first confirmed case :-)
And thank you for your thoughtful contributions in other issues -- it's greatly appreciated!
thank you also, @mszep
|
2025-04-01T04:34:50.161727
| 2015-02-27T02:09:50
|
59178800
|
{
"authors": [
"DerManoMann",
"dragonmantank",
"laurencei",
"lioannou",
"mtdowling",
"swekaj",
"taylorotwell"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8855",
"repo": "mtdowling/cron-expression",
"url": "https://github.com/mtdowling/cron-expression/issues/78"
}
|
gharchive/issue
|
No way to pass DateTime with custom TimeZone to isDue
In Laravel's scheduler component, we allow the user to specify a Timezone to evaluate the Cron against. However, cron-expression appears to always overwrite the timezone to the current application timezone, allowing no custom timezone to be set.
Is there any way this can be resolved?
I don't remember exactly why this was added really. This is what I got from the blame: https://github.com/mtdowling/cron-expression/commit/cbde1b5a2974c83a6926a7d387479f0e04bb93c7
I guess it's trying to normalize the timezone. I'm happy to make a change if needed.
It appears we can maybe work around it on our end as well by passing a string representation instead of a DateTime object, though it does seem intuitively like the Timezone of the given DateTime object should be respected.
Maybe I'm wrong there. Here is the PR someone is working on our side (https://github.com/laravel/framework/pull/7636) ... personally I kind of wonder if it should be fixed over here though.
I agree it should be fixed here. The library shouldn't be changing the timezone without giving the user the option to specify the timezone it changes to. It's very important to use the correct timezone when checking to see if a cron expression is due since the expressions themselves are not timezone-aware. If you create expressions for GMT-8:00, then the library will always be off by 8 hours if it's converting any DateTimes to GMT/UTC.
I remember a year or so ago, I modified my local copy of the library to accept an optional timezone in the CronExression::factory() method. That timezone was then used for any ::isDue() and similar calls. Unfortunately that work has been lost (I had forgotten I did that when I updated to using composer), but I'd be willing to recreate it.
I think it'd be most useful to allow providing a default timezone in the ::factory() method as well as allowing the user to override that by passing in a DateTime with the appropriate timezone set in ::isDue() and similar. If no default timezone is supplied, then the system's default is used where necessary.
I could probably work on this this weekend, if you'd like.
Looks like this is not handled consistently across all of the public API.
getRunDate() does just clone the given date/time if it is a DateTime instance. So, calling getMultipleRunDates() with a properly set up $currentTime will respect the timezone as given on $currentTime.
I remember that was something I was working towards as I also use this to evaluate against different timezones.
My stand is that if a DateTime instance is given, the timezone set in that instance should be used. For strings or 'now' the default timezone should be used.
Is there a suggested fix for this? Anything I can do?
It can be fixed. We need a PR that removes all of the timezone modifications from the library and any time a DateTime object will be modified, it's cloned.
We fixed it on our end for now by passing a string representation of the
time.
On Thu, Mar 26, 2015 at 11:38 AM, Michael Dowling<EMAIL_ADDRESS>wrote:
It can be fixed. We need a PR that removes all of the timezone
modifications from the library and any time a DateTime object will be
modified, it's cloned.
—
Reply to this email directly or view it on GitHub
https://github.com/mtdowling/cron-expression/issues/78#issuecomment-86609789
.
@mtdowling @dragonmantank - it looks like someone has made a PR for this issue here https://github.com/mtdowling/cron-expression/pull/115
Any chance of reviewing it - and potentially merging?
I've run into the same issue as @taylorotwell (ironically for inside a Laravel application) that I really need to support custom timezones...
@taylorotwell - this can be closed now as functionality was merged today: https://github.com/mtdowling/cron-expression/pull/134
This should have been corrected with #134, so closing.
|
2025-04-01T04:34:50.170939
| 2021-09-11T20:34:22
|
993936257
|
{
"authors": [
"mtgto",
"wqoq"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8856",
"repo": "mtgto/Unrar.swift",
"url": "https://github.com/mtgto/Unrar.swift/issues/3"
}
|
gharchive/issue
|
"Invalid Exclude" warnings when added to Xcode 13 Beta
I'm getting the following warnings when adding this package to a project in Xcode Beta 13:
Invalid Exclude '.../SourcePackages/checkouts/Unrar.swift/Sources/Cunrar/arcmem.hpp': File not found.
Invalid Exclude '.../SourcePackages/checkouts/Unrar.swift/Sources/Cunrar/arcmem.cpp': File not found.
My understanding is that it's because "arcmem.cpp" and "arcmem.hpp" are listed in the exclude section for the "Cunrar" target in Package.swift, but those files are not in the package.
My understanding is that it's because "arcmem.cpp" and "arcmem.hpp" are listed in the exclude section for the "Cunrar" target in Package.swift, but those files are not in the package.
Thanks for your report! I released v0.3.7.
|
2025-04-01T04:34:50.228485
| 2023-10-25T14:24:01
|
1961554978
|
{
"authors": [
"axkr",
"mtommila"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8860",
"repo": "mtommila/apfloat",
"url": "https://github.com/mtommila/apfloat/issues/30"
}
|
gharchive/issue
|
fresnelC, fresnelS for Apcomplex
Can you please review my fresnelC, fresnelS implementations:
https://github.com/axkr/symja_android_library/blob/3bf9f61f8dcc83b75ea2919d7ac2f6e10863ba25/symja_android_library/matheclipse-core/src/main/java/org/matheclipse/core/expression/ApcomplexNum.java#L222
https://github.com/axkr/symja_android_library/blob/3bf9f61f8dcc83b75ea2919d7ac2f6e10863ba25/symja_android_library/matheclipse-core/src/main/java/org/matheclipse/core/expression/ApcomplexNum.java#L245
Is it worth to move these algorithms to the apfloat library?
There seem to be somewhat simpler looking representations using 1F2 in
https://functions.wolfram.com/GammaBetaErf/FresnelC/26/01/01/
https://functions.wolfram.com/GammaBetaErf/FresnelS/26/01/01/
But your implementation of course works as well (representing the function first through erf and then erf through 1F1 I suppose
I'm not really competent to say anything about performance, accuracy or numerical stability
The only potential problem from my point of view are
new Apfloat(0.5)
new Apfloat(1.5)
new Apfloat(-0.5)
which might not at least in theory be accurate to more than approx. 16 decimal digits. Use the string constructor instead:
new Apfloat("0.5")
new Apfloat("1.5")
new Apfloat("-0.5")
I'm not sure about adding these to the apfloat library. There are of course very many similar cases where one special function can be represented through other similar or more generic special functions. In addition to Fresnel sin and cos, for example exp integral, log integral, sin integral, cos integral (and their hyperbolic versions), harmonic numbers, Bessel and Airy functions, elliptic integrals, ...
These are now in the apfloat library.
|
2025-04-01T04:34:50.331229
| 2023-12-04T00:39:02
|
2022803158
|
{
"authors": [
"choonyme",
"mtripg6666tdr",
"nakamuraos"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8861",
"repo": "mtripg6666tdr/nextjs-obfuscator",
"url": "https://github.com/mtripg6666tdr/nextjs-obfuscator/issues/23"
}
|
gharchive/issue
|
Build failed with webpack error when css filed imported
Thanks for the nice work. I'm on Nextjs 14.0.3 and am doing an "import '../styles/global.css'" in my /pages/_app.tsx file. This import caused the build process to failed with the messages below. I've tried even emptying the content of global.css and it's still causing error until I remove the import completely. Only nextjs-obfuscator 1.3.0 works so far but not the later versions. Any insight on this?
./src/styles/globals.css
TypeError: Class extends value undefined is not a constructor or null
at Function.getCssDependency (D:\app\node_modules\next\dist\compiled\mini-css-extract-plugin\index.js:1:8033)
at addDependencies (D:\app\node_modules\next\dist\compiled\mini-css-extract-plugin\loader.js:1:5836)
at handleExports (D:\app\node_modules\next\dist\compiled\mini-css-extract-plugin\loader.js:1:6477)
at D:\app\node_modules\next\dist\compiled\mini-css-extract-plugin\loader.js:1:7631
at D:\app\node_modules\next\dist\compiled\webpack\bundle5.js:28:814002
at D:\app\node_modules\next\dist\compiled\webpack\bundle5.js:28:131723
at symbolIterator (D:\app\node_modules\next\dist\compiled\neo-async\async.js:1:14444)
at timesSync (D:\app\node_modules\next\dist\compiled\neo-async\async.js:1:5027)
at Object.eachLimit (D:\app\node_modules\next\dist\compiled\neo-async\async.js:1:14208)
at D:\app\node_modules\next\dist\compiled\webpack\bundle5.js:28:129914
Import trace for requested module:
./src/styles/globals.css
Build failed because of webpack errors
Hello, I created a new project to look into the issue, but I couldn't replicate it.
Could you provide a reproducible repository or project, or instructions on how to reproduce? Thanks.
Apologies. My fault. Turns out I've used nextjs-obfuscator the wrong way. Was doing the following which was obsolete I guess.
const NextJSObfuscatorPlugin = require("nextjs-obfuscator");
const nextConfig = {
webpack: (config, {dev}) => {
if(!dev){
config.plugins.push(new NextJSObfuscatorPlugin({
...nextjsoptions
}, {
obfuscateFiles: {
app: true,
pages: true,
},
log: true,
}))
}
return config;
}
}
I'm glad that the issue was solved. If you have any other issues, please feel free to open a new issue.
Hi @mtripg6666tdr,
I have similar problem with this issue.
This repro repo: https://github.com/nakamuraos/next-js-obfuscator-boilerplate
Can you help me check it. Thank you.
Hi @mtripg6666tdr,
I have similar problem with this issue.
This repro repo: https://github.com/nakamuraos/next-js-obfuscator-boilerplate
Can you help me check it. Thank you.
Your repo seems to be configured in the way for nextjs-obfuscator v2.x or lower, so please re-configure it in the right way for v3.x. See the README as a reference.
Your repo seems to be configured in the way for nextjs-obfuscator v2.x or lower, so please re-configure it in the right way for v3.x. See the README as a reference.
Thank you. Just reconfig as v3 and It's working now.
Your repo seems to be configured in the way for nextjs-obfuscator v2.x or lower, so please re-configure it in the right way for v3.x. See the README as a reference.
Thank you. Just reconfig as v3 and It's working now.
|
2025-04-01T04:34:50.345234
| 2021-04-04T20:05:40
|
849951088
|
{
"authors": [
"chaosprint",
"matt-erhart",
"mtthw-meyer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8862",
"repo": "mtthw-meyer/libdaisy-rust",
"url": "https://github.com/mtthw-meyer/libdaisy-rust/issues/17"
}
|
gharchive/issue
|
Example Request: Low Latency Convolution
see http://blog.reverberate.ca/post/zero-latency-convolution/
This would be my learning project, figure other might like it too, but you could get it working much fast than me. It's useful for reverb & cabinet/mic simulation via impulse responses. I assume it's what all the impulse response (IR) guitar pedals are doing.
I want to get some DSP functions into the library (or a separate DSP crate) eventually. Once I get the core hardware enabled and working I want to do stuff like that and other things that can be ported from the cpp version of this library.
This is still beyond my current knowledge level. You can watch me experimenting with stuff over here https://github.com/mtthw-meyer/embedded-dsp
Cool! Check out https://www.youtube.com/channel/UCUR_LsXk7IYyueSnXcNextQ/playlists for more info on time-frequency analysis (using convolution) in the non-real time context.
This is still beyond my current knowledge level. You can watch me experimenting with stuff over here https://github.com/mtthw-meyer/embedded-dsp
Hi,
You can also check:
https://github.com/RustAudio/dasp
How can we create a variable to store the dsp engine?
I try to create the DSP instance during #[init], then return it to [local]:
let oscillator = SinOsc::new(440.0);
Local {
oscillator
}
But in the audio_handler:
let s = oscillator.next();
audio.push_stereo((s, s)).unwrap();
No sound is played.
|
2025-04-01T04:34:50.388951
| 2015-02-12T23:30:38
|
57536748
|
{
"authors": [
"QuinDennis",
"acosme",
"braoul",
"kentoii17",
"mrweix",
"shivrajsa"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8863",
"repo": "mudcube/MIDI.js",
"url": "https://github.com/mudcube/MIDI.js/issues/122"
}
|
gharchive/issue
|
midi js in android webview
Hello!
I am integrating midi.js in a phonegap app that is based on a webview.
For ios where the webview is based on Safari this works great! thanks!
For android however the webview is based on WebKit-based WebView (up to Android 4.4)
and Chromium (Android 4.4 and higher) .
All the sources are loaded but it just won't play a sound......
On Android 4.4 it works on Chrome Browser but not on the Chromium web view
Debugging also doesn't show any mistakes....loads the midi object correctly and display for api:
for android 4.1.2 (webkit webview) the api: "audiotag"
for android 4.4 (chromium) the api: "webaudio" (same as chrome on )
Any ideas if this can be fixed with a workaround or if this is a problem of midi.js?
Greetings! I have similar needs, midi.js in phonegap. I'll be generating local midi, so either need to pass a var (preferrably) or local file. Any ideas on how I should proceed, without having to modify midi.js?
Found it in the dev branch, midi.USE_XHR=false
I don't quite understand.... how do you propose this can be fixed? And where should one fix the midi.USE_XHR=false ?
https://github.com/mudcube/MIDI.js, select the dev branch, download, in plugin.js set USE_XHR=false.
midi.DEBUG = false;
//midi.USE_XHR = true;
midi.USE_XHR = false;
Then, the MIDIPlayer.html example works and playing notes in Chrome works. Haven't moved it to phonegap yet but believe it will work there also.
I'm not totally clear this solves your issue, but I'm good to go now! (I think)
Mmmh I tried it and still no sound on android.....
Please let me know if you have it working with Phonegap.
We also included the file acoustic_grand_piano-ogg.js within a
Oh, and yes, it works in Phonegap.
On 2/16/2015 5:09 PM, mrweix wrote:
Mmmh I tried it and still no sound on android.....
Please let me know if you have it working with Phonegap.
—
Reply to this email directly or view it on GitHub
https://github.com/mudcube/MIDI.js/issues/122#issuecomment-74586820.
--
Quin Dennis |<PHONE_NUMBER>
Thanks guys for all the input. I follow all the steps you recommend include the dev branch, include the soundfont files, everything works fine in Phonegap, but as soon as I bring it to an Android device: QUIET....I have no idea. Debugging and everything doesn't show any error. It just doesn't play a sound.... this is android 4.4 and 4.1 ...... on which android devices can you make it play a sound?
Hi, yes, we're running into issues on a NextBook 8 as well. My Samsung
G4 plays sounds but not as clearly as the browser, so we're still
researching, in the area of SoundFont files, etc. Will keep you in the
loop if/when we find something.
On 2/19/2015 1:38 PM, mrweix wrote:
Thanks guys for all the input. I follow all the steps you recommend
include the dev branch, include the soundfont files, everything works
fine in Phonegap, but as soon as I bring it to an Android device:
QUIET....I have no idea. Debugging and everything doesn't show any
error. It just doesn't play a sound.... this is android 4.4 and 4.1
...... on which android devices can you make it play a sound?
—
Reply to this email directly or view it on GitHub
https://github.com/mudcube/MIDI.js/issues/122#issuecomment-75120679.
--
Quin Dennis |<PHONE_NUMBER>
Great! Appreciate it!
I tried some more and no luck. It really comes down to a problem of midi.js talking to Chromium. On Chrome I got it working, but Chromium seems to have some slight changes the way ogg audio is handled in the webaudio api.....anybody else has an idea how to make this work?
I recently tried with phonegap with http://www.telerik.com for hybrid mobile app, I am using loop to play sequence of notes, it plays but not with perfect timing and some times some notes does not create sound. Anybody got it working perfectly with phonegap?
Hi, I also have a similar problem :
I would like to use the player of midi.js in a Crosswalk webview (which uses chrome for android) but it doesn't play any sound.
And I have no idea what is happening since the console doesn't show error..
Is there someone who figured out how to solve this problem? or does someone see what is happening?
Same problem here: Not work with ionic/cordova in android 4.1 and 4.4.
I saw in 'loader.js' line 50 "root.audioDetect", chose between 'webmidi', 'webaudio' and 'audiotag'. Always selects audiotag, but in android 5.0, it selects 'webmidi';
Finally seeing 'plugin.audiotag.js' line 41, when execute 'audio.play()' emits no sound ;
anyone?
Also same problem working on midi.js. It works fine when running it on desktop browser but when running it on android webview things were different, it plays but no sound. I'm new to android and hopefully you can help me guys.
|
2025-04-01T04:34:50.395093
| 2018-10-05T17:01:09
|
367293122
|
{
"authors": [
"muddasheep"
],
"license": "unlicense",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8864",
"repo": "muddasheep/hqtrilogy",
"url": "https://github.com/muddasheep/hqtrilogy/issues/209"
}
|
gharchive/issue
|
crash in hq2_fox
(god mode only)
AlexCorruptor
Hmmm
I miiiiight have a culprit(edited)
Are there anything that deals radiation/gASS damage?
Besides the hallway in hq2_fox (Rad.), Z A D I Z M and tetris questioning (gASS)
You know what?
Just for massive lols i will compile a small test map with all the dmg types
AlexCorruptorLast Tuesday at 8:17 PM
Alrighty, this does NOT make sence
I've compiled the vanilla-based map with ALL the possible vanilla DMGs
I've also labeled the damage slots, let's say
Got a crash on a number 16, which is Poison
And i've crashed because of the timed effect after taking the DMG
AlexCorruptorLast Tuesday at 9:37 PM
I have no clue why your NEURALGAS and RADIATION damages are not crashing for me(edited)
October 17, 2018
muddasheepLast Wednesday at 7:49 AM
did you compile it with the spirit18.fgd? there are different values
AlexCorruptorLast Wednesday at 7:49 AM
Nope
The timed effect was enough for me to raise an alarm
muddasheepLast Wednesday at 8:47 AM
are you testing it in vanilla HL?
AlexCorruptorLast Wednesday at 8:48 AM
Compiled for Vanilla HL, copied the result to HQ
As in the map itself
muddasheepLast Wednesday at 8:49 AM
HQ uses different values because of spirit of halflife, that's why you have to compile the map using the spirit18.fgd, otherwise the test results might be off
AlexCorruptorLast Wednesday at 8:49 AM
Can you check Radiation and Nerve gas in the Spirit of Half-life script?
muddasheepLast Wednesday at 8:49 AM
I'm on a work trip today, but I can recompile your test map tomorrow when I'm back home and check it out myself
yeah, but can't today
so it's probably a mix of god mode + poison/radiation right?
AlexCorruptorLast Wednesday at 8:50 AM
Here the JMF file, so i don't have to send it later
Attachment file type: unknown
Phuck.jmf
87.53 KB
So far in my testing Poison's duration effect crashed
And in HQ2 case it was Radiation and Nerve Gas DMG
|
2025-04-01T04:34:50.403400
| 2024-11-28T22:02:51
|
2703345079
|
{
"authors": [
"MaStr",
"johannesghd"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8865",
"repo": "muexxl/batcontrol",
"url": "https://github.com/muexxl/batcontrol/pull/73"
}
|
gharchive/pull-request
|
Rewrite re-evaluation trigger every 3 minutes and add fixed offset of 15 secs
Implements and closes https://github.com/muexxl/batcontrol/issues/63 and ensures the run is triggered on multiples of 3 minutes on the clock. (e.g. 03, 06, 09, 12, 15, 18, etc.)
To avoid API overload at the exact minute a delay of 15 seconds has been implemented.
The interval and the delay is currently not configurable, although maybe that would be wise to add to the config?
Thank you for your PR.
I tend to add a wait IF there is the need to call to the API and not shift the complete calculation.
And we can decide per API if there is the need for that shift. On a local EVCC instance, I don't need to wait extra seconds.
@johannesghd please review my latest commit on your branch. Yes, i did modified it that way..
Running for testing at my home
|
2025-04-01T04:34:50.438549
| 2018-08-10T21:01:29
|
349656235
|
{
"authors": [
"IssuehuntBot",
"jedwards1211",
"oliviertassinari"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8866",
"repo": "mui-org/material-ui",
"url": "https://github.com/mui-org/material-ui/issues/12472"
}
|
gharchive/issue
|
Feature request: ability to override the Input component of a TextField
[x] This is a v1.x issue.
[x] I have searched the issues of this repository and believe that this is not a duplicate.
Expected Behavior
redux-form-material-ui v5 has a quick-and-easy TextField adapter component that injects the necessary helperText, value, onChange, etc.
But to be able to use it for an address field, I need to override the Input component with something that does autocomplete (e.g. using react-places-autocomplete)
Current Behavior
Right now I'd have to basically duplicate the code in TextField, but with my own Input component instead.
Right now I'd have to basically duplicate the code in TextField, but with my own Input component instead.
@jedwards1211 What's wrong with that? The TextField is a simple wrapper component to solve 80% of the use cases. It doesn't aim at being more. Recomposing the text field is definitely a great pattern, it's allowing people to be closer to the actual DOM, something you might need here.
As far as I understand the problem, I'm very tempted to label the issue "won't fix".
Well, what do you see as the downside of allowing people to override the individual components?
It would be a small, very simple amount of code to support this so to me it seems like there's virtually no downside to it.
I can easily release a fork of TextField as it's own package but I thought it wasn't very justified to create additional fragmentation in the ecosystem for such a small change. Especially since it wouldn't automatically stay consistent with API changes here.
Though I hadn't thought that I may be able to just use Select for this, we'll see.
I'm happy to make the PR to do this btw, but I'm certainly not going to waste my time making PRs to this project anymore without any idea if they're likely to get merged
Well, what do you see as the downside of allowing people to override the individual components?
@jedwards1211 It's more about the direction we want to encourage people to go into. What's wrong about using FromControl, InputLabel, Input and FormHelperText?
Regarding the integration with redux-form, you might not need a library for it: https://github.com/mui-org/material-ui/issues/8377#issuecomment-331893521
But we migrated to react-final-form:
import React from 'react'
import PropTypes from 'prop-types'
import TextField from 'web/modules/components/TextField'
function RFTextField(props) {
const {
autoComplete,
helperText,
input: { name, ...input },
InputProps,
meta: { dirty, error, submitError, submitFailed },
...other
} = props
return (
<TextField
error={Boolean((dirty || submitFailed) && (error || submitError))}
{...input}
{...other}
id={name}
name={name}
InputProps={{
inputProps: {
autoComplete,
},
...InputProps,
}}
helperText={dirty || submitFailed ? error || submitError : helperText}
/>
)
}
RFTextField.propTypes = {
autoComplete: PropTypes.string,
helperText: PropTypes.node,
input: PropTypes.shape({
name: PropTypes.string.isRequired,
}).isRequired,
InputProps: PropTypes.object,
meta: PropTypes.shape({
dirty: PropTypes.bool.isRequired,
error: PropTypes.string,
submitError: PropTypes.string,
submitFailed: PropTypes.bool.isRequired,
}).isRequired,
}
export default RFTextField
Alright, encouraging one pattern doesn't mean we should prevent another when the overhead is low. Why not adding some InputComponent, SelectComponent, FormHelperTextComponent, FormControlComponent, and InputLabelComponent properties
The overhead is going to be about making:
the TextField source code harder to read
add some more bytes down the wire
Yeah it is true that it would add more bytes, that's a concrete downside.
Now that I read what TextField is doing, it's not as complicated as I thought so duplicating its behavior with a custom input for redux-form wouldn't be so bad. I was a bit worried at first that I would overlook something and burn time on it. Which is true to some degree, the error and helperText logic are a bit subtle.
I shouldn't have complained about whether a PR would get merged, because I always wind up with code I can use for myself even if it doesn't get merged. (I did release material-ui-popup-state btw)
Now that I read what TextField is doing, it's not as complicated as I thought so duplicating its behavior with a custom input for redux-form wouldn't be so bad.
@jedwards1211 Let me know what direction you want to take, I'm happy with both sides :).
I did release material-ui-popup-state btw
Awesome! I'm sorry I haven't answered yet on this topic. I have a backlog of items to handle that is growing. Do you want to link it in the documentation?
Yeah I'll make a PR for that soon!
@0maxxam0 funded this issue with $20. See it on IssueHunt
Ah man, I completely forgot about this 😅
I was recently trying to make some input components from react-stripe-elements look like MUI inputs, but it wasn't so simple because that lib actually displays its inputs inside iframes.
So I'm not sure how much flexibility what I was proposing here would buy us. In this case I was wishing the input underline was a public component I could use directly
@jedwards1211 Thanks for the context. I'm taking the weekend to go through all the issues and close anything that has a low ROI potential.
For this issue, we have added better integration documentation with Stripe since then.
Oh, I never even thought to look for that. Thanks for letting me know!
|
2025-04-01T04:34:50.443221
| 2019-04-04T18:34:56
|
429424105
|
{
"authors": [
"eps1lon",
"majames",
"oliviertassinari"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8867",
"repo": "mui-org/material-ui",
"url": "https://github.com/mui-org/material-ui/pull/15202"
}
|
gharchive/pull-request
|
[ToggleButton] align ToggleButton type prop with @types/react@latest
@types/react recently narrowed the type on the button element
this started causing issues with ToggleButton attempting to assign a looser type, string, to it
Fixes #15200
@majames Can you rebase the pull request on the next branch?
Can do!... but I thought to back port a fix you want to target master? https://github.com/mui-org/material-ui/blob/next/CONTRIBUTING.md#master-is-for-3x
@majames Could you rebase the pull request on the next branch?
It's already fixed in next. This is a backport specifically. If we backport we should use #15158 (without the @types/react bump.
@majames It's not an important bug fix as it only impacts the lab.
It's already fixed in next.
@eps1lon Oh nice!
@majames Thank you for taking the time!
|
2025-04-01T04:34:50.448780
| 2019-11-04T10:25:45
|
517061311
|
{
"authors": [
"eps1lon",
"oliviertassinari"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8868",
"repo": "mui-org/material-ui",
"url": "https://github.com/mui-org/material-ui/pull/18174"
}
|
gharchive/pull-request
|
[Slider] Improve UX for pointing device with limited accuracy
This is an alternative solution to #18074. It aims to find a tradeoff between:
Provide a tall enough interaction zone (e.g. to account for the size of a finger)
Provide a small enough interaction zone (e.g. to avoid unintended interactions with a too tall interaction zone or not to "waste" space on the screen)
Avoid overlapping other interactive elements (predictability)
Override simplicity
It is not an alternative as was extensively described in the original PR.
I don't understand your point. From what I understand #18174 and #18074 try to solve the same problem: from time to time when we try to interact with the slider on mobile, we have to give multiple tries, until our finger touches the rail.
Could you make an argument for each of these statements?
Do you mean, why are these dimensions are important?
Do you mean, why are these dimensions are important?
How are these addressed or apply here.
I don't understand your point.
I even made a video. I can't help you more without physically pointing the pointer for you. Nothing in my PR talks about mobile.
@eps1lon Oh, I think that I understand your point now, I was confused by the 48px touch target, for touch devices and the recording on a pointer device (the 7mm - 9mm recommendation of Google Material Design only applies to touch devices, right?) and the mention to a regression with v3.
So these two pull requests solve two different problems. #18074 is about increasing the thumb interaction zone from this zone:
to this zone:
Thanks for the patience, it took me some time to understand.
Yes I shouldn't have used this term. I just wanted to explain where the size comes from which lead to the misunderstanding that this targets mobile devices.
@eps1lon Ok awesome, sounds like a great change 👍.
|
2025-04-01T04:34:50.451121
| 2020-12-31T05:22:56
|
776796110
|
{
"authors": [
"hyeonhong",
"oliviertassinari"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8869",
"repo": "mui-org/material-ui",
"url": "https://github.com/mui-org/material-ui/pull/24205"
}
|
gharchive/pull-request
|
[DatePicker] Fix year only view, hide the current month
When you use the DatePicker and you select the year to be displayed only, the current month is shown with the year.
[x] I have followed (at least) the PR section of the contributing guide.
Closes #24195
Updated PickersCalendarHeader to show nothing in the case of displaying the year only.
@hyeonhong Thanks for raising the issue
@hyeonhong Thanks for raising the issue
|
2025-04-01T04:34:50.466905
| 2020-05-04T17:13:36
|
612032961
|
{
"authors": [
"Emiliano-Bucci",
"TheWhiteShade",
"alvamanu",
"elmeerr",
"mcancetin",
"oliviertassinari",
"schmod"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8870",
"repo": "mui/material-ui",
"url": "https://github.com/mui/material-ui/issues/20906"
}
|
gharchive/issue
|
[AutoComplete] - Load value asynchronously does not set selected value
When I have value after Autocomplete is rendered (originated from an async call), the value is not set as selected one.
[x] The issue is present in the latest release.
[x] I have searched the issues of this repository and believe that this is not a duplicate.
Current Behavior 😯
"selected value" is empty when value is loaded asynchronously
Expected Behavior 🤔
After I get data asynchronously, value should be selected
Steps to Reproduce 🕹
https://codesandbox.io/s/hardcore-wind-6pjlc
use getAsyncValue to see the issue
Context 🔦
When I have a list of options and I need to have a selected value (e.g from a list of countries, I want USA to be selected already), if my list of countries is an array of objects [{key: 1, value: "USA"}, {key: 1, value: "Japan"}, {key: 1, value: "Belgium"}] and, BEFORE rendering Autocomplete, I pass {key: 1, value: "USA"}, it works and USA is shown as selected but, if I call a api to retrieve a user and then pass user.country ( = {key: 1, value: "USA"}) (which will happen AFTER Autocomplete is rendered) , it does not work, no data is selected and gives me that controlled/uncontrolled warning.
As I just need to show what was selected before, I don't need a controlled autocomplete and I thought of using defaultValue to do such a thing...In this case I would pass only USA as defaultValue and everything would be fine. Now I can only manage to achieve this if I setup everything BEFORE rendering Autocomplete, which for the UI is weird as I won't be able to see the autocomplete until I have my data (which I also don't think it's necessary)
Basically, I need what getOptionSelected does but getOptionSelected only triggers onChange
I don't know if this is indeed an issue or if I just don't know how to use Autocomplete properly, if the later is true, I apologize in advance. I have been struggling a lot to use Autocomplete properly in my project and from time to time I end up with some kind of undesirable behavior.
Thanks for considering my request.
Regards,
Your Environment 🌎
Same as codesand
Tech
Version
Material-UI
latest
React
latest
Browser
Chrome
TypeScript
OS
Windows
@elmeerr You can find people facing the same challenge in older issues.
@oliviertassinari Thanks for taking the time but I've failed to find a use case like mine in older issues. The closest I was able to find was this https://github.com/mui-org/material-ui/issues/18748 but here his initial data is not loaded asynchronously, my example works with sync initial data, the problem is with async initial data. Reading the many issues related to inputValue / value / onInputChange / onChange / defaultValue and initial values I think it will be really helpfull to have a section in the docs with examples that covers the FAQs related to this properties...For example, is possible to see in the docs the function that handle changes assigned to the onChange from Autocomplete and also to onChange from TextField on renderInput...what is the reason behind? It's possible to find in the issues people saying "using onChange on TextField solves for me"...It can be my fault, I might be missing something..I just find confusing this part of the component
FWIW, I've run into this as well, and it appears to be a bug, or is an unintuitive behavior that (at the very least) needs to be mentioned/addressed in the demo showing how to use async requests with <Autocomplete>.
If <Autocomplete> doesn't work like any other input that you can initialize with a value, that needs to be mentioned somewhere.
@schmod hey Andrew, thanks for sharing your thought...were you able to work this around?
In my case, async values will only prepopulate when the component is set to multiple values, not to single value. Also, Autocomplete component has a state issue. Console is giving me a warning stating it's uncontrolled. It's out of my hands since it's part of the module.
Any help would be greatly appreciated.
I found a solution for this. First, you have to default the state value to null, not undefined. Then you have to determine in value prop what type of data you'll populate, like so:
const getValue = (value: AsyncACValueType) => !multiple ? value !== null ? Array.isArray(field.value) ? field.value[0] : field.value : [] : value !== null ? value : [];
then in value prop, add getValue(value)
I don't know why, but i have realized it's because the selected value must be obtained from the same object. I'm not very clear on it but i did this and it worked for me.
const [value, setValue] = useState([...options.filter((op1) => value.find(op2 => op1.id === op2.id ))]);
<Autocomplete
{...props}
value={value}
/>
@TheWhiteShade
When used in this way, if you have an onInputChange callback, it doesn't work. I think this component is not suitable for initialization with an asynchronous process and then searching.
This MUI guys have a paid library that doesn't handle this kind of stuff (in 2024); you should go to prison for engineering incompetency.
|
2025-04-01T04:34:50.473794
| 2022-10-02T08:29:54
|
1393686779
|
{
"authors": [
"TamirCode"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8871",
"repo": "mui/material-ui",
"url": "https://github.com/mui/material-ui/issues/34572"
}
|
gharchive/issue
|
[TextField] rendering input type datetime-local without value makes label go on top of placeholder text
Duplicates
[X] I have searched the existing issues
Latest version
[X] I have tested the latest version
Steps to reproduce 🕹
import { TextField } from "@mui/material"
export default function NativePickers() {
return (
<TextField
id="datetime-local"
label="Next appointment"
type="datetime-local"
// defaultValue="2017-05-24T10:30" // from demo
defaultValue=""
/>
)
}
https://codesandbox.io/s/mui-bug-input-datetime-local-mvf72n?file=/demo.js:0-263
I tried the following things:
defaultValue={undefined}
defaultValue={null}
defaultValue=""
not having defaultValue at all
the same behavior occurs with all the above.
using value instead of defaultValue results in the same behavior as well.
Current behavior 😯
When attempting not to have any initial value for datetime-local textfield, the label has a render issue and shows on top of the input. This is because it considers it as empty so it goes on top of it as usual. It needs a simple check specifically for datetime-local to always show on top because there will always be information there. This is the same for chrome, edge, and firefox.
Expected behavior 🤔
I want to have a textfield of type datetime-local without any initial value, so that it would just display like so:
(it shows like this because i selected it so the label moves up)
Context 🔦
I want to have a datetime-local textfield without an initial value, so it would show as mm/dd/yyyy --:-- --
Your environment 🌎
codesandbox
browsers: chrome, edge, firefox
I guess now I know what this does
InputLabelProps={{
shrink: true,
}}
|
2025-04-01T04:34:50.493376
| 2022-11-15T16:59:19
|
1450098810
|
{
"authors": [
"Vansh-Baghel",
"mpanjato-andri",
"siriwatknp",
"vimutti77"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8872",
"repo": "mui/material-ui",
"url": "https://github.com/mui/material-ui/issues/35159"
}
|
gharchive/issue
|
Allow using linear-gradient color in theme palette
Duplicates
[X] I have searched the existing issues
Latest version
[X] I have tested the latest version
Summary 💡
Allow using linear-gradient color in theme palette or allow to use it if it is non-default palette
Examples 🌈
I try to use linear-gradient in palette, but I got MUI: Unsupported 'linear-gradient(-39deg, #4991f8 0%, #4bc1ff 100%)' color.
https://codesandbox.io/s/determined-fermat-mikfcs?file=/src/App.tsx
Motivation 🔦
My client wants to use linear-gradient as one of background of button.
I see that the docs are lacking this info. We should mention that the palette.*.(light|main|dark) will be used to calculate channel colors (I think this should be fixed to only apply to the default palette)
@vimutti77 to make the background works, you need to use background, not bgcolor (this is how CSS works). https://codesandbox.io/s/silent-sun-onmj8f?file=/src/App.tsx
@vimutti77 Looks like you are trying to create a custom palette. Here is my suggestion:
All of the color tokens should live in colorSchemes.(light|dark).palette.* to get the benefit of the sx prop. When you write sx={{ bgcolor: 'custom.main' }}, the value will resolve correctly. This PR will fix the issue.
However, the sx prop does not work with background yet, so you can't do this sx={{ background: 'gradient.primary' }}.
I think the gradient should just be color tokens and theme each component to use the gradient selectively.
const GradientButton = ({ color = "primary" }: CustomButtonProps) => {
return (
<Button
sx={{
background: (theme) => theme.vars.palette.gradient[color].main,
"&:hover": {
background: (theme) => theme.vars.palette.gradient[color].dark
}
}}
variant="contained"
>
test
</Button>
);
};
I see that the docs are lacking this info. We should mention that the palette.*.(light|main|dark) will be used to calculate channel colors (I think this should be fixed to only apply to the default palette)
@vimutti77 to make the background works, you need to use background, not bgcolor (this is how CSS works). https://codesandbox.io/s/silent-sun-onmj8f?file=/src/App.tsx
Idk why but Experimental_CssVarsProvider and experimental_extendTheme stopped working in React JS.
I tried using it, but it didnt work.
Then I used this, and now it works fine :
`import {
createTheme
} from "@mui/material";
const theme = createTheme({
palette: {
gradient: {
main: "linear-gradient(-39deg, #4991f8 0%, #4bc1ff 100%)",
mainChannel: "0 0 0",
light: "linear-gradient(135deg, #4aaffd 0%, #4992f8 100%)",
lightChannel: "0 0 0",
dark: "linear-gradient(135deg, #4cc2ff 0%, #4aa0fa 100%)",
darkChannel: "0 0 0",
contrastText: "#fff",
contrastTextChannel: "0 0 0"
}
}
});
const Rewards = () => {
return (
<Box sx={{ width: "80%" }}>
<LinearProgress
sx={{ background: (theme) => theme.palette.gradient.main }}
variant="determinate"
value={progress}
/>
)
export default Rewards;
`
I see that the docs are lacking this info. We should mention that the palette.*.(light|main|dark) will be used to calculate channel colors (I think this should be fixed to only apply to the default palette)
@vimutti77 to make the background works, you need to use background, not bgcolor (this is how CSS works). https://codesandbox.io/s/silent-sun-onmj8f?file=/src/App.tsx
Experimental_CssVarsProvider and experimental_extendTheme didn't work for me who is using ReactJs.
I tried using it, but the site was just crashing.
Then I used createTheme and ThemeProvider, and now it works fine :
import { createTheme , Box, Typography, Button, ThemeProvider } from "@mui/material";
const theme = createTheme({
palette: {
gradient: {
main: "linear-gradient(-39deg, #4991f8 0%, #4bc1ff 100%)",
mainChannel: "0 0 0",
light: "linear-gradient(135deg, #4aaffd 0%, #4992f8 100%)",
lightChannel: "0 0 0",
dark: "linear-gradient(135deg, #4cc2ff 0%, #4aa0fa 100%)",
darkChannel: "0 0 0",
contrastText: "#fff",
contrastTextChannel: "0 0 0"
}
}
});
const Rewards = () => {
return (
<ThemeProvider theme={theme}>
<Box sx={{ width: "80%" }}>
<Button
sx={{ background: (theme) => theme.palette.gradient.main }}
variant="determinate"
value={progress}
/>
</Box>
</ThemeProvider>
)
export default Rewards;
@vimutti77 Looks like you are trying to create a custom palette. Here is my suggestion:
All of the color tokens should live in colorSchemes.(light|dark).palette.* to get the benefit of the sx prop. When you write sx={{ bgcolor: 'custom.main' }}, the value will resolve correctly. This PR will fix the issue.
However, the sx prop does not work with background yet, so you can't do this sx={{ background: 'gradient.primary' }}.
I think the gradient should just be color tokens and theme each component to use the gradient selectively.const GradientButton = ({ color = "primary" }: CustomButtonProps) => {
return (
<Button
sx={{
background: (theme) => theme.vars.palette.gradient[color].main,
"&:hover": {
background: (theme) => theme.vars.palette.gradient[color].dark
}
}}
variant="contained"
>
test
</Button>
);
};
Experimental_CssVarsProvider and experimental_extendTheme didn't work for me who is using ReactJs.
I tried using it, but the site was just crashing.
Then I used createTheme and ThemeProvider, and now it works fine :
import { createTheme , Box, Typography, Button, ThemeProvider } from "@mui/material";
const theme = createTheme({
palette: {
gradient: {
main: "linear-gradient(-39deg, #4991f8 0%, #4bc1ff 100%)",
mainChannel: "0 0 0",
light: "linear-gradient(135deg, #4aaffd 0%, #4992f8 100%)",
lightChannel: "0 0 0",
dark: "linear-gradient(135deg, #4cc2ff 0%, #4aa0fa 100%)",
darkChannel: "0 0 0",
contrastText: "#fff",
contrastTextChannel: "0 0 0"
}
}
});
const Rewards = () => {
return (
<ThemeProvider theme={theme}>
<Box sx={{ width: "80%" }}>
<Button
sx={{ background: (theme) => theme.palette.gradient.main }}
variant="determinate"
value={progress}
/>
</Box>
</ThemeProvider>
)
export default Rewards;
Experimental_CssVarsProvider and experimental_extendTheme didn't work for me who is using ReactJs.
I tried using it, but the site was just crashing.
Could you share a CodeSandbox to me so that I can take a look the error?
My bad, it is working. I was directly giving the color to the background rather than using sx. It does work.
https://codesandbox.io/s/affectionate-rui-tj5hnk?file=/src/App.js
Hello all, i want to use gradient in my icon material ? is it possible ?
For example for FavoriteIcon (import FavoriteIcon from '@mui/icons-material/Favorite'), i want to use gradient its color.
|
2025-04-01T04:34:50.497051
| 2023-03-27T14:08:20
|
1642162165
|
{
"authors": [
"canac",
"siriwatknp"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8873",
"repo": "mui/material-ui",
"url": "https://github.com/mui/material-ui/issues/36657"
}
|
gharchive/issue
|
[docs] Hidden down props should be exclusive not inclusive
Duplicates
[X] I have searched the existing issues
Related page
https://mui.com/material-ui/api/hidden/
Kind of issue
Other
Issue description
The docs for <Hidden /> say that the down props hide the element at "this size and down". However, the code shows that it only hides the element below this size (i.e. the docs say inclusive and the code says exclusive).
Context 🔦
No response
Do you want to submit a PR that fixes this?
@siriwatknp Sure! Feel free to wordsmith if desired.
|
2025-04-01T04:34:50.518330
| 2017-03-21T03:10:38
|
215616784
|
{
"authors": [
"mukulikadey",
"rekandiah"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8874",
"repo": "mukulikadey/SOEN341-Group1",
"url": "https://github.com/mukulikadey/SOEN341-Group1/issues/109"
}
|
gharchive/issue
|
Acceptance Cases
These are the acceptance cases for Sprints 1, 2, 3, 4, and 5
For Sprint 5: This is what we expect to test for sprint 5 given are expected specifications and requirement for the upcoming sprint.
Hours Spent: 4
Hours helping: 4
|
2025-04-01T04:34:50.520004
| 2022-11-04T05:06:30
|
1435559597
|
{
"authors": [
"mukundansundar"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8875",
"repo": "mukundansundar/proposals",
"url": "https://github.com/mukundansundar/proposals/pull/1"
}
|
gharchive/pull-request
|
initial commit
Signed-off-by: Mukundan Sundararajan<EMAIL_ADDRESS>
This proposal will not be part of dapr/proposals repo ... This will go into the dapr/community repo if it is a sandbox project or dapr CLI repo if we decide compose should be part of CLI from start.
The other listed items discussing dev day 1 exp will be part of the Dev Day 1 Experience project that I will create in Dapr org ...
|
2025-04-01T04:34:50.521413
| 2021-02-03T23:48:54
|
800793865
|
{
"authors": [
"IsaacEldridge"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8876",
"repo": "mulesoft/docs-studio",
"url": "https://github.com/mulesoft/docs-studio/pull/175"
}
|
gharchive/pull-request
|
DOCS-6650-generate-scaffold-of-updated-api-spec-ie
@fermujica Can you review this PR? I wasn't sure if we needed to use the word "scaffold" or not--seems to me that "generate a flow" is still accurate?
I'm going to close this pull request and resubmit--there's no info in the ticket I didn't see.
I'm going to close this pull request and start over--there was new info in the ticket I didn't see.
|
2025-04-01T04:34:50.569335
| 2019-03-18T02:23:04
|
422010264
|
{
"authors": [
"bochoven",
"ofirgalcon",
"rickheil"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8877",
"repo": "munkireport/munkireport-php",
"url": "https://github.com/munkireport/munkireport-php/issues/1237"
}
|
gharchive/issue
|
Default timeout too short
I find some modules like applications and fonts can take up to 90 seconds on some systems. I have an idle macmini server with SSD yet applications takes 80 secs.
My $0.02 as the author of a module (Sophos) that requires configuring a longer timeout: I think the default one is appropriate to prevent munki runs from getting unnecessarily long. Admins who need more time can easily add it with a config profile. Perhaps the solution is more in optimization of code and in documenting the ability to increase the timeout.
I’m with Rick on this issue
|
2025-04-01T04:34:50.661693
| 2022-05-12T04:47:21
|
1233438955
|
{
"authors": [
"marcelveldt",
"vk2him"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8878",
"repo": "music-assistant/hass-music-assistant",
"url": "https://github.com/music-assistant/hass-music-assistant/issues/145"
}
|
gharchive/issue
|
v1.0.17 - double left hand menus
Per the above screenshot, a double left hand menu displays if you follow these steps:
Click Music Assistant in left menu
Click on Album,Artists or Tracks
Click the back arrow next to Album,Artists or Tracks
Click arrow next to Music Assistant
Double menu is displayed
see this one: #73
#73 is a different problem?
Nope it is a side effect of the current workaround.
Refer #73
|
2025-04-01T04:34:50.665230
| 2022-10-30T15:44:33
|
1428866119
|
{
"authors": [
"erkr",
"gieljnssns"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8879",
"repo": "music-assistant/hass-music-assistant",
"url": "https://github.com/music-assistant/hass-music-assistant/issues/970"
}
|
gharchive/issue
|
All mass players stop playing when restarting home-assistant
What version of Music Assistant has the issue?
all
The problem
When I restart HA all my mass players stop playing music.
Isn't it possible to let the music play?
How to reproduce
Restart HA
Relevant log output
x
Additional information
x
What version of Home Assistant Core are your running
all
What type of installation are you running?
Home Assistant OS
On what type of hardware are you running?
Generic x86-64 (e.g. Intel NUC)
That is unavoidable. Integrations like MA are started and stopped by home assistant. So this is by design and not a bug. I will close the issue.
Best Eric
I thought so.
Isn't possible to run MA as an addon?
I can't judge if that is technically feasible, but you can create a feature request for that under discussions on GH!
|
2025-04-01T04:34:50.690556
| 2023-09-13T09:37:58
|
1894140824
|
{
"authors": [
"muzairkhattak",
"yxgnahz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8880",
"repo": "muzairkhattak/PromptSRC",
"url": "https://github.com/muzairkhattak/PromptSRC/issues/3"
}
|
gharchive/issue
|
About the accuracy on base-to-new generalization setting
Hi,
Thanks very much for your interesting and inspiring work and well-organized codebase! I am trying to reproduce your results in the paper, and here are the results I obtained for the base-to-new generalization setting:
| Imagenet | Caltech101 | Food101 | Dtd | Ucf101 | Oxford_Flowers | Oxford_Pets | FGVC | Standford_Cars | Sun397 | Eurosat | AVG
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
Base | 77.60 | 98.07 | 90.67 | 83.53 | 86.93 | 98.00 | 95.17 | 42.87 | 78.43 | 82.53 | 92.83 | 84.24
Std | 0.14 | 0.12 | 0.09 | 0.92 | 0.05 | 0.14 | 0.21 | 0.88 | 0.25 | 0.12 | 0.97 |
New | 70.30 | 93.87 | 91.57 | 59.63 | 78.93 | 77.10 | 97.33 | 36.60 | 75.30 | 78.80 | 72.93 | 75.67
Std | 0.08 | 0.05 | 0.09 | 3.52 | 2.07 | 0.14 | 0.49 | 0.42 | 0.59 | 0.41 | 3.92 |
H-mean | 73.77 | 95.92 | 91.12 | 69.58 | 82.74 | 86.30 | 96.24 | 39.49 | 76.83 | 80.62 | 81.69 | 79.73
Basically, I find that the accuracies on base classes are consistent with the results in the paper and their variances are quite small. However, when it comes to the accuracies on the new classes, the accuracies are lower than that reported in the paper (on average 75.67 vs. 76.10). After carefully checking the results, I find that the gap mainly comes from two datasets, DTD and EuraSAT, where the variances are much larger than other classes. Therefore, I wonder whether this phenomenon is normal or not? Did you observe larger variances for the two classes during testing on new classes?
Thanks in advance!
Hi @yxgnahz,
Thank you for showing interest in PromptSRC!
Regarding your query, we also observed a bit larger variances for these datasets, but the obtained results are same as mentioned in our main paper.
So I think, the reproduced performance should roughly match the main paper results. There can be slight variance in the results when reproducing results on different machines, but the shared results gap on average novel class performance is significant (75.67 vs 76.10). From your table, even for base classes, some datasets have different results, such as 78.43 % on Stanford cars vs reported 78.27 %.
In order to resolve the issue, can you kindly verify the following?
You can make sure that you are using the same versions of PyTorch and other packages as mentioned in our installation read-me files shown here. We have noticed in our experiments that using different PyTorch and CUDA version leads to fluctuating results. So this might help in obtaining the exact results.
You may kindly verify if you are using the provided scripts for running your experiments and not changing any hyper-parameters in the config files. For training in base-to-novel setting, we use this script which by default uses the vit_b16_c2_ep20_batch4_4+4ctx config file.
Optionally, can you also re-run your experiments using the commit version of 1af6e40b01e0c0c24fd8b6854be835209a0cde74 for the Dassl.pytorch library. I have noticed that we used slightly older version of the Dassl.pytorch. It should not be a problem, but you may just check this as well in-case.
I hope that would be helpful. Kindly let us know if this resolves your issue!
Thanks for your prompt reply! I followed your installation guide to set up the environments (torch 1.9.0+cu111) and used exactly the same script you provided for the reproduction. All the experiments are conducted on an Nvidia 3090 GPU. I also tried with the provided Dassl version, and I found the results unchanged. Therefore, could you please show me the exact variances you observed for these datasets?
Hi @yxgnahz,
Thank you for verifying the installation.
Actually I do not have access to the machine where all the checkpoints and results log files are stored. I will try to get the access, and share the variance details here as soon as possible.
Lastly, can you clarify if you are training your models from the scratch to reproduce the numbers or are you using the provided pre-trained weights to reproduce the results?
Thank you!
Hi @muzairkhattak ,
The results above are trained from scratch using the scripts you provided in the document. By the way, I just verified the official released checkpoint for base-to-new setting, and here are results (accuracy on new classes) I obtained by running the reproduce inference script:
DTD 63.08+/-3.15, Eurosat: 73.88+/-4.80
The averaged accuracies are very close to the results in the paper (less than 0.1%) and I think the minor gap may come from the difference between our hardware platform. Moreover, the variances from the released checkpoint are just consistent with what I have got in the models trained from scratch, which indicates that these two datasets indeed have large variances on new classes.
Hi @yxgnahz,
Thank you for sharing this update. Yes, I think these two datasets have higher variance as discussed.
If manageable for you, can you try to reproduce the results from scratch using another machine if that is available? This will be a double check.
Thank you once again!
Hi @yxgnahz,
I have just verified reproducing the main paper results by training models from scratch using this repository. I am able to obtain roughly same results as mentioned in our paper.
For example, the reproduced results for EuroSAT are 92.90% for base classes and 73.83% for novel classes. Similarly on DTD, the base class results are 83.53% and novel class results are 62.97%.
To facilitate the process of reproducing the results on these datasets, I have provided the training and testing log files through this link. You can verify and compare them with your log files.
I hope that would be helpful!
Thank you and kind regards!
Hi @muzairkhattak ,
Sorry I only have 3090 GPUs on my server. If I have chance, I will test this on another machine. I check the logs and I find all the settings are just the same. I believe the large variance leads to the performance gap in my reproduction and the value of the gap is normal considering the variance of the two datasets.
Thanks for your discussion!
|
2025-04-01T04:34:50.703680
| 2019-02-28T03:10:57
|
415426741
|
{
"authors": [
"gidfiddle",
"mviereck"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8881",
"repo": "mviereck/x11docker",
"url": "https://github.com/mviereck/x11docker/issues/133"
}
|
gharchive/issue
|
sometimes need delay before $Pid1pid is valid
x11docker --version: 5.4.4
On a rather slow Ubuntu 18.04 machine, the command
x11docker --nxagent ubuntu:bionic
always fails with a "Container startup seems to have failed." error, despite (usually) working well on a faster Fedora 29 machine. I traced the problem to the following statement in the x11docker script:
Pid1pid=$($Dockerexe inspect --format '{{.State.Pid}}' $Containername 2>>$Containerlogfile | rmcr)
sets Pid1pid to zero. I succeeded in working around this problem by putting this line in a loop like another one in the script:
for ((Count=1 ; Count<=20 ; Count++)); do
[line above]
if test x$Pid1pid != x && test $Pid1pid -ne 0; then break; fi
sleep 0.2
done
(The first test is needed because Fedora sometimes sets Pid1pid to a blank.) One pass through the loop usually suffices to get a valid Pid1pid.
I cannot explain why a delay is necessary here even after the container has been determined to be "ready" because the test
$Dockerexe exec $Containername sh -c :
has succeeded.
Could you please look into this issue and make an appropriate correction?
Thanks for a great piece of software.
Thank you for reporting!
I cannot explain why a delay is necessary here even after the container has been determined to be "ready" because the test
$Dockerexe exec $Containername sh -c :
has succeeded.
That is odd, indeed. I have made a commit with a repeated check like in your proposal, it should work now. Please try out.
Thanks for a great piece of software.
:-)
|
2025-04-01T04:34:50.706737
| 2018-08-31T11:11:18
|
355938915
|
{
"authors": [
"ratijas",
"yohad"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8882",
"repo": "mvoidex/hsdev",
"url": "https://github.com/mvoidex/hsdev/issues/91"
}
|
gharchive/issue
|
Cabal failes to install under clean haskell platform
Hi,
On a windows 10 I just installed haskell platform, them tried installing hsdev with the following result:
C:\Users\yotam>cabal install hsdev
Resolving dependencies...
cabal: Could not resolve dependencies:
[__0] trying: hsdev-<IP_ADDRESS> (user goal)
[__1] trying: hlint-2.1.10 (dependency of hsdev)
[__2] trying: haskell-src-exts-1.20.2 (dependency of hsdev)
[__3] next goal: ghc-syb-utils (dependency of hsdev)
[__3] rejecting: ghc-syb-utils-<IP_ADDRESS> (conflict: hsdev =>
ghc-syb-utils>=0.2.3 && <0.3)
[__3] trying: ghc-syb-utils-<IP_ADDRESS>
[__4] next goal: ghc (dependency of hsdev)
[__4] rejecting: ghc-8.4.3/installed-8.4..., ghc-8.4.3, ghc-8.4.1 (conflict:
ghc-syb-utils => ghc>=7.0 && <8.4)
[__4] rejecting: ghc-8.2.2, ghc-8.2.1 (conflict: hsdev => ghc==8.4.*)
After searching the rest of the dependency tree exhaustively, these were the
goals I've had most trouble fulfilling: hsdev, haskell-src-exts, ghc-mod,
hlint, base, ghc, ghc-syb-utils
Which I found odd, as it seems to me that I need to versions of ghc.
It just can not be build with the latest GHC (8.4.3 at the moment of writing), so use stack to get older lts snapshot and build hsdev inside that environment.
lts-11.22 / ghc-8.2.2 works well.
> stack --resolver=lts-11.22 install hsdev-<IP_ADDRESS> haddock-api-2.17.4 hdocs-0.5.2
Note that some other packages must be specified with fixed version. This is because stack refuses to resolve fuzzy dependency ranges by itself. There is some theory around predictable builds around it, but just skip it for now. Generally, when you see in Haskell package requirements version range like >= x.y.z it's probably a good idea to stick with x.y.z, as it was tested for sure and most likely won't cause troubles.
|
2025-04-01T04:34:50.709856
| 2021-03-31T19:29:39
|
847192072
|
{
"authors": [
"adsharma",
"mwhittaker"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8883",
"repo": "mwhittaker/quoracle",
"url": "https://github.com/mwhittaker/quoracle/pull/5"
}
|
gharchive/pull-request
|
Make Node a dataclass
Dataclasses make the code more compact and generate many convenience
functions.
Ah, I've been targeting Python 3.6, and I think dataclasses are a 3.7+ thing?
Backports are available: https://pypi.org/project/dataclasses/
In general, dataclasses eliminate so much boilerplate code that I use them wherever I can.
Yes - it saves only 10 lines in this diff. Not an earth shattering difference :)
I've addressed the bugs you found in the code and simplified the typing and default values.
I'll leave it here just in case you end up using dataclasses elsewhere and then want to convert this one for consistency.
|
2025-04-01T04:34:50.714313
| 2012-02-22T19:36:48
|
3342123
|
{
"authors": [
"candrews",
"mwrock"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8884",
"repo": "mwrock/RequestReduce",
"url": "https://github.com/mwrock/RequestReduce/issues/147"
}
|
gharchive/issue
|
Dashboard does not work on IIS 6
When a Request Reduce site is deployed to an IIS 6 server, the dashboard at /RequestReduceContent/Dashboard doesn't work - the URL returns a 404. The same site deployed to IIS 7 works fine.
Is there some special configuration that can be done so that the dashboard works on IIS 6?
Sorry for not getting back to you on this yet. I need to setup an IIS 6 environment to get a better idea of what is going on and then I'll let you know. I would definitely like to support IIS6.
|
2025-04-01T04:34:50.722827
| 2016-08-26T04:28:22
|
173367879
|
{
"authors": [
"mxcl",
"rlam3"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8885",
"repo": "mxcl/PromiseKit",
"url": "https://github.com/mxcl/PromiseKit/issues/518"
}
|
gharchive/issue
|
How to use PromiseKit.wrap?
I'm trying to use PromiseKit.wrap but I'm not able to get it to work. Is there a specific import that I'm not using that is causing this to happen?
//Module 'PromiseKit' has no member named 'wrap'
PromiseKit.wrap{
self.alamofireRequest1()
}.then { result in
alamofireRequest2(result)
}
This documentation is for the yet unreleased PromiseKit 4. This is mentioned at the top of the documentation.
Sorry for this, PMK4 is out very soon, and the documentation needed an update, so I combined the two.
The PMK3 equivalent is Promise { completionHandler in }.
@mxcl So in my case. should my self.alamofireRequest1 and self.alamofireRequest2 need to be promise functions too? Or is this method of using Promise to wrap two seperate async requests not the accepted way to use PromiseKit?
My problem involves Moya and PromiseKit. I'd like Moya to do the routing and the promisekit to handle the execution sequence. But there are not many examples out there using this approach.
If alamofireRequest1 returns Promise then you are using wrap wrongly. You should instead use firstly:
firstly {
alamofireRequest1()
}.then { result in
//…
}
// or you can just then of any promise:
alamofireRequest1().then { result in
//…
}
If they take a completion-handler then you should use wrap:
PromiseKit.wrap(alamofireRequest1).then { result in
//…
}
If you are using PromiseKit 3 however there is no wrap. You didn't say which PromiseKit you are using.
@mxcl both alamofireRequest1 and 2 are both returning Request not Promise Is there a way for me to wrap two requests and do them synchronously? And I'm using PromiseKit 3
Thanks!
If they return Request, then no, you need a completion handler variant to use with this initializer:
https://github.com/mxcl/PromiseKit/blob/master/Sources/Promise.swift#L92-L104
Closing due to lack of response. Let me know if you need further assistance.
@mxcl I have a function which I would like to return a request from. But it seems like the Promises are in different closures. Is there a way to get the genericRequest to be returned back after chaining? Thanks.
func alamofireRequest(url:String?,completionHandler:(response:Response<AnyObject,NSError>, error: NSError?) ->()) -> Request{
var genericRequest: Request?
firstly{
pp()
}
.then{
_ -> Promise<Request> in
print("BEGIN REQUST!!")
// Obtain new token
let user = User() // User is still grabbing the old token?
let URL = APIURLFactory().makeURLFor(url!)
// Debugging
print("alamofireRequest: \(URL)")
let mutableURLRequest = NSMutableURLRequest(URL: URL)
mutableURLRequest.HTTPMethod = "GET"
mutableURLRequest.setValue("application/json", forHTTPHeaderField: "Content-Type")
print("USER JWT:::::::;")
print(user.jwt_token)
mutableURLRequest.setValue("JWT \(user.jwt_token)", forHTTPHeaderField: "Authorization")
genericRequest = Alamofire.request(mutableURLRequest)
.validate()
.responseJSON{
response in
completionHandler(response: response, error: response.result.error)
}
return Promise{
fulfill, reject in
fulfill(genericRequest!)
}
}.then{
request in
return request <<<< This is not how we do it right?
}
return genericRequest! <<<< need to return here but nil keeps popping up
}
func pp() -> Promise<String>{
return Promise{ fulfill, reject in
if self.isCurrentAuthTokenIsExpired(){
print("Token is expired")
//FIXME: locksmith not updated prior to running
self.updateAuthTokenInLocksmithWithNewToken()// ERROR HERE
}
fulfill("UPDATED TOKEN")
}
}
Thanks!
What you are doing doesn't really make sense, as far as I can tell you want this:
func alamofireRequest(url: String) -> Promise<Response> {
return pp().then { _ -> Promise<Response> in
let user = User()
let URL = APIURLFactory().makeURLFor(url!)
let mutableURLRequest = NSMutableURLRequest(URL: URL)
mutableURLRequest.HTTPMethod = "GET"
mutableURLRequest.setValue("application/json", forHTTPHeaderField: "Content-Type")
mutableURLRequest.setValue("JWT \(user.jwt_token)", forHTTPHeaderField: "Authorization")
return PromiseKit.wrap(Alamofire.request(mutableURLRequest).validate().responseJSON)
}
}
|
2025-04-01T04:34:50.737677
| 2016-11-04T12:33:23
|
187322246
|
{
"authors": [
"coveralls",
"mxstbr",
"samit4me"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8886",
"repo": "mxstbr/react-boilerplate",
"url": "https://github.com/mxstbr/react-boilerplate/pull/1192"
}
|
gharchive/pull-request
|
Faq staying updated with boilerplate
Keeping your project up-to-date with react-boilerplate can be difficult and there is no recommended approach at the moment. There is a long term goal to change this (see #1180), so this PR is more or less to document the official response to the common question "how to keep my project up-to-date or in sync with the boilerplate". Would love to hear feedback!
Also added a TOC as I feel this makes it much easier to navigate.
Coverage remained the same at 98.605% when pulling ab3385e03580787eff0a443f035353908b388f3a on samit4me:faq-staying-updated-with-boilerplate into 3aafc0c593d2f51bda03538d7131dec7168b8475 on mxstbr:dev.
Coverage remained the same at 98.605% when pulling 05d28b46691418bdada7c597725f361f4304af9c on samit4me:faq-staying-updated-with-boilerplate into dd8f9a68382b640ab8c5d8caa9683f4afe6cad8c on mxstbr:dev.
LGTM, thanks!
|
2025-04-01T04:34:50.740262
| 2024-04-03T08:21:01
|
2222254200
|
{
"authors": [
"godind",
"jimmyendt"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8887",
"repo": "mxtommy/Kip",
"url": "https://github.com/mxtommy/Kip/issues/363"
}
|
gharchive/issue
|
feat: hide widget label
Hello,
I'm fairly new to using Signal K and KIP, but really loving it and all the hard work! But i have to admit I have no experience coding or anything.
I was wondering, would it be possible to make the Widget label removable? I know you can leave it blank, but I have the feeling that it does take up some space in the awesome boxes you can put the widgets in? is this correct?
it would be amazing to maximize space maybe?
i'm also on discord and maybe have some more ideas/questions if that is okay.
thanks in advance
Hi. Thanks for the feedback. It's possible but would require reworking all the Widgets. I am not convinced this use case would apply to most. It's would make pages hard to understand without any label. All marine package have labels for that purpose.
Thanks for the fast reply. Sounds like a lot of work and it's merely a question.
I was thinking about the wind-widget, or radial gauge widgets that would speak for itself and therefore maximizing the space.
can i ask you another question on discord maybe?
Absolutely!
|
2025-04-01T04:34:50.746495
| 2021-06-28T21:42:08
|
932017220
|
{
"authors": [
"hmreumann",
"myckhel"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8888",
"repo": "myckhel/laravel-chat-system",
"url": "https://github.com/myckhel/laravel-chat-system/pull/20"
}
|
gharchive/pull-request
|
Commented loadMigrationsFrom
Hallo @myckhel I tested it in a brand new laravel project and it worked fine.
When running php artisan migrate without publishing it only creates the base tables (users, password resets, and failed jobs).
After running php artisan vendor:publish --provider="Myckhel\ChatSystem\ChatSystemServiceProvider" --tag='migrations' , it copies all the migrations in the folder with the same name.
Running again the same command it doesn't duplicate the migrations.
Running php artisan migrate again, it creates the tables corresponding to the package.
I hope this works, i think it was easy, I've just commented the $this->loadMigrationsFrom(__DIR__.'/../database/migrations'); line in the boot function from the ChatSystemServiceProvider.php. It think it fixes #10
By the way i've learned a lot. Thanks for letting me contribute.
Hernán.
Thanks for your time.
Job well done.
|
2025-04-01T04:34:50.790188
| 2022-03-13T09:48:34
|
1167532190
|
{
"authors": [
"AnthonyAspen",
"mymmrac"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8889",
"repo": "mymmrac/telego",
"url": "https://github.com/mymmrac/telego/issues/20"
}
|
gharchive/issue
|
Typo in telego/pradicates
💬 Telego version
v0.10.2
👾 Issue description
AnyMassage instead of AnyMessage.
https://github.com/mymmrac/telego/blob/main/telegohandler/pradicates.go#L29
It's a minor typo, not sure worth opening a bug report
later when I feel myself more confident about how this library works, I'll create pull requests directly
⚡️ Expected behavior
AnyMessage
🧐 Code example
// AnyMassage is true if message isn't nil
func AnyMassage() Predicate {
return func(update telego.Update) bool {
return update.Message != nil
}
}
It's okay to create issues, even for small typos, it will help improve the library. And creating PR is also welcomed 🙂, will really appreciate it.
|
2025-04-01T04:34:50.830763
| 2016-10-09T09:38:25
|
181877449
|
{
"authors": [
"rjonaitis",
"vsergeev"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8890",
"repo": "myriadrf/LimeSuite",
"url": "https://github.com/myriadrf/LimeSuite/issues/71"
}
|
gharchive/issue
|
Minimum TX bandwidth limited to 5 MHz
The LMS7002 datasheet suggests the TXLPFL path can be tuned down to 2 MHz, but the code sets a lower limit of 5 MHz. Is there is an implementation detail behind the discrepancy?
It would be nice to use the lower TX filter bandwidths for more narrowband signals, if they're available.
The LMS7002 datasheet specifies TXLPFL frequencies in IF, while the code uses RF
Ah, so it looks like it's operating at a digital IF of at least 2.5 MHz?
|
2025-04-01T04:34:50.873988
| 2022-10-12T20:39:49
|
1406784501
|
{
"authors": [
"SteffenHeu",
"bsedio"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8891",
"repo": "mzmine/mzmine_documentation",
"url": "https://github.com/mzmine/mzmine_documentation/issues/72"
}
|
gharchive/issue
|
Windows for "Aducts" and "Modifications" do not appear for ion identity networking
Hello MZmine,
I am running MZmine 3.2.8 on a Mac OS 12.6 Monterey. When I open "ion identity networking" and "ion identity library:setup" the windows under "Aducts" and "Modifications" do not render, and hence no aducts can be selected and the ion identity networking module does not execute. A screenshot is attached.
Is this an error in MZmine3? Or am I doing something wrong that is preventing the window for selecting aducts to render?
Thank you very much,
Brian
there are no adducts/modifications present by default. click "reset positive/negative" and "reset" do use the default list.
if this issue persists, please use the MZmine repository to report it: https://github.com/mzmine/mzmine3/issues
this is just for the documentation.
|
2025-04-01T04:34:50.882429
| 2018-09-08T21:29:31
|
358329348
|
{
"authors": [
"mzolfaghari",
"sophia-wright-blue"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8892",
"repo": "mzolfaghari/ECO-efficient-video-understanding",
"url": "https://github.com/mzolfaghari/ECO-efficient-video-understanding/issues/17"
}
|
gharchive/issue
|
frames for Kinetics dataset
Hello,
In the script 'create_list_kinetics.m', you have the following path:
path_DB_rgb='/datasets/kinetics/train/db_frames//'
I'm assuming this folder contains the frames for the kinetics videos. Are the frames for the videos available online somewhere, or is there a script available to split the videos into the frames?
I tried running ''main.py' of your pytorch implementation and got the following error:
when running
---> 20 for i, (input, target) in enumerate(train_loader):
......
FileNotFoundError: [Errno 2] No such file or directory: '/kinetics/pumping_gas/ib5PzcBeYIc_000004_000014/0004.jpg'
Thanks,
Hi @sophia-wright-blue
Yes, you need to extract frames. For some datasets you can find the frames but for Kinetics I don't think you can find frames.
To extract frames you can use this script.
Please let me know if you still had problem with extracting frames.
Thank you so much for your response, and thank you for your patience with the basic questions.
I'm trying to use your PyTorch code from scratch. To clarify, here are the steps:
1). Download Kinetics dataset. To do this, I've found the following link:
https://github.com/activitynet/ActivityNet/tree/master/Crawler/Kinetics
Do you have a better script for doing this step?
2). Once I have downloaded all of the videos in a folder, I need to extract the frames from the videos. Here you have the script:
https://github.com/mzolfaghari/chained-multistream-networks/blob/master/scripts/extract_frames_frmRate.sh
I'd greatly appreciate your help in the exact command and software required to run the script to extract the frames for all the videos. I have the videos downloaded in a folder named kinetics/videos. I would like to extract the frames to the folder kinetics/frames.
3). Once the frames have been extracted, do we need to create the training and testing lists, by running the script:
https://github.com/mzolfaghari/ECO-efficient-video-understanding/blob/master/scripts/create_lists/create_list_kinetics.m
or is there an equivalent python script?
4). Once we have the frames extracted and lists created, we can run the script:
https://github.com/mzolfaghari/ECO-pytorch/blob/master/scripts/run_ECOLite_kinetics.sh
This would give us the trained model.
5). Use the trained model for inference on some test videos.
Once again, greatly appreciate your help and guidance with this.
@sophia-wright-blue
1- We used the same scripts!
2- Provided necessary scripts in this folder. Please check the code.
3- We don't have the python script for this.
4- Yes, after having the frames and list you can run the code and get the final model.
5- Yes.
thank you so much!
|
2025-04-01T04:34:50.891093
| 2022-12-13T15:11:26
|
1494459434
|
{
"authors": [
"Prince-Hamza",
"awais-amjed",
"mzusin",
"toolcool-org"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8893",
"repo": "mzusin/toolcool-range-slider",
"url": "https://github.com/mzusin/toolcool-range-slider/issues/4"
}
|
gharchive/issue
|
require is not defined when bundling code with node
it works fine with when used with react only. when i build with yarn build for nodejs server (because i have apis) . server shows a blank page .console says require is not defined.
The slider can't work in node.js only as it requires DOM. What is your use case?
Closing because there was no further response.
Getting this Error in Production.
I am using ReactJS v18
Slider works fine in dev mode but when I build for production and deploy it on Firebase I get this error.
Uncaught ReferenceError: require is not defined
I have tried to isolate the issue to this import
import 'toolcool-range-slider';
Removing this import, the code works fine in production but obviously the Slider is not displayed :D
I am using create-react-app
How to you import the slider to your component? Please share some code sample.
Following the guide here exactly: https://toolcool-range-slider.mzsoft.org/pages/react-typescript.html
Here is my code related to the slider:
import 'toolcool-range-slider/dist/plugins/tcrs-marks.min.js';
import 'toolcool-range-slider';
<tc-range-slider
id={`slider-${id}`}
data={data}
value={value}
ref={ sliderRef }
slider-bg={color1}
slider-bg-hover={color1}
slider-bg-fill={color1}
slider-height="2px"
slider-width={min768 ? "380px" : null}
slider-radius="2px"
pointer-width="18px"
pointer-height="18px"
pointer-bg="white"
pointer-bg-hover="white"
pointer-bg-focus="white"
pointer-border="0 solid transparent"
pointer-border-hover="0 solid transparent"
pointer-border-focus="0 solid transparent"
pointer-shadow="2px 7px 8px 0 rgba(0, 0, 0, 0.15)"
pointer-shadow-hover="2px 7px 8px 0 rgba(0, 0, 0, 0.15)"
pointer-shadow-focus="2px 7px 8px 0 rgba(0, 0, 0, 0.15)"
mousewheel-disabled="true"
marks="true"
min={minShown}
max={maxShown}
step={step}
marks-count={0}
marks-values-count={marksValueCount ?? (((maxShown - minShown) / step) + 1)}
marks-color="transparent"
marks-values-color="#A6A6A6"
/>
Produces Sliders like this in dev mode:
OK, I'll try to create a react app with the slider and reproduce this error.
ok Thanks
So I removed the package from package.json and imported it using the script from cdn in my index.html file like this
This works in Production.
So I think the issue is in the bundling process somewhere.
Yes, you're right, It looks like some conflict with babel package in create react app. I'm still investigating.
It's hard to believe, but it seems that create react app is no longer being maintained and kind of deprecated. They all recommend Vite as a replacement.
https://github.com/facebook/create-react-app/issues/13140
https://github.com/facebook/create-react-app/issues/12974
Here is a little more information
https://github.com/reactjs/react.dev/pull/5487#issuecomment-1409720741
Apparently, it is unlikely that someone will soon answer the ticket I opened.
|
2025-04-01T04:34:50.894043
| 2022-02-03T13:36:09
|
1123095735
|
{
"authors": [
"Aw3som3Guy",
"NdvJff"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8894",
"repo": "n00mkrad/cupscale",
"url": "https://github.com/n00mkrad/cupscale/issues/109"
}
|
gharchive/issue
|
An error occured during upscaling
Hi
I have all my attachements, 32gb of ram and a 3060rtx gpu
everytme i launch an upscale after 0.2sec it says " An error occured during upscaling"
do you know what could be the problem ?
cheers !
Have you tried changing the AI network? (Its the setting in the top left of the screen, a little ways above model selection.) I've found the middle choice (ESRGAN (ncnn)) has the best compatibility, and RealESRGAN has a massive speedup but you have to use a 4x model, no bigger no smaller.
|
2025-04-01T04:34:50.895773
| 2021-01-09T09:34:10
|
782571125
|
{
"authors": [
"n07734"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8895",
"repo": "n07734/community-health",
"url": "https://github.com/n07734/community-health/issues/4"
}
|
gharchive/issue
|
Top up report existing data
If you have older data already saved it would be good to be able to add newer data to it vs running it all again to also have new data.
Options:
Forwards pagination: would be more logical but tricky to implement as you would need the first request's pagination info and update pagination to start from and change the sorts in the query, then re save "origin" pagination info
Do existing style requests and get data until it is no longer getting fresh data, finding if data is new or not and de-duping could be challenging and fiddly.
Also remove need for preFetchedRepo key
done
|
2025-04-01T04:34:50.912139
| 2016-03-04T22:16:36
|
138601999
|
{
"authors": [
"Freundschaft",
"chaosmail",
"lorem--ipsum",
"samwzm"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8896",
"repo": "n3-charts/line-chart",
"url": "https://github.com/n3-charts/line-chart/issues/452"
}
|
gharchive/issue
|
Vertical line does not show all the dots
Hi, there,
I have the data sharing the same time, which would be a vertical bar. however, it seems only the bottom data point displays the dot, all the other data points do not display dots. I paste the sample html page below. Could anyone take a look?
By the way, good news is that if I put my mouse over that vertical bar, then all the dots will display.
thanks,
Sam
below is the sample html to show the issue:
<script src="./d3.min.js"></script>
<!-- Reference n3-line-chart -->
<script src="./LineChart.js"></script>
<link rel="stylesheet" href="./LineChart.css">
<script type="text/javascript">
angular.module('app', ['n3-line-chart'])
</script>
<script type="text/javascript">
angular
.module('app', ['n3-line-chart'])
.controller('MainCtrl', function($scope) {
$scope.data = {
dataset0: [
{x: 0, val_0: 0, val_1: 0, val_2: 0, val_3: 0},
{x: 0, val_0: 0, val_1: 0, val_2: 2, val_3: 0},
{x: 1, val_0: 0.993, val_1: 3.894, val_2: 6.47, val_3: 14.347},
{x: 1, val_0: 0.993, val_1: 3.894, val_2: 8.47, val_3: 14.347},
{x: 1, val_0: 0.993, val_1: 3.894, val_2: 10.47, val_3: 14.347},
{x: 2, val_0: 1.947, val_1: 7.174, val_2: 13.981, val_3: 19.991},
{x: 3, val_0: 2.823, val_1: 9.32, val_2: 14.608, val_3: 13.509},
{x: 4, val_0: 3.587, val_1: 9.996, val_2: 10.132, val_3: -1.167},
{x: 5, val_0: 4.207, val_1: 9.093, val_2: 2.117, val_3: -15.136},
{x: 6, val_0: 4.66, val_1: 6.755, val_2: -6.638, val_3: -19.923},
{x: 7, val_0: 4.927, val_1: 3.35, val_2: -13.074, val_3: -12.625}
]
};
$scope.options = {
series: [
{
axis: "y",
dataset: "dataset0",
key: "val_2",
label: "An area series",
color: "#1f77b4",
type: ['line', 'dot'],
id: 'mySeries0'
}
],
axes: {x: {key: "x"}}
};
});
</script>
<div ng-app="">
<p>Name : <input type="text" ng-model="name"></p>
<h1>Hello {{name}}</h1>
Yeah, currently we use the x value of the data point as reference - which does not allow multiple x values per series. I have to look into this in more detail if theoretically it could be changed. Thanks for reporting this!
+1 I'd also volounteer to check if it would be possible to change that
Okay so this is not gonna happen. Abscissas need to be distinct in order to keep things meaningful in the code, that's one of the few requirements. We can't implement a charting library that deals with every possible data structure. What's the use case for this data you posted @samwzm anyway ? Because to me it makes very little sense to have several different value for the same series on the same abscissa. I'd rather put that into different series.
I think what @samwzm wanted to do is build a scatter plot https://upload.wikimedia.org/wikipedia/commons/thumb/0/0f/Oldfaithful3.png/240px-Oldfaithful3.png
Ha, I hadn't thought of this. This can be useful indeed !
|
2025-04-01T04:34:50.948776
| 2015-06-23T07:03:21
|
90308076
|
{
"authors": [
"nDmitry",
"rarila"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8897",
"repo": "nDmitry/grunt-postcss",
"url": "https://github.com/nDmitry/grunt-postcss/pull/33"
}
|
gharchive/pull-request
|
Allow directory path and filename for annotation
(Split of #30 into two different PRs)
This adds the possibility to set the annotation either to a sourcemap filename (as postcss's default behavior) or a directory (path ends with ’’/’’ for use with multiple source files).
(See #30 for discussion on it)
Besides option complex logic I can see an issue in the implementation: If you specify a directory without a trailing slash you'll get something unexpected, so it's better to match the .map extension (or any extension), though it limits users on filename choice. We could also check if it's a directory or a file using fs module, but the given path may not exist at the time.
I understand that PostCSS accepts paths to filenames and directories, but it also accepts relative paths and it is only a library API that we don't have to expose as it is. So honestly I don't wanna merge this one.
|
2025-04-01T04:34:50.950549
| 2018-01-16T21:22:05
|
289060461
|
{
"authors": [
"nWidart",
"snipiba"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8898",
"repo": "nWidart/laravel-modules",
"url": "https://github.com/nWidart/laravel-modules/pull/463"
}
|
gharchive/pull-request
|
Update:
config.php added 'START_FILE' for replacement generator
stubs/json.stub - added $START_FILE$ to reflect config.php
Generators/ModuleGenerator.php - added function to return value from config/modules.php -> stubs.files.start ...
Hi,
I'm working the 3.0 release on master branch.
Re-reading your pull-request, I don't see where you replace the $START_FILE$ key with start.php?
|
2025-04-01T04:34:51.019106
| 2023-11-08T09:18:08
|
1983122426
|
{
"authors": [
"CLAassistant",
"junfengP"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8899",
"repo": "nacos-group/nacos-controller",
"url": "https://github.com/nacos-group/nacos-controller/pull/7"
}
|
gharchive/pull-request
|
feat: decouple config operation from nacos sdk
解除和Nacos SDK的强耦合
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
|
2025-04-01T04:34:51.073468
| 2016-05-06T12:04:49
|
153437470
|
{
"authors": [
"anonymau5",
"liamquin",
"nagadomi",
"panfengli"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8900",
"repo": "nagadomi/waifu2x",
"url": "https://github.com/nagadomi/waifu2x/issues/110"
}
|
gharchive/issue
|
Cuda issues.
Installed with the latest instructions and received the following. It seems that it is having trouble accessing CUDA although it is installed.
/usr/local/bin/luajit: /usr/local/share/lua/5.1/trepl/init.lua:383: lib/w2nn.lua:18: Failed to load CUDA modules. Please check the CUDA Settings.
stack traceback:
[C]: in function 'error'
/usr/local/share/lua/5.1/trepl/init.lua:383: in function 'require'
web.lua:7: in main chunk
[C]: in function 'dofile'
/usr/local/lib/luarocks/rocks/trepl/scm-1/bin/th:145: in main chunk
[C]: at 0x00404270
I added more informative error message at https://github.com/nagadomi/waifu2x/commit/958dcfaeb8dfd3c71df65282168b211f745ff970.
Could you retry and post the error message?
A little more information on this machine;
This machine uses a 550ti and has been running Waifu2x just fine for some time. I recently had a driver issue and had to purge and reinstall the nvidia drivers. The latest 352.63 drivers along with CUDA 7..5.18 is installed. There may be remnants of the older install still there.
/usr/local/bin/luajit: /usr/local/share/lua/5.1/trepl/init.lua:383: lib/w2nn.lua:18: Failed to load CUDA modules. Please check the CUDA Settings.
/usr/local/share/lua/5.1/trepl/init.lua:383: /usr/local/share/lua/5.1/cutorch/init.lua:2: cannot load '/usr/local/lib/lua/5.1/libcutorch.so'
stack traceback:
[C]: in function 'error'
/usr/local/share/lua/5.1/trepl/init.lua:383: in function 'require'
web.lua:7: in main chunk
[C]: in function 'dofile'
/usr/local/lib/luarocks/rocks/trepl/scm-1/bin/th:145: in main chunk
[C]: at 0x00404270
user@Mami:~/.Waifu2X$ sudo apt-get install libsnappy-dev
Reading package lists... Done
Building dependency tree
Reading state information... Done
libsnappy-dev is already the newest version.
0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded.
user@Mami:~/.Waifu2X$ sudo apt-get install libgraphicsmagick-dev
Reading package lists... Done
Building dependency tree
Reading state information... Done
Note, selecting 'libgraphicsmagick1-dev' instead of 'libgraphicsmagick-dev'
libgraphicsmagick1-dev is already the newest version.
0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded.
user@Mami:~/.Waifu2X$ sudo luarocks install graphicsmagick
Installing https://raw.githubusercontent.com/torch/rocks/master/graphicsmagick-1.scm-0.rockspec...
Using https://raw.githubusercontent.com/torch/rocks/master/graphicsmagick-1.scm-0.rockspec... switching to 'build' mode
Cloning into 'graphicsmagick'...
remote: Counting objects: 29, done.
remote: Compressing objects: 100% (29/29), done.
remote: Total 29 (delta 0), reused 19 (delta 0), pack-reused 0
Receiving objects: 100% (29/29), 788.37 KiB | 0 bytes/s, done.
Checking connectivity... done.
Updating manifest for /usr/local/lib/luarocks/rocks
graphicsmagick 1.scm-0 is now built and installed in /usr/local/ (license: BSD)
user@Mami:~/.Waifu2X$ sudo luarocks install lua-csnappy
Installing https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/lua-csnappy-0.1.3-1.src.rock...
Using https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/lua-csnappy-0.1.3-1.src.rock... switching to 'build' mode
gcc -O2 -fPIC -I/usr/local/include -c lsnappy.c -o lsnappy.o
gcc -shared -o snappy.so -L/usr/local/lib lsnappy.o
Updating manifest for /usr/local/lib/luarocks/rocks
lua-csnappy 0.1.3-1 is now built and installed in /usr/local/ (license: BSD)
user@Mami:~/.Waifu2X$ sudo luarocks install md5
Installing https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/md5-1.2-1.src.rock...
Using https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/md5-1.2-1.src.rock... switching to 'build' mode
gcc -O2 -fPIC -I/usr/local/include -c src/compat-5.2.c -o src/compat-5.2.o -Isrc/
gcc -O2 -fPIC -I/usr/local/include -c src/md5.c -o src/md5.o -Isrc/
gcc -O2 -fPIC -I/usr/local/include -c src/md5lib.c -o src/md5lib.o -Isrc/
gcc -shared -o md5/core.so -L/usr/local/lib src/compat-5.2.o src/md5.o src/md5lib.o
gcc -O2 -fPIC -I/usr/local/include -c src/compat-5.2.c -o src/compat-5.2.o -Isrc/
gcc -O2 -fPIC -I/usr/local/include -c src/des56.c -o src/des56.o -Isrc/
gcc -O2 -fPIC -I/usr/local/include -c src/ldes56.c -o src/ldes56.o -Isrc/
gcc -shared -o des56.so -L/usr/local/lib src/compat-5.2.o src/des56.o src/ldes56.o
Updating manifest for /usr/local/lib/luarocks/rocks
md5 1.2-1 is now built and installed in /usr/local/ (license: MIT/X11)
user@Mami:~/.Waifu2X$ sudo luarocks install uuid
Installing https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/uuid-0.2-1.src.rock...
Using https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/uuid-0.2-1.src.rock... switching to 'build' mode
Updating manifest for /usr/local/lib/luarocks/rocks
uuid 0.2-1 is now built and installed in /usr/local/ (license: Apache 2.0)
user@Mami:~/.Waifu2X$ sudo PREFIX=$HOME/torch/install luarocks install turbo
Installing https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/turbo-2.0-4.rockspec...
Using https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/turbo-2.0-4.rockspec... switching to 'build' mode
Cloning into 'turbo'...
remote: Counting objects: 165, done.
remote: Compressing objects: 100% (145/145), done.
remote: Total 165 (delta 28), reused 67 (delta 16), pack-reused 0
Receiving objects: 100% (165/165), 645.57 KiB | 0 bytes/s, done.
Resolving deltas: 100% (28/28), done.
Checking connectivity... done.
Note: checking out 'f43bffa558742548b4bba1f39ef2771d97d7041c'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b new_branch_name
Warning: variable CFLAGS was not passed in build_variables
make -C deps/http-parser library
make[1]: Entering directory /tmp/luarocks_turbo-2.0-4-6086/turbo/deps/http-parser' gcc -I. -DHTTP_PARSER_STRICT=0 -fPIC -Wall -Wextra -Werror -O3 -c http_parser.c -o libhttp_parser.o gcc -shared -Wl,-soname=libhttp_parser.so.2.1 -o libhttp_parser.so.2.1 libhttp_parser.o make[1]: Leaving directory /tmp/luarocks_turbo-2.0-4-6086/turbo/deps/http-parser'
gcc -Ideps/http-parser/ -shared -O3 -Wall -g -fPIC deps/http-parser/libhttp_parser.o deps/turbo_ffi_wrap.c -o libtffi_wrap.so -lcrypto -lssl
==== Installing Turbo.lua v2.0.4 to: ====
==== /home/user/torch/install/lib/lua/5.1 and ====
==== /home/user/torch/install/share/luajit-2.0.4 ====
mkdir -p /home/user/torch/install/lib
mkdir -p /home/user/torch/install/bin
mkdir -p /home/user/torch/install/share/lua/5.1/turbo
mkdir -p /home/user/torch/install/share/luajit-2.0.4/turbo
cp -r turbo/* /home/user/torch/install/share/lua/5.1/turbo
cp -r turbo.lua /home/user/torch/install/share/lua/5.1
cp -r turbovisor.lua /home/user/torch/install/share/lua/5.1
cp -r turbo/* /home/user/torch/install/share/luajit-2.0.4/turbo
cp -r turbo.lua /home/user/torch/install/share/luajit-2.0.4
cp -r turbovisor.lua /home/user/torch/install/share/luajit-2.0.4
install -m 0755 bin/turbovisor /home/user/torch/install/bin
==== Building 3rdparty modules ====
make -C deps/http-parser library
make[1]: Entering directory /tmp/luarocks_turbo-2.0-4-6086/turbo/deps/http-parser' gcc -shared -Wl,-soname=libhttp_parser.so.2.1 -o libhttp_parser.so.2.1 libhttp_parser.o make[1]: Leaving directory /tmp/luarocks_turbo-2.0-4-6086/turbo/deps/http-parser'
gcc -Ideps/http-parser/ -shared -O3 -Wall -g -fPIC deps/http-parser/libhttp_parser.o deps/turbo_ffi_wrap.c -o libtffi_wrap.so -lcrypto -lssl
==== Installing libtffi_wrap ====
test -f libtffi_wrap.so &&
install -m 0755 libtffi_wrap.so /home/user/torch/install/lib/libtffi_wrap.so.2.0.4 &&
ldconfig -n /home/user/torch/install/lib &&
ln -sf libtffi_wrap.so.2.0.4 /home/user/torch/install/lib/libtffi_wrap.so
==== Successfully installed Turbo.lua 2.0.4 to /home/user/torch/install ====
Updating manifest for /usr/local/lib/luarocks/rocks
turbo 2.0-4 is now built and installed in /usr/local/ (license: Apache 2.0)
user@Mami:~/.Waifu2X$ th waifu2x.lua
/usr/local/bin/luajit: /usr/local/share/lua/5.1/trepl/init.lua:383: lib/w2nn.lua:18: Failed to load CUDA modules. Please check the CUDA Settings.
/usr/local/share/lua/5.1/trepl/init.lua:383: /usr/local/share/lua/5.1/cutorch/init.lua:2: cannot load '/usr/local/lib/lua/5.1/libcutorch.so'
stack traceback:
[C]: in function 'error'
/usr/local/share/lua/5.1/trepl/init.lua:383: in function 'require'
waifu2x.lua:5: in main chunk
[C]: in function 'dofile'
/usr/local/lib/luarocks/rocks/trepl/scm-1/bin/th:145: in main chunk
[C]: at 0x00404270
It seems to be a cutorch(cuda module of torch data structure) issue.
Please try to reinstall that.
luarocks cutorch
luarocks cunn
user@Mami:~/.Waifu2X$ luarocks cutorch
Error: Unknown command: cutorch
user@Mami:~/.Waifu2X$ luarocks cunn
Error: Unknown command: cunn
user@Mami:~/.Waifu2X$
Ah, just noticed I need to do a sudo luarocks install cutorch Now I'm getting this.
user@Mami:~/.Waifu2X$ sudo luarocks install cutorch
Installing https://raw.githubusercontent.com/torch/rocks/master/cutorch-scm-1.rockspec...
Using https://raw.githubusercontent.com/torch/rocks/master/cutorch-scm-1.rockspec... switching to 'build' mode
Cloning into 'cutorch'...
remote: Counting objects: 107, done.
remote: Compressing objects: 100% (104/104), done.
remote: Total 107 (delta 8), reused 34 (delta 1), pack-reused 0
Receiving objects: 100% (107/107), 156.81 KiB | 0 bytes/s, done.
Resolving deltas: 100% (8/8), done.
Checking connectivity... done.
cmake -E make_directory build && cd build && cmake .. -DCMAKE_BUILD_TYPE=Release -DCMAKE_PREFIX_PATH="/usr/local/bin/.." -DCMAKE_INSTALL_PREFIX="/usr/local/lib/luarocks/rocks/cutorch/scm-1" && make -j$(getconf _NPROCESSORS_ONLN) install
-- The C compiler identification is GNU 4.8.4
-- The CXX compiler identification is GNU 4.8.4
-- Check for working C compiler: /usr/bin/cc
-- Check for working C compiler: /usr/bin/cc -- works
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++
-- Check for working CXX compiler: /usr/bin/c++ -- works
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Found Torch7 in /usr/local
CMake Error at /usr/share/cmake-2.8/Modules/FindCUDA.cmake:548 (message):
Specify CUDA_TOOLKIT_ROOT_DIR
Call Stack (most recent call first):
CMakeLists.txt:7 (FIND_PACKAGE)
-- Configuring incomplete, errors occurred!
See also "/tmp/luarocks_cutorch-scm-1-6154/cutorch/build/CMakeFiles/CMakeOutput.log".
Error: Build error: Failed building.
user@Mami:~/.Waifu2X$ sudo luarocks install cunn
Installing https://raw.githubusercontent.com/torch/rocks/master/cunn-scm-1.rockspec...
Using https://raw.githubusercontent.com/torch/rocks/master/cunn-scm-1.rockspec... switching to 'build' mode
Cloning into 'cunn'...
remote: Counting objects: 76, done.
remote: Compressing objects: 100% (70/70), done.
remote: Total 76 (delta 20), reused 17 (delta 2), pack-reused 0
Receiving objects: 100% (76/76), 93.95 KiB | 0 bytes/s, done.
Resolving deltas: 100% (20/20), done.
Checking connectivity... done.
cmake -E make_directory build && cd build && cmake .. -DCMAKE_BUILD_TYPE=Release -DCMAKE_PREFIX_PATH="/usr/local/bin/.." -DCMAKE_INSTALL_PREFIX="/usr/local/lib/luarocks/rocks/cunn/scm-1" && make -j$(getconf _NPROCESSORS_ONLN) install
-- The C compiler identification is GNU 4.8.4
-- The CXX compiler identification is GNU 4.8.4
-- Check for working C compiler: /usr/bin/cc
-- Check for working C compiler: /usr/bin/cc -- works
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++
-- Check for working CXX compiler: /usr/bin/c++ -- works
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Found Torch7 in /usr/local
CMake Error at /usr/share/cmake-2.8/Modules/FindCUDA.cmake:548 (message):
Specify CUDA_TOOLKIT_ROOT_DIR
Call Stack (most recent call first):
CMakeLists.txt:7 (FIND_PACKAGE)
-- Configuring incomplete, errors occurred!
See also "/tmp/luarocks_cunn-scm-1-480/cunn/build/CMakeFiles/CMakeOutput.log".
Error: Build error: Failed building.
Sorry, luarocks install cutorch and luarocks install cunn are correct.
If you use sudo, Maybe root does not read environment variables from user.
sudo su
export PATH=/usr/local/cuda/bin:$PATH
export LD_LIBRARY_PATH=/usr/local/cuda/lib64:$LD_LIBRARY_PATH
export CUDA_BIN_PATH=/usr/local/cuda/bin
luarocks install cutorch
luarocks install cunn
I recommend you to reinstall torch with non root environment.
See http://torch.ch/docs/getting-started.html
Unstall command: https://github.com/torch/ezinstall#updating-from-a-previous-version
Reinstalled torch with non root environment. Torch seems to have installed fine. No errors.
user@Mami:~/.Waifu2X$ th web.lua
/home/user/torch/install/bin/luajit: /home/user/torch/install/share/lua/5.1/trepl/init.lua:384: lib/w2nn.lua:18: Failed to load CUDA modules. Please check the CUDA Settings.
/home/user/torch/install/share/lua/5.1/trepl/init.lua:384: module 'cutorch' not found:No LuaRocks module found for cutorch
no field package.preload['cutorch']
no file 'lib/cutorch.lua'
no file '/home/user/.luarocks/share/lua/5.1/cutorch.lua'
no file '/home/user/.luarocks/share/lua/5.1/cutorch/init.lua'
no file '/home/user/torch/install/share/lua/5.1/cutorch.lua'
no file '/home/user/torch/install/share/lua/5.1/cutorch/init.lua'
no file './cutorch.lua'
no file '/home/user/torch/install/share/luajit-2.1.0-beta1/cutorch.lua'
no file '/usr/local/share/lua/5.1/cutorch.lua'
no file '/usr/local/share/lua/5.1/cutorch/init.lua'
no file '/home/user/.luarocks/lib/lua/5.1/cutorch.so'
no file '/home/user/torch/install/lib/lua/5.1/cutorch.so'
no file '/home/user/torch/install/lib/cutorch.so'
no file './cutorch.so'
no file '/usr/local/lib/lua/5.1/cutorch.so'
no file '/usr/local/lib/lua/5.1/loadall.so'
stack traceback:
[C]: in function 'error'
/home/user/torch/install/share/lua/5.1/trepl/init.lua:384: in function 'require'
web.lua:7: in main chunk
[C]: in function 'dofile'
...user/torch/install/lib/luarocks/rocks/trepl/scm-1/bin/th:145: in main chunk
[C]: at 0x00406670
I'm not sure why this newer version failed but I've rolled back to an older version with success.
It seems that cutorch is still not installed. I guess that CUDA is not detected from torch installer.
Is there nvcc?
nvcc --version
Then install cutorch
luarocks install cutorch
luarocks install cunn
I got an older version up and going. I'll copy the newer version and see if it'll run.
I found that FindCUDA didn't work if nvcc is in /bin - I had to put /usr/bin first in my PATH. This is because FindCUDA strips /bin/ from the directory when determining CUDA_HOME and if it's /bin then nothing is left.
I got everything built but then it wouldn't run; looks like my nvidia card is too old, as cutorch gives me Unsupported gpu architecture 'compute_52' - I gave up for now.
nvcc fatal : Unsupported gpu architecture 'compute_52'
I don't know your GPU model but this error is caused by old nvcc(CUDA 6.5?).
CUDA 7.5 is required.
Thank you for replying. Yes, 6.5, since 7.5 doesn't support my GPU - NVIDIA Corporation GT216GLM [Quadro FX 880M]. I did find a version that is patched (by NVidia) to support it, but it's a 1GByte download and unfortunately it also includes the kernel drivers, which will cause other problems on my system. So I gave up for now, will try on a different machine when I get the chance. I'd love to see waifu2x written as a GEGL operation by the way.
Bu fixing FindNVCC to work if nvcc is in /bin and not /usr/bin might be worthwhile as Linux systems are moving to have /usr/bin just be a symlink to /bin.
Try sudo bash /torch/update.sh
|
2025-04-01T04:34:51.079873
| 2020-06-05T23:53:37
|
632067473
|
{
"authors": [
"apalabrados",
"mcselasvegas",
"tinohager"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8901",
"repo": "nager/Nager.AmazonProductAdvertising",
"url": "https://github.com/nager/Nager.AmazonProductAdvertising/issues/140"
}
|
gharchive/issue
|
How to enable paging
Your demo site shows paging but I can not get it to work with the code from the repository. Is that something I need to enable?
Paging is spicify in the Request object:
SearchRequest searchRequest = new SearchRequest
{
Keywords = keyword,
ItemPage = **page**,
BrowseNodeId = browseNoseId,
Resources = new[]
{
"Images.Primary.Large",
"ItemInfo.Title",
"ItemInfo.Features",
"ItemInfo.TradeInInfo",
"ItemInfo.ByLineInfo",
"ItemInfo.ManufactureInfo",
}
};
So, for every request, you should set the page as you can see above.
https://github.com/nager/Nager.AmazonProductAdvertising/blob/9fdb77392138e8bf4b26f6f3813f6af453545d13/src/Nager.AmazonProductAdvertising/Model/SearchRequest.cs#L8
|
2025-04-01T04:34:51.081519
| 2022-06-10T13:02:03
|
1267514150
|
{
"authors": [
"dayaftereh",
"iilyak"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8902",
"repo": "nagix/chartjs-plugin-streaming",
"url": "https://github.com/nagix/chartjs-plugin-streaming/issues/165"
}
|
gharchive/issue
|
Streaming for non timebased data
I have a stream of data which doesn't have a timestamp. The rate of events is so high that I need a microseconds resolution to use timestamp based approach. I know that JS doesn't support microseconds in Date object This means calling new Date() will not work.
I have a monotonically increasing event id (integer). Is there a way I can implement a scrolling chart similar to the one provided by streaming plugin (with zoom feature) using linear scale instead of time or realtime?
+1
|
2025-04-01T04:34:51.086468
| 2023-12-07T07:59:22
|
2030119838
|
{
"authors": [
"Muni10"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8903",
"repo": "nais/nitro",
"url": "https://github.com/nais/nitro/issues/6"
}
|
gharchive/issue
|
Support new clusters in nitro
The current version of nitro does not support new clusters because of the etcd health check. We need something that handles this.
Added --newCluster flag to use for first time install
|
2025-04-01T04:34:51.109289
| 2020-05-24T06:54:53
|
623808631
|
{
"authors": [
"Anish-M-code",
"nandhabn"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8904",
"repo": "nandhabn/python-Flask-Chat-App",
"url": "https://github.com/nandhabn/python-Flask-Chat-App/issues/3"
}
|
gharchive/issue
|
Request to add Software License
Adding Software License makes the project usable by others for their own works.
It also protects your works from being used in unjust manner. It is an important part of developing opensource projects.
This source might help in choosing one : https://choosealicense.com/
LICENSE added. now you can use this project.
|
2025-04-01T04:34:51.110568
| 2021-03-12T08:43:53
|
829888050
|
{
"authors": [
"minhchienwikipedia",
"nandorojo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8905",
"repo": "nandorojo/moti",
"url": "https://github.com/nandorojo/moti/issues/35"
}
|
gharchive/issue
|
How to create animation component with moti?
Hi, I want to create a custom animation component with Moti, how can I do it?
Ex: I'm using LinearGradient I want to custom like Animated.createAnimatedComponent(LinearGradient) and then I can using animation for this component
https://moti.fyi/api/motify
|
2025-04-01T04:34:51.136065
| 2023-06-04T14:22:52
|
1740402249
|
{
"authors": [
"arsantiagolopez",
"axe312ger",
"nandorojo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8906",
"repo": "nandorojo/solito",
"url": "https://github.com/nandorojo/solito/issues/398"
}
|
gharchive/issue
|
Unable to get any starter running on iOS simulator
Maybe some updates broke sth? Looks like metro has issues.
I run MacOS + Node v18, freshly installed XCode + Simulator.
No matter which starter I pick, I get as soon I open the app in the simulator:
Unable to resolve "./App" from "index.js"
The web variant works fine for each starter.
I tried the expo starter, it works fine with my iOS Simulator.
Any ideas? 🤔
does that file exist
@nandorojo thx for super fast reply.
But index.js requries it? 🤔
import { registerRootComponent } from 'expo';
import App from './App';
// registerRootComponent calls AppRegistry.registerComponent('main', () => App);
// It also ensures that whether you load the app in Expo Go or in a native build,
// the environment is set up appropriately
registerRootComponent(App);
does App exist there? the error says it doesn’t.
I don't know, I executed npx create-solito-app@latest and wanted to see how it looks in the browser and in the iOS simulator.
I did zero changes to the files.
I am assuming, that the starters are ready to be used with device simulators. Am I wrong here?
you’re not wrong, but it would still help to check
is there any other expo port or app running currently? if so, maybe try clearing all your ports out
Ill restart my machine and close all the terminals 😇
So, some delays, but I did a fresh start, after restarting, some updates and making sure nothing else is running.
I still end up with the same error message.
Have to note that the troublesome index.js is created when executing npx expo run:ios
Full log
❯ npx create-solito-app@latest my-solito-app-fresh-gh-tailwind -t with-tailwind
🧑🎤 Creating solito app...
Creating a new solito app in my-solito-app-fresh-gh-tailwind...
my-solito-app-fresh-gh-tailwind folder created.
Copying template into my-solito-app-fresh-gh-tailwind...
Downloaded template into my-solito-app-fresh-gh-tailwind...
my-solito-app-fresh-gh-tailwind created!
Installing packages. This might take a couple of minutes.
> yarn install
➤ YN0000: ┌ Resolution step
➤ YN0002: │ @expo/cli@npm:0.6.2 doesn't provide expo-modules-autolinking (pe2591), requested by @expo/prebuild-config
➤ YN0002: │ app@workspace:packages/app doesn't provide react (p0cb3a), requested by @react-navigation/native-stack
➤ YN0002: │ app@workspace:packages/app doesn't provide react (p9a3d3), requested by @react-navigation/native
➤ YN0002: │ app@workspace:packages/app doesn't provide react-native (p0e2b8), requested by @react-navigation/native-stack
➤ YN0002: │ app@workspace:packages/app doesn't provide react-native (p34a6c), requested by @react-navigation/native
➤ YN0002: │ app@workspace:packages/app doesn't provide react-native-reanimated (pe7eec), requested by moti
➤ YN0002: │ app@workspace:packages/app doesn't provide react-native-safe-area-context (pecff1), requested by @react-navigation/native-stack
➤ YN0002: │ app@workspace:packages/app doesn't provide react-native-screens (p96773), requested by @react-navigation/native-stack
➤ YN0002: │ app@workspace:packages/app doesn't provide tailwindcss (pa880e), requested by nativewind
➤ YN0002: │ babel-preset-expo@npm:9.3.2 doesn't provide @babel/core (pe14e9), requested by @babel/plugin-proposal-object-rest-spread
➤ YN0002: │ babel-preset-expo@npm:9.3.2 doesn't provide @babel/core (pfe2df), requested by @babel/plugin-proposal-decorators
➤ YN0002: │ babel-preset-expo@npm:9.3.2 doesn't provide @babel/core (p9681f), requested by @babel/plugin-transform-react-jsx
➤ YN0002: │ babel-preset-expo@npm:9.3.2 doesn't provide @babel/core (p3ab61), requested by @babel/preset-env
➤ YN0002: │ expo-asset@npm:8.9.1 doesn't provide expo (p79f18), requested by expo-constants
➤ YN0002: │ expo-asset@npm:8.9.1 doesn't provide expo (p185f2), requested by expo-file-system
➤ YN0002: │ expo-linking@npm:4.0.1 doesn't provide expo (pe2296), requested by expo-constants
➤ YN0002: │ expo-splash-screen@npm:0.18.1 [72208] doesn't provide expo-modules-autolinking (p45296), requested by @expo/prebuild-config
➤ YN0002: │ moti@npm:0.24.2 [f87a9] doesn't provide react (p17866), requested by framer-motion
➤ YN0002: │ moti@npm:0.24.2 [f87a9] doesn't provide react-dom (p750ce), requested by framer-motion
➤ YN0002: │ nativewind@npm:2.0.9 [f87a9] doesn't provide react (p0b0c3), requested by use-sync-external-store
➤ YN0002: │ next-app@workspace:apps/next doesn't provide eslint (p29e84), requested by eslint-config-next
➤ YN0002: │ next-app@workspace:apps/next doesn't provide expo (p8f46e), requested by @expo/next-adapter
➤ YN0002: │ next-app@workspace:apps/next doesn't provide postcss (p2228b), requested by autoprefixer
➤ YN0002: │ next-app@workspace:apps/next doesn't provide react (p2df79), requested by @expo/next-adapter
➤ YN0002: │ next-app@workspace:apps/next doesn't provide react (p8afd9), requested by next
➤ YN0002: │ next-app@workspace:apps/next doesn't provide react-dom (pb1bef), requested by next
➤ YN0002: │ next-app@workspace:apps/next doesn't provide react-native-web (p9df76), requested by @expo/next-adapter
➤ YN0002: │ next-app@workspace:apps/next doesn't provide webpack (pa5629), requested by @expo/next-adapter
➤ YN0002: │ react-native-codegen@npm:0.71.5 doesn't provide @babel/preset-env (pdcb72), requested by jscodeshift
➤ YN0000: │ Some peer dependencies are incorrectly met; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code
➤ YN0000: └ Completed
➤ YN0000: ┌ Fetch step
➤ YN0013: │ yargs-parser@npm:18.1.3 can't be found in the cache and will be fetched from the remote registry
➤ YN0013: │ yargs-parser@npm:21.1.1 can't be found in the cache and will be fetched from the remote registry
➤ YN0013: │ yargs@npm:15.4.1 can't be found in the cache and will be fetched from the remote registry
➤ YN0013: │ yargs@npm:17.7.1 can't be found in the cache and will be fetched from the remote registry
➤ YN0013: │ yocto-queue@npm:0.1.0 can't be found in the cache and will be fetched from the remote registry
➤ YN0000: └ Completed in 4s 417ms
➤ YN0000: ┌ Link step
➤ YN0007: │ turbo@npm:1.8.8 must be built because it never has been before or the last one failed
➤ YN0000: └ Completed in 5s 522ms
➤ YN0000: Done with warnings in 10s 260ms
Success! Created my-solito-app-fresh-gh-tailwind at my-solito-app-fresh-gh-tailwind
Inside that directory, you can run several commands:
yarn web
Starts the development server for the Next.js site.
yarn native
We suggest that you begin by typing:
cd my-solito-app-fresh-gh-tailwind
yarn web
~/dev/food-diary via 11GiB/32GiB took 17s
❯ cd my-solito-app-fresh-gh-tailwind
~/dev/food-diary/my-solito-app-fresh-gh-tailwind via v18.16.0 via 14GiB/32GiB
❯ code .
~/dev/food-diary/my-solito-app-fresh-gh-tailwind via v18.16.0 via 14GiB/32GiB
❯ npm run
Scripts available in solito-with-tailwind via `npm run-script`:
native
cd apps/expo && yarn start
web
cd apps/next && yarn next
~/dev/food-diary/my-solito-app-fresh-gh-tailwind via v18.16.0 via 14GiB/32GiB
❯ npm run web
> web
> cd apps/next && yarn next
ready - started server on <IP_ADDRESS>:3000, url: http://localhost:3000
event - compiled client and server successfully in 1285 ms (413 modules)
wait - compiling / (client and server)...
event - compiled client and server successfully in 1251 ms (1348 modules)
^C
~/dev/food-diary/my-solito-app-fresh-gh-tailwind via v18.16.0 via 14GiB/32GiB took 14s
❯ npm run native
> native
> cd apps/expo && yarn start
Development build: Unable to get the default URI scheme for the project. Please make sure the expo-dev-client package is installed.
Starting project at /Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind/apps/expo
Starting Metro Bundler
▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄
█ ▄▄▄▄▄ █▄▀▀▄▄▀▀█▄█ ▄▄▄▄▄ █
█ █ █ ███▄█ ▀▀▄█ █ █ █
█ █▄▄▄█ ██▄▀▄▀███▀█ █▄▄▄█ █
█▄▄▄▄▄▄▄█ █ ▀▄▀ █ █▄▄▄▄▄▄▄█
█ ▄▀▄▄▀▄█▀ ▄▄▀▀█ ██▄▀▀▀▀▀▄█
█▀▀▀▀▀█▄▄▀▀ ▀█▄ ▄▄ ▀ ▀▀ █
█▀ ▄█▀ ▄▀▄▄█▄▄▀▄ ▀█▄▄▀██▀▄█
█ ▄▄ ▄▀▀██ ▄▄█ █▀ █▄ ▄ █
█▄████▄▄█ █ ▀▀▀▄█ ▄▄▄ █▄ ██
█ ▄▄▄▄▄ ██ █▄▀█▀▀ █▄█ ▄█▀ █
█ █ █ █ █▄█▄▄ ▄ ▄▄ █ █
█ █▄▄▄█ █▀▀▀█▄█ ▄▄▄█ █ █
█▄▄▄▄▄▄▄█▄▄███▄▄█▄▄▄██▄██▄█
› Metro waiting on http://localhost:8081
› Scan the QR code above with Expo Go (Android) or the Camera app (iOS)
› Press a │ open Android
› Press i │ open iOS simulator
› Press w │ open web
› Press j │ open debugger
› Press r │ reload app
› Press m │ toggle menu
› Press ? │ show all commands
Logs for your project will appear below. Press Ctrl+C to exit.
› Opening on iOS...
CommandError: No development build (com.solito.nativewind) for this project is installed. Please make and install a development build on the device first.
Learn more
› Stopped server
~/dev/food-diary/my-solito-app-fresh-gh-tailwind via v18.16.0 via 17GiB/32GiB took 41s
❯ npx expo run:ios
📝 iOS Bundle Identifier Learn more
✔ What would you like your iOS bundle identifier to be? … com.anonymous.solito-with-tailwind
✔ Created native project | gitignore skipped
✔ Added Metro config
✔ Updated package.json and added index.js entry point for iOS and Android
› Installing using yarn
> yarn install
» ios: icon: This is the image that your app uses on your home screen, you will need to configure it manually.
✔ Config synced
✔ Installed pods and initialized Xcode workspace.
To address this warning, either add output dependencies to the script phase, or configure it to run in every build by unchecking "Based on dependency analysis" in the script phase. (in target 'solitowithtailwind' from project 'solitowithtailwind')
Run script build phase 'Bundle React Native code and images' will be run during every build because it does not specify any outputs. To address this warning, either add output dependencies to the script phase, or configure it to run in every build by unchecking "Based on dependency analysis" in the script phase. (in target 'solitowithtailwind' from project 'solitowithtailwind')
Run script build phase '[CP-User] Generate app.config for prebuilt Constants.manifest' will be run during every build because it does not specify any outputs. To address this warning, either add output dependencies to the script phase, or configure it to run in every build by unchecking "Based on dependency analysis" in the script phase. (in target 'EXConstants' from project 'Pods')
› Build Succeeded
› 0 error(s), and 4 warning(s)
Starting Metro Bundler
▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄
█ ▄▄▄▄▄ █▄▄▄ ▀█▄█▄▄█▀▀▄ ███▄█▀█ ▄▄▄▄▄ █
█ █ █ ██▄▀ █ ▀ ▄▀▄▄▄ █▀▀ █ █ █ █ █
█ █▄▄▄█ ██▀▄ ▄███▀ ▀▄█▀█▀▀█ █▄█ █▄▄▄█ █
█▄▄▄▄▄▄▄█ ▀▄█ ▀▄▀ █ ▀ █▄█ ▀ ▀ █▄▄▄▄▄▄▄█
█ █ ▀▄██▄▀█▄▀▀ ▀▄ ▀███▀█▀█ █▀▄▀▀▀██▄█
█ ▀ ▄ ▀▄█ ▀█▄▄█▄ ██▀▄ ▄█▄▄▄█▀ ▄ █ █
█ ▄ █ █▄██▄██▀▀▄ ██▄█▀▀▄▀▄█ ▀ █▄▄▀ ▀▄▄█
█ █▀▄ ▄ ▄▀▄█▀▄█ ▄▄▄▄█▄██▄█▄▄▀ █ █▀▀▄▄█
██▀▄▀ █▄ ▀▀ ▄▄ ▀▀ ▀ █ █▄ ▄▀█ ▄▀█ ▄▀██
██▄ █▀▄▀▄██▀▄▀▄ ▄▄▄▄ █ ▄█▀█▄█▀▀██ █ █
█ ▀▄▄ ▄▄▄▀ ▀▄ ██ ▀▄ ▄▀▀▀▄█ ▄ ▄ ██▄██
█▄ ▀█▄ ▄▀▄▄█ ▀█ ▄ ██▀▄ ▄▄█▀▄██▄▄█▀▀█▄█
█▀ ▄▀▀▀▄▀▄ ▀ ▀▀ ▀██▄▄█▀█ ███ ▀█▄█▀ ▀███
█ ▄▀█▄ ▄ █ ██ ▄▄▄▀▄ ▄▄▄██▄▀▄█▀▀ ▄▀ ▄█
█▄██▄▄▄▄█ ▀▀█▀▄ ▀ ██ ▄▄▄▀█ ▄▄▄ ▄ ▄█
█ ▄▄▄▄▄ ██▄▄▄ ▀▄ █▀▄█▄▀▄▄ █ █▄█ ▀█▀▄█
█ █ █ █ █▀▀▀▄ ▄▄▄ ▀▀▀▀▀▄▀ ▄ ▄▄▀▀█▄█
█ █▄▄▄█ █▀▀▀ █ ▀ ██ █ █▄▀▄▄█▀▀ ▀▀▀ █
█▄▄▄▄▄▄▄█▄█████▄▄██▄▄████▄█▄█▄▄▄▄▄███▄█
› Metro waiting on com.anonymous.solito-with-tailwind://expo-development-client/?url=http%3A%2F%2F<IP_ADDRESS>%3A8081
› Scan the QR code above with Expo Go (Android) or the Camera app (iOS)
› Press a │ open Android
› Press i │ open iOS simulator
› Press w │ open web
› Press j │ open debugger
› Press r │ reload app
› Press m │ toggle menu
› Press ? │ show all commands
› Installing on iPhone 14 Pro Max
warning: Watchman `watch-project` returned a warning: Recrawled this watch 3 times, most recently because:
MustScanSubDirs UserDroppedTo resolve, please review the information on
https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl
To clear this warning, run:
`watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'`
warning: Watchman `watch-project` returned a warning: Recrawled this watch 3 times, most recently because:
MustScanSubDirs UserDroppedTo resolve, please review the information on
https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl
To clear this warning, run:
`watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'`
warning: Watchman `watch-project` returned a warning: Recrawled this watch 3 times, most recently because:
MustScanSubDirs UserDroppedTo resolve, please review the information on
https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl
To clear this warning, run:
`watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'`
warning: Watchman `watch-project` returned a warning: Recrawled this watch 3 times, most recently because:
MustScanSubDirs UserDroppedTo resolve, please review the information on
https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl
To clear this warning, run:
`watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'`
warning: Watchman `watch-project` returned a warning: Recrawled this watch 3 times, most recently because:
MustScanSubDirs UserDroppedTo resolve, please review the information on
https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl
To clear this warning, run:
`watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'`
warning: Watchman `query` returned a warning: Recrawled this watch 3 times, most recently because:
MustScanSubDirs UserDroppedTo resolve, please review the information on
https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl
To clear this warning, run:
`watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'`
Recrawled this watch 3 times, most recently because:
MustScanSubDirs UserDroppedTo resolve, please review the information on
https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl
To clear this warning, run:
`watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'`
› Opening on iPhone 14 Pro Max (com.anonymous.solito-with-tailwind)
› Opening com.anonymous.solito-with-tailwind://expo-development-client/?url=http%3A%2F%2F<IP_ADDRESS>%3A8081 on iPhone 14 Pro Max
› Opening the iOS simulator, this might take a moment.
› Logs for your project will appear below. Press Ctrl+C to exit.
iOS Bundling failed 257ms
Unable to resolve "./App" from "index.js"
Am I missing something thats needs to be done to configure expo for iOS correctly?
I'd guess non EAS usage is support?
🙈
i don’t think so, but maybe try EAS on your device?
I got it running! I executed npm run native from the project root dir. This does not work.
When I directly enter apps/expo, then do yarn and then execute npm run ios, the solito app opens in iOS simulator :)
hmmm okay good to know, thanks.
does yarn start in apps/expo work?
https://github.com/nandorojo/solito/blob/89c9dbb79d377ad663c326904bcfbed764caf6cd/example-monorepos/blank/package.json#L16
running yarn ios builds the dev client, but yarn native starts the dev server once the dev client is already on the device
does yarn start in apps/expo work?
https://github.com/nandorojo/solito/blob/89c9dbb79d377ad663c326904bcfbed764caf6cd/example-monorepos/blank/package.json#L16
Yes it does! :)
so yarn native from the root must work too, right?
@nandorojo sorry for no/late response.
Yes it should, at least everything was working fine the last time I was fiddling around.
Will soon restart working on that project, hopefully then this will be the last comment here :)
Thanks for your work!
Stumbled across this comment as I'm experiencing the same thing a couple months later.
Upon following the docs and running npx create-solito-app@latest, running yarn native gives me this error:
CommandError: No development build (com.solito.nativewind) for this project is installed. Please make and install a development build on the device first.
Learn more: https://docs.expo.dev/development/build/
The only way I got it to work was by following the steps suggested by @axe312ger
I got it running! I executed npm run native from the project root dir. This does not work.
When I directly enter apps/expo, then do yarn and then execute npm run ios, the solito app opens in iOS simulator :)
|
2025-04-01T04:34:51.150025
| 2018-05-14T23:31:14
|
323016586
|
{
"authors": [
"danhunsaker",
"fsecada01",
"vashchukmaksim"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8908",
"repo": "nanobox-io/nanobox-engine-python",
"url": "https://github.com/nanobox-io/nanobox-engine-python/issues/45"
}
|
gharchive/issue
|
python-3.6 and psycopg2 (Error: pg_config executable not found.)
I'm trying to setup an environment with nanobox python engine. If I use python-2.7 and then install a psycopg2 package (pip command within a container or with a requirements.txt) it works ok. But when I switch to python-3.6 I got an error "Error: pg_config executable not found.". This is a well known issue but I can't solve it with a nanobox container. I can't install any dependencies (python3-dev or libpq-dev) since there is none of these in Nanobox pkgsrc and even can't find a path to pg_conf.
Is there any way to use python3 with psycopg2 with nanobox?
My starting config:
run.config:
engine: python
engine.config:
runtime: python-3.6
extra_packages:
- nodejs
cache_dirs:
- client/node_modules
extra_path_dirs:
- client/node_modules/.bin
build_triggers:
- server/requirements.txt
- client/package.json
- client/package-lock.json
extra_steps:
- cd server && pip install -r requirements.txt
- cd client && npm install
fs_watch: true
data.postgresql:
image: nanobox/postgresql
config:
version: 9.6
`
in extra_packages, declare pkgin as a new dependency.
After that, write nanobox run in your command line, and write in pkgin install postgresql93-plpython'. Then write pkgin export > pkgin_dependencies.txt' or whatever extension is your notepad application. Then declare in extra_steps:
'pkgin im pkgin_dependencies.txt -y'
That should resolve your issue, hopefully.
Note that pkgin is already installed, since it's the package manager installing extra_packages in the first place, so you can skip adding it to your configuration.
|
2025-04-01T04:34:51.172828
| 2021-11-12T10:27:15
|
1051809585
|
{
"authors": [
"iGranDav",
"jeanphibaconnais"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8909",
"repo": "nantes-tech-communities/list-communities",
"url": "https://github.com/nantes-tech-communities/list-communities/pull/39"
}
|
gharchive/pull-request
|
Add CocoaHeads Nantes event
Adding a new Nantes event : CocoaHeads
Thanks @iGranDav for your contribution 🙏
@all-contributors please add @iGranDav for doc
|
2025-04-01T04:34:51.175994
| 2021-09-18T14:18:57
|
1000046349
|
{
"authors": [
"disasterdaddy"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8910",
"repo": "naofireblade/homebridge-weather-plus",
"url": "https://github.com/naofireblade/homebridge-weather-plus/issues/215"
}
|
gharchive/issue
|
Error "code": "CDN-0001", "message": "Invalid apiKey."
I installed version "homebridge-weather-plus v3.2.6" and get the error "code": "CDN-0001", "message": "Invalid apiKey during start. The config is:
{
"platform": "WeatherPlus",
"service": "weatherunderground",
"key": "ip9999999",
"stationId": "IGELXXX9",
"compatibility": "home"
}
I know that the apiKey and the stationId have correct values (not the value in the config here). What could be the clue for this problem?
Best regards
As reviewed via Google I had to generate a new apiKey via WU. Go to https://www.wunderground.com/member/settings, then select "API KEYS" and generate the new key. This key is only for your plugin. Copy it to your clipboard and paste it into your JSON plugin configuration. Restart your HomeBridge and everything will work. You will find your new values in the "Standard Room".
|
2025-04-01T04:34:51.195836
| 2023-03-13T10:01:51
|
1621146517
|
{
"authors": [
"angely-dev",
"codingnetworksb"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8911",
"repo": "napalm-automation-community/napalm-huawei-vrp",
"url": "https://github.com/napalm-automation-community/napalm-huawei-vrp/pull/23"
}
|
gharchive/pull-request
|
Add contextual diff computation mode
Problem
As you may know, the current implementation of _get_merge_diff() is not really sufficient.
Something as simple as:
Candidate
Running (extract)
#
interface GigabitEthernet0/0/2
undo portswitch
undo shutdown
l2 binding vsi CUST
#
bgp 1234
router-id <IP_ADDRESS>
ipv4-family vpn-instance SOME-VPN
import-route direct
#
#
interface GigabitEthernet0/0/1
description CustomerA
undo portswitch
l2 binding vsi CUST
#
interface GigabitEthernet0/0/2
description CustomerB
shutdown
#
bgp 1234
router-id <IP_ADDRESS>
ipv4-family vpn-instance SOME-VPN
import-route static
#
Using the basic Python sample:
from napalm import get_network_driver
driver = get_network_driver('huawei_vrp')
device = driver(hostname='switch', username='user', password='pass')
device.open()
device.load_merge_candidate(config=candidate)
print(device.compare_config())
Will result in:
undo shutdown
import-route direct
Inconsistencies:
New items from GigabitEthernet0/0/2 are not displayed
New items are displayed (if found) without any context
In other words, the diff is unable to compute changes per-context (per-indented block).
Proposed solution
I ended up doing DiffPlus, a lightweight module to compute an incremental and contextual diff between two indented configs. Taking about a hundred lines of code, it only relies on Python builtins and has no extra dependencies.
Though I did it originally for this specific use case, the module is generic enough to work with any indented config, theoretically. I ran it successfully on multiple Huawei platforms (S5732, NE20, NE05, CE6810) having from 2K to 16K number of lines.
Demo
The module is now on PyPI.
Diff only
Using the same configs as above:
from diffplus import IndentedConfig, IncrementalDiff
candidate = open('candidate.txt').read()
running = open('running.txt').read()
candidate = IndentedConfig(candidate, sanitize=True) # sanitize to remove comments, blank lines, etc.
running = IndentedConfig(running, sanitize=True)
diff = IncrementalDiff(candidate, running, colored=True) # colored mode is optional
print(diff)
Will output:
interface GigabitEthernet0/0/2
+ undo portswitch
+ undo shutdown
+ l2 binding vsi CUST
bgp 1234
ipv4-family vpn-instance SOME-VPN
+ import-route direct
Merging
Alternatively, we can merge to get a preview of the full config before applying it:
diff = IncrementalDiff(candidate, running, colored=True, merge=True)
Will output:
interface GigabitEthernet0/0/1
description CustomerA
undo portswitch
l2 binding vsi CUST
interface GigabitEthernet0/0/2
description CustomerB
shutdown
+ undo portswitch
+ undo shutdown
+ l2 binding vsi CUST
bgp 1234
router-id <IP_ADDRESS>
ipv4-family vpn-instance SOME-VPN
import-route static
+ import-route direct
Limitations
It is important to understand this module does NOT check syntax or semantic. By essence, it computes the diff between two indented texts, which makes it both simple and generic.
Also, deletions cannot be computed. I explain why here with an example. To sum up, it would require to re-implement the config logic in some way. Though it is tempting, that is far from easy and very platform-dependent.
The diff in NAPALM Cisco IOS actually fakes deletions this way: a no command is considered a deletion even if it won't negate any existing command. This is very debatable and does not make sense to me.
Changes in NAPALM Huawei VRP
Option 1: modify _get_merge_diff()
You may find this option too radical.
def _get_merge_diff(self):
running_config = self.get_config(retrieve='running')['running']
running_config = IndentedConfig(running_config, sanitize=True)
merge_candidate = IndentedConfig(self.merge_candidate, sanitize=True)
diff = IncrementalDiff(merge_candidate, running_config)
return diff
Option 2: add an optional_args
This option is more prudent. We let users choose the diff computation mode: either they stick with the actual mode (by default, for legacy purposes) or they experiment the contextual mode through an optional argument contextual_diff.
device = driver(
hostname='switch',
username='user',
password='pass',
optional_args={'contextual_diff': True}
)
This is what I implemented in the pull request and I'd be glad to have your view.
Any updates on my PR?
Any updates on my PR?
Hi @angely-dev sorry i haven't replied before. I will read and review your proposal in detail this weekend.
Thanks @codingnetworksb. I am available for any question or to resolve the conflict which just appeared after you recent activity.
Hi @angely-dev, this weekend i started to update the library because it was very outdated compared with original napalm library. I'm making sure it in complete compliance with getters structure, output, etc. I'm also implementing mocked data to for automatic testing when we make pull request.
I couldn't review your pull request. I will do it as soon as i finish updating, hope during the week or weekend. And i'll let you know at that time to check the conflicts.
Hi @angely-dev, i'm done until now with my updates. Could you check and fix the conflicts?
Thanks for maintaining the repo @codingnetworksb. Please allow some time to fix the conflicts which appeared.
So I checked and fixed the conflict @codingnetworksb. It was just a reformatting issue (see 502e930).
Hi @vladislav-tenishchev
The new code failed the Checks i have included recently. It's failing in Black format check. As napalm original project, we have included Black formatter check to this project, please format the code using Black formatter and push the code again.
We appreciate your contribution to this project.
So I did run the Black code formatter @codingnetworksb and committed the changes in my fork (https://github.com/napalm-automation-community/napalm-huawei-vrp/pull/23/commits/a8a8bf73e63673dd9f651113a286214edda6f1a3). It is OK for you at this point? Thanks!
Any updates @codingnetworksb? Or some concerns about the PR?
Not only I fixed the failed checks, but I also added a workflow in my project as well 😊 It includes the Black code formatting check and it ensures unit tests have passed (from Python 3.8 to 3.12). I also released the new version on PyPI.
Any updates @codingnetworksb? Or some concerns about the PR?
Not only I fixed the failed checks, but I also added a workflow in my project as well 😊 It includes the Black code formatting check and it ensures unit tests have passed (from Python 3.8 to 3.12). I also released the new version on PyPI.
No concern about the PR. I got distracted with other projects, but this weekend I'll catch up on these PRs
Hi @angely-dev ,
I was testing the PR in my local environment, using a Virtual NE40E and seems to work as you explain it. Too sad that deletions cannot be computed. I understood the complexity of this.
Thank you for documenting this PR so well.
I will approve and merge today.
Regards,
Hi @codingnetworksb,
Thanks for the feedback and for approving the PR. I hope users will take benefit of it despite the limitations.
I realized I didn't include an update of the README in the PR. We could update it this way:
Make a brief mention of the contextual_diff option In the Quick Start section and link it to the PR for reference:
from napalm import get_network_driver
driver = get_network_driver('huawei_vrp')
device = driver(
hostname='<IP_ADDRESS>',
username='admin',
password='this_is_not_a_secure_password',
optional_args={'contextual_diff': True} # enable contextual diff mode
)
device.open()
# Send Any CLI command
# …
Please refer to PR #23 for more information about the contextual_diff mode.
Or make it more explicit by adding a new subsection like this: Diff mechanism
I prefer the second way, but I let you choose if you are willing to update the README directly. Otherwise, I could do another PR just for this update.
Hi @angely-dev ,
Both options sounds good to me. Please submit a new PR just that update. If you create a new subsection make it as summary as possible.
|
2025-04-01T04:34:51.201834
| 2022-06-06T10:58:27
|
1261684687
|
{
"authors": [
"haesleinhuepf",
"jni"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8912",
"repo": "napari/napari",
"url": "https://github.com/napari/napari/issues/4623"
}
|
gharchive/issue
|
Viewer close event
🚀 Feature
The viewer should have an suscribable event that is fired once the viewer is closed.
Motivation
I would like to delete a temporary directory at the point in time when the viewer is closed. I am responsible for when to do this, because tempfile.mkdtemp is not capable of doing it.
The napari event documentation does not list a suitable event for this:
https://napari.org/guides/events_reference.html?highlight=event
Pitch
It would be cool if we could subscribe to an event such as
viewer.events.closed.connect(my_close_event)
Alternatives
Additional context
We've had some discussions about this before. @haesleinhuepf do you want this to happen before e.g. all the layers are cleaned up, or after, or don't care? As I recall someone else asked for this to happen before anything is actually done. In that case I might be inclined to call it "about_to_close" or "closing"
Hey @jni ,
I personally don't care. I think QT and/or the Windows API sends two events: closing(flag) and closed(), and I find the name intuitive. Using the flag it can be cancelled. Maybe it would be cool to give developers the same flexibility here...
Just adding a reference: http://www.java2s.com/Tutorial/CSharp/0470__Windows-Presentation-Foundation/WindowClosingandClosedevent.htm
Sorry for the sharp C 😜
|
2025-04-01T04:34:51.232539
| 2020-11-24T16:33:57
|
749863848
|
{
"authors": [
"Czaki",
"ziyangczi"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8913",
"repo": "napari/napari",
"url": "https://github.com/napari/napari/pull/1937"
}
|
gharchive/pull-request
|
Unpin qt version
Description
Because of this bug
https://bugreports.qt.io/browse/QTBUG-88688
Debian based image cannot be used to test Qt 5.15.2 release.
A possible option is to switch to the ubuntu based image. This is done in this PR. But this increase build time.
The possible workaround I to build a custom docker image for this. But then someone needs to maintain it.
The debug process is hidden in #1935 which is a register of the brute force approach.
Type of change
[ ] Bug-fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] This change requires a documentation update
References
How has this been tested?
[ ] example: the test suite for my feature covers cases x, y, and z
[ ] example: all tests pass with my change
Final checklist:
[ ] My PR is the minimum possible work for the desired functionality
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[ ] I have added tests that prove my fix is effective or that my feature works
I can look into building container for the test, thanks for the note!
I can look into building container for the test, thanks for the note!
I could prepare such script and instruction.
I suggest preparing a separate repository in napari and add a schedule to automatically rebuild every week and push on docker hub.
Until now, my experience is limited to preparing Dockerfile and manual upload to docker hub (for package imagecodecs).
I can look into building container for the test, thanks for the note!
I could prepare such script and instruction.
I suggest preparing a separate repository in napari and add a schedule to automatically rebuild every week and push on docker hub.
Until now, my experience is limited to preparing Dockerfile and manual upload to docker hub (for package imagecodecs).
I created a napari org in docker hub before to release docker version of napari, we can add another repo there, I can send an invite to you, can you share your account in an email to<EMAIL_ADDRESS>
I sent You mail.
|
2025-04-01T04:34:51.244689
| 2023-07-07T14:35:26
|
1793658352
|
{
"authors": [
"Shikhar9425",
"narayan954"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8914",
"repo": "narayan954/dummygram",
"url": "https://github.com/narayan954/dummygram/issues/703"
}
|
gharchive/issue
|
[FEATURE] update readme
Description
want to improve "vision" and "status" in readme section
Screenshots
No response
Additional information
No response
you can go ahead!
|
2025-04-01T04:34:51.262628
| 2024-10-14T15:51:04
|
2586442082
|
{
"authors": [
"CarloLepelaars",
"FBruzzesi",
"MarcoGorelli"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8915",
"repo": "narwhals-dev/narwhals",
"url": "https://github.com/narwhals-dev/narwhals/pull/1173"
}
|
gharchive/pull-request
|
feat: skew
This PR adds skew to Narwhals. Support is added for Polars, Pandas-like, Arrow and Dask.
Checklist
[x] Code follows style guide (ruff)
[x] Tests added
[x] Documented the changes
Thanks for the kind words! Doctest should be fixed now.
btw, if you wanted to just fix a typo somewhere in a separate pr (or, say, take https://github.com/narwhals-dev/narwhals/issues/1170), then once you're already a contributor, CI will always run automatically without me having to approve and run - just bringing this up in case it makes it easier for you
Hey @CarloLepelaars, thanks for the PR!
I left a few comments - the main challenge seems to be how different implementations are between pandas and polars native methods. However polars provide the formula it uses for the computation. It should be possible to reproduce that with native methods or using the series/expr methods that are already implemented in narwhals :)
This is indeed challenging @FBruzzesi. Currently I've made it so every backend returns the biased population skewness, but we can potentially include an option for the unbiased skewness.
Should we account for the len(ser) < 3 case and return 0?
Let's see, this is where Pandas diverges from the rest. To make it consistent we should only handle the case where len(data)==2. In that case Pandas and PyArrow can return 0. Do you also think that is the way to go?
Skewness for 2 elements:
SciPy: 0.000000
Pandas: nan
Polars: 0.000000
Skewness for 1 element:
SciPy: nan
Pandas: nan
Polars: nan
It may be worth checking that the numbers are same even when nulls are present
Good one! Can add a case in unary_test.py that has nulls.
Let's see, this is where Pandas diverges from the rest. To make it consistent we should only handle the case where len(data)==2. In that case Pandas and PyArrow can return 0. Do you also think that is the way to go?
Yes, we are trying to stick with polars api and results, so let's manually force that if needed!
Good one! Can add a case in unary_test.py that has nulls.
That would be great - if it is too much though, we can also make it in a follow up PR
I've covered the cases as discussed and made them consistent with Polars behavior. unary_test.py now also covers data with nan and cases where there are less than 3 rows.
Thanks for addressing the cases, the CI failure seems unrelated.
However I am still quite unsure that we are matching polars behavior. When counting number of elements for the base cases, we should ignore null values, then (pseudo code):
if n_not_nulls==0:
return None # same as pl.Series([]).skew() and pl.Series([None]).skew()
elif n_not_nulls==1:
return float("nan") # same as pl.Series([1]).skew() and pl.Series([1, None]).skew()
elif n_not_nulls==2:
return 0.0 # same as pl.Series([1, 2]).skew() and pl.Series([1, 2, None]).skew()
else:
return <compute_skew>
Implemented your suggestions for nan policy. There is only one edge case left for Dask, where it outputs nan instead of 0.0 with 2 non null elements. Not sure how to adjust _dask/expr.py to account for that.
Hey @CarloLepelaars, thanks for adjusting. CI is failing because in #1224 , compare_dicts was renamed to assert_equal_data.
Implemented your suggestions for nan policy. There is only one edge case left for Dask, where it outputs nan instead of 0.0 with 2 non null elements. Not sure how to adjust _dask/expr.py to account for that.
Regarding dask, I am not able to try it now, bif could definitly be a tricky one to get right! I am ok with marking it as xfail in tests for now
thanks both! should be good, will do another check but this should make it into the next release
Awesome, thank you both for working with me on this! Interesting trick to match the warning to Dask only.
just pushed a fix as the else part of m3 / (m2**1.5) if m2 != 0 else 0 wasn't right nor tested
will merge on green and this can enter the next release 🥦
|
2025-04-01T04:34:51.288253
| 2024-10-03T22:50:57
|
2565120405
|
{
"authors": [
"stephiescastle"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8916",
"repo": "nasa-jpl/explorer-1",
"url": "https://github.com/nasa-jpl/explorer-1/pull/650"
}
|
gharchive/pull-request
|
feature: add support for secondary nav on EDU Collections template
Checklist
[x] Include a description of your pull request and instructions for the reviewer to verify your work.
[x] Link to the issue if this PR is issue-specific.
[x] Create/update the corresponding story if this includes a UI component.
[ ] Create/update documentation. If not included, tell us why.
[x] List the environments / browsers in which you tested your changes.
[x] Tests, linting, or other required checks are passing.
[x] PR has an informative and human-readable title
PR titles are used to generate the change log in releases; good ones make that easier to scan.
PRs will be broadly categorized in the change log, but for even easier scanning, consider prefixing with a component name or other useful categorization, e.g., "BaseButton: fix layout bug", or "Storybook: Update dependencies".
[x] PR has been tagged with a SemVer label and a general category label, or skip-changelog.
These tags are used to do the aforementioned broad categorization in the change log and determine what the next release's version number should be.
Release Drafter will attempt to do the category labeling for you! Please double-check its work.
Description
Addresses feedback:
https://github.com/nasa-jpl/www/issues/617
Changes
Adds support for secondary nav on the EDU Collections template
Adds additional stories to demonstrate the use-case with different hero configurations
Instructions to test
make vue-storybook
View stories that pair secondary nav with:
Hero: http://localhost:6006/?path=/story/templates-edu-pageeducollectionsdetail--secondary-nav&globals=theme:ThemeEdu
Inline hero: http://localhost:6006/?path=/story/templates-edu-pageeducollectionsdetail--secondary-nav-hero-inline&globals=theme:ThemeEdu
No hero: http://localhost:6006/?path=/story/templates-edu-pageeducollectionsdetail--secondary-nav-no-hero&globals=theme:ThemeEdu
Tested in the following environments/browsers:
Operating System
[x] macOS
[ ] iOS
[ ] iPadOS
[ ] Windows
Browser
[x] Chrome
[x] Firefox ESR
[ ] Firefox
[ ] Safari
[ ] Edge
Note that https://github.com/nasa-jpl/www/issues/616 will need to be completed for the secondary nav to actually be used in the CMS
|
2025-04-01T04:34:51.292487
| 2022-01-18T20:39:05
|
1107341804
|
{
"authors": [
"astrogeco",
"skliper"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8917",
"repo": "nasa/cFE",
"url": "https://github.com/nasa/cFE/pull/2027"
}
|
gharchive/pull-request
|
Fix #2026, CFE_FS_ParseInputFileNameEx avoid uninit var
Describe the contribution
Fix #2026
Uses input buffer if default is NULL to set output to avoid uninitialized variables in unit test
Testing performed
Build/run unit tests with valgrind, eliminates warning
Expected behavior changes
Unit test only
System(s) tested on
Hardware: i5/Docker
OS: Ubuntu 18.04
Versions: Bundle main + this commit
Additional context
I wonder if this would have been caught sooner if we had a different bucket for "critical" static analysis checks.
Third party code
None
Contributor Info - All information REQUIRED for consideration of pull request
Jacob Hageman - NASA/GSFC
CCB:2022-01-19 APPROVED
|
2025-04-01T04:34:51.295844
| 2020-01-21T23:39:26
|
553214070
|
{
"authors": [
"astrogeco",
"skliper"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8918",
"repo": "nasa/cFS",
"url": "https://github.com/nasa/cFS/pull/42"
}
|
gharchive/pull-request
|
Integration Candidate 20200121
Describe the contribution
Updated all branches to IC-20200121
Testing performed
Current CI passes
Enhanced CI in #40 passes debug/with deprecated elements testing
Expected behavior changes
See submodules
System(s) tested on
CI testing only
Additional context
N/A
Code contibutions
The cFS repository is provided to bundle the cFS Framework. It is utilized for bundling submodules, continuous integration testing, and version management and does not contain any software. Code contributions should be directed to the appropriate submodule.
Contributor Info - All information REQUIRED for consideration of pull request
Jacob Hageman - NASA/GSFC
Changes directly applied to master. Closing PR
|
2025-04-01T04:34:51.364991
| 2022-03-30T21:52:55
|
1187032474
|
{
"authors": [
"AdrienVigne",
"natelindev"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8919",
"repo": "natelindev/tsdav",
"url": "https://github.com/natelindev/tsdav/issues/109"
}
|
gharchive/issue
|
How to create an adressBook with Davclient ?
Hello, I am searching for a solution to create an addressbook thanks to the davclient client and sync it with the server.
Could you help me please ?
Thanks
Just use makeCollection method with correct params and you should be good.
|
2025-04-01T04:34:51.372263
| 2022-07-25T17:23:14
|
1317141137
|
{
"authors": [
"Dimah4x",
"Yehuda-Edelstein",
"burak40"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8920",
"repo": "nathan-149/automated_youtube_channel",
"url": "https://github.com/nathan-149/automated_youtube_channel/issues/44"
}
|
gharchive/issue
|
Skipping all accounts
Facing this:
<instaloader.nodeiterator.NodeIterator object at 0x000001B5B8213A90>
Scraping From Account: funny_videos
Skipped acc funny_videosbecause of
user not found: 'funny_videos'
Scraping From Account: funny.homevideos
Skipped acc funny.homevideosbecause of
user not found: 'funny.homevideos'
Scraping From Account: try_not_to_laugh_funny_videos
Skipped acc try_not_to_laugh_funny_videosbecause of
user not found: 'try_not_to_laugh_funny_videos'
Scraping From Account: cattoseverywhere
Skipped acc cattoseverywherebecause of
user not found: 'cattoseverywhere'
Scraping From Account: cattobrat
Skipped acc cattobratbecause of
user not found: 'cattobrat'
Scraping From Account: memerzvidz
Skipped acc memerzvidzbecause of
user not found: 'memerzvidz'
Scraping From Account: fuckjerry.tv
Skipped acc fuckjerry.tvbecause of
user not found: 'fuckjerry.tv'
Scraped Videos!
Making Compilation...
Total Length: 0
Traceback (most recent call last):
File "C:\Users\Gebruiker\Desktop\automated_youtube_channel-master\main.py", line 144, in
attemptRoutine()
File "C:\Users\Gebruiker\Desktop\automated_youtube_channel-master\main.py", line 135, in attemptRoutine
routine()
File "C:\Users\Gebruiker\Desktop\automated_youtube_channel-master\main.py", line 100, in routine
makeCompilation(path = videoDirectory,
File "C:\Users\Gebruiker\Desktop\automated_youtube_channel-master\make_compilation.py", line 94, in makeCompilation
finalClip = concatenate_videoclips(videos, method="compose")
File "C:\Users\Gebruiker\AppData\Local\Programs\Python\Python39\lib\site-packages\moviepy\video\compositing\concatenate.py", line 75, in concatenate_videoclips
w = max(r[0] for r in sizes)
ValueError: max() arg is an empty sequence
C:\Users\Gebruiker\Desktop\automated_youtube_channel-master>
what instruction did you run?
I ran python3 main.py and got that same error
|
2025-04-01T04:34:51.379373
| 2024-10-15T23:54:33
|
2590198467
|
{
"authors": [
"nathanjmcdougall"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8921",
"repo": "nathanjmcdougall/usethis-python",
"url": "https://github.com/nathanjmcdougall/usethis-python/issues/26"
}
|
gharchive/issue
|
Set up a release pipeline
Motivation
We will want to release at some point. I'm keen to do it sooner rather than later, even if the repo is still private, because I'd like to start "eating my own dogfood", but the tool needs to run from a separate/isolate venv to avoid uninstalling itself while running. We can use uv tool run for this purpose but it looks like it only pulls from PyPI
Summary of desired feature
An automated release pipeline which publishes to PyPI.
Have made the repo public but still need to sort the release pipeline at some point. I've decided that there are probably a few pre-requisites before tackling that. Nothing to stop some manual releases before that point, although there are some steps which need to be taken first, e.g. setting up project information in pyproject.toml.
|
2025-04-01T04:34:51.395214
| 2023-08-30T11:27:32
|
1873478439
|
{
"authors": [
"dragon-dxw"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8922",
"repo": "nationalarchives/ds-caselaw-custom-api-client",
"url": "https://github.com/nationalarchives/ds-caselaw-custom-api-client/pull/363"
}
|
gharchive/pull-request
|
Add default user agent string to custom api client
Set the default user agent string to give the version of the custom api client, and have it overriddable by users of the client.
Add user-agent to APIClient
|
2025-04-01T04:34:51.398508
| 2022-09-19T08:59:52
|
1377630759
|
{
"authors": [
"alexfromk",
"dragon-dxw"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8923",
"repo": "nationalarchives/ds-caselaw-public-access-service",
"url": "https://github.com/nationalarchives/ds-caselaw-public-access-service/issues/73"
}
|
gharchive/issue
|
Atom API. Bug with parameter page greater than 2024
Hello,
we are trying to work with National Archives API, with service Atom.
We got the data by requests like this:
https://caselaw.nationalarchives.gov.uk/atom.xml?page=1&order=-date
All work fine with parameter "page" since 1 to 2024, but when page became greater 2024, service return HTTP 504.
For example,
GET https://caselaw.nationalarchives.gov.uk/atom.xml?page=2025&order=-date
response with http 504.
If it is possible, to fix this bug.
With best regards. Alex Kononenko
We had in fact scheduled today to look at that, and we've got a fix on staging, which should go live some day soon. You should also note significantly faster load times for the atom feed pages.
That should be fixed on the live site now. Thank you for letting us know about the problem -- whilst we did know about it, it gives us confidence that we're working on fixing things that people care about!
Hi, thank you very much.
I tested it, paging above 2024 work fine.
Alex Kononenko
|
2025-04-01T04:34:51.443702
| 2019-06-22T13:21:08
|
459473734
|
{
"authors": [
"gcolliso",
"morrisio"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8924",
"repo": "nats-io/docs",
"url": "https://github.com/nats-io/docs/pull/34"
}
|
gharchive/pull-request
|
Fix internal links
Uppercase internal links don't work when clicked in the 'Protocol Messages' tables of Client Protocol and Cluster Protocol pages. Changing links to lowercase (e.g. #PUB -> #pub) should make them clickable again.
Note: This problem does not occur on github.com, only when the files are rendered on github.io.
Also removing a number HTML anchors which should be made obsolete by this PR and adding a number of missing internal links.
lgtm - @morrisio thanks for the fix!
|
2025-04-01T04:34:51.446049
| 2019-04-12T18:33:16
|
432692630
|
{
"authors": [
"JensRantil",
"derekcollison"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8925",
"repo": "nats-io/gnatsd",
"url": "https://github.com/nats-io/gnatsd/issues/946"
}
|
gharchive/issue
|
Weird instructions to build
https://github.com/nats-io/gnatsd/#build states
Run go build inside the /nats-io/gnatsd directory.
. However, there is no /nats-io/gnatsd in the repo. I assume that /main.go contains the main function so, I propose simply writing
Run go build in root of the checked out repository.
.
It means the top of the repo directory, we will clarify the docs..
/cc @gcolliso
|
2025-04-01T04:34:51.462429
| 2015-09-26T14:11:03
|
108468212
|
{
"authors": [
"derekcollison",
"ronag"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8926",
"repo": "nats-io/node-nats",
"url": "https://github.com/nats-io/node-nats/issues/38"
}
|
gharchive/issue
|
Document callback signature in subscribe
I keep going back to subscribe to remind myself what the signature of the callback should be and I keep missing it...
Good idea.
|
2025-04-01T04:34:51.480034
| 2022-01-13T17:22:02
|
1102066343
|
{
"authors": [
"dutchshark",
"lampwins",
"nniehoff"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8927",
"repo": "nautobot/nautobot",
"url": "https://github.com/nautobot/nautobot/issues/1269"
}
|
gharchive/issue
|
Add BGP Support to Nautobot
Environment
Python version: 3.7
Nautobot version: 1.2.3
Proposed Functionality
I would like to model my BGP deployment in Nautobot. This should include ASNs as well as peering relationships at a minimum.
Use Case
As Nelly, a network engineer, I need to be able to model BGP configuration of my network in my source of truth to be able to completely automate the deployment of my network configuration.
Database Changes
Definitely, probably at least ASNs perhaps more. It's possible some of this may be better represented in a plugin.
External Dependencies
None
Perhaps use Peering Manager to prevent reinventing the wheel?
Closing with https://github.com/nautobot/nautobot-plugin-bgp-models
|
2025-04-01T04:34:51.485290
| 2021-05-31T04:48:28
|
907044458
|
{
"authors": [
"glennmatthews",
"waydh"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8928",
"repo": "nautobot/nautobot",
"url": "https://github.com/nautobot/nautobot/issues/512"
}
|
gharchive/issue
|
Cloning Device or IP with a tag applied results in a Server Error
Environment
Ubuntu 20.04
Python version: 3.8.5
Nautobot version: 1.0.2
Steps to Reproduce
Install NautoBot 1.0.2 per the directions on https://nautobot.readthedocs.io
Create a Site, Manufacturer, Device Role, Tag, and then a Device with all applied
Click clone on the created device and receive a Server Error
Remove the tag from the device (so there are no tags at all) click Clone and it works
Issue also occurs with IPs
Expected Behavior
Clone button works without error allowing the device to be cloned
Observed Behavior
Server Error like the following - the invalid UUID is not always 'a', it changes alphabet or number
Server Error
There was a problem with your request. Please contact an administrator.
The complete exception is provided below:
<class 'django.core.exceptions.ValidationError'>
['“a” is not a valid UUID.']
Python version: 3.8.5
Nautobot version: 1.0.2
If further assistance is required, please post to the #nautobot channel on NetworkToCode's Slack.
Thanks for the report! This may be related to #499.
|
2025-04-01T04:34:51.492733
| 2023-08-09T15:22:11
|
1843503222
|
{
"authors": [
"HanlinMiao",
"glennmatthews"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8929",
"repo": "nautobot/nautobot",
"url": "https://github.com/nautobot/nautobot/pull/4233"
}
|
gharchive/pull-request
|
Refine New UI Device Detail View
Closes: #4013
What's Changed
TODO
[x] Re-establishing the advanced tab into the base page layout (Metadata: Key-value panel, fields in legacy UI (created, updated, UUID))
[x] Finish actions menu (expands to an Edit link to the legacy UI for the time being)
[x] Device: Key-value panel, fields in legacy UI
[x] Management: Key-value panel, fields in the legacy UI
[x] Explanation of Change(s)
[x] Added change log fragment(s) (for more information see the documentation)
[x] Fixed incorrect related model url rendering in ObjectRetrieve view.
[ ] Attached Screenshots, Payload Example
[ ] Unit, Integration Tests
[ ] Documentation Updates (when adding/changing features)
[ ] Example Plugin Updates (when adding/changing features)
[ ] Outline Remaining Work, Constraints from Design
Looking at the first screenshot - definitely not blocking, but can we make the tags list wrap to a new line instead of scrolling horizontally when there are a ton of tags?
Looking at the advanced tab, can we fix capitalization of "Id" and "Url"?
Will make a follow-up story for more flexible rendering
Looking at the advanced tab, can we fix capitalization of "Id" and "Url"?
https://github.com/nautobot/nautobot/issues/4250
Looking at the first screenshot - definitely not blocking, but can we make the tags list wrap to a new line instead of scrolling horizontally when there are a ton of tags?
https://github.com/nautobot/nautobot/issues/4251
|
2025-04-01T04:34:51.537316
| 2019-07-04T07:44:39
|
464115506
|
{
"authors": [
"donggyu04",
"junoyoon"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8939",
"repo": "naver/ngrinder",
"url": "https://github.com/naver/ngrinder/pull/425"
}
|
gharchive/pull-request
|
Fix failed tests
[Merge into master]
Use network interface before monitor collection test.
Please use hotfix/** branch.
Move to #430
|
2025-04-01T04:34:51.549589
| 2019-05-08T07:25:47
|
441589938
|
{
"authors": [
"RoySRose",
"asahu4"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8940",
"repo": "naver/pinpoint",
"url": "https://github.com/naver/pinpoint/issues/5574"
}
|
gharchive/issue
|
Please confirm you offer the enterprise support for pinpoint.
Please confirm you offer the enterprise support for pinpoint. If yes please share more detail on support and charges .
Thanks, Ajay
Hello, @asahu4
I'm afraid we don't have any enterprise support business model, yet.
|
2025-04-01T04:34:51.551125
| 2020-02-17T08:32:50
|
566136540
|
{
"authors": [
"koo-taejin"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8941",
"repo": "naver/pinpoint",
"url": "https://github.com/naver/pinpoint/issues/6507"
}
|
gharchive/issue
|
Fix problem that do not set some fields when an exception occurs
related module
Kafka-Plugin
description
If service type has record statistics, destnationId value have to send for to draw servermap.
But, current ProducerSendInterceptor code do not set destinationId field when exception occurred while send() method invoked.
For this reason, servermap has not drawn when when exception occurred while send() method invoked.
This issue expect to have resolved via #6508, #6510
|
2025-04-01T04:34:51.565709
| 2019-05-02T06:06:10
|
439438295
|
{
"authors": [
"codecov-io",
"koo-taejin"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8942",
"repo": "naver/pinpoint",
"url": "https://github.com/naver/pinpoint/pull/5550"
}
|
gharchive/pull-request
|
[#4558] Add Stream function in profiler module using grpc
Add Stream function in profiler module using grpc.
Codecov Report
Merging #5550 into master will decrease coverage by 0.07%.
The diff coverage is 0%.
@@ Coverage Diff @@
## master #5550 +/- ##
==========================================
- Coverage 40.8% 40.72% -0.08%
==========================================
Files 2822 2826 +4
Lines 84907 85049 +142
Branches 11314 11328 +14
==========================================
- Hits 34644 34640 -4
- Misses 47239 47386 +147
+ Partials 3024 3023 -1
Impacted Files
Coverage Δ
...oint/profiler/receiver/grpc/GrpcStreamService.java
0% <0%> (ø)
.../receiver/grpc/PinpointClientResponseObserver.java
0% <0%> (ø)
...er/receiver/grpc/GrpcActiveThreadCountService.java
0% <0%> (ø)
...xt/thrift/CommandThriftToGrpcMessageConverter.java
61.53% <0%> (-38.47%)
:arrow_down:
...r/receiver/grpc/ActiveThreadCountStreamSocket.java
0% <0%> (ø)
...xt/thrift/CommandGrpcToThriftMessageConverter.java
31.25% <0%> (-40.18%)
:arrow_down:
.../profiler/receiver/grpc/GrpcCommandDispatcher.java
0% <0%> (ø)
:arrow_up:
...navercorp/pinpoint/rpc/common/SocketStateCode.java
82.92% <0%> (-2.44%)
:arrow_down:
...om/navercorp/pinpoint/rpc/codec/PacketDecoder.java
57.44% <0%> (-2.13%)
:arrow_down:
...point/rpc/client/DefaultPinpointClientHandler.java
71.24% <0%> (-1.64%)
:arrow_down:
... and 8 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 9e07e76...17bcff3. Read the comment docs.
|
2025-04-01T04:34:51.634232
| 2015-11-04T04:40:26
|
114968156
|
{
"authors": [
"melograph85",
"mlogan"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8961",
"repo": "ncb000gt/node.bcrypt.js",
"url": "https://github.com/ncb000gt/node.bcrypt.js/issues/369"
}
|
gharchive/issue
|
Consider XOR instead of equals for the comparator
The guide Salted Password Hashing - Doing it Right claims that XOR is better than == for a slow-equals comparison function to mitigate timing attacks:
The reason we need to use XOR instead of the "==" operator to compare integers is that "==" is usually translated/compiled/interpreted as a branch.... The branching makes the code execute in a different amount of time depending on the equality of the integers and the CPU's internal branch prediction state.
This module's implementation currently uses the != operator. Indeed, clang compiles this to code that branches, which can result in variable-time execution:
## ../src/bcrypt_node.cc:240
next:
cmp ebx, esi
jl continue
cmp eax, esi
jl continue
movzx edi, byte ptr [r15 + rsi]
movzx ecx, byte ptr [rbp + rsi - 176]
cmp ecx, edi
je continue
xor r8d, r8d
continue:
inc rsi
cmp esi, edx
jl next
(The above is edited output from running node-gyp's build command on bcrypt_node.cc with the additional -S -masm=intel options.)
The page mentioned has more information. I figured this might be worth bringing up.
That page has a lot of bad advice, such as using SHA256 for hashing passwords. (See: https://www.nccgroup.trust/us/about-us/newsroom-and-events/blog/2007/july/enough-with-the-rainbow-tables-what-you-need-to-know-about-secure-password-schemes/)
The particular suggestion of using timing attacks against password hashes is just bogus. Timing attacks work against HMACs because you can send anything you want as the signature. Once you guess the first byte, you hold that constant and work on the second byte. But you can't hold the first byte of the output of a cryptographic hash constant while probing for the second byte. If you could, you could just reverse the hash, and break most or all of modern crypto. (See: http://security.stackexchange.com/questions/9192/timing-attacks-on-password-hashes)
|
2025-04-01T04:34:51.656312
| 2018-10-17T16:35:20
|
371166039
|
{
"authors": [
"joe-graham",
"saidelike"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8962",
"repo": "nccgroup/asafw",
"url": "https://github.com/nccgroup/asafw/issues/5"
}
|
gharchive/issue
|
Serial debugging failing to start
This issue was found on a physical ASA 5505 running version 9.1(6).
When enabling gdbserver on the serial interface in asafw, the gdb script generated by asadbg hangs at target extended-remote [serial port].
Turning on gdb's remote debugging shows that gdbserver continually prints this string, preventing gdb from attaching:
It looks like a fix for this would be modifying the inittab file to run /tmp/start_cmd on a different tty. This will still cause gdbserver to attach to the serial interface, but will result in the error message being printed on the other tty as opposed to over the serial interface. Manually making this change on my end results in asadbg hanging (since it's waiting for the "Remote debugging over /dev/ttyS0" string) but eventually connecting over USB:
I'm willing to write the code to implement this change, but I'll hold off on implementing and making a pull request since this is a little more involved than my previous two issues.
Thanks for this.
Good finding about modifying /etc/inittab. So you effectively change: ttyS0::once:/tmp/run_cmd into ttyS1::once:/tmp/run_cmd? Automation could quite easily be tested by adding code to unpack_repack.sh in the custom function and invoking unpack_repack_bin.sh -c .... Once this works we could add a special case in the enable_gdb function for the firmware we know require that (e.g. "asa91"). Afaict most of the 9.1.x do require that.
I think if we can find a way to have gdb not hang at all, it will be even better. Not sure if we can tell gdb to not wait for the "Remote debugging using /dev/ttyS0" string. Tbh I didn't even know it worked like that.
What is interesting is that for asa924 for instance, we have this fix (note the comment which is not on asa91x):
# run on the console specified on the kernel start line
::once:/tmp/run_cmd
I am wondering if just changing to this would solve the gdb hanging problem.
Yes, I agree it involves a bit more changes that your previous PR but at the same time is a good exercise to be able to modify the scripts (and not too complex in this case imho) so if that is something that you are interested in, feel free to do so.
Note: In the past, we would use --bin-with-asa-to-inject to e.g. inject the /asa folder from 9.1.x into another firmware such as asa92*.bin but your way is a lot cleaner and does not change the firmware container.
Oh right thanks for explaining that after your fix, the hang is on the asadbg side, not gdb side. The reason there is a hang is because we timeout after 3 minutes in case we don't detect the "Remote debugging using /dev/ttyS0" string which indeed we don't detect following your fix. We should be able to work around that but not sure how yet either.
Good idea about the characters gdbserver may send over serial. I guess we could hexdump what we receive in comm.py instead of just printing ascii characters, to see if there is anything interesting.
In case there isn't any character send, it may be possible to check from comm.py if there is a listening gdbserver on the other end by sending special characters. So in case we read 0 bytes, we could send this special series of bytes from comm.py and check we get the right reply from the other end (gdbserver).
Definitely worth looking into. I'll get started on implementing the change for unpack_repack_bin.sh this weekend, and open an issue for the comm.py stuff over on the asadbg repo after I do a little more research into the gdbserver stuff. Thanks for your insight :)
|
2025-04-01T04:34:51.663488
| 2021-03-15T22:57:16
|
832264091
|
{
"authors": [
"nicl-nno",
"pep8speaks"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8963",
"repo": "nccr-itmo/FEDOT",
"url": "https://github.com/nccr-itmo/FEDOT/pull/252"
}
|
gharchive/pull-request
|
Better implementation of the caching
cache moved from memory to disk using shelve library
caching removed from nodes, now it is external object that can be used to 'fit' the nodes from pre-saved results
Hello @nicl-nno! Thanks for opening this PR. We checked the lines you've touched for PEP 8 issues, and found:
In the file fedot/core/composer/gp_composer/gp_composer.py:
Line 21:121: E501 line too long (121 > 120 characters)
In the file fedot/core/composer/optimisers/regularization.py:
Line 52:32: E713 test for membership should be 'not in'
|
2025-04-01T04:34:51.667106
| 2016-09-09T12:33:25
|
175998431
|
{
"authors": [
"nchammas",
"serialx"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8964",
"repo": "nchammas/flintrock",
"url": "https://github.com/nchammas/flintrock/pull/147"
}
|
gharchive/pull-request
|
Fix assume_yes not existing in _create_instances
I've encountered an error while doing test:
$ python3 -m flintrock --config config.yaml launch serialx-test
Requesting 3 spot instances at a max price of $0.311...
0 of 3 instances granted. Waiting...
0 of 3 instances granted. Waiting...
0 of 3 instances granted. Waiting...
0 of 3 instances granted. Waiting...
0 of 3 instances granted. Waiting...
0 of 3 instances granted. Waiting...
^CThere was a problem with the launch. Cleaning up...
Canceling spot instance requests...
Traceback (most recent call last):
File "/Users/serialx/workspace/flintrock/flintrock/ec2.py", line 685, in _create_instances
time.sleep(30)
KeyboardInterrupt
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/Cellar/python3/3.5.1/Frameworks/Python.framework/Versions/3.5/lib/python3.5/runpy.py", line 170, in _run_module_as_main
"__main__", mod_spec)
File "/usr/local/Cellar/python3/3.5.1/Frameworks/Python.framework/Versions/3.5/lib/python3.5/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/Users/serialx/workspace/flintrock/flintrock/__main__.py", line 8, in <module>
sys.exit(main())
File "/Users/serialx/workspace/flintrock/flintrock/flintrock.py", line 1034, in main
cli(obj={})
File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/core.py", line 716, in __call__
return self.main(*args, **kwargs)
File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/core.py", line 696, in main
rv = self.invoke(ctx)
File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/core.py", line 1060, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/core.py", line 889, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/core.py", line 534, in invoke
return callback(*args, **kwargs)
File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/decorators.py", line 17, in new_func
return f(get_current_context(), *args, **kwargs)
File "/Users/serialx/workspace/flintrock/flintrock/flintrock.py", line 339, in launch
instance_initiated_shutdown_behavior=ec2_instance_initiated_shutdown_behavior)
File "/Users/serialx/workspace/flintrock/flintrock/ec2.py", line 46, in wrapper
res = func(*args, **kwargs)
File "/Users/serialx/workspace/flintrock/flintrock/ec2.py", line 878, in launch
instance_initiated_shutdown_behavior=instance_initiated_shutdown_behavior)
File "/Users/serialx/workspace/flintrock/flintrock/ec2.py", line 759, in _create_instances
if not assume_yes:
NameError: name 'assume_yes' is not defined
This is an attempt to fix it.
Looks good to me. @serialx - Did you test this to confirm that the behavior is as expected for both --assume-yes and --no-assume-yes?
@nchammas To my knowledge, the code should work as expected. The codepath that triggers this bug is quite hard to reproduce, so this fix PR is not directly tested in that situation. Is there any place that you might want me to take extra testing?
Ah sorry, you're right. This comes up only when creating new instances fails for whatever reason.
Merging this in!
|
2025-04-01T04:34:51.676158
| 2018-12-03T10:04:00
|
386741300
|
{
"authors": [
"ArthurGing",
"ncw"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8965",
"repo": "ncw/rclone",
"url": "https://github.com/ncw/rclone/issues/2812"
}
|
gharchive/issue
|
Failed to create file system
What is the problem you are having with rclone?
Is there any way to refresh token of onedrive?
I can't do anything after token expired (it needs 60 minutes)
What is your rclone version (output from rclone version)
1.45
Which OS you are using and how many bits (eg Windows 7, 64 bit)
linux 64 bit
Which cloud storage system are you using? (eg Google Drive)
onedrive
The command you were trying to run (eg rclone copy /tmp remote:tmp)
lsl, copy, move
A log from the command with the -vv flag (eg output from rclone -vv copy /tmp remote:tmp)
Failed to create file system for "arthur:": failed to get root: InvalidAuthenticationToken: CompactToken validation failed with reason code: 80049228.
Then you wait > 1 hour and rclone gives you CompactToken validation failed with reason code: 80049228.
Is that right?
Feel free to XXX out anything that you think looks secret.
What kind of onedrive are you using? Personal, Business, Sharepoint, other?
If you do rclone config show arthur you should see your token, It should look something like this (it will all be one one line and the XXX are very long. Can you make sure it has a refresh_token in there?
token = {
"access_token":"XXX",
"token_type":"Bearer",
"refresh_token":"XXX",
"expiry":"2018-11-29T16:10:46.368783193Z"
}
Thank you for the prompt reply.
I guess Uploading files using 'copyto' may call api of cloud, so I am trying to get response with '--dump responses', but it show nothing. Is there any way to get response?
Best regards
I guess Uploading files using 'copyto' may call api of cloud, so I am trying to get response with '--dump responses', but it show nothing. Is there any way to get response?
When I try that I see the response. Don't forget -vv!
2018/12/05 10:51:57 DEBUG : HTTP RESPONSE (req 0xc0004aeb00)
2018/12/05 10:51:57 DEBUG : HTTP/1.1 201 Created
Content-Type: application/json; charset=utf-8
Date: Wed, 05 Dec 2018 10:51:57 GMT
P3p: CP="BUS CUR CONo FIN IVDo ONL OUR PHY SAMo TELo"
Strict-Transport-Security: max-age=31536000; includeSubDomains
Www-Authenticate: Bearer realm="OneDriveAPI", error="invalid_token", error_description="Invalid auth token"
X-Asmversion: UNKNOWN; 19.161.1128.2006
X-Msedge-Ref: Ref A: 48E0A333677C455EA636FA9904C82087 Ref B: LON21EDGE1014 Ref C: 2018-12-05T10:51:57Z
X-Msnserver: AM3PPF2A98B041F
{"createdBy":{"application":{"displayName":"rclone","id":"48211038"},"user":{"id":...
Thank you for your help.
I see the response after adding -vv, thanks a lot.
Great! I think we are done here so I'll close this issue.
|
2025-04-01T04:34:51.682097
| 2019-02-28T18:23:27
|
415760423
|
{
"authors": [
"brprice",
"ndmitchell"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8966",
"repo": "ndmitchell/ghcid",
"url": "https://github.com/ndmitchell/ghcid/issues/226"
}
|
gharchive/issue
|
Polling ignores requested frequency
The polling logic seems odd: we wait for the requested time only on the first iteration (in waitFiles) after each reload, and each subsequent one (in recheck) it waits a fixed 0.1.
I suspect that we should sleep in WaiterPoll branch of recheck's case.
I wonder if this is related to #182?
#182 wasn't using polling, so shouldn't be related. You would also have seen the 0.1s blips in the CPU graph if it was sleeping.
Reading the code carefully I think you're spot on though. There are two WaiterPoll branches, and I think they're the wrong way round 🤦♂️ . The corresponding WaiterNotify branch is first immediate, and then sleeping, so now I made WaiterPoll match that.
|
2025-04-01T04:34:51.699008
| 2018-06-18T08:29:59
|
333176356
|
{
"authors": [
"Steppschuh",
"codecov-io"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8967",
"repo": "neXenio/BLE-Indoor-Positioning",
"url": "https://github.com/neXenio/BLE-Indoor-Positioning/pull/90"
}
|
gharchive/pull-request
|
0075 render floor plan
WIP
Codecov Report
Merging #90 into dev will decrease coverage by 0.02%.
The diff coverage is 0%.
@@ Coverage Diff @@
## dev #90 +/- ##
============================================
- Coverage 35.28% 35.26% -0.03%
Complexity 204 204
============================================
Files 39 39
Lines 1278 1279 +1
Branches 123 123
============================================
Hits 451 451
- Misses 775 776 +1
Partials 52 52
Impacted Files
Coverage Δ
Complexity Δ
...io/bleindoorpositioning/location/LocationUtil.java
78.12% <0%> (-2.53%)
7 <0> (ø)
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update e6c387d...f98e9c9. Read the comment docs.
Implementation works, but still has issues (probably related to padding in the CanvasProjection)
|
2025-04-01T04:34:51.704592
| 2017-05-04T16:19:14
|
226335258
|
{
"authors": [
"JunjieW",
"nea"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8968",
"repo": "nea/MarkdownViewerPlusPlus",
"url": "https://github.com/nea/MarkdownViewerPlusPlus/issues/32"
}
|
gharchive/issue
|
CSS color is not rendered
Hi,
I was trying to add some customized CSS, but it wasn't rendered well. I'm not sure if it's caused by the HTMLRenderer or in this plugin. The exported html can be rendered as expected in Chrome, IE, Firefox.
body {
font-family: "Source Sans Pro",sans-serif;
font-size: 13px;
color: #3f3f3f
}
h1 { font-size: 25px; }
h2 { font-size: 23px; }
h3 { font-size: 21px; }
h4 { font-size: 19px; }
h5 { font-size: 16px; }
h1, h2, h3, h4, h5, h6 {
font-family: inherit;
}
blockquote {
background: #f9f9f9;
border-left: 10px solid #ccc;
padding-top: 5px;
padding-bottom: 5px;
quotes: "\201C""\201D""\2018""\2019";
}
code {
font-family: Monaco, Consolas, "Andale Mono", "DejaVu Sans Mono", monospace;
font-size: 95%;
white-space: pre;
white-space: pre-wrap;
white-space: -moz-pre-wrap;
white-space: -o-pre-wrap;
background: rgb(200,200,200);
display: inline-block;
}
pre code {
font-family: Monaco, Consolas, "Andale Mono", "DejaVu Sans Mono", monospace;
font-size: 95%;
line-height: 140%;
white-space: pre;
white-space: pre-wrap;
white-space: -moz-pre-wrap;
white-space: -o-pre-wrap;
background: #faf8f0;
display: block;
padding: 0.5em 1em;
border: 1px solid #bebab0;
}
BTW there is another html rendering solution provided by this repo,
I was trying to improve that plugin then found yours, while that one renders HTML without problem since it is using IE core, it doesn't provide functionality for HTML and PDF export, which is a very useful for me : ) ...
Hmm, actually it works for CSS property background-color but not background
Hey @JunjieW
Thanks for the example.
Yeah, I just the IE import originally but wanted to keep everything contained with as little external dependencies as possible, why I changed to the HTML Renderer. It is not the most active development but a very good basic library.
But because of that, many things are probably just not implemented in the lib. I will check on the source and see why background-color works but not background and maybe I can create a pull-request.
Cheers
|
2025-04-01T04:34:51.711576
| 2019-10-15T16:39:14
|
507359694
|
{
"authors": [
"janxter",
"neagle"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8969",
"repo": "neagle/gokibitz",
"url": "https://github.com/neagle/gokibitz/issues/195"
}
|
gharchive/issue
|
Bug kifu listing page hanging
https://gokibitz.com/kifu has been hanging up in Firefox and Chrome on multiple computers for me over the last few days.
Sorry to hear you're having trouble! Could you open your browser's console and paste any error messages you see there? The service is still up and running, and I haven't been able to reproduce an error on my end.
I'm not getting anything in the Chrome console. After an initial "Page Unresponsive", it still reads "Waiting for GoKibitz.com...." an hour later while displaying either the background color or everything but the kifu list.
Firefox prompts to wait or close, although I'll have to check the exact message on that system later.
On one of the systems I've tried this on (Windows 10 Pro), the page loads just fine in Edge (which I've only used for this one purpose).
This only started a couple of days ago and is taking place both at home and my office.
If I ctrl-alt-del, kill Chrome, and then restore, it finally displays both the kifu list and comments, but I'm logged out. I reload, and it logs me back in but the problem returns.
So I killed it, went to the main page, logged out, and was then able to view kifu. Once I log back in, the problem resumes.
I'm user HarveyCrichton and I'm pretty sure the last game I uploaded was https://gokibitz.com/kifu/ByyM3ZeFr
Hm—so you had two SGFs uploaded that were truncated at random points.
(Example: "smog山人: if your opponent makes some mistake you can profit from it :). From move 1: D16 Q4 D4 F17 D14 C17 D17 D18 C16 E18 C18 B17 B18 B16 B15 \nsmog山人: if he connects solid you can still play hane if you want. From move 1: D16 Q4 D4 F17 D14 C17 D17 D18 C16 C18 B17 \nsmog山人: if he blocks you there, you're back to the joseki xd. From move 1: D16 Q4 D4 F17 D")
The malformed SGFs were causing a surprisingly complete meltdown of functions that tried to retrieve and parse them. Obviously this shows that the uploader is insufficiently guarded against invalid SGF files: I'll see if I can address that on my side.
Out of curiosity, were you cutting/pasting the text of your SGFs? Is there any chance the truncation was caused by user error?
(You should be able to log back in now. I've marked the two invalid SGFs as deleted.)
As it happens, I just linked to the URL instead of downloading from OGS and uploading to GK from my computer for the last few games. Maybe some sketchy network issue took place.
I likely hand-edited text in the game after it was in GK, but that seems unlikely to impact the sgf in your db.
Hmmm—maybe the hand-editing thing could be a point of vulnerability. Another thing to investigate. If I ever get the chance to do GoKibitz 2.0, I'm going to store SGFs in the DB in a less fragile way.
Thanks for your excellent debugging; sorry again for the inconvenience. Let me know if you have any other issues!
|
2025-04-01T04:34:51.713485
| 2021-11-12T00:42:04
|
1051474801
|
{
"authors": [
"brianzzzasd",
"dimfeld",
"glennmichael123"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8970",
"repo": "near/borsh-js",
"url": "https://github.com/near/borsh-js/pull/40"
}
|
gharchive/pull-request
|
Change TextDecoder resolution to not rely on global
global does not exist in the browser without Node polyfills, and we shouldn't assume their presence.
Fixes #38
@volovyk-s thank you thank you thank you! you have no idea how merging this fix solved a lot of problems for me 😭
@volovyk-s thank you sir 🙇
|
2025-04-01T04:34:51.741625
| 2022-02-22T11:29:44
|
1146792703
|
{
"authors": [
"ntotten",
"radomird"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8971",
"repo": "nearform/get-jwks",
"url": "https://github.com/nearform/get-jwks/pull/115"
}
|
gharchive/pull-request
|
Added type definitions and tests for types
This PR is a continuation of the work done by @ntotten in this PR: https://github.com/nearform/get-jwks/pull/108
This PR contains all the work from https://github.com/nearform/get-jwks/pull/108 + I've added tests for types.
Closes #109
Hey, thanks for finishing this. I just came back to do this and saw it is already released. Thanks again!
@ntotten no problem 😃
|
2025-04-01T04:34:51.744072
| 2019-11-26T07:08:03
|
528531915
|
{
"authors": [
"AnaisUrlichs",
"icerove"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8972",
"repo": "nearprotocol/near-explorer",
"url": "https://github.com/nearprotocol/near-explorer/issues/131"
}
|
gharchive/issue
|
NearKat blocking the support section
Story
As a user looking at the explorer, and don't get what's going on, I would like to get some help. But "Oh no!" The NearKat is blocking the support section.
Acceptance Criteria
[x] The user can read the support section clearly
[x] NearKat does not reduce accessibility. (It's a malicious NearKat)
fix it in the PR and wait to be merged
|
2025-04-01T04:34:51.748076
| 2019-11-15T08:19:14
|
523321798
|
{
"authors": [
"icerove",
"vgrichina"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8973",
"repo": "nearprotocol/near-shell",
"url": "https://github.com/nearprotocol/near-shell/pull/189"
}
|
gharchive/pull-request
|
Fix tx-status to handle account id better
Fixes https://github.com/nearprotocol/near-shell/issues/187
It is really a goo idea to separate the command when you fix it.
Please also update the readme, I think it is not including the new command you added.
It is really a goo idea to separate the command when you fix it.
Seemed like good opportunity to do this. If command was more complex I'd split the change in 2 separate (first move, then fix).
Please also update the readme, I think it is not including the new command you added.
I didn't add any new command, not sure what do you mean.
near generate-key and near repl is not added. Also, still not clear about accountid problem I left above?
near generate-key and near repl is not added.
these aren't part of this PR
near generate-key and near repl is not added. Also, still not clear about accountid problem I left above?
https://github.com/nearprotocol/near-shell/commit/1c139dd33cb5ec0cd15cbf8e6070efdc21e8d753
|
2025-04-01T04:34:51.766752
| 2023-11-27T02:18:06
|
2011360874
|
{
"authors": [
"aktech",
"dharhas",
"nkaretnikov"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8974",
"repo": "nebari-dev/argo-jupyter-scheduler",
"url": "https://github.com/nebari-dev/argo-jupyter-scheduler/pull/3"
}
|
gharchive/pull-request
|
Generate HTML output and send it to Slack
Reference Issues or PRs
What does this implement/fix?
Put a x in the boxes that apply
[ ] Bug fix (non-breaking change which fixes an issue)
[x] New feature (non-breaking change which adds a feature)
[ ] Breaking change (fix or feature that would cause existing features not to work as expected)
[ ] Documentation Update
[ ] Code style update (formatting, renaming)
[ ] Refactoring (no functional changes, no API changes)
[ ] Build related changes
[ ] Other (please describe):
Testing
[ ] Did you test the pull request locally?
[ ] Did you add new tests?
Documentation
Access-centered content checklist
Text styling
[ ] The content is written with plain language (where relevant).
[ ] If there are headers, they use the proper header tags (with only one level-one header: H1 or # in markdown).
[ ] All links describe where they link to (for example, check the Nebari website).
[ ] This content adheres to the Nebari style guides.
Non-text content
[ ] All content is represented as text (for example, images need alt text, and videos need captions or descriptive transcripts).
[ ] If there are emojis, there are not more than three in a row.
[ ] Don't use flashing GIFs or videos.
[ ] If the content were to be read as plain text, it still makes sense, and no information is missing.
Any other comments?
Summary:
adds a send_to_slack step to scheduled and one-time workflows
it uses the Slack API to send HTML output to a specified Slack channel
added a call to jupyter nbconvert to generate HTML
configured via Parameters SLACK_TOKEN and SLACK_CHANNEL in Notebook Jobs in the web UI, which are accessible via envs in the code
see the Slack API docs on how to configure a bot to send a file to a channel -- this needs to be done first for the bot/sending functionality to work
this new step is integrated with update_job_status_failure, so it will be visible in the UI if it fails
the Slack script also has some printing and additional validation, so an exception will be raised on failure, which will cause the job to fail
cmd_args generation is changed because (1) two commands are now called there and (2) it's passed to /bin/sh as a string anyway, so no point in keeping that in a list
changed *path functions to return Path objects since that's more flexible, in case callers want to modify these paths.
So how will this be configured? i.e. "send to slack" is not a feature that all nebari / jupyter-scheduler users will need. Also someone else might want to send it to mattermost or another rest api. Is there a way to make this a bit more generic.
@dharhas
So how will this be configured? i.e. "send to slack" is not a feature that all nebari / jupyter-scheduler users will need.
Currently, it'll only execute this task if you provide SLACK_TOKEN and SLACK_CHANNEL as Parameters when scheduling the notebook. If you don't provide this, nothing will be sent.
Is there a way to make this a bit more generic.
Technically, we can turn this into "specify a random shell command and I'll execute it", but I don't think it's a good design.
Users might run into issues with string escaping
This prevents us from doing API-specific checking of whether the request was successful or not.
I'd suggest we add support for additional APIs separately, on a case by case basis.
Is there an example (a screenshot maybe) of slack output in a channel or something?
Also, would be nice to add some docs for this.
@nkaretnikov lets add docs also I think we need to make sure runs are timestamped.
Are they also saved to disk as well as sent to slack? "send to slack" needs to be optional.
Is there an example (a screenshot maybe) of slack output in a channel or something?
Slack previews HTML as source code here. I think they don't render it by default for security reasons. I've looked and I'm not sure there's a way to render it. Once you download it, it's valid HTML.
@aktech I've tested and reviewed this. PTAL
@aktech PTAL. Made the changes you suggested, added more info to the internals section of README. Tested to make sure it's working and the backtraces are logged to a file.
I went ahead and merged this since it'd be nice to have as part of the current Nebari release, see https://github.com/nebari-dev/nebari/issues/2195#issuecomment-1886155081.
|
2025-04-01T04:34:51.768778
| 2022-10-05T11:36:22
|
1397651943
|
{
"authors": [
"pavithraes"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8975",
"repo": "nebari-dev/nebari-demo",
"url": "https://github.com/nebari-dev/nebari-demo/issues/5"
}
|
gharchive/issue
|
How will participants log-in to Nebari?
We'd like the participants at PyData NYC to get hands-on time with Nebari. This issue is to track the logistics behind providing participants with login details.
@viniciusdc had noted that we can configure Keycloak to allow random passwords.
Done: https://github.com/nebari-dev/nebari-demo#login-for-the-first-time
|
2025-04-01T04:34:51.772080
| 2024-10-11T14:41:01
|
2581549183
|
{
"authors": [
"MatteoGheza",
"nebojsatomic"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8976",
"repo": "nebojsatomic/IDE21",
"url": "https://github.com/nebojsatomic/IDE21/issues/53"
}
|
gharchive/issue
|
Add language Italian
I'm working on adding the italian translation, to both src/legacy/dev-application/languages/en.php and src/legacy/languages/en.php.
@MatteoGheza Thank you for participation, but at the moment, only work on src/dev-application/languages/creator/it.php by updating the current translation, if some strings are not translated properly,
or add src/dev-application/languages/it.php file, copy and paste the contents of sr.php from the same folder, and translate in that file.
Do not touch src/legacy/dev-application/languages/en.php and src/legacy/languages/en.php.
you have the translation guide already written in repo's Readme.md
Thank you for the feedback, and sorry for the double PR.
Have a nice day.
@MatteoGheza feel free to contact me on linkedin
also, did you check the it.php file in creator folder? are all the strings translated properly?
|
2025-04-01T04:34:51.818272
| 2022-07-17T00:45:25
|
1306950801
|
{
"authors": [
"Ahmad-A0",
"ErdemOzgen",
"anjola-adeuyi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8977",
"repo": "neetcode-gh/leetcode",
"url": "https://github.com/neetcode-gh/leetcode/pull/509"
}
|
gharchive/pull-request
|
Remove All Adjacent Duplicates in String II - python and JavaScript
Solution using stack data structure in python and javascript
Thanks, @anjola-adeuyi!
Thanks @Ahmad-A0
|
2025-04-01T04:34:51.820886
| 2022-08-28T17:35:28
|
1353421107
|
{
"authors": [
"Ahmad-A0",
"MaratKhakim"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8978",
"repo": "neetcode-gh/leetcode",
"url": "https://github.com/neetcode-gh/leetcode/pull/958"
}
|
gharchive/pull-request
|
Kotlin: 124. Binary Tree Maximum Path Sum
File(s) Modified: 124-Binary-Tree-Maximum-Path-Sum.kt.
Language(s) Used: Kotlin
Submission URL: _https://leetcode.com/submissions/detail/785676279/_
Thanks, @MaratKhakim!
|
2025-04-01T04:34:51.836373
| 2017-08-24T15:40:46
|
252643924
|
{
"authors": [
"ghost",
"negokaz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8979",
"repo": "negokaz/vscode-live-server-preview",
"url": "https://github.com/negokaz/vscode-live-server-preview/issues/2"
}
|
gharchive/issue
|
Doesn't Display locally saved graphics and pictures
When I add a picture to my Code which lies locally on some Drive it isn't displayed.
Closed this issue because lacking information to resolve.
|
2025-04-01T04:34:51.900828
| 2021-03-22T12:51:35
|
837686391
|
{
"authors": [
"cryptoBeliever",
"kodty"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8980",
"repo": "nemgrouplimited/symbol-desktop-wallet",
"url": "https://github.com/nemgrouplimited/symbol-desktop-wallet/issues/1358"
}
|
gharchive/issue
|
Harvesting - Wallet not showing income from 25% harvester service fee
Reproduce:
set up a node leaving automatic harvesting on
set the beneficiary account to be the same address as the node account
credit the node account (which is already converted to multisig) with some xym and allow it to harvest
go to the wallet and select the cosigner account
open the harvesting page of the wallet and select the multisig account above
In the list of blocks harvested, it will show the income from the 75%, but the 25% beneficiary income does not show in the wallet while it does show in the explorer.
Fixed in integration branch.
Before:
After fix:
|
2025-04-01T04:34:51.903542
| 2023-10-23T18:52:38
|
1957817055
|
{
"authors": [
"igorssilva"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8981",
"repo": "nemo-ufes/frameweb-vp-plugin",
"url": "https://github.com/nemo-ufes/frameweb-vp-plugin/pull/44"
}
|
gharchive/pull-request
|
36 generate code from application model
Closes #36
Add:
Support for generation of code for application classes and interfaces;
New default templates.
Fix:
Template errors
Fixed some template errors and added a more robust template validation on imports.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.