added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:34:50.053543
2017-01-19T07:52:22
201787317
{ "authors": [ "artm", "janjon", "mspae", "thesolotraveller" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8849", "repo": "mspae/react-wavesurfer", "url": "https://github.com/mspae/react-wavesurfer/issues/28" }
gharchive/issue
plugins/timeline Can not introduce Introduction method import Timeline from 'react-wavesurfer/lib/plugins/timeline'; error: React.createElement: type should not be null, undefined, boolean, or number. It should be a string (for DOM elements) or a ReactClass (for composite components Hello, can you post your code? Which transpilation library and config are you using? Can you maybe create a codepen which reproduces the issue. Thanks! I'm using electron-compile for transpilation. I'm 95% sure it uses babel for the actual work. I get a different error when trying to import a plugin: import Regions from "react-wavesurfer/lib/plugins/regions"; Uncaught ReferenceError: WaveSurfer is not defined at WaveSurfer.Regions.init.wavesurfer.wavesurfer (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:288) at Object.<anonymous> (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:290) at __webpack_require__ (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:30) at Object.module.exports (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:73) at __webpack_require__ (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:30) at Object.defineProperty.value (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:50) at /home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:53 at webpackUniversalModuleDefinition (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:3) at Object.<anonymous> (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:10) at Object.<anonymous> (/home/artm/src/surok/node_modules/react-wavesurfer/lib/plugins/regions.js:831) ah, and apparently electron-compile uses webpack for which you're describing some extra confguration, I'll have a look into that. I had a look and now I realize that electron-complile doesn't use webpack, the webpack in the stack trace comes form your code, sorry about the confusion. I don't understand the instructions though, how can I achieve the same effect if I'm using babel for transpilation? Ok, I've figured it out. In my case (transpilation by babel via electron-compile) the following requires/imports section works: window.WaveSurfer = require("wavesurfer.js"); let Regions = require("react-wavesurfer/lib/plugins/regions").default; let Minimap = require("react-wavesurfer/lib/plugins/minimap").default; import React from "react"; import Wavesurfer from "react-wavesurfer"; import Segmentator from "./segmentator"; Hi ! In my case regions is not loading and it is causing all other components to not load in main app. Here is the error it shows on including - import Regions from 'react-wavesurfer/lib/plugins/regions'; bundle.js:42575 Uncaught ReferenceError: WaveSurfer is not defined at WaveSurfer.Regions.init.wavesurfer.wavesurfer (bundle.js:42575) at Object. (bundle.js:42577) at webpack_require (bundle.js:42317) at Object.module.exports (bundle.js:42360) at webpack_require (bundle.js:42317) at Object.defineProperty.value (bundle.js:42337) at bundle.js:42340 at webpackUniversalModuleDefinition (bundle.js:42290) at Object. (bundle.js:42297) at webpack_require (bundle.js:20) Please refer to this: https://github.com/mspae/react-wavesurfer#prerequisites-and-common-pitfalls – If this doesn't fix it I'll reopen the issue.
2025-04-01T04:34:50.064760
2019-09-15T08:38:33
493711254
{ "authors": [ "mstksg", "rnhmjoj" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8850", "repo": "mstksg/uncertain", "url": "https://github.com/mstksg/uncertain/issues/4" }
gharchive/issue
Initialize values from a covariance matrix Sometimes you have a set of correlated numbers (for example parameters estimate of a fitted curve) that are given by a vector and a covariance matrix. I don't think it's possibile to handle these with the current interface will look into this! :) thanks for the idea!
2025-04-01T04:34:50.123842
2021-04-20T16:03:45
862974064
{ "authors": [ "camchenry", "kettanaito", "marcosvega91" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8852", "repo": "mswjs/data", "url": "https://github.com/mswjs/data/issues/78" }
gharchive/issue
Error when calling findFirst even though nothing has changed First, thanks for the great library. This solves the exact problem I have been working on, and we're already using MSW so this is a natural fit. I have been running into an issue where the relations seem to "disappear" from the database, even though nothing has changed? I created a CodeSandbox which consistently crashes: https://codesandbox.io/s/mswjs-relation-undefined-error-wkcfp?file=/src/App.tsx import { factory, oneOf, primaryKey } from "@mswjs/data"; import * as faker from "faker"; const db = factory({ user: { id: primaryKey(faker.datatype.uuid), firstName: String }, userObject: { id: primaryKey(faker.datatype.uuid), data: String, user: oneOf("user") } }); // Seed the database const seededUser = db.user.create({ id: "ab4f631c-cca4-498f-a5aa-4828352a7c69", firstName: "Test" }); db.userObject.create({ user: seededUser, data: "test data 1 - associated with user" }); const queryObject = () => { const object = db.userObject.findFirst({ where: { user: { id: { equals: "ab4f631c-cca4-498f-a5aa-4828352a7c69" } } } }); console.log("query:", { object, user: object.user }); }; // This will work OK, user is not undefined queryObject(); // This will NOT work ok setTimeout(() => { console.log("deferred query"); queryObject(); }, 1000); I make the same queryObject call twice: once immediately after creation, and another 1 second after seeding the database. My expectation is that both invocations of the function should do exactly the same thing. However, what happens is that the first call works as expected, while the second invocation crashes with the error TypeError: can't access property "__type", actualValue is undefined. Willing to help out with this, let me know what questions you have. Thanks in advance. (I apologize for all of the issues, but I am desperately trying to figure out this issue, and I have had no luck. I'd love to use this project.) Hi @camchenry thanks for raising the issue :). I think that making the property enumerable should solve the issue. I have already done it in the same PR as for the other issue because I need it for deleting internal properties. @marcosvega91 Sounds great! I will look forward to when that releases then. Thanks for raising this, @camchenry. You've been immensely helpful in uncovering these behaviors. That's indeed an issue and we should fix it in the upcoming release. @marcosvega91, I'll take a look at the internal properties removal pull request once more, would love to get it merged. It's a superb piece of work already, so huge thanks! I have one concern, but we'll figure it out. The issue itself is addressed in #67, but we need to add a test for it. @camchenry, please, would you be interested in contributing the relevant test? We can support you throughout the process and have this not only fixed, but bullet-proof against any future regressions? @kettanaito Sure! I'll try to get around to it today and put up a PR. With the test added, is there anything left to address to mak this as done, @camchenry? @kettanaito i think with the tests written and the manual testing I've done, I think the original issue should be fixed 🤞 @kettanaito Any chance this will make it into a new release sometime soon? This was somewhat of a blocker in my project, but I'd like to give it another go and see if I can uncover any more issues. Released in 0.3.0. Please give it a try, your feedback is highly valuable!
2025-04-01T04:34:50.143928
2018-07-09T23:22:29
339640451
{ "authors": [ "deronnax", "mszep", "nitrocode" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8853", "repo": "mszep/pandoc_resume", "url": "https://github.com/mszep/pandoc_resume/issues/49" }
gharchive/issue
Thank You Thank you @mszep for starting this project because it helped get me create multiple resumes, cover letters, interviews, which lead to multiple job offers, and finally a job! I enjoyed adding my own contributions to it and hope it helps others. :smile: Great to hear @nitrocode, and congrats on your new job! I had hoped this project might help someone in this way, but you're the first confirmed case :-) And thank you for your thoughtful contributions in other issues -- it's greatly appreciated! thank you also, @mszep
2025-04-01T04:34:50.161727
2015-02-27T02:09:50
59178800
{ "authors": [ "DerManoMann", "dragonmantank", "laurencei", "lioannou", "mtdowling", "swekaj", "taylorotwell" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8855", "repo": "mtdowling/cron-expression", "url": "https://github.com/mtdowling/cron-expression/issues/78" }
gharchive/issue
No way to pass DateTime with custom TimeZone to isDue In Laravel's scheduler component, we allow the user to specify a Timezone to evaluate the Cron against. However, cron-expression appears to always overwrite the timezone to the current application timezone, allowing no custom timezone to be set. Is there any way this can be resolved? I don't remember exactly why this was added really. This is what I got from the blame: https://github.com/mtdowling/cron-expression/commit/cbde1b5a2974c83a6926a7d387479f0e04bb93c7 I guess it's trying to normalize the timezone. I'm happy to make a change if needed. It appears we can maybe work around it on our end as well by passing a string representation instead of a DateTime object, though it does seem intuitively like the Timezone of the given DateTime object should be respected. Maybe I'm wrong there. Here is the PR someone is working on our side (https://github.com/laravel/framework/pull/7636) ... personally I kind of wonder if it should be fixed over here though. I agree it should be fixed here. The library shouldn't be changing the timezone without giving the user the option to specify the timezone it changes to. It's very important to use the correct timezone when checking to see if a cron expression is due since the expressions themselves are not timezone-aware. If you create expressions for GMT-8:00, then the library will always be off by 8 hours if it's converting any DateTimes to GMT/UTC. I remember a year or so ago, I modified my local copy of the library to accept an optional timezone in the CronExression::factory() method. That timezone was then used for any ::isDue() and similar calls. Unfortunately that work has been lost (I had forgotten I did that when I updated to using composer), but I'd be willing to recreate it. I think it'd be most useful to allow providing a default timezone in the ::factory() method as well as allowing the user to override that by passing in a DateTime with the appropriate timezone set in ::isDue() and similar. If no default timezone is supplied, then the system's default is used where necessary. I could probably work on this this weekend, if you'd like. Looks like this is not handled consistently across all of the public API. getRunDate() does just clone the given date/time if it is a DateTime instance. So, calling getMultipleRunDates() with a properly set up $currentTime will respect the timezone as given on $currentTime. I remember that was something I was working towards as I also use this to evaluate against different timezones. My stand is that if a DateTime instance is given, the timezone set in that instance should be used. For strings or 'now' the default timezone should be used. Is there a suggested fix for this? Anything I can do? It can be fixed. We need a PR that removes all of the timezone modifications from the library and any time a DateTime object will be modified, it's cloned. We fixed it on our end for now by passing a string representation of the time. On Thu, Mar 26, 2015 at 11:38 AM, Michael Dowling<EMAIL_ADDRESS>wrote: It can be fixed. We need a PR that removes all of the timezone modifications from the library and any time a DateTime object will be modified, it's cloned. — Reply to this email directly or view it on GitHub https://github.com/mtdowling/cron-expression/issues/78#issuecomment-86609789 . @mtdowling @dragonmantank - it looks like someone has made a PR for this issue here https://github.com/mtdowling/cron-expression/pull/115 Any chance of reviewing it - and potentially merging? I've run into the same issue as @taylorotwell (ironically for inside a Laravel application) that I really need to support custom timezones... @taylorotwell - this can be closed now as functionality was merged today: https://github.com/mtdowling/cron-expression/pull/134 This should have been corrected with #134, so closing.
2025-04-01T04:34:50.170939
2021-09-11T20:34:22
993936257
{ "authors": [ "mtgto", "wqoq" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8856", "repo": "mtgto/Unrar.swift", "url": "https://github.com/mtgto/Unrar.swift/issues/3" }
gharchive/issue
"Invalid Exclude" warnings when added to Xcode 13 Beta I'm getting the following warnings when adding this package to a project in Xcode Beta 13: Invalid Exclude '.../SourcePackages/checkouts/Unrar.swift/Sources/Cunrar/arcmem.hpp': File not found. Invalid Exclude '.../SourcePackages/checkouts/Unrar.swift/Sources/Cunrar/arcmem.cpp': File not found. My understanding is that it's because "arcmem.cpp" and "arcmem.hpp" are listed in the exclude section for the "Cunrar" target in Package.swift, but those files are not in the package. My understanding is that it's because "arcmem.cpp" and "arcmem.hpp" are listed in the exclude section for the "Cunrar" target in Package.swift, but those files are not in the package. Thanks for your report! I released v0.3.7.
2025-04-01T04:34:50.228485
2023-10-25T14:24:01
1961554978
{ "authors": [ "axkr", "mtommila" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8860", "repo": "mtommila/apfloat", "url": "https://github.com/mtommila/apfloat/issues/30" }
gharchive/issue
fresnelC, fresnelS for Apcomplex Can you please review my fresnelC, fresnelS implementations: https://github.com/axkr/symja_android_library/blob/3bf9f61f8dcc83b75ea2919d7ac2f6e10863ba25/symja_android_library/matheclipse-core/src/main/java/org/matheclipse/core/expression/ApcomplexNum.java#L222 https://github.com/axkr/symja_android_library/blob/3bf9f61f8dcc83b75ea2919d7ac2f6e10863ba25/symja_android_library/matheclipse-core/src/main/java/org/matheclipse/core/expression/ApcomplexNum.java#L245 Is it worth to move these algorithms to the apfloat library? There seem to be somewhat simpler looking representations using 1F2 in https://functions.wolfram.com/GammaBetaErf/FresnelC/26/01/01/ https://functions.wolfram.com/GammaBetaErf/FresnelS/26/01/01/ But your implementation of course works as well (representing the function first through erf and then erf through 1F1 I suppose I'm not really competent to say anything about performance, accuracy or numerical stability The only potential problem from my point of view are new Apfloat(0.5) new Apfloat(1.5) new Apfloat(-0.5) which might not at least in theory be accurate to more than approx. 16 decimal digits. Use the string constructor instead: new Apfloat("0.5") new Apfloat("1.5") new Apfloat("-0.5") I'm not sure about adding these to the apfloat library. There are of course very many similar cases where one special function can be represented through other similar or more generic special functions. In addition to Fresnel sin and cos, for example exp integral, log integral, sin integral, cos integral (and their hyperbolic versions), harmonic numbers, Bessel and Airy functions, elliptic integrals, ... These are now in the apfloat library.
2025-04-01T04:34:50.331229
2023-12-04T00:39:02
2022803158
{ "authors": [ "choonyme", "mtripg6666tdr", "nakamuraos" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8861", "repo": "mtripg6666tdr/nextjs-obfuscator", "url": "https://github.com/mtripg6666tdr/nextjs-obfuscator/issues/23" }
gharchive/issue
Build failed with webpack error when css filed imported Thanks for the nice work. I'm on Nextjs 14.0.3 and am doing an "import '../styles/global.css'" in my /pages/_app.tsx file. This import caused the build process to failed with the messages below. I've tried even emptying the content of global.css and it's still causing error until I remove the import completely. Only nextjs-obfuscator 1.3.0 works so far but not the later versions. Any insight on this? ./src/styles/globals.css TypeError: Class extends value undefined is not a constructor or null at Function.getCssDependency (D:\app\node_modules\next\dist\compiled\mini-css-extract-plugin\index.js:1:8033) at addDependencies (D:\app\node_modules\next\dist\compiled\mini-css-extract-plugin\loader.js:1:5836) at handleExports (D:\app\node_modules\next\dist\compiled\mini-css-extract-plugin\loader.js:1:6477) at D:\app\node_modules\next\dist\compiled\mini-css-extract-plugin\loader.js:1:7631 at D:\app\node_modules\next\dist\compiled\webpack\bundle5.js:28:814002 at D:\app\node_modules\next\dist\compiled\webpack\bundle5.js:28:131723 at symbolIterator (D:\app\node_modules\next\dist\compiled\neo-async\async.js:1:14444) at timesSync (D:\app\node_modules\next\dist\compiled\neo-async\async.js:1:5027) at Object.eachLimit (D:\app\node_modules\next\dist\compiled\neo-async\async.js:1:14208) at D:\app\node_modules\next\dist\compiled\webpack\bundle5.js:28:129914 Import trace for requested module: ./src/styles/globals.css Build failed because of webpack errors Hello, I created a new project to look into the issue, but I couldn't replicate it. Could you provide a reproducible repository or project, or instructions on how to reproduce? Thanks. Apologies. My fault. Turns out I've used nextjs-obfuscator the wrong way. Was doing the following which was obsolete I guess. const NextJSObfuscatorPlugin = require("nextjs-obfuscator"); const nextConfig = { webpack: (config, {dev}) => { if(!dev){ config.plugins.push(new NextJSObfuscatorPlugin({ ...nextjsoptions }, { obfuscateFiles: { app: true, pages: true, }, log: true, })) } return config; } } I'm glad that the issue was solved. If you have any other issues, please feel free to open a new issue. Hi @mtripg6666tdr, I have similar problem with this issue. This repro repo: https://github.com/nakamuraos/next-js-obfuscator-boilerplate Can you help me check it. Thank you. Hi @mtripg6666tdr, I have similar problem with this issue. This repro repo: https://github.com/nakamuraos/next-js-obfuscator-boilerplate Can you help me check it. Thank you. Your repo seems to be configured in the way for nextjs-obfuscator v2.x or lower, so please re-configure it in the right way for v3.x. See the README as a reference. Your repo seems to be configured in the way for nextjs-obfuscator v2.x or lower, so please re-configure it in the right way for v3.x. See the README as a reference. Thank you. Just reconfig as v3 and It's working now. Your repo seems to be configured in the way for nextjs-obfuscator v2.x or lower, so please re-configure it in the right way for v3.x. See the README as a reference. Thank you. Just reconfig as v3 and It's working now.
2025-04-01T04:34:50.345234
2021-04-04T20:05:40
849951088
{ "authors": [ "chaosprint", "matt-erhart", "mtthw-meyer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8862", "repo": "mtthw-meyer/libdaisy-rust", "url": "https://github.com/mtthw-meyer/libdaisy-rust/issues/17" }
gharchive/issue
Example Request: Low Latency Convolution see http://blog.reverberate.ca/post/zero-latency-convolution/ This would be my learning project, figure other might like it too, but you could get it working much fast than me. It's useful for reverb & cabinet/mic simulation via impulse responses. I assume it's what all the impulse response (IR) guitar pedals are doing. I want to get some DSP functions into the library (or a separate DSP crate) eventually. Once I get the core hardware enabled and working I want to do stuff like that and other things that can be ported from the cpp version of this library. This is still beyond my current knowledge level. You can watch me experimenting with stuff over here https://github.com/mtthw-meyer/embedded-dsp Cool! Check out https://www.youtube.com/channel/UCUR_LsXk7IYyueSnXcNextQ/playlists for more info on time-frequency analysis (using convolution) in the non-real time context. This is still beyond my current knowledge level. You can watch me experimenting with stuff over here https://github.com/mtthw-meyer/embedded-dsp Hi, You can also check: https://github.com/RustAudio/dasp How can we create a variable to store the dsp engine? I try to create the DSP instance during #[init], then return it to [local]: let oscillator = SinOsc::new(440.0); Local { oscillator } But in the audio_handler: let s = oscillator.next(); audio.push_stereo((s, s)).unwrap(); No sound is played.
2025-04-01T04:34:50.388951
2015-02-12T23:30:38
57536748
{ "authors": [ "QuinDennis", "acosme", "braoul", "kentoii17", "mrweix", "shivrajsa" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8863", "repo": "mudcube/MIDI.js", "url": "https://github.com/mudcube/MIDI.js/issues/122" }
gharchive/issue
midi js in android webview Hello! I am integrating midi.js in a phonegap app that is based on a webview. For ios where the webview is based on Safari this works great! thanks! For android however the webview is based on WebKit-based WebView (up to Android 4.4) and Chromium (Android 4.4 and higher) . All the sources are loaded but it just won't play a sound...... On Android 4.4 it works on Chrome Browser but not on the Chromium web view Debugging also doesn't show any mistakes....loads the midi object correctly and display for api: for android 4.1.2 (webkit webview) the api: "audiotag" for android 4.4 (chromium) the api: "webaudio" (same as chrome on ) Any ideas if this can be fixed with a workaround or if this is a problem of midi.js? Greetings! I have similar needs, midi.js in phonegap. I'll be generating local midi, so either need to pass a var (preferrably) or local file. Any ideas on how I should proceed, without having to modify midi.js? Found it in the dev branch, midi.USE_XHR=false I don't quite understand.... how do you propose this can be fixed? And where should one fix the midi.USE_XHR=false ? https://github.com/mudcube/MIDI.js, select the dev branch, download, in plugin.js set USE_XHR=false. midi.DEBUG = false; //midi.USE_XHR = true; midi.USE_XHR = false; Then, the MIDIPlayer.html example works and playing notes in Chrome works. Haven't moved it to phonegap yet but believe it will work there also. I'm not totally clear this solves your issue, but I'm good to go now! (I think) Mmmh I tried it and still no sound on android..... Please let me know if you have it working with Phonegap. We also included the file acoustic_grand_piano-ogg.js within a Oh, and yes, it works in Phonegap. On 2/16/2015 5:09 PM, mrweix wrote: Mmmh I tried it and still no sound on android..... Please let me know if you have it working with Phonegap. — Reply to this email directly or view it on GitHub https://github.com/mudcube/MIDI.js/issues/122#issuecomment-74586820. -- Quin Dennis |<PHONE_NUMBER> Thanks guys for all the input. I follow all the steps you recommend include the dev branch, include the soundfont files, everything works fine in Phonegap, but as soon as I bring it to an Android device: QUIET....I have no idea. Debugging and everything doesn't show any error. It just doesn't play a sound.... this is android 4.4 and 4.1 ...... on which android devices can you make it play a sound? Hi, yes, we're running into issues on a NextBook 8 as well. My Samsung G4 plays sounds but not as clearly as the browser, so we're still researching, in the area of SoundFont files, etc. Will keep you in the loop if/when we find something. On 2/19/2015 1:38 PM, mrweix wrote: Thanks guys for all the input. I follow all the steps you recommend include the dev branch, include the soundfont files, everything works fine in Phonegap, but as soon as I bring it to an Android device: QUIET....I have no idea. Debugging and everything doesn't show any error. It just doesn't play a sound.... this is android 4.4 and 4.1 ...... on which android devices can you make it play a sound? — Reply to this email directly or view it on GitHub https://github.com/mudcube/MIDI.js/issues/122#issuecomment-75120679. -- Quin Dennis |<PHONE_NUMBER> Great! Appreciate it! I tried some more and no luck. It really comes down to a problem of midi.js talking to Chromium. On Chrome I got it working, but Chromium seems to have some slight changes the way ogg audio is handled in the webaudio api.....anybody else has an idea how to make this work? I recently tried with phonegap with http://www.telerik.com for hybrid mobile app, I am using loop to play sequence of notes, it plays but not with perfect timing and some times some notes does not create sound. Anybody got it working perfectly with phonegap? Hi, I also have a similar problem : I would like to use the player of midi.js in a Crosswalk webview (which uses chrome for android) but it doesn't play any sound. And I have no idea what is happening since the console doesn't show error.. Is there someone who figured out how to solve this problem? or does someone see what is happening? Same problem here: Not work with ionic/cordova in android 4.1 and 4.4. I saw in 'loader.js' line 50 "root.audioDetect", chose between 'webmidi', 'webaudio' and 'audiotag'. Always selects audiotag, but in android 5.0, it selects 'webmidi'; Finally seeing 'plugin.audiotag.js' line 41, when execute 'audio.play()' emits no sound ; anyone? Also same problem working on midi.js. It works fine when running it on desktop browser but when running it on android webview things were different, it plays but no sound. I'm new to android and hopefully you can help me guys.
2025-04-01T04:34:50.395093
2018-10-05T17:01:09
367293122
{ "authors": [ "muddasheep" ], "license": "unlicense", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8864", "repo": "muddasheep/hqtrilogy", "url": "https://github.com/muddasheep/hqtrilogy/issues/209" }
gharchive/issue
crash in hq2_fox (god mode only) AlexCorruptor Hmmm I miiiiight have a culprit(edited) Are there anything that deals radiation/gASS damage? Besides the hallway in hq2_fox (Rad.), Z A D I Z M and tetris questioning (gASS) You know what? Just for massive lols i will compile a small test map with all the dmg types AlexCorruptorLast Tuesday at 8:17 PM Alrighty, this does NOT make sence I've compiled the vanilla-based map with ALL the possible vanilla DMGs I've also labeled the damage slots, let's say Got a crash on a number 16, which is Poison And i've crashed because of the timed effect after taking the DMG AlexCorruptorLast Tuesday at 9:37 PM I have no clue why your NEURALGAS and RADIATION damages are not crashing for me(edited) October 17, 2018 muddasheepLast Wednesday at 7:49 AM did you compile it with the spirit18.fgd? there are different values AlexCorruptorLast Wednesday at 7:49 AM Nope The timed effect was enough for me to raise an alarm muddasheepLast Wednesday at 8:47 AM are you testing it in vanilla HL? AlexCorruptorLast Wednesday at 8:48 AM Compiled for Vanilla HL, copied the result to HQ As in the map itself muddasheepLast Wednesday at 8:49 AM HQ uses different values because of spirit of halflife, that's why you have to compile the map using the spirit18.fgd, otherwise the test results might be off AlexCorruptorLast Wednesday at 8:49 AM Can you check Radiation and Nerve gas in the Spirit of Half-life script? muddasheepLast Wednesday at 8:49 AM I'm on a work trip today, but I can recompile your test map tomorrow when I'm back home and check it out myself yeah, but can't today so it's probably a mix of god mode + poison/radiation right? AlexCorruptorLast Wednesday at 8:50 AM Here the JMF file, so i don't have to send it later Attachment file type: unknown Phuck.jmf 87.53 KB So far in my testing Poison's duration effect crashed And in HQ2 case it was Radiation and Nerve Gas DMG
2025-04-01T04:34:50.403400
2024-11-28T22:02:51
2703345079
{ "authors": [ "MaStr", "johannesghd" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8865", "repo": "muexxl/batcontrol", "url": "https://github.com/muexxl/batcontrol/pull/73" }
gharchive/pull-request
Rewrite re-evaluation trigger every 3 minutes and add fixed offset of 15 secs Implements and closes https://github.com/muexxl/batcontrol/issues/63 and ensures the run is triggered on multiples of 3 minutes on the clock. (e.g. 03, 06, 09, 12, 15, 18, etc.) To avoid API overload at the exact minute a delay of 15 seconds has been implemented. The interval and the delay is currently not configurable, although maybe that would be wise to add to the config? Thank you for your PR. I tend to add a wait IF there is the need to call to the API and not shift the complete calculation. And we can decide per API if there is the need for that shift. On a local EVCC instance, I don't need to wait extra seconds. @johannesghd please review my latest commit on your branch. Yes, i did modified it that way.. Running for testing at my home
2025-04-01T04:34:50.438549
2018-08-10T21:01:29
349656235
{ "authors": [ "IssuehuntBot", "jedwards1211", "oliviertassinari" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8866", "repo": "mui-org/material-ui", "url": "https://github.com/mui-org/material-ui/issues/12472" }
gharchive/issue
Feature request: ability to override the Input component of a TextField [x] This is a v1.x issue. [x] I have searched the issues of this repository and believe that this is not a duplicate. Expected Behavior redux-form-material-ui v5 has a quick-and-easy TextField adapter component that injects the necessary helperText, value, onChange, etc. But to be able to use it for an address field, I need to override the Input component with something that does autocomplete (e.g. using react-places-autocomplete) Current Behavior Right now I'd have to basically duplicate the code in TextField, but with my own Input component instead. Right now I'd have to basically duplicate the code in TextField, but with my own Input component instead. @jedwards1211 What's wrong with that? The TextField is a simple wrapper component to solve 80% of the use cases. It doesn't aim at being more. Recomposing the text field is definitely a great pattern, it's allowing people to be closer to the actual DOM, something you might need here. As far as I understand the problem, I'm very tempted to label the issue "won't fix". Well, what do you see as the downside of allowing people to override the individual components? It would be a small, very simple amount of code to support this so to me it seems like there's virtually no downside to it. I can easily release a fork of TextField as it's own package but I thought it wasn't very justified to create additional fragmentation in the ecosystem for such a small change. Especially since it wouldn't automatically stay consistent with API changes here. Though I hadn't thought that I may be able to just use Select for this, we'll see. I'm happy to make the PR to do this btw, but I'm certainly not going to waste my time making PRs to this project anymore without any idea if they're likely to get merged Well, what do you see as the downside of allowing people to override the individual components? @jedwards1211 It's more about the direction we want to encourage people to go into. What's wrong about using FromControl, InputLabel, Input and FormHelperText? Regarding the integration with redux-form, you might not need a library for it: https://github.com/mui-org/material-ui/issues/8377#issuecomment-331893521 But we migrated to react-final-form: import React from 'react' import PropTypes from 'prop-types' import TextField from 'web/modules/components/TextField' function RFTextField(props) { const { autoComplete, helperText, input: { name, ...input }, InputProps, meta: { dirty, error, submitError, submitFailed }, ...other } = props return ( <TextField error={Boolean((dirty || submitFailed) && (error || submitError))} {...input} {...other} id={name} name={name} InputProps={{ inputProps: { autoComplete, }, ...InputProps, }} helperText={dirty || submitFailed ? error || submitError : helperText} /> ) } RFTextField.propTypes = { autoComplete: PropTypes.string, helperText: PropTypes.node, input: PropTypes.shape({ name: PropTypes.string.isRequired, }).isRequired, InputProps: PropTypes.object, meta: PropTypes.shape({ dirty: PropTypes.bool.isRequired, error: PropTypes.string, submitError: PropTypes.string, submitFailed: PropTypes.bool.isRequired, }).isRequired, } export default RFTextField Alright, encouraging one pattern doesn't mean we should prevent another when the overhead is low. Why not adding some InputComponent, SelectComponent, FormHelperTextComponent, FormControlComponent, and InputLabelComponent properties The overhead is going to be about making: the TextField source code harder to read add some more bytes down the wire Yeah it is true that it would add more bytes, that's a concrete downside. Now that I read what TextField is doing, it's not as complicated as I thought so duplicating its behavior with a custom input for redux-form wouldn't be so bad. I was a bit worried at first that I would overlook something and burn time on it. Which is true to some degree, the error and helperText logic are a bit subtle. I shouldn't have complained about whether a PR would get merged, because I always wind up with code I can use for myself even if it doesn't get merged. (I did release material-ui-popup-state btw) Now that I read what TextField is doing, it's not as complicated as I thought so duplicating its behavior with a custom input for redux-form wouldn't be so bad. @jedwards1211 Let me know what direction you want to take, I'm happy with both sides :). I did release material-ui-popup-state btw Awesome! I'm sorry I haven't answered yet on this topic. I have a backlog of items to handle that is growing. Do you want to link it in the documentation? Yeah I'll make a PR for that soon! @0maxxam0 funded this issue with $20. See it on IssueHunt Ah man, I completely forgot about this 😅 I was recently trying to make some input components from react-stripe-elements look like MUI inputs, but it wasn't so simple because that lib actually displays its inputs inside iframes. So I'm not sure how much flexibility what I was proposing here would buy us. In this case I was wishing the input underline was a public component I could use directly @jedwards1211 Thanks for the context. I'm taking the weekend to go through all the issues and close anything that has a low ROI potential. For this issue, we have added better integration documentation with Stripe since then. Oh, I never even thought to look for that. Thanks for letting me know!
2025-04-01T04:34:50.443221
2019-04-04T18:34:56
429424105
{ "authors": [ "eps1lon", "majames", "oliviertassinari" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8867", "repo": "mui-org/material-ui", "url": "https://github.com/mui-org/material-ui/pull/15202" }
gharchive/pull-request
[ToggleButton] align ToggleButton type prop with @types/react@latest @types/react recently narrowed the type on the button element this started causing issues with ToggleButton attempting to assign a looser type, string, to it Fixes #15200 @majames Can you rebase the pull request on the next branch? Can do!... but I thought to back port a fix you want to target master? https://github.com/mui-org/material-ui/blob/next/CONTRIBUTING.md#master-is-for-3x @majames Could you rebase the pull request on the next branch? It's already fixed in next. This is a backport specifically. If we backport we should use #15158 (without the @types/react bump. @majames It's not an important bug fix as it only impacts the lab. It's already fixed in next. @eps1lon Oh nice! @majames Thank you for taking the time!
2025-04-01T04:34:50.448780
2019-11-04T10:25:45
517061311
{ "authors": [ "eps1lon", "oliviertassinari" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8868", "repo": "mui-org/material-ui", "url": "https://github.com/mui-org/material-ui/pull/18174" }
gharchive/pull-request
[Slider] Improve UX for pointing device with limited accuracy This is an alternative solution to #18074. It aims to find a tradeoff between: Provide a tall enough interaction zone (e.g. to account for the size of a finger) Provide a small enough interaction zone (e.g. to avoid unintended interactions with a too tall interaction zone or not to "waste" space on the screen) Avoid overlapping other interactive elements (predictability) Override simplicity It is not an alternative as was extensively described in the original PR. I don't understand your point. From what I understand #18174 and #18074 try to solve the same problem: from time to time when we try to interact with the slider on mobile, we have to give multiple tries, until our finger touches the rail. Could you make an argument for each of these statements? Do you mean, why are these dimensions are important? Do you mean, why are these dimensions are important? How are these addressed or apply here. I don't understand your point. I even made a video. I can't help you more without physically pointing the pointer for you. Nothing in my PR talks about mobile. @eps1lon Oh, I think that I understand your point now, I was confused by the 48px touch target, for touch devices and the recording on a pointer device (the 7mm - 9mm recommendation of Google Material Design only applies to touch devices, right?) and the mention to a regression with v3. So these two pull requests solve two different problems. #18074 is about increasing the thumb interaction zone from this zone: to this zone: Thanks for the patience, it took me some time to understand. Yes I shouldn't have used this term. I just wanted to explain where the size comes from which lead to the misunderstanding that this targets mobile devices. @eps1lon Ok awesome, sounds like a great change 👍.
2025-04-01T04:34:50.451121
2020-12-31T05:22:56
776796110
{ "authors": [ "hyeonhong", "oliviertassinari" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8869", "repo": "mui-org/material-ui", "url": "https://github.com/mui-org/material-ui/pull/24205" }
gharchive/pull-request
[DatePicker] Fix year only view, hide the current month When you use the DatePicker and you select the year to be displayed only, the current month is shown with the year. [x] I have followed (at least) the PR section of the contributing guide. Closes #24195 Updated PickersCalendarHeader to show nothing in the case of displaying the year only. @hyeonhong Thanks for raising the issue @hyeonhong Thanks for raising the issue
2025-04-01T04:34:50.466905
2020-05-04T17:13:36
612032961
{ "authors": [ "Emiliano-Bucci", "TheWhiteShade", "alvamanu", "elmeerr", "mcancetin", "oliviertassinari", "schmod" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8870", "repo": "mui/material-ui", "url": "https://github.com/mui/material-ui/issues/20906" }
gharchive/issue
[AutoComplete] - Load value asynchronously does not set selected value When I have value after Autocomplete is rendered (originated from an async call), the value is not set as selected one. [x] The issue is present in the latest release. [x] I have searched the issues of this repository and believe that this is not a duplicate. Current Behavior 😯 "selected value" is empty when value is loaded asynchronously Expected Behavior 🤔 After I get data asynchronously, value should be selected Steps to Reproduce 🕹 https://codesandbox.io/s/hardcore-wind-6pjlc use getAsyncValue to see the issue Context 🔦 When I have a list of options and I need to have a selected value (e.g from a list of countries, I want USA to be selected already), if my list of countries is an array of objects [{key: 1, value: "USA"}, {key: 1, value: "Japan"}, {key: 1, value: "Belgium"}] and, BEFORE rendering Autocomplete, I pass {key: 1, value: "USA"}, it works and USA is shown as selected but, if I call a api to retrieve a user and then pass user.country ( = {key: 1, value: "USA"}) (which will happen AFTER Autocomplete is rendered) , it does not work, no data is selected and gives me that controlled/uncontrolled warning. As I just need to show what was selected before, I don't need a controlled autocomplete and I thought of using defaultValue to do such a thing...In this case I would pass only USA as defaultValue and everything would be fine. Now I can only manage to achieve this if I setup everything BEFORE rendering Autocomplete, which for the UI is weird as I won't be able to see the autocomplete until I have my data (which I also don't think it's necessary) Basically, I need what getOptionSelected does but getOptionSelected only triggers onChange I don't know if this is indeed an issue or if I just don't know how to use Autocomplete properly, if the later is true, I apologize in advance. I have been struggling a lot to use Autocomplete properly in my project and from time to time I end up with some kind of undesirable behavior. Thanks for considering my request. Regards, Your Environment 🌎 Same as codesand Tech Version Material-UI latest React latest Browser Chrome TypeScript OS Windows @elmeerr You can find people facing the same challenge in older issues. @oliviertassinari Thanks for taking the time but I've failed to find a use case like mine in older issues. The closest I was able to find was this https://github.com/mui-org/material-ui/issues/18748 but here his initial data is not loaded asynchronously, my example works with sync initial data, the problem is with async initial data. Reading the many issues related to inputValue / value / onInputChange / onChange / defaultValue and initial values I think it will be really helpfull to have a section in the docs with examples that covers the FAQs related to this properties...For example, is possible to see in the docs the function that handle changes assigned to the onChange from Autocomplete and also to onChange from TextField on renderInput...what is the reason behind? It's possible to find in the issues people saying "using onChange on TextField solves for me"...It can be my fault, I might be missing something..I just find confusing this part of the component FWIW, I've run into this as well, and it appears to be a bug, or is an unintuitive behavior that (at the very least) needs to be mentioned/addressed in the demo showing how to use async requests with <Autocomplete>. If <Autocomplete> doesn't work like any other input that you can initialize with a value, that needs to be mentioned somewhere. @schmod hey Andrew, thanks for sharing your thought...were you able to work this around? In my case, async values will only prepopulate when the component is set to multiple values, not to single value. Also, Autocomplete component has a state issue. Console is giving me a warning stating it's uncontrolled. It's out of my hands since it's part of the module. Any help would be greatly appreciated. I found a solution for this. First, you have to default the state value to null, not undefined. Then you have to determine in value prop what type of data you'll populate, like so: const getValue = (value: AsyncACValueType) => !multiple ? value !== null ? Array.isArray(field.value) ? field.value[0] : field.value : [] : value !== null ? value : []; then in value prop, add getValue(value) I don't know why, but i have realized it's because the selected value must be obtained from the same object. I'm not very clear on it but i did this and it worked for me. const [value, setValue] = useState([...options.filter((op1) => value.find(op2 => op1.id === op2.id ))]); <Autocomplete {...props} value={value} /> @TheWhiteShade When used in this way, if you have an onInputChange callback, it doesn't work. I think this component is not suitable for initialization with an asynchronous process and then searching. This MUI guys have a paid library that doesn't handle this kind of stuff (in 2024); you should go to prison for engineering incompetency.
2025-04-01T04:34:50.473794
2022-10-02T08:29:54
1393686779
{ "authors": [ "TamirCode" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8871", "repo": "mui/material-ui", "url": "https://github.com/mui/material-ui/issues/34572" }
gharchive/issue
[TextField] rendering input type datetime-local without value makes label go on top of placeholder text Duplicates [X] I have searched the existing issues Latest version [X] I have tested the latest version Steps to reproduce 🕹 import { TextField } from "@mui/material" export default function NativePickers() { return ( <TextField id="datetime-local" label="Next appointment" type="datetime-local" // defaultValue="2017-05-24T10:30" // from demo defaultValue="" /> ) } https://codesandbox.io/s/mui-bug-input-datetime-local-mvf72n?file=/demo.js:0-263 I tried the following things: defaultValue={undefined} defaultValue={null} defaultValue="" not having defaultValue at all the same behavior occurs with all the above. using value instead of defaultValue results in the same behavior as well. Current behavior 😯 When attempting not to have any initial value for datetime-local textfield, the label has a render issue and shows on top of the input. This is because it considers it as empty so it goes on top of it as usual. It needs a simple check specifically for datetime-local to always show on top because there will always be information there. This is the same for chrome, edge, and firefox. Expected behavior 🤔 I want to have a textfield of type datetime-local without any initial value, so that it would just display like so: (it shows like this because i selected it so the label moves up) Context 🔦 I want to have a datetime-local textfield without an initial value, so it would show as mm/dd/yyyy --:-- -- Your environment 🌎 codesandbox browsers: chrome, edge, firefox I guess now I know what this does InputLabelProps={{ shrink: true, }}
2025-04-01T04:34:50.493376
2022-11-15T16:59:19
1450098810
{ "authors": [ "Vansh-Baghel", "mpanjato-andri", "siriwatknp", "vimutti77" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8872", "repo": "mui/material-ui", "url": "https://github.com/mui/material-ui/issues/35159" }
gharchive/issue
Allow using linear-gradient color in theme palette Duplicates [X] I have searched the existing issues Latest version [X] I have tested the latest version Summary 💡 Allow using linear-gradient color in theme palette or allow to use it if it is non-default palette Examples 🌈 I try to use linear-gradient in palette, but I got MUI: Unsupported 'linear-gradient(-39deg, #4991f8 0%, #4bc1ff 100%)' color. https://codesandbox.io/s/determined-fermat-mikfcs?file=/src/App.tsx Motivation 🔦 My client wants to use linear-gradient as one of background of button. I see that the docs are lacking this info. We should mention that the palette.*.(light|main|dark) will be used to calculate channel colors (I think this should be fixed to only apply to the default palette) @vimutti77 to make the background works, you need to use background, not bgcolor (this is how CSS works). https://codesandbox.io/s/silent-sun-onmj8f?file=/src/App.tsx @vimutti77 Looks like you are trying to create a custom palette. Here is my suggestion: All of the color tokens should live in colorSchemes.(light|dark).palette.* to get the benefit of the sx prop. When you write sx={{ bgcolor: 'custom.main' }}, the value will resolve correctly. This PR will fix the issue. However, the sx prop does not work with background yet, so you can't do this sx={{ background: 'gradient.primary' }}. I think the gradient should just be color tokens and theme each component to use the gradient selectively. const GradientButton = ({ color = "primary" }: CustomButtonProps) => { return ( <Button sx={{ background: (theme) => theme.vars.palette.gradient[color].main, "&:hover": { background: (theme) => theme.vars.palette.gradient[color].dark } }} variant="contained" > test </Button> ); }; I see that the docs are lacking this info. We should mention that the palette.*.(light|main|dark) will be used to calculate channel colors (I think this should be fixed to only apply to the default palette) @vimutti77 to make the background works, you need to use background, not bgcolor (this is how CSS works). https://codesandbox.io/s/silent-sun-onmj8f?file=/src/App.tsx Idk why but Experimental_CssVarsProvider and experimental_extendTheme stopped working in React JS. I tried using it, but it didnt work. Then I used this, and now it works fine : `import { createTheme } from "@mui/material"; const theme = createTheme({ palette: { gradient: { main: "linear-gradient(-39deg, #4991f8 0%, #4bc1ff 100%)", mainChannel: "0 0 0", light: "linear-gradient(135deg, #4aaffd 0%, #4992f8 100%)", lightChannel: "0 0 0", dark: "linear-gradient(135deg, #4cc2ff 0%, #4aa0fa 100%)", darkChannel: "0 0 0", contrastText: "#fff", contrastTextChannel: "0 0 0" } } }); const Rewards = () => { return ( <Box sx={{ width: "80%" }}> <LinearProgress sx={{ background: (theme) => theme.palette.gradient.main }} variant="determinate" value={progress} /> ) export default Rewards; ` I see that the docs are lacking this info. We should mention that the palette.*.(light|main|dark) will be used to calculate channel colors (I think this should be fixed to only apply to the default palette) @vimutti77 to make the background works, you need to use background, not bgcolor (this is how CSS works). https://codesandbox.io/s/silent-sun-onmj8f?file=/src/App.tsx Experimental_CssVarsProvider and experimental_extendTheme didn't work for me who is using ReactJs. I tried using it, but the site was just crashing. Then I used createTheme and ThemeProvider, and now it works fine : import { createTheme , Box, Typography, Button, ThemeProvider } from "@mui/material"; const theme = createTheme({ palette: { gradient: { main: "linear-gradient(-39deg, #4991f8 0%, #4bc1ff 100%)", mainChannel: "0 0 0", light: "linear-gradient(135deg, #4aaffd 0%, #4992f8 100%)", lightChannel: "0 0 0", dark: "linear-gradient(135deg, #4cc2ff 0%, #4aa0fa 100%)", darkChannel: "0 0 0", contrastText: "#fff", contrastTextChannel: "0 0 0" } } }); const Rewards = () => { return ( <ThemeProvider theme={theme}> <Box sx={{ width: "80%" }}> <Button sx={{ background: (theme) => theme.palette.gradient.main }} variant="determinate" value={progress} /> </Box> </ThemeProvider> ) export default Rewards; @vimutti77 Looks like you are trying to create a custom palette. Here is my suggestion: All of the color tokens should live in colorSchemes.(light|dark).palette.* to get the benefit of the sx prop. When you write sx={{ bgcolor: 'custom.main' }}, the value will resolve correctly. This PR will fix the issue. However, the sx prop does not work with background yet, so you can't do this sx={{ background: 'gradient.primary' }}. I think the gradient should just be color tokens and theme each component to use the gradient selectively.const GradientButton = ({ color = "primary" }: CustomButtonProps) => { return ( <Button sx={{ background: (theme) => theme.vars.palette.gradient[color].main, "&:hover": { background: (theme) => theme.vars.palette.gradient[color].dark } }} variant="contained" > test </Button> ); }; Experimental_CssVarsProvider and experimental_extendTheme didn't work for me who is using ReactJs. I tried using it, but the site was just crashing. Then I used createTheme and ThemeProvider, and now it works fine : import { createTheme , Box, Typography, Button, ThemeProvider } from "@mui/material"; const theme = createTheme({ palette: { gradient: { main: "linear-gradient(-39deg, #4991f8 0%, #4bc1ff 100%)", mainChannel: "0 0 0", light: "linear-gradient(135deg, #4aaffd 0%, #4992f8 100%)", lightChannel: "0 0 0", dark: "linear-gradient(135deg, #4cc2ff 0%, #4aa0fa 100%)", darkChannel: "0 0 0", contrastText: "#fff", contrastTextChannel: "0 0 0" } } }); const Rewards = () => { return ( <ThemeProvider theme={theme}> <Box sx={{ width: "80%" }}> <Button sx={{ background: (theme) => theme.palette.gradient.main }} variant="determinate" value={progress} /> </Box> </ThemeProvider> ) export default Rewards; Experimental_CssVarsProvider and experimental_extendTheme didn't work for me who is using ReactJs. I tried using it, but the site was just crashing. Could you share a CodeSandbox to me so that I can take a look the error? My bad, it is working. I was directly giving the color to the background rather than using sx. It does work. https://codesandbox.io/s/affectionate-rui-tj5hnk?file=/src/App.js Hello all, i want to use gradient in my icon material ? is it possible ? For example for FavoriteIcon (import FavoriteIcon from '@mui/icons-material/Favorite'), i want to use gradient its color.
2025-04-01T04:34:50.497051
2023-03-27T14:08:20
1642162165
{ "authors": [ "canac", "siriwatknp" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8873", "repo": "mui/material-ui", "url": "https://github.com/mui/material-ui/issues/36657" }
gharchive/issue
[docs] Hidden down props should be exclusive not inclusive Duplicates [X] I have searched the existing issues Related page https://mui.com/material-ui/api/hidden/ Kind of issue Other Issue description The docs for <Hidden /> say that the down props hide the element at "this size and down". However, the code shows that it only hides the element below this size (i.e. the docs say inclusive and the code says exclusive). Context 🔦 No response Do you want to submit a PR that fixes this? @siriwatknp Sure! Feel free to wordsmith if desired.
2025-04-01T04:34:50.518330
2017-03-21T03:10:38
215616784
{ "authors": [ "mukulikadey", "rekandiah" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8874", "repo": "mukulikadey/SOEN341-Group1", "url": "https://github.com/mukulikadey/SOEN341-Group1/issues/109" }
gharchive/issue
Acceptance Cases These are the acceptance cases for Sprints 1, 2, 3, 4, and 5 For Sprint 5: This is what we expect to test for sprint 5 given are expected specifications and requirement for the upcoming sprint. Hours Spent: 4 Hours helping: 4
2025-04-01T04:34:50.520004
2022-11-04T05:06:30
1435559597
{ "authors": [ "mukundansundar" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8875", "repo": "mukundansundar/proposals", "url": "https://github.com/mukundansundar/proposals/pull/1" }
gharchive/pull-request
initial commit Signed-off-by: Mukundan Sundararajan<EMAIL_ADDRESS> This proposal will not be part of dapr/proposals repo ... This will go into the dapr/community repo if it is a sandbox project or dapr CLI repo if we decide compose should be part of CLI from start. The other listed items discussing dev day 1 exp will be part of the Dev Day 1 Experience project that I will create in Dapr org ...
2025-04-01T04:34:50.521413
2021-02-03T23:48:54
800793865
{ "authors": [ "IsaacEldridge" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8876", "repo": "mulesoft/docs-studio", "url": "https://github.com/mulesoft/docs-studio/pull/175" }
gharchive/pull-request
DOCS-6650-generate-scaffold-of-updated-api-spec-ie @fermujica Can you review this PR? I wasn't sure if we needed to use the word "scaffold" or not--seems to me that "generate a flow" is still accurate? I'm going to close this pull request and resubmit--there's no info in the ticket I didn't see. I'm going to close this pull request and start over--there was new info in the ticket I didn't see.
2025-04-01T04:34:50.569335
2019-03-18T02:23:04
422010264
{ "authors": [ "bochoven", "ofirgalcon", "rickheil" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8877", "repo": "munkireport/munkireport-php", "url": "https://github.com/munkireport/munkireport-php/issues/1237" }
gharchive/issue
Default timeout too short I find some modules like applications and fonts can take up to 90 seconds on some systems. I have an idle macmini server with SSD yet applications takes 80 secs. My $0.02 as the author of a module (Sophos) that requires configuring a longer timeout: I think the default one is appropriate to prevent munki runs from getting unnecessarily long. Admins who need more time can easily add it with a config profile. Perhaps the solution is more in optimization of code and in documenting the ability to increase the timeout. I’m with Rick on this issue
2025-04-01T04:34:50.661693
2022-05-12T04:47:21
1233438955
{ "authors": [ "marcelveldt", "vk2him" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8878", "repo": "music-assistant/hass-music-assistant", "url": "https://github.com/music-assistant/hass-music-assistant/issues/145" }
gharchive/issue
v1.0.17 - double left hand menus Per the above screenshot, a double left hand menu displays if you follow these steps: Click Music Assistant in left menu Click on Album,Artists or Tracks Click the back arrow next to Album,Artists or Tracks Click arrow next to Music Assistant Double menu is displayed see this one: #73 #73 is a different problem? Nope it is a side effect of the current workaround. Refer #73
2025-04-01T04:34:50.665230
2022-10-30T15:44:33
1428866119
{ "authors": [ "erkr", "gieljnssns" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8879", "repo": "music-assistant/hass-music-assistant", "url": "https://github.com/music-assistant/hass-music-assistant/issues/970" }
gharchive/issue
All mass players stop playing when restarting home-assistant What version of Music Assistant has the issue? all The problem When I restart HA all my mass players stop playing music. Isn't it possible to let the music play? How to reproduce Restart HA Relevant log output x Additional information x What version of Home Assistant Core are your running all What type of installation are you running? Home Assistant OS On what type of hardware are you running? Generic x86-64 (e.g. Intel NUC) That is unavoidable. Integrations like MA are started and stopped by home assistant. So this is by design and not a bug. I will close the issue. Best Eric I thought so. Isn't possible to run MA as an addon? I can't judge if that is technically feasible, but you can create a feature request for that under discussions on GH!
2025-04-01T04:34:50.690556
2023-09-13T09:37:58
1894140824
{ "authors": [ "muzairkhattak", "yxgnahz" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8880", "repo": "muzairkhattak/PromptSRC", "url": "https://github.com/muzairkhattak/PromptSRC/issues/3" }
gharchive/issue
About the accuracy on base-to-new generalization setting Hi, Thanks very much for your interesting and inspiring work and well-organized codebase! I am trying to reproduce your results in the paper, and here are the results I obtained for the base-to-new generalization setting:  | Imagenet | Caltech101 | Food101 | Dtd | Ucf101 | Oxford_Flowers | Oxford_Pets | FGVC | Standford_Cars | Sun397 | Eurosat | AVG -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- Base | 77.60 | 98.07 | 90.67 | 83.53 | 86.93 | 98.00 | 95.17 | 42.87 | 78.43 | 82.53 | 92.83 | 84.24 Std | 0.14 | 0.12 | 0.09 | 0.92 | 0.05 | 0.14 | 0.21 | 0.88 | 0.25 | 0.12 | 0.97 |   New | 70.30 | 93.87 | 91.57 | 59.63 | 78.93 | 77.10 | 97.33 | 36.60 | 75.30 | 78.80 | 72.93 | 75.67 Std | 0.08 | 0.05 | 0.09 | 3.52 | 2.07 | 0.14 | 0.49 | 0.42 | 0.59 | 0.41 | 3.92 |   H-mean | 73.77 | 95.92 | 91.12 | 69.58 | 82.74 | 86.30 | 96.24 | 39.49 | 76.83 | 80.62 | 81.69 | 79.73 Basically, I find that the accuracies on base classes are consistent with the results in the paper and their variances are quite small. However, when it comes to the accuracies on the new classes, the accuracies are lower than that reported in the paper (on average 75.67 vs. 76.10). After carefully checking the results, I find that the gap mainly comes from two datasets, DTD and EuraSAT, where the variances are much larger than other classes. Therefore, I wonder whether this phenomenon is normal or not? Did you observe larger variances for the two classes during testing on new classes? Thanks in advance! Hi @yxgnahz, Thank you for showing interest in PromptSRC! Regarding your query, we also observed a bit larger variances for these datasets, but the obtained results are same as mentioned in our main paper. So I think, the reproduced performance should roughly match the main paper results. There can be slight variance in the results when reproducing results on different machines, but the shared results gap on average novel class performance is significant (75.67 vs 76.10). From your table, even for base classes, some datasets have different results, such as 78.43 % on Stanford cars vs reported 78.27 %. In order to resolve the issue, can you kindly verify the following? You can make sure that you are using the same versions of PyTorch and other packages as mentioned in our installation read-me files shown here. We have noticed in our experiments that using different PyTorch and CUDA version leads to fluctuating results. So this might help in obtaining the exact results. You may kindly verify if you are using the provided scripts for running your experiments and not changing any hyper-parameters in the config files. For training in base-to-novel setting, we use this script which by default uses the vit_b16_c2_ep20_batch4_4+4ctx config file. Optionally, can you also re-run your experiments using the commit version of 1af6e40b01e0c0c24fd8b6854be835209a0cde74 for the Dassl.pytorch library. I have noticed that we used slightly older version of the Dassl.pytorch. It should not be a problem, but you may just check this as well in-case. I hope that would be helpful. Kindly let us know if this resolves your issue! Thanks for your prompt reply! I followed your installation guide to set up the environments (torch 1.9.0+cu111) and used exactly the same script you provided for the reproduction. All the experiments are conducted on an Nvidia 3090 GPU. I also tried with the provided Dassl version, and I found the results unchanged. Therefore, could you please show me the exact variances you observed for these datasets? Hi @yxgnahz, Thank you for verifying the installation. Actually I do not have access to the machine where all the checkpoints and results log files are stored. I will try to get the access, and share the variance details here as soon as possible. Lastly, can you clarify if you are training your models from the scratch to reproduce the numbers or are you using the provided pre-trained weights to reproduce the results? Thank you! Hi @muzairkhattak , The results above are trained from scratch using the scripts you provided in the document. By the way, I just verified the official released checkpoint for base-to-new setting, and here are results (accuracy on new classes) I obtained by running the reproduce inference script: DTD 63.08+/-3.15, Eurosat: 73.88+/-4.80 The averaged accuracies are very close to the results in the paper (less than 0.1%) and I think the minor gap may come from the difference between our hardware platform. Moreover, the variances from the released checkpoint are just consistent with what I have got in the models trained from scratch, which indicates that these two datasets indeed have large variances on new classes. Hi @yxgnahz, Thank you for sharing this update. Yes, I think these two datasets have higher variance as discussed. If manageable for you, can you try to reproduce the results from scratch using another machine if that is available? This will be a double check. Thank you once again! Hi @yxgnahz, I have just verified reproducing the main paper results by training models from scratch using this repository. I am able to obtain roughly same results as mentioned in our paper. For example, the reproduced results for EuroSAT are 92.90% for base classes and 73.83% for novel classes. Similarly on DTD, the base class results are 83.53% and novel class results are 62.97%. To facilitate the process of reproducing the results on these datasets, I have provided the training and testing log files through this link. You can verify and compare them with your log files. I hope that would be helpful! Thank you and kind regards! Hi @muzairkhattak , Sorry I only have 3090 GPUs on my server. If I have chance, I will test this on another machine. I check the logs and I find all the settings are just the same. I believe the large variance leads to the performance gap in my reproduction and the value of the gap is normal considering the variance of the two datasets. Thanks for your discussion!
2025-04-01T04:34:50.703680
2019-02-28T03:10:57
415426741
{ "authors": [ "gidfiddle", "mviereck" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8881", "repo": "mviereck/x11docker", "url": "https://github.com/mviereck/x11docker/issues/133" }
gharchive/issue
sometimes need delay before $Pid1pid is valid x11docker --version: 5.4.4 On a rather slow Ubuntu 18.04 machine, the command x11docker --nxagent ubuntu:bionic always fails with a "Container startup seems to have failed." error, despite (usually) working well on a faster Fedora 29 machine. I traced the problem to the following statement in the x11docker script: Pid1pid=$($Dockerexe inspect --format '{{.State.Pid}}' $Containername 2>>$Containerlogfile | rmcr) sets Pid1pid to zero. I succeeded in working around this problem by putting this line in a loop like another one in the script: for ((Count=1 ; Count<=20 ; Count++)); do [line above] if test x$Pid1pid != x && test $Pid1pid -ne 0; then break; fi sleep 0.2 done (The first test is needed because Fedora sometimes sets Pid1pid to a blank.) One pass through the loop usually suffices to get a valid Pid1pid. I cannot explain why a delay is necessary here even after the container has been determined to be "ready" because the test $Dockerexe exec $Containername sh -c : has succeeded. Could you please look into this issue and make an appropriate correction? Thanks for a great piece of software. Thank you for reporting! I cannot explain why a delay is necessary here even after the container has been determined to be "ready" because the test $Dockerexe exec $Containername sh -c : has succeeded. That is odd, indeed. I have made a commit with a repeated check like in your proposal, it should work now. Please try out. Thanks for a great piece of software. :-)
2025-04-01T04:34:50.706737
2018-08-31T11:11:18
355938915
{ "authors": [ "ratijas", "yohad" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8882", "repo": "mvoidex/hsdev", "url": "https://github.com/mvoidex/hsdev/issues/91" }
gharchive/issue
Cabal failes to install under clean haskell platform Hi, On a windows 10 I just installed haskell platform, them tried installing hsdev with the following result: C:\Users\yotam>cabal install hsdev Resolving dependencies... cabal: Could not resolve dependencies: [__0] trying: hsdev-<IP_ADDRESS> (user goal) [__1] trying: hlint-2.1.10 (dependency of hsdev) [__2] trying: haskell-src-exts-1.20.2 (dependency of hsdev) [__3] next goal: ghc-syb-utils (dependency of hsdev) [__3] rejecting: ghc-syb-utils-<IP_ADDRESS> (conflict: hsdev => ghc-syb-utils>=0.2.3 && <0.3) [__3] trying: ghc-syb-utils-<IP_ADDRESS> [__4] next goal: ghc (dependency of hsdev) [__4] rejecting: ghc-8.4.3/installed-8.4..., ghc-8.4.3, ghc-8.4.1 (conflict: ghc-syb-utils => ghc>=7.0 && <8.4) [__4] rejecting: ghc-8.2.2, ghc-8.2.1 (conflict: hsdev => ghc==8.4.*) After searching the rest of the dependency tree exhaustively, these were the goals I've had most trouble fulfilling: hsdev, haskell-src-exts, ghc-mod, hlint, base, ghc, ghc-syb-utils Which I found odd, as it seems to me that I need to versions of ghc. It just can not be build with the latest GHC (8.4.3 at the moment of writing), so use stack to get older lts snapshot and build hsdev inside that environment. lts-11.22 / ghc-8.2.2 works well. > stack --resolver=lts-11.22 install hsdev-<IP_ADDRESS> haddock-api-2.17.4 hdocs-0.5.2 Note that some other packages must be specified with fixed version. This is because stack refuses to resolve fuzzy dependency ranges by itself. There is some theory around predictable builds around it, but just skip it for now. Generally, when you see in Haskell package requirements version range like >= x.y.z it's probably a good idea to stick with x.y.z, as it was tested for sure and most likely won't cause troubles.
2025-04-01T04:34:50.709856
2021-03-31T19:29:39
847192072
{ "authors": [ "adsharma", "mwhittaker" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8883", "repo": "mwhittaker/quoracle", "url": "https://github.com/mwhittaker/quoracle/pull/5" }
gharchive/pull-request
Make Node a dataclass Dataclasses make the code more compact and generate many convenience functions. Ah, I've been targeting Python 3.6, and I think dataclasses are a 3.7+ thing? Backports are available: https://pypi.org/project/dataclasses/ In general, dataclasses eliminate so much boilerplate code that I use them wherever I can. Yes - it saves only 10 lines in this diff. Not an earth shattering difference :) I've addressed the bugs you found in the code and simplified the typing and default values. I'll leave it here just in case you end up using dataclasses elsewhere and then want to convert this one for consistency.
2025-04-01T04:34:50.714313
2012-02-22T19:36:48
3342123
{ "authors": [ "candrews", "mwrock" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8884", "repo": "mwrock/RequestReduce", "url": "https://github.com/mwrock/RequestReduce/issues/147" }
gharchive/issue
Dashboard does not work on IIS 6 When a Request Reduce site is deployed to an IIS 6 server, the dashboard at /RequestReduceContent/Dashboard doesn't work - the URL returns a 404. The same site deployed to IIS 7 works fine. Is there some special configuration that can be done so that the dashboard works on IIS 6? Sorry for not getting back to you on this yet. I need to setup an IIS 6 environment to get a better idea of what is going on and then I'll let you know. I would definitely like to support IIS6.
2025-04-01T04:34:50.722827
2016-08-26T04:28:22
173367879
{ "authors": [ "mxcl", "rlam3" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8885", "repo": "mxcl/PromiseKit", "url": "https://github.com/mxcl/PromiseKit/issues/518" }
gharchive/issue
How to use PromiseKit.wrap? I'm trying to use PromiseKit.wrap but I'm not able to get it to work. Is there a specific import that I'm not using that is causing this to happen? //Module 'PromiseKit' has no member named 'wrap' PromiseKit.wrap{ self.alamofireRequest1() }.then { result in alamofireRequest2(result) } This documentation is for the yet unreleased PromiseKit 4. This is mentioned at the top of the documentation. Sorry for this, PMK4 is out very soon, and the documentation needed an update, so I combined the two. The PMK3 equivalent is Promise { completionHandler in }. @mxcl So in my case. should my self.alamofireRequest1 and self.alamofireRequest2 need to be promise functions too? Or is this method of using Promise to wrap two seperate async requests not the accepted way to use PromiseKit? My problem involves Moya and PromiseKit. I'd like Moya to do the routing and the promisekit to handle the execution sequence. But there are not many examples out there using this approach. If alamofireRequest1 returns Promise then you are using wrap wrongly. You should instead use firstly: firstly { alamofireRequest1() }.then { result in //… } // or you can just then of any promise: alamofireRequest1().then { result in //… } If they take a completion-handler then you should use wrap: PromiseKit.wrap(alamofireRequest1).then { result in //… } If you are using PromiseKit 3 however there is no wrap. You didn't say which PromiseKit you are using. @mxcl both alamofireRequest1 and 2 are both returning Request not Promise Is there a way for me to wrap two requests and do them synchronously? And I'm using PromiseKit 3 Thanks! If they return Request, then no, you need a completion handler variant to use with this initializer: https://github.com/mxcl/PromiseKit/blob/master/Sources/Promise.swift#L92-L104 Closing due to lack of response. Let me know if you need further assistance. @mxcl I have a function which I would like to return a request from. But it seems like the Promises are in different closures. Is there a way to get the genericRequest to be returned back after chaining? Thanks. func alamofireRequest(url:String?,completionHandler:(response:Response<AnyObject,NSError>, error: NSError?) ->()) -> Request{ var genericRequest: Request? firstly{ pp() } .then{ _ -> Promise<Request> in print("BEGIN REQUST!!") // Obtain new token let user = User() // User is still grabbing the old token? let URL = APIURLFactory().makeURLFor(url!) // Debugging print("alamofireRequest: \(URL)") let mutableURLRequest = NSMutableURLRequest(URL: URL) mutableURLRequest.HTTPMethod = "GET" mutableURLRequest.setValue("application/json", forHTTPHeaderField: "Content-Type") print("USER JWT:::::::;") print(user.jwt_token) mutableURLRequest.setValue("JWT \(user.jwt_token)", forHTTPHeaderField: "Authorization") genericRequest = Alamofire.request(mutableURLRequest) .validate() .responseJSON{ response in completionHandler(response: response, error: response.result.error) } return Promise{ fulfill, reject in fulfill(genericRequest!) } }.then{ request in return request <<<< This is not how we do it right? } return genericRequest! <<<< need to return here but nil keeps popping up } func pp() -> Promise<String>{ return Promise{ fulfill, reject in if self.isCurrentAuthTokenIsExpired(){ print("Token is expired") //FIXME: locksmith not updated prior to running self.updateAuthTokenInLocksmithWithNewToken()// ERROR HERE } fulfill("UPDATED TOKEN") } } Thanks! What you are doing doesn't really make sense, as far as I can tell you want this: func alamofireRequest(url: String) -> Promise<Response> { return pp().then { _ -> Promise<Response> in let user = User() let URL = APIURLFactory().makeURLFor(url!) let mutableURLRequest = NSMutableURLRequest(URL: URL) mutableURLRequest.HTTPMethod = "GET" mutableURLRequest.setValue("application/json", forHTTPHeaderField: "Content-Type") mutableURLRequest.setValue("JWT \(user.jwt_token)", forHTTPHeaderField: "Authorization") return PromiseKit.wrap(Alamofire.request(mutableURLRequest).validate().responseJSON) } }
2025-04-01T04:34:50.737677
2016-11-04T12:33:23
187322246
{ "authors": [ "coveralls", "mxstbr", "samit4me" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8886", "repo": "mxstbr/react-boilerplate", "url": "https://github.com/mxstbr/react-boilerplate/pull/1192" }
gharchive/pull-request
Faq staying updated with boilerplate Keeping your project up-to-date with react-boilerplate can be difficult and there is no recommended approach at the moment. There is a long term goal to change this (see #1180), so this PR is more or less to document the official response to the common question "how to keep my project up-to-date or in sync with the boilerplate". Would love to hear feedback! Also added a TOC as I feel this makes it much easier to navigate. Coverage remained the same at 98.605% when pulling ab3385e03580787eff0a443f035353908b388f3a on samit4me:faq-staying-updated-with-boilerplate into 3aafc0c593d2f51bda03538d7131dec7168b8475 on mxstbr:dev. Coverage remained the same at 98.605% when pulling 05d28b46691418bdada7c597725f361f4304af9c on samit4me:faq-staying-updated-with-boilerplate into dd8f9a68382b640ab8c5d8caa9683f4afe6cad8c on mxstbr:dev. LGTM, thanks!
2025-04-01T04:34:50.740262
2024-04-03T08:21:01
2222254200
{ "authors": [ "godind", "jimmyendt" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8887", "repo": "mxtommy/Kip", "url": "https://github.com/mxtommy/Kip/issues/363" }
gharchive/issue
feat: hide widget label Hello, I'm fairly new to using Signal K and KIP, but really loving it and all the hard work! But i have to admit I have no experience coding or anything. I was wondering, would it be possible to make the Widget label removable? I know you can leave it blank, but I have the feeling that it does take up some space in the awesome boxes you can put the widgets in? is this correct? it would be amazing to maximize space maybe? i'm also on discord and maybe have some more ideas/questions if that is okay. thanks in advance Hi. Thanks for the feedback. It's possible but would require reworking all the Widgets. I am not convinced this use case would apply to most. It's would make pages hard to understand without any label. All marine package have labels for that purpose. Thanks for the fast reply. Sounds like a lot of work and it's merely a question. I was thinking about the wind-widget, or radial gauge widgets that would speak for itself and therefore maximizing the space. can i ask you another question on discord maybe? Absolutely!
2025-04-01T04:34:50.746495
2021-06-28T21:42:08
932017220
{ "authors": [ "hmreumann", "myckhel" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8888", "repo": "myckhel/laravel-chat-system", "url": "https://github.com/myckhel/laravel-chat-system/pull/20" }
gharchive/pull-request
Commented loadMigrationsFrom Hallo @myckhel I tested it in a brand new laravel project and it worked fine. When running php artisan migrate without publishing it only creates the base tables (users, password resets, and failed jobs). After running php artisan vendor:publish --provider="Myckhel\ChatSystem\ChatSystemServiceProvider" --tag='migrations' , it copies all the migrations in the folder with the same name. Running again the same command it doesn't duplicate the migrations. Running php artisan migrate again, it creates the tables corresponding to the package. I hope this works, i think it was easy, I've just commented the $this->loadMigrationsFrom(__DIR__.'/../database/migrations'); line in the boot function from the ChatSystemServiceProvider.php. It think it fixes #10 By the way i've learned a lot. Thanks for letting me contribute. Hernán. Thanks for your time. Job well done.
2025-04-01T04:34:50.790188
2022-03-13T09:48:34
1167532190
{ "authors": [ "AnthonyAspen", "mymmrac" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8889", "repo": "mymmrac/telego", "url": "https://github.com/mymmrac/telego/issues/20" }
gharchive/issue
Typo in telego/pradicates 💬 Telego version v0.10.2 👾 Issue description AnyMassage instead of AnyMessage. https://github.com/mymmrac/telego/blob/main/telegohandler/pradicates.go#L29 It's a minor typo, not sure worth opening a bug report later when I feel myself more confident about how this library works, I'll create pull requests directly ⚡️ Expected behavior AnyMessage 🧐 Code example // AnyMassage is true if message isn't nil func AnyMassage() Predicate { return func(update telego.Update) bool { return update.Message != nil } } It's okay to create issues, even for small typos, it will help improve the library. And creating PR is also welcomed 🙂, will really appreciate it.
2025-04-01T04:34:50.830763
2016-10-09T09:38:25
181877449
{ "authors": [ "rjonaitis", "vsergeev" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8890", "repo": "myriadrf/LimeSuite", "url": "https://github.com/myriadrf/LimeSuite/issues/71" }
gharchive/issue
Minimum TX bandwidth limited to 5 MHz The LMS7002 datasheet suggests the TXLPFL path can be tuned down to 2 MHz, but the code sets a lower limit of 5 MHz. Is there is an implementation detail behind the discrepancy? It would be nice to use the lower TX filter bandwidths for more narrowband signals, if they're available. The LMS7002 datasheet specifies TXLPFL frequencies in IF, while the code uses RF Ah, so it looks like it's operating at a digital IF of at least 2.5 MHz?
2025-04-01T04:34:50.873988
2022-10-12T20:39:49
1406784501
{ "authors": [ "SteffenHeu", "bsedio" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8891", "repo": "mzmine/mzmine_documentation", "url": "https://github.com/mzmine/mzmine_documentation/issues/72" }
gharchive/issue
Windows for "Aducts" and "Modifications" do not appear for ion identity networking Hello MZmine, I am running MZmine 3.2.8 on a Mac OS 12.6 Monterey. When I open "ion identity networking" and "ion identity library:setup" the windows under "Aducts" and "Modifications" do not render, and hence no aducts can be selected and the ion identity networking module does not execute. A screenshot is attached. Is this an error in MZmine3? Or am I doing something wrong that is preventing the window for selecting aducts to render? Thank you very much, Brian there are no adducts/modifications present by default. click "reset positive/negative" and "reset" do use the default list. if this issue persists, please use the MZmine repository to report it: https://github.com/mzmine/mzmine3/issues this is just for the documentation.
2025-04-01T04:34:50.882429
2018-09-08T21:29:31
358329348
{ "authors": [ "mzolfaghari", "sophia-wright-blue" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8892", "repo": "mzolfaghari/ECO-efficient-video-understanding", "url": "https://github.com/mzolfaghari/ECO-efficient-video-understanding/issues/17" }
gharchive/issue
frames for Kinetics dataset Hello, In the script 'create_list_kinetics.m', you have the following path: path_DB_rgb='/datasets/kinetics/train/db_frames//' I'm assuming this folder contains the frames for the kinetics videos. Are the frames for the videos available online somewhere, or is there a script available to split the videos into the frames? I tried running ''main.py' of your pytorch implementation and got the following error: when running ---> 20 for i, (input, target) in enumerate(train_loader): ...... FileNotFoundError: [Errno 2] No such file or directory: '/kinetics/pumping_gas/ib5PzcBeYIc_000004_000014/0004.jpg' Thanks, Hi @sophia-wright-blue Yes, you need to extract frames. For some datasets you can find the frames but for Kinetics I don't think you can find frames. To extract frames you can use this script. Please let me know if you still had problem with extracting frames. Thank you so much for your response, and thank you for your patience with the basic questions. I'm trying to use your PyTorch code from scratch. To clarify, here are the steps: 1). Download Kinetics dataset. To do this, I've found the following link: https://github.com/activitynet/ActivityNet/tree/master/Crawler/Kinetics Do you have a better script for doing this step? 2). Once I have downloaded all of the videos in a folder, I need to extract the frames from the videos. Here you have the script: https://github.com/mzolfaghari/chained-multistream-networks/blob/master/scripts/extract_frames_frmRate.sh I'd greatly appreciate your help in the exact command and software required to run the script to extract the frames for all the videos. I have the videos downloaded in a folder named kinetics/videos. I would like to extract the frames to the folder kinetics/frames. 3). Once the frames have been extracted, do we need to create the training and testing lists, by running the script: https://github.com/mzolfaghari/ECO-efficient-video-understanding/blob/master/scripts/create_lists/create_list_kinetics.m or is there an equivalent python script? 4). Once we have the frames extracted and lists created, we can run the script: https://github.com/mzolfaghari/ECO-pytorch/blob/master/scripts/run_ECOLite_kinetics.sh This would give us the trained model. 5). Use the trained model for inference on some test videos. Once again, greatly appreciate your help and guidance with this. @sophia-wright-blue 1- We used the same scripts! 2- Provided necessary scripts in this folder. Please check the code. 3- We don't have the python script for this. 4- Yes, after having the frames and list you can run the code and get the final model. 5- Yes. thank you so much!
2025-04-01T04:34:50.891093
2022-12-13T15:11:26
1494459434
{ "authors": [ "Prince-Hamza", "awais-amjed", "mzusin", "toolcool-org" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8893", "repo": "mzusin/toolcool-range-slider", "url": "https://github.com/mzusin/toolcool-range-slider/issues/4" }
gharchive/issue
require is not defined when bundling code with node it works fine with when used with react only. when i build with yarn build for nodejs server (because i have apis) . server shows a blank page .console says require is not defined. The slider can't work in node.js only as it requires DOM. What is your use case? Closing because there was no further response. Getting this Error in Production. I am using ReactJS v18 Slider works fine in dev mode but when I build for production and deploy it on Firebase I get this error. Uncaught ReferenceError: require is not defined I have tried to isolate the issue to this import import 'toolcool-range-slider'; Removing this import, the code works fine in production but obviously the Slider is not displayed :D I am using create-react-app How to you import the slider to your component? Please share some code sample. Following the guide here exactly: https://toolcool-range-slider.mzsoft.org/pages/react-typescript.html Here is my code related to the slider: import 'toolcool-range-slider/dist/plugins/tcrs-marks.min.js'; import 'toolcool-range-slider'; <tc-range-slider id={`slider-${id}`} data={data} value={value} ref={ sliderRef } slider-bg={color1} slider-bg-hover={color1} slider-bg-fill={color1} slider-height="2px" slider-width={min768 ? "380px" : null} slider-radius="2px" pointer-width="18px" pointer-height="18px" pointer-bg="white" pointer-bg-hover="white" pointer-bg-focus="white" pointer-border="0 solid transparent" pointer-border-hover="0 solid transparent" pointer-border-focus="0 solid transparent" pointer-shadow="2px 7px 8px 0 rgba(0, 0, 0, 0.15)" pointer-shadow-hover="2px 7px 8px 0 rgba(0, 0, 0, 0.15)" pointer-shadow-focus="2px 7px 8px 0 rgba(0, 0, 0, 0.15)" mousewheel-disabled="true" marks="true" min={minShown} max={maxShown} step={step} marks-count={0} marks-values-count={marksValueCount ?? (((maxShown - minShown) / step) + 1)} marks-color="transparent" marks-values-color="#A6A6A6" /> Produces Sliders like this in dev mode: OK, I'll try to create a react app with the slider and reproduce this error. ok Thanks So I removed the package from package.json and imported it using the script from cdn in my index.html file like this This works in Production. So I think the issue is in the bundling process somewhere. Yes, you're right, It looks like some conflict with babel package in create react app. I'm still investigating. It's hard to believe, but it seems that create react app is no longer being maintained and kind of deprecated. They all recommend Vite as a replacement. https://github.com/facebook/create-react-app/issues/13140 https://github.com/facebook/create-react-app/issues/12974 Here is a little more information https://github.com/reactjs/react.dev/pull/5487#issuecomment-1409720741 Apparently, it is unlikely that someone will soon answer the ticket I opened.
2025-04-01T04:34:50.894043
2022-02-03T13:36:09
1123095735
{ "authors": [ "Aw3som3Guy", "NdvJff" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8894", "repo": "n00mkrad/cupscale", "url": "https://github.com/n00mkrad/cupscale/issues/109" }
gharchive/issue
An error occured during upscaling Hi I have all my attachements, 32gb of ram and a 3060rtx gpu everytme i launch an upscale after 0.2sec it says " An error occured during upscaling" do you know what could be the problem ? cheers ! Have you tried changing the AI network? (Its the setting in the top left of the screen, a little ways above model selection.) I've found the middle choice (ESRGAN (ncnn)) has the best compatibility, and RealESRGAN has a massive speedup but you have to use a 4x model, no bigger no smaller.
2025-04-01T04:34:50.895773
2021-01-09T09:34:10
782571125
{ "authors": [ "n07734" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8895", "repo": "n07734/community-health", "url": "https://github.com/n07734/community-health/issues/4" }
gharchive/issue
Top up report existing data If you have older data already saved it would be good to be able to add newer data to it vs running it all again to also have new data. Options: Forwards pagination: would be more logical but tricky to implement as you would need the first request's pagination info and update pagination to start from and change the sorts in the query, then re save "origin" pagination info Do existing style requests and get data until it is no longer getting fresh data, finding if data is new or not and de-duping could be challenging and fiddly. Also remove need for preFetchedRepo key done
2025-04-01T04:34:50.912139
2016-03-04T22:16:36
138601999
{ "authors": [ "Freundschaft", "chaosmail", "lorem--ipsum", "samwzm" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8896", "repo": "n3-charts/line-chart", "url": "https://github.com/n3-charts/line-chart/issues/452" }
gharchive/issue
Vertical line does not show all the dots Hi, there, I have the data sharing the same time, which would be a vertical bar. however, it seems only the bottom data point displays the dot, all the other data points do not display dots. I paste the sample html page below. Could anyone take a look? By the way, good news is that if I put my mouse over that vertical bar, then all the dots will display. thanks, Sam below is the sample html to show the issue: <script src="./d3.min.js"></script> <!-- Reference n3-line-chart --> <script src="./LineChart.js"></script> <link rel="stylesheet" href="./LineChart.css"> <script type="text/javascript"> angular.module('app', ['n3-line-chart']) </script> <script type="text/javascript"> angular .module('app', ['n3-line-chart']) .controller('MainCtrl', function($scope) { $scope.data = { dataset0: [ {x: 0, val_0: 0, val_1: 0, val_2: 0, val_3: 0}, {x: 0, val_0: 0, val_1: 0, val_2: 2, val_3: 0}, {x: 1, val_0: 0.993, val_1: 3.894, val_2: 6.47, val_3: 14.347}, {x: 1, val_0: 0.993, val_1: 3.894, val_2: 8.47, val_3: 14.347}, {x: 1, val_0: 0.993, val_1: 3.894, val_2: 10.47, val_3: 14.347}, {x: 2, val_0: 1.947, val_1: 7.174, val_2: 13.981, val_3: 19.991}, {x: 3, val_0: 2.823, val_1: 9.32, val_2: 14.608, val_3: 13.509}, {x: 4, val_0: 3.587, val_1: 9.996, val_2: 10.132, val_3: -1.167}, {x: 5, val_0: 4.207, val_1: 9.093, val_2: 2.117, val_3: -15.136}, {x: 6, val_0: 4.66, val_1: 6.755, val_2: -6.638, val_3: -19.923}, {x: 7, val_0: 4.927, val_1: 3.35, val_2: -13.074, val_3: -12.625} ] }; $scope.options = { series: [ { axis: "y", dataset: "dataset0", key: "val_2", label: "An area series", color: "#1f77b4", type: ['line', 'dot'], id: 'mySeries0' } ], axes: {x: {key: "x"}} }; }); </script> <div ng-app=""> <p>Name : <input type="text" ng-model="name"></p> <h1>Hello {{name}}</h1> Yeah, currently we use the x value of the data point as reference - which does not allow multiple x values per series. I have to look into this in more detail if theoretically it could be changed. Thanks for reporting this! +1 I'd also volounteer to check if it would be possible to change that Okay so this is not gonna happen. Abscissas need to be distinct in order to keep things meaningful in the code, that's one of the few requirements. We can't implement a charting library that deals with every possible data structure. What's the use case for this data you posted @samwzm anyway ? Because to me it makes very little sense to have several different value for the same series on the same abscissa. I'd rather put that into different series. I think what @samwzm wanted to do is build a scatter plot https://upload.wikimedia.org/wikipedia/commons/thumb/0/0f/Oldfaithful3.png/240px-Oldfaithful3.png Ha, I hadn't thought of this. This can be useful indeed !
2025-04-01T04:34:50.948776
2015-06-23T07:03:21
90308076
{ "authors": [ "nDmitry", "rarila" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8897", "repo": "nDmitry/grunt-postcss", "url": "https://github.com/nDmitry/grunt-postcss/pull/33" }
gharchive/pull-request
Allow directory path and filename for annotation (Split of #30 into two different PRs) This adds the possibility to set the annotation either to a sourcemap filename (as postcss's default behavior) or a directory (path ends with ’’/’’ for use with multiple source files). (See #30 for discussion on it) Besides option complex logic I can see an issue in the implementation: If you specify a directory without a trailing slash you'll get something unexpected, so it's better to match the .map extension (or any extension), though it limits users on filename choice. We could also check if it's a directory or a file using fs module, but the given path may not exist at the time. I understand that PostCSS accepts paths to filenames and directories, but it also accepts relative paths and it is only a library API that we don't have to expose as it is. So honestly I don't wanna merge this one.
2025-04-01T04:34:50.950549
2018-01-16T21:22:05
289060461
{ "authors": [ "nWidart", "snipiba" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8898", "repo": "nWidart/laravel-modules", "url": "https://github.com/nWidart/laravel-modules/pull/463" }
gharchive/pull-request
Update: config.php added 'START_FILE' for replacement generator stubs/json.stub - added $START_FILE$ to reflect config.php Generators/ModuleGenerator.php - added function to return value from config/modules.php -> stubs.files.start ... Hi, I'm working the 3.0 release on master branch. Re-reading your pull-request, I don't see where you replace the $START_FILE$ key with start.php?
2025-04-01T04:34:51.019106
2023-11-08T09:18:08
1983122426
{ "authors": [ "CLAassistant", "junfengP" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8899", "repo": "nacos-group/nacos-controller", "url": "https://github.com/nacos-group/nacos-controller/pull/7" }
gharchive/pull-request
feat: decouple config operation from nacos sdk 解除和Nacos SDK的强耦合 Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
2025-04-01T04:34:51.073468
2016-05-06T12:04:49
153437470
{ "authors": [ "anonymau5", "liamquin", "nagadomi", "panfengli" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8900", "repo": "nagadomi/waifu2x", "url": "https://github.com/nagadomi/waifu2x/issues/110" }
gharchive/issue
Cuda issues. Installed with the latest instructions and received the following. It seems that it is having trouble accessing CUDA although it is installed. /usr/local/bin/luajit: /usr/local/share/lua/5.1/trepl/init.lua:383: lib/w2nn.lua:18: Failed to load CUDA modules. Please check the CUDA Settings. stack traceback: [C]: in function 'error' /usr/local/share/lua/5.1/trepl/init.lua:383: in function 'require' web.lua:7: in main chunk [C]: in function 'dofile' /usr/local/lib/luarocks/rocks/trepl/scm-1/bin/th:145: in main chunk [C]: at 0x00404270 I added more informative error message at https://github.com/nagadomi/waifu2x/commit/958dcfaeb8dfd3c71df65282168b211f745ff970. Could you retry and post the error message? A little more information on this machine; This machine uses a 550ti and has been running Waifu2x just fine for some time. I recently had a driver issue and had to purge and reinstall the nvidia drivers. The latest 352.63 drivers along with CUDA 7..5.18 is installed. There may be remnants of the older install still there. /usr/local/bin/luajit: /usr/local/share/lua/5.1/trepl/init.lua:383: lib/w2nn.lua:18: Failed to load CUDA modules. Please check the CUDA Settings. /usr/local/share/lua/5.1/trepl/init.lua:383: /usr/local/share/lua/5.1/cutorch/init.lua:2: cannot load '/usr/local/lib/lua/5.1/libcutorch.so' stack traceback: [C]: in function 'error' /usr/local/share/lua/5.1/trepl/init.lua:383: in function 'require' web.lua:7: in main chunk [C]: in function 'dofile' /usr/local/lib/luarocks/rocks/trepl/scm-1/bin/th:145: in main chunk [C]: at 0x00404270 user@Mami:~/.Waifu2X$ sudo apt-get install libsnappy-dev Reading package lists... Done Building dependency tree Reading state information... Done libsnappy-dev is already the newest version. 0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded. user@Mami:~/.Waifu2X$ sudo apt-get install libgraphicsmagick-dev Reading package lists... Done Building dependency tree Reading state information... Done Note, selecting 'libgraphicsmagick1-dev' instead of 'libgraphicsmagick-dev' libgraphicsmagick1-dev is already the newest version. 0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded. user@Mami:~/.Waifu2X$ sudo luarocks install graphicsmagick Installing https://raw.githubusercontent.com/torch/rocks/master/graphicsmagick-1.scm-0.rockspec... Using https://raw.githubusercontent.com/torch/rocks/master/graphicsmagick-1.scm-0.rockspec... switching to 'build' mode Cloning into 'graphicsmagick'... remote: Counting objects: 29, done. remote: Compressing objects: 100% (29/29), done. remote: Total 29 (delta 0), reused 19 (delta 0), pack-reused 0 Receiving objects: 100% (29/29), 788.37 KiB | 0 bytes/s, done. Checking connectivity... done. Updating manifest for /usr/local/lib/luarocks/rocks graphicsmagick 1.scm-0 is now built and installed in /usr/local/ (license: BSD) user@Mami:~/.Waifu2X$ sudo luarocks install lua-csnappy Installing https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/lua-csnappy-0.1.3-1.src.rock... Using https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/lua-csnappy-0.1.3-1.src.rock... switching to 'build' mode gcc -O2 -fPIC -I/usr/local/include -c lsnappy.c -o lsnappy.o gcc -shared -o snappy.so -L/usr/local/lib lsnappy.o Updating manifest for /usr/local/lib/luarocks/rocks lua-csnappy 0.1.3-1 is now built and installed in /usr/local/ (license: BSD) user@Mami:~/.Waifu2X$ sudo luarocks install md5 Installing https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/md5-1.2-1.src.rock... Using https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/md5-1.2-1.src.rock... switching to 'build' mode gcc -O2 -fPIC -I/usr/local/include -c src/compat-5.2.c -o src/compat-5.2.o -Isrc/ gcc -O2 -fPIC -I/usr/local/include -c src/md5.c -o src/md5.o -Isrc/ gcc -O2 -fPIC -I/usr/local/include -c src/md5lib.c -o src/md5lib.o -Isrc/ gcc -shared -o md5/core.so -L/usr/local/lib src/compat-5.2.o src/md5.o src/md5lib.o gcc -O2 -fPIC -I/usr/local/include -c src/compat-5.2.c -o src/compat-5.2.o -Isrc/ gcc -O2 -fPIC -I/usr/local/include -c src/des56.c -o src/des56.o -Isrc/ gcc -O2 -fPIC -I/usr/local/include -c src/ldes56.c -o src/ldes56.o -Isrc/ gcc -shared -o des56.so -L/usr/local/lib src/compat-5.2.o src/des56.o src/ldes56.o Updating manifest for /usr/local/lib/luarocks/rocks md5 1.2-1 is now built and installed in /usr/local/ (license: MIT/X11) user@Mami:~/.Waifu2X$ sudo luarocks install uuid Installing https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/uuid-0.2-1.src.rock... Using https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/uuid-0.2-1.src.rock... switching to 'build' mode Updating manifest for /usr/local/lib/luarocks/rocks uuid 0.2-1 is now built and installed in /usr/local/ (license: Apache 2.0) user@Mami:~/.Waifu2X$ sudo PREFIX=$HOME/torch/install luarocks install turbo Installing https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/turbo-2.0-4.rockspec... Using https://raw.githubusercontent.com/rocks-moonscript-org/moonrocks-mirror/master/turbo-2.0-4.rockspec... switching to 'build' mode Cloning into 'turbo'... remote: Counting objects: 165, done. remote: Compressing objects: 100% (145/145), done. remote: Total 165 (delta 28), reused 67 (delta 16), pack-reused 0 Receiving objects: 100% (165/165), 645.57 KiB | 0 bytes/s, done. Resolving deltas: 100% (28/28), done. Checking connectivity... done. Note: checking out 'f43bffa558742548b4bba1f39ef2771d97d7041c'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b new_branch_name Warning: variable CFLAGS was not passed in build_variables make -C deps/http-parser library make[1]: Entering directory /tmp/luarocks_turbo-2.0-4-6086/turbo/deps/http-parser' gcc -I. -DHTTP_PARSER_STRICT=0 -fPIC -Wall -Wextra -Werror -O3 -c http_parser.c -o libhttp_parser.o gcc -shared -Wl,-soname=libhttp_parser.so.2.1 -o libhttp_parser.so.2.1 libhttp_parser.o make[1]: Leaving directory /tmp/luarocks_turbo-2.0-4-6086/turbo/deps/http-parser' gcc -Ideps/http-parser/ -shared -O3 -Wall -g -fPIC deps/http-parser/libhttp_parser.o deps/turbo_ffi_wrap.c -o libtffi_wrap.so -lcrypto -lssl ==== Installing Turbo.lua v2.0.4 to: ==== ==== /home/user/torch/install/lib/lua/5.1 and ==== ==== /home/user/torch/install/share/luajit-2.0.4 ==== mkdir -p /home/user/torch/install/lib mkdir -p /home/user/torch/install/bin mkdir -p /home/user/torch/install/share/lua/5.1/turbo mkdir -p /home/user/torch/install/share/luajit-2.0.4/turbo cp -r turbo/* /home/user/torch/install/share/lua/5.1/turbo cp -r turbo.lua /home/user/torch/install/share/lua/5.1 cp -r turbovisor.lua /home/user/torch/install/share/lua/5.1 cp -r turbo/* /home/user/torch/install/share/luajit-2.0.4/turbo cp -r turbo.lua /home/user/torch/install/share/luajit-2.0.4 cp -r turbovisor.lua /home/user/torch/install/share/luajit-2.0.4 install -m 0755 bin/turbovisor /home/user/torch/install/bin ==== Building 3rdparty modules ==== make -C deps/http-parser library make[1]: Entering directory /tmp/luarocks_turbo-2.0-4-6086/turbo/deps/http-parser' gcc -shared -Wl,-soname=libhttp_parser.so.2.1 -o libhttp_parser.so.2.1 libhttp_parser.o make[1]: Leaving directory /tmp/luarocks_turbo-2.0-4-6086/turbo/deps/http-parser' gcc -Ideps/http-parser/ -shared -O3 -Wall -g -fPIC deps/http-parser/libhttp_parser.o deps/turbo_ffi_wrap.c -o libtffi_wrap.so -lcrypto -lssl ==== Installing libtffi_wrap ==== test -f libtffi_wrap.so && install -m 0755 libtffi_wrap.so /home/user/torch/install/lib/libtffi_wrap.so.2.0.4 && ldconfig -n /home/user/torch/install/lib && ln -sf libtffi_wrap.so.2.0.4 /home/user/torch/install/lib/libtffi_wrap.so ==== Successfully installed Turbo.lua 2.0.4 to /home/user/torch/install ==== Updating manifest for /usr/local/lib/luarocks/rocks turbo 2.0-4 is now built and installed in /usr/local/ (license: Apache 2.0) user@Mami:~/.Waifu2X$ th waifu2x.lua /usr/local/bin/luajit: /usr/local/share/lua/5.1/trepl/init.lua:383: lib/w2nn.lua:18: Failed to load CUDA modules. Please check the CUDA Settings. /usr/local/share/lua/5.1/trepl/init.lua:383: /usr/local/share/lua/5.1/cutorch/init.lua:2: cannot load '/usr/local/lib/lua/5.1/libcutorch.so' stack traceback: [C]: in function 'error' /usr/local/share/lua/5.1/trepl/init.lua:383: in function 'require' waifu2x.lua:5: in main chunk [C]: in function 'dofile' /usr/local/lib/luarocks/rocks/trepl/scm-1/bin/th:145: in main chunk [C]: at 0x00404270 It seems to be a cutorch(cuda module of torch data structure) issue. Please try to reinstall that. luarocks cutorch luarocks cunn user@Mami:~/.Waifu2X$ luarocks cutorch Error: Unknown command: cutorch user@Mami:~/.Waifu2X$ luarocks cunn Error: Unknown command: cunn user@Mami:~/.Waifu2X$ Ah, just noticed I need to do a sudo luarocks install cutorch Now I'm getting this. user@Mami:~/.Waifu2X$ sudo luarocks install cutorch Installing https://raw.githubusercontent.com/torch/rocks/master/cutorch-scm-1.rockspec... Using https://raw.githubusercontent.com/torch/rocks/master/cutorch-scm-1.rockspec... switching to 'build' mode Cloning into 'cutorch'... remote: Counting objects: 107, done. remote: Compressing objects: 100% (104/104), done. remote: Total 107 (delta 8), reused 34 (delta 1), pack-reused 0 Receiving objects: 100% (107/107), 156.81 KiB | 0 bytes/s, done. Resolving deltas: 100% (8/8), done. Checking connectivity... done. cmake -E make_directory build && cd build && cmake .. -DCMAKE_BUILD_TYPE=Release -DCMAKE_PREFIX_PATH="/usr/local/bin/.." -DCMAKE_INSTALL_PREFIX="/usr/local/lib/luarocks/rocks/cutorch/scm-1" && make -j$(getconf _NPROCESSORS_ONLN) install -- The C compiler identification is GNU 4.8.4 -- The CXX compiler identification is GNU 4.8.4 -- Check for working C compiler: /usr/bin/cc -- Check for working C compiler: /usr/bin/cc -- works -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working CXX compiler: /usr/bin/c++ -- Check for working CXX compiler: /usr/bin/c++ -- works -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Found Torch7 in /usr/local CMake Error at /usr/share/cmake-2.8/Modules/FindCUDA.cmake:548 (message): Specify CUDA_TOOLKIT_ROOT_DIR Call Stack (most recent call first): CMakeLists.txt:7 (FIND_PACKAGE) -- Configuring incomplete, errors occurred! See also "/tmp/luarocks_cutorch-scm-1-6154/cutorch/build/CMakeFiles/CMakeOutput.log". Error: Build error: Failed building. user@Mami:~/.Waifu2X$ sudo luarocks install cunn Installing https://raw.githubusercontent.com/torch/rocks/master/cunn-scm-1.rockspec... Using https://raw.githubusercontent.com/torch/rocks/master/cunn-scm-1.rockspec... switching to 'build' mode Cloning into 'cunn'... remote: Counting objects: 76, done. remote: Compressing objects: 100% (70/70), done. remote: Total 76 (delta 20), reused 17 (delta 2), pack-reused 0 Receiving objects: 100% (76/76), 93.95 KiB | 0 bytes/s, done. Resolving deltas: 100% (20/20), done. Checking connectivity... done. cmake -E make_directory build && cd build && cmake .. -DCMAKE_BUILD_TYPE=Release -DCMAKE_PREFIX_PATH="/usr/local/bin/.." -DCMAKE_INSTALL_PREFIX="/usr/local/lib/luarocks/rocks/cunn/scm-1" && make -j$(getconf _NPROCESSORS_ONLN) install -- The C compiler identification is GNU 4.8.4 -- The CXX compiler identification is GNU 4.8.4 -- Check for working C compiler: /usr/bin/cc -- Check for working C compiler: /usr/bin/cc -- works -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working CXX compiler: /usr/bin/c++ -- Check for working CXX compiler: /usr/bin/c++ -- works -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Found Torch7 in /usr/local CMake Error at /usr/share/cmake-2.8/Modules/FindCUDA.cmake:548 (message): Specify CUDA_TOOLKIT_ROOT_DIR Call Stack (most recent call first): CMakeLists.txt:7 (FIND_PACKAGE) -- Configuring incomplete, errors occurred! See also "/tmp/luarocks_cunn-scm-1-480/cunn/build/CMakeFiles/CMakeOutput.log". Error: Build error: Failed building. Sorry, luarocks install cutorch and luarocks install cunn are correct. If you use sudo, Maybe root does not read environment variables from user. sudo su export PATH=/usr/local/cuda/bin:$PATH export LD_LIBRARY_PATH=/usr/local/cuda/lib64:$LD_LIBRARY_PATH export CUDA_BIN_PATH=/usr/local/cuda/bin luarocks install cutorch luarocks install cunn I recommend you to reinstall torch with non root environment. See http://torch.ch/docs/getting-started.html Unstall command: https://github.com/torch/ezinstall#updating-from-a-previous-version Reinstalled torch with non root environment. Torch seems to have installed fine. No errors. user@Mami:~/.Waifu2X$ th web.lua /home/user/torch/install/bin/luajit: /home/user/torch/install/share/lua/5.1/trepl/init.lua:384: lib/w2nn.lua:18: Failed to load CUDA modules. Please check the CUDA Settings. /home/user/torch/install/share/lua/5.1/trepl/init.lua:384: module 'cutorch' not found:No LuaRocks module found for cutorch no field package.preload['cutorch'] no file 'lib/cutorch.lua' no file '/home/user/.luarocks/share/lua/5.1/cutorch.lua' no file '/home/user/.luarocks/share/lua/5.1/cutorch/init.lua' no file '/home/user/torch/install/share/lua/5.1/cutorch.lua' no file '/home/user/torch/install/share/lua/5.1/cutorch/init.lua' no file './cutorch.lua' no file '/home/user/torch/install/share/luajit-2.1.0-beta1/cutorch.lua' no file '/usr/local/share/lua/5.1/cutorch.lua' no file '/usr/local/share/lua/5.1/cutorch/init.lua' no file '/home/user/.luarocks/lib/lua/5.1/cutorch.so' no file '/home/user/torch/install/lib/lua/5.1/cutorch.so' no file '/home/user/torch/install/lib/cutorch.so' no file './cutorch.so' no file '/usr/local/lib/lua/5.1/cutorch.so' no file '/usr/local/lib/lua/5.1/loadall.so' stack traceback: [C]: in function 'error' /home/user/torch/install/share/lua/5.1/trepl/init.lua:384: in function 'require' web.lua:7: in main chunk [C]: in function 'dofile' ...user/torch/install/lib/luarocks/rocks/trepl/scm-1/bin/th:145: in main chunk [C]: at 0x00406670 I'm not sure why this newer version failed but I've rolled back to an older version with success. It seems that cutorch is still not installed. I guess that CUDA is not detected from torch installer. Is there nvcc? nvcc --version Then install cutorch luarocks install cutorch luarocks install cunn I got an older version up and going. I'll copy the newer version and see if it'll run. I found that FindCUDA didn't work if nvcc is in /bin - I had to put /usr/bin first in my PATH. This is because FindCUDA strips /bin/ from the directory when determining CUDA_HOME and if it's /bin then nothing is left. I got everything built but then it wouldn't run; looks like my nvidia card is too old, as cutorch gives me Unsupported gpu architecture 'compute_52' - I gave up for now. nvcc fatal : Unsupported gpu architecture 'compute_52' I don't know your GPU model but this error is caused by old nvcc(CUDA 6.5?). CUDA 7.5 is required. Thank you for replying. Yes, 6.5, since 7.5 doesn't support my GPU - NVIDIA Corporation GT216GLM [Quadro FX 880M]. I did find a version that is patched (by NVidia) to support it, but it's a 1GByte download and unfortunately it also includes the kernel drivers, which will cause other problems on my system. So I gave up for now, will try on a different machine when I get the chance. I'd love to see waifu2x written as a GEGL operation by the way. Bu fixing FindNVCC to work if nvcc is in /bin and not /usr/bin might be worthwhile as Linux systems are moving to have /usr/bin just be a symlink to /bin. Try sudo bash /torch/update.sh
2025-04-01T04:34:51.079873
2020-06-05T23:53:37
632067473
{ "authors": [ "apalabrados", "mcselasvegas", "tinohager" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8901", "repo": "nager/Nager.AmazonProductAdvertising", "url": "https://github.com/nager/Nager.AmazonProductAdvertising/issues/140" }
gharchive/issue
How to enable paging Your demo site shows paging but I can not get it to work with the code from the repository. Is that something I need to enable? Paging is spicify in the Request object: SearchRequest searchRequest = new SearchRequest { Keywords = keyword, ItemPage = **page**, BrowseNodeId = browseNoseId, Resources = new[] { "Images.Primary.Large", "ItemInfo.Title", "ItemInfo.Features", "ItemInfo.TradeInInfo", "ItemInfo.ByLineInfo", "ItemInfo.ManufactureInfo", } }; So, for every request, you should set the page as you can see above. https://github.com/nager/Nager.AmazonProductAdvertising/blob/9fdb77392138e8bf4b26f6f3813f6af453545d13/src/Nager.AmazonProductAdvertising/Model/SearchRequest.cs#L8
2025-04-01T04:34:51.081519
2022-06-10T13:02:03
1267514150
{ "authors": [ "dayaftereh", "iilyak" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8902", "repo": "nagix/chartjs-plugin-streaming", "url": "https://github.com/nagix/chartjs-plugin-streaming/issues/165" }
gharchive/issue
Streaming for non timebased data I have a stream of data which doesn't have a timestamp. The rate of events is so high that I need a microseconds resolution to use timestamp based approach. I know that JS doesn't support microseconds in Date object This means calling new Date() will not work. I have a monotonically increasing event id (integer). Is there a way I can implement a scrolling chart similar to the one provided by streaming plugin (with zoom feature) using linear scale instead of time or realtime? +1
2025-04-01T04:34:51.086468
2023-12-07T07:59:22
2030119838
{ "authors": [ "Muni10" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8903", "repo": "nais/nitro", "url": "https://github.com/nais/nitro/issues/6" }
gharchive/issue
Support new clusters in nitro The current version of nitro does not support new clusters because of the etcd health check. We need something that handles this. Added --newCluster flag to use for first time install
2025-04-01T04:34:51.109289
2020-05-24T06:54:53
623808631
{ "authors": [ "Anish-M-code", "nandhabn" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8904", "repo": "nandhabn/python-Flask-Chat-App", "url": "https://github.com/nandhabn/python-Flask-Chat-App/issues/3" }
gharchive/issue
Request to add Software License Adding Software License makes the project usable by others for their own works. It also protects your works from being used in unjust manner. It is an important part of developing opensource projects. This source might help in choosing one : https://choosealicense.com/ LICENSE added. now you can use this project.
2025-04-01T04:34:51.110568
2021-03-12T08:43:53
829888050
{ "authors": [ "minhchienwikipedia", "nandorojo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8905", "repo": "nandorojo/moti", "url": "https://github.com/nandorojo/moti/issues/35" }
gharchive/issue
How to create animation component with moti? Hi, I want to create a custom animation component with Moti, how can I do it? Ex: I'm using LinearGradient I want to custom like Animated.createAnimatedComponent(LinearGradient) and then I can using animation for this component https://moti.fyi/api/motify
2025-04-01T04:34:51.136065
2023-06-04T14:22:52
1740402249
{ "authors": [ "arsantiagolopez", "axe312ger", "nandorojo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8906", "repo": "nandorojo/solito", "url": "https://github.com/nandorojo/solito/issues/398" }
gharchive/issue
Unable to get any starter running on iOS simulator Maybe some updates broke sth? Looks like metro has issues. I run MacOS + Node v18, freshly installed XCode + Simulator. No matter which starter I pick, I get as soon I open the app in the simulator: Unable to resolve "./App" from "index.js" The web variant works fine for each starter. I tried the expo starter, it works fine with my iOS Simulator. Any ideas? 🤔 does that file exist @nandorojo thx for super fast reply. But index.js requries it? 🤔 import { registerRootComponent } from 'expo'; import App from './App'; // registerRootComponent calls AppRegistry.registerComponent('main', () => App); // It also ensures that whether you load the app in Expo Go or in a native build, // the environment is set up appropriately registerRootComponent(App); does App exist there? the error says it doesn’t. I don't know, I executed npx create-solito-app@latest and wanted to see how it looks in the browser and in the iOS simulator. I did zero changes to the files. I am assuming, that the starters are ready to be used with device simulators. Am I wrong here? you’re not wrong, but it would still help to check is there any other expo port or app running currently? if so, maybe try clearing all your ports out Ill restart my machine and close all the terminals 😇 So, some delays, but I did a fresh start, after restarting, some updates and making sure nothing else is running. I still end up with the same error message. Have to note that the troublesome index.js is created when executing npx expo run:ios Full log ❯ npx create-solito-app@latest my-solito-app-fresh-gh-tailwind -t with-tailwind 🧑‍🎤 Creating solito app... Creating a new solito app in my-solito-app-fresh-gh-tailwind... my-solito-app-fresh-gh-tailwind folder created. Copying template into my-solito-app-fresh-gh-tailwind... Downloaded template into my-solito-app-fresh-gh-tailwind... my-solito-app-fresh-gh-tailwind created! Installing packages. This might take a couple of minutes. > yarn install ➤ YN0000: ┌ Resolution step ➤ YN0002: │ @expo/cli@npm:0.6.2 doesn't provide expo-modules-autolinking (pe2591), requested by @expo/prebuild-config ➤ YN0002: │ app@workspace:packages/app doesn't provide react (p0cb3a), requested by @react-navigation/native-stack ➤ YN0002: │ app@workspace:packages/app doesn't provide react (p9a3d3), requested by @react-navigation/native ➤ YN0002: │ app@workspace:packages/app doesn't provide react-native (p0e2b8), requested by @react-navigation/native-stack ➤ YN0002: │ app@workspace:packages/app doesn't provide react-native (p34a6c), requested by @react-navigation/native ➤ YN0002: │ app@workspace:packages/app doesn't provide react-native-reanimated (pe7eec), requested by moti ➤ YN0002: │ app@workspace:packages/app doesn't provide react-native-safe-area-context (pecff1), requested by @react-navigation/native-stack ➤ YN0002: │ app@workspace:packages/app doesn't provide react-native-screens (p96773), requested by @react-navigation/native-stack ➤ YN0002: │ app@workspace:packages/app doesn't provide tailwindcss (pa880e), requested by nativewind ➤ YN0002: │ babel-preset-expo@npm:9.3.2 doesn't provide @babel/core (pe14e9), requested by @babel/plugin-proposal-object-rest-spread ➤ YN0002: │ babel-preset-expo@npm:9.3.2 doesn't provide @babel/core (pfe2df), requested by @babel/plugin-proposal-decorators ➤ YN0002: │ babel-preset-expo@npm:9.3.2 doesn't provide @babel/core (p9681f), requested by @babel/plugin-transform-react-jsx ➤ YN0002: │ babel-preset-expo@npm:9.3.2 doesn't provide @babel/core (p3ab61), requested by @babel/preset-env ➤ YN0002: │ expo-asset@npm:8.9.1 doesn't provide expo (p79f18), requested by expo-constants ➤ YN0002: │ expo-asset@npm:8.9.1 doesn't provide expo (p185f2), requested by expo-file-system ➤ YN0002: │ expo-linking@npm:4.0.1 doesn't provide expo (pe2296), requested by expo-constants ➤ YN0002: │ expo-splash-screen@npm:0.18.1 [72208] doesn't provide expo-modules-autolinking (p45296), requested by @expo/prebuild-config ➤ YN0002: │ moti@npm:0.24.2 [f87a9] doesn't provide react (p17866), requested by framer-motion ➤ YN0002: │ moti@npm:0.24.2 [f87a9] doesn't provide react-dom (p750ce), requested by framer-motion ➤ YN0002: │ nativewind@npm:2.0.9 [f87a9] doesn't provide react (p0b0c3), requested by use-sync-external-store ➤ YN0002: │ next-app@workspace:apps/next doesn't provide eslint (p29e84), requested by eslint-config-next ➤ YN0002: │ next-app@workspace:apps/next doesn't provide expo (p8f46e), requested by @expo/next-adapter ➤ YN0002: │ next-app@workspace:apps/next doesn't provide postcss (p2228b), requested by autoprefixer ➤ YN0002: │ next-app@workspace:apps/next doesn't provide react (p2df79), requested by @expo/next-adapter ➤ YN0002: │ next-app@workspace:apps/next doesn't provide react (p8afd9), requested by next ➤ YN0002: │ next-app@workspace:apps/next doesn't provide react-dom (pb1bef), requested by next ➤ YN0002: │ next-app@workspace:apps/next doesn't provide react-native-web (p9df76), requested by @expo/next-adapter ➤ YN0002: │ next-app@workspace:apps/next doesn't provide webpack (pa5629), requested by @expo/next-adapter ➤ YN0002: │ react-native-codegen@npm:0.71.5 doesn't provide @babel/preset-env (pdcb72), requested by jscodeshift ➤ YN0000: │ Some peer dependencies are incorrectly met; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code ➤ YN0000: └ Completed ➤ YN0000: ┌ Fetch step ➤ YN0013: │ yargs-parser@npm:18.1.3 can't be found in the cache and will be fetched from the remote registry ➤ YN0013: │ yargs-parser@npm:21.1.1 can't be found in the cache and will be fetched from the remote registry ➤ YN0013: │ yargs@npm:15.4.1 can't be found in the cache and will be fetched from the remote registry ➤ YN0013: │ yargs@npm:17.7.1 can't be found in the cache and will be fetched from the remote registry ➤ YN0013: │ yocto-queue@npm:0.1.0 can't be found in the cache and will be fetched from the remote registry ➤ YN0000: └ Completed in 4s 417ms ➤ YN0000: ┌ Link step ➤ YN0007: │ turbo@npm:1.8.8 must be built because it never has been before or the last one failed ➤ YN0000: └ Completed in 5s 522ms ➤ YN0000: Done with warnings in 10s 260ms Success! Created my-solito-app-fresh-gh-tailwind at my-solito-app-fresh-gh-tailwind Inside that directory, you can run several commands: yarn web Starts the development server for the Next.js site. yarn native We suggest that you begin by typing: cd my-solito-app-fresh-gh-tailwind yarn web ~/dev/food-diary via 11GiB/32GiB took 17s ❯ cd my-solito-app-fresh-gh-tailwind ~/dev/food-diary/my-solito-app-fresh-gh-tailwind via  v18.16.0 via 14GiB/32GiB ❯ code . ~/dev/food-diary/my-solito-app-fresh-gh-tailwind via  v18.16.0 via 14GiB/32GiB ❯ npm run Scripts available in solito-with-tailwind via `npm run-script`: native cd apps/expo && yarn start web cd apps/next && yarn next ~/dev/food-diary/my-solito-app-fresh-gh-tailwind via  v18.16.0 via 14GiB/32GiB ❯ npm run web > web > cd apps/next && yarn next ready - started server on <IP_ADDRESS>:3000, url: http://localhost:3000 event - compiled client and server successfully in 1285 ms (413 modules) wait - compiling / (client and server)... event - compiled client and server successfully in 1251 ms (1348 modules) ^C ~/dev/food-diary/my-solito-app-fresh-gh-tailwind via  v18.16.0 via 14GiB/32GiB took 14s ❯ npm run native > native > cd apps/expo && yarn start Development build: Unable to get the default URI scheme for the project. Please make sure the expo-dev-client package is installed. Starting project at /Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind/apps/expo Starting Metro Bundler ▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄ █ ▄▄▄▄▄ █▄▀▀▄▄▀▀█▄█ ▄▄▄▄▄ █ █ █ █ ███▄█ ▀▀▄█ █ █ █ █ █▄▄▄█ ██▄▀▄▀███▀█ █▄▄▄█ █ █▄▄▄▄▄▄▄█ █ ▀▄▀ █ █▄▄▄▄▄▄▄█ █ ▄▀▄▄▀▄█▀ ▄▄▀▀█ ██▄▀▀▀▀▀▄█ █▀▀▀▀▀█▄▄▀▀ ▀█▄ ▄▄ ▀ ▀▀ █ █▀ ▄█▀ ▄▀▄▄█▄▄▀▄ ▀█▄▄▀██▀▄█ █ ▄▄ ▄▀▀██ ▄▄█ █▀ █▄ ▄ █ █▄████▄▄█ █ ▀▀▀▄█ ▄▄▄ █▄ ██ █ ▄▄▄▄▄ ██ █▄▀█▀▀ █▄█ ▄█▀ █ █ █ █ █ █▄█▄▄ ▄ ▄▄ █ █ █ █▄▄▄█ █▀▀▀█▄█ ▄▄▄█ █ █ █▄▄▄▄▄▄▄█▄▄███▄▄█▄▄▄██▄██▄█ › Metro waiting on http://localhost:8081 › Scan the QR code above with Expo Go (Android) or the Camera app (iOS) › Press a │ open Android › Press i │ open iOS simulator › Press w │ open web › Press j │ open debugger › Press r │ reload app › Press m │ toggle menu › Press ? │ show all commands Logs for your project will appear below. Press Ctrl+C to exit. › Opening on iOS... CommandError: No development build (com.solito.nativewind) for this project is installed. Please make and install a development build on the device first. Learn more › Stopped server ~/dev/food-diary/my-solito-app-fresh-gh-tailwind via  v18.16.0 via 17GiB/32GiB took 41s ❯ npx expo run:ios 📝 iOS Bundle Identifier Learn more ✔ What would you like your iOS bundle identifier to be? … com.anonymous.solito-with-tailwind ✔ Created native project | gitignore skipped ✔ Added Metro config ✔ Updated package.json and added index.js entry point for iOS and Android › Installing using yarn > yarn install » ios: icon: This is the image that your app uses on your home screen, you will need to configure it manually. ✔ Config synced ✔ Installed pods and initialized Xcode workspace. To address this warning, either add output dependencies to the script phase, or configure it to run in every build by unchecking "Based on dependency analysis" in the script phase. (in target 'solitowithtailwind' from project 'solitowithtailwind') Run script build phase 'Bundle React Native code and images' will be run during every build because it does not specify any outputs. To address this warning, either add output dependencies to the script phase, or configure it to run in every build by unchecking "Based on dependency analysis" in the script phase. (in target 'solitowithtailwind' from project 'solitowithtailwind') Run script build phase '[CP-User] Generate app.config for prebuilt Constants.manifest' will be run during every build because it does not specify any outputs. To address this warning, either add output dependencies to the script phase, or configure it to run in every build by unchecking "Based on dependency analysis" in the script phase. (in target 'EXConstants' from project 'Pods') › Build Succeeded › 0 error(s), and 4 warning(s) Starting Metro Bundler ▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄▄ █ ▄▄▄▄▄ █▄▄▄ ▀█▄█▄▄█▀▀▄ ███▄█▀█ ▄▄▄▄▄ █ █ █ █ ██▄▀ █ ▀ ▄▀▄▄▄ █▀▀ █ █ █ █ █ █ █▄▄▄█ ██▀▄ ▄███▀ ▀▄█▀█▀▀█ █▄█ █▄▄▄█ █ █▄▄▄▄▄▄▄█ ▀▄█ ▀▄▀ █ ▀ █▄█ ▀ ▀ █▄▄▄▄▄▄▄█ █ █ ▀▄██▄▀█▄▀▀ ▀▄ ▀███▀█▀█ █▀▄▀▀▀██▄█ █ ▀ ▄ ▀▄█ ▀█▄▄█▄ ██▀▄ ▄█▄▄▄█▀ ▄ █ █ █ ▄ █ █▄██▄██▀▀▄ ██▄█▀▀▄▀▄█ ▀ █▄▄▀ ▀▄▄█ █ █▀▄ ▄ ▄▀▄█▀▄█ ▄▄▄▄█▄██▄█▄▄▀ █ █▀▀▄▄█ ██▀▄▀ █▄ ▀▀ ▄▄ ▀▀ ▀ █ █▄ ▄▀█ ▄▀█ ▄▀██ ██▄ █▀▄▀▄██▀▄▀▄ ▄▄▄▄ █ ▄█▀█▄█▀▀██ █ █ █ ▀▄▄ ▄▄▄▀ ▀▄ ██ ▀▄ ▄▀▀▀▄█ ▄ ▄ ██▄██ █▄ ▀█▄ ▄▀▄▄█ ▀█ ▄ ██▀▄ ▄▄█▀▄██▄▄█▀▀█▄█ █▀ ▄▀▀▀▄▀▄ ▀ ▀▀ ▀██▄▄█▀█ ███ ▀█▄█▀ ▀███ █ ▄▀█▄ ▄ █ ██ ▄▄▄▀▄ ▄▄▄██▄▀▄█▀▀ ▄▀ ▄█ █▄██▄▄▄▄█ ▀▀█▀▄ ▀ ██ ▄▄▄▀█ ▄▄▄ ▄ ▄█ █ ▄▄▄▄▄ ██▄▄▄ ▀▄ █▀▄█▄▀▄▄ █ █▄█ ▀█▀▄█ █ █ █ █ █▀▀▀▄ ▄▄▄ ▀▀▀▀▀▄▀ ▄ ▄▄▀▀█▄█ █ █▄▄▄█ █▀▀▀ █ ▀ ██ █ █▄▀▄▄█▀▀ ▀▀▀ █ █▄▄▄▄▄▄▄█▄█████▄▄██▄▄████▄█▄█▄▄▄▄▄███▄█ › Metro waiting on com.anonymous.solito-with-tailwind://expo-development-client/?url=http%3A%2F%2F<IP_ADDRESS>%3A8081 › Scan the QR code above with Expo Go (Android) or the Camera app (iOS) › Press a │ open Android › Press i │ open iOS simulator › Press w │ open web › Press j │ open debugger › Press r │ reload app › Press m │ toggle menu › Press ? │ show all commands › Installing on iPhone 14 Pro Max warning: Watchman `watch-project` returned a warning: Recrawled this watch 3 times, most recently because: MustScanSubDirs UserDroppedTo resolve, please review the information on https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl To clear this warning, run: `watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'` warning: Watchman `watch-project` returned a warning: Recrawled this watch 3 times, most recently because: MustScanSubDirs UserDroppedTo resolve, please review the information on https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl To clear this warning, run: `watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'` warning: Watchman `watch-project` returned a warning: Recrawled this watch 3 times, most recently because: MustScanSubDirs UserDroppedTo resolve, please review the information on https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl To clear this warning, run: `watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'` warning: Watchman `watch-project` returned a warning: Recrawled this watch 3 times, most recently because: MustScanSubDirs UserDroppedTo resolve, please review the information on https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl To clear this warning, run: `watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'` warning: Watchman `watch-project` returned a warning: Recrawled this watch 3 times, most recently because: MustScanSubDirs UserDroppedTo resolve, please review the information on https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl To clear this warning, run: `watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'` warning: Watchman `query` returned a warning: Recrawled this watch 3 times, most recently because: MustScanSubDirs UserDroppedTo resolve, please review the information on https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl To clear this warning, run: `watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'` Recrawled this watch 3 times, most recently because: MustScanSubDirs UserDroppedTo resolve, please review the information on https://facebook.github.io/watchman/docs/troubleshooting.html#recrawl To clear this warning, run: `watchman watch-del '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind' ; watchman watch-project '/Users/bene/dev/food-diary/my-solito-app-fresh-gh-tailwind'` › Opening on iPhone 14 Pro Max (com.anonymous.solito-with-tailwind) › Opening com.anonymous.solito-with-tailwind://expo-development-client/?url=http%3A%2F%2F<IP_ADDRESS>%3A8081 on iPhone 14 Pro Max › Opening the iOS simulator, this might take a moment. › Logs for your project will appear below. Press Ctrl+C to exit. iOS Bundling failed 257ms Unable to resolve "./App" from "index.js" Am I missing something thats needs to be done to configure expo for iOS correctly? I'd guess non EAS usage is support? 🙈 i don’t think so, but maybe try EAS on your device? I got it running! I executed npm run native from the project root dir. This does not work. When I directly enter apps/expo, then do yarn and then execute npm run ios, the solito app opens in iOS simulator :) hmmm okay good to know, thanks. does yarn start in apps/expo work? https://github.com/nandorojo/solito/blob/89c9dbb79d377ad663c326904bcfbed764caf6cd/example-monorepos/blank/package.json#L16 running yarn ios builds the dev client, but yarn native starts the dev server once the dev client is already on the device does yarn start in apps/expo work? https://github.com/nandorojo/solito/blob/89c9dbb79d377ad663c326904bcfbed764caf6cd/example-monorepos/blank/package.json#L16 Yes it does! :) so yarn native from the root must work too, right? @nandorojo sorry for no/late response. Yes it should, at least everything was working fine the last time I was fiddling around. Will soon restart working on that project, hopefully then this will be the last comment here :) Thanks for your work! Stumbled across this comment as I'm experiencing the same thing a couple months later. Upon following the docs and running npx create-solito-app@latest, running yarn native gives me this error: CommandError: No development build (com.solito.nativewind) for this project is installed. Please make and install a development build on the device first. Learn more: https://docs.expo.dev/development/build/ The only way I got it to work was by following the steps suggested by @axe312ger I got it running! I executed npm run native from the project root dir. This does not work. When I directly enter apps/expo, then do yarn and then execute npm run ios, the solito app opens in iOS simulator :)
2025-04-01T04:34:51.150025
2018-05-14T23:31:14
323016586
{ "authors": [ "danhunsaker", "fsecada01", "vashchukmaksim" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8908", "repo": "nanobox-io/nanobox-engine-python", "url": "https://github.com/nanobox-io/nanobox-engine-python/issues/45" }
gharchive/issue
python-3.6 and psycopg2 (Error: pg_config executable not found.) I'm trying to setup an environment with nanobox python engine. If I use python-2.7 and then install a psycopg2 package (pip command within a container or with a requirements.txt) it works ok. But when I switch to python-3.6 I got an error "Error: pg_config executable not found.". This is a well known issue but I can't solve it with a nanobox container. I can't install any dependencies (python3-dev or libpq-dev) since there is none of these in Nanobox pkgsrc and even can't find a path to pg_conf. Is there any way to use python3 with psycopg2 with nanobox? My starting config: run.config: engine: python engine.config: runtime: python-3.6 extra_packages: - nodejs cache_dirs: - client/node_modules extra_path_dirs: - client/node_modules/.bin build_triggers: - server/requirements.txt - client/package.json - client/package-lock.json extra_steps: - cd server && pip install -r requirements.txt - cd client && npm install fs_watch: true data.postgresql: image: nanobox/postgresql config: version: 9.6 ` in extra_packages, declare pkgin as a new dependency. After that, write nanobox run in your command line, and write in pkgin install postgresql93-plpython'. Then write pkgin export > pkgin_dependencies.txt' or whatever extension is your notepad application. Then declare in extra_steps: 'pkgin im pkgin_dependencies.txt -y' That should resolve your issue, hopefully. Note that pkgin is already installed, since it's the package manager installing extra_packages in the first place, so you can skip adding it to your configuration.
2025-04-01T04:34:51.172828
2021-11-12T10:27:15
1051809585
{ "authors": [ "iGranDav", "jeanphibaconnais" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8909", "repo": "nantes-tech-communities/list-communities", "url": "https://github.com/nantes-tech-communities/list-communities/pull/39" }
gharchive/pull-request
Add CocoaHeads Nantes event Adding a new Nantes event : CocoaHeads Thanks @iGranDav for your contribution 🙏 @all-contributors please add @iGranDav for doc
2025-04-01T04:34:51.175994
2021-09-18T14:18:57
1000046349
{ "authors": [ "disasterdaddy" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8910", "repo": "naofireblade/homebridge-weather-plus", "url": "https://github.com/naofireblade/homebridge-weather-plus/issues/215" }
gharchive/issue
Error "code": "CDN-0001", "message": "Invalid apiKey." I installed version "homebridge-weather-plus v3.2.6" and get the error "code": "CDN-0001", "message": "Invalid apiKey during start. The config is: { "platform": "WeatherPlus", "service": "weatherunderground", "key": "ip9999999", "stationId": "IGELXXX9", "compatibility": "home" } I know that the apiKey and the stationId have correct values (not the value in the config here). What could be the clue for this problem? Best regards As reviewed via Google I had to generate a new apiKey via WU. Go to https://www.wunderground.com/member/settings, then select "API KEYS" and generate the new key. This key is only for your plugin. Copy it to your clipboard and paste it into your JSON plugin configuration. Restart your HomeBridge and everything will work. You will find your new values in the "Standard Room".
2025-04-01T04:34:51.195836
2023-03-13T10:01:51
1621146517
{ "authors": [ "angely-dev", "codingnetworksb" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8911", "repo": "napalm-automation-community/napalm-huawei-vrp", "url": "https://github.com/napalm-automation-community/napalm-huawei-vrp/pull/23" }
gharchive/pull-request
Add contextual diff computation mode Problem As you may know, the current implementation of _get_merge_diff() is not really sufficient. Something as simple as: Candidate Running (extract) # interface GigabitEthernet0/0/2 undo portswitch undo shutdown l2 binding vsi CUST # bgp 1234 router-id <IP_ADDRESS> ipv4-family vpn-instance SOME-VPN import-route direct # # interface GigabitEthernet0/0/1 description CustomerA undo portswitch l2 binding vsi CUST # interface GigabitEthernet0/0/2 description CustomerB shutdown # bgp 1234 router-id <IP_ADDRESS> ipv4-family vpn-instance SOME-VPN import-route static # Using the basic Python sample: from napalm import get_network_driver driver = get_network_driver('huawei_vrp') device = driver(hostname='switch', username='user', password='pass') device.open() device.load_merge_candidate(config=candidate) print(device.compare_config()) Will result in: undo shutdown import-route direct Inconsistencies: New items from GigabitEthernet0/0/2 are not displayed New items are displayed (if found) without any context In other words, the diff is unable to compute changes per-context (per-indented block). Proposed solution I ended up doing DiffPlus, a lightweight module to compute an incremental and contextual diff between two indented configs. Taking about a hundred lines of code, it only relies on Python builtins and has no extra dependencies. Though I did it originally for this specific use case, the module is generic enough to work with any indented config, theoretically. I ran it successfully on multiple Huawei platforms (S5732, NE20, NE05, CE6810) having from 2K to 16K number of lines. Demo The module is now on PyPI. Diff only Using the same configs as above: from diffplus import IndentedConfig, IncrementalDiff candidate = open('candidate.txt').read() running = open('running.txt').read() candidate = IndentedConfig(candidate, sanitize=True) # sanitize to remove comments, blank lines, etc. running = IndentedConfig(running, sanitize=True) diff = IncrementalDiff(candidate, running, colored=True) # colored mode is optional print(diff) Will output: interface GigabitEthernet0/0/2 + undo portswitch + undo shutdown + l2 binding vsi CUST bgp 1234 ipv4-family vpn-instance SOME-VPN + import-route direct Merging Alternatively, we can merge to get a preview of the full config before applying it: diff = IncrementalDiff(candidate, running, colored=True, merge=True) Will output: interface GigabitEthernet0/0/1 description CustomerA undo portswitch l2 binding vsi CUST interface GigabitEthernet0/0/2 description CustomerB shutdown + undo portswitch + undo shutdown + l2 binding vsi CUST bgp 1234 router-id <IP_ADDRESS> ipv4-family vpn-instance SOME-VPN import-route static + import-route direct Limitations It is important to understand this module does NOT check syntax or semantic. By essence, it computes the diff between two indented texts, which makes it both simple and generic. Also, deletions cannot be computed. I explain why here with an example. To sum up, it would require to re-implement the config logic in some way. Though it is tempting, that is far from easy and very platform-dependent. The diff in NAPALM Cisco IOS actually fakes deletions this way: a no command is considered a deletion even if it won't negate any existing command. This is very debatable and does not make sense to me. Changes in NAPALM Huawei VRP Option 1: modify _get_merge_diff() You may find this option too radical. def _get_merge_diff(self): running_config = self.get_config(retrieve='running')['running'] running_config = IndentedConfig(running_config, sanitize=True) merge_candidate = IndentedConfig(self.merge_candidate, sanitize=True) diff = IncrementalDiff(merge_candidate, running_config) return diff Option 2: add an optional_args This option is more prudent. We let users choose the diff computation mode: either they stick with the actual mode (by default, for legacy purposes) or they experiment the contextual mode through an optional argument contextual_diff. device = driver( hostname='switch', username='user', password='pass', optional_args={'contextual_diff': True} ) This is what I implemented in the pull request and I'd be glad to have your view. Any updates on my PR? Any updates on my PR? Hi @angely-dev sorry i haven't replied before. I will read and review your proposal in detail this weekend. Thanks @codingnetworksb. I am available for any question or to resolve the conflict which just appeared after you recent activity. Hi @angely-dev, this weekend i started to update the library because it was very outdated compared with original napalm library. I'm making sure it in complete compliance with getters structure, output, etc. I'm also implementing mocked data to for automatic testing when we make pull request. I couldn't review your pull request. I will do it as soon as i finish updating, hope during the week or weekend. And i'll let you know at that time to check the conflicts. Hi @angely-dev, i'm done until now with my updates. Could you check and fix the conflicts? Thanks for maintaining the repo @codingnetworksb. Please allow some time to fix the conflicts which appeared. So I checked and fixed the conflict @codingnetworksb. It was just a reformatting issue (see 502e930). Hi @vladislav-tenishchev The new code failed the Checks i have included recently. It's failing in Black format check. As napalm original project, we have included Black formatter check to this project, please format the code using Black formatter and push the code again. We appreciate your contribution to this project. So I did run the Black code formatter @codingnetworksb and committed the changes in my fork (https://github.com/napalm-automation-community/napalm-huawei-vrp/pull/23/commits/a8a8bf73e63673dd9f651113a286214edda6f1a3). It is OK for you at this point? Thanks! Any updates @codingnetworksb? Or some concerns about the PR? Not only I fixed the failed checks, but I also added a workflow in my project as well 😊 It includes the Black code formatting check and it ensures unit tests have passed (from Python 3.8 to 3.12). I also released the new version on PyPI. Any updates @codingnetworksb? Or some concerns about the PR? Not only I fixed the failed checks, but I also added a workflow in my project as well 😊 It includes the Black code formatting check and it ensures unit tests have passed (from Python 3.8 to 3.12). I also released the new version on PyPI. No concern about the PR. I got distracted with other projects, but this weekend I'll catch up on these PRs Hi @angely-dev , I was testing the PR in my local environment, using a Virtual NE40E and seems to work as you explain it. Too sad that deletions cannot be computed. I understood the complexity of this. Thank you for documenting this PR so well. I will approve and merge today. Regards, Hi @codingnetworksb, Thanks for the feedback and for approving the PR. I hope users will take benefit of it despite the limitations. I realized I didn't include an update of the README in the PR. We could update it this way: Make a brief mention of the contextual_diff option In the Quick Start section and link it to the PR for reference: from napalm import get_network_driver driver = get_network_driver('huawei_vrp') device = driver( hostname='<IP_ADDRESS>', username='admin', password='this_is_not_a_secure_password', optional_args={'contextual_diff': True} # enable contextual diff mode ) device.open() # Send Any CLI command # … Please refer to PR #23 for more information about the contextual_diff mode. Or make it more explicit by adding a new subsection like this: Diff mechanism I prefer the second way, but I let you choose if you are willing to update the README directly. Otherwise, I could do another PR just for this update. Hi @angely-dev , Both options sounds good to me. Please submit a new PR just that update. If you create a new subsection make it as summary as possible.
2025-04-01T04:34:51.201834
2022-06-06T10:58:27
1261684687
{ "authors": [ "haesleinhuepf", "jni" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8912", "repo": "napari/napari", "url": "https://github.com/napari/napari/issues/4623" }
gharchive/issue
Viewer close event 🚀 Feature The viewer should have an suscribable event that is fired once the viewer is closed. Motivation I would like to delete a temporary directory at the point in time when the viewer is closed. I am responsible for when to do this, because tempfile.mkdtemp is not capable of doing it. The napari event documentation does not list a suitable event for this: https://napari.org/guides/events_reference.html?highlight=event Pitch It would be cool if we could subscribe to an event such as viewer.events.closed.connect(my_close_event) Alternatives Additional context We've had some discussions about this before. @haesleinhuepf do you want this to happen before e.g. all the layers are cleaned up, or after, or don't care? As I recall someone else asked for this to happen before anything is actually done. In that case I might be inclined to call it "about_to_close" or "closing" Hey @jni , I personally don't care. I think QT and/or the Windows API sends two events: closing(flag) and closed(), and I find the name intuitive. Using the flag it can be cancelled. Maybe it would be cool to give developers the same flexibility here... Just adding a reference: http://www.java2s.com/Tutorial/CSharp/0470__Windows-Presentation-Foundation/WindowClosingandClosedevent.htm Sorry for the sharp C 😜
2025-04-01T04:34:51.232539
2020-11-24T16:33:57
749863848
{ "authors": [ "Czaki", "ziyangczi" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8913", "repo": "napari/napari", "url": "https://github.com/napari/napari/pull/1937" }
gharchive/pull-request
Unpin qt version Description Because of this bug https://bugreports.qt.io/browse/QTBUG-88688 Debian based image cannot be used to test Qt 5.15.2 release. A possible option is to switch to the ubuntu based image. This is done in this PR. But this increase build time. The possible workaround I to build a custom docker image for this. But then someone needs to maintain it. The debug process is hidden in #1935 which is a register of the brute force approach. Type of change [ ] Bug-fix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [ ] This change requires a documentation update References How has this been tested? [ ] example: the test suite for my feature covers cases x, y, and z [ ] example: all tests pass with my change Final checklist: [ ] My PR is the minimum possible work for the desired functionality [ ] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [ ] I have added tests that prove my fix is effective or that my feature works I can look into building container for the test, thanks for the note! I can look into building container for the test, thanks for the note! I could prepare such script and instruction. I suggest preparing a separate repository in napari and add a schedule to automatically rebuild every week and push on docker hub. Until now, my experience is limited to preparing Dockerfile and manual upload to docker hub (for package imagecodecs). I can look into building container for the test, thanks for the note! I could prepare such script and instruction. I suggest preparing a separate repository in napari and add a schedule to automatically rebuild every week and push on docker hub. Until now, my experience is limited to preparing Dockerfile and manual upload to docker hub (for package imagecodecs). I created a napari org in docker hub before to release docker version of napari, we can add another repo there, I can send an invite to you, can you share your account in an email to<EMAIL_ADDRESS> I sent You mail.
2025-04-01T04:34:51.244689
2023-07-07T14:35:26
1793658352
{ "authors": [ "Shikhar9425", "narayan954" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8914", "repo": "narayan954/dummygram", "url": "https://github.com/narayan954/dummygram/issues/703" }
gharchive/issue
[FEATURE] update readme Description want to improve "vision" and "status" in readme section Screenshots No response Additional information No response you can go ahead!
2025-04-01T04:34:51.262628
2024-10-14T15:51:04
2586442082
{ "authors": [ "CarloLepelaars", "FBruzzesi", "MarcoGorelli" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8915", "repo": "narwhals-dev/narwhals", "url": "https://github.com/narwhals-dev/narwhals/pull/1173" }
gharchive/pull-request
feat: skew This PR adds skew to Narwhals. Support is added for Polars, Pandas-like, Arrow and Dask. Checklist [x] Code follows style guide (ruff) [x] Tests added [x] Documented the changes Thanks for the kind words! Doctest should be fixed now. btw, if you wanted to just fix a typo somewhere in a separate pr (or, say, take https://github.com/narwhals-dev/narwhals/issues/1170), then once you're already a contributor, CI will always run automatically without me having to approve and run - just bringing this up in case it makes it easier for you Hey @CarloLepelaars, thanks for the PR! I left a few comments - the main challenge seems to be how different implementations are between pandas and polars native methods. However polars provide the formula it uses for the computation. It should be possible to reproduce that with native methods or using the series/expr methods that are already implemented in narwhals :) This is indeed challenging @FBruzzesi. Currently I've made it so every backend returns the biased population skewness, but we can potentially include an option for the unbiased skewness. Should we account for the len(ser) < 3 case and return 0? Let's see, this is where Pandas diverges from the rest. To make it consistent we should only handle the case where len(data)==2. In that case Pandas and PyArrow can return 0. Do you also think that is the way to go? Skewness for 2 elements: SciPy: 0.000000 Pandas: nan Polars: 0.000000 Skewness for 1 element: SciPy: nan Pandas: nan Polars: nan It may be worth checking that the numbers are same even when nulls are present Good one! Can add a case in unary_test.py that has nulls. Let's see, this is where Pandas diverges from the rest. To make it consistent we should only handle the case where len(data)==2. In that case Pandas and PyArrow can return 0. Do you also think that is the way to go? Yes, we are trying to stick with polars api and results, so let's manually force that if needed! Good one! Can add a case in unary_test.py that has nulls. That would be great - if it is too much though, we can also make it in a follow up PR I've covered the cases as discussed and made them consistent with Polars behavior. unary_test.py now also covers data with nan and cases where there are less than 3 rows. Thanks for addressing the cases, the CI failure seems unrelated. However I am still quite unsure that we are matching polars behavior. When counting number of elements for the base cases, we should ignore null values, then (pseudo code): if n_not_nulls==0: return None # same as pl.Series([]).skew() and pl.Series([None]).skew() elif n_not_nulls==1: return float("nan") # same as pl.Series([1]).skew() and pl.Series([1, None]).skew() elif n_not_nulls==2: return 0.0 # same as pl.Series([1, 2]).skew() and pl.Series([1, 2, None]).skew() else: return <compute_skew> Implemented your suggestions for nan policy. There is only one edge case left for Dask, where it outputs nan instead of 0.0 with 2 non null elements. Not sure how to adjust _dask/expr.py to account for that. Hey @CarloLepelaars, thanks for adjusting. CI is failing because in #1224 , compare_dicts was renamed to assert_equal_data. Implemented your suggestions for nan policy. There is only one edge case left for Dask, where it outputs nan instead of 0.0 with 2 non null elements. Not sure how to adjust _dask/expr.py to account for that. Regarding dask, I am not able to try it now, bif could definitly be a tricky one to get right! I am ok with marking it as xfail in tests for now thanks both! should be good, will do another check but this should make it into the next release Awesome, thank you both for working with me on this! Interesting trick to match the warning to Dask only. just pushed a fix as the else part of m3 / (m2**1.5) if m2 != 0 else 0 wasn't right nor tested will merge on green and this can enter the next release 🥦
2025-04-01T04:34:51.288253
2024-10-03T22:50:57
2565120405
{ "authors": [ "stephiescastle" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8916", "repo": "nasa-jpl/explorer-1", "url": "https://github.com/nasa-jpl/explorer-1/pull/650" }
gharchive/pull-request
feature: add support for secondary nav on EDU Collections template Checklist [x] Include a description of your pull request and instructions for the reviewer to verify your work. [x] Link to the issue if this PR is issue-specific. [x] Create/update the corresponding story if this includes a UI component. [ ] Create/update documentation. If not included, tell us why. [x] List the environments / browsers in which you tested your changes. [x] Tests, linting, or other required checks are passing. [x] PR has an informative and human-readable title PR titles are used to generate the change log in releases; good ones make that easier to scan. PRs will be broadly categorized in the change log, but for even easier scanning, consider prefixing with a component name or other useful categorization, e.g., "BaseButton: fix layout bug", or "Storybook: Update dependencies". [x] PR has been tagged with a SemVer label and a general category label, or skip-changelog. These tags are used to do the aforementioned broad categorization in the change log and determine what the next release's version number should be. Release Drafter will attempt to do the category labeling for you! Please double-check its work. Description Addresses feedback: https://github.com/nasa-jpl/www/issues/617 Changes Adds support for secondary nav on the EDU Collections template Adds additional stories to demonstrate the use-case with different hero configurations Instructions to test make vue-storybook View stories that pair secondary nav with: Hero: http://localhost:6006/?path=/story/templates-edu-pageeducollectionsdetail--secondary-nav&globals=theme:ThemeEdu Inline hero: http://localhost:6006/?path=/story/templates-edu-pageeducollectionsdetail--secondary-nav-hero-inline&globals=theme:ThemeEdu No hero: http://localhost:6006/?path=/story/templates-edu-pageeducollectionsdetail--secondary-nav-no-hero&globals=theme:ThemeEdu Tested in the following environments/browsers: Operating System [x] macOS [ ] iOS [ ] iPadOS [ ] Windows Browser [x] Chrome [x] Firefox ESR [ ] Firefox [ ] Safari [ ] Edge Note that https://github.com/nasa-jpl/www/issues/616 will need to be completed for the secondary nav to actually be used in the CMS
2025-04-01T04:34:51.292487
2022-01-18T20:39:05
1107341804
{ "authors": [ "astrogeco", "skliper" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8917", "repo": "nasa/cFE", "url": "https://github.com/nasa/cFE/pull/2027" }
gharchive/pull-request
Fix #2026, CFE_FS_ParseInputFileNameEx avoid uninit var Describe the contribution Fix #2026 Uses input buffer if default is NULL to set output to avoid uninitialized variables in unit test Testing performed Build/run unit tests with valgrind, eliminates warning Expected behavior changes Unit test only System(s) tested on Hardware: i5/Docker OS: Ubuntu 18.04 Versions: Bundle main + this commit Additional context I wonder if this would have been caught sooner if we had a different bucket for "critical" static analysis checks. Third party code None Contributor Info - All information REQUIRED for consideration of pull request Jacob Hageman - NASA/GSFC CCB:2022-01-19 APPROVED
2025-04-01T04:34:51.295844
2020-01-21T23:39:26
553214070
{ "authors": [ "astrogeco", "skliper" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8918", "repo": "nasa/cFS", "url": "https://github.com/nasa/cFS/pull/42" }
gharchive/pull-request
Integration Candidate 20200121 Describe the contribution Updated all branches to IC-20200121 Testing performed Current CI passes Enhanced CI in #40 passes debug/with deprecated elements testing Expected behavior changes See submodules System(s) tested on CI testing only Additional context N/A Code contibutions The cFS repository is provided to bundle the cFS Framework. It is utilized for bundling submodules, continuous integration testing, and version management and does not contain any software. Code contributions should be directed to the appropriate submodule. Contributor Info - All information REQUIRED for consideration of pull request Jacob Hageman - NASA/GSFC Changes directly applied to master. Closing PR
2025-04-01T04:34:51.364991
2022-03-30T21:52:55
1187032474
{ "authors": [ "AdrienVigne", "natelindev" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8919", "repo": "natelindev/tsdav", "url": "https://github.com/natelindev/tsdav/issues/109" }
gharchive/issue
How to create an adressBook with Davclient ? Hello, I am searching for a solution to create an addressbook thanks to the davclient client and sync it with the server. Could you help me please ? Thanks Just use makeCollection method with correct params and you should be good.
2025-04-01T04:34:51.372263
2022-07-25T17:23:14
1317141137
{ "authors": [ "Dimah4x", "Yehuda-Edelstein", "burak40" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8920", "repo": "nathan-149/automated_youtube_channel", "url": "https://github.com/nathan-149/automated_youtube_channel/issues/44" }
gharchive/issue
Skipping all accounts Facing this: <instaloader.nodeiterator.NodeIterator object at 0x000001B5B8213A90> Scraping From Account: funny_videos Skipped acc funny_videosbecause of user not found: 'funny_videos' Scraping From Account: funny.homevideos Skipped acc funny.homevideosbecause of user not found: 'funny.homevideos' Scraping From Account: try_not_to_laugh_funny_videos Skipped acc try_not_to_laugh_funny_videosbecause of user not found: 'try_not_to_laugh_funny_videos' Scraping From Account: cattoseverywhere Skipped acc cattoseverywherebecause of user not found: 'cattoseverywhere' Scraping From Account: cattobrat Skipped acc cattobratbecause of user not found: 'cattobrat' Scraping From Account: memerzvidz Skipped acc memerzvidzbecause of user not found: 'memerzvidz' Scraping From Account: fuckjerry.tv Skipped acc fuckjerry.tvbecause of user not found: 'fuckjerry.tv' Scraped Videos! Making Compilation... Total Length: 0 Traceback (most recent call last): File "C:\Users\Gebruiker\Desktop\automated_youtube_channel-master\main.py", line 144, in attemptRoutine() File "C:\Users\Gebruiker\Desktop\automated_youtube_channel-master\main.py", line 135, in attemptRoutine routine() File "C:\Users\Gebruiker\Desktop\automated_youtube_channel-master\main.py", line 100, in routine makeCompilation(path = videoDirectory, File "C:\Users\Gebruiker\Desktop\automated_youtube_channel-master\make_compilation.py", line 94, in makeCompilation finalClip = concatenate_videoclips(videos, method="compose") File "C:\Users\Gebruiker\AppData\Local\Programs\Python\Python39\lib\site-packages\moviepy\video\compositing\concatenate.py", line 75, in concatenate_videoclips w = max(r[0] for r in sizes) ValueError: max() arg is an empty sequence C:\Users\Gebruiker\Desktop\automated_youtube_channel-master> what instruction did you run? I ran python3 main.py and got that same error
2025-04-01T04:34:51.379373
2024-10-15T23:54:33
2590198467
{ "authors": [ "nathanjmcdougall" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8921", "repo": "nathanjmcdougall/usethis-python", "url": "https://github.com/nathanjmcdougall/usethis-python/issues/26" }
gharchive/issue
Set up a release pipeline Motivation We will want to release at some point. I'm keen to do it sooner rather than later, even if the repo is still private, because I'd like to start "eating my own dogfood", but the tool needs to run from a separate/isolate venv to avoid uninstalling itself while running. We can use uv tool run for this purpose but it looks like it only pulls from PyPI Summary of desired feature An automated release pipeline which publishes to PyPI. Have made the repo public but still need to sort the release pipeline at some point. I've decided that there are probably a few pre-requisites before tackling that. Nothing to stop some manual releases before that point, although there are some steps which need to be taken first, e.g. setting up project information in pyproject.toml.
2025-04-01T04:34:51.395214
2023-08-30T11:27:32
1873478439
{ "authors": [ "dragon-dxw" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8922", "repo": "nationalarchives/ds-caselaw-custom-api-client", "url": "https://github.com/nationalarchives/ds-caselaw-custom-api-client/pull/363" }
gharchive/pull-request
Add default user agent string to custom api client Set the default user agent string to give the version of the custom api client, and have it overriddable by users of the client. Add user-agent to APIClient
2025-04-01T04:34:51.398508
2022-09-19T08:59:52
1377630759
{ "authors": [ "alexfromk", "dragon-dxw" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8923", "repo": "nationalarchives/ds-caselaw-public-access-service", "url": "https://github.com/nationalarchives/ds-caselaw-public-access-service/issues/73" }
gharchive/issue
Atom API. Bug with parameter page greater than 2024 Hello, we are trying to work with National Archives API, with service Atom. We got the data by requests like this: https://caselaw.nationalarchives.gov.uk/atom.xml?page=1&order=-date All work fine with parameter "page" since 1 to 2024, but when page became greater 2024, service return HTTP 504. For example, GET https://caselaw.nationalarchives.gov.uk/atom.xml?page=2025&order=-date response with http 504. If it is possible, to fix this bug. With best regards. Alex Kononenko We had in fact scheduled today to look at that, and we've got a fix on staging, which should go live some day soon. You should also note significantly faster load times for the atom feed pages. That should be fixed on the live site now. Thank you for letting us know about the problem -- whilst we did know about it, it gives us confidence that we're working on fixing things that people care about! Hi, thank you very much. I tested it, paging above 2024 work fine. Alex Kononenko
2025-04-01T04:34:51.443702
2019-06-22T13:21:08
459473734
{ "authors": [ "gcolliso", "morrisio" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8924", "repo": "nats-io/docs", "url": "https://github.com/nats-io/docs/pull/34" }
gharchive/pull-request
Fix internal links Uppercase internal links don't work when clicked in the 'Protocol Messages' tables of Client Protocol and Cluster Protocol pages. Changing links to lowercase (e.g. #PUB -> #pub) should make them clickable again. Note: This problem does not occur on github.com, only when the files are rendered on github.io. Also removing a number HTML anchors which should be made obsolete by this PR and adding a number of missing internal links. lgtm - @morrisio thanks for the fix!
2025-04-01T04:34:51.446049
2019-04-12T18:33:16
432692630
{ "authors": [ "JensRantil", "derekcollison" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8925", "repo": "nats-io/gnatsd", "url": "https://github.com/nats-io/gnatsd/issues/946" }
gharchive/issue
Weird instructions to build https://github.com/nats-io/gnatsd/#build states Run go build inside the /nats-io/gnatsd directory. . However, there is no /nats-io/gnatsd in the repo. I assume that /main.go contains the main function so, I propose simply writing Run go build in root of the checked out repository. . It means the top of the repo directory, we will clarify the docs.. /cc @gcolliso
2025-04-01T04:34:51.462429
2015-09-26T14:11:03
108468212
{ "authors": [ "derekcollison", "ronag" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8926", "repo": "nats-io/node-nats", "url": "https://github.com/nats-io/node-nats/issues/38" }
gharchive/issue
Document callback signature in subscribe I keep going back to subscribe to remind myself what the signature of the callback should be and I keep missing it... Good idea.
2025-04-01T04:34:51.480034
2022-01-13T17:22:02
1102066343
{ "authors": [ "dutchshark", "lampwins", "nniehoff" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8927", "repo": "nautobot/nautobot", "url": "https://github.com/nautobot/nautobot/issues/1269" }
gharchive/issue
Add BGP Support to Nautobot Environment Python version: 3.7 Nautobot version: 1.2.3 Proposed Functionality I would like to model my BGP deployment in Nautobot. This should include ASNs as well as peering relationships at a minimum. Use Case As Nelly, a network engineer, I need to be able to model BGP configuration of my network in my source of truth to be able to completely automate the deployment of my network configuration. Database Changes Definitely, probably at least ASNs perhaps more. It's possible some of this may be better represented in a plugin. External Dependencies None Perhaps use Peering Manager to prevent reinventing the wheel? Closing with https://github.com/nautobot/nautobot-plugin-bgp-models
2025-04-01T04:34:51.485290
2021-05-31T04:48:28
907044458
{ "authors": [ "glennmatthews", "waydh" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8928", "repo": "nautobot/nautobot", "url": "https://github.com/nautobot/nautobot/issues/512" }
gharchive/issue
Cloning Device or IP with a tag applied results in a Server Error Environment Ubuntu 20.04 Python version: 3.8.5 Nautobot version: 1.0.2 Steps to Reproduce Install NautoBot 1.0.2 per the directions on https://nautobot.readthedocs.io Create a Site, Manufacturer, Device Role, Tag, and then a Device with all applied Click clone on the created device and receive a Server Error Remove the tag from the device (so there are no tags at all) click Clone and it works Issue also occurs with IPs Expected Behavior Clone button works without error allowing the device to be cloned Observed Behavior Server Error like the following - the invalid UUID is not always 'a', it changes alphabet or number Server Error There was a problem with your request. Please contact an administrator. The complete exception is provided below: <class 'django.core.exceptions.ValidationError'> ['“a” is not a valid UUID.'] Python version: 3.8.5 Nautobot version: 1.0.2 If further assistance is required, please post to the #nautobot channel on NetworkToCode's Slack. Thanks for the report! This may be related to #499.
2025-04-01T04:34:51.492733
2023-08-09T15:22:11
1843503222
{ "authors": [ "HanlinMiao", "glennmatthews" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8929", "repo": "nautobot/nautobot", "url": "https://github.com/nautobot/nautobot/pull/4233" }
gharchive/pull-request
Refine New UI Device Detail View Closes: #4013 What's Changed TODO [x] Re-establishing the advanced tab into the base page layout (Metadata: Key-value panel, fields in legacy UI (created, updated, UUID)) [x] Finish actions menu (expands to an Edit link to the legacy UI for the time being) [x] Device: Key-value panel, fields in legacy UI [x] Management: Key-value panel, fields in the legacy UI [x] Explanation of Change(s) [x] Added change log fragment(s) (for more information see the documentation) [x] Fixed incorrect related model url rendering in ObjectRetrieve view. [ ] Attached Screenshots, Payload Example [ ] Unit, Integration Tests [ ] Documentation Updates (when adding/changing features) [ ] Example Plugin Updates (when adding/changing features) [ ] Outline Remaining Work, Constraints from Design Looking at the first screenshot - definitely not blocking, but can we make the tags list wrap to a new line instead of scrolling horizontally when there are a ton of tags? Looking at the advanced tab, can we fix capitalization of "Id" and "Url"? Will make a follow-up story for more flexible rendering Looking at the advanced tab, can we fix capitalization of "Id" and "Url"? https://github.com/nautobot/nautobot/issues/4250 Looking at the first screenshot - definitely not blocking, but can we make the tags list wrap to a new line instead of scrolling horizontally when there are a ton of tags? https://github.com/nautobot/nautobot/issues/4251
2025-04-01T04:34:51.537316
2019-07-04T07:44:39
464115506
{ "authors": [ "donggyu04", "junoyoon" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8939", "repo": "naver/ngrinder", "url": "https://github.com/naver/ngrinder/pull/425" }
gharchive/pull-request
Fix failed tests [Merge into master] Use network interface before monitor collection test. Please use hotfix/** branch. Move to #430
2025-04-01T04:34:51.549589
2019-05-08T07:25:47
441589938
{ "authors": [ "RoySRose", "asahu4" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8940", "repo": "naver/pinpoint", "url": "https://github.com/naver/pinpoint/issues/5574" }
gharchive/issue
Please confirm you offer the enterprise support for pinpoint. Please confirm you offer the enterprise support for pinpoint. If yes please share more detail on support and charges . Thanks, Ajay Hello, @asahu4 I'm afraid we don't have any enterprise support business model, yet.
2025-04-01T04:34:51.551125
2020-02-17T08:32:50
566136540
{ "authors": [ "koo-taejin" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8941", "repo": "naver/pinpoint", "url": "https://github.com/naver/pinpoint/issues/6507" }
gharchive/issue
Fix problem that do not set some fields when an exception occurs related module Kafka-Plugin description If service type has record statistics, destnationId value have to send for to draw servermap. But, current ProducerSendInterceptor code do not set destinationId field when exception occurred while send() method invoked. For this reason, servermap has not drawn when when exception occurred while send() method invoked. This issue expect to have resolved via #6508, #6510
2025-04-01T04:34:51.565709
2019-05-02T06:06:10
439438295
{ "authors": [ "codecov-io", "koo-taejin" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8942", "repo": "naver/pinpoint", "url": "https://github.com/naver/pinpoint/pull/5550" }
gharchive/pull-request
[#4558] Add Stream function in profiler module using grpc Add Stream function in profiler module using grpc. Codecov Report Merging #5550 into master will decrease coverage by 0.07%. The diff coverage is 0%. @@ Coverage Diff @@ ## master #5550 +/- ## ========================================== - Coverage 40.8% 40.72% -0.08% ========================================== Files 2822 2826 +4 Lines 84907 85049 +142 Branches 11314 11328 +14 ========================================== - Hits 34644 34640 -4 - Misses 47239 47386 +147 + Partials 3024 3023 -1 Impacted Files Coverage Δ ...oint/profiler/receiver/grpc/GrpcStreamService.java 0% <0%> (ø) .../receiver/grpc/PinpointClientResponseObserver.java 0% <0%> (ø) ...er/receiver/grpc/GrpcActiveThreadCountService.java 0% <0%> (ø) ...xt/thrift/CommandThriftToGrpcMessageConverter.java 61.53% <0%> (-38.47%) :arrow_down: ...r/receiver/grpc/ActiveThreadCountStreamSocket.java 0% <0%> (ø) ...xt/thrift/CommandGrpcToThriftMessageConverter.java 31.25% <0%> (-40.18%) :arrow_down: .../profiler/receiver/grpc/GrpcCommandDispatcher.java 0% <0%> (ø) :arrow_up: ...navercorp/pinpoint/rpc/common/SocketStateCode.java 82.92% <0%> (-2.44%) :arrow_down: ...om/navercorp/pinpoint/rpc/codec/PacketDecoder.java 57.44% <0%> (-2.13%) :arrow_down: ...point/rpc/client/DefaultPinpointClientHandler.java 71.24% <0%> (-1.64%) :arrow_down: ... and 8 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 9e07e76...17bcff3. Read the comment docs.
2025-04-01T04:34:51.634232
2015-11-04T04:40:26
114968156
{ "authors": [ "melograph85", "mlogan" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8961", "repo": "ncb000gt/node.bcrypt.js", "url": "https://github.com/ncb000gt/node.bcrypt.js/issues/369" }
gharchive/issue
Consider XOR instead of equals for the comparator The guide Salted Password Hashing - Doing it Right claims that XOR is better than == for a slow-equals comparison function to mitigate timing attacks: The reason we need to use XOR instead of the "==" operator to compare integers is that "==" is usually translated/compiled/interpreted as a branch.... The branching makes the code execute in a different amount of time depending on the equality of the integers and the CPU's internal branch prediction state. This module's implementation currently uses the != operator. Indeed, clang compiles this to code that branches, which can result in variable-time execution: ## ../src/bcrypt_node.cc:240 next: cmp ebx, esi jl continue cmp eax, esi jl continue movzx edi, byte ptr [r15 + rsi] movzx ecx, byte ptr [rbp + rsi - 176] cmp ecx, edi je continue xor r8d, r8d continue: inc rsi cmp esi, edx jl next (The above is edited output from running node-gyp's build command on bcrypt_node.cc with the additional -S -masm=intel options.) The page mentioned has more information. I figured this might be worth bringing up. That page has a lot of bad advice, such as using SHA256 for hashing passwords. (See: https://www.nccgroup.trust/us/about-us/newsroom-and-events/blog/2007/july/enough-with-the-rainbow-tables-what-you-need-to-know-about-secure-password-schemes/) The particular suggestion of using timing attacks against password hashes is just bogus. Timing attacks work against HMACs because you can send anything you want as the signature. Once you guess the first byte, you hold that constant and work on the second byte. But you can't hold the first byte of the output of a cryptographic hash constant while probing for the second byte. If you could, you could just reverse the hash, and break most or all of modern crypto. (See: http://security.stackexchange.com/questions/9192/timing-attacks-on-password-hashes)
2025-04-01T04:34:51.656312
2018-10-17T16:35:20
371166039
{ "authors": [ "joe-graham", "saidelike" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8962", "repo": "nccgroup/asafw", "url": "https://github.com/nccgroup/asafw/issues/5" }
gharchive/issue
Serial debugging failing to start This issue was found on a physical ASA 5505 running version 9.1(6). When enabling gdbserver on the serial interface in asafw, the gdb script generated by asadbg hangs at target extended-remote [serial port]. Turning on gdb's remote debugging shows that gdbserver continually prints this string, preventing gdb from attaching: It looks like a fix for this would be modifying the inittab file to run /tmp/start_cmd on a different tty. This will still cause gdbserver to attach to the serial interface, but will result in the error message being printed on the other tty as opposed to over the serial interface. Manually making this change on my end results in asadbg hanging (since it's waiting for the "Remote debugging over /dev/ttyS0" string) but eventually connecting over USB: I'm willing to write the code to implement this change, but I'll hold off on implementing and making a pull request since this is a little more involved than my previous two issues. Thanks for this. Good finding about modifying /etc/inittab. So you effectively change: ttyS0::once:/tmp/run_cmd into ttyS1::once:/tmp/run_cmd? Automation could quite easily be tested by adding code to unpack_repack.sh in the custom function and invoking unpack_repack_bin.sh -c .... Once this works we could add a special case in the enable_gdb function for the firmware we know require that (e.g. "asa91"). Afaict most of the 9.1.x do require that. I think if we can find a way to have gdb not hang at all, it will be even better. Not sure if we can tell gdb to not wait for the "Remote debugging using /dev/ttyS0" string. Tbh I didn't even know it worked like that. What is interesting is that for asa924 for instance, we have this fix (note the comment which is not on asa91x): # run on the console specified on the kernel start line ::once:/tmp/run_cmd I am wondering if just changing to this would solve the gdb hanging problem. Yes, I agree it involves a bit more changes that your previous PR but at the same time is a good exercise to be able to modify the scripts (and not too complex in this case imho) so if that is something that you are interested in, feel free to do so. Note: In the past, we would use --bin-with-asa-to-inject to e.g. inject the /asa folder from 9.1.x into another firmware such as asa92*.bin but your way is a lot cleaner and does not change the firmware container. Oh right thanks for explaining that after your fix, the hang is on the asadbg side, not gdb side. The reason there is a hang is because we timeout after 3 minutes in case we don't detect the "Remote debugging using /dev/ttyS0" string which indeed we don't detect following your fix. We should be able to work around that but not sure how yet either. Good idea about the characters gdbserver may send over serial. I guess we could hexdump what we receive in comm.py instead of just printing ascii characters, to see if there is anything interesting. In case there isn't any character send, it may be possible to check from comm.py if there is a listening gdbserver on the other end by sending special characters. So in case we read 0 bytes, we could send this special series of bytes from comm.py and check we get the right reply from the other end (gdbserver). Definitely worth looking into. I'll get started on implementing the change for unpack_repack_bin.sh this weekend, and open an issue for the comm.py stuff over on the asadbg repo after I do a little more research into the gdbserver stuff. Thanks for your insight :)
2025-04-01T04:34:51.663488
2021-03-15T22:57:16
832264091
{ "authors": [ "nicl-nno", "pep8speaks" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8963", "repo": "nccr-itmo/FEDOT", "url": "https://github.com/nccr-itmo/FEDOT/pull/252" }
gharchive/pull-request
Better implementation of the caching cache moved from memory to disk using shelve library caching removed from nodes, now it is external object that can be used to 'fit' the nodes from pre-saved results Hello @nicl-nno! Thanks for opening this PR. We checked the lines you've touched for PEP 8 issues, and found: In the file fedot/core/composer/gp_composer/gp_composer.py: Line 21:121: E501 line too long (121 > 120 characters) In the file fedot/core/composer/optimisers/regularization.py: Line 52:32: E713 test for membership should be 'not in'
2025-04-01T04:34:51.667106
2016-09-09T12:33:25
175998431
{ "authors": [ "nchammas", "serialx" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8964", "repo": "nchammas/flintrock", "url": "https://github.com/nchammas/flintrock/pull/147" }
gharchive/pull-request
Fix assume_yes not existing in _create_instances I've encountered an error while doing test: $ python3 -m flintrock --config config.yaml launch serialx-test Requesting 3 spot instances at a max price of $0.311... 0 of 3 instances granted. Waiting... 0 of 3 instances granted. Waiting... 0 of 3 instances granted. Waiting... 0 of 3 instances granted. Waiting... 0 of 3 instances granted. Waiting... 0 of 3 instances granted. Waiting... ^CThere was a problem with the launch. Cleaning up... Canceling spot instance requests... Traceback (most recent call last): File "/Users/serialx/workspace/flintrock/flintrock/ec2.py", line 685, in _create_instances time.sleep(30) KeyboardInterrupt During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/Cellar/python3/3.5.1/Frameworks/Python.framework/Versions/3.5/lib/python3.5/runpy.py", line 170, in _run_module_as_main "__main__", mod_spec) File "/usr/local/Cellar/python3/3.5.1/Frameworks/Python.framework/Versions/3.5/lib/python3.5/runpy.py", line 85, in _run_code exec(code, run_globals) File "/Users/serialx/workspace/flintrock/flintrock/__main__.py", line 8, in <module> sys.exit(main()) File "/Users/serialx/workspace/flintrock/flintrock/flintrock.py", line 1034, in main cli(obj={}) File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/core.py", line 716, in __call__ return self.main(*args, **kwargs) File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/core.py", line 696, in main rv = self.invoke(ctx) File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/core.py", line 1060, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/core.py", line 889, in invoke return ctx.invoke(self.callback, **ctx.params) File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/core.py", line 534, in invoke return callback(*args, **kwargs) File "/Users/serialx/.virtualenvs/flintrock/lib/python3.5/site-packages/click/decorators.py", line 17, in new_func return f(get_current_context(), *args, **kwargs) File "/Users/serialx/workspace/flintrock/flintrock/flintrock.py", line 339, in launch instance_initiated_shutdown_behavior=ec2_instance_initiated_shutdown_behavior) File "/Users/serialx/workspace/flintrock/flintrock/ec2.py", line 46, in wrapper res = func(*args, **kwargs) File "/Users/serialx/workspace/flintrock/flintrock/ec2.py", line 878, in launch instance_initiated_shutdown_behavior=instance_initiated_shutdown_behavior) File "/Users/serialx/workspace/flintrock/flintrock/ec2.py", line 759, in _create_instances if not assume_yes: NameError: name 'assume_yes' is not defined This is an attempt to fix it. Looks good to me. @serialx - Did you test this to confirm that the behavior is as expected for both --assume-yes and --no-assume-yes? @nchammas To my knowledge, the code should work as expected. The codepath that triggers this bug is quite hard to reproduce, so this fix PR is not directly tested in that situation. Is there any place that you might want me to take extra testing? Ah sorry, you're right. This comes up only when creating new instances fails for whatever reason. Merging this in!
2025-04-01T04:34:51.676158
2018-12-03T10:04:00
386741300
{ "authors": [ "ArthurGing", "ncw" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8965", "repo": "ncw/rclone", "url": "https://github.com/ncw/rclone/issues/2812" }
gharchive/issue
Failed to create file system What is the problem you are having with rclone? Is there any way to refresh token of onedrive? I can't do anything after token expired (it needs 60 minutes) What is your rclone version (output from rclone version) 1.45 Which OS you are using and how many bits (eg Windows 7, 64 bit) linux 64 bit Which cloud storage system are you using? (eg Google Drive) onedrive The command you were trying to run (eg rclone copy /tmp remote:tmp) lsl, copy, move A log from the command with the -vv flag (eg output from rclone -vv copy /tmp remote:tmp) Failed to create file system for "arthur:": failed to get root: InvalidAuthenticationToken: CompactToken validation failed with reason code: 80049228. Then you wait > 1 hour and rclone gives you CompactToken validation failed with reason code: 80049228. Is that right? Feel free to XXX out anything that you think looks secret. What kind of onedrive are you using? Personal, Business, Sharepoint, other? If you do rclone config show arthur you should see your token, It should look something like this (it will all be one one line and the XXX are very long. Can you make sure it has a refresh_token in there? token = { "access_token":"XXX", "token_type":"Bearer", "refresh_token":"XXX", "expiry":"2018-11-29T16:10:46.368783193Z" } Thank you for the prompt reply. I guess Uploading files using 'copyto' may call api of cloud, so I am trying to get response with '--dump responses', but it show nothing. Is there any way to get response? Best regards I guess Uploading files using 'copyto' may call api of cloud, so I am trying to get response with '--dump responses', but it show nothing. Is there any way to get response? When I try that I see the response. Don't forget -vv! 2018/12/05 10:51:57 DEBUG : HTTP RESPONSE (req 0xc0004aeb00) 2018/12/05 10:51:57 DEBUG : HTTP/1.1 201 Created Content-Type: application/json; charset=utf-8 Date: Wed, 05 Dec 2018 10:51:57 GMT P3p: CP="BUS CUR CONo FIN IVDo ONL OUR PHY SAMo TELo" Strict-Transport-Security: max-age=31536000; includeSubDomains Www-Authenticate: Bearer realm="OneDriveAPI", error="invalid_token", error_description="Invalid auth token" X-Asmversion: UNKNOWN; 19.161.1128.2006 X-Msedge-Ref: Ref A: 48E0A333677C455EA636FA9904C82087 Ref B: LON21EDGE1014 Ref C: 2018-12-05T10:51:57Z X-Msnserver: AM3PPF2A98B041F {"createdBy":{"application":{"displayName":"rclone","id":"48211038"},"user":{"id":... Thank you for your help. I see the response after adding -vv, thanks a lot. Great! I think we are done here so I'll close this issue.
2025-04-01T04:34:51.682097
2019-02-28T18:23:27
415760423
{ "authors": [ "brprice", "ndmitchell" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8966", "repo": "ndmitchell/ghcid", "url": "https://github.com/ndmitchell/ghcid/issues/226" }
gharchive/issue
Polling ignores requested frequency The polling logic seems odd: we wait for the requested time only on the first iteration (in waitFiles) after each reload, and each subsequent one (in recheck) it waits a fixed 0.1. I suspect that we should sleep in WaiterPoll branch of recheck's case. I wonder if this is related to #182? #182 wasn't using polling, so shouldn't be related. You would also have seen the 0.1s blips in the CPU graph if it was sleeping. Reading the code carefully I think you're spot on though. There are two WaiterPoll branches, and I think they're the wrong way round 🤦‍♂️ . The corresponding WaiterNotify branch is first immediate, and then sleeping, so now I made WaiterPoll match that.
2025-04-01T04:34:51.699008
2018-06-18T08:29:59
333176356
{ "authors": [ "Steppschuh", "codecov-io" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8967", "repo": "neXenio/BLE-Indoor-Positioning", "url": "https://github.com/neXenio/BLE-Indoor-Positioning/pull/90" }
gharchive/pull-request
0075 render floor plan WIP Codecov Report Merging #90 into dev will decrease coverage by 0.02%. The diff coverage is 0%. @@ Coverage Diff @@ ## dev #90 +/- ## ============================================ - Coverage 35.28% 35.26% -0.03% Complexity 204 204 ============================================ Files 39 39 Lines 1278 1279 +1 Branches 123 123 ============================================ Hits 451 451 - Misses 775 776 +1 Partials 52 52 Impacted Files Coverage Δ Complexity Δ ...io/bleindoorpositioning/location/LocationUtil.java 78.12% <0%> (-2.53%) 7 <0> (ø) Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update e6c387d...f98e9c9. Read the comment docs. Implementation works, but still has issues (probably related to padding in the CanvasProjection)
2025-04-01T04:34:51.704592
2017-05-04T16:19:14
226335258
{ "authors": [ "JunjieW", "nea" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8968", "repo": "nea/MarkdownViewerPlusPlus", "url": "https://github.com/nea/MarkdownViewerPlusPlus/issues/32" }
gharchive/issue
CSS color is not rendered Hi, I was trying to add some customized CSS, but it wasn't rendered well. I'm not sure if it's caused by the HTMLRenderer or in this plugin. The exported html can be rendered as expected in Chrome, IE, Firefox. body { font-family: "Source Sans Pro",sans-serif; font-size: 13px; color: #3f3f3f } h1 { font-size: 25px; } h2 { font-size: 23px; } h3 { font-size: 21px; } h4 { font-size: 19px; } h5 { font-size: 16px; } h1, h2, h3, h4, h5, h6 { font-family: inherit; } blockquote { background: #f9f9f9; border-left: 10px solid #ccc; padding-top: 5px; padding-bottom: 5px; quotes: "\201C""\201D""\2018""\2019"; } code { font-family: Monaco, Consolas, "Andale Mono", "DejaVu Sans Mono", monospace; font-size: 95%; white-space: pre; white-space: pre-wrap; white-space: -moz-pre-wrap; white-space: -o-pre-wrap; background: rgb(200,200,200); display: inline-block; } pre code { font-family: Monaco, Consolas, "Andale Mono", "DejaVu Sans Mono", monospace; font-size: 95%; line-height: 140%; white-space: pre; white-space: pre-wrap; white-space: -moz-pre-wrap; white-space: -o-pre-wrap; background: #faf8f0; display: block; padding: 0.5em 1em; border: 1px solid #bebab0; } BTW there is another html rendering solution provided by this repo, I was trying to improve that plugin then found yours, while that one renders HTML without problem since it is using IE core, it doesn't provide functionality for HTML and PDF export, which is a very useful for me : ) ... Hmm, actually it works for CSS property background-color but not background Hey @JunjieW Thanks for the example. Yeah, I just the IE import originally but wanted to keep everything contained with as little external dependencies as possible, why I changed to the HTML Renderer. It is not the most active development but a very good basic library. But because of that, many things are probably just not implemented in the lib. I will check on the source and see why background-color works but not background and maybe I can create a pull-request. Cheers
2025-04-01T04:34:51.711576
2019-10-15T16:39:14
507359694
{ "authors": [ "janxter", "neagle" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8969", "repo": "neagle/gokibitz", "url": "https://github.com/neagle/gokibitz/issues/195" }
gharchive/issue
Bug kifu listing page hanging https://gokibitz.com/kifu has been hanging up in Firefox and Chrome on multiple computers for me over the last few days. Sorry to hear you're having trouble! Could you open your browser's console and paste any error messages you see there? The service is still up and running, and I haven't been able to reproduce an error on my end. I'm not getting anything in the Chrome console. After an initial "Page Unresponsive", it still reads "Waiting for GoKibitz.com...." an hour later while displaying either the background color or everything but the kifu list. Firefox prompts to wait or close, although I'll have to check the exact message on that system later. On one of the systems I've tried this on (Windows 10 Pro), the page loads just fine in Edge (which I've only used for this one purpose). This only started a couple of days ago and is taking place both at home and my office. If I ctrl-alt-del, kill Chrome, and then restore, it finally displays both the kifu list and comments, but I'm logged out. I reload, and it logs me back in but the problem returns. So I killed it, went to the main page, logged out, and was then able to view kifu. Once I log back in, the problem resumes. I'm user HarveyCrichton and I'm pretty sure the last game I uploaded was https://gokibitz.com/kifu/ByyM3ZeFr Hm—so you had two SGFs uploaded that were truncated at random points. (Example: "smog山人: if your opponent makes some mistake you can profit from it :). From move 1: D16 Q4 D4 F17 D14 C17 D17 D18 C16 E18 C18 B17 B18 B16 B15 \nsmog山人: if he connects solid you can still play hane if you want. From move 1: D16 Q4 D4 F17 D14 C17 D17 D18 C16 C18 B17 \nsmog山人: if he blocks you there, you're back to the joseki xd. From move 1: D16 Q4 D4 F17 D") The malformed SGFs were causing a surprisingly complete meltdown of functions that tried to retrieve and parse them. Obviously this shows that the uploader is insufficiently guarded against invalid SGF files: I'll see if I can address that on my side. Out of curiosity, were you cutting/pasting the text of your SGFs? Is there any chance the truncation was caused by user error? (You should be able to log back in now. I've marked the two invalid SGFs as deleted.) As it happens, I just linked to the URL instead of downloading from OGS and uploading to GK from my computer for the last few games. Maybe some sketchy network issue took place. I likely hand-edited text in the game after it was in GK, but that seems unlikely to impact the sgf in your db. Hmmm—maybe the hand-editing thing could be a point of vulnerability. Another thing to investigate. If I ever get the chance to do GoKibitz 2.0, I'm going to store SGFs in the DB in a less fragile way. Thanks for your excellent debugging; sorry again for the inconvenience. Let me know if you have any other issues!
2025-04-01T04:34:51.713485
2021-11-12T00:42:04
1051474801
{ "authors": [ "brianzzzasd", "dimfeld", "glennmichael123" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8970", "repo": "near/borsh-js", "url": "https://github.com/near/borsh-js/pull/40" }
gharchive/pull-request
Change TextDecoder resolution to not rely on global global does not exist in the browser without Node polyfills, and we shouldn't assume their presence. Fixes #38 @volovyk-s thank you thank you thank you! you have no idea how merging this fix solved a lot of problems for me 😭 @volovyk-s thank you sir 🙇
2025-04-01T04:34:51.741625
2022-02-22T11:29:44
1146792703
{ "authors": [ "ntotten", "radomird" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8971", "repo": "nearform/get-jwks", "url": "https://github.com/nearform/get-jwks/pull/115" }
gharchive/pull-request
Added type definitions and tests for types This PR is a continuation of the work done by @ntotten in this PR: https://github.com/nearform/get-jwks/pull/108 This PR contains all the work from https://github.com/nearform/get-jwks/pull/108 + I've added tests for types. Closes #109 Hey, thanks for finishing this. I just came back to do this and saw it is already released. Thanks again! @ntotten no problem 😃
2025-04-01T04:34:51.744072
2019-11-26T07:08:03
528531915
{ "authors": [ "AnaisUrlichs", "icerove" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8972", "repo": "nearprotocol/near-explorer", "url": "https://github.com/nearprotocol/near-explorer/issues/131" }
gharchive/issue
NearKat blocking the support section Story As a user looking at the explorer, and don't get what's going on, I would like to get some help. But "Oh no!" The NearKat is blocking the support section. Acceptance Criteria [x] The user can read the support section clearly [x] NearKat does not reduce accessibility. (It's a malicious NearKat) fix it in the PR and wait to be merged
2025-04-01T04:34:51.748076
2019-11-15T08:19:14
523321798
{ "authors": [ "icerove", "vgrichina" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8973", "repo": "nearprotocol/near-shell", "url": "https://github.com/nearprotocol/near-shell/pull/189" }
gharchive/pull-request
Fix tx-status to handle account id better Fixes https://github.com/nearprotocol/near-shell/issues/187 It is really a goo idea to separate the command when you fix it. Please also update the readme, I think it is not including the new command you added. It is really a goo idea to separate the command when you fix it. Seemed like good opportunity to do this. If command was more complex I'd split the change in 2 separate (first move, then fix). Please also update the readme, I think it is not including the new command you added. I didn't add any new command, not sure what do you mean. near generate-key and near repl is not added. Also, still not clear about accountid problem I left above? near generate-key and near repl is not added. these aren't part of this PR near generate-key and near repl is not added. Also, still not clear about accountid problem I left above? https://github.com/nearprotocol/near-shell/commit/1c139dd33cb5ec0cd15cbf8e6070efdc21e8d753
2025-04-01T04:34:51.766752
2023-11-27T02:18:06
2011360874
{ "authors": [ "aktech", "dharhas", "nkaretnikov" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8974", "repo": "nebari-dev/argo-jupyter-scheduler", "url": "https://github.com/nebari-dev/argo-jupyter-scheduler/pull/3" }
gharchive/pull-request
Generate HTML output and send it to Slack Reference Issues or PRs What does this implement/fix? Put a x in the boxes that apply [ ] Bug fix (non-breaking change which fixes an issue) [x] New feature (non-breaking change which adds a feature) [ ] Breaking change (fix or feature that would cause existing features not to work as expected) [ ] Documentation Update [ ] Code style update (formatting, renaming) [ ] Refactoring (no functional changes, no API changes) [ ] Build related changes [ ] Other (please describe): Testing [ ] Did you test the pull request locally? [ ] Did you add new tests? Documentation Access-centered content checklist Text styling [ ] The content is written with plain language (where relevant). [ ] If there are headers, they use the proper header tags (with only one level-one header: H1 or # in markdown). [ ] All links describe where they link to (for example, check the Nebari website). [ ] This content adheres to the Nebari style guides. Non-text content [ ] All content is represented as text (for example, images need alt text, and videos need captions or descriptive transcripts). [ ] If there are emojis, there are not more than three in a row. [ ] Don't use flashing GIFs or videos. [ ] If the content were to be read as plain text, it still makes sense, and no information is missing. Any other comments? Summary: adds a send_to_slack step to scheduled and one-time workflows it uses the Slack API to send HTML output to a specified Slack channel added a call to jupyter nbconvert to generate HTML configured via Parameters SLACK_TOKEN and SLACK_CHANNEL in Notebook Jobs in the web UI, which are accessible via envs in the code see the Slack API docs on how to configure a bot to send a file to a channel -- this needs to be done first for the bot/sending functionality to work this new step is integrated with update_job_status_failure, so it will be visible in the UI if it fails the Slack script also has some printing and additional validation, so an exception will be raised on failure, which will cause the job to fail cmd_args generation is changed because (1) two commands are now called there and (2) it's passed to /bin/sh as a string anyway, so no point in keeping that in a list changed *path functions to return Path objects since that's more flexible, in case callers want to modify these paths. So how will this be configured? i.e. "send to slack" is not a feature that all nebari / jupyter-scheduler users will need. Also someone else might want to send it to mattermost or another rest api. Is there a way to make this a bit more generic. @dharhas So how will this be configured? i.e. "send to slack" is not a feature that all nebari / jupyter-scheduler users will need. Currently, it'll only execute this task if you provide SLACK_TOKEN and SLACK_CHANNEL as Parameters when scheduling the notebook. If you don't provide this, nothing will be sent. Is there a way to make this a bit more generic. Technically, we can turn this into "specify a random shell command and I'll execute it", but I don't think it's a good design. Users might run into issues with string escaping This prevents us from doing API-specific checking of whether the request was successful or not. I'd suggest we add support for additional APIs separately, on a case by case basis. Is there an example (a screenshot maybe) of slack output in a channel or something? Also, would be nice to add some docs for this. @nkaretnikov lets add docs also I think we need to make sure runs are timestamped. Are they also saved to disk as well as sent to slack? "send to slack" needs to be optional. Is there an example (a screenshot maybe) of slack output in a channel or something? Slack previews HTML as source code here. I think they don't render it by default for security reasons. I've looked and I'm not sure there's a way to render it. Once you download it, it's valid HTML. @aktech I've tested and reviewed this. PTAL @aktech PTAL. Made the changes you suggested, added more info to the internals section of README. Tested to make sure it's working and the backtraces are logged to a file. I went ahead and merged this since it'd be nice to have as part of the current Nebari release, see https://github.com/nebari-dev/nebari/issues/2195#issuecomment-1886155081.
2025-04-01T04:34:51.768778
2022-10-05T11:36:22
1397651943
{ "authors": [ "pavithraes" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8975", "repo": "nebari-dev/nebari-demo", "url": "https://github.com/nebari-dev/nebari-demo/issues/5" }
gharchive/issue
How will participants log-in to Nebari? We'd like the participants at PyData NYC to get hands-on time with Nebari. This issue is to track the logistics behind providing participants with login details. @viniciusdc had noted that we can configure Keycloak to allow random passwords. Done: https://github.com/nebari-dev/nebari-demo#login-for-the-first-time
2025-04-01T04:34:51.772080
2024-10-11T14:41:01
2581549183
{ "authors": [ "MatteoGheza", "nebojsatomic" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8976", "repo": "nebojsatomic/IDE21", "url": "https://github.com/nebojsatomic/IDE21/issues/53" }
gharchive/issue
Add language Italian I'm working on adding the italian translation, to both src/legacy/dev-application/languages/en.php and src/legacy/languages/en.php. @MatteoGheza Thank you for participation, but at the moment, only work on src/dev-application/languages/creator/it.php by updating the current translation, if some strings are not translated properly, or add src/dev-application/languages/it.php file, copy and paste the contents of sr.php from the same folder, and translate in that file. Do not touch src/legacy/dev-application/languages/en.php and src/legacy/languages/en.php. you have the translation guide already written in repo's Readme.md Thank you for the feedback, and sorry for the double PR. Have a nice day. @MatteoGheza feel free to contact me on linkedin also, did you check the it.php file in creator folder? are all the strings translated properly?
2025-04-01T04:34:51.818272
2022-07-17T00:45:25
1306950801
{ "authors": [ "Ahmad-A0", "ErdemOzgen", "anjola-adeuyi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8977", "repo": "neetcode-gh/leetcode", "url": "https://github.com/neetcode-gh/leetcode/pull/509" }
gharchive/pull-request
Remove All Adjacent Duplicates in String II - python and JavaScript Solution using stack data structure in python and javascript Thanks, @anjola-adeuyi! Thanks @Ahmad-A0
2025-04-01T04:34:51.820886
2022-08-28T17:35:28
1353421107
{ "authors": [ "Ahmad-A0", "MaratKhakim" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8978", "repo": "neetcode-gh/leetcode", "url": "https://github.com/neetcode-gh/leetcode/pull/958" }
gharchive/pull-request
Kotlin: 124. Binary Tree Maximum Path Sum File(s) Modified: 124-Binary-Tree-Maximum-Path-Sum.kt. Language(s) Used: Kotlin Submission URL: _https://leetcode.com/submissions/detail/785676279/_ Thanks, @MaratKhakim!
2025-04-01T04:34:51.836373
2017-08-24T15:40:46
252643924
{ "authors": [ "ghost", "negokaz" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8979", "repo": "negokaz/vscode-live-server-preview", "url": "https://github.com/negokaz/vscode-live-server-preview/issues/2" }
gharchive/issue
Doesn't Display locally saved graphics and pictures When I add a picture to my Code which lies locally on some Drive it isn't displayed. Closed this issue because lacking information to resolve.
2025-04-01T04:34:51.900828
2021-03-22T12:51:35
837686391
{ "authors": [ "cryptoBeliever", "kodty" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8980", "repo": "nemgrouplimited/symbol-desktop-wallet", "url": "https://github.com/nemgrouplimited/symbol-desktop-wallet/issues/1358" }
gharchive/issue
Harvesting - Wallet not showing income from 25% harvester service fee Reproduce: set up a node leaving automatic harvesting on set the beneficiary account to be the same address as the node account credit the node account (which is already converted to multisig) with some xym and allow it to harvest go to the wallet and select the cosigner account open the harvesting page of the wallet and select the multisig account above In the list of blocks harvested, it will show the income from the 75%, but the 25% beneficiary income does not show in the wallet while it does show in the explorer. Fixed in integration branch. Before: After fix:
2025-04-01T04:34:51.903542
2023-10-23T18:52:38
1957817055
{ "authors": [ "igorssilva" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8981", "repo": "nemo-ufes/frameweb-vp-plugin", "url": "https://github.com/nemo-ufes/frameweb-vp-plugin/pull/44" }
gharchive/pull-request
36 generate code from application model Closes #36 Add: Support for generation of code for application classes and interfaces; New default templates. Fix: Template errors Fixed some template errors and added a more robust template validation on imports.