id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
456554659
Upload more files at the same time Is it possible to upload more files at once? Like adding them to a queue? The queue is there, but it will upload one file at a time.
gharchive/issue
2019-06-15T16:41:33
2025-04-01T06:45:02.124427
{ "authors": [ "moisespr123", "sproott" ], "repo": "moisespr123/GoogleDriveUploadTool", "url": "https://github.com/moisespr123/GoogleDriveUploadTool/issues/16", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
126213728
[Issue#12] Arbitrary up-to-date checker MOJOs Initial implementation, documentation and integration tests for Issue#12. Please pull these changes and release a new build containing these enhancements. Thanks! 12 weeks since PR raised and successful test build - any chance of this getting accepted? Karl, I looked at your link but the page makes no mention of any Maven plug-in. I checked out the source from SVN and couldn't find one. So I've rebased to leave you with a single, conflict-less commit to review and hopefully merge. Thanks, --A Karl, any chance we can get this merged without further ado? I'm now taking a lot of heat over the inordinate delays in getting these contributions accepted and this is seriously undermining the case I've been passionately making for us to do the honourable thing, engage with the Open Source communities and make useful contributions rather than doing private forks that benefit no-one. Thanks! Thank you for the merge, Karl. @demonfiddler, I know it has been a while, but other than the ITs in this project, I found no other source on the WWW showing any examples of this plugin goal. Are you sure that this is actually working as intended? All ITs just execute validate, none of them actually uses any of the generated properties later in the build. I know, the log output is checked after the ITs, but maybe that is not enough. Either way, I am having problems using this goal. Somehow, it always evaluates to the same property value. Maybe I misunderstand the documentation, but somehow I cannot get the evaluation to return value, only else. I know that this is not Stack Overflow, but I am not super confident that anyone would find it there, given the fact that the plugin goal seems to be rarely used by anyone. Alexander, Please forgive the lack of response to your inquiry—I’ve just found it languishing along with 2,400+ unread emails. It was indeed a long time ago (eight years) that I worked on this code, and I have since then retired. You raise an interesting question about how the tests work, only checking the log file. I don’t recall how I arrived at this pattern but at a guess it’s likely I a copied a pattern already in use elsewhere in the existing code. It certainly sounds like a more robust assertion would check the property value itself. All I can say for sure is that at TIBCO we were using the plug-in goal in production builds but your query makes me wonder whether it was in fact working correctly. If it isn’t that would be a great embarrassment to myself, for as you have seen, I thought I had gone to great lengths to ensure comprehensive test coverage. Looking at src/main/java/org/codehaus/mojo/buildhelper/AbstractUpToDatePropertyMojo.java lines 110–113, it looks like the code to set the up-to-date or out-of-date property value is present, so the question reduces to whether that code is working correctly. I am now extremely busy on private projects and unfortunately don’t really have the time to spend on this, for which I offer you my sincerest apologies. Might I suggest that you submit a GitHub issue with a reproducer test? If the functionality is still of interest to you, perhaps you could consider doing a PR? I don’t know whether @khmarbaise is still active on the project. I wish you well, Adrian Price From: Alexander Kriegisch @.> Sent: 21 December 2023 09:18 To: mojohaus/build-helper-maven-plugin @.> Cc: Adrian Price @.>; Mention @.> Subject: Re: [mojohaus/build-helper-maven-plugin] [Issue#12] Arbitrary up-to-date checker MOJOs (#13) @demonfiddler https://github.com/demonfiddler , I know it has been a while, but other than the ITs in this project, I found no other source on the WWW showing any examples of this plugin goal. Are you sure that this is actually working as intended? All ITs just execute validate, none of them actually uses any of the generated properties later in the build. I know, the log output is checked after the ITs, but maybe that is not enough. Either way, I am having problems using this goal. Somehow, it always evaluates to the same property value. Maybe I misunderstand the documentation, but somehow I cannot get the evaluation to return value, only else. I know that this is not Stack Overflow, but I am not super confident that anyone would find it there, given the fact that the plugin goal seems to be rarely used by anyone. — Reply to this email directly, view it on GitHub https://github.com/mojohaus/build-helper-maven-plugin/pull/13#issuecomment-1865918697 , or unsubscribe https://github.com/notifications/unsubscribe-auth/ADK47OZ5UXBZJWEZRMYQXMTYKP5FPAVCNFSM4BYNC7VKU5DIOJSWCZC7NNSXTN2JONZXKZKDN5WW2ZLOOQ5TCOBWGU4TCOBWHE3Q . You are receiving this because you were mentioned.Message ID: @.***> @demonfiddler, I am trying to remember the context in which I tried to use these mojos, but I cannot even find the original message you quoted in your e-mail response. Did I comment on this PR, the corresponding issue, lines of code in a commit or elsewhere?In one of the OSS projects I contribute to, I must have used a workaround due to the lack of response here, but if I had managed to solve the problem using Build Helper, I would have commented here. I’m not entirely sure how you entered the dialogue. I was merely responding to an email from GitHub, linked to my PR for this functionality. From: Alexander Kriegisch @.> Sent: 20 May 2024 12:58 To: mojohaus/build-helper-maven-plugin @.> Cc: Adrian Price @.>; Mention @.> Subject: Re: [mojohaus/build-helper-maven-plugin] [Issue#12] Arbitrary up-to-date checker MOJOs (#13) @demonfiddler https://github.com/demonfiddler , I am trying to remember the context in which I tried to use these mojos, but I cannot even find the original message you quoted in your e-mail response. Did I comment on this PR, the corresponding issue, lines of code in a commit or elsewhere?In one of the OSS projects I contribute to, I must have used a workaround due to the lack of response here, but if I had managed to solve the problem using Build Helper, I would have commented here. — Reply to this email directly, view it on GitHub https://github.com/mojohaus/build-helper-maven-plugin/pull/13#issuecomment-2120312684 , or unsubscribe https://github.com/notifications/unsubscribe-auth/ADK47O4F52GTTZ3NSSYJEC3ZDHQK3AVCNFSM4BYNC7VKU5DIOJSWCZC7NNSXTN2JONZXKZKDN5WW2ZLOOQ5TEMJSGAZTCMRWHA2A . You are receiving this because you were mentioned.Message ID: @.***>
gharchive/pull-request
2016-01-12T16:11:57
2025-04-01T06:45:02.158804
{ "authors": [ "demonfiddler", "kriegaex" ], "repo": "mojohaus/build-helper-maven-plugin", "url": "https://github.com/mojohaus/build-helper-maven-plugin/pull/13", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1197296147
Will this work on Monterey 12.3.1 I have a hackintosh with opencore Monterey 12.3.1 also using whatevergreen.kext and Nvidia kepler patcher. Will this work on my machine? I cannot open simulator on xcode. It is just a black screen Nope, this will not fix your problem, sorry. This extension is obsolete, WhateverGreen replaces it.
gharchive/issue
2022-04-08T13:06:48
2025-04-01T06:45:02.189150
{ "authors": [ "AK-3696", "mologie" ], "repo": "mologie/NVWebDriverLibValFix", "url": "https://github.com/mologie/NVWebDriverLibValFix/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
931750735
[Solved] Executable requires libtiff.dylib to be installed (v0.1.0.dev) Problem For our 0.1.0 release I built a universal binary and bundled it into an appbundle for distribution. I then discovered that the application didn't run on Macs with x86_64 architecture. The error report suggested that the issue was with a dynamic library that was not installed on the machine. Library not loaded: /usr/local/opt/libtiff/lib/libtiff.5.dylib Referenced from: /Users/USER/Downloads/MoloVol.app/Contents/MacOS/MoloVol Reason: image not found Using otool on my executable I was getting this output: $ otool -L MoloVol.app/Contents/MacOS/MoloVol MoloVol.app/Contents/MacOS/MoloVol: ... /usr/local/opt/libtiff/lib/libtiff.5.dylib (compatibility version 13.0.0, current version 13.0.0) I discovered that installing libtiff using Homebrew fixed the app crashing. However, it would be unreasonable to expect users to install unrelated libraries for the app to work. I was confused about why I was encountering the issue in the first place. I had built wxWidgets specifically as static library, so I wasn't expecting dylib dependencies to pop up. Solution Eventually I discovered a solution on this web page (I've archived a snap shot of the page with the Wayback Machine if the link ever breaks). I had to change the relevant section a little, but I added the following code to my Makefile: WXCONFIGLIBS := $(shell wx-config --libs) WXCONFIGLIBS := $(WXCONFIGLIBS:-ltiff=/usr/local/opt/libtiff//lib/libtiff.a) LDFLAGS := $(WXCONFIGLIBS) This section statically links my executable against libtiff. All that was left to do was add $(LDFLAGS) to the linking line in my Makefile in place of `wx-config --libs`. Inspecting the resulting executable with otool confirms that the executable no longer asks for the dylib. Opening the app on the x86_64 machine works as desired. Alternative Solution Before I discovered how to statically link libtiff, I constructed a whole working solution by bundling the libtiff with the appbundle. I believe this is even explained in the resource I've linked above. In essence, the idea is to place all required dylibs inside the bundle directory. Most places I found created a Frameworks directory inside the Contents directory. In my case I needed two dylibs in this directory: libtiff.5.dylib and libjpeg.9.dylib because the latter was a dependency of the former. As part of the appbundle creation I needed to copy the dylibs into the Frameworks folder, then inform the executable to look for the libtiff library there, instead of the usual place. The relevant make file section looked something like this: BUNDLE := bin/MoloVol.app LIBNAME_LIBTIFF := libtiff.5.dylib LIBDIR_LIBTIFF := /usr/local/opt/libtiff/lib LIBNAME_JPEG := libjpeg.9.dylib LIBDIR_JPEG := /usr/local/opt/jpeg/lib FWDIR := $(BUNDLE)/Contents/Frameworks FWDIRREL := @executable_path/../Frameworks ... appbundle: ... cp $(LIBDIR_LIBTIFF)/$(LIBNAME_LIBTIFF) $(FWDIR) cp $(LIBDIR_JPEG)/$(LIBNAME_JPEG) $(FWDIR) install_name_tool -change "$(LIBDIR_LIBTIFF)/$(LIBNAME_LIBTIFF)" "$(FWDIRREL)/$(LIBNAME_LIBTIFF)" $(BUNDLE)/Contents/MacOS/MoloVol install_name_tool -change "$(LIBDIR_JPEG)/$(LIBNAME_JPEG)" "$(FWDIRREL)/$(LIBNAME_JPEG)" $(FWDIR)/$(LIBNAME_LIBTIFF) In order to make the executable look in the right place for the dylib, I had to use install_name_tool with the -change switch as seen above. The first argument is the original dylib path, the second argument is the new path relative to the executable and the last argument is the executable path. Since libtiff also had a dependency (libjpeg), I needed to do this twice, but the second time with libtiff as executable. Using otool -L path/to/executable I could check whether the dylib path was successfully changed. There are many places that explain how to do this, but I did find parts of this guide to be most useful. Ultimately, the problem with this approach is that now we're distributing not only our code, but also third-party binaries. This leads to all kinds of licensing requirements that are best left untouched. Failed Attempts Before I got to any solution, I tried to create fat/universal binaries of the dylibs so I could include them in the appbundle. However, this ended up not being necessary, since I could just use the binaries already installed on my system. For whatever reason, only the x86_64 build requires the dylib, whereas the arm64 build manages to ignore it when linking. That means no universal libraries are required. After my original solution stopped working, I came back to this issue and finally found a proper solution. The solution comes from this 2014 blog post on the wxWidgets website. This is the relevant quote: [...] sometimes you may want to avoid using the system libraries, e.g. because you want to build a statically linked version of your program with as few dependencies as possible. This was always possible by explicitly disabling the use of each and every library with -with-libfoo=builtin, but this was relatively tiresome and it was easy to forget a library or two. To remedy this, I’ve just added a new --disable-sys-libs configure option which does exactly what it says: when it is specified on configure command line, only the built-in versions of the third party libraries will be used. So the solution is simply to use the option --disable-sys-libs when configuring wxWidgets. No more dependencies on local dylibs.
gharchive/issue
2021-06-28T16:33:24
2025-04-01T06:45:02.203345
{ "authors": [ "jmaglic" ], "repo": "molovol/MoloVol", "url": "https://github.com/molovol/MoloVol/issues/96", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
273525327
Getting last Tuesday... without if conditions I'm trying to get last X day of either current week or previous week... just like php strotime("last tuesday" Should return the date for the Tuesday last week if you're on the Monday of this week or if it's now Wed you should get Tuesday of this week... I'm not sure how to do this without if statements of figuring out if you've passed a specific day of the week already. See the bigger discussion in #2522
gharchive/issue
2017-11-13T18:09:37
2025-04-01T06:45:02.205195
{ "authors": [ "icambron", "raptor235" ], "repo": "moment/moment", "url": "https://github.com/moment/moment/issues/4304", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1480190910
morgan1bathe@Momentum-automation-center Created By morgan1bathe from @Momentum-automation-center @ArthusLiang Hi Arthus! I see the Circleci tests failed again. Is this a cause for concern or should I push this through? @ArthusLiang based on the failed checks, is there anything I can do?
gharchive/pull-request
2022-12-06T21:10:40
2025-04-01T06:45:02.206721
{ "authors": [ "ArthusLiang", "morgan1bathe" ], "repo": "momentum-design/momentum-ui", "url": "https://github.com/momentum-design/momentum-ui/pull/1404", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1470907878
Misc cleanup, polish Adds Pastel for console colorization Adds more error spew if something goes wrong Remove dup comments/dead code Add barebones usage section Changed syntax, lumper will now accept an infile bsp and dump an outfile bsp specified on console Since the old Program.cs is gone now, should be just close this or do you want to add some of it to the Lumper.UI? The args stuff would be nice in the UI project. I'll go ahead and stick it in UI
gharchive/pull-request
2022-12-01T08:43:05
2025-04-01T06:45:02.208910
{ "authors": [ "BIRD311", "sapphonie" ], "repo": "momentum-mod/lumper", "url": "https://github.com/momentum-mod/lumper/pull/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2044745694
⚠️ haccp has degraded performance In effe098, haccp (https://www.haccp.or.kr) experienced degraded performance: HTTP code: 400 Response time: 5604 ms Resolved: haccp performance has improved in 310ea57 after 10 minutes.
gharchive/issue
2023-12-16T13:14:19
2025-04-01T06:45:02.212378
{ "authors": [ "momo31" ], "repo": "momo31/upptime", "url": "https://github.com/momo31/upptime/issues/536", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
724395820
Create option for a script that loads from file Currently for my dev workflow, I install the example.meta.js file and then manually add this line to the file via the browser editor: // @require file:///path\to\dist\example.user.js Can you create a new option that automatically generates a *.load-local.js (can be any name) file that's basically the meta file and that require line? I'm not using the proxy script option for dev because it requires 2-3 reloads after every change vs 1 reload when loading from file (on Tampermonkey/Chrome). I also can't specify the @require in the headers option because it'll get included in the user.js file as well, causing it to require itself. 1 reload when loading from file (on Tampermonkey/Chrome) If it is true, it's sure a great news!!! Thank you for the information! I think the proxy script feature can be extended to support local file:// protocol in order to make use of the 1-reload behavior. Yea, it seems Tampermonkey doesn't cache local files so it gets refreshed every page load I just need to make sure webpack dev server writes to file devServer: { writeToDisk: true, }, plugins: [ new WebpackUserscript(...) ] and enable "Allow access to file URLs" under Extensions -> Tampermonkey you don't have to do this manually, just use baseUrl: require('url').pathToFileURL(<devServer.contentBase>).href in your proxyScript config you don't have to do this manually, just use baseUrl: require('url').pathToFileURL(<devServer.contentBase>).href in your proxyScript config Thanks! I'm able to resolve this with this config: import url from 'url' const baseUrl = url.pathToFileURL(path.resolve(__dirname, 'dist')).href export default { devServer: { writeToDisk: true, inline: false, }, output: { filename: 'example.user.js', }, plugins: [ new WebpackUserscript({ proxyScript: { enable: (process.env.NODE_ENV === 'development'), baseUrl: baseUrl, }, }), ], } And then installing http://localhost:8080/example.proxy.user.js
gharchive/issue
2020-10-19T08:27:53
2025-04-01T06:45:02.217261
{ "authors": [ "Trinovantes", "momocow", "tao-cumplido" ], "repo": "momocow/webpack-userscript", "url": "https://github.com/momocow/webpack-userscript/issues/40", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
588061825
NTRs related to COVID-19/coronavirus diagnostics Looking to add some new terms regarding coronavirus diagnostic screenings and evaluations: From a general perspective we may benefit from a term such as, term: coronavinae infectious disease screening ( refers to MONDO_0005719) parent: disease screening definition: Screening for virus diseases caused by the coronavirus genus. The following terms could be considered children of 'coronavinae infectious disease screening' or we may choose to put them under separate parents. I've suggested some options for both. Long and short, these are some pieces folks have been using for initial coronavirus diagnosis. term: screening for travel related infection (open to other term name options here!) parent: coronavinae infectious disease screening OR disease screening definition: Evaluation of a patient's recent domestic or international travel history, often conducted alongside presentation of acute illness. Ref:https://www.doh.wa.gov/ForPublicHealthandHealthcareProviders/HealthcareProfessionsandFacilities/HealthcareAssociatedInfections/InfectionControlAssessmentandResponse/TravelrelatedInfectionScreening. (still looking for publications to help define this term better, many relate to screening asymptomatic individuals in the airports following travel) term: radiography of the chest parent: 'therapeutic imaging procedure' (MAXO_0000005), but we likely need a broad radiography imaging procedure term there in general (see below) definition: Radiographic imaging of the chest. synonyms: chest x-ray, chest xray, chest radiography, chest radiograph xref: NCIT:C38103 note, this seems like something we should have, am I just missing it somewhere? term: radiography imaging procedure parent: therapeutic imaging procedure definition: Conventional radiography produces a single image of an object by measuring the attenuation of an x-ray beam passing through it. xref: NCIT:C38101 Ref: Wernick, M. N., Wirjadi, O., Chapman, D., Zhong, Z., Galatsanos, N. P., Yang, Y., ... & Muehleman, C. (2003). Multiple-image radiography. Physics in Medicine & Biology, 48(23), 3875. For more of the diagnostic tools utilized for evaluating specimens: Term: Real-time PCR specimen test. (not certain if 'test' is the most appropriate choice, what other options might we have?) Parent: clinical laboratory procedure Definition: A sensitive DNA analysis tool that can be used for genotyping, quantifying viral load, and gene copy number assays for specimens including whole blood, plasma, serum, and sputum. In RT-PCR, an RNA transcript is reverse transcribed into cDNA and amplified to qualitatively detect gene expression. Synonyms: quantitative PCR, qPCR, RT-PCR, real-time polymerase chain reaction xRefs: MMO:0000655, OBI_0000911 References: https://doi.org/10.1016/B978-0-12-405914-6.00003-2 , https://doi.org/10.1016/B978-0-12-802211-5.00009-6 I'll keep scoping around to see what other approaches folks are using. +1 but Minor comment: radiography of the chest is not a 'therapeutic imaging procedure' screening for travel related infection => This is probably a suboptimal term because it combines a procedure with a role/intention. coronavinae infectious disease screening => Perhaps simply "infectious disease screening" is sufficient. I am worried that we will wind up with one term for each microbe. Alternatively, consultation with infectious disease physician? maybe the issue is the 'therapeutic imaging procedure' label? Maybe we could make that more generic or call it 'therapeutic or diagnostic imaging procedure'? The only reason I used the therapeutic in the label is to distinguish it from general microscopy that might not be specific for clinical use. FTR, I am trying to compile some terms here. It is kind of sloppy right now, but feel free to add stuff. It is just sort of a scratch sheet to figure out what types of terms we might need. The items on the spreadsheet look all good to me. I would even add chloroquine even though it might not be as effective as some have claimed, because we might want to represent trials... The only reason I said I wouldn't add is because I thought we were staying away from adding specific drugs. Should we add it anyways? I see, sorry, you were right! I wasn't thinking... @pnrobinson @LCCarmody Should I add 'infectious disease screening' , 'radiography imaging procedure', 'radiography of the chest', and 'Real-time PCR specimen test' to the spreadsheet and then start filling some of those spreadsheet terms in more? Also, are we interested in any term related to questionnaires about recent travel or is that best left out? Sure, add them. it may start to become more obvious how to extend the structure of MAXO to accommodate. I think adding questionnaire's is fine, but I would like to stay away from what NCIt did where they added every question on questionnaires. Maybe a general 'travel risk assessment questionaire' or something? Most of these terms have been added in the new release (Yay!). I will add the others to separate tickets and eventually close this ticket. Feel free to add more to the google doc or individual tickets. Just trying to contain this a bit. Closing. Nearly all terms have been added (thanks for the suggestions!). The remaining ones are in #168 #169 #166.
gharchive/issue
2020-03-25T23:57:18
2025-04-01T06:45:02.230780
{ "authors": [ "LCCarmody", "laurenechan", "pnrobinson" ], "repo": "monarch-initiative/MAxO", "url": "https://github.com/monarch-initiative/MAxO/issues/158", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
610263902
Pickle or otherwise save random walks so they can be reused Per convo in N2V meeting This would be useful to reuse walks and speed up runtime for different learning tasks or different classifiers, etc @realmarcin @justaddcoffee Is this desired? If so, I can look into this @deepakunni3 -- definitely we should be able to pickle/unpickle the walks data and the embeddings, it would be great to have you think about this! Yeah, the walk can be pretty computationally painful - it'd be good to have the option to save off walks and reuse them! Okay, I'll start working on this 👍
gharchive/issue
2020-04-30T17:38:27
2025-04-01T06:45:02.233532
{ "authors": [ "deepakunni3", "justaddcoffee", "pnrobinson" ], "repo": "monarch-initiative/N2V", "url": "https://github.com/monarch-initiative/N2V/issues/139", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1977960215
KDM6B https://pubmed.ncbi.nlm.nih.gov/37196654/ added
gharchive/issue
2023-11-05T20:54:52
2025-04-01T06:45:02.242952
{ "authors": [ "pnrobinson" ], "repo": "monarch-initiative/phenopacket-store", "url": "https://github.com/monarch-initiative/phenopacket-store/issues/44", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
416734812
Add cli parameter for monitor mode refresh rate Please make the new refresh rate option configurable at launch using -m -r for example. Available in v0.9-beta.10.
gharchive/issue
2019-03-04T10:32:58
2025-04-01T06:45:02.254056
{ "authors": [ "Memphizzz", "moncho" ], "repo": "moncho/dry", "url": "https://github.com/moncho/dry/issues/94", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
272628056
Request for merchants & services page We would like to add WalletMonero to your services list on the page: https://getmonero.org/community/merchants/ Previously we introduced our service to Monero community, which resulted in good reviews: https://www.reddit.com/r/Monero/comments/78nvn6/walletmonero_new_monero_light_wallet/ Kindly use the following link: https://walletmonero.com?utm_source=getmonero_org&utm_medium=list&utm_content=monero_list&utm_campaign=getmonero_org In case of any questions don't hesitate to contact me: support@walletmonero.com +merchant Added in #464
gharchive/issue
2017-11-09T16:30:37
2025-04-01T06:45:02.899627
{ "authors": [ "QuickBASIC", "erciccione", "walletmonero" ], "repo": "monero-project/monero-site", "url": "https://github.com/monero-project/monero-site/issues/473", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1014261709
moneropedia add p2pool sidechain merge-mining First PR for a Moneropedia page after translation changes, so let me know if it's missing anything. Adds pages for P2Pool, sidechain, and merge mining. A lot of the text is borrowed from announcement blog post but I attempted to adjust for appropriate tone. Feedback is appreciated. Resolves #1845 Should this be closed since it's included in #2074? yep @erciccione
gharchive/pull-request
2021-10-03T04:51:31
2025-04-01T06:45:02.901199
{ "authors": [ "QuickBASIC", "erciccione", "plowsof" ], "repo": "monero-project/monero-site", "url": "https://github.com/monero-project/monero-site/pull/1858", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
217695289
Spanish translation Spanish Translation. Everything should be right. Im planning in translating everything to spanish. You can't just edit the text in the tree to create a translation, unfortunately - with the redesign we're sorting out a translation setup that will led you edit the files directly.
gharchive/pull-request
2017-03-28T21:17:21
2025-04-01T06:45:02.902163
{ "authors": [ "fluffypony", "romanturbo" ], "repo": "monero-project/monero-site", "url": "https://github.com/monero-project/monero-site/pull/243", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2044831414
Adding in app Is it possible that the codec's of this code of yours can be inserted into any other app which is built on Java? There are not decoders in this app other than just https://github.com/androidx/media
gharchive/issue
2023-12-16T15:51:55
2025-04-01T06:45:02.944119
{ "authors": [ "devilxsam", "moneytoo" ], "repo": "moneytoo/Player", "url": "https://github.com/moneytoo/Player/issues/543", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
1637685502
CLOUDP-167800: Migrate Custom DNS to v2 client Proposed changes Migrate Custom DNS methods to v2 client. Jira ticket: CLOUDP-167800 Closes #[issue number] Checklist [x] I have signed the MongoDB CLA [ ] I have added tests that prove my fix is effective or that my feature works [ ] I have added any necessary documentation in document requirements section listed in CONTRIBUTING.md (if appropriate) [ ] I have addressed the @mongodb/docs-cloud-team comments (if appropriate) [ ] I have updated test/README.md (if an e2e test has been added) [x] I have run make fmt and formatted my code Further comments It will be good to get info what type of testing were done locally. Verification will be much easier.
gharchive/pull-request
2023-03-23T14:43:41
2025-04-01T06:45:02.997431
{ "authors": [ "andmatei", "wtrocki" ], "repo": "mongodb/mongodb-atlas-cli", "url": "https://github.com/mongodb/mongodb-atlas-cli/pull/1797", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2723227273
CLOUDP-280747: Build a layer to handle --format Proposed changes support for --format support content type support go templates support for --output --output is required when not piping stdout and content type is binary fixed bug where all flags were set even though the user did not set them + added more tests Jira ticket: CLOUDP-280747 LGTM Do you think it's fine to skip docs review here, since it's a merge from master?
gharchive/pull-request
2024-12-06T14:44:49
2025-04-01T06:45:03.000499
{ "authors": [ "jeroenvervaeke" ], "repo": "mongodb/mongodb-atlas-cli", "url": "https://github.com/mongodb/mongodb-atlas-cli/pull/3453", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
220177823
possible issue in delayOnNextBySelector? I'm trying to create an Observable[Int] that delays every element by its amount, e.g., [100, 200, 300] => first element is emitted after 100 ms, after that the second after 200 ms, etc I have tried def delayBySelf(in: Observable[Int]) = in.delayOnNextBySelector(x => Observable.evalDelayed(x.millis, ())) but this doesn't seem to work - all the elements are emitted instantly. Nevermind, I did a mistake :)
gharchive/issue
2017-04-07T10:42:24
2025-04-01T06:45:03.034643
{ "authors": [ "clayrat" ], "repo": "monix/monix", "url": "https://github.com/monix/monix/issues/345", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1012967238
Update list.json Update company participant Don't waste our time
gharchive/pull-request
2021-10-01T06:52:28
2025-04-01T06:45:03.035918
{ "authors": [ "monizb", "ravi-hash" ], "repo": "monizb/hacktoberfest-swag-list", "url": "https://github.com/monizb/hacktoberfest-swag-list/pull/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1364036906
fix: update dependencies to solve the problem of garbled CJK characters Ref: https://github.com/fiatjaf/nostr-tools/pull/20 Update dependencies to solve the problem of garbled CJK characters: hey there, thank you this has been merged. closing.
gharchive/pull-request
2022-09-07T03:26:35
2025-04-01T06:45:03.044442
{ "authors": [ "bieberg0n", "monlovesmango" ], "repo": "monlovesmango/astral", "url": "https://github.com/monlovesmango/astral/pull/39", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
517418906
[BUG] Memory leak at SKRoundRect#Transform? https://github.com/mono/SkiaSharp/blob/master/binding/Binding/SKRoundRect.cs public SKRoundRect Transform (SKMatrix matrix) { var destHandle = SkiaApi.sk_rrect_new (); if (SkiaApi.sk_rrect_transform (Handle, ref matrix, destHandle)) { return new SKRoundRect (destHandle, true); } return null; } I think before return null;, we need to call SkiaApi.sk_rrect_delete (destHandle);. Good catch! Thanks! Fixed in #995 Looking at the new code: https://github.com/mono/SkiaSharp/pull/995/files I think the memory leak hasn't been fixed, because there's no call to sk_rrect_delete. When sk_rrect_transform returns false, destHandle is still not released. @mattleibow Please see the above comment ^^ Oh no. I think something got lost in a merge. Let me fix that. Should be there now.
gharchive/issue
2019-11-04T21:37:56
2025-04-01T06:45:03.054547
{ "authors": [ "mattleibow", "ngocdaothanh" ], "repo": "mono/SkiaSharp", "url": "https://github.com/mono/SkiaSharp/issues/994", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
648119463
Fix clang warnings around implicit float conversions /Users/hugh/Documents/GitHub/libgdiplus/src/graphics.c:2647:30: warning: implicit conversion from 'int' to 'float' changes value from 1000000032 to 1.0E+9 [-Wimplicit-int-float-conversion] Closing, as libgdiplus is deprecated
gharchive/pull-request
2020-06-30T12:08:05
2025-04-01T06:45:03.055824
{ "authors": [ "hughbe" ], "repo": "mono/libgdiplus", "url": "https://github.com/mono/libgdiplus/pull/655", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
199413025
Is there any way to prevent multiselect styles from loading? I want to write styles from scratch, I need a very customized design and I don't want to overwrite the default styles with my own styles. Is there any way to disable the default stylesheet completely? My guess would be to extend the Multiselect: <template> // custom multiselect template </template> <script> import Multiselect from 'vue-multiselect' export default { extends: Multiselect, // your custom logic here } </script> <style> // your custom styles here </style> This however, would require creating a new template that you will be styling. However you might try not providing a template here, just the styles. This might result in overriding just the styles, not the template. But that’s just a guess. Let me know if that helps! :) Thanks for the response. I've already tried this solution before posting on Github. It doesn't prevent default styles from loading. Even if you extend the component and provide a custom template and CSS then original styles are still injected into the page. This way I can only override the original styles (which I can do without extending the component as well), but the original styles are still loaded. But it seems that it has nothing to do with the vue-multiselect component, it's a vue.js core issue, or even a vue-loader/vuify issue. I will talk with Linus or someone other from the core team to find a solution. Thanks! :) I'm also looking for some easy solution for this problem. Currently i have to override styles with !important tag which I dislike very much I think the best solution would be if you guys would drop .vue component in favour of plain js file and render function and provide styles in separate file (.css, .scss) then the customization will be much more easier :) For SSR I’m working on separating the styles from the component so that you have to import them yourself. This should solve all problems regarding using custom styles. @sqal @wujekbogdan beta.14 released and should fix this problem. Sorry you had to wait so long :( Please read: https://github.com/monterail/vue-multiselect/blob/2.0/README.md Breaking change regarding the styles.
gharchive/issue
2017-01-08T09:44:58
2025-04-01T06:45:03.187257
{ "authors": [ "shentao", "sqal", "wujekbogdan" ], "repo": "monterail/vue-multiselect", "url": "https://github.com/monterail/vue-multiselect/issues/201", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1489406405
🛑 THORN Data Sync Service (China) is down In c0bd208, THORN Data Sync Service (China) (https://syncpoint.thorn.red/latency) was down: HTTP code: 0 Response time: 0 ms Resolved: THORN Data Sync Service (China) is back up in 72a3ea2.
gharchive/issue
2022-12-11T04:18:38
2025-04-01T06:45:03.217782
{ "authors": [ "Alecyrus" ], "repo": "mooncyan/Status", "url": "https://github.com/mooncyan/Status/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1782047582
Failed to push the docker image to Cloud Foundry I am having an issue when I deploy the Docker image to cloud foundry (provided by SAP BTP) What were you trying to do? In the terminal, I executed the following command. cf push x9-moov --docker-image moov/imagecashletter:latest What did you expect to see? I expected it to deploy the server and start. What did you see? The app was deployed, but it could not be started. Here is the error log. I increased the timeout to 3 minutes and still time out for the health checkup. Any thoughts on how can I solve this issue? How can we reproduce the problem? SAP BTP provides a trial account with a Cloud Foundry environment. Do you have the output that's easier to read? Red on blue is quite hard, but it looks like the ports for health checks are wrong? Got it fixed. The issue was with the health check port. By default, the Cloud Foundry checked on the 8080 port with no path. Had to make it to ping 8083 and the '/ping' path.
gharchive/issue
2023-06-30T07:34:31
2025-04-01T06:45:03.239029
{ "authors": [ "adamdecaf", "krisho007" ], "repo": "moov-io/imagecashletter", "url": "https://github.com/moov-io/imagecashletter/issues/301", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1689633106
Feature disconnects covers and fixes #469 does clenup integrated into Server implementation also contains #743 Also fixed an issue of race that caused exception on connect, that is very easy to reproduce: import paho.mqtt.client as mqtt mqttBroker = "127.0.0.1" client = mqtt.Client(client_id=b"Temperature_Inside") client.connect(mqttBroker) res = client.publish("/TEMPERATURE", randNumber) that causes: 20:33:56.213 [nioEventLoopGroup-13-1] DEBUG io.moquette.broker.MQTTConnection -- Received MQTT message, type: CONNECT 20:33:56.214 [nioEventLoopGroup-13-1] DEBUG io.moquette.broker.PostOffice -- Routing cmd [CONN] for session [Temperature_Inside] to event processor 7 20:33:56.215 [nioEventLoopGroup-13-1] DEBUG io.moquette.broker.metrics.MQTTMessageLogger -- C->B PUBLISH <null> to topics </TEMPERATURE> 20:33:56.215 [nioEventLoopGroup-13-1] DEBUG io.moquette.broker.MQTTConnection -- Received MQTT message, type: PUBLISH 20:33:56.216 [nioEventLoopGroup-13-1] ERROR io.moquette.broker.NewNettyMQTTHandler -- Error processing protocol message: PUBLISH java.lang.NullPointerException: Cannot invoke "String.hashCode()" because the return value of "io.moquette.broker.SessionCommand.getSessionId()" is null at io.moquette.broker.PostOffice.routeCommand(PostOffice.java:632) at io.moquette.broker.MQTTConnection.processPublish(MQTTConnection.java:427) at io.moquette.broker.MQTTConnection.handleMessage(MQTTConnection.java:80) at io.moquette.broker.NewNettyMQTTHandler.channelRead(NewNettyMQTTHandler.java:58) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) .... The disconnection feature works like this (traces): ======> MQTT Clients: ({:id Temperature_Inside, :address 127.0.0.1, :port 54641}) ======> MQTT Advanced Broker disconnecting ......... 20:05:45.234 [main] DEBUG com.dkdhub.mqtt_broker.AdvancedBroker -- Will disconnect client Temperature_Inside 20:05:45.234 [main] DEBUG io.moquette.broker.SessionRegistry -- Disconnecting client: Temperature_Inside 20:05:45.235 [main] DEBUG io.moquette.broker.SessionRegistry -- Remove session state for client Temperature_Inside 20:05:45.235 [main] DEBUG io.moquette.broker.SessionRegistry -- Client Temperature_Inside successfully disconnected from broker 20:05:45.235 [main] DEBUG com.dkdhub.mqtt_broker.AdvancedBroker -- Disconnect result: true ======> MQTT Clients: () 20:05:45.236 [nioEventLoopGroup-3-1] INFO io.moquette.broker.metrics.MQTTMessageLogger -- Channel Inactive 20:05:45.236 [nioEventLoopGroup-3-1] DEBUG io.moquette.broker.MQTTConnection -- Notifying connection lost event 20:05:45.237 [nioEventLoopGroup-3-1] DEBUG io.moquette.broker.SessionEventLoopGroup -- Routing cmd [CONN LOST] for session [Temperature_Inside] to event processor 7 20:05:45.237 [Session Executor 7] DEBUG io.moquette.broker.MQTTConnection -- Cleaning Temperature_Inside 20:05:45.237 [Session Executor 7] DEBUG io.moquette.interception.BrokerInterceptor -- Notifying unexpected MQTT client disconnection to interceptor CId=Temperature_Inside, username=null, interceptorId=3456 20:05:45.238 [pool-2-thread-1] DEBUG basic -- MQTT: client #object[clojure.core$_PLUS_ 0x46f9cda1 clojure.core$_PLUS_@46f9cda1] Temperature_Inside #object[clojure.core$_PLUS_ 0x46f9cda1 clojure.core$_PLUS_@46f9cda1] disconnected (lost) We maintain a clojure library that wraps moquette, so the tests are written in clojure (actually this functionality is at trunk now and awaits for upstream updates), however they are way simple, so you may take a look by yourself - https://github.com/dkdhub/clj-mqtt-broker/blob/3d59a2244afec43ae2a3522ac9173ae683a9dc3d/test/basic.clj#L129 Looking forward to hearing from you! @andsel may I ask you to take a look at this please? Or to assign this request to someone who may address this? Small note in regards to CONNECT+PUBLISH issue: the root cause is that the implementation mainly relies on NettyUtils.clientID(channel) call, that requires channel to be associated with the connection. At the moment of CONNECT the new association creation is implemented by firing out the FutureTask instance and since immediate publishing is not restricted by MQTT spec - the message can not be processed and should be ignored. It is enough to return PostOffice.RouteResult.failed for the case. That oneline fix is embedded into this request. Thank you. Hi @source-c thanks for your contribution. To ease the review, discussion an tracking of the issues I would ask a couple things. Please: split the NPE PR, it's already covered by #743 so document well how to reproduce the problem in that PR remove the 0.17-SNAPSHOT-1 changes Thank you, @andsel. Both PRs were modified in accordance to your comments. Thanks @source-c I'll review in the next days, thank you for contributing to the project. I'll comment on them, if I have any doubts. @source-c #743 has been merged on main so you can rebase this, please Hi @source-c yes it's possible, but I have to review it and had no time in the past weeks :-( I'll try to do it during next weekend Seems, you cannot reproduce that because of implementation differences: org.eclipse.paho.client.mqttv3.MqttClient: aClient.connect(options, null, null).waitForCompletion(getTimeToWait()); when paho.mqtt.client does just: self.connect_async(host, port, keepalive, bind_address, bind_port, clean_start, properties) you see there is no await at pythonic client (which is used very broadly). So, when one uses such approach: client = mqtt.Client(client_id=cid) client.connect(mqttBroker) res = client.publish(topic, value) both CONNECT and PUBLISH may (and will) go almost simultaniously. On the other hand we have association creation at broker, that yields FutureTask called by CONNECT that came. For the next packetcontains PUBLISH, broker searches for client id (at processPublish) using getClientId() which is in turn just a wrapper around the code: NettyUtils.clientID(channel); and this channel in turn is yet not completely initialized. As the result we are getting an NPE there. I've tried also with Java Paho Async client and HiveMQ async, but neither of the two exposes such behavior. Reading the spec fro MQTT 3.1.1 it doesn't prohibit such to send other control packets before having received the CONN_ACK, requirement [MQTT-3.1.4-5] Clients are allowed to send further Control Packets immediately after sending a CONNECT Packet; Clients need not wait for a CONNACK Packet to arrive from the Server. If the Server rejects the CONNECT, it MUST NOT process any data sent by the Client after the CONNECT Packet [MQTT-3.1.4-5]. although it describes this non normative comment: Clients typically wait for a CONNACK Packet, However, if the Client exploits its freedom to send Control Packets before it receives a CONNACK, it might simplify the Client implementation as it does not have to police the connected state. The Client accepts that any data that it sends before it receives a CONNACK packet from the Server will not be processed if the Server rejects the connection. So it seems that if the broker receives the PUB before sending the CONN_ACK, and the connection is accepted, then it has to process the PUB as if it's a valid. I think that to implement such behavior we should move the assignment of clientId to channel, which happens in the session loop while processing the connect https://github.com/moquette-io/moquette/blob/ec4e84a666ab4b8cd4bf9431da3ea6c3b5fe6fc1/broker/src/main/java/io/moquette/broker/MQTTConnection.java#L217 , to he point before enqueuing the command: https://github.com/moquette-io/moquette/blob/ec4e84a666ab4b8cd4bf9431da3ea6c3b5fe6fc1/broker/src/main/java/io/moquette/broker/MQTTConnection.java#L183 In this case: if the CONNECT result with a positive result, then the PUBLISH is routed to the correct session loop and given the serialization of the commands, when the session loop start to process the PUBLISH all the session data is setup and both the session and the connection have valid data. if the CONNECT result with a negative result, and closes the channel, when the session loop encounter any command (PUB, SUB etc) related to that channel/clientId, it doesn't have a channel/session created and should log an error or warning. I think that to implement such behavior we should move the assignment of clientId to channel, which happens in the session loop while processing the connect moquette/broker/src/main/java/io/moquette/broker/MQTTConnection.java Line 217 in ec4e84a NettyUtils.clientID(channel, clientId); , to he point before enqueuing the command: moquette/broker/src/main/java/io/moquette/broker/MQTTConnection.java Line 183 in ec4e84a final String sessionId = clientId; In this case: if the CONNECT result with a positive result, then the PUBLISH is routed to the correct session loop and given the serialization of the commands, when the session loop start to process the PUBLISH all the session data is setup and both the session and the connection have valid data. if the CONNECT result with a negative result, and closes the channel, when the session loop encounter any command (PUB, SUB etc) related to that channel/clientId, it doesn't have a channel/session created and should log an error or warning. Of course, implementing that in such a way follows the spec more precisely. But it doesn't neglect that empty clientId case should be either be covered at the routing. At least to identify a case of collision/race if any. Also, I suppose changing that behaviour is to be done separately to avoid mess, does it? However it's not clear to me why SessionRegistry.closeSession is not enough for your use case. There are a lot use cases when forcing disconnect is used under hard pressure and in accordance to security policies. Some of use cases already discussed at #747 BTW. changing that behaviour is to be done separately to avoid mess, does it? Yes, it's better separate in another PR There are a lot use cases when forcing disconnect I understand, but it's not clear to me if the intended effect of the new method should be like a standard close; which shutdown the MQTTConnection and clean the Session status (means topic subscriptions and queues) if needed (clean session = false). Or does it has to remove all session status also if the session was a non clean session? In the second case every reference to the session is brutally removed, while in the first is like a normal disconnection flow. In the second case every reference to the session is brutally removed, while in the first is like a normal disconnection flow. Unfortunatelly, second case is most prefered. We should remove all the session internals disregading the its state. The state might be not clean, the channel might not be operational, server (broker) might be overhelmed by unwanted (sometimes broken) requests. Forcing session drop (i.e. kicking off the client) we perserve server stability among other reasons, trying to keep it operational for all other clients. However you are right - this should decided end user, so I remade cleanup contract and now session cleanup is to be specified by outer parameter explicitly.
gharchive/pull-request
2023-04-29T17:40:08
2025-04-01T06:45:03.274457
{ "authors": [ "andsel", "source-c" ], "repo": "moquette-io/moquette", "url": "https://github.com/moquette-io/moquette/pull/744", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2011638495
enhancement: add support to publish message to specific clients Release notes What does this PR do? add support to publish message to specific clients It's a fixed version of https://github.com/moquette-io/moquette/pull/786 Why is it important/What is the impact to the user? The original request is that one request will be sent to topic "/request", but the response should be sent back on another topic "/response". With these modifications, the embedding broker could distinguish all connections and support such send/resp mechanism without MQTT5 version. Checklist [ ] My code follows the style guidelines of this project [ ] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [ ] I have made corresponding change to the default configuration files (and/or docker env variables) [ ] I have added tests that prove my fix is effective or that my feature works [ ] I have updated the Changelog if it's a feature or a fix that has to be reported Author's Checklist [ ] How to test this PR locally Related issues Use cases Hi @maxwang789 thanks for your interest in the project. I think that this a deviation from MQTT protocol, all client subscribed to topics that matches the topic on which publish is done, has to be notified of a message. The MQTT 3.1 and MQTT 3.1.1 doesn't specify anything of this shape, so could be considered a break of the protocol. Hi @maxwang789 I understand your reason, but internalPublish is an API method that Server provides to it's consumers. So the ask here is to modify a published API intended to respect the MQTT protocol to satisfy a specific request that's outside the protocol specification. If you provide any number of valid use cases would be helpful , because as I can understand it so satisfy your specific requirements. Having more context, for example which devices claims to be MQTT compliant (while in reality they aren't) and if it solves a specific use case or is more general, would help understand if it worthwhile to go outside of specs or not. For example, I don't know if other brokers implements such customizations.
gharchive/pull-request
2023-11-27T07:10:33
2025-04-01T06:45:03.283301
{ "authors": [ "andsel", "maxwang789" ], "repo": "moquette-io/moquette", "url": "https://github.com/moquette-io/moquette/pull/797", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
122186594
fix markdown format. cc @morishitter thanks :)
gharchive/pull-request
2015-12-15T03:39:29
2025-04-01T06:45:03.301535
{ "authors": [ "appleboy", "morishitter" ], "repo": "morishitter/gulp-cssfmt", "url": "https://github.com/morishitter/gulp-cssfmt/pull/2", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
214388857
Make it working with Inferno.js Make it working with Inferno.js Thanks!
gharchive/pull-request
2017-03-15T13:29:19
2025-04-01T06:45:03.304837
{ "authors": [ "AlgoTrader", "moroshko" ], "repo": "moroshko/react-autosuggest", "url": "https://github.com/moroshko/react-autosuggest/pull/342", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1159969295
Morpheus inserts __typename to disambiguate unions. Description When selecting fragments on a union, Morpheus inserts a __typename field in the answer. If I have the following schema: union Test = Foo | Bar type Foo { foo: Int! } type Bar { bar: Int! } type Query { getFooOrBar(id: Int!): Test } and I send the following query: query { getFooOrBar(id: 1) { ... on Foo { foo } ... on Bar { bar } } } I am expecting to receive (in the case of Foo): data: getFooOrBar: foo: 1 but instead I receive: data: getFooOrBar: __typename: Foo foo: 1 Minimal program import Data.Aeson (encode) import Data.ByteString.Lazy as B (putStr) import Data.Morpheus.Document import Data.Morpheus.Types import Data.Morpheus [gqlDocument| union Test = Foo | Bar type Foo { foo: Int! } type Bar { bar: Int! } type Query { getFooOrBar(id: Int!): Test } |] queryRoot :: Query (Resolver QUERY () IO) queryRoot = Query {getFooOrBar} where getFooOrBar (Arg i) = pure $ case i of 1 -> Just $ TestFoo $ Foo $ pure 1 2 -> Just $ TestBar $ Bar $ pure 2 _ -> Nothing main = do let service :: GQLRequest -> IO GQLResponse service = interpreter $ RootResolver queryRoot Undefined Undefined send q = service $ GQLRequest { operationName = Nothing , variables = Nothing , query = q } B.putStr . encode =<< send "\ \query {\ \ getFooOrBar(id: 1) {\ \ ... on Foo {\ \ foo\ \ }\ \ ... on Bar {\ \ bar\ \ }\ \ }\ \}" thanks @nicuveo, why is it a problem? It's not a huge problem, but it technically goes against the spec; this is a very pedantic issue, I'm sorry. ^^ The spec says it explicitly in § 2.4: An operation selects the set of information it needs, and will receive exactly that information and nothing more, avoiding over‐fetching and under‐fetching data. Additionally, there are two parts of the spec that describe how selection sets are to be handled, and they explicitly mention starting with an empty mapping and adding one entry per requested field: § 6.3.2 gives the instructions on how to create the set of "grouped fields" [§ 6.3] http://spec.graphql.org/June2018/#sec-Executing-Selection-Sets then explains how to construct the answer from the grouped fields @nicuveo all right. then how can we fix the problem with client? probably some kind of config disambigueTypes:: Boolean
gharchive/issue
2022-03-04T19:23:04
2025-04-01T06:45:03.310742
{ "authors": [ "nalchevanidze", "nicuveo" ], "repo": "morpheusgraphql/morpheus-graphql", "url": "https://github.com/morpheusgraphql/morpheus-graphql/issues/687", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
270419067
Add script to use an existing codebase, for core or client development Great work, so far @mortenson It would be nice if this environment could be easily used for drupal core development. That means the complete web folder should be a repository. Then it's much easier to generate patches and stuff like that. Running tests within the container will need to covered by this issue, it appears that there's some DX problems that need to be addressed. Running tests is documented here: https://github.com/mortenson/drupal-docker-lite#running-phpunit-tests Next is adding repository support for ./ddl start I just tried: Using an existing codebase You can use an existing codebase by cloning it into a "code" subdirectory of this project. For >example: git clone git@github.com:mortenson/drupal-docker-lite.git your-project cd your-project git clone code .>/ddl.sh start I cloned drupal core into the code folder and after ddl.sh start http://drupal.ddl/ pops up and throws 403 Forbidden Ok, moved files into code/docroot and it works fine @chrfritsch Core is a special case, I plan to write a new script for core/contrib contributors. 😄 @chrfritsch I went a little crazy with https://github.com/mortenson/drupal-docker-lite/commit/255ae6600c968432f95c7de48e4bb3134a263e79 - now when you run ddl.sh start it prompts you to enter a composer project, an existing repository (!!), or you can enter contrib which is a special keyword that will download/setup Drupal core's dev branch. If you have time to play around with it I'd appreciate any feedback. Thanks!
gharchive/issue
2017-11-01T19:13:57
2025-04-01T06:45:03.336327
{ "authors": [ "chrfritsch", "mortenson" ], "repo": "mortenson/drupal-docker-lite", "url": "https://github.com/mortenson/drupal-docker-lite/issues/4", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2103830903
🛑 Bama is down In d922348, Bama (https://bama.design) was down: HTTP code: 0 Response time: 0 ms Resolved: Bama is back up in 5d6bf8c after 12 minutes.
gharchive/issue
2024-01-27T21:59:04
2025-04-01T06:45:03.340412
{ "authors": [ "mortonpepper" ], "repo": "mortonpepper/upptime-upptime", "url": "https://github.com/mortonpepper/upptime-upptime/issues/1070", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2049495083
Updating the Flash Attention version to fix cross entropy loss The cross entropy loss of Flash Attention v2.3.2 (and lower) throws illegal memory access error when used with large (device train microbatch size X sequence length X vocabulary). To fix this we had reverted back to FA v1's CE loss in this PR (https://github.com/mosaicml/llm-foundry/pull/795). However, we discovered that for very large (device train microbatch size X sequence length X vocabulary), FA v1's CE loss runs into numerical precision errors, causing divergence. The newer versions of Flash Attention (v2.3.3 and higher) seem to have solved both of these problems, and hence in this PR, we update the repo to start using FA v2.3.6 (the latest version) instead of FA v2.3.2. The blue loss curve below corresponds to the training run using FA v2.3.6's CE loss, and the pink curve corresponds to training run using FA v1's CE loss. throughput, memory, and loss before and after: LGTM, please add a PR description explaining stuff Done.
gharchive/pull-request
2023-12-19T21:54:26
2025-04-01T06:45:03.344787
{ "authors": [ "ShashankMosaicML", "dakinggg" ], "repo": "mosaicml/llm-foundry", "url": "https://github.com/mosaicml/llm-foundry/pull/812", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
302532283
Move atomtyping step into a specific function Dicussed in #157, but to repeat the important bit: Currently there's a bit of code at the beginning of Forcefield.createSystem that manages the atomtyping step, whereas this should more cleanly done by being sectioned off to a different function. I'm open to naming the function something more elegant, I'm not happy with Forcefield.run_atomtyping but couldn't come up with something better. I'm not sure if it should be a helper function, or whatever Forcefield._run_atomtyping would be called. It seems possible for a user to want to call it directly without creating the entire openmm.System or parmed.Structure. I don't think a test is necessary for this, as I'm just moving code around. But I also couldn't think of how a useful test would be structured. I could verify that the returned openmm.Topology has atomtype information, but this should cause later tests to crash anyway. I added use_residue_map as an argument to Forcefield.apply. I don't know why it wasn't there before, and don't see any reason for it to be buried inside these other functions. Do you want this in a test or also in the function itself? I don't think the docstring I added is great (please suggest fixes/improvements!) but it's more descriptive. Personally I like the name of the argument - it's not great that it describes a variable that's totally internal - but it's a difficult feature to describe in an argument alone. And coveralls is whining that I added one un-tested line but only three tested lines 🙄 Doc string LGTM! Would be easy to add a test case that triggers the error to get that coverage up 😁 I'm having trouble seeing a good way to test for missing atomtypes. If an atom in the topology truly lacks atomtypes defined in the forcefield, I think errors in find_atomtypes would be raised. I could try to manually pop a.id from a random atom to force it, I guess? https://docs.pytest.org/en/latest/assert.html#assertions-about-expected-exceptions In this case I just don't know how to raise that error. I get how to use pytest to verify that a desired error has been raised. don't know how to raise that error I would just call createSystem without the atomtyping flag and it should yell at you But that argument is gone now (the atomtyping happens before createSystem is called). We can try applying an inappropriate force field to a molecule, but that triggers errors elsewhere that are already test for: def test_missing_atomtypes(): ethane = pmd.load_file(get_fn('pf6.mol2'), structure=True) oplsaa = Forcefield(name='oplsaa') topo, NULL = generate_topology(ethane) with pytest.raises(ValueError): > oplsaa.run_atomtyping(topo) foyer/tests/test_forcefield.py:137: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ foyer/forcefield.py:369: in run_atomtyping find_atomtypes(residue, forcefield=self) foyer/atomtyper.py:43: in find_atomtypes _resolve_atomtypes(topology) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ topology = <Topology; 1 chains, 1 residues, 7 atoms, 6 bonds> def _resolve_atomtypes(topology): """Determine the final atomtypes from the white- and blacklists. """ for atom in topology.atoms(): atomtype = [rule_name for rule_name in atom.whitelist - atom.blacklist] if len(atomtype) == 1: atom.id = atomtype[0] elif len(atomtype) > 1: raise FoyerError("Found multiple types for atom {} ({}): {}.".format( atom.index, atom.element.name, atomtype)) else: raise FoyerError("Found no types for atom {} ({}).".format( > atom.index, atom.element.name)) E foyer.exceptions.FoyerError: Found no types for atom 0 (phosphorus). foyer/atomtyper.py:104: FoyerError Ok good point - sorry I misread how you had implemented this. I think we're good then since we already guarantee that all atoms got a type before calling createSystem which is the whole point One of the travis builds on mac got hung up on something, but the second most recent test did not, and the most recent commit is minor. Should be fine. Somebody else should give it a once-over before merging, but from my perspective this is good to go. Do you have admin access on travis? To restart builds etc? No on upstream I don't think Did I miss anything or is this good to go? LGTM. @ctk3b anything else, or are we good to merge? Sorry I missed this - LGTM!
gharchive/pull-request
2018-03-06T01:50:22
2025-04-01T06:45:03.354384
{ "authors": [ "ctk3b", "mattwthompson", "summeraz" ], "repo": "mosdef-hub/foyer", "url": "https://github.com/mosdef-hub/foyer/pull/159", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
772616239
[wip]support netpoll in dev branch mosn/easygo@27b5582 add field to desc, avoid calling file.Fd which will set the conn to blocking mode remove fd copy in NewDesc, prevent pre close file during GC, before the connection actually close call ReadOnce again until readAgain == false, if there is still data to read when TLS is enabled. clear read timeout timer before doRead, set timer after doRead to enable/disable trigger for read timeout event to send heartbeat request todo test heartbeat... Codecov Report Merging #1523 (771c67e) into develop (a520ff2) will increase coverage by 0.27%. The diff coverage is 55.66%. @@ Coverage Diff @@ ## develop #1523 +/- ## =========================================== + Coverage 58.98% 59.25% +0.27% =========================================== Files 334 337 +3 Lines 26974 27034 +60 =========================================== + Hits 15910 16019 +109 + Misses 9287 9233 -54 - Partials 1777 1782 +5 Impacted Files Coverage Δ pkg/network/eventloop.go 10.00% <12.50%> (+10.00%) :arrow_up: pkg/network/connection.go 41.59% <12.69%> (-0.02%) :arrow_down: pkg/upstream/servicediscovery/dubbod/pub.go 68.96% <42.85%> (-6.48%) :arrow_down: pkg/track/track.go 86.53% <86.53%> (ø) pkg/stream/xprotocol/conn.go 38.01% <87.50%> (+2.42%) :arrow_up: pkg/track/buffer.go 90.90% <90.90%> (ø) pkg/proxy/downstream.go 47.87% <100.00%> (+0.88%) :arrow_up: pkg/stream/xprotocol/stream.go 31.14% <100.00%> (+3.56%) :arrow_up: pkg/trace/sofa/http/tracer.go 83.87% <100.00%> (ø) pkg/trace/sofa/xprotocol/span.go 63.77% <100.00%> (+1.79%) :arrow_up: ... and 13 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update f45b080...44c3ad0. Read the comment docs. modify mtls/crypto/README
gharchive/pull-request
2020-12-22T03:53:09
2025-04-01T06:45:03.395971
{ "authors": [ "cch123", "codecov-io", "taoyuanyuan" ], "repo": "mosn/mosn", "url": "https://github.com/mosn/mosn/pull/1523", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1346373538
支持 connpool_binding 的 GoAway 1.1 在 upstream 收到了 GoAway 请求后会将自己从 conn_pool 里面删除掉,这样下次就会用到新的链接 1.2 在 upstream 收到了 GoAway 标记自己已经从 upstream 删除了,下次收到关链接的 event 不再删除 downstream 的链接 Issues associated with this PR Your PR should present related issues you want to solve. Solutions You should show your solutions about the issues in your PR, including the overall solutions, details and the changes. At this time, Chinese is allowed to describe these. UT result Unit Test is needed if the code is changed, your unit test should cover boundary cases, corner cases, and some exceptional cases. And you need to show the UT result. Benchmark If your code involves the processing of every request, you should give the Benchmark Result. Code Style Make sure Goimports has run Show Golint result #2114 @Dennis8274 @zonghaishang cc,看是否对你们的binding有影响哟 没影响,connpool_binding.go 暂时没用到
gharchive/pull-request
2022-08-22T12:51:09
2025-04-01T06:45:03.400160
{ "authors": [ "EraserTime", "taoyuanyuan", "zonghaishang" ], "repo": "mosn/mosn", "url": "https://github.com/mosn/mosn/pull/2115", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1630741686
deps: updates wazero to 1.0.0-rc.2 This updates wazero to 1.0.0-rc.2 which notably to mosn: is the final release candidate before 1.0 next Friday. improves instantiation performance (startup time) Benchmark If your code involves the processing of every request, you should give the Benchmark Result. $ go test -run='^$' -tags wasmer -bench '^Benchmark_.*' ./test/integrate/proxywasm -count=6 > old.txt $ go test -run='^$' -tags wasmer -bench '^Benchmark_.*' ./test/integrate/proxywasm -count=6 > new.txt benchstat old.txt new.txt goos: darwin goarch: arm64 pkg: mosn.io/mosn/test/integrate/proxywasm │ old.txt │ new.txt │ │ sec/op │ sec/op vs base │ _BaseCase-12 62.26µ ± 0% 62.16µ ± 2% ~ (p=0.699 n=6) _ProxyWasmV1_wasmer-12 84.92µ ± 4% 86.13µ ± 2% ~ (p=0.132 n=6) _ProxyWasmV1_wazero-12 76.44µ ± 2% 74.53µ ± 2% -2.50% (p=0.004 n=6) geomean 73.94µ 73.62µ -0.43% Good job @mathetake in getting 2.5% execution improvement within a single release candidate! this is out of date as latest is 1.2.0 this is out of date as latest is 1.2.0 @codefromthecrypt Will you update this PR or open a new one? @spacewander I can update this after https://github.com/mosn/proxy-wasm-go-host/pull/38 merges. thanks for the reminder. @spacewander ready
gharchive/pull-request
2023-03-19T02:05:17
2025-04-01T06:45:03.404627
{ "authors": [ "codefromthecrypt", "spacewander" ], "repo": "mosn/mosn", "url": "https://github.com/mosn/mosn/pull/2254", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
349300313
Warn against using this module Since the module is no longer needed, the README should make clear that this module is not needed for new projects with recent AWS SDK versions. I agree that it would be good to communicate that the types are currently out of date - this is something I intend to fix, though can't commit to a timeline - so I'll gladly add a notice to the README. However, I wouldn't say that this module is "not needed". The AWS SDK doesn't ship with Flow types, only TypeScript definitions. As long as this is the case, there is room for a 3rd party library of Flow definitions to fill the gap.
gharchive/pull-request
2018-08-09T21:38:51
2025-04-01T06:45:03.446516
{ "authors": [ "esbenvb", "motiz88" ], "repo": "motiz88/aws-sdk-flow-decls", "url": "https://github.com/motiz88/aws-sdk-flow-decls/pull/117", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1455679781
Trusted Players are not affected by GM Theme option I just realized that users with the trusted player level of permission are not affected by the GM Theme option. Well, from what I saw they are, but not in all aspects, so the UI looks very strange and unreadable. Maybe an option to enable the GM theme just for certain roles could enhance that? gonna use this thread to say that my GM theme option, actually, does not enable the theme for any user at all, I wonder if I'm doing something wrong or it's just a bug Is there any errors in the console? If your player is logged in does it change the setting when you change a setting? I made a change in the last update that I only pass in the most recent update and not all of them. However it should sync when they login to the current theme. Can you send me a photo of your settings for libThemer Only these errors/warnings related to lib-themer The only thing I noticed in that regard is that when I disable the GM Theme option, the player user can see the button to configure a theme (in the settings tab), aside from that I didn't notice anything of the sort My settings as requested: Alright, so it looks like the variable array that should be storing the theme isn't an array and is causing the .concat function of the array to break. It also appears that is an issue with my localization repeating over and over for you. Thats gotta be annoying, is that only libThemer doing that for you, where the hint text is repeated like that? No no, I guess it's a bug from another module, a refresh solves it, so it's fine So this array breaking is something I would need to solve on my end? Its a bug, I am not sure what caused it, I think I have an idea how to fix it but I want to test it before I just tell you to make sure I don't break more things. gonna use this thread to say that my GM theme option, actually, does not enable the theme for any user at all, I wonder if I'm doing something wrong or it's just a bug Just wanted to mention that it is working for me on both a Player and a Trusted Player. After checking 'Enable GM Theme' I had to relog both the GM and the player to get it to work (initially I just did the player but that wasn't enough), but it does work as expected. I am not using the custom theme directory, I just used the Manage Presets thing. Here is my module list of things that are working fine with it in case it's a conflict (DnD5e system). 2022-12-30-package-list.txt
gharchive/issue
2022-11-18T19:27:27
2025-04-01T06:45:03.462766
{ "authors": [ "ducky00", "mouse0270", "niuserre" ], "repo": "mouse0270/lib-themer", "url": "https://github.com/mouse0270/lib-themer/issues/35", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
92561209
公開終了日時に過去の日時を設定できる MTではエラーを表示しますが、アプリではどうしますか? Steps: 記事の作成 設定画面で、終了日時を設定 過去の日時を入力して、完了 MTに送信 Expected: (エラーを表示?) Observed: そのまま保存/送信できる。MTで編集画面を開くと、「公開終了日は、未来の日時を指定してください。」とエラーが表示される 修正しました verified fix.
gharchive/issue
2015-07-02T06:40:08
2025-04-01T06:45:03.467710
{ "authors": [ "cheebow", "cmatsuo" ], "repo": "movabletype/smartphone-app", "url": "https://github.com/movabletype/smartphone-app/issues/104", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1327217854
EleutherAI How easy would it be to swap out SalesForce CodeGen models with those from EleutherAI? GPT-Neo, GPT-J, and GPT-NeoX models are also trained on GitHub. GPT-J would be straightforward, since that's the same architecture as CodeGen. FasterTransformer has a guide here; you'd just need to convert the model: https://github.com/NVIDIA/FasterTransformer/blob/main/docs/gptj_guide.md#download-the-model I'm not sure what the architecture of NeoX-20B is. If it's identical to GPT-J but bigger, it could also be converted to FT. I don't expect GPT-J to be very good as a code assistant, though -- the CodeGen paper evaluated it (Table 3): https://arxiv.org/pdf/2203.13474.pdf And it's about the same size as CodeGen 6B, so you'd probably be better off just using that. I saw SaleForce and assumed incorrectly that their models were nonfree. No reason to use EleutherAI in this case.
gharchive/issue
2022-08-03T13:26:52
2025-04-01T06:45:03.475005
{ "authors": [ "TechnologyClassroom", "moyix" ], "repo": "moyix/fauxpilot", "url": "https://github.com/moyix/fauxpilot/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
149266878
Validation password fields Fields: password and repeat-password. How to validate this? And how to return errors in form from backend? Password and repeat-password. How to validate this? This is likely to need #79 to be addressed. And how to return errors in form from backend? This lib is intended to use client-side validation only. We don't have use cases for propagating errors from server side validation, so if you want this to happen that's likely to require contribution :) Side note: Of course you can always add two password fields in your schema and perform validation when the form is submitted, but you then admittedly have no way to render error messages contextually to the form (meaning notifying the user outside of the form itself). We don't have use cases for propagating errors from server side validation. Hmm. Example: user registers and enters a username already contained in the database, it is logical to return an error in form. I know there are legitimate use cases for password validation in the real world, just here we're dealing with a client side librariy which has no plan on supporting server side rendering nor validation anytime sound. I'm just suggesting that if you want to see this happening, you should probably start to get involved by participating in the reflexion process how should it be done, eg by providing feedback on the issue I've just filed for your specific need in #145? Thanks for answers! Also filed #146 to initiate a conversation about async external validation. @Chachaproper I've started an experiment in #163 about that, feedback welcome. The fix has landed as part of v0.30.0.
gharchive/issue
2016-04-18T20:46:54
2025-04-01T06:45:03.692456
{ "authors": [ "Chachaproper", "n1k0" ], "repo": "mozilla-services/react-jsonschema-form", "url": "https://github.com/mozilla-services/react-jsonschema-form/issues/144", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
116122937
Cannot transfer ownership for Themes Prerequisites: Build identifier: Mozilla/5.0 (Windows NT 6.1; WOW64; rv:45.0) Gecko/20100101 Firefox/45.0 Steps to reproduce: 1.Submit a new theme 2.Go to theme edit listing page and try to transfer ownership Expected results: You can transfer ownership without problems Actual results: You cant transfer ownership(Button is greyed out) Please see screencast for this bug : http://screencast.com/t/OQSPJX88wC Is there a javascript error on this page? There is no error in console Verified as fixed on AMO-dev FF48.0.2(Win 7) Postfix video:
gharchive/issue
2015-11-10T14:59:14
2025-04-01T06:45:04.177231
{ "authors": [ "ValentinaPC", "andymckay", "vcarciu" ], "repo": "mozilla/olympia", "url": "https://github.com/mozilla/olympia/issues/898", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
71349715
Addon name included in error message (bug 1119792) When an error for adding an addon with the same name is raised, the addon name will also be returned in the error message. Hello @yashmehrotra, did you mean to close the PR straight away? When someone submits the form, he already knows the name, as it's the one being used to submit, right? Thanks for your interest anyway! @magopian - I closed the PR as it failed the build ( I need to make changes in the test files too ). So I plan to make those changes and then send a new PR. As for the bug, so according to you the bug isn't legit ? I'm not saying it's not legit, just that it's a bit surprising. I guess it would still make sense to have that extra bit of information. Thanks for taking the time to work on this! Please let me know if I can help. I think I got it, apart from the change I made, I have to make the same change in the test files for the tests to pass. Will send a new PR upon completion. Thanks for the help.
gharchive/pull-request
2015-04-27T17:24:57
2025-04-01T06:45:04.180110
{ "authors": [ "magopian", "yashmehrotra" ], "repo": "mozilla/olympia", "url": "https://github.com/mozilla/olympia/pull/529", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
629367389
getDocument(...).then is not a function The following code loads the PDF file successfully from a filepicker into a byte array, but fails PDF.js with Uncaught TypeError: pdfjsLib.getDocument(...).then is not a function Code: <html> <head> <script src="https://mozilla.github.io/pdf.js/build/pdf.js"></script> </head> <body> <input type="file" id="file"> </body> <script> pdfjsLib.GlobalWorkerOptions.workerSrc = 'https://mozilla.github.io/pdf.js/build/pdf.worker.js'; document.getElementById('file').onchange = function(event) { var file = event.target.files[0]; var fileReader = new FileReader(); fileReader.onload = function() { var typedarray = new Uint8Array(this.result); console.log(typedarray); pdfjsLib.getDocument(typedarray).then(function(pdf) { console.log(pdf); }); }; fileReader.readAsArrayBuffer(file); } </script> </html> Mostly coming from the main answer of https://stackoverflow.com/questions/22048395/how-to-open-a-local-pdf-in-pdfjs-using-file-input. How to make it work? PS: I have seen a few similar questions like https://github.com/mozilla/pdf.js/issues/9075 closed because they lacked a minimal reproducible example. As shown in the JSDocs, getDocument returns a PDFDocumentLoadingTask instance: https://github.com/mozilla/pdf.js/blob/8fc1126b5a2b6d2fa0cfa98a061882ef21c1abc9/src/display/api.js#L169-L180 which has a promise property which contains the information you want here: https://github.com/mozilla/pdf.js/blob/8fc1126b5a2b6d2fa0cfa98a061882ef21c1abc9/src/display/api.js#L453-L459 According to the code/docs quoted above, the following thus works: const loadingTask = pdfjsLib.getDocument(/* data here */); loadingTask.promise.then(pdfDocument => { // The document is loaded here... }); Generally speaking, note that linking to code at https://mozilla.github.io/pdf.js/build/ is never a good idea (since those files change with every merge). Please always use official releases instead, see https://github.com/mozilla/pdf.js/releases Finally, when opening an issue, please provide all of the information requested in https://github.com/mozilla/pdf.js/blob/master/.github/ISSUE_TEMPLATE.md Thanks @Snuffleupagus for your answer. Finally just the .promise. was missing, I'll remember this. More generally, is it mandatory to use asynchronous code? For example: pdfjsLib.getDocument(typedarray).promise.then(function(pdf) { for (var i = 1; i <= pdf.numPages; i++) { pdf.getPage(i).then(function(page) { var viewport = page.getViewport({scale: 1}); var context = canvas.getContext('2d'); ... does not work because it would require multiple canvas: Error: Cannot use the same canvas during multiple render() operations. Is there a way to use synchronous code instead? page = pdf.getPage(i); in replacement of pdf.getPage(i).then(function(page) { is not working. More generally, is it mandatory to use asynchronous code? Yes; note also that generally speaking a lot of modern JavaScript code is asynchronous. does not work because it would require multiple canvas: Error: Cannot use the same canvas during multiple render() operations. The solution is simple though, either: create a new canvas for each page, or wait for one page to finish rendering before starting the next one. Is there a way to use synchronous code instead? No, for all intents and purposes all PDF.js API-methods are asynchronous at this point in time. Thanks @Snuffleupagus for your reply. Just to conclude, how would you transform: for (var i = 1; i <= pdf.numPages; i++) { pdf.getPage(i).then(function(page) { ... }); } to make it "wait for one page to finish rendering before starting the next one"?
gharchive/issue
2020-06-02T17:06:40
2025-04-01T06:45:04.190594
{ "authors": [ "Snuffleupagus", "josephernest" ], "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/issues/11960", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1114049444
Creating a pdf through MacOS by using a collection of images can lead into the images being rendered as black boxes in pdf.js Attach (recommended) or Link to PDF file here: test conglomerate 6 2.pdf Configuration: Web browser and its version: Safari 5.17 Operating system and its version: MacOS PDF.js version: 2.12 Is a browser extension: No Steps to reproduce the problem: Create a pdf in MacOS by using several pictures through find and generating the pdf through the Image Viewer (as shown in this link) Open the generated pdf in pdf.js and progressively browse through the file while zooming in and out Sometimes images from the pdf will render as entire black boxes What is the expected behavior? (add screenshot) The images should be rendered with no issue or whatsoever What went wrong? (add screenshot) I hope you could assist me on this issue Thank you in advance! WFM, using PDF.js version 2.13.77 in both Firefox Nightly 98 and Google Chrome Beta 98. Hence this unfortunately looks like a browser, or possible OS/driver, bug rather than a PDF.js library one. Web browser and its version: Safari 5.17 According to https://en.wikipedia.org/wiki/Safari_version_history#Safari_5 that version number doesn't seem to exist, and in any case it'd be a very old and outdated browser Please refer to https://github.com/mozilla/pdf.js/wiki/Frequently-Asked-Questions#faq-support PDF.js version: 2.12 Sorry, but that's not a complete version number; which PDF.js version are you actually using? Hey I have edited with the pertinent info required Unfortunately, given that it's specific to just one browser it doesn't seem like a bug in the PDF.js library but rather e.g. a canvas-related bug the browser itself. Please note that Firefox is the primary development target for this library.
gharchive/issue
2022-01-25T15:43:56
2025-04-01T06:45:04.198779
{ "authors": [ "Snuffleupagus", "richardba" ], "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/issues/14494", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
104278501
Hyper links are not supported in PDF JS version 1.1.1 Hi , I am using pdf js stable version 1.1.1 . Hyperlinks are not supported in pdf js 1.1.1 and we are able to open the hyper links from version (1.1.114) . Can you please suggest any changes to be made in 1.1.1 version to enable hyper links , as of now we can not change the files of other version. No changes have been made to the Link annotation code recently as far as I know (except for #6281, which appears to be unrelated). Version 1.1.1 is pretty old; the latest stable version is 1.1.215. You should really consider updating, and otherwise you would have to go through the last commits manually to find out which commit fixed the situation for you. Closing as answered.
gharchive/issue
2015-09-01T14:26:50
2025-04-01T06:45:04.201521
{ "authors": [ "gsampath127", "timvandermeij" ], "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/issues/6408", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
148762351
How do we get code metadata though. I love our repository harvesting. It's so great. I am a huge fan. :100: :100: But, as per the synopsis and, well, the title of this project... we need to be able to collect Software. Harvesting repositories from sites like DataCite, Figshare, etc is nice, as these sites have the capacity to host research code. but how can we specifically filter these repositories so we are only showing research software? I leaned a lot about the metadata fields at CodeMeta. And I think that to solve this problem (at least for Figshare), it may be possible to use DataCite entirely. I wasn't too familiar with DataCites query language (SOLR) nor did I make our DataCite client, but I figured out how to do a few things I don't think we knew we could do. For starters, say we wanted to do a search for biology on DataCite, BUT, filter the search to only show github related repositories (http://search.datacite.org/api?wt=json&q=biology&fq=relatedIdentifier:github) to get more perspective on software in datacite. We can do that with that query, for example. What I noticed is that resourceType seems to often not be used. BUT, resourceTypeGeneral is always filled. and it is ALWAYS filled with Software in the example I just gave, and many other different search criteria. So, that being said, how does this apply to a research repository that could /not/ contain code, like Figshare? Github only contains code so that's easy. Well I'm glad you asked. What I'm thinking is we can combine the publisher field with the resourceTypeGeneral field to literally solve this problem COMPLETELY. So basically, we need a query that ensures that resourceTypeGeneral:Software AND publisher:Figshare (well, at least in this one use case) So, here's that query.. which returns Figshare published results, and returns software related results. Boom. http://search.datacite.org/api?wt=json&q=science&fq=publisher:Figshare&fq=resourceTypeGeneral:Software&start=0 So... since any Figshare repo's that have a citable DOI get a DataCite DOI (see here & here,) that solves that problem. I'm not sure if Zenodo works in the same way. Taking a really quick look at their API, specifically "Metadata Formats" seems promising. Super promising. guys.. I think we just solved the big problem cc literally the entire universe @mok4ry @amb8805 @Lettuceman44 @acabunoc Thats awesome!! Lets see what we can do with it!
gharchive/issue
2016-04-15T20:33:57
2025-04-01T06:45:04.278880
{ "authors": [ "amb8805", "lukecoy" ], "repo": "mozillascience/software-discovery-dashboard", "url": "https://github.com/mozillascience/software-discovery-dashboard/issues/71", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
627663828
Navigation control position Hello! script.js contains code, which supposed to set up position for zoom buttons. However the code below does not works. map.addControl(new mapboxgl.NavigationControl({ position: 'top-left' })); Consider to switch to the following instead: map.addControl(new mapboxgl.NavigationControl(), "top-left"); Thanks for pointing this out. It used to work, but broke sometime in the past five years with a Mapbox GL JS update. Since the exact position doesn't matter, I just removed the location specification.
gharchive/issue
2020-05-30T06:27:05
2025-04-01T06:45:04.319993
{ "authors": [ "mpetroff", "utilmind" ], "repo": "mpetroff/print-maps", "url": "https://github.com/mpetroff/print-maps/issues/24", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2239876135
本地小程序能正常跑,到了真机上时好时坏 问题描述 首先感谢作者做了这么好的大量的工作。我用官方的 demo 跑起来没有任何问题。 但在自己的新坑上build出来在微信开发者工具能完好运行。上传到小程序体验版之后发现首次打开一直在 splash 页面,然后退出来以后再进入偶尔能成功。然后在里面体验几分钟出来重新打开后又一直在splash 页面。重试也不成功。 不知道是一些三方包不支持还是其他问题。 希望能得到解答,后续上线我会付费,谢谢! 环境信息 Flutter 版本号:3.16.7 (通过 fvm 工具安装) MPFlutter 版本号:2.5.1 操作系统:MacBook Pro sonoma v14.3.1 IDE:vscode 依赖以及build脚本 三方包: dependencies: flutter: sdk: flutter flutter_localizations: sdk: flutter cupertino_icons: ^1.0.2 go_router: ^13.2.0 # 路由 dio: ^5.4.2+1 # 网络请求 get: ^4.6.6 # 状态管理 flutter_svg: ^2.0.9 # svg lottie: ^3.0.0 # lottie动效 # cached_network_image: ^3.3.0 # 图片缓存 share_plus: ^7.2.1 # 分享 package_info_plus: ^5.0.1 # 获取原生端的包信息 device_info_plus: ^9.1.2 # 获取设备信息 app_links: ^3.5.0 # 分享跳转 pull_to_refresh: ^2.0.0 # 下拉刷新 shared_preferences: ^2.2.2 # 存储 flutter_staggered_grid_view: ^0.7.0 # 瀑布流 carousel_slider: ^4.2.1 # banner connectivity_plus: ^6.0.1 # 判断网络连接 logger: ^2.2.0 # 日志 fluttertoast: ^8.2.4 # 弹框 flutter_easyloading: ^3.0.5 # loading 弹框 flutter_screenutil: ^5.9.0 # 屏幕适配 flutter_native_splash: ^2.3.10 # 启动闪屏 flutter_keyboard_visibility: ^6.0.0 # 检测键盘的可见性 mpflutter_build_tools: ^2.5.0 mpflutter_wechat_api: ^2.2.2 mpflutter_wechat_mapview: ^0.0.3 mpflutter_core: ^2.5.1 audioplayers: ^5.0.0 # 音频播放 permission_handler: ^11.3.1 # 动态权限 dependency_overrides: audioplayers_web: git: url: https://github.com/swiftdo/audioplayers.git ref: mpflutter-v5.2.1-inner path: packages/audioplayers_web dev_dependencies: flutter_test: sdk: flutter # The "flutter_lints" package below contains a set of recommended lints to # encourage good coding practices. The lint set provided by the package is # activated in the `analysis_options.yaml` file located at the root of your # package. See that file for information about deactivating specific lint # rules and activating additional ones. flutter_lints: ^2.0.0 flutter_launcher_icons: "^0.13.1" build_runner: ^2.4.0 # For information on the generic Dart part of this file, see the # following page: https://dart.dev/tools/pub/pubspec # The following section is specific to Flutter packages. flutter: # The following line ensures that the Material Icons font is # included with your application, so that you can use the icons in # the material Icons class. uses-material-design: true build 文件: import 'package:mpflutter_build_tools/main.dart' as build_tools; void main(List<String> arguments) async { final buildArgs = [...arguments, '--wechat']; build_tools.addNonCompatiblesPackage('flutter_native_splash'); build_tools.useMiniTex = true; build_tools.useNoFontCanvasKit = true; build_tools.main(buildArgs); } 最好能提供一下产物,请私聊微信提供。
gharchive/issue
2024-04-12T11:13:07
2025-04-01T06:45:04.322899
{ "authors": [ "PonyCui", "maogpt" ], "repo": "mpflutter/mpflutter", "url": "https://github.com/mpflutter/mpflutter/issues/564", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
223117380
Integrate Facebook Workplace Driver Description: Add configuration for Worplace Integration: https://developers.facebook.com/docs/workplace/integrations/custom-integrations/bots That would be great! I'll try to open up a workplace account to integrate it, but I guess it just works with the normal Facebook drivers out of the box. I just tested using the Facebook driver and replacing YOUR-FACEBOOK-PAGE-TOKEN-HERE with YOUR-FACEBOOK-APP-TOKEN-HERE and everything works. I do not know if it is better to create a specific driver but all functionality works. I leave you to choose :) Nevertheless you can now update the documentation to put Facebook Workplace. I am available if you want more information, because maybe you do not have the possibility to have a Workplace instance to do the tests. That's great news! Thanks for testing it.
gharchive/issue
2017-04-20T16:00:52
2025-04-01T06:45:04.332772
{ "authors": [ "davaxi", "mpociot" ], "repo": "mpociot/botman", "url": "https://github.com/mpociot/botman/issues/341", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2174091401
Backup to USB? Dear Mr-Manuel, I have used rpi-clone to clone my raspberry pi SD cards in the past, so that I can just pop in the cloned SD card if the original got corrupted. There is no need for an external computer in this scenario. Rpi-clone will not work with Venus OS however, so I am very interested in your backup script. I see that your script is designed to create a backup on a remote computer. Can it be used instead to clone the SD card to another SD card located locally on USB? I think the path to the USB drive on Venus would be /run/media/sda1, but I am not sure this would work if your script is expecting a remote connection. Please advise. Thank you in advance for the clarification. P.S. I also use your dbus_serialbattery script on my system and it is awesome. Just change the mount path to your USB mount. Just change the mount path to your USB mount. Like this? Yes. You also have to disable the mounting process else the script tries to mount that path. https://github.com/mr-manuel/raspberry-pi-backup/blob/3b3089299e3319f55cc16b6d9b900c5c219185ca/raspberry-pi-backup/backup.sh#L57-L59
gharchive/issue
2024-03-07T15:10:58
2025-04-01T06:45:04.388297
{ "authors": [ "lfhoward", "mr-manuel" ], "repo": "mr-manuel/raspberry-pi-backup", "url": "https://github.com/mr-manuel/raspberry-pi-backup/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1425773887
Nulls in the keyset I'm having an issue where whenever I try to paginate after a specific record, my result set comes back empty. var reference = after.HasValue ? await query.Where((e) => e.Id == after.Value).FirstOrDefaultAsync() : null; var pagination = query .OrderBy((e) => e.ModifiedUtc) .ThenBy((e) => e.Name) .KeysetPaginate( (builder) => builder .Ascending((e) => e.ModifiedUtc) .Ascending((e) => e.Name), KeysetPaginationDirection.Forward, reference ); var result = pagination.Query .Select((e) => new Dto { Id = e.GlobalId, Name = e.Name, LastModified = e.ModifiedUtc, }) .Take(Math.Clamp(size, size, 100)) The only thing noteworthy about the test data I'm using is that they all have the same ModifiedUtc value. Though their names are different. Is there any good way to troubleshoot whatever is going on here? Thank you for this really cool library, looking forward to mastering using it! Cleaned up the redundancy, no impact. Here's a sample of the query being generated when I try to pass a reference: SELECT TOP(4) [t].[id] AS [Id], [t].[name] AS [Name], [t].[last_modified] AS [LastModified] FROM [mymodel] AS [t] WHERE ([t].[organization] = 'SOMEGUID') AND ((CASE WHEN [t].[last_modified] >= @__p_1 THEN CAST(1 AS bit) ELSE CAST(0 AS bit) END & (CASE WHEN [t].[last_modified] > @__p_1 THEN CAST(1 AS bit) ELSE CAST(0 AS bit) END | (CASE WHEN [t].[last_modified] = @__p_2 THEN CAST(1 AS bit) ELSE CAST(0 AS bit) END & CASE WHEN [t].[name] > @__p_3 THEN CAST(1 AS bit) ELSE CAST(0 AS bit) END))) = CAST(1 AS bit)) ORDER BY [t].[last_modified], [t].[name] I feel like it's taking the reference and creating some kind of impossible filter scenario. Again, the one anomaly in the data is the last_modified being the same between all the records. But that is something I hoped to see mitigated by the second sort on name. Oh, and also, my reference entity is the last record from the previous page. Does that make sense? Yes, that's the idea. A quick test I did with a model that has the same date column but different names columns worked completely fine. And the generated sql you're showing looks very weird to me. So this is proving hard to reproduce on my side. If you can provide a repro repo that I can just run and inspect I might be able to help further. Hmm, tricky for me to do at this point. Is there anything specific you might suggest I look at? @mrahhal -- The only part that's missing is a .ToListAsync(). I feel like there's something wrong going on with how the WHEREs are being generated when there's multiple sort columns. But I just can't seem to figure out how/why. Especially that my attempt at a repro worked fine. So won't be able to proceed here without a repro. Also to point out, this library only generates the dynamic C# expressions, and EF Core translates those. You can try and show me the generated expressions, not the sql (like I did above, they should be logged to the same place just before the generated sql log). That might be easier to understand, and can tell me if there's anything wrong with the generated expressions themselves. I wonder if it's because the date field that I'm trying to include in the sort order is nullable? I didn't know that. This is why a repro is important 😅 I can actually see there are no results returned if I change it to nullable and make them all null (giving them the same value, as your first comment seemed to suggest, returned results). I'll investigate. So, this occurs when the reference has a null value in its keyset. Comparison operators don't really work with NULLs in SQL and they invalidate the whole WHERE clause, resulting in no results. I missed to account for this. Unfortunately, this isn't an easy problem to solve, and the best I can say right now is you'll want to avoid having nullable columns in your keyset. This is a problem related to SQL itself, compounded by the fact that the SQL standard does not define a default ordering of NULLs, but mainly because NULL is a very special value in SQL. I was trying to figure out if I can confidently always do something with nulls in the reference in a predictable way, but I can't even begin to foresee all the different problematic cases. There are a lot of caveats. WHERE Created < "2022" will never return rows with NULL in Created, no matter if the DB sorts NULLs first or last. So this is looking like an unsolvable problem to me. Only thing I can recommend, again, is to not have nullable columns in the keyset. And I'm now beginning to wonder if I should actually throw an exception if that is detected, as I can't think of a single correct use case that can work here. Also, this is pretty much the same as https://github.com/mrahhal/MR.EntityFrameworkCore.KeysetPagination/issues/18, but I didn't investigate it at the time. And I also just realized that I don't have good test coverage on this particular issue. I only simply tested that the DB didn't scream at me when there are nulls involved.. which obviously wasn't enough to detect this. Awesome! Yes, apologies, the repro was a little difficult for me to make without knowing what was causing the problem. I might have ended up creating a repro without the issue and come up with nothing! 😆 I think our conversation did help and I was able to keep poking at it on my end. What if you made it so that if a column was detected as nullable, the query was augmented with where [column] is not null, and then that just ensures that the data is never tainted with null values? This is obviously a workaround and would require documentation, but at least people could still paginate on those columns if they so decided to accept the limitation? First of all you have to decide if you want null values to appear at the beginning or at the end of the sorted values. I am using SQL-Server, which uses NULLS-First (that is NULL value are lower than any other value), so I am doing the same. Once you know where to place null values, you can generate ad-hoc comparison expressions between "field" and "value" in all the possible cases: Field not nullable, value not null, any comparison operator => “normal” case, you are already handling it. Field nullable, value not null, ‘>’ operator => “normal” case, you are already handling it. Field nullable, value not null, ‘>=’ operator => “normal” case, you are already handling it. Field nullable, value null, ‘>’ operator => (constant expression true). Field nullable, value null, ‘>=’ operator => (field is null) (Expression.Equal(fieldExpr, nullExpr). Field nullable, value not null, ‘<’ operator => add “OR (field is NULL)” to the expression you generate for the “normal case”. Field nullable, value not null, ‘<=’ operator => add “OR (field is NULL)” to the expression you generate for the “normal case”. Field nullable, value null, ‘<’ operator => (constant expression false). Field nullable, value null, ‘<=’ operator => field is not null (Expression.NotEqual(fieldExpr, nullExpr). I believe if you modify your private static BinaryExpression MakeComparisonExpression() to handle the above cases, it will work. What if you made it so that if a column was detected as nullable, the query was augmented with where [column] is not null, and then that just ensures that the data is never tainted with null values? This will work only for very specific cases, it's riddled with caveats (see @raffaele-cappelli's comment). @raffaele-cappelli, this is assuming I can tell what the sorting of NULLs will be from this library, which I can't do. Sure, I can make it an option for the user to provide, but this makes it that much more error prone and harder to consume as a library. The alternative of trying to somehow get the registered data provider and map that to a well known sorting order also doesn't appeal to me (especially that this library exposes its apis as extension methods on top of IQueryable). This is also assuming that we can tell if a column is nullable without going back to check the table's schema, which isn't the case at all. For exmaple, apps that aren't annotated for NRTs. And even for apps that are, users can always configure nullable columns irrelevant of the .net nullability context. One workaround I can suggest is to create a (non-nullable) computed column that coalesces nulls into some value that makes sense to your specific use case (and results in the sorting you expect), and then use that in the keyset instead. If this works just fine as I'm expecting here, I think this would be the best way to go about solving this, as opposed to introducing flaky complicated code in the library to solve a niche scenario. That's a great idea. I verified that using a computed column works properly and is predictable. It also allows you to deterministically specify the sorting of NULLs, irrelevant of the database impl. I'll be adding an example to the tests and sample. In addition, soon I'll release a new version with an analyzer that detects nullable properties in the keyset and warns about it, pointing to the workaround. This might be better than an exception, as you'll be able to suppress it if you really really want to. This is now possible in the latest version without a computed column, through expressing the coalescing in the keyset itself. Explained here.
gharchive/issue
2022-10-27T14:54:01
2025-04-01T06:45:04.414927
{ "authors": [ "atrauzzi", "mrahhal", "raffaele-cappelli" ], "repo": "mrahhal/MR.EntityFrameworkCore.KeysetPagination", "url": "https://github.com/mrahhal/MR.EntityFrameworkCore.KeysetPagination/issues/24", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
109451178
[Node 4] Upgrade NAN to v2 and use Node 4 socketwatcher commit Also update the engines field in package.json. Fixes #181 Fixes #169 Test Plan: Ran a very basic test on OS X listening on en0 (had to use sudo, which I believe is normal) and was able to pick up packets and see which MAC addr they were from. Crap, I didn't see you already ported it. It was fun nevertheless (https://github.com/mranney/node_pcap/pull/184) :beers: :+1: The examples don't work here, either. We should look into that.. did the API differs from the examples on purpose and they simply weren't updated? I rebuilt node_pcap with Node 0.10.38 and had no luck running the examples either. Alas :confused: However, it works just fine. So it doesn't really matter IMHO. A new set of examples could be delivered in a different PR. I cannot figure out how to get this to build in node 4, is anyone else still having issues? What OS, version, architecture? disregard, i figured it out. The new version was not pushed to NPM, so i just included a link directly to this repo
gharchive/pull-request
2015-10-02T08:33:12
2025-04-01T06:45:04.419790
{ "authors": [ "KenanSulayman", "hortinstein", "ide" ], "repo": "mranney/node_pcap", "url": "https://github.com/mranney/node_pcap/pull/182", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
419176698
Stuck when trying to view the stripe openApi definitions When trying to load the following openAPI definitions the RapiDoc is stuck (page not responding): https://api.apis.guru/v2/specs/stripe.com/2019-02-19/swagger.json I have tried to load this in the following: https://mrin9.github.io/RapiDoc/example2.html Thanks, Guy I am getting some issue with circular refs in the spec, will work on a fix soon this should be fixed with the latest commit. You may give it a try by pulling latest code to build and test yarn install yarn serve then navigate to http://localhost:8080 Reopening the issue, The fix broke some other functionalities The spec seems to have circular references, The latest commit ensures that it does not get into an un-ending loop. But it may ignore certain parts of the spec. Can you verify again if it shows like you expect it to be? caused due to a bug in json-schema-ref-parser https://github.com/APIDevTools/json-schema-ref-parser/issues/56 have introduced another attribute resolve-circular-refs which defaults to true. If you are facing any performance issue you might turn off resolving circular refs. stripe open-api spec can now be loaded by setting this attribute to 'false' I am closing this issue as the actual reason is at https://github.com/APIDevTools/json-schema-ref-parser/issues/56 but now you have a workaround Reopen it if you still face problem you still cant use https://mrin9.github.io/RapiDoc/example2.html to open stripe open-api spec as this example attempts to resolve all the circular refs in the spec Thanks for the quick fix. Really love the RapiDoc. @LevinGuy released version 4.2.1 this version replaces json-schema-ref-parser with json-refs In our testing we have found json-refs to be performing better than json-schema-ref-parser when handling large files. It did not had an issue in resolving https://api.apis.guru/v2/specs/stripe.com/2019-02-19/swagger.json the drawback is its size. @mrin9 Thanks, looks good 👍
gharchive/issue
2019-03-10T11:49:19
2025-04-01T06:45:04.532185
{ "authors": [ "LevinGuy", "mrin9" ], "repo": "mrin9/RapiDoc", "url": "https://github.com/mrin9/RapiDoc/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1510210201
[Meta]: Add usage example of nvim-tree Category Docs Description -- ~/.config/nvim/lua/init.lua local nvimtreemappings = require("nvim-tree.keymap").keymaps local keystodisable = vim.tbl_map(function(e) return e.key end, nvimtreemappings) require("nvim-tree").setup({ ... on_attach = function(bufnr) -- Set up Legendary plugin keybindigns keymaps = vim.tbl_map(function(e) local entry = {} entry.mode = "n" entry[1] = e.key -- Key if type(e.key) == "table" then entry[1] = e.key[1] end entry[2] = e.callback entry.description = "Nvim-tree: " .. e.desc entry.opts = { buffer = bufnr, noremap = true } return entry end, nvimtreemappings) require("legendary").setup({ keymaps = keymaps }) end, remove_keymaps = keystodisable ... https://user-images.githubusercontent.com/8136158/209453360-27d30933-94fa-42a0-8a39-f7ee64fb0c4d.mp4 Demo video. (CC BY 4.0). Note: the Nvim-tree: toggle is bound by Legendary separately. The above demostrates integration with nvim-tree/nvim-tree.lua module. Note: the said plugin should export a stable API, see related issue here: https://github.com/nvim-tree/nvim-tree.lua/issues/1858 Hey there, this is awesome, thanks for filing the issue! This may be better suited actually for #184, nvim-tree is probably a good candidate for a first one for a proof of concept. I'm not sure exactly what they will look like, but I have a few ideas. The best idea I have currently is along the lines of: Legendary.nvim must be loaded before other plugins. We have a set of plugins, or plugins can be provided by the other plugin itself, at legendary.plugins.[plugin name]. This way, plugins can be looked up dynamically; that is, if they wanted to, nvim-tree could provide its own legendary.nvim integration by adding lua/legendary/plugins/nvim_tree.lua in their repo, just for an example. The plugin is just a plugin-specific script that must handle registering its keymaps, commands, autocmds, and functions. It may set up autocmds to do so on BufEnter for example, or it may call some custom logic to determine the correct buffer to attach to. So for this example, and nvim-tree / legendary.nvim plugin would look like: User's config: require('legendary').setup({ plugins = { nvim_tree = true, } }) File: lua/legendary/plugins/nvim_tree.lua vim.api.nvim_create_autocmd('FileType', { pattern = 'NvimTree', callback = function() local nvimtreemappings = require("nvim-tree.keymap").keymaps local keystodisable = vim.tbl_map(function(e) return e.key end, nvimtreemappings) -- Set up Legendary plugin keybindigns local keymaps = vim.tbl_map(function(e) -- just set up a description-only keymap local entry = {} entry.mode = "n" entry[1] = e.key -- Key if type(e.key) == "table" then entry[1] = e.key[1] end entry.description = "Nvim-tree: " .. e.desc entry.opts = { buffer = bufnr, noremap = true } return entry end, nvimtreemappings) require('legendary').keymaps(keymaps) end }) cc @olimorris for input @mrjones2014 I would suggest to take a look at others nvim extensions that provide kind of extensions APIs to learn the best way, .e.g: Telescope.nvim#extensions @mrjones2014 I would suggest to take a look at others nvim extensions that provide kind of APIs for sub-extensions to learn the best way to do that, .e.g: Telescope.nvim#extensions nvim-tree provides a callback :help nvim-tree.on_attach to allow custom mappings for the buffer. Similar to gitsigns. nvim-tree is moving to that on_attach mechanism being the preferred means of creating mappings: https://github.com/nvim-tree/nvim-tree.lua/pull/1579. which-key and friends will be functional and the help screen will be greatly improved. The existing mechanism :help nvim-tree.view.mappings will still be functional but deprecated. Going to close this issue in favor of #184 since we now have a concrete implementation plan for it. @mrjones2014 Would be best if you update USAGE_EXAMPLES.md when #183 is closed. Best. Yeah I will definitely add documentation. Most likely I will create a separate PLUGINS.md file.
gharchive/issue
2022-12-24T23:50:14
2025-04-01T06:45:04.543793
{ "authors": [ "alex-courtis", "hinell", "mrjones2014" ], "repo": "mrjones2014/legendary.nvim", "url": "https://github.com/mrjones2014/legendary.nvim/issues/289", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2523753264
Blooks - CigRig CigRig CigRig
gharchive/issue
2024-09-13T02:37:45
2025-04-01T06:45:04.546204
{ "authors": [ "CigRig-10" ], "repo": "mrmojololo/blooket-admin-hack", "url": "https://github.com/mrmojololo/blooket-admin-hack/issues/994", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2038377475
React native support Hello, would you consider React native support for this library? It would be really amazing not only for me :-). Thank you, Krystof Hello, React native support is not planned at the moment.
gharchive/issue
2023-12-12T18:55:27
2025-04-01T06:45:04.561304
{ "authors": [ "Krelyshy", "mroczis" ], "repo": "mroczis/netmonster-core", "url": "https://github.com/mroczis/netmonster-core/issues/91", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
149990111
put definintions to definetlypyped Hello! Please add this definitons to definetlyTyped repo Or, if free for PR, i can move this definitions and update this project I'm a little busy to move and updated the definitions. You are free to update and submit them to DefinitelyTyped or Typings repository :) @mrpatiwi, ok, no problems)))
gharchive/issue
2016-04-21T07:25:34
2025-04-01T06:45:04.563043
{ "authors": [ "mrpatiwi", "stepancar" ], "repo": "mrpatiwi/typedgram-bot", "url": "https://github.com/mrpatiwi/typedgram-bot/issues/4", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1957495163
'Remote branch HEAD not found in upstream origin' error on build I am building mruby with specified mgems. After performing rake deep_clean and trying to build it again, I encounter a 'Remote branch HEAD not found in upstream origin' error, and the build fails. From what I've investigated, it seems that the issue is caused by the branch name written in build_config.rb.lock changing from master to HEAD. Reproduction Steps I have placed the code to reproduce the issue on https://github.com/buty4649/mruby-remote-branch-bug. Please use this for reference. Clone the repository from https://github.com/buty4649/mruby-remote-branch-bug and move to that directory.$ git clone https://github.com/buty4649/mruby-remote-branch-bug $ cd mruby-remote-branch-bug At this point, it is set to branch: master.$ grep branch build_config.rb.lock branch: HEAD Proceed with the build.$ MRUBY_CONFIG=$(pwd)/build_config.rb MRUBY_BUILD_DIR=$(pwd)/build rake -f /path/to/mruby/Rakefile After the build completes, it changes to branch: HEAD.$ git diff | cat diff --git a/build_config.rb.lock b/build_config.rb.lock index db02184..642f97b 100644 --- a/build_config.rb.lock +++ b/build_config.rb.lock @@ -7,6 +7,6 @@ builds: host: https://github.com/mattn/mruby-json.git: url: https://github.com/mattn/mruby-json.git - branch: master + branch: HEAD commit: f99d9428025469f2400f93c53b185f65f963e507 version: 0.0.0 Cleanup build directory.$ MRUBY_CONFIG=$(pwd)/build_config.rb MRUBY_BUILD_DIR=$(pwd)/build rake -f /path/to/mruby/Rakefile deep_clean Try building again, and the issue should reproduce.$ MRUBY_CONFIG=$(pwd)/build_config.rb MRUBY_BUILD_DIR=$(pwd)/build rake -f /path/to/mruby/Rakefile -- snip -- fatal: Remote branch HEAD not found in upstream origin rake aborted! Command failed with status (128): [git clone --recursive --branch "HEAD" "htt...] -- snip --- This problem should have been solved by me in #4899, but I overlooked it. :thinking: I will try to come up with a solution. I have sent #6092 as a proposed fix to this issue. Thank you for the report. I also tested #6092 in my local environment. The bug I reported was fixed. Thank you for the correction! 😄
gharchive/issue
2023-10-23T15:56:39
2025-04-01T06:45:04.568800
{ "authors": [ "buty4649", "dearblue" ], "repo": "mruby/mruby", "url": "https://github.com/mruby/mruby/issues/6087", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2494738132
Teach me github Mr. Sadeghi, Can you teach me how to work with Github? I have just made a My_space_game, One_file_projects, and Platformer repository. Hi dear, Yes. First you must understand some git commands. Mr. Sadeghi, Can you teach me how to work with Github? I have just made a My_space_game, One_file_projects, and Platformer repository. Hi dear, Yes. First you must understand some git commands.
gharchive/issue
2024-08-29T14:56:29
2025-04-01T06:45:04.575313
{ "authors": [ "alipnm", "msa-sadeghi" ], "repo": "msa-sadeghi/ali_pishnamazi", "url": "https://github.com/msa-sadeghi/ali_pishnamazi/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
711603603
[Bug]: Broken link in contributing.md Describe the bug Broken link 'Code of Conduct' in contributing.md To Reproduce Steps to reproduce the behavior: Go to '10-Easy-Steps/contributing.md' Click on 'contributing.md' Scroll to 'line 7' Edit link to the correct link of Code of Conduct (https://github.com/msandfor/10-Easy-Steps/blob/master/code_of_conduct.md) Expected behavior The link should go to Code of Conduct (https://github.com/msandfor/10-Easy-Steps/blob/master/code_of_conduct.md) Screenshots If applicable, add screenshots to help explain your problem. I will fix this issue
gharchive/issue
2020-09-30T03:28:56
2025-04-01T06:45:04.580233
{ "authors": [ "Keatnuxsuo" ], "repo": "msandfor/10-Easy-Steps", "url": "https://github.com/msandfor/10-Easy-Steps/issues/56", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
183254167
not able to install 0.1.24 on any version other than 0.10.x Hey and just a small note, I know we should use version 0.2, but for now the time is limited so sorry for my question. when I tried to install the new version 0.1.24 it never worked with any of the new versions of node for example this is the output from 0.12 2372 info preinstall mariasql@0.1.24 2373 verbose readDependencies loading dependencies from /var/www/node_modules/mariasql/package.json 2374 silly prepareForInstallMany adding lru-cache@2.3.1 from mariasql dependencies 2375 verbose readDependencies loading dependencies from /var/www/node_modules/mariasql/package.json 2376 silly cache add args [ 'lru-cache@2.3.1', null ] 2377 verbose cache add spec lru-cache@2.3.1 2378 silly cache add parsed spec { raw: 'lru-cache@2.3.1', 2378 silly cache add scope: null, 2378 silly cache add name: 'lru-cache', 2378 silly cache add rawSpec: '2.3.1', 2378 silly cache add spec: '2.3.1', 2378 silly cache add type: 'version' } 2379 silly addNamed lru-cache@2.3.1 2380 verbose addNamed "2.3.1" is a plain semver version for lru-cache 2381 silly mapToRegistry name lru-cache 2382 silly mapToRegistry using default registry 2383 silly mapToRegistry registry https://registry.npmjs.org/ 2384 silly mapToRegistry data { raw: 'lru-cache', 2384 silly mapToRegistry scope: null, 2384 silly mapToRegistry name: 'lru-cache', 2384 silly mapToRegistry rawSpec: '', 2384 silly mapToRegistry spec: 'latest', 2384 silly mapToRegistry type: 'tag' } 2385 silly mapToRegistry uri https://registry.npmjs.org/lru-cache 2386 verbose addNameVersion registry:https://registry.npmjs.org/lru-cache not in flight; fetching 2387 verbose request uri https://registry.npmjs.org/lru-cache 2388 verbose request no auth needed 2389 info attempt registry request try #1 at 7:35:21 AM 2390 verbose etag "F01T17JD32FTQ78T5MY5SOKXF" 2391 http request GET https://registry.npmjs.org/lru-cache 2392 verbose write writing to /var/www/node_modules/sequelize/node_modules/moment/package.json 2393 info preinstall moment@2.5.1 2394 verbose readDependencies loading dependencies from /var/www/node_modules/sequelize/node_modules/moment/package.json 2395 verbose readDependencies loading dependencies from /var/www/node_modules/sequelize/node_modules/moment/package.json 2396 silly install resolved [] 2397 verbose about to build /var/www/node_modules/sequelize/node_modules/moment 2398 info build /var/www/node_modules/sequelize/node_modules/moment 2399 info linkStuff moment@2.5.1 2400 silly linkStuff moment@2.5.1 has /var/www/node_modules/sequelize/node_modules as its parent node_modules 2401 verbose linkBins moment@2.5.1 2402 verbose linkMans moment@2.5.1 2403 verbose rebuildBundles moment@2.5.1 2404 info install moment@2.5.1 2405 info postinstall moment@2.5.1 2406 verbose unlock done using /root/.npm/_locks/moment-e505e6653c698ac9.lock for /var/www/node_modules/sequelize/node_modules/moment 2407 verbose about to build /var/www/node_modules/sequelize 2408 info build /var/www/node_modules/sequelize 2409 info linkStuff sequelize@1.7.9 2410 silly linkStuff sequelize@1.7.9 has /var/www/node_modules as its parent node_modules 2411 verbose linkBins sequelize@1.7.9 2412 verbose link bins [ { sequelize: 'bin/sequelize' }, 2412 verbose link bins '/var/www/node_modules/.bin', 2412 verbose link bins false ] 2413 verbose linkMans sequelize@1.7.9 2414 verbose rebuildBundles sequelize@1.7.9 2415 silly gentlyRm /var/www/node_modules/.bin/sequelize is being purged 2416 verbose gentlyRm don't care about contents; nuking /var/www/node_modules/.bin/sequelize 2417 verbose rebuildBundles [ 'bluebird', 'commander', 'moment' ] 2418 info install sequelize@1.7.9 2419 info postinstall sequelize@1.7.9 2420 verbose unlock done using /root/.npm/_locks/sequelize-27ffed5e38c43cfb.lock for /var/www/node_modules/sequelize 2421 http 304 https://registry.npmjs.org/lru-cache 2422 verbose headers { date: 'Sun, 16 Oct 2016 07:35:27 GMT', 2422 verbose headers via: '1.1 varnish', 2422 verbose headers 'cache-control': 'max-age=300', 2422 verbose headers etag: '"F01T17JD32FTQ78T5MY5SOKXF"', 2422 verbose headers age: '82', 2422 verbose headers connection: 'keep-alive', 2422 verbose headers 'x-served-by': 'cache-fra1237-FRA', 2422 verbose headers 'x-cache': 'HIT', 2422 verbose headers 'x-cache-hits': '6', 2422 verbose headers 'x-timer': 'S1476603327.126595,VS0,VE0', 2422 verbose headers vary: 'Accept-Encoding' } 2423 silly get cb [ 304, 2423 silly get { date: 'Sun, 16 Oct 2016 07:35:27 GMT', 2423 silly get via: '1.1 varnish', 2423 silly get 'cache-control': 'max-age=300', 2423 silly get etag: '"F01T17JD32FTQ78T5MY5SOKXF"', 2423 silly get age: '82', 2423 silly get connection: 'keep-alive', 2423 silly get 'x-served-by': 'cache-fra1237-FRA', 2423 silly get 'x-cache': 'HIT', 2423 silly get 'x-cache-hits': '6', 2423 silly get 'x-timer': 'S1476603327.126595,VS0,VE0', 2423 silly get vary: 'Accept-Encoding' } ] 2424 verbose etag https://registry.npmjs.org/lru-cache from cache 2425 verbose get saving lru-cache to /root/.npm/registry.npmjs.org/lru-cache/.cache.json 2426 verbose correctMkdir /root/.npm correctMkdir not in flight; initializing 2427 silly cache afterAdd lru-cache@2.3.1 2428 verbose afterAdd /root/.npm/lru-cache/2.3.1/package/package.json not in flight; writing 2429 verbose correctMkdir /root/.npm correctMkdir not in flight; initializing 2430 verbose afterAdd /root/.npm/lru-cache/2.3.1/package/package.json written 2431 silly install resolved [ { name: 'lru-cache', 2431 silly install resolved description: 'A cache object that deletes the least-recently-used items.', 2431 silly install resolved version: '2.3.1', 2431 silly install resolved author: { name: 'Isaac Z. Schlueter', email: 'i@izs.me' }, 2431 silly install resolved scripts: { test: 'tap test --gc' }, 2431 silly install resolved main: 'lib/lru-cache.js', 2431 silly install resolved repository: 2431 silly install resolved { type: 'git', 2431 silly install resolved url: 'git://github.com/isaacs/node-lru-cache.git' }, 2431 silly install resolved devDependencies: { tap: '', weak: '' }, 2431 silly install resolved license: 2431 silly install resolved { type: 'MIT', 2431 silly install resolved url: 'http://github.com/isaacs/node-lru-cache/raw/master/LICENSE' }, 2431 silly install resolved contributors: 2431 silly install resolved [ [Object], 2431 silly install resolved [Object], 2431 silly install resolved [Object], 2431 silly install resolved [Object], 2431 silly install resolved [Object], 2431 silly install resolved [Object], 2431 silly install resolved [Object] ], 2431 silly install resolved bugs: { url: 'https://github.com/isaacs/node-lru-cache/issues' }, 2431 silly install resolved _id: 'lru-cache@2.3.1', 2431 silly install resolved dist: 2431 silly install resolved { shasum: 'b3adf6b3d856e954e2c390e6cef22081245a53d6', 2431 silly install resolved tarball: 'https://registry.npmjs.org/lru-cache/-/lru-cache-2.3.1.tgz' }, 2431 silly install resolved _from: 'lru-cache@2.3.1', 2431 silly install resolved _npmVersion: '1.3.8', 2431 silly install resolved _npmUser: { name: 'isaacs', email: 'i@izs.me' }, 2431 silly install resolved maintainers: [ [Object] ], 2431 silly install resolved directories: {}, 2431 silly install resolved _shasum: 'b3adf6b3d856e954e2c390e6cef22081245a53d6', 2431 silly install resolved _resolved: 'https://registry.npmjs.org/lru-cache/-/lru-cache-2.3.1.tgz', 2431 silly install resolved readme: 'ERROR: No README data found!', 2431 silly install resolved homepage: 'https://github.com/isaacs/node-lru-cache#readme' } ] 2432 info install lru-cache@2.3.1 into /var/www/node_modules/mariasql 2433 info installOne lru-cache@2.3.1 2434 verbose installOne of lru-cache to /var/www/node_modules/mariasql not in flight; installing 2435 verbose correctMkdir /root/.npm/_locks correctMkdir not in flight; initializing 2436 verbose lock using /root/.npm/_locks/lru-cache-4862eead3cff6bbf.lock for /var/www/node_modules/mariasql/node_modules/lru-cache 2437 silly install write writing lru-cache 2.3.1 to /var/www/node_modules/mariasql/node_modules/lru-cache 2438 verbose unbuild node_modules/mariasql/node_modules/lru-cache 2439 silly gentlyRm /var/www/node_modules/mariasql/node_modules/lru-cache is being purged from base /var/www 2440 verbose gentlyRm don't care about contents; nuking /var/www/node_modules/mariasql/node_modules/lru-cache 2441 verbose tar unpack /root/.npm/lru-cache/2.3.1/package.tgz 2442 verbose tar unpacking to /var/www/node_modules/mariasql/node_modules/lru-cache 2443 silly gentlyRm /var/www/node_modules/mariasql/node_modules/lru-cache is being purged 2444 verbose gentlyRm don't care about contents; nuking /var/www/node_modules/mariasql/node_modules/lru-cache 2445 silly gunzTarPerm modes [ '755', '644' ] 2446 silly gunzTarPerm extractEntry package.json 2447 silly gunzTarPerm extractEntry .npmignore 2448 silly gunzTarPerm extractEntry README.md 2449 silly gunzTarPerm extractEntry LICENSE 2450 silly gunzTarPerm extractEntry bench.js 2451 silly gunzTarPerm extractEntry AUTHORS 2452 silly gunzTarPerm extractEntry lib/lru-cache.js 2453 silly gunzTarPerm extractEntry test/basic.js 2454 silly gunzTarPerm extractEntry test/foreach.js 2455 silly gunzTarPerm extractEntry test/memory-leak.js 2456 verbose write writing to /var/www/node_modules/mariasql/node_modules/lru-cache/package.json 2457 info preinstall lru-cache@2.3.1 2458 verbose readDependencies loading dependencies from /var/www/node_modules/mariasql/node_modules/lru-cache/package.json 2459 verbose readDependencies loading dependencies from /var/www/node_modules/mariasql/node_modules/lru-cache/package.json 2460 silly install resolved [] 2461 verbose about to build /var/www/node_modules/mariasql/node_modules/lru-cache 2462 info build /var/www/node_modules/mariasql/node_modules/lru-cache 2463 info linkStuff lru-cache@2.3.1 2464 silly linkStuff lru-cache@2.3.1 has /var/www/node_modules/mariasql/node_modules as its parent node_modules 2465 verbose linkBins lru-cache@2.3.1 2466 verbose linkMans lru-cache@2.3.1 2467 verbose rebuildBundles lru-cache@2.3.1 2468 info install lru-cache@2.3.1 2469 info postinstall lru-cache@2.3.1 2470 verbose unlock done using /root/.npm/_locks/lru-cache-4862eead3cff6bbf.lock for /var/www/node_modules/mariasql/node_modules/lru-cache 2471 verbose about to build /var/www/node_modules/mariasql 2472 info build /var/www/node_modules/mariasql 2473 info linkStuff mariasql@0.1.24 2474 silly linkStuff mariasql@0.1.24 has /var/www/node_modules as its parent node_modules 2475 verbose linkBins mariasql@0.1.24 2476 verbose linkMans mariasql@0.1.24 2477 verbose rebuildBundles mariasql@0.1.24 2478 verbose rebuildBundles [ 'lru-cache' ] 2479 info install mariasql@0.1.24 2480 verbose unsafe-perm in lifecycle false 2481 info mariasql@0.1.24 Failed to exec install script 2482 verbose unlock done using /root/.npm/_locks/mariasql-6148fa5430cf29a9.lock for /var/www/node_modules/mariasql 2483 verbose stack Error: mariasql@0.1.24 install: `node-gyp rebuild` 2483 verbose stack Exit status 1 2483 verbose stack at EventEmitter.<anonymous> (/usr/local/lib/node_modules/npm/lib/utils/lifecycle.js:217:16) 2483 verbose stack at EventEmitter.emit (events.js:110:17) 2483 verbose stack at ChildProcess.<anonymous> (/usr/local/lib/node_modules/npm/lib/utils/spawn.js:24:14) 2483 verbose stack at ChildProcess.emit (events.js:110:17) 2483 verbose stack at maybeClose (child_process.js:1019:16) 2483 verbose stack at Process.ChildProcess._handle.onexit (child_process.js:1091:5) 2484 verbose pkgid mariasql@0.1.24 2485 verbose cwd /var/www 2486 error Linux 4.4.24-moby 2487 error argv "/usr/local/bin/node" "/usr/local/bin/npm" "install" "--save" "--loglevel" "silly" 2488 error node v0.12.16 2489 error npm v2.15.1 2490 error code ELIFECYCLE 2491 error mariasql@0.1.24 install: `node-gyp rebuild` 2491 error Exit status 1 2492 error Failed at the mariasql@0.1.24 install script 'node-gyp rebuild'. 2492 error This is most likely a problem with the mariasql package, 2492 error not with npm itself. 2492 error Tell the author that this fails on your system: 2492 error node-gyp rebuild 2492 error You can get information on how to open an issue for this project with: 2492 error npm bugs mariasql 2492 error Or if that isn't available, you can get their info via: 2492 error 2492 error npm owner ls mariasql 2492 error There is likely additional logging output above. 2493 verbose exit [ 1, true ] 2494 verbose unbuild node_modules/mariasql 2495 info preuninstall mariasql@0.1.24 2496 info uninstall mariasql@0.1.24 2497 verbose unbuild rmStuff mariasql@0.1.24 from /var/www/node_modules 2498 info postuninstall mariasql@0.1.24 2499 silly gentlyRm /var/www/node_modules/mariasql is being purged from base /var/www 2500 verbose gentlyRm don't care about contents; nuking /var/www/node_modules/mariasql 2501 silly vacuum-fs purging /var/www/node_modules/mariasql 2502 silly vacuum-fs quitting because other entries in /var/www/node_modules any idea how to fix this ? Yep, upgrade. :-) @mscdex but isnt 0.1.24 should support all the new node versions ? No, mariasql v0.1.x and older generally pre-dates nan, which is the library that allows for better cross-V8/node compatibility. ok, thanks for the info .. another small question, do you know if there is any issue between sequelizejs 1.7.x and the new version of mariasql ? I couldn't tell you, I don't use sequelizejs. no worries, thanks for the help again. @linuxjuggler Can you the actual console output when installing mariasql? @mscdex mariasql is added directly to package.json file and which version you want the output for ? plus for which node ? cause I have been trying 0.10.x, 0.12.x, 4.x, and 6.x using the official docker images Whichever environment you were using when you initially reported the issue. then its the one I just posted in the top What you posted was the npm debug log, what we need is the actual console output. Unfortunately this is necessary for cases like this because npm's debug log file does not contain some important information, such as the stdout/stderr from make, compilers, etc. ok, Will do that in the evening when I get back home this one on node 4 node_1 | > mariasql@0.1.24 install /var/www/node_modules/mariasql node_1 | > node-gyp rebuild node_1 | node_1 | gyp info it worked if it ends with ok node_1 | gyp info using node-gyp@3.4.0 node_1 | gyp info using node@4.6.0 | linux | x64 node_1 | gyp http GET https://nodejs.org/download/release/v4.6.0/node-v4.6.0-headers.tar.gz node_1 | gyp http 200 https://nodejs.org/download/release/v4.6.0/node-v4.6.0-headers.tar.gz node_1 | gyp http GET https://nodejs.org/download/release/v4.6.0/SHASUMS256.txt node_1 | gyp http 200 https://nodejs.org/download/release/v4.6.0/SHASUMS256.txt node_1 | gyp info spawn /usr/bin/python2 node_1 | gyp info spawn args [ '/usr/local/lib/node_modules/npm/node_modules/node-gyp/gyp/gyp_main.py', node_1 | gyp info spawn args 'binding.gyp', node_1 | gyp info spawn args '-f', node_1 | gyp info spawn args 'make', node_1 | gyp info spawn args '-I', node_1 | gyp info spawn args '/var/www/node_modules/mariasql/build/config.gypi', node_1 | gyp info spawn args '-I', node_1 | gyp info spawn args '/usr/local/lib/node_modules/npm/node_modules/node-gyp/addon.gypi', node_1 | gyp info spawn args '-I', node_1 | gyp info spawn args '/root/.node-gyp/4.6.0/include/node/common.gypi', node_1 | gyp info spawn args '-Dlibrary=shared_library', node_1 | gyp info spawn args '-Dvisibility=default', node_1 | gyp info spawn args '-Dnode_root_dir=/root/.node-gyp/4.6.0', node_1 | gyp info spawn args '-Dnode_gyp_dir=/usr/local/lib/node_modules/npm/node_modules/node-gyp', node_1 | gyp info spawn args '-Dnode_lib_file=node.lib', node_1 | gyp info spawn args '-Dmodule_root_dir=/var/www/node_modules/mariasql', node_1 | gyp info spawn args '--depth=.', node_1 | gyp info spawn args '--no-parallel', node_1 | gyp info spawn args '--generator-output', node_1 | gyp info spawn args 'build', node_1 | gyp info spawn args '-Goutput_dir=.' ] node_1 | gyp info spawn make node_1 | gyp info spawn args [ 'BUILDTYPE=Release', '-C', 'build' ] node_1 | make: Entering directory '/var/www/node_modules/mariasql/build' node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/aes.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/aestables.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/algebra.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/arc4.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/asn.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/bftables.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/blowfish.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/coding.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/des.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/dh.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/dsa.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/file.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/hash.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/hc128.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/integer.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/md2.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/md4.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/md5.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/misc.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/rabbit.o node_1 | ../deps/libmariadbclient/extra/yassl/taocrypt/src/rabbit.cpp: In member function 'void TaoCrypt::Rabbit::Process(TaoCrypt::byte*, const byte*, TaoCrypt::word32)': node_1 | ../deps/libmariadbclient/extra/yassl/taocrypt/src/rabbit.cpp:234:29: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing] node_1 | *(word32*)(buffer+ 0) = LITTLE32(workCtx_.x[0] ^ node_1 | ^ node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/random.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/ripemd.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/rsa.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/sha.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/tftables.o node_1 | CXX(target) Release/obj.target/taocrypt/deps/libmariadbclient/extra/yassl/taocrypt/src/twofish.o node_1 | AR(target) Release/obj.target/deps/libmariadbclient/extra/yassl/taocrypt/taocrypt.a node_1 | COPY Release/taocrypt.a node_1 | CXX(target) Release/obj.target/yassl/deps/libmariadbclient/extra/yassl/src/buffer.o node_1 | CXX(target) Release/obj.target/yassl/deps/libmariadbclient/extra/yassl/src/cert_wrapper.o node_1 | CXX(target) Release/obj.target/yassl/deps/libmariadbclient/extra/yassl/src/crypto_wrapper.o node_1 | CXX(target) Release/obj.target/yassl/deps/libmariadbclient/extra/yassl/src/handshake.o node_1 | CXX(target) Release/obj.target/yassl/deps/libmariadbclient/extra/yassl/src/lock.o node_1 | CXX(target) Release/obj.target/yassl/deps/libmariadbclient/extra/yassl/src/log.o node_1 | CXX(target) Release/obj.target/yassl/deps/libmariadbclient/extra/yassl/src/socket_wrapper.o node_1 | CXX(target) Release/obj.target/yassl/deps/libmariadbclient/extra/yassl/src/ssl.o node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:35:0: warning: "X509_LOOKUP_add_dir" redefined node_1 | #define X509_LOOKUP_add_dir yaX509_LOOKUP_add_dir node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/x509.h:581:0, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:156, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/x509_vfy.h:312:0: note: this is the location of the previous definition node_1 | # define X509_LOOKUP_add_dir(x,name,type) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:36:0: warning: "X509_LOOKUP_load_file" redefined node_1 | #define X509_LOOKUP_load_file yaX509_LOOKUP_load_file node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/x509.h:581:0, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:156, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/x509_vfy.h:309:0: note: this is the location of the previous definition node_1 | # define X509_LOOKUP_load_file(x,name,type) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:64:0: warning: "SSL_get_cipher" redefined node_1 | #define SSL_get_cipher yaSSL_get_cipher node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1792:0: note: this is the location of the previous definition node_1 | # define SSL_get_cipher(s) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:65:0: warning: "SSL_get_cipher_name" redefined node_1 | #define SSL_get_cipher_name yaSSL_get_cipher_name node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1798:0: note: this is the location of the previous definition node_1 | # define SSL_get_cipher_name(s) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:76:0: warning: "SSL_CTX_set_session_cache_mode" redefined node_1 | #define SSL_CTX_set_session_cache_mode yaSSL_CTX_set_session_cache_mode node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:2483:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_set_session_cache_mode(ctx,m) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:85:0: warning: "SSL_CTX_set_options" redefined node_1 | #define SSL_CTX_set_options yaSSL_CTX_set_options node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:785:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_set_options(ctx,op) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:92:0: warning: "SSL_CTX_sess_accept" redefined node_1 | #define SSL_CTX_sess_accept yaSSL_CTX_sess_accept node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1208:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_accept(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:93:0: warning: "SSL_CTX_sess_connect" redefined node_1 | #define SSL_CTX_sess_connect yaSSL_CTX_sess_connect node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1202:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_connect(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:94:0: warning: "SSL_CTX_sess_accept_good" redefined node_1 | #define SSL_CTX_sess_accept_good yaSSL_CTX_sess_accept_good node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1212:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_accept_good(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:95:0: warning: "SSL_CTX_sess_connect_good" redefined node_1 | #define SSL_CTX_sess_connect_good yaSSL_CTX_sess_connect_good node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1204:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_connect_good(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:96:0: warning: "SSL_CTX_sess_accept_renegotiate" redefined node_1 | #define SSL_CTX_sess_accept_renegotiate yaSSL_CTX_sess_accept_renegotiate node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1210:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_accept_renegotiate(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from .ibmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:97:0: warning: "SSL_CTX_sess_connect_renegotiate" redefined node_1 | #define SSL_CTX_sess_connect_renegotiate yaSSL_CTX_sess_connect_renegotiate node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1206:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_connect_renegotiate(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:98:0: warning: "SSL_CTX_sess_hits" redefined node_1 | #define SSL_CTX_sess_hits yaSSL_CTX_sess_hits node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1214:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_hits(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:99:0: warning: "SSL_CTX_sess_cb_hits" redefined node_1 | #define SSL_CTX_sess_cb_hits yaSSL_CTX_sess_cb_hits node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1216:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_cb_hits(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:100:0: warning: "SSL_CTX_sess_cache_full" redefined node_1 | #define SSL_CTX_sess_cache_full yaSSL_CTX_sess_cache_full node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1222:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_cache_full(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:101:0: warning: "SSL_CTX_sess_misses" redefined node_1 | #define SSL_CTX_sess_misses yaSSL_CTX_sess_misses node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1218:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_misses(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:102:0: warning: "SSL_CTX_sess_timeouts" redefined node_1 | #define SSL_CTX_sess_timeouts yaSSL_CTX_sess_timeouts node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1220:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_timeouts(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:103:0: warning: "SSL_CTX_sess_number" redefined node_1 | #define SSL_CTX_sess_number yaSSL_CTX_sess_number node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1200:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_number(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:104:0: warning: "SSL_CTX_sess_get_cache_size" redefined node_1 | #define SSL_CTX_sess_get_cache_size yaSSL_CTX_sess_get_cache_size node_1 | ^ node_1 | 'fIn file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:2481:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_get_cache_size(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:110:0: warning: "SSL_CTX_get_session_cache_mode" redefined node_1 | #define SSL_CTX_get_session_cache_mode yaSSL_CTX_get_session_cache_mode node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:2485:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_get_session_cache_mode(ctx) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:111:0: warning: "SSL_session_reused" redefined node_1 | #define SSL_session_reused yaSSL_session_reused node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1990:0: note: this is the location of the previous definition node_1 | # define SSL_session_reused(ssl) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:117:0: warning: "SSL_want_read" redefined node_1 | #define SSL_want_read yaSSL_want_read node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1413:0: note: this is the location of the previous definition node_1 | # define SSL_want_read(s) (SSL_want(s) == SSL_READING) node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:118:0: warning: "SSL_want_write" redefined node_1 | #define SSL_want_write yaSSL_want_write node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1414:0: note: this is the location of the previous definition node_1 | # define SSL_want_write(s) (SSL_want(s) == SSL_WRITING) node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:131:0: warning: "SSL_CTX_sess_set_cache_size" redefined node_1 | #define SSL_CTX_sess_set_cache_size yaSSL_CTX_sess_set_cache_size node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:2479:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_sess_set_cache_size(ctx,t) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:132:0: warning: "SSL_CTX_set_tmp_dh" redefined node_1 | #define SSL_CTX_set_tmp_dh yaSSL_CTX_set_tmp_dh node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:2002:0: note: this is the location of the previous definition node_1 | # define SSL_CTX_set_tmp_dh(ctx,dh) \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:133:0: warning: "OpenSSL_add_all_algorithms" redefined node_1 | #define OpenSSL_add_all_algorithms yaOpenSSL_add_all_algorithms node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/x509.h:73:0, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:156, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/evp.h:912:0: note: this is the location of the previous definition node_1 | # define OpenSSL_add_all_algorithms() \ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:135:0: warning: "SSLeay_add_ssl_algorithms" redefined node_1 | #define SSLeay_add_ssl_algorithms yaSSLeay_add_ssl_algorithms node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:1781:0: note: this is the location of the previous definition node_1 | # define SSLeay_add_ssl_algorithms() SSL_library_init() node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:168:0: warning: "sk_GENERAL_NAME_num" redefined node_1 | #define sk_GENERAL_NAME_num yask_GENERAL_NAME_num node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/crypto.h:129:0, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/comp.h:5, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:149, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/safestack.h:877:0: note: this is the location of the previous definition node_1 | # define sk_GENERAL_NAME_num(st) SKM_sk_num(GENERAL_NAME, (st)) node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:169:0: warning: "sk_GENERAL_NAME_value" redefined node_1 | #define sk_GENERAL_NAME_value yask_GENERAL_NAME_value node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/crypto.h:129:0, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/comp.h:5, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:149, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/safestack.h:878:0: note: this is the location of the previous definition node_1 | # define sk_GENERAL_NAME_value(st, i) SKM_sk_value(GENERAL_NAME, (st), (i)) node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:180:0: warning: "X509_get_notBefore" redefined node_1 | #define X509_get_notBefore yaX509_get_notBefore node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:156:0, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/x509.h:593:0: note: this is the location of the previous definition node_1 | # define X509_get_notBefore(x) ((x)->cert_info->validity->notBefore) node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:30:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/prefix_ssl.h:181:0: warning: "X509_get_notAfter" redefined node_1 | #define X509_get_notAfter yaX509_get_notAfter node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:156:0, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/x509.h:594:0: note: this is the location of the previous definition node_1 | # define X509_get_notAfter(x) ((x)->cert_info->validity->notAfter) node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:34:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/opensslv.h:26:0: warning: "OPENSSL_VERSION_NUMBER" redefined node_1 | #define OPENSSL_VERSION_NUMBER 0x0090700f node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/crypto.h:130:0, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/comp.h:5, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:149, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/opensslv.h:33:0: note: this is the location of the previous definition node_1 | # define OPENSSL_VERSION_NUMBER 0x100020afL node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:457:0: warning: "STACK_OF" redefined node_1 | #define STACK_OF(x) x node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/crypto.h:129:0, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/comp.h:5, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:149, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/safestack.h:88:0: note: this is the location of the previous definition node_1 | # define STACK_OF(type) struct stack_st_##type node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:492:0: warning: "OPENSSL_malloc" redefined node_1 | #define OPENSSL_malloc(x) malloc(x) node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/comp.h:5:0, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:149, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/crypto.h:381:0: note: this is the location of the previous definition node_1 | # define OPENSSL_malloc(num) CRYPTO_malloc((int)num,__FILE__,__LINE__) node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:493:0: warning: "OPENSSL_free" redefined node_1 | #define OPENSSL_free(x) free(x) node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/comp.h:5:0, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:149, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/crypto.h:390:0: note: this is the location of the previous definition node_1 | # define OPENSSL_free(addr) CRYPTO_free(addr) node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:539:0: warning: "SSL_DEFAULT_CIPHER_LIST" redefined node_1 | #define SSL_DEFAULT_CIPHER_LIST "" /* default all */ node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33:0: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:341:0: note: this is the location of the previous definition node_1 | # define SSL_DEFAULT_CIPHER_LIST "ALL:!EXPORT:!LOW:!aNULL:!eNULL:!SSLv2" node_1 | ^ node_1 | In file included from /root/.node-gyp/4.6.0/include/node/openssl/x509.h:96:0, node_1 | from /root/.node-gyp/4.6.0/include/node/openssl/ssl.h:156, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:33: node_1 | /root/.node-gyp/4.6.0/include/node/openssl/rsa.h:181:18: error: expected identifier before numeric constant node_1 | # define RSA_F4 0x10001L node_1 | ^ node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/rsa.h:25:8: note: in expansion of macro 'RSA_F4' node_1 | enum { RSA_F4 = 1 }; node_1 | ^ node_1 | /root/.node-gyp/4.6.0/include/node/openssl/rsa.h:181:18: error: expected '}' before numeric constant node_1 | # define RSA_F4 0x10001L node_1 | ^ node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/rsa.h:25:8: note: in expansion of macro 'RSA_F4' node_1 | enum { RSA_F4 = 1 }; node_1 | ^ node_1 | /root/.node-gyp/4.6.0/include/node/openssl/rsa.h:181:18: error: expected unqualified-id before numeric constant node_1 | # define RSA_F4 0x10001L node_1 | ^ node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/rsa.h:25:8: note: in expansion of macro 'RSA_F4' node_1 | enum { RSA_F4 = 1 }; node_1 | ^ node_1 | In file included from ../deps/libmariadbclient/extra/yassl/include/openssl/ssl.h:35:0, node_1 | from ../deps/libmariadbclient/extra/yassl/include/cert_wrapper.hpp:38, node_1 | from ../deps/libmariadbclient/extra/yassl/include/yassl_int.hpp:31, node_1 | from ../deps/libmariadbclient/extra/yassl/src/ssl.cpp:35: node_1 | ../deps/libmariadbclient/extra/yassl/include/openssl/rsa.h:25:19: error: expected declaration before '}' token node_1 | enum { RSA_F4 = 1 }; node_1 | ^ node_1 | deps/libmariadbclient/extra/yassl/yassl.target.mk:133: recipe for target 'Release/obj.target/yassl/deps/libmariadbclient/extra/yassl/src/ssl.o' failed node_1 | make: *** [Release/obj.target/yassl/deps/libmariadbclient/extra/yassl/src/ssl.o] Error 1 node_1 | make: Leaving directory '/var/www/node_modules/mariasql/build' node_1 | gyp ERR! build error node_1 | gyp ERR! stack Error: `make` failed with exit code: 2 node_1 | gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/npm/node_modules/node-gyp/lib/build.js:276:23) node_1 | gyp ERR! stack at emitTwo (events.js:87:13) node_1 | gyp ERR! stack at ChildProcess.emit (events.js:172:7) node_1 | gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:200:12) node_1 | gyp ERR! System Linux 4.4.24-moby node_1 | gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/npm/node_modules/node-gyp/bin/node-gyp.js" "rebuild" node_1 | gyp ERR! cwd /var/www/node_modules/mariasql node_1 | gyp ERR! node -v v4.6.0 node_1 | gyp ERR! node-gyp -v v3.4.0 node_1 | gyp ERR! not ok node_1 | npm info mariasql@0.1.24 Failed to exec install script node_1 | npm ERR! Linux 4.4.24-moby node_1 | npm ERR! argv "/usr/local/bin/node" "/usr/local/bin/npm" "install" node_1 | npm ERR! node v4.6.0 node_1 | npm ERR! npm v2.15.9 node_1 | npm ERR! code ELIFECYCLE node_1 | node_1 | npm ERR! mariasql@0.1.24 install: `node-gyp rebuild` node_1 | npm ERR! Exit status 1 node_1 | npm ERR! node_1 | npm ERR! Failed at the mariasql@0.1.24 install script 'node-gyp rebuild'. node_1 | npm ERR! This is most likely a problem with the mariasql package, node_1 | npm ERR! not with npm itself. node_1 | npm ERR! Tell the author that this fails on your system: node_1 | npm ERR! node-gyp rebuild node_1 | npm ERR! You can get information on how to open an issue for this project with: node_1 | npm ERR! npm bugs mariasql node_1 | npm ERR! Or if that isn't available, you can get their info via: node_1 | npm ERR! node_1 | npm ERR! npm owner ls mariasql node_1 | npm ERR! There is likely additional logging output above. node_1 | npm info preuninstall mariasql@0.1.24 node_1 | npm info uninstall mariasql@0.1.24 node_1 | npm info postuninstall mariasql@0.1.24 node_1 | node_1 | npm ERR! Please include the following file with any support request: node_1 | npm ERR! /var/www/npm-debug.log Assuming you're using gcc/g++, what does gcc -v show? its gcc version 4.9.2 (Debian 4.9.2-10) am using node:4 from docker https://hub.docker.com/_/node/ I can duplicate the issue with node v4.6.0 and mariasql v0.1.24, but I'm not sure offhand at what point the problem got fixed, since mariasql v0.2.6 installed just fine with node 4.6.0. My suggestion would be to upgrade, since v0.1.x is not maintained anymore. understood .. I wish this can be an option, but sadly its not .. any way thanks
gharchive/issue
2016-10-16T07:41:15
2025-04-01T06:45:04.605006
{ "authors": [ "linuxjuggler", "mscdex" ], "repo": "mscdex/node-mariasql", "url": "https://github.com/mscdex/node-mariasql/issues/163", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2189926841
🛑 Do53 Roost IPv4 TCP is down In a6f992c, Do53 Roost IPv4 TCP (http://107.189.10.142:9203) was down: HTTP code: 0 Response time: 0 ms Resolved: Do53 Roost IPv4 TCP is back up in 29fd955 after 9 minutes.
gharchive/issue
2024-03-16T11:39:26
2025-04-01T06:45:04.627699
{ "authors": [ "mschirrmeister" ], "repo": "mschirrmeister/upptime-loopx", "url": "https://github.com/mschirrmeister/upptime-loopx/issues/2962", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2216658508
🛑 Do53 Roost IPv4 TCP is down In 38ea7e2, Do53 Roost IPv4 TCP (http://107.189.10.142:9203) was down: HTTP code: 0 Response time: 0 ms Resolved: Do53 Roost IPv4 TCP is back up in 7ac5e93 after 23 minutes.
gharchive/issue
2024-03-30T20:43:12
2025-04-01T06:45:04.630094
{ "authors": [ "mschirrmeister" ], "repo": "mschirrmeister/upptime-loopx", "url": "https://github.com/mschirrmeister/upptime-loopx/issues/4575", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2275772259
CUDNN_STATUS_NOT_SUPPORTED (fid) 1104:~/wav/pytorch-fid-master/src$ python -m pytorch_fid pytorch-fid-master/real pytorch-fid-master/fake --device cuda:0 anaconda3/envs/fid/lib/python3.8/site-packages/torch/nn/modules/conv.py:456: UserWarning: Plan failed with a cudnnException: CUDNN_BACKEND_EXECUTION_PLAN_DESCRIPTOR: cudnnFinalize Descriptor Failed cudnn_status: CUDNN_STATUS_NOT_SUPPORTED (Triggered internally at /opt/conda/conda-bld/pytorch_1712608843393/work/aten/src/ATen/native/cudnn/Conv_v8.cpp:919.) return F.conv2d(input, weight, bias, self.stride, FID: 206.58618781110312 请问一下运行时遇到这个错误怎么解决 I also see this same message. Although I'm not sure whether it has affected my output: conv.py:456: UserWarning: Plan failed with a cudnnException: CUDNN_BACKEND_EXECUTION_PLAN_DESCRIPTOR: cudnnFinalize Descriptor Failed cudnn_status: CUDNN_STATUS_NOT_SUPPORTED (Triggered internally at /opt/conda/conda-bld/pytorch_1712608843393/work/aten/src/ATen/native/cudnn/Conv_v8.cpp:919.) return F.conv2d(input, weight, bias, self.stride,
gharchive/issue
2024-05-02T15:03:46
2025-04-01T06:45:04.638432
{ "authors": [ "guo-king666", "shreyas1230" ], "repo": "mseitzer/pytorch-fid", "url": "https://github.com/mseitzer/pytorch-fid/issues/111", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1802124798
Does set logit_bias to 100 lead the model choice between 'A' and 'B' essentially random? logit_bias={ '32': 100, # 'A' token '33': 100, # 'B' token }, I asked GPT4 this issue, the following is its answer: Not random! It essentially just excludes any other tokens. Unlike the explanation in the gpt-4 response, I think as @mshumer already said it just limits the output to either 'A' or 'B'.
gharchive/issue
2023-07-13T04:17:58
2025-04-01T06:45:04.640476
{ "authors": [ "Phq-art", "mshumer", "vegetax" ], "repo": "mshumer/gpt-prompt-engineer", "url": "https://github.com/mshumer/gpt-prompt-engineer/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1092650198
High cpu usage 100% cpu spikes are not unusual to see when I open a package.json The extension has a unusually high startup time from time to time I'm not too surprised, this is an extension I made a while back and never really got to a great working state as there are certainly performance issues with it. I don't really use or maintain it anymore, so I'm happy to leave this issue open but I likely won't work on it myself. @shiftgeist Can you try again with the latest released version 2.0.3? The new version is definitly a huge improvement towards performance. I'm only seeing startup times up to 700 ms with the average beeing more like 400 ms. CPU usage while opening a package.json has also gone down. Great work!
gharchive/issue
2022-01-03T16:31:58
2025-04-01T06:45:04.642758
{ "authors": [ "mskelton", "rentalhost", "shiftgeist" ], "repo": "mskelton/vscode-npm-outdated", "url": "https://github.com/mskelton/vscode-npm-outdated/issues/7", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
233380496
Jump variable casting [x] Implement feature / fix bug [x] Add tests [x] Update changelog Changed the casting of jump variables so that they are cast to native python types (i.e. floats and ints). The problem was list(choices) keeps the types in the list intact and only changes the container to a python list. choices.tolist() changes the container AND the datatypes to native python types. 👍 Fantastic!
gharchive/pull-request
2017-06-03T16:38:14
2025-04-01T06:45:04.644949
{ "authors": [ "RobertArbon", "cxhernandez" ], "repo": "msmbuilder/osprey", "url": "https://github.com/msmbuilder/osprey/pull/224", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
565311350
Update nginx.conf Fix missing option inside nginx.conf [error] 1#1: [lua] init_master.lua:113: init_master(): auto-ssl: dict auto_ssl_settings could not be found. Please add it to your configuration: `lua_shared_dict auto_ssl_settings 64k;` Inactive GitHub user. 😞
gharchive/pull-request
2020-02-14T12:56:54
2025-04-01T06:45:04.654300
{ "authors": [ "mdzidic" ], "repo": "msumpter/docker-lua-resty-auto-ssl", "url": "https://github.com/msumpter/docker-lua-resty-auto-ssl/pull/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1780366187
Fix OpSum to TTN converter for multiple on-site operators In the spirit of better late than never, a fix for #62. I still need to figure out the problem with the manual svd compression so we can default to that and remove the current redundancy, but this might take some time so I just patched the current working version for now. The issue was that the finite state machine representation just overwrote the on-site entry every time an on-site term was added; I just didn't think of on-site sums writing it. Wrapping everything in a Sum{...} fixes it. I added a small test. Codecov Report Merging #95 (6a5fb6d) into main (34fd5dc) will increase coverage by 0.04%. The diff coverage is 100.00%. :exclamation: Your organization is not using the GitHub App Integration. As a result you may experience degraded service beginning May 15th. Please install the Github App Integration for your organization. Read more. @@ Coverage Diff @@ ## main #95 +/- ## ========================================== + Coverage 79.25% 79.30% +0.04% ========================================== Files 61 61 Lines 3442 3450 +8 ========================================== + Hits 2728 2736 +8 Misses 714 714 Impacted Files Coverage Δ src/treetensornetworks/opsum_to_ttn.jl 90.11% <100.00%> (+0.23%) :arrow_up: Thanks for circling back to this @leburgel! Do you remember what still needed to be done to make the OpSum to TTN conversion feature complete with OpSum to MPO conversion in ITensors.jl? I seem to recall that it doesn't support QNs, and also that it may have had some issues for longer range interactions, even in the dense case. The default current default implementation should work for any long range interactions (at least for anything I've tried so far), the only problem is that the resulting network has a bond dimension that is larger than necessary. There is an issue when I manually compress the network by merging interaction channels (similar to what is done in ITensors.jl) that I haven't figured out, so I've just switched off the compression by default for now until this is fixed. So right now everything should work correctly as is, it's just not optimal. I should have time to try and fix this soon though. Everything with QN conservation errors by default right now I think. For fermions, there's definitely things in the converter that are just plain broken. It's hard for me to tell how easily this can be fixed though, since I'm not really familiar with how ITensors.jl handles fermions. For regular symmetries however, I don't think there should be any fundamental issues. I've just never tested any of the tree tensor network code with QN number support so this errors more as a precaution. It should be a matter of trying it and seeing whether things break at some point. I can also try to have a look at this soon.
gharchive/pull-request
2023-06-29T08:28:52
2025-04-01T06:45:04.719431
{ "authors": [ "codecov-commenter", "leburgel", "mtfishman" ], "repo": "mtfishman/ITensorNetworks.jl", "url": "https://github.com/mtfishman/ITensorNetworks.jl/pull/95", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1694678336
🛑 https://reports.mtit.pna.ps is down In 46bdee6, https://reports.mtit.pna.ps (https://reports.mtit.pna.ps) was down: HTTP code: 0 Response time: 0 ms Resolved: https://reports.mtit.pna.ps is back up in 9ee73d4.
gharchive/issue
2023-05-03T19:34:46
2025-04-01T06:45:04.743656
{ "authors": [ "mtitservice" ], "repo": "mtitservice/site", "url": "https://github.com/mtitservice/site/issues/10601", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1703242832
🛑 https://db.hcys.ps is down In f6c6437, https://db.hcys.ps (https://db.hcys.ps) was down: HTTP code: 0 Response time: 0 ms Resolved: https://db.hcys.ps is back up in 63a006c.
gharchive/issue
2023-05-10T06:53:55
2025-04-01T06:45:04.746789
{ "authors": [ "mtitservice" ], "repo": "mtitservice/site", "url": "https://github.com/mtitservice/site/issues/11680", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1709472209
🛑 https://takaful.salfeet.plo.ps is down In 61f33e1, https://takaful.salfeet.plo.ps (https://takaful.salfeet.plo.ps) was down: HTTP code: 0 Response time: 0 ms Resolved: https://takaful.salfeet.plo.ps is back up in 6e2cc13.
gharchive/issue
2023-05-15T07:19:25
2025-04-01T06:45:04.749791
{ "authors": [ "mtitservice" ], "repo": "mtitservice/site", "url": "https://github.com/mtitservice/site/issues/12369", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1623298407
🛑 https://salfeet.plo.ps is down In eb6820f, https://salfeet.plo.ps (https://salfeet.plo.ps) was down: HTTP code: 0 Response time: 0 ms Resolved: https://salfeet.plo.ps is back up in bbe645e.
gharchive/issue
2023-03-14T11:39:25
2025-04-01T06:45:04.752938
{ "authors": [ "mtitservice" ], "repo": "mtitservice/site", "url": "https://github.com/mtitservice/site/issues/2927", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1646893414
🛑 https://reports.mtit.pna.ps is down In 02dba72, https://reports.mtit.pna.ps (https://reports.mtit.pna.ps) was down: HTTP code: 0 Response time: 0 ms Resolved: https://reports.mtit.pna.ps is back up in c4b63dd.
gharchive/issue
2023-03-30T04:31:47
2025-04-01T06:45:04.755862
{ "authors": [ "mtitservice" ], "repo": "mtitservice/site", "url": "https://github.com/mtitservice/site/issues/5108", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1669322268
🛑 https://reports.mtit.pna.ps is down In 5e6a591, https://reports.mtit.pna.ps (https://reports.mtit.pna.ps) was down: HTTP code: 0 Response time: 0 ms Resolved: https://reports.mtit.pna.ps is back up in bd6d1ee.
gharchive/issue
2023-04-15T11:40:50
2025-04-01T06:45:04.758773
{ "authors": [ "mtitservice" ], "repo": "mtitservice/site", "url": "https://github.com/mtitservice/site/issues/7638", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1689586278
🛑 https://takaful.salfeet.plo.ps is down In 00a9f4c, https://takaful.salfeet.plo.ps (https://takaful.salfeet.plo.ps) was down: HTTP code: 0 Response time: 0 ms Resolved: https://takaful.salfeet.plo.ps is back up in 90b618f.
gharchive/issue
2023-04-29T14:54:05
2025-04-01T06:45:04.761922
{ "authors": [ "mtitservice" ], "repo": "mtitservice/site", "url": "https://github.com/mtitservice/site/issues/9795", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1183775785
[Update] Update Nimlang Binaries WTF Bin with this gem of a screenshot Added in 522fd55
gharchive/issue
2022-03-28T17:45:33
2025-04-01T06:45:04.878753
{ "authors": [ "HuskyHacks", "mttaggart" ], "repo": "mttaggart/wtfbins", "url": "https://github.com/mttaggart/wtfbins/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2613244740
How to host a live 3D view with Flask, vite, react-three-fiber with my custom data? When i run visualize_collection(vector_store._collection), i just get the raw data in web page. I have no idea about web develop, could you show me how to visualize the data with vite as your screenshot demonstration this might help you https://github.com/mtybadger/chromaviz/issues/15 if you've downloaded the project and run the test file, you probably want to visualize an existing database you have? in which case, you need a different script to point to a persistent chromadB database (SQLLite file) the article above shows how to do that. when its working, all you see are coloured data points with lines between them, where if you hover your mouse over any node/data point you see the id, document, and metadata it represents. NOTE: the collection you point it to must have 50+ documents, otherwise you'll get aa 500 internal server error - see here: https://github.com/mtybadger/chromaviz/issues/6
gharchive/issue
2024-10-25T06:49:51
2025-04-01T06:45:04.881513
{ "authors": [ "WithFoxSquirrel", "coffeecodeconverter" ], "repo": "mtybadger/chromaviz", "url": "https://github.com/mtybadger/chromaviz/issues/13", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2359447729
Versions shown in POST object (and GET status) are incorrect curl -X 'POST' \ 'http://127.0.0.1:8000/api/taxii2/demo_database/collections/blog/objects/' \ -H 'accept: application/taxii+json' \ -H 'Authorization: Basic cmVhZF93cml0ZV91c2VyOnRlc3RpbmcxMjM=' \ -H 'Content-Type: application/json' \ -H 'X-CSRFTOKEN: PeGDYiBqxZSTpfQglf9Ns2wjXqxR1DimGbvF809eZftyoPgbG7OtY35bBYcBaKZm' \ -d '{ "objects": [ { "type": "attack-pattern", "spec_version": "2.1", "id": "attack-pattern--6b948b5a-3c09-5365-b48a-da95c3964cb5", "created_by_ref": "identity--d2916708-57b9-5636-8689-62f049e9f727", "created": "2020-01-01T11:21:07.478851Z", "modified": "2020-01-01T11:21:07.478851Z", "name": "Spear Phishing", "description": "Used for tutorial content", "object_marking_refs": ["marking-definition--34098fce-860f-48ae-8e50-ebd3cc5e41da"] }, { "type": "attack-pattern", "spec_version": "2.1", "id": "attack-pattern--6b948b5a-3c09-5365-b48a-da95c3964cb5", "created_by_ref": "identity--d2916708-57b9-5636-8689-62f049e9f727", "created": "2020-01-02T11:21:07.478851Z", "modified": "2020-01-02T11:21:07.478851Z", "name": "Spear Phishing Updated ONCE", "description": "Used for tutorial content", "object_marking_refs": ["marking-definition--34098fce-860f-48ae-8e50-ebd3cc5e41da"] }, { "type": "attack-pattern", "spec_version": "2.1", "id": "attack-pattern--6b948b5a-3c09-5365-b48a-da95c3964cb5", "created_by_ref": "identity--d2916708-57b9-5636-8689-62f049e9f727", "created": "2020-01-03T11:21:07.478851Z", "modified": "2020-01-03T11:21:07.478851Z", "name": "Spear Phishing Updated TWICE", "description": "Used for tutorial content", "object_marking_refs": ["marking-definition--34098fce-860f-48ae-8e50-ebd3cc5e41da"] } ] }' prints { "id": "da62e209-cf48-40f3-a687-b89b0cde5546", "status": "pending", "total_count": 3, "success_count": 0, "successes": [], "failure_count": 0, "failures": [], "pending_count": 3, "pendings": [ { "message": null, "version": "2024-06-18 09:48:30.879800+00:00", "id": "attack-pattern--6b948b5a-3c09-5365-b48a-da95c3964cb5" }, { "message": null, "version": "2024-06-18 09:48:30.879800+00:00", "id": "attack-pattern--6b948b5a-3c09-5365-b48a-da95c3964cb5" }, { "message": null, "version": "2024-06-18 09:48:30.879800+00:00", "id": "attack-pattern--6b948b5a-3c09-5365-b48a-da95c3964cb5" } ], "request_timestamp": "2024-06-18T09:48:30.879800Z" } See versions show request time. They should show each modified property for the objects posted. The exception to this is objects with no modified time, in which case the object _record_modified time is used (which I think is the current behaviour). Are you able to post objects successfully? Since this push i cannot get it to work correctly
gharchive/issue
2024-06-18T09:55:17
2025-04-01T06:45:04.894969
{ "authors": [ "himynamesdave" ], "repo": "muchdogesec/arango_taxii_server", "url": "https://github.com/muchdogesec/arango_taxii_server/issues/17", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2365955693
WebUI Filter Tag issue Hi, I'm using v2.17.1 on my linux server and if I enter the WebUi with /browser, I see >19k of filter tags, which makes it impossible to use the page for a while. Best Frank @FrankDaze are you adding/having a separate repository? where all these tags are coming from? I have only this setting in the env file: GALLERIES=[{"name":"model-gallery", "url":"github:go-skynet/model-gallery/index.yaml"}, {"url": "github:go-skynet/model-gallery/huggingface.yaml","name":"huggingface"}] Hi, Same problem, same configuration, the '/browse/' completely freeze the browser @FrankDaze - @heralight I'm sorry but that repository isn't maintained, so you should adapt the GALLERIES to drop the huggingface one. I'd suggest to delete it from the configuration. The only one maintained now is the LocalAI one which ships pre-configured in the images. However the problem that a large gallery might crash the UI is still an issue, but that needs more enhancements like pagination, which is tracked in #2156 @mudler thank you! @FrankDaze , to fix it I change my .env to default as: LOCALAI_GALLERIES=[{"name":"localai", "url":"github:mudler/LocalAI/gallery/index.yaml@master"}] #GALLERIES=[{"name":"model-gallery", "url":"github:go-skynet/model-gallery/index.yaml"}] Thanks it's working now.
gharchive/issue
2024-06-21T07:56:17
2025-04-01T06:45:04.930580
{ "authors": [ "FrankDaze", "heralight", "mudler" ], "repo": "mudler/LocalAI", "url": "https://github.com/mudler/LocalAI/issues/2619", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
428852483
[Tooltip/Popper]: keepMounted doesn't work when open updated before Portal updates [x] This is not a v0.x issue. [x] I have searched the issues of this repository and believe that this is not a duplicate. Expected Behavior 🤔 When the PopperProps.keepMounted prop is set to false, I expect the <Tooltip /> / <Popper /> / <Portal /> component to not exist in the DOM whenever the Tooltip is closed (open prop set to false). If the open prop changes from true to false before the <Portal /> component gets its' mount node and re-renders, I'd expect the Tooltip to never exist on the DOM (or at least immediately clean itself up after attaching to the DOM) Current Behavior 😯 Not that! If the open prop is changed from false to true, and then set back to false within the next update phase (or the setState callback), the Tooltip is added to the DOM and not removed, even though open is false. Steps to Reproduce 🕹 Link: https://codesandbox.io/embed/4jjq50051x Open DevTools so you can see the React DOM root (It's nested inside the iframe, so you might want to search for "Index-root- to find it faster) Press the "TOGGLE TOOLTIP" button - This will set the state to open, and then set the state to false on the next update (setState callback) Note that the Tooltip Portal element is now on the DOM, but not visible on the screen If you comment out line 34, and uncomment lines 36-38, this works fine because the closing setState is now after the next update phase, allowing the react-transition-group Transition component to fully render and set its' internal state to ENTERED. When the component updates to in: false, it will then invoke MUI's Grow and Popper's handleExited functions and remove the Portal from the DOM. Context 🔦 We have an RxJS stream which is indirectly used in hiding/showing a component Tooltip. This can cause the Tooltip to have open: true and then open: false in the update phase immediately following, which causes this issue. We use the fact that the Tooltip unmounts in our functional tests, when we're waiting (or testing) for the removal of the Tooltip. Therefore, this bug causes our functional tests to fail. It obviously also causes our DOM to be littered with redundant tooltips which shouldn't exist. Your Environment 🌎 Tech Version Material-UI v3.9.3 React v16.8.6 Browser Chrome 73 TypeScript v3.4.1 @Sharakai We most likely have the same issue with the Modal component. What do you think of this change? --- a/packages/material-ui/src/Popper/Popper.js +++ b/packages/material-ui/src/Popper/Popper.js @@ -66,23 +66,6 @@ class Popper extends React.Component { this.handleClose(); } - static getDerivedStateFromProps(nextProps) { - if (nextProps.open) { - return { - exited: false, - }; - } - - if (!nextProps.transition) { - // Otherwise let handleExited take care of marking exited. - return { - exited: true, - }; - } - - return null; - } - handleOpen = () => { const { anchorEl, modifiers, open, placement, popperOptions = {}, disablePortal } = this.props; const popperNode = this.tooltipRef.current; @@ -126,6 +109,10 @@ class Popper extends React.Component { } }; + handleEntered = () => { + this.setState({ exited: false }); + }; + handleExited = () => { this.setState({ exited: true }); this.handleClose(); @@ -174,6 +161,7 @@ class Popper extends React.Component { childProps.TransitionProps = { in: open, onExited: this.handleExited, + onEntered: this.handleEntered, }; } It seems to solve the problem and to help in the upcoming class -> hook migration cc @joshwooding. It would be great to add a failing test case for your codesandbox as well as migrate the Modal component if this solution is satisfactory. Do you want to give it a shot? :) This would definitely simplify the hook migration :) Modal might be slightly more difficult though Right you are with the Modal component. I've made a codesandbox to repro that, too: https://codesandbox.io/s/2xolvp57n0 I've added a failing test case for both Popper and Modal, as well as a few others to ensure the keepMounted prop works as expected for both components. @oliviertassinari The proposed change works fine, though I'm still trying to grok how... @joshwooding If there's anything I can do to help with the hooks migration, I'd be more than happy to help. I'll make this change sans-hooks, of course, but happy to look at switching either/or later on. @Sharakai More than happy for you to help. I've created an umbrella issue with the components we have left (#15231). I will warn you that some of the ones left are quite difficult
gharchive/issue
2019-04-03T16:05:32
2025-04-01T06:45:04.950039
{ "authors": [ "Sharakai", "joshwooding", "oliviertassinari" ], "repo": "mui-org/material-ui", "url": "https://github.com/mui-org/material-ui/issues/15180", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
988368102
Combine multiple pseudo selector Description: Support multiple pesudo selectors Current: Using makestyle, if we want similar CSS properties on active and hover, we have to repeat our code, Expected: There should be a way in which we can combine multiple pseudo selectors Tip, you can use the same comma syntax as with CSS
gharchive/issue
2021-09-04T20:58:04
2025-04-01T06:45:04.952433
{ "authors": [ "oliviertassinari", "umerjaved178" ], "repo": "mui-org/material-ui", "url": "https://github.com/mui-org/material-ui/issues/28140", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
363766652
Introduce StepIconComponent prop in StepLabel Add a StepIconComponent prop in StepLabel to overwrite StepIcon standard behaviour with the provided component. The component will be instantiated with the completed/active/error props. Unit tests are OK. fix #12987 To fix CI run in the workspace root: yarn docs:api yarn prettier @eps1lon Thanks for all the input. I'm a bit uncomfortable with this proposal, as similar props of StepLabel are using nodes (icon, optional) and not components. I'm afraid this could cause confusion on how to use this new prop. The simplest way to achieve the same behaviour would be to clone the icon to inject props in the StepIcon component, but that would introduce a slight behavioural change. What's your opinion ? @semos It's a great first pull request on Material-UI 👌🏻. Thank you for working on it!
gharchive/pull-request
2018-09-25T21:07:08
2025-04-01T06:45:04.955612
{ "authors": [ "eps1lon", "oliviertassinari", "semos" ], "repo": "mui-org/material-ui", "url": "https://github.com/mui-org/material-ui/pull/13003", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1433429482
Stability of CssVarsProvider / extendTheme Hello! Our organization is interested in using the recent CssVarsProvider / extendTheme functionality documented here. However, the names of the exports still including "experimental" gives us pause. Are these features still considered experimental? Is there a roadmap to stabilization? Thank you! Are these features still considered experimental? To be honest, Yes but it is in a good shape to try it out. I don't think there will be a big change in the near future (will make it stable in the next major release of v6). I'd encourage you to try it out and let us know how it goes because it is an opt-in feature without breaking changes so I think it should not impact your business logics. Is there a roadmap to stabilization? We are incrementally migrating https://mui.com to use the new APIs to stress test the feature and we have been fixing known issues along the way. So we are stabilizing it. Thank you @siriwatknp !!!
gharchive/issue
2022-11-02T16:18:48
2025-04-01T06:45:04.959125
{ "authors": [ "aaronlademann-wf", "siriwatknp" ], "repo": "mui/material-ui", "url": "https://github.com/mui/material-ui/issues/34989", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1844595678
[Tabs] style overrides function gets incorrectly typed props object if variant='scrollable' Duplicates [X] I have searched the existing issues Latest version [X] I have tested the latest version Steps to reproduce 🕹 Link to live example: https://codesandbox.io/s/mui-tabs-style-override-crash-p9wfjm-p9wfjm?file=/package.json Downgrading the package version in the live example will "fix" the error. Steps: Use a Tabs component, set its variant to scrollable. Create a custom theme which uses a function to override Tabs styling. This function uses the props parameter to decide which styles to apply. The ownerState.orientation field is deconstructed out of the object. Current behavior 😯 The rendering crashes, stating that _ref.ownerState is undefined (where _ref is the name of the props parameter after downleveling). This happens because the props object does not have the ownerState prop, even though its typings suggest that it does have that prop. Expected behavior 🤔 The props object should have the ownerState prop, as the typings describe. Context 🔦 This appears to happen because TabsScrollbarSize is being rendered. It uses the same style overrides as the other Tabs components, but it does not get an ownerState prop. Thus, it is undefined. Your environment 🌎 This error started appearing during routine package upgrades. In one package, it began to occur after upgrading from 5.14.2 to 5.14.4. But in others, the upgrade to 5.14.4 did not trigger the same issue, and only appeared when we forced nested dependencies to upgrade as well. We are not completely sure, but believe that it is related to @mui/system. Downgrading that package appears to resolve the issue on the latest version of @mui/material. "resolutions": { "@mui/system": "5.14.1" } It happens in all browsers. npx @mui/envinfo System: OS: Windows 10 10.0.22621 Binaries: Node: 18.17.0 - C:\Program Files\nodejs\node.EXE Yarn: 1.22.19 - C:\Program Files\nodejs\yarn.CMD npm: 9.6.7 - C:\Program Files\nodejs\npm.CMD Browsers: Chrome: Not Found Edge: Spartan (44.22621.1992.0), Chromium (115.0.1901.200) npmPackages: @emotion/react: 11.11.1 => 11.11.1 @emotion/styled: 11.11.0 => 11.11.0 @mui/base: 5.0.0-beta.10 @mui/core-downloads-tracker: 5.14.4 @mui/material: 5.14.4 => 5.14.4 @mui/private-theming: 5.14.4 @mui/styled-engine: 5.13.2 @mui/system: 5.14.4 @mui/types: 7.2.4 @mui/utils: 5.14.4 @types/react: 18.2.18 => 18.2.18 react: 18.2.0 => 18.2.0 react-dom: 18.2.0 => 18.2.0 typescript: 5.1.6 => 5.1.6 The proposed solution above works fine on my end. 👍 Would you like to submit a PR?
gharchive/issue
2023-08-10T07:34:31
2025-04-01T06:45:04.966695
{ "authors": [ "pvdstel", "siriwatknp" ], "repo": "mui/material-ui", "url": "https://github.com/mui/material-ui/issues/38406", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2004950666
[material-ui][docs] Improve documentation for "complementary" components Feedback received: Please include some example code in the demo. There are four different components listed at the bottom of this page, the API docs for three of them have no details about what they do or how to use them, they link back to this page with a "see our demo" message but then are ONLY mentioned at the very bottom in the link to the page I was just on. Documentation should be better than this. Show me how to use your system, I can't read the minds of strangers, let alone strangers who are obviously several orders of magnitude more experienced than I am if they built this library. from https://mui.com/material-ui/react-accordion/# I can empathize with this frustration. Material UI has lots of these complementary components (or maybe supplementary is a better term?) that don't have their own demo pages, because you would only ever use them in the context of their parent component. But they're not consistently documented in the parent demo docs either, leading to this awful UX loop: Look for info on Accordion Actions component usage on the Accordion page Find the link to the Accordion Actions API page and click through The API page, in turn, redirects you back to the Accordion demo page, which has no info about Accordion Actions other than the link to the API page Not fun! I think we've largely resolved this issue in the Joy and Base docs by consistently documenting these complementary components in their respective parent component demo pages. We need to comb through the Material UI docs to find these undocumented components and show users what to do with them. I'd be interested in resolving this issue, can I help? Here are the components I've found that seem to lack clear documentation for complementary components: INPUTS [ ] Transfer list DATA DISPLAY [ ] List [ ] Table FEEDBACK [ ] Dialog [ ] Snackbar SURFACES [ ] Accordion #40284 [ ] Card NAGIVATION [ ] Menu [ ] Stepper [ ] Tabs Created an issue for the tabs complementary components: #40558 An issue was created for the Stepper complementary components: https://github.com/mui/material-ui/issues/41454
gharchive/issue
2023-11-21T18:58:51
2025-04-01T06:45:04.977353
{ "authors": [ "Babak1v9", "DiegoAndai", "anle9650", "samuelsycamore" ], "repo": "mui/material-ui", "url": "https://github.com/mui/material-ui/issues/39952", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1890452409
[base-ui]: extended getInputProps with React.Ref in useAutocomplete [x] I have followed (at least) the PR section of the contributing guide. The type of the UseAutocompleteReturnValue.getInputProps (in packages\mui-base\src\useAutocomplete\useAutocomplete.d.ts) needed to be extended and include the ref: React.Ref field. @michaldudak @sai6855 can you please help me out, why my 3 checks failed? @DarhkVoyd The test_static fails because you did not run yarn docs:api. See the Contributing doc for required steps when submitting a PR. As for the test_types, it seems that your change has already helped find issues in our demos :) The docs\data\base\components\autocomplete\AutocompleteIntroduction\system\index.tsx and docs\data\base\components\autocomplete\AutocompleteIntroduction\tailwind\index.tsx incorrectly set their input refs (as getInputProps override whatever is set before). Since the demos work fine, it seems that setting the refs manually is not needed, so you can delete the offending lines. Make sure to run yarn docs:typescript:formatted afterward. @michaldudak I have made necessary changes. @michaldudak can you please guide me what's next? I would like to do more, where can I start. @DarhkVoyd feel free to browse the issues. We've got the good first issue and ready to take labels you can check out first.
gharchive/pull-request
2023-09-11T12:56:23
2025-04-01T06:45:04.982424
{ "authors": [ "DarhkVoyd", "michaldudak" ], "repo": "mui/material-ui", "url": "https://github.com/mui/material-ui/pull/38919", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2088992156
[docs] Fix broken styles on Base UI page Fixed the /base-ui page that got broken after changing the Base UI class prefix to base- Preview: https://deploy-preview-40683--material-ui.netlify.app/base-ui The demos of the docs were broken to. Great to see this fix.
gharchive/pull-request
2024-01-18T20:20:42
2025-04-01T06:45:04.984061
{ "authors": [ "michaldudak", "oliviertassinari" ], "repo": "mui/material-ui", "url": "https://github.com/mui/material-ui/pull/40683", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2611241395
[AvatarGroup] Fix spacing prop [x] I have followed (at least) the PR section of the contributing guide. Fixes #44172 The bug is still present. This change only fixes the spacing for the "surplus" avatar. You can compare the behavior between this PR's version and v5 by changing adding the spacing prop in the docs demos: https://deploy-preview-44202--material-ui.netlify.app/material-ui/react-avatar/#grouped https://v5.mui.com/material-ui/react-avatar/#grouped
gharchive/pull-request
2024-10-24T11:11:18
2025-04-01T06:45:04.986921
{ "authors": [ "aarongarciah", "navedqb" ], "repo": "mui/material-ui", "url": "https://github.com/mui/material-ui/pull/44202", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2492485883
Replace endsWith with includes to check adapter path Fix this problem https://github.com/mui/material-ui/issues/43471#issuecomment-2315676829 [x] I have followed (at least) the PR section of the contributing guide. https://github.com/mui/pigment-css/issues/223 This issue starts after this is merged! @siriwatknp
gharchive/pull-request
2024-08-28T15:59:24
2025-04-01T06:45:05.033987
{ "authors": [ "Nayeem-XTREME", "siriwatknp" ], "repo": "mui/pigment-css", "url": "https://github.com/mui/pigment-css/pull/214", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1195992670
Add design doc (draft version) Signed-off-by: Ying Mo morningspace@yahoo.com Description of your changes This PR is used to track the design of Ansible provider. I have: [x] Read and followed Crossplane's contribution process. [ ] Run make reviewable to ensure this PR is ready for review. [ ] Added backport release-x.y labels to auto-backport this PR if necessary. How has this code been tested @morningspace Great documentation! this rase multiple features I will try to create associated ticket design schema recommendation: After a successful requeuing, a new reconciliation will be triggered and the first step of this new reconciliation will again be a connection I think we have to bring up the notion of the runner in the schema . The connect generate a runner that will implement a CRUD Connect -> runner -> Observe -> create/update/delete For doc: Adding a list with implement and not implemented feature , we can add an emoticon in front unimplemented features
gharchive/pull-request
2022-04-07T12:42:33
2025-04-01T06:45:05.060120
{ "authors": [ "fahedouch", "morningspace" ], "repo": "multicloudlab/crossplane-provider-ansible", "url": "https://github.com/multicloudlab/crossplane-provider-ansible/pull/27", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1813401116
Upgrade bootstrap.sh Correct vagrant.sh I made a cleanup to make it more consistent and readable. Please note the new 'order' and numbering of the installation process, e.g. moved insite up to step 4. please check the NEST DESKTOP and INSITE installation. please have a quick look at the changes to NEST installation I would suggest to change the path to plan XML file when generating the run_on_loacl.sh so it runs the Cosim_NestDesktop_Insite usecase. For example in line 218 instead of --action-plan \${CO_SIM_MODULES_ROOT_PATH}/EBRAINS_WorkflowConfigurations/usecase/local/plans/cosim_alpha_brunel_local.xml it should be --action-plan \${CO_SIM_USE_CASE_ROOT_PATH}/userland/configs/local/plans/cosim_alpha_brunel_local.xml perfect. If Marcel does not see any problems with INSITE we can merge. I added the change to the correct action-xml as suggested by Fahad. I will test and review the automatic script generation then on the VM.
gharchive/pull-request
2023-07-20T07:58:35
2025-04-01T06:45:05.086087
{ "authors": [ "babsey", "mfahdaz", "sontheimer" ], "repo": "multiscale-cosim/Cosim_NestDesktop_Insite", "url": "https://github.com/multiscale-cosim/Cosim_NestDesktop_Insite/pull/19", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
620445935
Local module directory Adding a default search path for custom modules in the local directory. So close. I've noticed that if I configure an alternative local_directory_path, if I leave off the trailing slash, it does not work, though the other local config does.
gharchive/pull-request
2020-05-18T19:18:06
2025-04-01T06:45:05.210029
{ "authors": [ "joncrain" ], "repo": "munkireport/munkireport-php", "url": "https://github.com/munkireport/munkireport-php/pull/1347", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
206854673
Revamp of MiqViq's USB module MiqViq's keyboard and mouse module, but revamped for all USB devices tx!!
gharchive/pull-request
2017-02-10T17:20:36
2025-04-01T06:45:05.210884
{ "authors": [ "bochoven", "tuxudo" ], "repo": "munkireport/munkireport-php", "url": "https://github.com/munkireport/munkireport-php/pull/601", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
230791863
Improve identification of separate diffs when only indentation is different This is a minor quibble, but it's impacting me due to some changes I'm having to make to a large number of files, and I otherwise think this plugin is excellent, so I thought I'd suggest an improvement... I'm attempting to make a mostly similar block of code optional via a command line parameter in a bunch of files. I want to have a file which has been modified to the left, so I can then diff repeatedly with unmodified files on the right, moving over the change which is mostly the if statement and the increase in indent level, but avoiding an overwrite where the two files do actually differ. The diff identification algorithm is finding a few lines further down than I'd like, that happen to match, skipping over multiple line breaks which would normally trigger a separate diff. This prevents use of the copy-to-right button to move the much larger sections of code which are identical except for the indent, as they always seem to have one line which has a difference. So, I have to manually cut-and-paste between the two panes to avoid overwriting the few lines which are different in each file. What I'd like to see, is something more sophisticated than the current "ignore whitespace" checkbox when it comes to detection of when code which is different only in the indent, used to: be more sophisticated in synching left/right scrolling during a diff, if the code only differs in the indent level for a large section of contiguous lines, with a single begin if, end if type statement above and below, treat that as a match for scroll synchronization purposes, but still show it as a diff which can be moved via copy-right/left buttons. detect this situation, and instead of attempting to find one line which happens to match to separate one diff from the next, look for the largest block which matches other than a difference in indent, and then use the blank lines which correlate in each to split up into separate diffs. This would allow me to move over the large blocks of code which differ only in indent level via copy-right, while avoiding a copy of the lines which differ both in indent and in content. I uploaded a picture which shows the problem, which hopefully will help you visualize. http://tinypic.com/r/2dwfdj6/9 This package uses a diff algorithm. What you are suggesting sounds like it might be fixed by switching the algorithm to a more robust one. My hope is this issue would go away for you once https://github.com/mupchrch/split-diff/issues/114 has been fully accomplished. Is that a fair assumption?
gharchive/issue
2017-05-23T17:50:19
2025-04-01T06:45:05.216473
{ "authors": [ "michael-crawford", "mupchrch" ], "repo": "mupchrch/split-diff", "url": "https://github.com/mupchrch/split-diff/issues/109", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1832093957
OPPO105 not working correctly via DLNA Hi, all cabled, I'm not a fan of wifi so it's reserved for portable devices such as Phones and Tablets only! I suppose that doesn't fully negate a networking issue though. ... Actually, I just removed the DLNA Player Provider again and re-added it and it's working again. I did capture this from the logs though when playing the first track. 2023-08-01 14:17:00.946 INFO (MainThread) [music_assistant] Loaded player provider UPnP/DLNA Player provider 2023-08-01 14:17:31.960 ERROR (MainThread) [asyncio] Task exception was never retrieved future: <Task finished name='Task-96956' coro=<TracksController.add_item_to_library() done, defined at /usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/media/tracks.py:114> exception=AssertionError()> Traceback (most recent call last): File "/usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/media/tracks.py", line 160, in add_item_to_library library_item = await self._add_library_item(item) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/media/tracks.py", line 375, in _add_library_item return await self.update_item_in_library(cur_item.item_id, item) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/media/tracks.py", line 200, in update_item_in_library await self._set_track_album( File "/usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/media/tracks.py", line 420, in _set_track_album db_album = await self.mass.music.albums.get( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/media/albums.py", line 81, in get album.artists = [ ^ File "/usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/media/albums.py", line 82, in await self.mass.music.artists.get( File "/usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/media/base.py", line 200, in get await add_task File "/usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/media/artists.py", line 66, in add_item_to_library await self.mass.metadata.get_artist_metadata(item) File "/usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/metadata.py", line 119, in get_artist_metadata artist.mbid = await self.get_artist_mbid(artist) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/music_assistant/server/controllers/metadata.py", line 252, in get_artist_mbid if mbid := await provider.get_musicbrainz_artist_id( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/music_assistant/server/providers/musicbrainz/init.py", line 110, in get_musicbrainz_artist_id if mbid := await self._search_artist_by_track( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/music_assistant/server/providers/musicbrainz/init.py", line 161, in _search_artist_by_track assert trackname or track_isrc AssertionError I don't however get any indication in Music Assistant that the Oppo105 is playing though, the player shows idle and I cannot see the queue of items setup for playing, but it is working it's way through an album that I have queued. Actual scratch that, it's repeat playing track 1 over and over. Originally posted by @gadgetbazza in https://github.com/music-assistant/hass-music-assistant/discussions/1139#discussioncomment-6606344 @gadgetbazza does this issue still exist in beta63 ? @gadgetbazza Please confirm this has been fixed Hi, Updated to B67 just now and can confirm it's playing through tracks now without issue. Thanks so much! As an aside though, when I click on Albums in the nav and then selecting an album there is no content in the main pain (nav and player status still in place and working). So currently cannot navigate music via the Albums route. Please open a new issue about that and advise what music provider you are using and include some screenshots. Let us know if the tracks and albums appear under the artist and whether the tracks appear under tracks. Have a look in the logs when you navigate to the Album view. You could try to restart the addon just in case that fixes it.
gharchive/issue
2023-08-01T22:03:52
2025-04-01T06:45:05.361155
{ "authors": [ "OzGav", "gadgetbazza", "marcelveldt" ], "repo": "music-assistant/hass-music-assistant", "url": "https://github.com/music-assistant/hass-music-assistant/issues/1421", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }