id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1534532951
Update/fix XML docs for generated types Rationale The XML documentation of the generated types - at least the types which are directly used by developers - should match the documentation which is available on learn.microsoft.com and in the IDE when the "native" types are used. I haven't checked all types, but I've found differences in: CallerArgumentExpressionAttribute class summary class remarks ctor paramter property summary Index class summary class remarks various small differences in members I guess this is somewhat linked to #88. For my use, I don't really care if the docs match the types. I have XML doc generation enabled in my libs, and I get warnings on any public code that doesn't have XML comments. I'd be happy with anything, even empty, but I agree that it should be whatever MS has. would adding #pragma warning disable to all the generated files suffice? That would solve my problem, sure.
gharchive/issue
2023-01-16T08:54:29
2025-04-01T04:33:04.517672
{ "authors": [ "0xced", "SimonCropp", "cremor", "gregsdennis" ], "repo": "Sergio0694/PolySharp", "url": "https://github.com/Sergio0694/PolySharp/issues/44", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
255822713
NO BMI Source in Github There is no source in github, samples/bmi https://github.com/ServiceComb/ServiceComb-Java-Chassis/tree/master/samples/bmi The BMI sample currently is still in PR. You can get the code with the following commands: git clone https://github.com/ServiceComb/ServiceComb-Java-Chassis cd ServiceComb-Java-Chassis git fetch origin pull/129/head:sample git checkout sample @zhuhoudong The url should work now. Thanks for your feedback.
gharchive/issue
2017-09-07T05:51:24
2025-04-01T04:33:04.545494
{ "authors": [ "eric-lee-ltk", "zhuhoudong" ], "repo": "ServiceComb/ServiceComb-Java-Chassis", "url": "https://github.com/ServiceComb/ServiceComb-Java-Chassis/issues/137", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
615404750
Added 5 unit cases for delete method. Used pytest framework for unit testing. Created dataset to test using pytest fixture. Tested valid and invalid scenarios too. Check this issue https://github.com/SforAiDl/decepticonlp/issues/22 Acknowledged. Discussion thread continued in #22
gharchive/pull-request
2020-05-10T14:36:32
2025-04-01T04:33:04.565567
{ "authors": [ "rajaswa", "rohts-patil", "someshsingh22" ], "repo": "SforAiDl/decepticonlp", "url": "https://github.com/SforAiDl/decepticonlp/pull/23", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2461716182
[BUG]: Starting a new game with 1 duplicant bugs it out Which Mod? Duplicant Stat Selector Description of the bug/crash Upon embarking on a new game with 1 dupe, the duplicant spawns but the game is seemingly stuck (not frozen). Pressing the ESC key 3+ times is required to unstuck it. The regular "Alert" popup (check screenshot below) does NOT appear at the beginning. Note that I don't get the bug with 2 or 4+ dupes. Steps To Reproduce New Game > Any Game Mode > Any Asteroid Style > Start Game Click on "Remove Slot" on 2 duplicants to only have 1 remaining Embark This is 100% reproductible in my end, the bug always happens. Player.log Player.log fixed with release 7.3.1
gharchive/issue
2024-08-12T19:15:39
2025-04-01T04:33:04.570329
{ "authors": [ "Sad-theFaceless", "Sgt-Imalas" ], "repo": "Sgt-Imalas/Sgt_Imalas-Oni-Mods", "url": "https://github.com/Sgt-Imalas/Sgt_Imalas-Oni-Mods/issues/126", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1352898022
Reopening Again Issue With The Timebar Size. Describe your problem here. I Just Want To Make The Timebar To Look Like This With The Source Code. Can You Help Me? Are you modding a build from source or with Lua? Source What is your build target? Neko, HashLink, or other build system Did you edit anything in this build? If so, mention or summarize your changes. For The Better Style. theres no need to reopen the issue you can just press the 3 dots and press edit for the fix go to assets/shared/images/timebar.png and replace it theres no need to duplicate the issue you can just press the 3 dots and press edit for the fix go to assets/shared/images/timebar.png and replace it I Just Trying To Find To All Mods, But It's Old Classic TimeBar I Just Trying To Find To All Mods, But It's Old Classic TimeBar What do you mean What do you mean Mods Only Have This Old Time Bar: you can just replace it because it's a sprite Make the sprite longer
gharchive/issue
2022-08-27T01:32:21
2025-04-01T04:33:04.590663
{ "authors": [ "Mastercopper2346", "StefanBETA2008", "TheKitBoi" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/10326", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1376888992
how do you add images to the menu Describe your problem here. i want to add a image to the main menu but i dont know how to, is there like a code or something? Are you modding a build from source or with Lua? Source What is your build target? Windows x64 Did you edit anything in this build? If so, mention or summarize your changes. MainMenuState.hx make a var then asses = new FlxSprite().loadGraphic(Paths.images('image')); asses.x += 0; asses.y += 0; add(asses); make a variable for your sprite: var sprite:FlxSprite; sprite = new FlxSprite(X, Y).loadGraphic(Paths.image('yourimage')); add(sprite); change X and Y into the coordinates you want your sprite to be change yourimage to the sprite's graphic change the sprite variable to be called whatever you want For easier movement var sprite:FlxSprite; sprite = new FlxSprite(0, 0).loadGraphic(Paths.image('yourimage')); sprite.x += X; //for negative coordinates use -= instead sprite.y += Y; add(sprite); change X into the x position you want your sprite to go same for the Y thanks it works :D yo why does it go off the screen once I scrolled to options yo why does it go off the screen once I scrolled to options Camera movement Just do sprite.scrollFactor.set();, changing sprite to the variable that is your sprite thanks bro Help please, I'm using psych engine 0.7.3 and getting this error when compiling: source/states/MainMenuState.hx:232: characters 3-9 : Unexpected sprite Help please, I'm using psych engine 0.7.3 and getting this error when compiling: source/states/MainMenuState.hx:232: characters 3-9 : Unexpected sprite change sprite to your sprite variable Like if you have a var overlay:FlxSprite, for example, and wanted to use that, change it to overlay.scrollFactor.set(); Помогите пожалуйста, я использую Psych Engine 0.7.3 и получаю эту ошибку при компиляции: source/states/MainMenuState.hx:232: символы 3-9 : Неожиданный спрайт Измените spriteпеременную спрайта. Например, если у вас есть var overlay:FlxSprite, и вы хотите использовать его, измените его наoverlay.scrollFactor.set(); It looks like I'm too stupid for this, please write a code example var sprite:FlxSprite; sprite = new FlxSprite(0, 0).loadGraphic(Paths.image('yourimage')); sprite.x += X; //for negative coordinates use -= instead sprite.y += Y; sprite.scrollFactor.set(); add(sprite);
gharchive/issue
2022-09-17T22:40:10
2025-04-01T04:33:04.600287
{ "authors": [ "BernardoGP4504", "DisIsNOt", "Velikiycho", "raxyd", "uhhhhuhhhhhhokay" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/10571", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2599685214
Weekend 1 freeze after cutscene Describe your bug here. After loading up Weekend 1 from the Story Mode menu and watching the cutscene, the application stops responding. Command Prompt/Terminal logs (if existing) No response Are you modding a build from source or with Lua? Lua What is your build target? Windows Did you edit anything in this build? If so, mention or summarize your changes. No. To bypass this glitch use the 32-bit version. This version doesn't support videos, so it skips the cutscenes and thus the week is playable.
gharchive/issue
2024-10-19T22:49:05
2025-04-01T04:33:04.602570
{ "authors": [ "BluCoolKnux", "DacroyleYT" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/15763", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1070538013
Chart's Editor Playtest but outside Chart Editor What feature do you want to get added on the base engine? (I'll try my best to explain this i don't speak english very well lol) Basically just like the Chart Editor has as playtest function to test the chart from an exact position, but what i suggest is adding that same feature but in base game. This could be useful for viewing events or functionality of custom notetypes, because this can't be viewed in the Chart Editor's playtest. For example, let's say your song lasts 5 minutes, and at the end you have some sort of event, Character Change, Custom Event, Arrows Move, etc. Or you have a custom notetype you want to check if it works, but you'll have to wait 5 minutes. This could be solved by adding support for playtesting in-game, i think it would be a really nice time-saving feature! the annoying thing about the playtest feature is that it puts you in a seperate state from playstate where the only thing you can see is the notes/ just put the event at the start of the song I have a different plan for that.
gharchive/issue
2021-12-03T12:23:43
2025-04-01T04:33:04.605222
{ "authors": [ "4wdthinks", "AetherLynx", "MayoALT", "ShadowMario" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/1633", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1159148140
Character Bug Describe your bug here. Hello, I am here to tell you about this bug that happens to me very often and it is very annoying when I finish creating my character and everything is fine with the png, the XML and the sprites/frames but when I finish configuring the character and I put it on the chart editor and when I give it to play for a moment it loads but then it closes Command Prompt/Terminal logs (if existing) Does not exist Are you modding a build from source or with Lua? Source What is your build target? Windows Did you edit anything in this build? If so, mention or summarize your changes. it's new Never had any issues with any character so do you mind posting the files here so I can check them out? I doubt this is a bug but I'll test it out anyway maybe you just coded it wrong Never had any issues with any character so do you mind posting the files here so I can check them out? I doubt this is a bug but I'll test it out anyway https://drive.google.com/file/d/1KpFHKmOAeapDpcmiWWYIwjq_L_FJV2ce/view (png) https://drive.google.com/file/d/1ZnAIv751Ux3fPU7ST3iBJ1VOYoyZZ5dO/view (XML) and the json works correctly Send the json
gharchive/issue
2022-03-04T01:38:09
2025-04-01T04:33:04.609212
{ "authors": [ "frantastic24", "thedeveloper112206", "unnamedmask3" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/6203", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1292428928
Idk with the source code Describe your problem here. when I complie the build with the source code u know the build it does not appear in the export folder, only the mods, assets, plugins and manifest folders are exported, not the program or game, this did not happen to me with other versions such as 0.5 or 0.52 I don;t know if it is a problem or something but I don't know I'm new in coding yeah Are you modding a build from source or with Lua? Source What is your build target? Windows x64 Did you edit anything in this build? If so, mention or summarize your changes. no, i just export all with the source code, I didn't touch anything lime build windows -64 -clean make sure that it does not errors in the console nor stops if it does, please screenshot it and show the error here I tried 3 times and it didn't work Duplicate. Read the readme file.
gharchive/issue
2022-07-03T21:52:06
2025-04-01T04:33:04.612647
{ "authors": [ "Raltyro", "SG-ProBlue", "ShadowMario" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/9481", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1993089330
Rename EditorPlayState to EditorPlaySubState I think after updates EditorPlayState become a substate so yeah no idk lmao this might be a dumb pr but this says everything just do "EditorGameplay.hx" just do "EditorGameplay.hx" PlayStateEditorSubstate Renames it to EditorPlaySubState Doesn't put it in the substates folder This doesn't follow the naming scheme pretty much. Renames it to EditorPlaySubState Doesn't put it in the substates folder options are substates they're not in substates folder CustomFadeTransition is a substate it's not in the substates folder... This doesn't follow the naming scheme pretty much. think Play"State" and EditorPlay"SubState" Its called like that because it's a copy of PlayState, not because it is/isn't a state
gharchive/pull-request
2023-11-14T16:20:59
2025-04-01T04:33:04.616256
{ "authors": [ "CaptainBaldi", "EyeDaleHim", "KarimAkra", "ShadowMario", "crowplexus", "mcagabe19" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/pull/13610", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1516253569
Failing does not trigger a restart Triggering a level fail will not restart the current level. The menu is still accessible so it is still possible to exit the game by normal means. Losing should prematurely win the level (IIRC this is what original ZB does) Implemented in the private asset repository
gharchive/issue
2023-01-02T10:40:01
2025-04-01T04:33:04.632701
{ "authors": [ "ShamblesSM" ], "repo": "ShamblesSM/OpenSMCE", "url": "https://github.com/ShamblesSM/OpenSMCE/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
869910426
Cruise Control Malfunction , LCA Malfunction when attempting Engage I installed the fork and rebooted. EON loads Stock Addition and I get a different GUI but when trying to enable/engage self driving i get a crouse control Malfunction error on my 2020 corolla Do you get this on stock openpilot? I don't see any errors about this, check the /data/community/crashes dir for a latest file. If there's nothing about a malfunction in the file, check some of the other log files. Thanks. I didn't get this error on stock openpilot before I installed the fork. I tried reinstalling the fork and same error happens (I see the new GUI but ACC doesn't engage). I see "vehicle online" on EON but it says "connect error" in red or "connect online" in yellow. I installed DP and similar error happens. I will check logs next Try reflashing panda as well if you get this on other forks as well. Is this fixed now on my fork?
gharchive/issue
2021-04-28T12:48:29
2025-04-01T04:33:04.634774
{ "authors": [ "ShaneSmiskol", "hwhorf" ], "repo": "ShaneSmiskol/openpilot", "url": "https://github.com/ShaneSmiskol/openpilot/issues/402", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1100822355
Update Pages “index” Automatically generated by Netlify CMS 👷 Deploy Preview for gracious-knuth-20d29a processing. 🔨 Explore the source changes: f09593915a01f63772a8ed0e1ccdffd35631f4a8 🔍 Inspect the deploy log: https://app.netlify.com/sites/gracious-knuth-20d29a/deploys/61df4ce0d84d260007dff9c5
gharchive/pull-request
2022-01-12T21:49:19
2025-04-01T04:33:04.681332
{ "authors": [ "ShannonLCapper" ], "repo": "ShannonLCapper/netlify-cms-gatsby-spike", "url": "https://github.com/ShannonLCapper/netlify-cms-gatsby-spike/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
282723311
Added the spo site list command solving #115 Added the spo site list command solving #115 Coverage remained the same at 100.0% when pulling ee0262cfbaa8ee2ea0d046727622f34dda54afde on waldekmastykarz:spo-site-list into bd2361710599e1a1ea401468c9cb092fbf986a17 on SharePoint:dev. Merged manually
gharchive/pull-request
2017-12-17T20:59:25
2025-04-01T04:33:04.700707
{ "authors": [ "coveralls", "waldekmastykarz" ], "repo": "SharePoint/office365-cli", "url": "https://github.com/SharePoint/office365-cli/pull/129", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
562197863
NodeJs download link not working The nodejs download link returns 404 https://nodejs.org/dist/latest-v10.x/node-v10.18.1-x64.msi Document Details ⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking. ID: 5bee09a8-022e-1b5a-4aab-d6e7c6216808 Version Independent ID: 49edce1e-f2be-e989-917f-c6b389363e2f Content: Set up your SharePoint Framework development environment Content Source: docs/spfx/set-up-your-development-environment.md Product: sharepoint Technology: sharepoint-framework GitHub Login: @spdevdocs Microsoft Alias: spdevdocs The latest v10.x version of the NodeJs is v10.19.0, please download file using the below link. https://nodejs.org/dist/latest-v10.x/node-v10.19.0-x64.msi Dennis Guo Microsoft SharePoint Community Support
gharchive/issue
2020-02-09T15:59:12
2025-04-01T04:33:04.705962
{ "authors": [ "Dennis-LZ", "gxy-dev" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/issues/5310", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
249508181
Documents are not renderred correctly in Production Category [ ] Question [ ] Typo [x ] Bug [ ] Additional article idea Expected or Desired Behavior Documents are rendered correctly as designed. Observed Behavior All the documents under SharePoint Framework (e.g. https://dev.office.com/sharepoint/docs/spfx/set-up-your-developer-tenant) are not rendered currently in production. There's corrupted section at bottom of each page. Steps to Reproduce browser to https://dev.office.com/sharepoint/docs/spfx/set-up-your-developer-tenant scroll down to bottom of the page thx @aaronmi for reporting. Issue bypassed for the CMS owners - hopefully, will get fixed asap. thx @aaronmi for reporting this. Fixed now in the production CMS side.
gharchive/issue
2017-08-10T23:57:33
2025-04-01T04:33:04.709859
{ "authors": [ "VesaJuvonen", "aaronmi" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/issues/776", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
430359355
Small spelling correction Category [x] Content fix [ ] New article What's in this Pull Request? On the page https://docs.microsoft.com/en-us/sharepoint/dev/spfx/web-parts/get-started/use-fabric-react-components Spelled "of" instead of "fo" Guidance No major update. Only a small spelling correction. thx @nanddeepn, appreciated
gharchive/pull-request
2019-04-08T09:56:57
2025-04-01T04:33:04.712215
{ "authors": [ "VesaJuvonen", "nanddeepn" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/pull/3748", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
485461109
Removing note that the site footer hasn't rolled out WW yet Category [X] Content fix [ ] New article Related issues: N/A What's in this Pull Request? According to MC 186030 (https://admin.microsoft.com/AdminPortal/home?switchtomodern=true#/MessageCenter?id=MC186030) it has rolled out WW by the end of July 2019 already, so the notice about it not being available yet can be removed. thx @KoenZomers, appreciated.
gharchive/pull-request
2019-08-26T21:26:53
2025-04-01T04:33:04.714251
{ "authors": [ "KoenZomers", "VesaJuvonen" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/pull/4538", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1266790216
Validation: Fix missing description This PR addresses the following validation issues and includes an updated ms.date: missing description Docs Build status updates of commit 1bff4f4: :warning: Validation status: warnings File Status Preview URL Details docs/schema/property-element-property-bag.md :warning:Warning View Details docs/schema/property-element-property-bag.md Line 2, Column 1: [Warning: ms-prod-and-service - See documentation] Only one of the following attributes can exist: 'ms.prod', 'ms.service'. Use ms.prod for on-premise products, or ms.service for cloud services. For more details, please refer to the build report. If you see build warnings/errors with permission issues, it might be due to single sign-on (SSO) enabled on Microsoft's GitHub organizations. Please follow instructions here to re-authorize your GitHub account to Docs Build. Note: Broken links written as relative paths are included in the above build report. For broken links written as absolute paths or external URLs, see the broken link report. Note: Your PR may contain errors or warnings unrelated to the files you changed. This happens when external dependencies like GitHub alias, Microsoft alias, cross repo links are updated. Please use these instructions to resolve them. For any questions, please:Try searching the docs.microsoft.com contributor guidesPost your question in the Docs support channel This PR addresses the following validation issues and includes an updated ms.date: missing description Global effort to fix build validation errors @VesaJuvonen The Content & Learning team is fixing build validation errors on docs.microsoft.com for the rest of H2. This will eliminate potential accessibility, security, and usability issues. This PR includes only build validation fixes and does not change other content. Please review within five business days and merge or comment in the PR with any changes you'd like to see. Thank you!
gharchive/pull-request
2022-06-10T00:07:10
2025-04-01T04:33:04.724127
{ "authors": [ "VesaJuvonen", "annmariehitchcock" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/pull/8112", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
465591658
Looking for to Integrate an ICal Feed in SharePoint Online which is Link Based || SharePoint Calendar I am looking for a way to integrate an ICal Calendar Feed in SharePoint Online Which will Keep Synchronizing with SharePoint using the ICal feed. Originally the feed belongs from a Third Party Human Resource Software, Using this I am trying to keep the calendar synched with the updates for e.g. Holidays! Can someone help me with how this could be achieved! Thank you Please, can someone help with this? Thank you @tejasthite the React Calendar Feed web part sample will retrieve iCal feeds and present them as if they were native events in SharePoint. However, it does not sync with local SharePoint content. If you want to sync content, I would suggest that you try using Flow instead and create a scheduled flow to regularly sync the content. Let me know if you aren't familiar with Flow, I'd be happy to help. Yes, I would need help with this, how can I get this accomplished? thank you I tried the steps from the Readme.md File, I am getting the following errors || any help will be appreciated! Thank you C:\Users\tejast>npm install npm WARN ajv-keywords@3.2.0 requires a peer of ajv@^6.0.0 but none is installed. You must install peer dependencies yourself. npm WARN gulp-karma@0.0.5 requires a peer of karma@>=0.10 <=0.13 but none is installed. You must install peer dependencies yourself. npm WARN slick-carousel@1.8.1 requires a peer of jquery@>=1.8.0 but none is installed. You must install peer dependencies yourself. npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@1.2.4 (node_modules\fsevents): npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.2.4: wanted {"os":"darwin","arch":"any"} (current: {"os":"win32","arch":"x64"}) Then When i try the npm audit fix - I get the Below || Please can someone help Thank you C:\Users\tejast>npm audit fix npm WARN deprecated left-pad@1.3.0: use String.prototype.padStart() npm WARN deprecated kleur@2.0.2: Please upgrade to kleur@3 or migrate to 'ansi-colors' if you prefer the old syntax. Visit https://github.com/lukeed/kleur/releases/tag/v3.0.0\ for migration path(s). npm WARN deprecated browserslist@1.7.7: Browserslist 2 could fail on reading Browserslist >3.0 config used in other tools. npm WARN deprecated flatten@1.0.2: I wrote this module a very long time ago; you should use something else. deasync@0.1.15 install C:\Users\tejast\node_modules\deasync node ./build.js win32-x64-node-10 exists; testing Binary is fine; exiting core-js@2.6.9 postinstall C:\Users\tejast\node_modules\babel-register\node_modules\core-js node scripts/postinstall || echo "ignore" Thank you for using core-js ( https://github.com/zloirock/core-js ) for polyfilling JavaScript standard library! The project needs your help! Please consider supporting of core-js on Open Collective or Patreon: https://opencollective.com/core-js https://www.patreon.com/zloirock Also, the author of core-js ( https://github.com/zloirock ) is looking for a good job -) core-js@2.6.9 postinstall C:\Users\tejast\node_modules\babel-runtime\node_modules\core-js node scripts/postinstall || echo "ignore" Thank you for using core-js ( https://github.com/zloirock/core-js ) for polyfilling JavaScript standard library! The project needs your help! Please consider supporting of core-js on Open Collective or Patreon: https://opencollective.com/core-js https://www.patreon.com/zloirock Also, the author of core-js ( https://github.com/zloirock ) is looking for a good job -) npm WARN rollback Rolling back node-pre-gyp@0.12.0 failed (this is probably harmless): EPERM: operation not permitted, rmdir 'C:\Users\tejast\node_modules\fsevents\node_modules' npm WARN slick-carousel@1.8.1 requires a peer of jquery@>=1.8.0 but none is installed. You must install peer dependencies yourself. npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself. npm WARN ajv-keywords@3.4.1 requires a peer of ajv@^6.9.1 but none is installed. You must install peer dependencies yourself. npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself. npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself. npm WARN office-ui-fabric-react@6.199.0 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself. npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself. npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself. npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself. npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@1.2.9 (node_modules\fsevents): npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.2.9: wanted {"os":"darwin","arch":"any"} (current: {"os":"win32","arch":"x64"}) @microsoft/sp-build-web@1.8.2 @microsoft/sp-webpart-workbench@1.8.2 added 313 packages from 120 contributors, removed 282 packages, updated 196 packages and moved 14 packages in 340.523s fixed 9448 of 11201 vulnerabilities in 682936 scanned packages 1746 vulnerabilities required manual review and could not be updated 1 package update for 7 vulns involved breaking changes (use npm audit fix --force to install breaking changes; or refer to npm audit for steps to fix these manually) @hugoabernier - i was successfully able to load the Web part and attach the URL for my iCal Calendar Feed but for some reason it keeps me saying - "Oops, something went wrong! We can't display your events at the moment. Please try again later." Any help would be appreciated? Thank you @hugoabernier it would be great if you could help with this. Thank you! @tejasthite do you mind sharing the iCal URL so that I can try to diagnose the issue? @hugoabernier is there a way I could send you a Private Message with the Link? Thank you Yes send me a direct message on my Twitter account @hugoabernier dropped you the link on twitter, Thank you @hugoabernier Thank you Very Much, I will give it a try and update. @hugoabernier - Wil I need to reclone the WebPart Sample after which the changes you have made recently? Thank you @hugoabernier - I have been able to Pull the Events Now Thank you Very Much for your help, How do I now integrate this with My SharePoint Online Calendar? That's great @tejasthite ! Do you have access to Flow? If so, may I suggest that you close this issue and we'll continue this one offline via Twitter DM? I'll come back later and update this issue with our resolution so that anyone else who needs this type of solution can benefit from it. @hugoabernier - Yes I do have Flow. It would be always great to have the solution for the solution proposed with React Calendar Feed asap, I truly need this as well. But at the same time if Flow is going faster right now then I am happy to implement that as well. Thank you very much again. "let me how to proceed Further"? @hugoabernier , Please could you update on the above? Thank you @hugoabernier, Thank you very much for the Time from Your End - and the Microsoft Flow, Excellent work with this. I appreciate all your help. Using Flow it works Perfect. Thank you @tejasthite I'm glad it worked out. If anyone is looking for the solution we used, we documented the approach in this blog post and provided the step-by-step instructions in this one
gharchive/issue
2019-07-09T06:21:11
2025-04-01T04:33:04.750688
{ "authors": [ "hugoabernier", "tejasthite" ], "repo": "SharePoint/sp-dev-fx-webparts", "url": "https://github.com/SharePoint/sp-dev-fx-webparts/issues/909", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
489527291
Icons does not display in peoples directory I got the latest code and solution and added some properties to view in card in peoples directory. and bundled it and deployed. But i cannot view any of the fabric icons ? Do u have any idea why>]? Screenshot I am also facing this issue. Hi, are you still facing this issue? Are you able to share the code you modified?
gharchive/issue
2019-09-05T04:54:43
2025-04-01T04:33:04.753498
{ "authors": [ "bcameron1231", "ddmarsden", "shwanthi" ], "repo": "SharePoint/sp-starter-kit", "url": "https://github.com/SharePoint/sp-starter-kit/issues/317", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
979845839
xlsx.read not work in chrome use In-Browser demo https://oss.sheetjs.com/ use same csv file , firefox show right result however chrome didn`t work fine. firefox result: chrome result: click test.csv to download test file test.csv thanks
gharchive/issue
2021-08-26T05:18:44
2025-04-01T04:33:04.767273
{ "authors": [ "mario-mui" ], "repo": "SheetJS/sheetjs", "url": "https://github.com/SheetJS/sheetjs/issues/2368", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
676099918
Adda option for creating new user data action need to be performed with js What acton do you need to be performed with js Want to create user and store their data in local json file. All from front end only. User fills the form data and that data is submitted to json file. addUser.html is the form which want to convert the details into json file yes @kunalprakash1309 Assign this task to me. I want to try on it hey @kunalprakash1309 are you still working on this?
gharchive/issue
2020-08-10T12:31:19
2025-04-01T04:33:04.898683
{ "authors": [ "ShivSt", "kunalprakash1309" ], "repo": "ShivSt/demoWebsite", "url": "https://github.com/ShivSt/demoWebsite/issues/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
157872515
How do you get the XML's response & request Can you please explain what the easiest way to get a response/request XML/JSON strings without debugging the code? any logs for that? couldn't find it. Logging is the responsibility of the given adapters. You can look at the carrier you care about to see how it is logged. Or debug the request. In general, reporting an issue should follow our guideline for contributing and should be well described problem with the code. Not simply a question.
gharchive/issue
2016-06-01T10:21:58
2025-04-01T04:33:04.931567
{ "authors": [ "kmcphillips", "megetron" ], "repo": "Shopify/active_shipping", "url": "https://github.com/Shopify/active_shipping/issues/377", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1282815735
[Bug]: Deploy App (without running dev server) > App Url in Partners links to 404 page Please confirm that you have: [X] Searched existing issues to see if your issue is a duplicate. (If you’ve found a duplicate issue, feel free to add additional information in a comment on it.) [X] Reproduced the issue in the latest CLI version. In which of these areas are you experiencing a problem? Extension Expected behavior Typically, the App Url would be set to an ngrok url if you ran npm run dev or yarn dev before deploying and creating the app through the CLI. In this case, if you don't ever run an app locally, a default App Url appears to be set. When a defauIt URL is used, it should resolve to a page that does not 404. In addition to this, it would be helpful if this linked to a page to tell me that I need to either deploy the App to a public URL or run the app locally in order to generate a functional merchant install link. Actual behavior The link for the App Url in the Partners app is https://shopify.github.io/shopify-cli/help/start-app/ which goes to a 404 page. Stack trace No response Reproduction steps Create a new app using yarn create @shopify/app --template node. Scaffold a new extension using yarn scaffold extension. Deploy WITHOUT running yarn dev. Use yarn deploy to deploy and create a new app. Navigate to the partners link and you'll see the App URL field is set to https://shopify.github.io/shopify-cli/help/start-app/ which is a 404 page. Operating System Mac OS Monterey Shopify CLI version (check your project's package.json if you're not sure) 3.0.24 Shell zsh Node version (run node -v if you're not sure) v18.0.0 What language and version are you using in your application? Node Hi @jamesvidler -- I believe this issue is resolved via https://github.com/Shopify/cli/pull/153. Do let us know if I'm missing something and your need is a different one.
gharchive/issue
2022-06-23T18:57:32
2025-04-01T04:33:04.939067
{ "authors": [ "jamesvidler", "shauns" ], "repo": "Shopify/cli", "url": "https://github.com/Shopify/cli/issues/47", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
813542930
Issues with indexes While generating a new db migration, @sambostock and I noticed that the descending order was dropped from the ["task_name", "created_at"] index when the migration ran. Some of our adopters also seem to not have the order on the index present: https://github.com/Shopify/arrive-server/blob/master/db/schema.rb#L393 https://github.com/Shopify/accio/blob/master/db/schema.rb#L192 @etiennebarrie had also flagged that the index on just task_name is potentially redundant - we should double check whether we need it or not. Some updates on this: Confirmed that our task_name index is redundant and can be dropped. Will open a PR Adopters not having the sort order present on the index is actually not a bug, but instead due to the fact that most of our apps are using MySQL 5.7, and ordering is only present on MySQL 8.0+ 🤦‍♀️ The sort order disappearing from SQLite3 when running a new migration is actually the result of a bug in the SQLite3 Adapter in Rails.
gharchive/issue
2021-02-22T14:17:55
2025-04-01T04:33:04.947210
{ "authors": [ "adrianna-chang-shopify" ], "repo": "Shopify/maintenance_tasks", "url": "https://github.com/Shopify/maintenance_tasks/issues/342", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
57246928
Producer: implement AsyncClose @wvanbergen :+1: tested this on a local app.
gharchive/pull-request
2015-02-10T22:21:39
2025-04-01T04:33:04.986536
{ "authors": [ "eapache", "wvanbergen" ], "repo": "Shopify/sarama", "url": "https://github.com/Shopify/sarama/pull/258", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1836967262
upgrade prettier plugins Description This PR upgrades prettier plugin versions so we can upgrade web's version to ^3 Also if you're in the process of making breaking changes - @typescript/eslint-plugin v6 recently got released and we'll need to update to use that if we want to support Typescript 5.2 or later.
gharchive/pull-request
2023-08-04T15:36:11
2025-04-01T04:33:05.002314
{ "authors": [ "BPScott", "m4thieulavoie" ], "repo": "Shopify/web-configs", "url": "https://github.com/Shopify/web-configs/pull/397", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
379308808
Add Croissant Makers The concept of Croissant Makers: Give $20/10s. That's fun to implement into my quarter-second clock! :( Costs $75 to hire Has some sort of funny tooltip, just because. Made this in Commit 49.
gharchive/issue
2018-11-09T20:22:07
2025-04-01T04:33:05.024907
{ "authors": [ "Shrubhog" ], "repo": "Shrubhog/expresso", "url": "https://github.com/Shrubhog/expresso/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
124087696
Network AVROTROS missing Network AVROTROS missing in network_timezones.txt and also uploaded a logo for this network. 2015-12-28 10:58:19 INFO Thread-51 :: Network was not found in the network time zones: AVROTROS Thanks for the pull request! Before a real human comes by, please make sure your PR has all of the below criteria checked [x] Give a description on what the PR is for. [x] Make sure your PR is based on the DEVELOP branch [x] Don't send big changes all at once. Split up big PRs into multiple smaller PRs that are easier to manage and review Please make sure you also read contribution guide and followed all of the steps. Thanks! The logo for the AVROTROS network. Do you want me to create a pr at https://github.com/SickRage/SickRage/tree/master/gui/slick/images/network/ for this?? @aboogerd the logo goes on the main repo like you had it
gharchive/pull-request
2015-12-28T16:25:35
2025-04-01T04:33:05.065882
{ "authors": [ "SickRage2", "aboogerd", "duramato" ], "repo": "SickRage/sickrage.github.io", "url": "https://github.com/SickRage/sickrage.github.io/pull/22", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2687957887
Collecting the sustainable materials for building Below are the acceptance criteria for this issue which is an enhancement: The game should identifies the specific locations where players can get sustainable materials. Only sustainable materials should be able to collect for this quest. Points will be rewarded for collecting materials without damaging the environment. Done
gharchive/issue
2024-11-24T16:10:57
2025-04-01T04:33:05.069283
{ "authors": [ "Siddhartha-atu" ], "repo": "Siddhartha-atu/Zelda_Githubactions", "url": "https://github.com/Siddhartha-atu/Zelda_Githubactions/issues/6", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1780904602
Multi browser support Since the extension has turned into a glorified script and style injector, adapting LiChess Tools for other browsers might be trivial. I would still need help from people using said browsers. I use Brave, which is a version of Chrome. Edge is a version of Chrome. What else is there? Firefox? Safari? I don't like them, so how could I test them properly? I don't have much affinity in the programming part of extensions, but I can try to help test in firefox By the way, if you want to launch the .xpi file first before publishing https://addons.mozilla.org also works The problem is with the content_scripts manifest property. While Firefox seems to support Manifest V3, it doesn't support the "world" setting which allows a script to be loaded in the context of the web page. I will have to research how to manage this in FF. adding dubious tag because Firefox has 1.5% of browser share as opposed to Chrome+Edge which is 65%. I had no idea Firefox is so lame. :) I will not do it. Firefox is small and stupid and Safari asks me for money. For future Mac users who come across this issue and are looking into Safari compatibility: you should know, you can convert Chrome extensions to Safari extensions using Xcode on your Mac (no developer account needed). Just Google it and follow the instructions. Also, remember to enable unsigned extensions in Safari. New developments make this a possibility. Firefox version implemented in v2.3.141
gharchive/issue
2023-06-29T14:07:52
2025-04-01T04:33:05.072875
{ "authors": [ "KeatsPeeks", "SergioGlorias", "Siderite" ], "repo": "Siderite/lichessTools", "url": "https://github.com/Siderite/lichessTools/issues/34", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1563031201
possible overflow bugs https://github.com/SierraWireless/octave-orp/blob/3ccf4406a2c0cd9f5f8c081b2bf340e575158806/clients/c/src/orpUtils.c#L141-L150 At Line 144, there is no guarantee that message->status * -1 < sizeof(statusStr), which may lead to the following bug: ==257646==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000005 (pc 0x0000005371b0 bp 0x7fff256688e0 sp 0x7fff25668058 T0) ==257646==The signal is caused by a READ memory access. ==257646==Hint: address points to the zero page. #0 0x5371b0 in __sanitizer::internal_strlen(char const*) (/home/parallels/octave-orp/clients/c/bin/orp+0x5371b0) #1 0x4c57d6 in printf_common(void*, char const*, __va_list_tag*) (/home/parallels/octave-orp/clients/c/bin/orp+0x4c57d6) #2 0x4c6c4e in printf (/home/parallels/octave-orp/clients/c/bin/orp+0x4c6c4e) #3 0x56730c in orp_MessagePrint /home/parallels/octave-orp/clients/c/src/orpUtils.c:144:17 #4 0x5630f7 in orp_HdlcDeframe /home/parallels/octave-orp/clients/c/src/orpClient.c:446:9 https://github.com/SierraWireless/octave-orp/blob/3ccf4406a2c0cd9f5f8c081b2bf340e575158806/clients/c/src/orpProtocol.c#L1014-L1018 At Line 1017, offset may equal pktLen - 1. Thus, pktBuf[offset + 1] equals pktBuf[pktLen], which leads to the following overflow bug when calling strtoul: ==261454==ERROR: AddressSanitizer: SEGV on unknown address 0x6060002d0000 (pc 0x7efd0d5ef17b bp 0x000000000000 sp 0x7ffe53ff0c80 T0) ==261454==The signal is caused by a READ memory access. #0 0x7efd0d5ef17b in __GI_____strtoul_l_internal /build/glibc-sMfBJT/glibc-2.31/stdlib/../stdlib/strtol_l.c:432:28 #1 0x55baad in orp_ProtocolDecode_v1 /home/parallels/octave-orp/clients/c/src/orpProtocol.c:1017:42 Similar issues may happen at Line 995, Line 1006, and other places where pktBuf[offset + 1] is used. https://github.com/SierraWireless/octave-orp/blob/3ccf4406a2c0cd9f5f8c081b2bf340e575158806/clients/c/src/orpProtocol.c#L1054-L1059 At Line 1058, pktBuf + ORP_OFFSET_VARLENGTH may overrun the buffer. ==261221==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x6020000000d4 at pc 0x0000004c5908 bp 0x7ffd3ad375a0 sp 0x7ffd3ad36d20 READ of size 1 at 0x6020000000d4 thread T0 #0 0x4c5907 in printf_common(void*, char const*, __va_list_tag*) (/home/parallels/octave-orp/clients/c/bin/orp+0x4c5907) #1 0x4c6c4e in printf (/home/parallels/octave-orp/clients/c/bin/orp+0x4c6c4e) #2 0x55bec8 in orp_ProtocolDecode_v1 /home/parallels/octave-orp/clients/c/src/orpProtocol.c:1056:9
gharchive/issue
2023-01-30T19:04:47
2025-04-01T04:33:05.080803
{ "authors": [ "harryreps" ], "repo": "SierraWireless/octave-orp", "url": "https://github.com/SierraWireless/octave-orp/issues/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
134555747
Create a tab-separated text format for creating translations that are in the .po file The .po file is a bit impenetrable to the average translator. Create a simple table view with the English source text in the first column, and the other languages in the second-third-etc. columns One such file per foreign language would in principle also do the trick, but as the number of terms to be translated grows, it may be easier to keep things consistent if we would see all languages in a single table Create a script for converting the tab-separated text to new/updated .po files for each language There are programs for editing '.po' files that supposedly make translating easier, you might want to look into them. I have no experience with them myself, so I am not sure if they work with djangos .po files. I have not used these apps myself, but here are a few examples: https://poedit.net/ https://localise.biz/free/poeditor Thanks @henrinie , I'll check it out. @ocrasborn, are the suggestions Henri made sufficient? If so, can we close this issue? Yes, let's close this for now, I think Poedit will do the job fine.
gharchive/issue
2016-02-18T11:33:54
2025-04-01T04:33:05.107326
{ "authors": [ "henrinie", "ocrasborn", "vanlummelhuizen" ], "repo": "Signbank/NGT-signbank", "url": "https://github.com/Signbank/NGT-signbank/issues/132", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
523298537
BREAKING CHANGE: Rename AppendToStream to camelCase from PascalCase Seriously...sadface. :tada: This issue has been resolved in version 1.0.0 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
gharchive/issue
2019-11-15T07:16:07
2025-04-01T04:33:05.196360
{ "authors": [ "YannickMeeus" ], "repo": "Silly-Goose-Software/event-sauced-ts", "url": "https://github.com/Silly-Goose-Software/event-sauced-ts/issues/255", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2535413632
一直在加载 load project 已经无法使用了 用v0.2.4的版本尝试一下 麻烦了,还有一个问题,就是每一次打开项目,有概率无法加载, 其他插件也停止了,没有任何信息,需要启动,触发进度才可以继续操作 用了最新的版本还会吗,有复现的gif吗 可以查看一下 vscode 的 output 日志有没有什么报错之类的 目前没发现错误,正在load project 这是正常情况, 非正常情况就是一直 load project 你这个不是最新版本吧的截图吧 这是最新的 你本地开代理了吗,因为 unpkg 和 jsdelivr需要梯子才能拉下来 有代理,但不是全局的 duplicated: #89
gharchive/issue
2024-09-19T06:33:32
2025-04-01T04:33:05.201658
{ "authors": [ "Simon-He95", "xiaojia21190" ], "repo": "Simon-He95/common-intellisense", "url": "https://github.com/Simon-He95/common-intellisense/issues/85", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
173375171
Hosting a second demo? Hi there, I would be happy to host a second demo if you want. Thank you, Timo The demo has data donated by the admins of two separate servers. If you want your server to be included there, feel free to shoot me an email. My email is in my profile. If you want to run it in your server and demo a different language, you can include the link to there in the email. ~Simon I mailed you, if you like we can discuss the details via email?
gharchive/issue
2016-08-26T05:50:47
2025-04-01T04:33:05.219838
{ "authors": [ "SimonOrJ", "ntimo" ], "repo": "SimonOrJ/CoreProtect-Lookup-Web-Interface", "url": "https://github.com/SimonOrJ/CoreProtect-Lookup-Web-Interface/issues/8", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
928426957
When running on Docker I just get 502 I've deployed the docker image to k8s for the ui and when I try to access the contianer on port 9000 I just get 502. I have tried running the docker image locally but to the same result. Any logs that you can share to try and debug? docker-compose logs ui Was this solved by the new version of the API server?
gharchive/issue
2021-06-23T16:12:24
2025-04-01T04:33:05.250498
{ "authors": [ "dillu24", "tomjohnburton" ], "repo": "SimplyVC/panic_polkadot", "url": "https://github.com/SimplyVC/panic_polkadot/issues/36", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1871505562
add the docker file The docker file has been added with a system to document what was simulated and the results. Rejected as we can't remove the check for complex arithmetic. The code needs to be run in an environment where the complex variables are setup correctly.
gharchive/pull-request
2023-08-29T11:51:02
2025-04-01T04:33:05.251379
{ "authors": [ "njsaugat", "pmanandhar1452" ], "repo": "SimulateAnything/wentworth-simulate-anything-acoustic-levitation-v0", "url": "https://github.com/SimulateAnything/wentworth-simulate-anything-acoustic-levitation-v0/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2448215183
Documentation of WG meeting Mon 5th Aug 2024 #37 Document the WG meeting on 5th August 2024, and add it to the Archives via the meeting summary tool https://archive-oracle.netlify.app/ Done https://snet-ambassadors.gitbook.io/singularitynet-archive/timeline/2024/august-2024/week-32#ai-ethics-wg Transferred to Vani and sent for payment
gharchive/issue
2024-08-05T10:36:31
2025-04-01T04:33:05.253131
{ "authors": [ "VanCardui" ], "repo": "SingularityNet-Ambassador-Program/AI-Ethics-Workgroup", "url": "https://github.com/SingularityNet-Ambassador-Program/AI-Ethics-Workgroup/issues/39", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
216999375
Add Julia language definition This PR adds a definition file for the Julia language. It uses the Julia standard of 4 spaces for indentation. With this, the Julia files in Julia's standard library seem to be highlighted correctly. The only issue is names that contain numbers: the numbers are highlighted within the name. For example, in the type Int64, Int will appear normally (i.e. not highlighted) but 64 is highlighted as a number. It's part of the identifier, so it would be nice if Int64 (as just one of many examples) could be homogeneously colored. Is there a way I can specify that in the configuration file? Oddly enough, despite the aforementioned issue, the unsigned literal0x0000, for example, has the leading 0 highlighted but not the trailing zeros. Unfortunately not in a generic way. wsedit's highlighter works like this: Check all active -l* rules, take the longest match. If no rule matches, highlight on character class. You could of course mark Int64 as a keyword, but I don't think that's what you meant. I was thinking about adding an option to disable character class highlighting altogether, but I didn't have any good reason to do so yet. ... Now that I think about it, maybe switching from character-based highlighting to a style where the first character of a "word" determines its colour would be a good idea, but it doesn't look like a quick 5 minute fix. (Hell, it might even boost performance in some cases) Cool, thanks for the explanation. An advantage for Julia of the current behavior is that Julia allows numeric literal juxtaposition for multiplication. That is, 2x means 2*x. Yeah, I think I've seen a few languages do that kind of thing. It's definitely cool to write, but it doesn't make my job any easier =) So maybe "try to match the longest identifier ([A-Za-z_][A-Za-z0-9_]*) and make sure to paint it all white"? That would at least work for Julia.
gharchive/pull-request
2017-03-25T19:53:44
2025-04-01T04:33:05.259066
{ "authors": [ "SirBoonami", "ararslan" ], "repo": "SirBoonami/wsedit", "url": "https://github.com/SirBoonami/wsedit/pull/17", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
230898476
[Linux] Failure to Launch I'm running into a bit of a problem, but perhaps it's just user error and I'm just overlooking something. I'm using Ubuntu 17.04, and I've installed the following packages which I believe are the same as the dependencies listed on the download page of the TrueCraft website: mono-complete gtk-sharp2 libwebkitgtk-3.0-0 libsdl-mixer1.2 I'm able to use the launcher just fine; I was able to sign in, select single player, and generate a world. The problem started once I attempted to launch the game, it simply returned me to the launcher. I checked around the issues for this problem, but could only find it for people using Windows, and your recommendation was to use Linux. Here's the output from mono: Image format: http://i.imgur.com/VeVlNTJ.png What I assume is the important part: Cannot connect to server socket err = No such file or directory Cannot connect to server request channel jack server is not running or cannot be started JackShmReadWritePtr::~JackShmReadWritePtr - Init not done for 4294967295, skipping unlock JackShmReadWritePtr::~JackShmReadWritePtr - Init not done for 4294967295, skipping unlock X Error of failed request: BadValue (integer parameter out of range for operation) Major opcode of failed request: 154 (GLX) Minor opcode of failed request: 3 (X_GLXCreateContext) Value in failed request: 0x0 Serial number of failed request: 78 Current serial number in output stream: 79 AL lib: (EE) alc_cleanup: 2 devices not closed Any help is appreciated :) can you try running the server and client separately? Run the server normally, and run the client with mono TrueCraft.Client.exe localhost LakenH. Appreciate the quick response :) I was able to run the server fine, but get the identical error when running the client with that command. What graphics driver are you using? Using a NVIDIA one, version 375.66, but I see that there are more recent versions; should I try updating it? Try nouveau. Only slightly better, lol. This time it at least launches, but still throws a similar error. However, after the first key press, the game crashes and gives quite the error. Output: https://gist.github.com/LakenH/39f7aa3b966bbdc67633031ae0bdc0c9 Ouch. I'm not sure what the issue is, but seeing as the client is going to be overhauled soon it's not worth looking into. Understandable, thanks for your time! After some further testing, it seems to only throw that error and crash once I open the inventory. I'm still able to maneuver with WASD and such. So this might be two separate issues.
gharchive/issue
2017-05-24T01:55:45
2025-04-01T04:33:05.265827
{ "authors": [ "LakenH", "SirCmpwn" ], "repo": "SirCmpwn/TrueCraft", "url": "https://github.com/SirCmpwn/TrueCraft/issues/267", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
180485384
Add bind mapper This pull request creates a bind mapping subsystem (with unit tests), and integrates it into aerc. Some default binds are hardcoded, but can be overwritten in aerc.conf. I'm a little uncertain about how I've named the bind mapping system, so if you have any better ideas, or other suggestions for improvements, I'm open to them. Potential way to provide vim-like behaviour: At ui.c:187 Create a queue of pending input events. Add events from tb_peek_event to input queue. Pop events from the queue and process them one at a time. When handling a command generated from bind_handle_key_event, determine which input events it would generate, and add them to the input queue. Generating basic key presses will be easy. Support for <CR> etc. will need work. Perhaps existing logic from bind.c could be leveraged. Processing continues until there are no more events. There's a gotcha with this though: a bad user configuration could generate an infinite loop. This needs to be caught somehow. An additional optimisation: Add a rerender_required flag to state, and make rerender simply set it. Then at the end of the ui_tick, if it's been set, unset it and actually rerender the ui. This will prevent spurious rerenders. That sounds perfect. This now handles key events as intended, however the new functionality (bind_translate_key_name) doesn't have any unit tests yet, and there's nothing to prevent an infinite loop. How do you make a binding that i.e. types in Ctrl+C? In the [input] section of aerc.conf: f = <Ctrl+c> Cool. LGTM. Do you want this merged now or do you want to add tests and loop checking first? I'll add another unit test or two, then merge. I think loop checking can wait, as it's not that likely to happen by accident, and I'd rather give it some thought and put in an elegant safeguard, than hack something in now and never bother to improve it. Sounds good. A new unit test uncovered that I'd missed support for <Meta+key>. That's now been resolved, and this is ready for merging. Thanks!
gharchive/pull-request
2016-10-01T23:11:10
2025-04-01T04:33:05.273426
{ "authors": [ "SirCmpwn", "eXeC64" ], "repo": "SirCmpwn/aerc", "url": "https://github.com/SirCmpwn/aerc/pull/11", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1671242751
Supporting knowledge of other participants In some of OFMC's examples, it is specified when a participant has knowledge of another participant. Is this something that should be supported? Seems to be kinda supported already
gharchive/issue
2023-04-17T13:55:18
2025-04-01T04:33:05.278479
{ "authors": [ "PhilipFlyvholm" ], "repo": "SirSorensen/Seculizer", "url": "https://github.com/SirSorensen/Seculizer/issues/49", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1004704132
Downloads links in menu and content I've added links to downloads in the menu and in relevant MD files. Fixes #53 Looking at it I do believe we need to add some sort of indicator to point out its a different link then the others. But we can add that later @markvanaalst : I agree. Also, since they are on another site, I was thinking perhaps the visualization used in the waffle menu that shows they go somewhere else? That would draw attention, but also be informative. I'm going to take a look into that today when I get a chance.
gharchive/pull-request
2021-09-22T19:42:12
2025-04-01T04:33:05.285165
{ "authors": [ "jst-cyr" ], "repo": "Sitecore/developer-portal", "url": "https://github.com/Sitecore/developer-portal/pull/64", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
838625182
Support import and export declarations Hope to support ES Module. acorn supported sourceType option. Can Sval supports this option or enhance the extensibility of options? Any progress on this? @curly210102 @TheShermanTanker Supported from 0.5.1
gharchive/issue
2021-03-23T11:27:12
2025-04-01T04:33:05.302661
{ "authors": [ "Siubaak", "TheShermanTanker", "curly210102" ], "repo": "Siubaak/sval", "url": "https://github.com/Siubaak/sval/issues/68", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1624957888
Support to pull URLs from CSS This PR fixes #68 , #212 , #142 I did not add a test for this new code. I tested this by pulling scraping resources from https://www.purdue.edu/itap/iamo/p/login.css with a depth of 1. This can be tried by executing: cargo run https://www.purdue.edu/itap/iamo/p/login.css --depth 1 I have added a new struct named FileType which handles html, css, or other. This is a drop in replacement for the previously used is_html function in parser.rs. Edit: There is a slight issue, where if the request for the CSS file is 302, and that CSS file has a relative URL path, then the parser will not update this accordingly. Looking to fix. Edit 2: Auditing this bug, it appears that other files will suffer from not being to download relative dependencies on a 302 redirect. Perhaps this should go into another PR? (e.g. request /dir1/dir2/hello.html 302 -> /dir1/hello.html, hello.html contains dependency mystuff.css, where its real path is at /dir1/mystuff.css) Thanks you very much! I won't have time to review this in the next couple of days, but be sure that I'm really excited about this and I will give it a spin and a review as soon as I can! Another item is the use of __querystring__ in line 13 of url_helper.rs. This may panic on an instance of a request where the parameter utilizes special characters like = or ;...maybe I should adjust for url encoding? maybe I should adjust for url encoding? That's a good solution Unfortunately I've become way too busy to get back to this. I hope to revisit this in the future! Such a useful library...if I get a chance, I'll be sure to open another pr. No problem, I totally get it. You went way further than I never did with this. Thanks you very much :) I gave this a spin again. I managed to fix an old issue where books.toscrape.com did not have the star ratings because we where missing a font so I'm quite happy with that. I'm trying to fix this: Another item is the use of querystring in line 13 of url_helper.rs. This may panic on an instance of a request where the parameter utilizes special characters like = or ;...maybe I should adjust for url encoding? e.g.: thread '' panicked at 'Couldn't create folder ...some directory\fonts.googleapis.com\css__querystring__family=Archivo+Narrow:400,700: The directory name is invalid. (os error 267)', logger.rs:42:9 but cargo run https://www.purdue.edu/itap/iamo/p/login.css --depth 1 does not seems to provide this panick anymore
gharchive/pull-request
2023-03-15T07:59:44
2025-04-01T04:33:05.320520
{ "authors": [ "Skallwar", "dsgallups" ], "repo": "Skallwar/suckit", "url": "https://github.com/Skallwar/suckit/pull/213", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
354235773
Build failed Hello, there is a question when I build the QtMvvm with Qt Creator 5.9 'qpmx' 不是内部或外部命令,也不是可运行的程序 Project ERROR: qpmx initialization failed. Check the compilation log for details. Did you install all the build dependencies found in the readme? Especially qpmx? If yes, ensure that qpmx and qpm are in the PATH Also Qt 5.9 is not supported, as the QML-APIs used are only available from Qt 5.10 on. Can I close this issue?
gharchive/issue
2018-08-27T08:55:51
2025-04-01T04:33:05.370009
{ "authors": [ "Skycoder42", "wxddegithub" ], "repo": "Skycoder42/QtMvvm", "url": "https://github.com/Skycoder42/QtMvvm/issues/3", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
218059637
Fix plugin timer request not properly initialized connects #333 Don’t know how to write unit test to guard against this particular kind of bug without refactoring the transport interface. @rickmak New issue #338 opened for that. Merge via cb71a30
gharchive/pull-request
2017-03-30T02:01:03
2025-04-01T04:33:05.376754
{ "authors": [ "cheungpat", "rickmak" ], "repo": "SkygearIO/skygear-server", "url": "https://github.com/SkygearIO/skygear-server/pull/337", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
855802956
Unable to start controller The latest version doesn't start because of the following error argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout 2021-04-12T10:06:16.393Z INFO controller-runtime.metricsmetrics server is starting to listen {"addr": ":8080"} argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout 2021-04-12T10:06:16.393Z ERROR setup unable to read configuration {"error": "strconv.ParseBool: parsing \"\\\"true\\\"\": invalid syntax"} argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout github.com/go-logr/zapr.(*zapLogger).Error argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout /go/pkg/mod/github.com/go-logr/zapr@v0.2.0/zapr.go:132 argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout main.main argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout /workspace/main.go:74 argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout runtime.main argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout /usr/local/go/src/runtime/proc.go:204 After some investigation this seems related to the quote clause in the Secret.yaml. We should also understand why the installation fails as per https://github.com/Skyscanner/argocd-progressive-rollout/runs/2322327460?check_suite_focus=true but the action results in a pass.
gharchive/issue
2021-04-12T10:06:54
2025-04-01T04:33:05.409477
{ "authors": [ "maruina" ], "repo": "Skyscanner/argocd-progressive-rollout", "url": "https://github.com/Skyscanner/argocd-progressive-rollout/issues/51", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
141790850
Move Routes out of Providers Move the Routes out from regular Providers, and into a separate directory Routes. The Router component has to provide a helper provider abstract class to simplify and unify adding of routes to the Route Container. This helper must define a register method and implement the Pimple\ServiceProviderInterface, so the user doesn't need to deal with this, but just extends this helper when adding new Route definition loader classes. The register method must call the define method, which must be declared as abstract in the helper abstract class. The define method must add route info to a local property as array. The array must contain: uri - can be regex, all rules from Router component apply method - HTTP method in upper case action - the route definition, callable After define method call the register method must add all new definitions to the Route Container. Resolved.
gharchive/issue
2016-03-18T07:02:08
2025-04-01T04:33:05.425767
{ "authors": [ "slax0rr" ], "repo": "SlaxWeb/Framework", "url": "https://github.com/SlaxWeb/Framework/issues/47", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1316005525
[Betterdiscord] Submit the theme to the official website Suggestion Title^ Benefit Easier access This suggestion is unique [X] I have searched the issue tracker and did not find an issue describing my suggestion, especially not one that has been rejected. You may use the editor below to elaborate further. No response Please read the README. You will see why this theme is not uploaded there. might I ask if it's the ¨Original Work¨ requirement that's the problem?
gharchive/issue
2022-07-24T20:49:57
2025-04-01T04:33:05.427879
{ "authors": [ "Slddev", "isaktexas" ], "repo": "Slddev/ModernX-Theme", "url": "https://github.com/Slddev/ModernX-Theme/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
85354705
Exception while updating neighbours Minecraft - 1.7.10 --------Involved Mods------ Ender IO - 2.2.8.381 Sanguimancy - 1.1.9-29 Mekinism - 8.1.3.231 --------Crash Report------- Pastebin - http://pastebin.com/menKawd7 I can't tell you what was going on before this crash happened as this is from one of the servers using the Resonant Rise pack. I am still trying to recreate this issue, just thought you might have some insight on this one. This issue is between Sanguimancy and Mekanism...EIO is just triggering it. cool just checking.
gharchive/issue
2015-06-05T00:56:41
2025-04-01T04:33:05.430358
{ "authors": [ "GodNero", "tterrag1098" ], "repo": "SleepyTrousers/EnderIO", "url": "https://github.com/SleepyTrousers/EnderIO/issues/2543", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
326265050
[Request] Processing (killing) of Soul Vial'ed mobs A Machine that when powered and supplied with soul vials with mobs within, kills the creature as if it had been spawned in to the world and slain by any other means of killing. Could also have different toggle able settings to allow for player only drops, experience and adding looting effect somehow, all theses effect would make it use more energy or time of course Thank you for your suggestion, but we manage suggestions on http://ideas.theideawall.com/EnderIO This ticket will be closed because we want the list of open tickets to reflect known issues. Ok i will create a copy of the suggestion, on the mention site.
gharchive/issue
2018-05-24T19:46:48
2025-04-01T04:33:05.432648
{ "authors": [ "HenryLoenwind", "andr9528" ], "repo": "SleepyTrousers/EnderIO", "url": "https://github.com/SleepyTrousers/EnderIO/issues/4659", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
152794093
Show Block Name in GuiExternalConnectionSelector in 1.8.9. Implemens #1984 See #3287 for 1.7.10 PR. Oh, and something I noticed on your screenshot: Your text is touching the buttons, it needs to move down 1 pixel. Text is now 1 pixel lower and rendered behind buttons. Didn't look like i had to do some GL-stuff. Is it just that image or is the background a solid black now? Its just the background. if you look VERY closely you can see a coal ore on the right. I think i just took the screenshot at night. ok, that's fine. It may also be that the monitors at work are not really configured for anything. Actually, I know they aren't---I set them up. ;) 2 things: First, Block.getUnlocalizedName() is notoriously unreliable. It doesn't even work for all vanilla blocks, let alone modded blocks. Second, I expanded it a bit, I hope you don't mind: It's great. Thanks a lot for improving it! Especially that late at night. It's now even better, than I could have imagined. That's the wonder of collaboration; the result is better than what any participant could have accomplished on their own. But I'm still not sure if we are making clear what each button represents. So far we only have: Its position The name of the direction The name of the block The picture of the block Maybe we should add a couple more hints....lol, just kidding ;) I think it looks best with normal sized text in the buttons, and unicode text below.
gharchive/pull-request
2016-05-03T14:53:32
2025-04-01T04:33:05.437826
{ "authors": [ "HenryLoenwind", "MaPePeR", "tterrag1098" ], "repo": "SleepyTrousers/EnderIO", "url": "https://github.com/SleepyTrousers/EnderIO/pull/3288", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
1713829448
Update English documentation and screenshot to match the latest version of the software Update this page, including screenshots to match the current software version. Got it, thanks! Le mer. 17 mai 2023 à 12:45, Andras Lasso @.***> a écrit : Assigned #27 https://github.com/Slicer/SlicerLanguagePacks/issues/27 to @MohamedAllali https://github.com/MohamedAllali. — Reply to this email directly, view it on GitHub https://github.com/Slicer/SlicerLanguagePacks/issues/27#event-9268934383, or unsubscribe https://github.com/notifications/unsubscribe-auth/AWFEO4M4IU5EML6EA5VGXMTXGTB4RANCNFSM6AAAAAAYFBFQIM . You are receiving this because you were assigned.Message ID: @.***> Good evening, Mr. Andras I hope everything is going well. I want to ask you a question related to SlicerLanguagePacks update, isn't it that Qt toolkit is now integrated with SlicerLanguagePacks, so there is no need to download and install, my question to you is it better to delete the download and install Qt toolkit part of HowToUse https://github.com/Slicer/SlicerLanguagePacks/blob/main/HowToUse.md#:~:text=this language selector.-,Download and install Qt toolkit,is done%2C the lrelease file will be located in "/usr/bin/" .,-SlicerLanguagePacks/HowToUse.md . Le mer. 17 mai 2023 à 13:48, Mohamed Alalli BILAL @.***> a écrit : Got it, thanks! Le mer. 17 mai 2023 à 12:45, Andras Lasso @.***> a écrit : Assigned #27 https://github.com/Slicer/SlicerLanguagePacks/issues/27 to @MohamedAllali https://github.com/MohamedAllali. — Reply to this email directly, view it on GitHub https://github.com/Slicer/SlicerLanguagePacks/issues/27#event-9268934383, or unsubscribe https://github.com/notifications/unsubscribe-auth/AWFEO4M4IU5EML6EA5VGXMTXGTB4RANCNFSM6AAAAAAYFBFQIM . You are receiving this because you were assigned.Message ID: @.***> Ok, thanks for the clarification. Happy Weekend. Mohamed. Le sam. 20 mai 2023 à 14:15, Andras Lasso @.***> a écrit : Correct. Manual installation of Qt is not needed anymore (at least in recent Slicer Preview Releases), so it should be removed from the instructions. — Reply to this email directly, view it on GitHub https://github.com/Slicer/SlicerLanguagePacks/issues/27#issuecomment-1555921185, or unsubscribe https://github.com/notifications/unsubscribe-auth/AWFEO4KHCXSZLVPEPIDJ7RLXHDGZBANCNFSM6AAAAAAYFBFQIM . You are receiving this because you were mentioned.Message ID: @.***> @lassoan I've updated How to use, you can see from here, wait for your comment if there's anything that needs to be modified. done
gharchive/issue
2023-05-17T12:44:45
2025-04-01T04:33:05.490658
{ "authors": [ "MohamedAllali", "lassoan" ], "repo": "Slicer/SlicerLanguagePacks", "url": "https://github.com/Slicer/SlicerLanguagePacks/issues/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
193979240
Support for Symfony 3 Will this support for symfony 3 version in the future? Is this repo even still maintained ?
gharchive/issue
2016-12-07T07:44:19
2025-04-01T04:33:05.491637
{ "authors": [ "SlimenTN", "rainercedric23" ], "repo": "SlikNL/DompdfBundle", "url": "https://github.com/SlikNL/DompdfBundle/issues/15", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1167237672
Use POWER_SAVING_LEGACY by default BNO users experience strange delays with new power saving options. Revert to legacy to mitigate the issue and for further investigation. Closing my pull request in favor of #125.
gharchive/pull-request
2022-03-12T09:39:10
2025-04-01T04:33:05.494351
{ "authors": [ "deiteris" ], "repo": "SlimeVR/SlimeVR-Tracker-ESP", "url": "https://github.com/SlimeVR/SlimeVR-Tracker-ESP/pull/126", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1057619752
🛑 My blog(github.io) is down In 35d3d9c, My blog(github.io) (https://smartbrave.github.io/) was down: HTTP code: 0 Response time: 0 ms Resolved: My blog(github.io) is back up in a0eaf29.
gharchive/issue
2021-11-18T17:33:27
2025-04-01T04:33:05.514686
{ "authors": [ "SmartBrave" ], "repo": "SmartBrave/uptime", "url": "https://github.com/SmartBrave/uptime/issues/1155", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
999590635
🛑 My blog(CDN) is down In 7e45363, My blog(CDN) (https://sbrave.cn) was down: HTTP code: 0 Response time: 0 ms Resolved: My blog(CDN) is back up in 7f2e277.
gharchive/issue
2021-09-17T17:49:39
2025-04-01T04:33:05.517134
{ "authors": [ "SmartBrave" ], "repo": "SmartBrave/uptime", "url": "https://github.com/SmartBrave/uptime/issues/331", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1751262187
🛑 My blog(CDN) is down In 9b72292, My blog(CDN) (https://sbrave.cn) was down: HTTP code: 0 Response time: 0 ms Resolved: My blog(CDN) is back up in 3f6ded8.
gharchive/issue
2023-06-11T05:58:29
2025-04-01T04:33:05.519577
{ "authors": [ "SmartBrave" ], "repo": "SmartBrave/uptime", "url": "https://github.com/SmartBrave/uptime/issues/3969", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2656666690
Avoid native handling for zwave devices with odd endpoint layouts Check all that apply Type of Change [ ] WWST Certification Request If this is your first time contributing code: [ ] I have reviewed the README.md file [ ] I have reviewed the CODE_OF_CONDUCT.md file [ ] I have signed the CLA [ ] I plan on entering a WWST Certification Request or have entered a request through the WWST Certification console at developer.smartthings.com [x] Bug fix [ ] New feature [ ] Refactor Checklist [x] I have performed a self-review of my code [x] I have commented my code in hard-to-understand areas [x] I have verified my changes with unit tests Description of Change The native handlers for switch on/off and setLevel commands only send zwave commands to the root endpoint, 0, and these devices are overriding the base drivers component to endpoint mapping. Summary of Completed Tests @lelandblue this may be something that we want to skip the beta soak time to help users experiencing the issue sooner rather than later.
gharchive/pull-request
2024-11-13T20:20:23
2025-04-01T04:33:05.530740
{ "authors": [ "cjswedes" ], "repo": "SmartThingsCommunity/SmartThingsEdgeDrivers", "url": "https://github.com/SmartThingsCommunity/SmartThingsEdgeDrivers/pull/1752", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2426447323
🛑 MPC Homepage is down In 37075ed, MPC Homepage (https://www.minorplanetcenter.net) was down: HTTP code: 0 Response time: 0 ms Resolved: MPC Homepage is back up in e6182a2 after 4 minutes.
gharchive/issue
2024-07-24T02:20:20
2025-04-01T04:33:05.568990
{ "authors": [ "ChrisMoriarty" ], "repo": "Smithsonian/upptime", "url": "https://github.com/Smithsonian/upptime/issues/2854", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2435638193
🛑 NEOCP is down In a1563a1, NEOCP (https://minorplanetcenter.net/iau/NEO/toconfirm_tabular.html) was down: HTTP code: 0 Response time: 0 ms Resolved: NEOCP is back up in bea5b90 after 4 minutes.
gharchive/issue
2024-07-29T14:50:14
2025-04-01T04:33:05.571544
{ "authors": [ "ChrisMoriarty" ], "repo": "Smithsonian/upptime", "url": "https://github.com/Smithsonian/upptime/issues/3883", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2440153319
🛑 MPC Homepage is down In 7a84076, MPC Homepage (https://www.minorplanetcenter.net) was down: HTTP code: 0 Response time: 0 ms Resolved: MPC Homepage is back up in 7726613 after 4 minutes.
gharchive/issue
2024-07-31T14:06:39
2025-04-01T04:33:05.574041
{ "authors": [ "ChrisMoriarty" ], "repo": "Smithsonian/upptime", "url": "https://github.com/Smithsonian/upptime/issues/4460", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
127416809
Fixed a bug where string couldn't be app arguments lib.load_string failed to properly call Lua's loadstring built-in. Merged into max-next.
gharchive/pull-request
2016-01-19T10:39:19
2025-04-01T04:33:05.591949
{ "authors": [ "aequabit", "eugeneia" ], "repo": "SnabbCo/snabbswitch", "url": "https://github.com/SnabbCo/snabbswitch/pull/717", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
312431787
UIViewController in TabBarViewController hidden tabbar item.snp.bottom.equalToSuperview() invalid Bottom let logoView = OWSeriesLogoView() view.addSubview(logoView) logoView.snp.makeConstraints { (maker) in maker.top.equalTo(menuView.snp.bottom) maker.left.right.equalToSuperview() maker.bottom.equalToSuperview() } Issue Info Info Value Platform e.g. ios Platform Version e.g. 9.0 SnapKit Version e.g. 4.0.0 Issue Description UIViewController in TabBarViewController hidden tabbar item.snp.bottom.equalToSuperview() invalid There are 44 blanks from the bottom Not really sure I can help here your sample code is rather incomplete and the description lacking some details. Are you trying to create auto layout pinned to items in the tab bar? I'm not sure if that is legal in auto layout.
gharchive/issue
2018-04-09T08:02:43
2025-04-01T04:33:05.595305
{ "authors": [ "ljs19850321", "robertjpayne" ], "repo": "SnapKit/SnapKit", "url": "https://github.com/SnapKit/SnapKit/issues/505", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2058307562
Change geographic location in opera proxy for android Hello, thank you again for this VPN :) I have a question please, why opera does not protect its VPN i.e. why opera does not put a protection mechanism to prevent the use of its proxy outside of opera browser? or maybe there is ads when using this proxy so opera can finance its vpn ? I have another question please: when using the Opera browser, the Opera browser automatically chooses the best VPN location for faster browsing, can Standalone Opera VPN client do this by choosing the default location? fastest geographic location automatically? and also how we can choose the geographical location when we use opera-proxy for android because we cannot add command line arguments to the shortcut like in windows 10 os? Thanks ! Another question please : we know that opera was bought by a Chinese consortium in 2016, I do not accuse this company which is certainly honest but we know that the Chinese communist party likes to control and monitor people and likes to spy on people and imposes on Chinese companies to follow their orders, so is there a security risk when using opera proxy when for example the Chinese communist party orders opera to give out user passwords? Another quetion please : hola -proxy has option " peer proxies on residental IPs (consumer ISP) ", I read that the tuxler vpn https://www.tuxlervpn.com/ allows us to use the real IP addresses of users and if a malicious user takes our IP address and does illegal activity, we can be punished and go to prison....is it the same thing with hola proxy when we use hola -proxy has option " peer proxies on residential IPs" ? hola proxy allows people to use our real ip address ? I have a question please, why opera does not protect its VPN i.e. why opera does not put a protection mechanism to prevent the use of its proxy outside of opera browser? or maybe there is ads when using this proxy so opera can finance its vpn ? They actually do protect it from unauthorized usage. There is a login and password required for interaction with their API, and sophisticated digest authentication scheme required to sign requests. But since these credentials are baked into Opera browser, it is possible to analyze compiled machine code and figure out how browser interacts with API to get access. But this is as much as they can do to provide public service in their software without user registration. I have another question please: when using the Opera browser, the Opera browser automatically chooses the best VPN location for faster browsing, can Standalone Opera VPN client do this by choosing the default location? fastest geographic location automatically? If someone will implement this, I'll be glad to accept pull request. But in general they offer just three continents (America, Europe, Asia) and it's quite evident which is closer to you. and also how we can choose the geographical location when we use opera-proxy for android because we cannot add command line arguments to the shortcut like in windows 10 os? Thanks ! If you're running opera-proxy on Android using something like Termux, you can add command line parameters the same way after name of executable. Another question please : we know that opera was bought by a Chinese consortium in 2016, I do not accuse this company which is certainly honest but we know that the Chinese communist party likes to control and monitor people and likes to spy on people and imposes on Chinese companies to follow their orders, so is there a security risk when using opera proxy when for example the Chinese communist party orders opera to give out user passwords? Any VPN company, Internet service provider or hosting provider may listen users traffic, regardless Chinese or not. Even European hosting providers follow such orders and even actively interfere with service's traffic to perform man-in-the-middle attack. However, in Opera case such traffic interception is kind of fruitless. Almost all sites are HTTPS now and they won't be able to perform decryption (unless they have TLS certs of that site or cooperating with them certificate authority). Without decryption all they can do is to learn visited domains and your connecting IP address. So I would say you're not risking your passwords more than usual. Another quetion please : hola -proxy has option " peer proxies on residental IPs (consumer ISP) ", I read that the tuxler vpn https://www.tuxlervpn.com/ allows us to use the real IP addresses of users and if a malicious user takes our IP address and does illegal activity, we can be punished and go to prison....is it the same thing with hola proxy when we use hola -proxy has option " peer proxies on residential IPs" ? hola proxy allows people to use our real ip address ? Right now Hola forwards traffic of free users mostly through datacenter IP addresses because residential traffic is quite expensive. However in general Hola allows the use of residential IP addresses for some cases. Otherwise it will not be possible to access services with geo-restricted video content like Netflix or Hulu, which do not expect user to have datacenter IP address. Such IP sharing was a public concern and caused many complaints: http://adios-hola.org/ hola-proxy does not allow your IP to be used by others. It was specifically one of goals of such alternative implementation. Such function does not exists in hola-proxy and never will be. Ohhh very very interesting and useful information! Thanks a lot for your precious help and for all efforts to make many great awesome softwares 🥇 💯 ! We suggest that you put a pop-up window that launches automatically when we use one of your softwares to make easily a donation via paypal or buy me a coffee or bitcoin.....when we use one of your magnificent softwares so it will allow people who can makes donations to support these fantastic projects and also display messages to encourage people to share your softwares on social networks like Twitter, Facebook, Snap....it will give even more visibility to this amazing softwares.... I also suggest making a discussion section so you can include these important information and revelations that you so kindly made to me around the internal functioning of opera proxy, hola proxy.....in order to definitively answer any future questions and curiosity of users....thank you again very much for all your efforts to invent such useful softwares 🥇 !
gharchive/issue
2023-12-28T11:09:33
2025-04-01T04:33:05.611874
{ "authors": [ "Snawoot", "trimechee" ], "repo": "Snawoot/opera-proxy", "url": "https://github.com/Snawoot/opera-proxy/issues/42", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
599057421
Replace programming language icons Used icons of programming languages are outdated in my opinion. I propose replacing them with Material, it will be much nicer to look aesthetically. If you can send me some examples of icons that could look better than the current icons I can consider replacing them. What resolution should the icons be in? .png (What is the resolution? 512x512, 256x256, 64x54) ot .svg (vector)? @Snazzah @dmsass 512x512 For reference, here are the current icons on the application. I do agree that there lacks any consistency in style, I am open to any changes. My suggestions javascript: https://upload.wikimedia.org/wikipedia/commons/thumb/9/99/Unofficial_JavaScript_logo_2.svg/512px-Unofficial_JavaScript_logo_2.svg.png python: https://upload.wikimedia.org/wikipedia/commons/thumb/c/c3/Python-logo-notext.svg/512px-Python-logo-notext.svg.png rust: https://www.rust-lang.org/logos/rust-logo-512x512-blk.png text: https://upload.wikimedia.org/wikipedia/commons/thumb/a/ae/Icon-txt.svg/512px-Icon-txt.svg.png typescript: https://www.typescriptlang.org/icons/icon-512x512.png unknown: lets use twemoji cause why not https://discord.com/assets/aef26397c9a6a3afee9c857c5e6f3317.svg v: https://vlang.io/img/v-logo.png sass because cool sass: https://sass-lang.com/assets/img/logos/logo-b6e1ef6e.svg @botatooo I replaced the V logo and Python logo The JS and TS logos you provided are a bit less consistent and might clip the text when using the big_icon setting. I'm considering replacing the ? unknown logo with a blank file icon Rust is more or less the same as it is already, maybe add a color background like for JS and TS? (Same for Sass) i’ve found this for javascript, seems to fit with html and css’s design for typescript idk if we could fix that because that’s the logo they use in their marketing and stuffs From what I've seen, the first suggestion for the JS logo is indeed the most commonly used icon, but I see no benefit of using it over the better suited logo we currently have that allows it to be cut off a bit at the edges. my fork has custom consistent icons that i think are really nice :l if you want i can just zip em up and upload @TorchedSammy Those icons look great! Could you upload them in a zip? Here they are, ill probably make a separate repo for convenience @TorchedSammy we could just make a folder in the repo to add the icons into for anyone that wants to use their own application ID @Snazzah oh yea, feel free to and for any new icons ya can just ask me if you'd like @TorchedSammy Also, speaking of your fork, would it be alright if you pushed some changes here? I've been meaning to get this package updated a bit anyways. @Snazzah yep, I can open another PR with the changes in a while
gharchive/issue
2020-04-13T18:35:05
2025-04-01T04:33:05.622812
{ "authors": [ "Calvin-rgb", "FichteFoll", "Snazzah", "TorchedSammy", "botatooo", "dmsass" ], "repo": "Snazzah/SublimeDiscordRP", "url": "https://github.com/Snazzah/SublimeDiscordRP/issues/49", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2640987615
ss_mkt Parameter Causing Anomaly Detection in DuckDuckGo Results Description When using DuckDuckGo's API for querying, adding the ss_mkt parameter seems to trigger anomaly detection. Specifically, if ss_mkt=FR (or any other country code variant) is appended to the request URL, the response includes a script with window.execDeep = function() {DDG.deep.anomalyDetectionBlock({...})}, which appears to be blocking or altering the query response. Steps to Reproduce Use the following working URL to perform a query (e.g., what is weather today): https://links.duckduckgo.com/d.js?q=what is weather today&kl=wt-wt&l=us-en&p=1&s=0&a=h_&dl=en&ct=FR&df=m&vqd=4-130790525484318490675696036619647809147&bing_market=en-US&p_ent=&dp=a2VtvYQbwBUAnJjgW_GROFNV_mc6SfdjMK9PLqH42pqZ3fsCcOsNuDJagf0YUf4NazjDeNsiE8uQBCPRHf0rDV9gnMMek-we5JwQZ87pMeFh4-7e5KVEkLAYafYyGqW7.FOFBUBx1to6nRs7UUN6QSA&sjs=forecast&perf_id=2d03f98f2d7a55da&parent_perf_id=7fe2f5d308add94a&host_region=eun&sp=1&dfrsp=1&bpa=1&baa=1&bcca=1&btaa=1&wrap=1&aps=0&bccaexp=b&biaexp=b&btaaexp=b&eclsexp=b&litexp=b&msvrtexp=b Add the ss_mkt parameter, e.g., ss_mkt=FR, to the URL: https://links.duckduckgo.com/d.js?q=what is weather today&kl=wt-wt&l=us-en&p=1&s=0&a=h_&dl=en&ct=FR&df=m&vqd=4-130790525484318490675696036619647809147&bing_market=en-US&p_ent=&dp=a2VtvYQbwBUAnJjgW_GROFNV_mc6SfdjMK9PLqH42pqZ3fsCcOsNuDJagf0YUf4NazjDeNsiE8uQBCPRHf0rDV9gnMMek-we5JwQZ87pMeFh4-7e5KVEkLAYafYyGqW7.FOFBUBx1to6nRs7UUN6QSA&sjs=forecast&perf_id=2d03f98f2d7a55da&parent_perf_id=7fe2f5d308add94a&host_region=eun&sp=1&dfrsp=1&bpa=1&baa=1&bcca=1&btaa=1&wrap=1&aps=0&bccaexp=b&biaexp=b&btaaexp=b&eclsexp=b&litexp=b&msvrtexp=b&ss_mkt=FR 3. Observe the response which includes window.execDeep = function() {DDG.deep.anomalyDetectionBlock({...})} instead of the expected data. Expected Behavior Adding the ss_mkt parameter should not trigger anomaly detection, and the query should return results as expected. Actual Behavior Adding ss_mkt=FR results in the DuckDuckGo response containing a script with DDG.deep.anomalyDetectionBlock, which seems to block or modify the response. Environment Query Tool: DuckDuckGo API Affected Parameter: ss_mkt Additional Information This anomaly is observed specifically when setting the ss_mkt parameter to a value such as FR. Other values may produce similar results. Request If the current DuckDuckGo API does not support the ss_mkt parameter, please consider removing or disabling it to avoid triggering anomaly detection blocks. I have the same issue using langchain and the duck duck search library, please consider removing the parameter from the code. Temporary fix import { search, SafeSearchType } from 'duck-duck-scrape'; // import * as DDG from 'duck-duck-scrape'; const searchResults = await search('node.js', { safeSearch: SafeSearchType.STRICT }, { uri_modifier: (rawUrl) => { const url = new URL(rawUrl); url.searchParams.delete("ss_mkt"); return url.toString(); } ); Hello, Is there any reference where the have mentioned about replacing ss_mkt by bing_market?. For me replacing ss_mkt by bing_market flag does not give any answer. Thank you. Looking at doing a DDG request now does seem the query parameter was replaced with bing_market, also assuming the values were somewhat changed since the default is "us" with what im receiving being "en-US". Don't know of other changes but that seems to be the most obvious one. @flaviogoncalves If you just make a PR and summarize your changes I can review and merge It seems that new version is not fixing the issue. I can do one query per like ~15-30 minutes Not sure how else to fix this other than adding common browser headers to each request, I don't know what else I really can do to solve this. You might just need to deal with proxies whenever you hit an anomaly.
gharchive/issue
2024-11-07T13:07:34
2025-04-01T04:33:05.632406
{ "authors": [ "Snazzah", "Tomas2D", "dzek69", "flaviogoncalves", "sarfudheen" ], "repo": "Snazzah/duck-duck-scrape", "url": "https://github.com/Snazzah/duck-duck-scrape/issues/140", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
376992691
Update kubernetes to newer version apiVersion no longer beta update to using ClusterIP for newer networking @jondubois Hey John, I'll make those changes shortly. I only add the server and deployment to the name to make it easier for me to remember what's going on in our production environment. There is no need to do it here. Sorry about that. As far as ClusterIP is concerned, it's the way you're supposed to do internal networking in Kubernetes. The other kind is going to be deprecated like loadbalancer is. That's at least my understanding. Thanks! oops... wrong account.. sorry lol @jondubois I fixed those changes you asked about. Thanks! @jondubois hey man I was thinking if the docker images shared the same persistent storage claim could you replicate state servers and remove the single point of failure? You could use ClusterIP to handle the communication. What do you think? scc-state doesn't use persistent storage; it's all in memory. scc-state is not really a single point of failure because SCC can keep working without it. There will only be partial disruption of the service if both scc-state and an scc-broker fail at the exact same time; and even if this unlikely scenario were to happen, the disruption should only last a few seconds. I will update the scc-guide to remove the phrase 'single point of failure' because it sound like it's a problem but it's not. You always need to make a trade-off between consistency or availability and SCC leans slightly more towards consistency but it's still very resilient in terms of availability. @happilymarrieddad Your PR looks good now. It might take a bit of time to merge because I want to test it first. @jondubois no worries @happilymarrieddad Thanks for this contribution :) Sorry it took so long; I was busy and didn't get the time to test it until recently. After testing it on GKE, I changed it to expose the socketcluster (worker) service using NodePort otherwise it doesn't work on GKE. This could also be an issue on other platforms. @jondubois cool thanks! Yea, I don't know.. I use kops and AWS for production. I played around a little bit on GKE but it was too much magic for me. Thanks!
gharchive/pull-request
2018-11-02T22:57:53
2025-04-01T04:33:05.656541
{ "authors": [ "0xsegfault", "happilymarrieddad", "jondubois" ], "repo": "SocketCluster/socketcluster", "url": "https://github.com/SocketCluster/socketcluster/pull/453", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2152763903
Add option for skipping instrument rescaling when not available As described in #47. This PR adds an option error_on_missing to read_nda (and read) that allows the parser to continue even when instrument scaling is not possible due to to a missing multiplier. The current behaviour is preserved by the default value of error_on_missing = True. This allows the package to be used more easily downstream without needing to wait for new multipliers to be added and released. Before merging this, there is no test data available to contributors (presumably you have some somewhere!) so consider this PR as untested! (It works for me with my own data, of course...) -- see #48 for my thoughts on how best to do this. See #47
gharchive/pull-request
2024-02-25T13:22:15
2025-04-01T04:33:05.700249
{ "authors": [ "ml-evs" ], "repo": "Solid-Energy-Systems/NewareNDA", "url": "https://github.com/Solid-Energy-Systems/NewareNDA/pull/49", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
293812820
Microsoft Core Dependency Injection as a DI for SolrNet.Cloud Hi, I have asked question in SO https://stackoverflow.com/questions/48564227/can-i-use-microsoft-core-dependency-injection-with-solrnet-cloud-core about ability to use Microsoft Core Dependency Injection as a DI for SolrNet.Cloud But I can assume that there is no support of Microsoft DI in Solr.Net.Cloud. Are there any plans to introduce support of Microsoft Core DI in Solr.Net.Cloud in future? @Fatalityap Sure, we want to add also Microsoft DI support for SolrNet.Cloud . Please feel free to start the work on it and help us getting it supported. Ok, I have talked with my management. Maybe we can write support for it. At least my management do not mind for now. Did this ever get progressed? I am facing the same issue. Also looking for Microsoft Core DI support. Any news on this?
gharchive/issue
2018-02-02T08:44:27
2025-04-01T04:33:05.708848
{ "authors": [ "Fatalityap", "Wintermute79", "gjunge", "technicaldogsbody" ], "repo": "SolrNet/SolrNet", "url": "https://github.com/SolrNet/SolrNet/issues/381", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1272463521
Non-recoverable error on launch Mod list: {big ones K2+SE} mod-list.zip First save: {file was too big to upload on github} https://drive.google.com/file/d/1eyUhUnJWvyoELBNjYKQ-s0_fZ1OVQ6os/view?usp=sharing Error message: Second save: {same mods} FILIP MOD 002.zip Error message: Error is shown on both saves seconds after loading of map , only new mod that was added to modpack during playthrough , nothing removed ... Edit: Only thing i can think of that only mod that is somewhat modying something about trains is LTN but that dont do anything with signals only with train stations Fixed by commits 61ba57df22043e855c62db69909f7dd5b5e960af and c2000a77fc919b98ae3aee9961c2c5b84d4444a0 (version 1.1.1). This rail network is crazy, and there's a lot of weird things going on. (The mod also identified plenty of issues :)) The way this railway network is made causes a lot of exploration to get enough rails in the graph. This makes the update take a few seconds (!). Makes it also a good test-case for potential optimization. Al;so makes me think that a non-realtime (single use) annotation (through some button) could be useful for such large and interconnected networks. Very nice test case for the mod overall, I'll be using it a lot. Thank you for reporting this.
gharchive/issue
2022-06-15T16:18:22
2025-04-01T04:33:05.796901
{ "authors": [ "Sopel97", "TheAsKo" ], "repo": "Sopel97/railway_signalling_overseer", "url": "https://github.com/Sopel97/railway_signalling_overseer/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
951014372
Reading log guidance Hi I wanted to test various mitigation settings to compare performance and was wondering what was the best way going about to do that from inspecting logs, given a line sequence like: 2021-07-22 20:29:10.780 AnimationLockLatencyHandler 16f4: C2S_ActionRequest(0175): actionId=08cf sequence=0e22 delay=1358ms 2021-07-22 20:29:10.963 AnimationLockLatencyHandler 16f4: S2C_ActionEffect(0102): actionId=08cf sourceSequence=0e22 rtt=188ms pingLatency=33ms delay=75ms wait=600ms->487ms next=19:29:11.4507252 2021-07-22 20:29:11.451 AnimationLockLatencyHandler 16f4: S2C_ActionEffect(0102): actionId=0007 sourceSequence=0000 serverOriginated wait=108ms->100ms next=19:29:11.5517470 2021-07-22 20:29:12.214 AnimationLockLatencyHandler 16f4: C2S_ActionRequest(0175): actionId=1e8e8c sequence=0e23 delay=666ms Also sometimes the wait value is not 600, even for non-single digit actionids (which I read on a comment somewhere to ignore) 2021-07-22 20:28:30.228 AnimationLockLatencyHandler 16f4: S2C_ActionEffect(0102): actionId=4976 sourceSequence=0e01 rtt=204ms pingLatency=34ms delay=75ms wait=349ms->220ms next=19:28:30.4481509 2021-07-22 20:28:30.510 AnimationLockLatencyHandler 16f4: C2S_ActionRequest(0175): actionId=4975 sequence=0e02 delay=76ms What values are we trying to minimize here? is delay=X what the program "adjusts" our latency to be in terms of performing the actions locally? actionId states which action you used, including auto attacks, weaponskills, spells, and abilities. When auto attack happens, wait value may be any value between previous action's animation lock duration and 100ms. delay is the simulated server response time. Note that it's not ping; the server may take longer than usual even when your ping itself is stable, like when you were moving around. Thanks for responding. So in essence if we want to see whether the options such as Use Delay Detection and Use Latency Correction are better or worse, we want to minimize the wait value? (to ensure the lowest amount of time between animation lock and next action)? Yes, that's correct.
gharchive/issue
2021-07-22T20:13:51
2025-04-01T04:33:05.800275
{ "authors": [ "Soreepeong", "perks" ], "repo": "Soreepeong/XivAlexander", "url": "https://github.com/Soreepeong/XivAlexander/issues/117", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2554103335
🛑 Marcel Web is down In ecc1b3b, Marcel Web ($MARCEL_WEB) was down: HTTP code: 0 Response time: 0 ms Resolved: Marcel Web is back up in 82c42db after 3 hours, 1 minute.
gharchive/issue
2024-09-28T08:14:05
2025-04-01T04:33:05.844339
{ "authors": [ "Sundypha" ], "repo": "Source-Graphics-GmbH/upptime", "url": "https://github.com/Source-Graphics-GmbH/upptime/issues/1122", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1138130860
register @JuliaRegistrator register Error while trying to register: "Tag with name v0.4.0 already exists and points to a different commit"
gharchive/issue
2022-02-15T03:28:51
2025-04-01T04:33:05.861640
{ "authors": [ "JuliaRegistrator", "shinaoka" ], "repo": "SpM-lab/SparseIR.jl", "url": "https://github.com/SpM-lab/SparseIR.jl/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1887268758
🛑 Yiffed Main is down In 75f4c9a, Yiffed Main (https://yiffed.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: Yiffed Main is back up in 2d7b0a8 after 10 minutes.
gharchive/issue
2023-09-08T09:08:45
2025-04-01T04:33:05.927751
{ "authors": [ "SparksTheFolf" ], "repo": "SparksTheFolf/STF-Uptime-Status", "url": "https://github.com/SparksTheFolf/STF-Uptime-Status/issues/246", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1901358194
🛑 Yiffed Main is down In f29b1ae, Yiffed Main (https://yiffed.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: Yiffed Main is back up in 9ea6436 after 36 minutes.
gharchive/issue
2023-09-18T16:49:50
2025-04-01T04:33:05.930091
{ "authors": [ "SparksTheFolf" ], "repo": "SparksTheFolf/STF-Uptime-Status", "url": "https://github.com/SparksTheFolf/STF-Uptime-Status/issues/897", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2177601144
Using Flags Hi Is there a way to use Enum Lookup Tables with Flags? I tried (of course) already and even though it works with an individual Flag, it fails when using multiple Flags as the foreign key reference doesn't handle bitwise added values. So is there any way to make this work? Possibly you would create overload for this method of "Flagged" case: EnumLookupExtension.cs#L129 Now it creates row per enum. You should create row per each combination of flags. Then any bitwise would be found in foreign table. I will try to go that way. Just didn't want to "invent the wheel" if it was already there and I just missed it. I don't claim credit for the GetCombinations method (google was my friend), but in case someone else is searching this then here's my change to GetEnumData and the added GetCombinations method: private static object[] GetEnumData(Type enumType, Type concreteType, bool useNumberLookup, bool usesDescription, Dictionary<int, string> enumValueDescriptions) { if (enumType.IsDefined(typeof(FlagsAttribute), false)) { return GetCombinations((Enum)Activator.CreateInstance(enumType)) .Select(x => { var instance = Activator.CreateInstance(concreteType); concreteType .GetProperty("Id") ?.SetValue(instance, x.Key); if (useNumberLookup) { concreteType .GetProperty("Name") ?.SetValue(instance, x.Value.ToString()); } if (usesDescription) { concreteType .GetProperty("Description") ?.SetValue(instance, enumValueDescriptions[x.Key]); } return instance; }) .ToArray(); } return Enum .GetValues(enumType) .OfType<object>() .Select(x => { var instance = Activator.CreateInstance(concreteType); concreteType .GetProperty("Id") ?.SetValue(instance, x); if (useNumberLookup) { concreteType .GetProperty("Name") ?.SetValue(instance, x.ToString()); } if (usesDescription) { concreteType .GetProperty("Description") ?.SetValue(instance, enumValueDescriptions[(int)x]); } return instance; }) .ToArray(); } private static Dictionary<int, string> GetCombinations(this Enum concreteEnum) { var fields = concreteEnum .GetType() .GetFields() .Where(f => f.Name != "value__") .DistinctBy(f => Convert.ToInt32(f.GetRawConstantValue())); var result = fields.ToDictionary(f => Convert.ToInt32(f.GetRawConstantValue()), f => f.Name); var max = Enum .GetValues(concreteEnum.GetType()) .Cast<int>() .Max(); var upperBound = max * 2; for (var i = 0; i <= upperBound; i += 2) { var s = Convert .ToString(i, 2) .PadLeft(Math.Abs(i - max), '0'); var bits = s .Select(chs => chs == '1') .Reverse() .ToArray(); if (!result.ContainsKey(i)) { var newComb = string.Empty; for (var j = 1; j < bits.Length; j++) { var idx = 1 << j; if (bits[j] && result.TryGetValue(idx, out var value)) { newComb = newComb + value + " | "; } } newComb = newComb.Trim([' ', '|']); if (!result.ContainsValue(newComb) && !string.IsNullOrEmpty(newComb)) { result.Add(i, newComb); } } } return result; }
gharchive/issue
2024-03-10T08:28:18
2025-04-01T04:33:05.934355
{ "authors": [ "AllanMichaelsen", "john-bartu" ], "repo": "SpatialFocus/EntityFrameworkCore.Extensions", "url": "https://github.com/SpatialFocus/EntityFrameworkCore.Extensions/issues/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
522726953
Fail during make install Environment System: window10 WSL Centos7 scipy==1.1.0 install pillow==5.1.0 Problem I have already chanege the 'CV_BGR2GRAY' to 'COLOR_BGR2GRAY' in file st_tissue_recognition/src/util.cpp then cmake and make install again but still get Error in step make install Error Information: [ 87%] Linking CXX executable recognize-tissue /mnt/e/BioInfo/APP/Anaconda3/bin/../lib/gcc/x86_64-conda_cos6-linux-gnu/7.3.0/../../../../x86_64-conda_cos6-linux-gnu/bin/ld: warning: libz.so.1, needed by /usr/local/lib64/libopencv_core.so.4.1.2, not found (try using -rpath or -rpath-link) /mnt/e/BioInfo/APP/Anaconda3/bin/../lib/gcc/x86_64-conda_cos6-linux-gnu/7.3.0/../../../../x86_64-conda_cos6-linux-gnu/bin/ld: warning: libjpeg.so.9, needed by /usr/local/lib64/libopencv_imgcodecs.so.4.1.2, not found (try using -rpath or -rpath-link) /mnt/e/BioInfo/APP/Anaconda3/bin/../lib/gcc/x86_64-conda_cos6-linux-gnu/7.3.0/../../../../x86_64-conda_cos6-linux-gnu/bin/ld: warning: libpng15.so.15, needed by /usr/local/lib64/libopencv_imgcodecs.so.4.1.2, not found (try using -rpath or -rpath-link) /mnt/e/BioInfo/APP/Anaconda3/bin/../lib/gcc/x86_64-conda_cos6-linux-gnu/7.3.0/../../../../x86_64-conda_cos6-linux-gnu/bin/ld: CMakeFiles/recognize-tissue.dir/main.cpp.o: in function `main': main.cpp:(.text+0xfd): undefined reference to `cv::imread(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, int)' ..... /mnt/e/BioInfo/APP/Anaconda3/bin/../lib/gcc/x86_64-conda_cos6-linux-gnu/7.3.0/../../../../x86_64-conda_cos6-linux-gnu/bin/ld: /usr/local/lib64/libopencv_imgcodecs.so.4.1.2: undefined reference to `png_set_filter@PNG15_0' collect2: error: ld returned 1 exit status make[2]: *** [src/CMakeFiles/recognize-tissue.dir/build.make:100: src/recognize-tissue] Error 1 make[1]: *** [CMakeFiles/Makefile2:169: src/CMakeFiles/recognize-tissue.dir/all] Error 2 I think it may cause by cross-compiled which I'm not so familiar with。 The only one thing that I'm sure is the openCV has install successfuly, because: (base) [Robin@SC-201910280935 python-module]$ python Python 3.7.4 (default, Aug 13 2019, 20:35:49) [GCC 7.3.0] :: Anaconda, Inc. on linux Type "help", "copyright", "credits" or "license" for more information. >>> import cv2 >>> cv2.__version__ '4.1.1' >>> This is the first time for me to use issues in Github,and I am extremely thristy to your reply to my message~Thx~ I find a solution:Change the the 'CV_BGR2GRAY' to 'COLOR_BGR2GRAY' in file st_tissue_recognition/src/util.cpp, cd into the /python-mudule/ then python setup.py build python setup.py install it works for me Thanks for reporting back the solution!
gharchive/issue
2019-11-14T09:24:03
2025-04-01T04:33:05.939183
{ "authors": [ "BioAIEvolu", "ludvb" ], "repo": "SpatialTranscriptomicsResearch/st_tissue_recognition", "url": "https://github.com/SpatialTranscriptomicsResearch/st_tissue_recognition/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2423920112
BED-4598: added group entities to open api spec Description Added endpoint definitions for group entity endpoints to the open api spec Motivation and Context This PR addresses: BED-4141 / BED-4598 Types of changes Chore (a change that does not modify the application functionality) Checklist: [x] I have met the contributing prerequisites Assigned myself to this PR Added the appropriate labels Associated an issue: https://github.com/SpecterOps/BloodHound/issues/672 Read the Contributing guide: https://github.com/SpecterOps/BloodHound/wiki/Contributing [x] I have ensured that related documentation is up-to-date Open API docs Code comments (GoDocs / JSDocs) [x] I have followed proper test practices Added/updated tests to cover my changes All new and existing tests passed I also added a README with some helpful information about paths, and removed a couple duplicate files
gharchive/pull-request
2024-07-22T23:17:54
2025-04-01T04:33:05.945519
{ "authors": [ "sircodemane" ], "repo": "SpecterOps/BloodHound", "url": "https://github.com/SpecterOps/BloodHound/pull/724", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
210292583
Poor performance with modsecurity enabled As suggested within issues https://github.com/SpiderLabs/ModSecurity/issues/1318 I'll open a new one here. With MS enabled with the current codebase the performance test with "ab" from apache2-utils is not going over 300req/s. With a virtual machine on Virtualbox with 4 cores and 4gb ram: root@nginx:~# ab -n 10000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 1000 requests Completed 2000 requests Completed 3000 requests Completed 4000 requests Completed 5000 requests Completed 6000 requests Completed 7000 requests Completed 8000 requests Completed 9000 requests Completed 10000 requests Finished 10000 requests Server Software: nginx/1.11.9 Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 612 bytes Concurrency Level: 20 Time taken for tests: 73.444 seconds Complete requests: 10000 Failed requests: 0 Keep-Alive requests: 9909 Total transferred: 8499545 bytes HTML transferred: 6120000 bytes Requests per second: 136.16 [#/sec] (mean) Time per request: 146.889 [ms] (mean) Time per request: 7.344 [ms] (mean, across all concurrent requests) Transfer rate: 113.02 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.0 0 0 Processing: 14 147 31.7 141 1308 Waiting: 14 147 31.7 141 1308 Total: 14 147 31.7 141 1308 Percentage of the requests served within a certain time (ms) 50% 141 66% 147 75% 151 80% 154 90% 166 95% 188 98% 207 99% 232 100% 1308 (longest request) And without: root@nginx:~# /opt/nginx/sbin/nginx root@nginx:~# ab -n 10000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 1000 requests Completed 2000 requests Completed 3000 requests Completed 4000 requests Completed 5000 requests Completed 6000 requests Completed 7000 requests Completed 8000 requests Completed 9000 requests Completed 10000 requests Finished 10000 requests Server Software: nginx/1.11.9 Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 612 bytes Concurrency Level: 20 Time taken for tests: 0.197 seconds Complete requests: 10000 Failed requests: 0 Keep-Alive requests: 9909 Total transferred: 8499545 bytes HTML transferred: 6120000 bytes Requests per second: 50755.75 [#/sec] (mean) Time per request: 0.394 [ms] (mean) Time per request: 0.020 [ms] (mean, across all concurrent requests) Transfer rate: 42128.99 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.0 0 0 Processing: 0 0 0.3 0 5 Waiting: 0 0 0.3 0 5 Total: 0 0 0.3 0 5 Percentage of the requests served within a certain time (ms) 50% 0 66% 0 75% 0 80% 0 90% 0 95% 1 98% 1 99% 2 100% 5 (longest request) I tested this also with Nginx+ but then the tool stucks a 90%, no mather how many requests (20 cores and 64GB ram): This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 100 requests Completed 200 requests Completed 300 requests Completed 400 requests Completed 500 requests Completed 600 requests Completed 700 requests Completed 800 requests Completed 900 requests apr_pollset_poll: The timeout specified has expired (70007) Total of 999 requests completed But with just 100 it's ok: Server Software: nginx Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 162 bytes Concurrency Level: 20 Time taken for tests: 0.243 seconds Complete requests: 100 Failed requests: 0 Non-2xx responses: 100 Keep-Alive requests: 100 Total transferred: 31000 bytes HTML transferred: 16200 bytes Requests per second: 410.71 [#/sec] (mean) Time per request: 48.696 [ms] (mean) Time per request: 2.435 [ms] (mean, across all concurrent requests) Transfer rate: 124.34 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.7 0 3 Processing: 3 42 34.6 58 99 Waiting: 3 42 34.6 58 99 Total: 3 42 34.6 60 99 Percentage of the requests served within a certain time (ms) 50% 60 66% 66 75% 70 80% 74 90% 86 95% 92 98% 97 99% 99 100% 99 (longest request) Perhaps others can post their results too. Ah, I also tried to disable logging but it doens't affect the performance at all. Ok, I disabled modsec_audit.log and now it seems better the more requests I send :) `root@waf-1-a-02:~# ab -n 100 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient).....done Server Software: nginx Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 162 bytes Concurrency Level: 20 Time taken for tests: 0.175 seconds Complete requests: 100 Failed requests: 0 Non-2xx responses: 100 Keep-Alive requests: 100 Total transferred: 31000 bytes HTML transferred: 16200 bytes Requests per second: 570.71 [#/sec] (mean) Time per request: 35.044 [ms] (mean) Time per request: 1.752 [ms] (mean, across all concurrent requests) Transfer rate: 172.77 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.7 0 3 Processing: 4 29 20.5 25 76 Waiting: 3 29 20.5 25 76 Total: 4 30 20.6 26 77 Percentage of the requests served within a certain time (ms) 50% 26 66% 44 75% 48 80% 51 90% 59 95% 65 98% 72 99% 77 100% 77 (longest request) root@waf-1-a-02:~# ab -n 1000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 100 requests Completed 200 requests Completed 300 requests Completed 400 requests Completed 500 requests Completed 600 requests Completed 700 requests Completed 800 requests Completed 900 requests Completed 1000 requests Finished 1000 requests Server Software: nginx Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 162 bytes Concurrency Level: 20 Time taken for tests: 1.458 seconds Complete requests: 1000 Failed requests: 0 Non-2xx responses: 1000 Keep-Alive requests: 996 Total transferred: 309980 bytes HTML transferred: 162000 bytes Requests per second: 685.68 [#/sec] (mean) Time per request: 29.168 [ms] (mean) Time per request: 1.458 [ms] (mean, across all concurrent requests) Transfer rate: 207.57 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.2 0 3 Processing: 2 29 18.6 31 84 Waiting: 2 29 18.6 31 84 Total: 2 29 18.6 33 85 Percentage of the requests served within a certain time (ms) 50% 33 66% 39 75% 41 80% 51 90% 54 95% 54 98% 62 99% 72 100% 85 (longest request) root@waf-1-a-02:~# ab -n 2000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 200 requests Completed 400 requests Completed 600 requests Completed 800 requests Completed 1000 requests Completed 1200 requests Completed 1400 requests Completed 1600 requests Completed 1800 requests Completed 2000 requests Finished 2000 requests Server Software: nginx Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 162 bytes Concurrency Level: 20 Time taken for tests: 2.007 seconds Complete requests: 2000 Failed requests: 0 Non-2xx responses: 2000 Keep-Alive requests: 1988 Total transferred: 619940 bytes HTML transferred: 324000 bytes Requests per second: 996.74 [#/sec] (mean) Time per request: 20.065 [ms] (mean) Time per request: 1.003 [ms] (mean, across all concurrent requests) Transfer rate: 301.72 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.2 0 3 Processing: 2 20 23.0 3 94 Waiting: 2 20 23.0 3 94 Total: 3 20 23.0 3 94 Percentage of the requests served within a certain time (ms) 50% 3 66% 43 75% 44 80% 44 90% 58 95% 60 98% 61 99% 70 100% 94 (longest request) root@waf-1-a-02:~# ab -n 3000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 300 requests Completed 600 requests Completed 900 requests Completed 1200 requests Completed 1500 requests Completed 1800 requests Completed 2100 requests Completed 2400 requests Completed 2700 requests Completed 3000 requests Finished 3000 requests Server Software: nginx Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 162 bytes Concurrency Level: 20 Time taken for tests: 2.702 seconds Complete requests: 3000 Failed requests: 0 Non-2xx responses: 3000 Keep-Alive requests: 2981 Total transferred: 929905 bytes HTML transferred: 486000 bytes Requests per second: 1110.42 [#/sec] (mean) Time per request: 18.011 [ms] (mean) Time per request: 0.901 [ms] (mean, across all concurrent requests) Transfer rate: 336.13 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.1 0 3 Processing: 2 18 20.5 3 90 Waiting: 2 18 20.5 3 90 Total: 2 18 20.5 3 91 Percentage of the requests served within a certain time (ms) 50% 3 66% 13 75% 42 80% 43 90% 44 95% 58 98% 59 99% 60 100% 91 (longest request) root@waf-1-a-02:~# ab -n 10000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 1000 requests Completed 2000 requests Completed 3000 requests Completed 4000 requests Completed 5000 requests Completed 6000 requests Completed 7000 requests Completed 8000 requests Completed 9000 requests Completed 10000 requests Finished 10000 requests Server Software: nginx Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 162 bytes Concurrency Level: 20 Time taken for tests: 5.262 seconds Complete requests: 10000 Failed requests: 0 Non-2xx responses: 10000 Keep-Alive requests: 9913 Total transferred: 3099565 bytes HTML transferred: 1620000 bytes Requests per second: 1900.24 [#/sec] (mean) Time per request: 10.525 [ms] (mean) Time per request: 0.526 [ms] (mean, across all concurrent requests) Transfer rate: 575.19 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.1 0 3 Processing: 2 11 16.2 3 91 Waiting: 2 10 16.2 3 91 Total: 2 11 16.2 3 91 Percentage of the requests served within a certain time (ms) 50% 3 66% 3 75% 5 80% 6 90% 43 95% 45 98% 57 99% 59 100% 91 (longest request)` Hi @mimugmail, @zimmerle, I have created a separate repository that can be used to easily spin up reproducible environments (thanks goes to Vagrant and SaltStack) suitable for running benchmarks against nginx with libmodsecurity connector module: https://github.com/defanator/modsecurity-performance It also can be extended to include more sophisticated things, if anyone would like to. For now I'm seeing more than 4x performance drop when libmodsecurity is turned on without any actual rules configured, and more than 170x performance drop when libmodsecurity is turned on with OWASP CRS v3.0.0 loaded (I'm not including any absolute numbers here for obvious reasons; you can find my own results in the repo - I'll add a bit more details about the hardware used as well). Main goal is to give an ability for anyone to easily reproduce the tests against particular combination of libmodsecurity, ModSecurity-nginx connector module, and nginx. @mimugmail - can I ask you to share your modsecurity.conf you were using while testing? (if it differs from the default one here - https://github.com/SpiderLabs/ModSecurity/blob/master/modsecurity.conf-recommended) `# -- Rule engine initialization ---------------------------------------------- Enable ModSecurity, attaching it to every transaction. Use detection only to start with, because that minimises the chances of post-installation disruption. SecRuleEngine On -- Request body handling --------------------------------------------------- Allow ModSecurity to access request bodies. If you don't, ModSecurity won't be able to see any POST parameters, which opens a large security hole for attackers to exploit. SecRequestBodyAccess On Enable XML request body parser. Initiate XML Processor in case of xml content-type SecRule REQUEST_HEADERS:Content-Type "(?:text|application)/xml" "id:'200000',phase:1,t:none,t:lowercase,pass,nolog,ctl:requestBodyProcessor=XML" Enable JSON request body parser. Initiate JSON Processor in case of JSON content-type; change accordingly if your application does not use 'application/json' SecRule REQUEST_HEADERS:Content-Type "application/json" "id:'200001',phase:1,t:none,t:lowercase,pass,nolog,ctl:requestBodyProcessor=JSON" Maximum request body size we will accept for buffering. If you support file uploads then the value given on the first line has to be as large as the largest file you are willing to accept. The second value refers to the size of data, with files excluded. You want to keep that value as low as practical. SecRequestBodyLimit 536870912 SecRequestBodyNoFilesLimit 536870912 Store up to 128 KB of request body data in memory. When the multipart parser reachers this limit, it will start using your hard disk for storage. That is slow, but unavoidable. SecRequestBodyInMemoryLimit 134217728 What do do if the request body size is above our configured limit. Keep in mind that this setting will automatically be set to ProcessPartial when SecRuleEngine is set to DetectionOnly mode in order to minimize disruptions when initially deploying ModSecurity. SecRequestBodyLimitAction ProcessPartial Verify that we've correctly processed the request body. As a rule of thumb, when failing to process a request body you should reject the request (when deployed in blocking mode) or log a high-severity alert (when deployed in detection-only mode). SecRule REQBODY_ERROR "!@eq 0" "id:'200002', phase:2,t:none,log,deny,status:400,msg:'Failed to parse request body.',logdata:'%{reqbody_error_msg}',severity:2" By default be strict with what we accept in the multipart/form-data request body. If the rule below proves to be too strict for your environment consider changing it to detection-only. You are encouraged not to remove it altogether. SecRule MULTIPART_STRICT_ERROR "!@eq 0" "id:'200003',phase:2,t:none,log,deny,status:400, msg:'Multipart request body failed strict validation: PE %{REQBODY_PROCESSOR_ERROR}, BQ %{MULTIPART_BOUNDARY_QUOTED}, BW %{MULTIPART_BOUNDARY_WHITESPACE}, DB %{MULTIPART_DATA_BEFORE}, DA %{MULTIPART_DATA_AFTER}, HF %{MULTIPART_HEADER_FOLDING}, LF %{MULTIPART_LF_LINE}, SM %{MULTIPART_MISSING_SEMICOLON}, IQ %{MULTIPART_INVALID_QUOTING}, IP %{MULTIPART_INVALID_PART}, IH %{MULTIPART_INVALID_HEADER_FOLDING}, FL %{MULTIPART_FILE_LIMIT_EXCEEDED}'" Did we see anything that might be a boundary? SecRule MULTIPART_UNMATCHED_BOUNDARY "!@eq 0" "id:'200004',phase:2,t:none,log,deny,msg:'Multipart parser detected a possible unmatched boundary.'" PCRE Tuning We want to avoid a potential RegEx DoS condition SecPcreMatchLimit 1000 SecPcreMatchLimitRecursion 1000 Some internal errors will set flags in TX and we will need to look for these. All of these are prefixed with "MSC_". The following flags currently exist: MSC_PCRE_LIMITS_EXCEEDED: PCRE match limits were exceeded. SecRule TX:/^MSC_/ "!@streq 0" "id:'200005',phase:2,t:none,deny,msg:'ModSecurity internal error flagged: %{MATCHED_VAR_NAME}'" -- Response body handling -------------------------------------------------- Allow ModSecurity to access response bodies. You should have this directive enabled in order to identify errors and data leakage issues. Do keep in mind that enabling this directive does increases both memory consumption and response latency. SecResponseBodyAccess On Which response MIME types do you want to inspect? You should adjust the configuration below to catch documents but avoid static files (e.g., images and archives). SecResponseBodyMimeType text/plain text/html text/xml Buffer response bodies of up to 512 KB in length. SecResponseBodyLimit 524288 What happens when we encounter a response body larger than the configured limit? By default, we process what we have and let the rest through. That's somewhat less secure, but does not break any legitimate pages. SecResponseBodyLimitAction ProcessPartial -- Filesystem configuration ------------------------------------------------ The location where ModSecurity stores temporary files (for example, when it needs to handle a file upload that is larger than the configured limit). This default setting is chosen due to all systems have /tmp available however, this is less than ideal. It is recommended that you specify a location that's private. SecTmpDir /tmp/ The location where ModSecurity will keep its persistent data. This default setting is chosen due to all systems have /tmp available however, it too should be updated to a place that other users can't access. SecDataDir /tmp/ -- File uploads handling configuration ------------------------------------- The location where ModSecurity stores intercepted uploaded files. This location must be private to ModSecurity. You don't want other users on the server to access the files, do you? #SecUploadDir /opt/modsecurity/var/upload/ By default, only keep the files that were determined to be unusual in some way (by an external inspection script). For this to work you will also need at least one file inspection rule. #SecUploadKeepFiles RelevantOnly Uploaded files are by default created with permissions that do not allow any other user to access them. You may need to relax that if you want to interface ModSecurity to an external program (e.g., an anti-virus). #SecUploadFileMode 0600 -- Debug log configuration ------------------------------------------------- The default debug log configuration is to duplicate the error, warning and notice messages from the error log. #SecDebugLog /var/log/modsec_debug.log #SecDebugLogLevel 4 -- Audit log configuration ------------------------------------------------- Log the transactions that are marked by a rule, as well as those that trigger a server error (determined by a 5xx or 4xx, excluding 404, level response status codes). #SecAuditEngine Off SecAuditEngine RelevantOnly #SecAuditLogRelevantStatus "^(?:5|4(?!04))" SecAuditLogRelevantStatus "403" Log everything we know about a transaction. #SecAuditLogParts ABIJDEFHZ SecAuditLogParts ABCDEFHIJKZ Use a single file for logging. This is much easier to look at, but assumes that you will use the audit log only ocassionally. SecAuditLogType Serial SecAuditLog /var/log/modsec_audit.log Specify the path for concurrent audit logging. #SecAuditLogStorageDir /opt/modsecurity/var/audit/ -- Miscellaneous ----------------------------------------------------------- Use the most commonly used application/x-www-form-urlencoded parameter separator. There's probably only one application somewhere that uses something else so don't expect to change this value. SecArgumentSeparator & Settle on version 0 (zero) cookies, as that is what most applications use. Using an incorrect cookie version may open your installation to evasion attacks (against the rules that examine named cookies). SecCookieFormat 0 Specify your Unicode Code Point. This mapping is used by the t:urlDecodeUni transformation function to properly map encoded data to your language. Properly setting these directives helps to reduce false positives and negatives. SecUnicodeMapFile unicode.mapping 20127 Improve the quality of ModSecurity by sharing information about your current ModSecurity version and dependencies versions. The following information will be shared: ModSecurity version, Web Server version, APR version, PCRE version, Lua version, Libxml2 version, Anonymous unique id for host. SecStatusEngine On` @mimugmail, @zimmerle, I've spent some time running simple benchmarks on multi-core VM, and found something interesting about SecAuditEngine: https://github.com/defanator/modsecurity-performance/wiki#2017-03-05 @mimugmail, I've just updated your comments to get proper formatting for preformatted text (commands output) - JFYI. @defanator Thanks :) Just did a copy+paste. This behavior fits perfectly with my measurements. Have you already tried concurrent logging? I changed to: SecAuditLogType Concurrent SecAuditLog /var/log/modsec_audit.log SecAuditLogStorageDir /opt/nginx/logs/audit But performance is still around 144req/s Hi @mimugmail and @defanator, Do you mind to test the branch v3/dev/speedup? I did some performance improvements on v3/dev/speedup. There are space for more, but it will be good to count in your tests to confirm that I am on the right track. Oh btw, please give it a try with the SecAuditLog disabled, going to look into the audit log generation tomorrow. @zimmerle, please check https://github.com/defanator/modsecurity-performance/wiki#2017-03-10 I ran a few cycles with SecAuditEngine RelevantOnly, and then switched to SecAuditEngine Off - numbers were nearly the same, in the margin of error. However, in the worst case (full OWASP CRS v3.0.0 is loaded) average RPS is more than 10% greater, and latency is more than 10% lower than in previous run (v3/master as of 20170307). SecAuditEngine RelevantOnly seems to affect only multi-worker configurations (probably some locking issues - I did not have time to test further). @zimmerle With SecAuditEngine On `root@nginx:/opt# ab -n 10000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 1000 requests Completed 2000 requests Completed 3000 requests Completed 4000 requests Completed 5000 requests Completed 6000 requests Completed 7000 requests Completed 8000 requests Completed 9000 requests Completed 10000 requests Finished 10000 requests Server Software: nginx/1.11.9 Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 612 bytes Concurrency Level: 20 Time taken for tests: 80.526 seconds Complete requests: 10000 Failed requests: 0 Keep-Alive requests: 9909 Total transferred: 8499545 bytes HTML transferred: 6120000 bytes Requests per second: 124.18 [#/sec] (mean) Time per request: 161.052 [ms] (mean) Time per request: 8.053 [ms] (mean, across all concurrent requests) Transfer rate: 103.08 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.0 0 0 Processing: 15 161 32.7 159 1517 Waiting: 15 161 32.7 159 1517 Total: 15 161 32.7 159 1517 Percentage of the requests served within a certain time (ms) 50% 159 66% 161 75% 163 80% 165 90% 169 95% 175 98% 184 99% 215 100% 1517 (longest request) root@nginx:/opt# ab -n 1000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 100 requests Completed 200 requests Completed 300 requests Completed 400 requests Completed 500 requests Completed 600 requests Completed 700 requests Completed 800 requests Completed 900 requests Completed 1000 requests Finished 1000 requests Server Software: nginx/1.11.9 Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 612 bytes Concurrency Level: 20 Time taken for tests: 8.606 seconds Complete requests: 1000 Failed requests: 0 Keep-Alive requests: 1000 Total transferred: 850000 bytes HTML transferred: 612000 bytes Requests per second: 116.19 [#/sec] (mean) Time per request: 172.127 [ms] (mean) Time per request: 8.606 [ms] (mean, across all concurrent requests) Transfer rate: 96.45 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.0 0 0 Processing: 7 171 18.3 169 243 Waiting: 7 171 18.3 169 243 Total: 8 171 18.2 169 243 Percentage of the requests served within a certain time (ms) 50% 169 66% 173 75% 176 80% 177 90% 183 95% 199 98% 211 99% 226 100% 243 (longest request)` WIth SecAuditEngine Off results are the same ... @mimugmail @defanator thanks! @mimugmail how many workers do you have? @zimmerle meeep you got me, now it's 100 (was 1) 👍 SecAuditEngine Off: `root@nginx:/opt# ab -n 10000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 1000 requests Completed 2000 requests Completed 3000 requests Completed 4000 requests Completed 5000 requests Completed 6000 requests Completed 7000 requests Completed 8000 requests Completed 9000 requests Completed 10000 requests Finished 10000 requests Server Software: nginx/1.11.9 Server Hostname: 127.0.0.1 Server Port: 80 Document Path: / Document Length: 612 bytes Concurrency Level: 20 Time taken for tests: 25.751 seconds Complete requests: 10000 Failed requests: 0 Keep-Alive requests: 9909 Total transferred: 8499545 bytes HTML transferred: 6120000 bytes Requests per second: 388.34 [#/sec] (mean) Time per request: 51.501 [ms] (mean) Time per request: 2.575 [ms] (mean, across all concurrent requests) Transfer rate: 322.34 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.8 0 28 Processing: 7 51 71.6 33 1140 Waiting: 0 50 71.5 32 1140 Total: 7 51 72.0 33 1151 Percentage of the requests served within a certain time (ms) 50% 33 66% 42 75% 49 80% 56 90% 92 95% 148 98% 235 99% 352 100% 1151 (longest request) `SecAuditEngine On: `root@nginx:/opt# ab -n 10000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 1000 requests Completed 2000 requests Completed 3000 requests Completed 4000 requests Completed 5000 requests Completed 6000 requests Completed 7000 requests Completed 8000 requests Completed 9000 requests apr_pollset_poll: The timeout specified has expired (70007) Total of 9981 requests completed` Nginx 1.11.9, 4GB RAM, 4 CPUs Also with just 1000: `root@nginx:/opt# ab -n 1000 -c 20 -k http://127.0.0.1:80/ This is ApacheBench, Version 2.3 <$Revision: 1604373 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking 127.0.0.1 (be patient) Completed 100 requests Completed 200 requests Completed 300 requests Completed 400 requests Completed 500 requests Completed 600 requests Completed 700 requests Completed 800 requests Completed 900 requests apr_pollset_poll: The timeout specified has expired (70007) Total of 981 requests completed` @mimugmail Great! Thanks ;) @mimugmail, running 100 workers on 4-core machine serving CPU-heavy load (like ModSecurity) won't lead to any benefits. I don't see much sense in setting worker_processes to a number greater than total number of available cores (for this particular pattern). Also, your latest results do not include any summary for SecAuditEngine On case due to timeouts detected by ab tool. @defanator Yep, I set it to 4, no difference. Funny thing is that between 1000 and 10000 checks it always stucks at the last 105. With just 100 I get: Requests per second: 206.24 [#/sec] (mean) the branch v3/dev/speedup was merged into v3/master. 2G KVM VPS 1vcore ab -n 10000 -c 20 -k http://theo.in/ This is ApacheBench, Version 2.3 <$Revision: 1706008 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking theo.in (be patient) Completed 1000 requests Completed 2000 requests Completed 3000 requests Completed 4000 requests Completed 5000 requests Completed 6000 requests Completed 7000 requests Completed 8000 requests Completed 9000 requests Completed 10000 requests Finished 10000 requests Server Software: XtendWeb-nginx Server Hostname: theo.in Server Port: 80 Document Path: / Document Length: 623 bytes Concurrency Level: 20 Time taken for tests: 156.437 seconds Complete requests: 10000 Failed requests: 0 Keep-Alive requests: 9908 Total transferred: 8149540 bytes HTML transferred: 6230000 bytes Requests per second: 63.92 [#/sec] (mean) Time per request: 312.873 [ms] (mean) Time per request: 15.644 [ms] (mean, across all concurrent requests) Transfer rate: 50.87 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 3 28.2 0 465 Processing: 190 310 74.8 297 688 Waiting: 190 310 74.8 297 688 Total: 190 313 80.7 298 959 Percentage of the requests served within a certain time (ms) 50% 298 66% 327 75% 346 80% 359 90% 401 95% 468 98% 556 99% 599 100% 959 (longest request) # nginx -V nginx version: nginx/1.11.12 built by gcc 4.8.5 20150623 (Red Hat 4.8.5-11) (GCC) built with LibreSSL 2.5.1 TLS SNI support enabled configure arguments: --prefix=/etc/nginx --sbin-path=/usr/sbin/nginx --modules-path=/etc/nginx/modules --with-pcre=./pcre-8.40 --with-pcre-jit --with-zlib=./zlib-1.2.11 --with-openssl=./libressl-2.5.1 --conf-path=/etc/nginx/nginx.conf --error-log-path=/var/log/nginx/error_log --http-log-path=/var/log/nginx/access_log --pid-path=/var/run/nginx.pid --lock-path=/var/run/nginx.lock --http-client-body-temp-path=/var/cache/nginx/client_temp --http-proxy-temp-path=/var/cache/nginx/proxy_temp --http-fastcgi-temp-path=/var/cache/nginx/fastcgi_temp --http-uwsgi-temp-path=/var/cache/nginx/uwsgi_temp --http-scgi-temp-path=/var/cache/nginx/scgi_temp --user=nobody --group=nobody --with-http_ssl_module --with-http_realip_module --with-http_addition_module --with-http_sub_module --with-http_dav_module --with-http_flv_module --with-http_mp4_module --with-http_gunzip_module --with-http_gzip_static_module --with-http_random_index_module --with-http_secure_link_module --with-http_stub_status_module --with-http_auth_request_module --add-dynamic-module=naxsi-http2/naxsi_src --with-file-aio --with-threads --with-stream --with-stream_ssl_module --with-http_slice_module --with-compat --with-http_v2_module --with-http_geoip_module=dynamic --add-dynamic-module=ngx_pagespeed-release-1.11.33.4-beta --add-dynamic-module=/usr/local/rvm/gems/ruby-2.3.1/gems/passenger-5.1.2/src/nginx_module --add-dynamic-module=ngx_brotli --add-dynamic-module=echo-nginx-module-0.60 --add-dynamic-module=headers-more-nginx-module-0.32 --add-dynamic-module=ngx_http_redis-0.3.8 --add-dynamic-module=redis2-nginx-module --add-dynamic-module=srcache-nginx-module-0.31 --add-dynamic-module=ngx_devel_kit-0.3.0 --add-dynamic-module=set-misc-nginx-module-0.31 --add-dynamic-module=ModSecurity-nginx --with-cc-opt='-O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector-strong --param=ssp-buffer-size=4 -grecord-gcc-switches -m64 -mtune=generic' --with-ld-opt=-Wl,-E nginx_conf.txt A couple of benchmarks with recent libmodsecurity sources: https://github.com/defanator/modsecurity-performance/wiki#2017-04-11-1 (single worker), https://github.com/defanator/modsecurity-performance/wiki#2017-04-11-2 (multi worker). Noticeable improvement is that multi-worker setup does not lock all the processing inside single worker (as it was before [1]), so I was able to reach almost 1k RPS with the full OWASP CRS v3.0.0 loaded, with 6 workers. [1] https://github.com/defanator/modsecurity-performance/wiki#2017-03-05 I am closing this issue in favor of @defanator wiki. I would encourage you guys to keep the work monitoring the performance. whenever you find a problem with it, please re-open this issue. @defanator did you get a chance to re-run the benchmark again? @luengnat I run it periodically, just updated the results page with most recent build here: https://github.com/defanator/modsecurity-performance/wiki#2017-12-11
gharchive/issue
2017-02-26T06:51:17
2025-04-01T04:33:06.083986
{ "authors": [ "AnoopAlias", "defanator", "luengnat", "mimugmail", "zimmerle" ], "repo": "SpiderLabs/ModSecurity-nginx", "url": "https://github.com/SpiderLabs/ModSecurity-nginx/issues/35", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
21181969
Adding parameters to transformation functions MODSEC-210: It would be very useful to have the possibility to send (string) parameters to transformation functions, like t:replace(a,b) The syntax is only an example, it could be completely different. As we have a parser on v3, we could easily have that implemented per-transformation. Said that, we no longer need to have such issue.
gharchive/issue
2013-10-17T20:18:39
2025-04-01T04:33:06.087523
{ "authors": [ "rcbarnett-zz", "zimmerle" ], "repo": "SpiderLabs/ModSecurity", "url": "https://github.com/SpiderLabs/ModSecurity/issues/364", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
604422252
what version of discord.js is this? Also, you dont mention anywhere in the readme that you need to npm install ytdl and yt-search and all the other things V12 Discord.js and also it's common sense to download the npms unless you just wanna fork without any knowledge or experience of a Discord bot. @Ravenwulf running npm install installs every packages needed for the bot since package.json contains all the needed information. It'll also install the correct version of said package. Simply rum npm install as stated in the README file.
gharchive/issue
2020-04-22T03:19:07
2025-04-01T04:33:06.114757
{ "authors": [ "ItsAuric", "Ravenwulf", "Spimy" ], "repo": "Spimy/muse", "url": "https://github.com/Spimy/muse/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
624229604
Synchronize Gradle wrappers among builds Now that we have custom pull scripts (see #534), we can sync the Gradle wrapper configs and JARs automatically upon pull instead of manually updating ./gradle/wrapper, ./tools/smoke-tests/gradle/wrapper, and ./base-validating-builders/gradle/wrapper. This issue relates to #578. No longer relevant to this repo and is addressed in mc-java.
gharchive/issue
2020-05-25T11:09:00
2025-04-01T04:33:06.134791
{ "authors": [ "alexander-yevsyukov", "dmdashenkov" ], "repo": "SpineEventEngine/base", "url": "https://github.com/SpineEventEngine/base/issues/538", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
489797376
Publish 1.0.8-SNAPSHOT This PR advances the version of the library to 1.0.8-SNAPSHOT. @alexander-yevsyukov, PTAL.
gharchive/pull-request
2019-09-05T14:26:09
2025-04-01T04:33:06.135787
{ "authors": [ "dmdashenkov" ], "repo": "SpineEventEngine/time", "url": "https://github.com/SpineEventEngine/time/pull/71", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
270055264
[beta] Transpiled error class is not working as expected. Catching errors is broken with the message: TypeError: _this4.addToMessage is not a function Even though babel is trying to create the error classes here: https://github.com/SpireTeam/whitelister/blob/feature/bundle/dist/whitelister.js#L206 See codepen: https://codepen.io/anon/pen/eeNMWY?editors=1111#anon-login Turns out this had to do with extending a global when using babel, specifically Error. Fixed in 0.1.0-beta2 using the babel plugin referenced on MDN here.
gharchive/issue
2017-10-31T17:48:52
2025-04-01T04:33:06.138378
{ "authors": [ "davidpaulhunt" ], "repo": "SpireTeam/whitelister", "url": "https://github.com/SpireTeam/whitelister/issues/10", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1780996587
Added top contributors in ABOUTUS page Fixes Issue Closes #1656 Changes proposed Screenshots Checklist [x ] You've linked this PR to the correct issue. [x ] You have checked that the code is working correctly. [ ] You ⭐️ the repository! Note to reviewers @Spyware007 do add level in this PR ,it will be kind of you : |)
gharchive/pull-request
2023-06-29T14:52:40
2025-04-01T04:33:06.176093
{ "authors": [ "ManishaSwain8" ], "repo": "Spyware007/Animating-Buttons", "url": "https://github.com/Spyware007/Animating-Buttons/pull/1674", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1954191412
Incorporar métodos de eliminación a EntityPersist Incorporar métodos de eliminación a EntityPersist. se incorporo primer metodo de eliminación DeleteIds
gharchive/issue
2023-10-20T12:29:24
2025-04-01T04:33:06.176914
{ "authors": [ "ivancas84" ], "repo": "SqlOrganize/SqlOrganizeCs", "url": "https://github.com/SqlOrganize/SqlOrganizeCs/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1659431107
Testing, please ignore. (qdocvn) Bug bounty test - please ignore.... Please DO NOT APPROVE THIS! (qdocvn) Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.alveranuno seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
gharchive/pull-request
2023-04-08T07:30:54
2025-04-01T04:33:06.181454
{ "authors": [ "CLAassistant", "alveranuno" ], "repo": "Squarespace/cldr-engine-java", "url": "https://github.com/Squarespace/cldr-engine-java/pull/2", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
70929627
Disconnect cables with wrench Add the ability to disconnect multipart cables with buildwraft wrenches. We could store this in another flag and merge the two when checking canBeVisited and canVisitTo. I like the idea, but then the cable has to be a multipart cable to use this feature, which can be frustrating when you want to use it on a non-multipart cable. I know, we could add a custom handler to the activate method to convert it to a multipart block - but this seems less than ideal. The alternative would be an 'advanced cable' which has more functionality but I feel we would have to make it really worthwhile for it to exist. That's an interesting idea. I dunno though... An alternative would be to dye cables instead - so different coloured cables cannot connect to each other. The most sensible thing to do seems to be to add a IColoredNode or something which controls if a connection should occur. That can probably wait until 0.2.1 though. Yea I'd say we hold onto this one until 0.2.1. I've thought about this a lot and I'm not sure I see the point in this: I can't see any benefit of being able to separate networks. It would be a really cool feature to add but I don't want to add a load of unneeded complexity. I might open if someone can think of an application of this: separating cables without FMP installed would be good.
gharchive/issue
2015-04-25T15:46:22
2025-04-01T04:33:06.190657
{ "authors": [ "ElvishJerricco", "SquidDev" ], "repo": "SquidDev-CC/CC-Tweaks", "url": "https://github.com/SquidDev-CC/CC-Tweaks/issues/19", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1126739843
Add Delete to context menu I hadn't any issues with the dirty indicator for many months. I think it's already safe to have Delete in the Context Menu. If the repo is dirty, show ~ "The repository has local changes. Do you want to move it to the trash?" VS Code has this windows when deleting a file, we can have a similar text / look and feel: It's possible to use the trash pkg. Added in 1.6.0. Note however that we send it to trash as there is no reliable package for it in npm. Maybe copy VSCode code for it?
gharchive/issue
2022-02-08T03:13:04
2025-04-01T04:33:06.193002
{ "authors": [ "SrBrahma" ], "repo": "SrBrahma/GitHub-Repository-Manager", "url": "https://github.com/SrBrahma/GitHub-Repository-Manager/issues/34", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
257754074
can't use that module with angular-cli optimized builds Hi I am using angular/cli=1.4.2, typescript=2.5.2 and latest version of Angular2-Toaster I am trying to compile my app like ng build -e=prod -aot --build-optimizer, but I have these errors: my component problematic line in toaster.js It looks like easy issues tp fix, but I just wodering why BodyOutputType.Default expression fails I think I just need to override configs, but maybe you want to fix that issue This is very curious...the BodyOutputType IS defaulted intentionally. I suspect this is due to the optimize flag attempting dead code elimination and it may believe the property isn't needed. I will try to reproduce. This is currently fixed in 5.0.0-beta.1 and will be released in the full 5.0.0 release. Released in 5.0.0.
gharchive/issue
2017-09-14T15:07:02
2025-04-01T04:33:06.206824
{ "authors": [ "Stabzs", "ToGoBananas" ], "repo": "Stabzs/Angular2-Toaster", "url": "https://github.com/Stabzs/Angular2-Toaster/issues/128", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
234941034
Add PUT method to action executions API Allow action executions to be updated over API. The API will only allow status and result to be updated for incomplete action executions. This patch is required for 1) mistral->st2 callback on workflow completion and 2) st2 response to update st2.ask action execution. LGTM.
gharchive/pull-request
2017-06-09T21:17:16
2025-04-01T04:33:06.213285
{ "authors": [ "Kami", "m4dcoder" ], "repo": "StackStorm/st2", "url": "https://github.com/StackStorm/st2/pull/3461", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
616718281
Do not lookup release from package cloud if fork in circle config.yml if package cloud token is not set use hard coded value of 1 fixes #752 PR from forks are not able to be run through unit testing. This is because the package cloud token is not set. Here we are not going to lookup the release number in package cloud if it is a fork. This was stolen from https://github.com/StackStorm/st2-packages/blob/29a493e33de928e6df452a9102e80aaeadc5ec53/.circle/buildenv_st2.sh#L45-L51 Looks like it passed.
gharchive/pull-request
2020-05-12T14:48:55
2025-04-01T04:33:06.217044
{ "authors": [ "guzzijones" ], "repo": "StackStorm/st2web", "url": "https://github.com/StackStorm/st2web/pull/753", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2756434625
node22.x not working There was a typo in the previous release. There is the node18.x twice and no node22.x Please fix support for node22.x Thanks! Please upgrade to 0.15.1 as it contains a fix.
gharchive/issue
2024-12-23T17:04:56
2025-04-01T04:33:06.218151
{ "authors": [ "buggy", "xpert4" ], "repo": "StackToolbox/aws-sam-webpack-plugin", "url": "https://github.com/StackToolbox/aws-sam-webpack-plugin/issues/101", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
182804424
Bad require in stack bin file. In stack/bin/stack file the require in the following line fails: https://github.com/Stackvana/stack/blob/master/bin/stack#L39 I think it should be var viewPresenter = #require('../lib/viewPresenter'); Ahh! Sorry about that, will patch now and publish.
gharchive/issue
2016-10-13T14:24:48
2025-04-01T04:33:06.219583
{ "authors": [ "Marak", "labajo" ], "repo": "Stackvana/stack", "url": "https://github.com/Stackvana/stack/issues/10", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
361095748
LND won't start after updating to v0.5-beta I've tried updating to v0.5-beta following the guide but LND fails to start: Sep 15 11:24:49 pideypideypideypi systemd[1]: Started LND Lightning Daemon. Sep 15 11:24:49 pideypideypideypi systemd[1]: lnd.service: Main process exited, code=exited, status=1/FAILURE Sep 15 11:24:49 pideypideypideypi systemd[1]: lnd.service: Unit entered failed s tate. Sep 15 11:24:49 pideypideypideypi systemd[1]: lnd.service: Failed with result 'e xit-code'. No problems after reverting to v0.4.2-beta. I'm not sure how to pull more detailed logs. I have the same issue, fwiw. Although I'm on "0.5 release candidate" (can't update to "proper" 0.5 due to the missing link in the LND repository, as is also described here: https://github.com/Stadicus/guides/issues/210). Also, the "logs" folder under /home/bitcoin/.lnd is showing as empty (0 bytes), so apparently no log is being written after I try to restart the LND client and it fails repeatedly. Please try to run lnd manually with the bitcoin user to see the terminal output. Can you start lnd from the root account, just by typing lnd? Seems to be possible in general, although it gives out loadConfig: either bitcoin.active or litecoin.active must be set to 1 (true). It's probably not relevant to the issue in the OP though (except if they followed the same older version of the guide, and it's somehow connected to the issue), so hopefully this doesn't derail the thread too much :) That might be the issue. Do you have the option bitcoin.active=1 set in your lnd.config? My lnd.conf (no reference, might have some stupid shit in it) # RaspiBolt LND Mainnet: lnd configuration # /home/bitcoin/.lnd/lnd.conf [Application Options] debuglevel=debug maxpendingchannels=5 alias=Stadicus RaspiBolt color=#68F442 restlisten=localhost:8080 rpclisten=localhost:10009 [Bitcoin] bitcoin.active=1 bitcoin.mainnet=1 bitcoin.node=bitcoind [autopilot] autopilot.active=0 I think I found a fix, when running 0.5 rc2. In the release notes, changes are needed the .conf for lnd and bitcoind. After I made these changes and restarted both, lnd would start. Users running a bitcoind backend will need to modify their existing ZMQ-related settings. Previously, lnd listened on one shared port for both blocks and transactions. 0.5 requires two distinct ports to be set, such that they are isolated. Users should remove the old bitcoind.zmqpath entry in lnd.conf, and, for example, replace it with: bitcoind.zmqpubrawblock=tcp://127.0.0.1:28332 bitcoind.zmqpubrawtx=tcp://127.0.0.1:28333 Also ensure that bitcoind is exposing distinct ZMQ ports in bitcoin.conf: zmqpubrawblock=tcp://127.0.0.1:28332 zmqpubrawtx=tcp://127.0.0.1:28333 Was there a breaking change that I did not cover in the FAQ update section? Nope, I upgraded using the updated FAQ and everything went fine.
gharchive/issue
2018-09-18T00:52:39
2025-04-01T04:33:06.226893
{ "authors": [ "Stadicus", "pardus79", "raindogdance", "whatthemoses" ], "repo": "Stadicus/guides", "url": "https://github.com/Stadicus/guides/issues/216", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }