id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1534532951 | Update/fix XML docs for generated types
Rationale
The XML documentation of the generated types - at least the types which are directly used by developers - should match the documentation which is available on learn.microsoft.com and in the IDE when the "native" types are used.
I haven't checked all types, but I've found differences in:
CallerArgumentExpressionAttribute
class summary
class remarks
ctor paramter
property summary
Index
class summary
class remarks
various small differences in members
I guess this is somewhat linked to #88.
For my use, I don't really care if the docs match the types. I have XML doc generation enabled in my libs, and I get warnings on any public code that doesn't have XML comments. I'd be happy with anything, even empty, but I agree that it should be whatever MS has.
would adding #pragma warning disable to all the generated files suffice?
That would solve my problem, sure.
| gharchive/issue | 2023-01-16T08:54:29 | 2025-04-01T04:33:04.517672 | {
"authors": [
"0xced",
"SimonCropp",
"cremor",
"gregsdennis"
],
"repo": "Sergio0694/PolySharp",
"url": "https://github.com/Sergio0694/PolySharp/issues/44",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
255822713 | NO BMI Source in Github
There is no source in github,
samples/bmi
https://github.com/ServiceComb/ServiceComb-Java-Chassis/tree/master/samples/bmi
The BMI sample currently is still in PR. You can get the code with the following commands:
git clone https://github.com/ServiceComb/ServiceComb-Java-Chassis
cd ServiceComb-Java-Chassis
git fetch origin pull/129/head:sample
git checkout sample
@zhuhoudong The url should work now. Thanks for your feedback.
| gharchive/issue | 2017-09-07T05:51:24 | 2025-04-01T04:33:04.545494 | {
"authors": [
"eric-lee-ltk",
"zhuhoudong"
],
"repo": "ServiceComb/ServiceComb-Java-Chassis",
"url": "https://github.com/ServiceComb/ServiceComb-Java-Chassis/issues/137",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
615404750 | Added 5 unit cases for delete method.
Used pytest framework for unit testing. Created dataset to test using pytest fixture.
Tested valid and invalid scenarios too.
Check this issue
https://github.com/SforAiDl/decepticonlp/issues/22
Acknowledged.
Discussion thread continued in #22
| gharchive/pull-request | 2020-05-10T14:36:32 | 2025-04-01T04:33:04.565567 | {
"authors": [
"rajaswa",
"rohts-patil",
"someshsingh22"
],
"repo": "SforAiDl/decepticonlp",
"url": "https://github.com/SforAiDl/decepticonlp/pull/23",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2461716182 | [BUG]: Starting a new game with 1 duplicant bugs it out
Which Mod?
Duplicant Stat Selector
Description of the bug/crash
Upon embarking on a new game with 1 dupe, the duplicant spawns but the game is seemingly stuck (not frozen). Pressing the ESC key 3+ times is required to unstuck it. The regular "Alert" popup (check screenshot below) does NOT appear at the beginning.
Note that I don't get the bug with 2 or 4+ dupes.
Steps To Reproduce
New Game > Any Game Mode > Any Asteroid Style > Start Game
Click on "Remove Slot" on 2 duplicants to only have 1 remaining
Embark
This is 100% reproductible in my end, the bug always happens.
Player.log
Player.log
fixed with release 7.3.1
| gharchive/issue | 2024-08-12T19:15:39 | 2025-04-01T04:33:04.570329 | {
"authors": [
"Sad-theFaceless",
"Sgt-Imalas"
],
"repo": "Sgt-Imalas/Sgt_Imalas-Oni-Mods",
"url": "https://github.com/Sgt-Imalas/Sgt_Imalas-Oni-Mods/issues/126",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1352898022 | Reopening Again Issue With The Timebar Size.
Describe your problem here.
I Just Want To Make The Timebar To Look Like This With The Source Code.
Can You Help Me?
Are you modding a build from source or with Lua?
Source
What is your build target?
Neko, HashLink, or other build system
Did you edit anything in this build? If so, mention or summarize your changes.
For The Better Style.
theres no need to reopen the issue
you can just press the 3 dots and press edit
for the fix
go to assets/shared/images/timebar.png
and replace it
theres no need to duplicate the issue you can just press the 3 dots and press edit for the fix go to assets/shared/images/timebar.png and replace it
I Just Trying To Find To All Mods, But It's Old Classic TimeBar
I Just Trying To Find To All Mods, But It's Old Classic TimeBar
What do you mean
What do you mean
Mods Only Have This Old Time Bar:
you can just replace it because it's a sprite
Make the sprite longer
| gharchive/issue | 2022-08-27T01:32:21 | 2025-04-01T04:33:04.590663 | {
"authors": [
"Mastercopper2346",
"StefanBETA2008",
"TheKitBoi"
],
"repo": "ShadowMario/FNF-PsychEngine",
"url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/10326",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1376888992 | how do you add images to the menu
Describe your problem here.
i want to add a image to the main menu but i dont know how to, is there like a code or something?
Are you modding a build from source or with Lua?
Source
What is your build target?
Windows x64
Did you edit anything in this build? If so, mention or summarize your changes.
MainMenuState.hx
make a var
then
asses = new FlxSprite().loadGraphic(Paths.images('image'));
asses.x += 0;
asses.y += 0;
add(asses);
make a variable for your sprite:
var sprite:FlxSprite;
sprite = new FlxSprite(X, Y).loadGraphic(Paths.image('yourimage'));
add(sprite);
change X and Y into the coordinates you want your sprite to be
change yourimage to the sprite's graphic
change the sprite variable to be called whatever you want
For easier movement
var sprite:FlxSprite;
sprite = new FlxSprite(0, 0).loadGraphic(Paths.image('yourimage'));
sprite.x += X; //for negative coordinates use -= instead
sprite.y += Y;
add(sprite);
change X into the x position you want your sprite to go
same for the Y
thanks it works :D
yo why does it go off the screen once I scrolled to options
yo why does it go off the screen once I scrolled to options
Camera movement
Just do sprite.scrollFactor.set();, changing sprite to the variable that is your sprite
thanks bro
Help please, I'm using psych engine 0.7.3 and getting this error when compiling:
source/states/MainMenuState.hx:232: characters 3-9 : Unexpected sprite
Help please, I'm using psych engine 0.7.3 and getting this error when compiling: source/states/MainMenuState.hx:232: characters 3-9 : Unexpected sprite
change sprite to your sprite variable
Like if you have a var overlay:FlxSprite, for example, and wanted to use that, change it to overlay.scrollFactor.set();
Помогите пожалуйста, я использую Psych Engine 0.7.3 и получаю эту ошибку при компиляции: source/states/MainMenuState.hx:232: символы 3-9 : Неожиданный спрайт
Измените spriteпеременную спрайта. Например, если у вас есть var overlay:FlxSprite, и вы хотите использовать его, измените его наoverlay.scrollFactor.set();
It looks like I'm too stupid for this, please write a code example
var sprite:FlxSprite;
sprite = new FlxSprite(0, 0).loadGraphic(Paths.image('yourimage'));
sprite.x += X; //for negative coordinates use -= instead
sprite.y += Y;
sprite.scrollFactor.set();
add(sprite);
| gharchive/issue | 2022-09-17T22:40:10 | 2025-04-01T04:33:04.600287 | {
"authors": [
"BernardoGP4504",
"DisIsNOt",
"Velikiycho",
"raxyd",
"uhhhhuhhhhhhokay"
],
"repo": "ShadowMario/FNF-PsychEngine",
"url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/10571",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2599685214 | Weekend 1 freeze after cutscene
Describe your bug here.
After loading up Weekend 1 from the Story Mode menu and watching the cutscene, the application stops responding.
Command Prompt/Terminal logs (if existing)
No response
Are you modding a build from source or with Lua?
Lua
What is your build target?
Windows
Did you edit anything in this build? If so, mention or summarize your changes.
No.
To bypass this glitch use the 32-bit version. This version doesn't support videos, so it skips the cutscenes and thus the week is playable.
| gharchive/issue | 2024-10-19T22:49:05 | 2025-04-01T04:33:04.602570 | {
"authors": [
"BluCoolKnux",
"DacroyleYT"
],
"repo": "ShadowMario/FNF-PsychEngine",
"url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/15763",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1070538013 | Chart's Editor Playtest but outside Chart Editor
What feature do you want to get added on the base engine?
(I'll try my best to explain this i don't speak english very well lol)
Basically just like the Chart Editor has as playtest function to test the chart from an exact position, but what i suggest is adding that same feature but in base game.
This could be useful for viewing events or functionality of custom notetypes, because this can't be viewed in the Chart Editor's playtest.
For example, let's say your song lasts 5 minutes, and at the end you have some sort of event, Character Change, Custom Event, Arrows Move, etc. Or you have a custom notetype you want to check if it works, but you'll have to wait 5 minutes.
This could be solved by adding support for playtesting in-game, i think it would be a really nice time-saving feature!
the annoying thing about the playtest feature is that it puts you in a seperate state from playstate where the only thing you can see is the notes/
just put the event at the start of the song
I have a different plan for that.
| gharchive/issue | 2021-12-03T12:23:43 | 2025-04-01T04:33:04.605222 | {
"authors": [
"4wdthinks",
"AetherLynx",
"MayoALT",
"ShadowMario"
],
"repo": "ShadowMario/FNF-PsychEngine",
"url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/1633",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1159148140 | Character Bug
Describe your bug here.
Hello, I am here to tell you about this bug that happens to me very often and it is very annoying when I finish creating my character and everything is fine with the png, the XML and the sprites/frames but when I finish configuring the character and I put it on the chart editor and when I give it to play for a moment it loads but then it closes
Command Prompt/Terminal logs (if existing)
Does not exist
Are you modding a build from source or with Lua?
Source
What is your build target?
Windows
Did you edit anything in this build? If so, mention or summarize your changes.
it's new
Never had any issues with any character so do you mind posting the files here so I can check them out? I doubt this is a bug but I'll test it out anyway
maybe you just coded it wrong
Never had any issues with any character so do you mind posting the files here so I can check them out? I doubt this is a bug but I'll test it out anyway
https://drive.google.com/file/d/1KpFHKmOAeapDpcmiWWYIwjq_L_FJV2ce/view (png)
https://drive.google.com/file/d/1ZnAIv751Ux3fPU7ST3iBJ1VOYoyZZ5dO/view (XML)
and the json works correctly
Send the json
| gharchive/issue | 2022-03-04T01:38:09 | 2025-04-01T04:33:04.609212 | {
"authors": [
"frantastic24",
"thedeveloper112206",
"unnamedmask3"
],
"repo": "ShadowMario/FNF-PsychEngine",
"url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/6203",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1292428928 | Idk with the source code
Describe your problem here.
when I complie the build with the source code u know
the build it does not appear in the export folder, only the mods, assets, plugins and manifest folders are exported, not the program or game, this did not happen to me with other versions such as 0.5 or 0.52
I don;t know if it is a problem or something but I don't know I'm new in coding yeah
Are you modding a build from source or with Lua?
Source
What is your build target?
Windows x64
Did you edit anything in this build? If so, mention or summarize your changes.
no, i just export all with the source code, I didn't touch anything
lime build windows -64 -clean
make sure that it does not errors in the console nor stops
if it does, please screenshot it and show the error here
I tried 3 times and it didn't work
Duplicate.
Read the readme file.
| gharchive/issue | 2022-07-03T21:52:06 | 2025-04-01T04:33:04.612647 | {
"authors": [
"Raltyro",
"SG-ProBlue",
"ShadowMario"
],
"repo": "ShadowMario/FNF-PsychEngine",
"url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/9481",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1993089330 | Rename EditorPlayState to EditorPlaySubState
I think after updates EditorPlayState become a substate so yeah no idk lmao this might be a dumb pr but this says everything
just do "EditorGameplay.hx"
just do "EditorGameplay.hx"
PlayStateEditorSubstate
Renames it to EditorPlaySubState
Doesn't put it in the substates folder
This doesn't follow the naming scheme pretty much.
Renames it to EditorPlaySubState Doesn't put it in the substates folder
options are substates
they're not in substates folder
CustomFadeTransition is a substate
it's not in the substates folder...
This doesn't follow the naming scheme pretty much.
think Play"State" and EditorPlay"SubState"
Its called like that because it's a copy of PlayState, not because it is/isn't a state
| gharchive/pull-request | 2023-11-14T16:20:59 | 2025-04-01T04:33:04.616256 | {
"authors": [
"CaptainBaldi",
"EyeDaleHim",
"KarimAkra",
"ShadowMario",
"crowplexus",
"mcagabe19"
],
"repo": "ShadowMario/FNF-PsychEngine",
"url": "https://github.com/ShadowMario/FNF-PsychEngine/pull/13610",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1516253569 | Failing does not trigger a restart
Triggering a level fail will not restart the current level. The menu is still accessible so it is still possible to exit the game by normal means.
Losing should prematurely win the level (IIRC this is what original ZB does)
Implemented in the private asset repository
| gharchive/issue | 2023-01-02T10:40:01 | 2025-04-01T04:33:04.632701 | {
"authors": [
"ShamblesSM"
],
"repo": "ShamblesSM/OpenSMCE",
"url": "https://github.com/ShamblesSM/OpenSMCE/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
869910426 | Cruise Control Malfunction , LCA Malfunction when attempting Engage
I installed the fork and rebooted. EON loads Stock Addition and I get a different GUI but when trying to enable/engage self driving i get a crouse control Malfunction error on my 2020 corolla
Do you get this on stock openpilot? I don't see any errors about this, check the /data/community/crashes dir for a latest file. If there's nothing about a malfunction in the file, check some of the other log files.
Thanks. I didn't get this error on stock openpilot before I installed the fork. I tried reinstalling the fork and same error happens (I see the new GUI but ACC doesn't engage). I see "vehicle online" on EON but it says "connect error" in red or "connect online" in yellow. I installed DP and similar error happens. I will check logs next
Try reflashing panda as well if you get this on other forks as well.
Is this fixed now on my fork?
| gharchive/issue | 2021-04-28T12:48:29 | 2025-04-01T04:33:04.634774 | {
"authors": [
"ShaneSmiskol",
"hwhorf"
],
"repo": "ShaneSmiskol/openpilot",
"url": "https://github.com/ShaneSmiskol/openpilot/issues/402",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1100822355 | Update Pages “index”
Automatically generated by Netlify CMS
👷 Deploy Preview for gracious-knuth-20d29a processing.
🔨 Explore the source changes: f09593915a01f63772a8ed0e1ccdffd35631f4a8
🔍 Inspect the deploy log: https://app.netlify.com/sites/gracious-knuth-20d29a/deploys/61df4ce0d84d260007dff9c5
| gharchive/pull-request | 2022-01-12T21:49:19 | 2025-04-01T04:33:04.681332 | {
"authors": [
"ShannonLCapper"
],
"repo": "ShannonLCapper/netlify-cms-gatsby-spike",
"url": "https://github.com/ShannonLCapper/netlify-cms-gatsby-spike/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
282723311 | Added the spo site list command solving #115
Added the spo site list command solving #115
Coverage remained the same at 100.0% when pulling ee0262cfbaa8ee2ea0d046727622f34dda54afde on waldekmastykarz:spo-site-list into bd2361710599e1a1ea401468c9cb092fbf986a17 on SharePoint:dev.
Merged manually
| gharchive/pull-request | 2017-12-17T20:59:25 | 2025-04-01T04:33:04.700707 | {
"authors": [
"coveralls",
"waldekmastykarz"
],
"repo": "SharePoint/office365-cli",
"url": "https://github.com/SharePoint/office365-cli/pull/129",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
562197863 | NodeJs download link not working
The nodejs download link returns 404
https://nodejs.org/dist/latest-v10.x/node-v10.18.1-x64.msi
Document Details
⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.
ID: 5bee09a8-022e-1b5a-4aab-d6e7c6216808
Version Independent ID: 49edce1e-f2be-e989-917f-c6b389363e2f
Content: Set up your SharePoint Framework development environment
Content Source: docs/spfx/set-up-your-development-environment.md
Product: sharepoint
Technology: sharepoint-framework
GitHub Login: @spdevdocs
Microsoft Alias: spdevdocs
The latest v10.x version of the NodeJs is v10.19.0, please download file using the below link.
https://nodejs.org/dist/latest-v10.x/node-v10.19.0-x64.msi
Dennis Guo
Microsoft SharePoint Community Support
| gharchive/issue | 2020-02-09T15:59:12 | 2025-04-01T04:33:04.705962 | {
"authors": [
"Dennis-LZ",
"gxy-dev"
],
"repo": "SharePoint/sp-dev-docs",
"url": "https://github.com/SharePoint/sp-dev-docs/issues/5310",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
249508181 | Documents are not renderred correctly in Production
Category
[ ] Question
[ ] Typo
[x ] Bug
[ ] Additional article idea
Expected or Desired Behavior
Documents are rendered correctly as designed.
Observed Behavior
All the documents under SharePoint Framework (e.g. https://dev.office.com/sharepoint/docs/spfx/set-up-your-developer-tenant) are not rendered currently in production. There's corrupted section at bottom of each page.
Steps to Reproduce
browser to https://dev.office.com/sharepoint/docs/spfx/set-up-your-developer-tenant
scroll down to bottom of the page
thx @aaronmi for reporting. Issue bypassed for the CMS owners - hopefully, will get fixed asap.
thx @aaronmi for reporting this. Fixed now in the production CMS side.
| gharchive/issue | 2017-08-10T23:57:33 | 2025-04-01T04:33:04.709859 | {
"authors": [
"VesaJuvonen",
"aaronmi"
],
"repo": "SharePoint/sp-dev-docs",
"url": "https://github.com/SharePoint/sp-dev-docs/issues/776",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
430359355 | Small spelling correction
Category
[x] Content fix
[ ] New article
What's in this Pull Request?
On the page https://docs.microsoft.com/en-us/sharepoint/dev/spfx/web-parts/get-started/use-fabric-react-components
Spelled "of" instead of "fo"
Guidance
No major update. Only a small spelling correction.
thx @nanddeepn, appreciated
| gharchive/pull-request | 2019-04-08T09:56:57 | 2025-04-01T04:33:04.712215 | {
"authors": [
"VesaJuvonen",
"nanddeepn"
],
"repo": "SharePoint/sp-dev-docs",
"url": "https://github.com/SharePoint/sp-dev-docs/pull/3748",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
485461109 | Removing note that the site footer hasn't rolled out WW yet
Category
[X] Content fix
[ ] New article
Related issues:
N/A
What's in this Pull Request?
According to MC 186030 (https://admin.microsoft.com/AdminPortal/home?switchtomodern=true#/MessageCenter?id=MC186030) it has rolled out WW by the end of July 2019 already, so the notice about it not being available yet can be removed.
thx @KoenZomers, appreciated.
| gharchive/pull-request | 2019-08-26T21:26:53 | 2025-04-01T04:33:04.714251 | {
"authors": [
"KoenZomers",
"VesaJuvonen"
],
"repo": "SharePoint/sp-dev-docs",
"url": "https://github.com/SharePoint/sp-dev-docs/pull/4538",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1266790216 | Validation: Fix missing description
This PR addresses the following validation issues and includes an updated ms.date:
missing description
Docs Build status updates of commit 1bff4f4:
:warning: Validation status: warnings
File
Status
Preview URL
Details
docs/schema/property-element-property-bag.md
:warning:Warning
View
Details
docs/schema/property-element-property-bag.md
Line 2, Column 1: [Warning: ms-prod-and-service - See documentation] Only one of the following attributes can exist: 'ms.prod', 'ms.service'. Use ms.prod for on-premise products, or ms.service for cloud services.
For more details, please refer to the build report.
If you see build warnings/errors with permission issues, it might be due to single sign-on (SSO) enabled on Microsoft's GitHub organizations. Please follow instructions here to re-authorize your GitHub account to Docs Build.
Note: Broken links written as relative paths are included in the above build report. For broken links written as absolute paths or external URLs, see the broken link report.
Note: Your PR may contain errors or warnings unrelated to the files you changed. This happens when external dependencies like GitHub alias, Microsoft alias, cross repo links are updated. Please use these instructions to resolve them.
For any questions, please:Try searching the docs.microsoft.com contributor guidesPost your question in the Docs support channel
This PR addresses the following validation issues and includes an updated ms.date:
missing description
Global effort to fix build validation errors
@VesaJuvonen
The Content & Learning team is fixing build validation errors on docs.microsoft.com for the rest of H2. This will eliminate potential accessibility, security, and usability issues. This PR includes only build validation fixes and does not change other content. Please review within five business days and merge or comment in the PR with any changes you'd like to see. Thank you!
| gharchive/pull-request | 2022-06-10T00:07:10 | 2025-04-01T04:33:04.724127 | {
"authors": [
"VesaJuvonen",
"annmariehitchcock"
],
"repo": "SharePoint/sp-dev-docs",
"url": "https://github.com/SharePoint/sp-dev-docs/pull/8112",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
465591658 | Looking for to Integrate an ICal Feed in SharePoint Online which is Link Based || SharePoint Calendar
I am looking for a way to integrate an ICal Calendar Feed in SharePoint Online Which will Keep Synchronizing with SharePoint using the ICal feed.
Originally the feed belongs from a Third Party Human Resource Software, Using this I am trying to keep the calendar synched with the updates for e.g. Holidays!
Can someone help me with how this could be achieved!
Thank you
Please, can someone help with this? Thank you
@tejasthite the React Calendar Feed web part sample will retrieve iCal feeds and present them as if they were native events in SharePoint. However, it does not sync with local SharePoint content.
If you want to sync content, I would suggest that you try using Flow instead and create a scheduled flow to regularly sync the content.
Let me know if you aren't familiar with Flow, I'd be happy to help.
Yes, I would need help with this, how can I get this accomplished? thank you
I tried the steps from the Readme.md File, I am getting the following errors || any help will be appreciated!
Thank you
C:\Users\tejast>npm install
npm WARN ajv-keywords@3.2.0 requires a peer of ajv@^6.0.0 but none is installed. You must install peer dependencies yourself.
npm WARN gulp-karma@0.0.5 requires a peer of karma@>=0.10 <=0.13 but none is installed. You must install peer dependencies yourself.
npm WARN slick-carousel@1.8.1 requires a peer of jquery@>=1.8.0 but none is installed. You must install peer dependencies yourself.
npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@1.2.4 (node_modules\fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.2.4: wanted {"os":"darwin","arch":"any"} (current: {"os":"win32","arch":"x64"})
Then When i try the npm audit fix - I get the Below || Please can someone help
Thank you
C:\Users\tejast>npm audit fix
npm WARN deprecated left-pad@1.3.0: use String.prototype.padStart()
npm WARN deprecated kleur@2.0.2: Please upgrade to kleur@3 or migrate to 'ansi-colors' if you prefer the old syntax. Visit https://github.com/lukeed/kleur/releases/tag/v3.0.0\ for migration path(s).
npm WARN deprecated browserslist@1.7.7: Browserslist 2 could fail on reading Browserslist >3.0 config used in other tools.
npm WARN deprecated flatten@1.0.2: I wrote this module a very long time ago; you should use something else.
deasync@0.1.15 install C:\Users\tejast\node_modules\deasync
node ./build.js
win32-x64-node-10 exists; testing
Binary is fine; exiting
core-js@2.6.9 postinstall C:\Users\tejast\node_modules\babel-register\node_modules\core-js
node scripts/postinstall || echo "ignore"
Thank you for using core-js ( https://github.com/zloirock/core-js ) for polyfilling JavaScript standard library!
The project needs your help! Please consider supporting of core-js on Open Collective or Patreon:
https://opencollective.com/core-js
https://www.patreon.com/zloirock
Also, the author of core-js ( https://github.com/zloirock ) is looking for a good job -)
core-js@2.6.9 postinstall C:\Users\tejast\node_modules\babel-runtime\node_modules\core-js
node scripts/postinstall || echo "ignore"
Thank you for using core-js ( https://github.com/zloirock/core-js ) for polyfilling JavaScript standard library!
The project needs your help! Please consider supporting of core-js on Open Collective or Patreon:
https://opencollective.com/core-js
https://www.patreon.com/zloirock
Also, the author of core-js ( https://github.com/zloirock ) is looking for a good job -)
npm WARN rollback Rolling back node-pre-gyp@0.12.0 failed (this is probably harmless): EPERM: operation not permitted, rmdir 'C:\Users\tejast\node_modules\fsevents\node_modules'
npm WARN slick-carousel@1.8.1 requires a peer of jquery@>=1.8.0 but none is installed. You must install peer dependencies yourself.
npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself.
npm WARN ajv-keywords@3.4.1 requires a peer of ajv@^6.9.1 but none is installed. You must install peer dependencies yourself.
npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself.
npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself.
npm WARN office-ui-fabric-react@6.199.0 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself.
npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself.
npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself.
npm WARN @uifabric/utilities@6.41.6 requires a peer of @types/react-dom@>=16.3.0 <17.0.0 but none is installed. You must install peer dependencies yourself.
npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@1.2.9 (node_modules\fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.2.9: wanted {"os":"darwin","arch":"any"} (current: {"os":"win32","arch":"x64"})
@microsoft/sp-build-web@1.8.2
@microsoft/sp-webpart-workbench@1.8.2
added 313 packages from 120 contributors, removed 282 packages, updated 196 packages and moved 14 packages in 340.523s
fixed 9448 of 11201 vulnerabilities in 682936 scanned packages
1746 vulnerabilities required manual review and could not be updated
1 package update for 7 vulns involved breaking changes
(use npm audit fix --force to install breaking changes; or refer to npm audit for steps to fix these manually)
@hugoabernier - i was successfully able to load the Web part and attach the URL for my iCal Calendar Feed but for some reason it keeps me saying - "Oops, something went wrong! We can't display your events at the moment. Please try again later."
Any help would be appreciated? Thank you
@hugoabernier it would be great if you could help with this. Thank you!
@tejasthite do you mind sharing the iCal URL so that I can try to diagnose the issue?
@hugoabernier is there a way I could send you a Private Message with the Link? Thank you
Yes send me a direct message on my Twitter account
@hugoabernier dropped you the link on twitter, Thank you
@hugoabernier Thank you Very Much, I will give it a try and update.
@hugoabernier - Wil I need to reclone the WebPart Sample after which the changes you have made recently? Thank you
@hugoabernier - I have been able to Pull the Events Now Thank you Very Much for your help, How do I now integrate this with My SharePoint Online Calendar?
That's great @tejasthite ! Do you have access to Flow? If so, may I suggest that you close this issue and we'll continue this one offline via Twitter DM? I'll come back later and update this issue with our resolution so that anyone else who needs this type of solution can benefit from it.
@hugoabernier - Yes I do have Flow. It would be always great to have the solution for the solution proposed with React Calendar Feed asap, I truly need this as well.
But at the same time if Flow is going faster right now then I am happy to implement that as well.
Thank you very much again. "let me how to proceed Further"?
@hugoabernier , Please could you update on the above? Thank you
@hugoabernier, Thank you very much for the Time from Your End - and the Microsoft Flow, Excellent work with this. I appreciate all your help. Using Flow it works Perfect.
Thank you
@tejasthite I'm glad it worked out. If anyone is looking for the solution we used, we documented the approach in this blog post and provided the step-by-step instructions in this one
| gharchive/issue | 2019-07-09T06:21:11 | 2025-04-01T04:33:04.750688 | {
"authors": [
"hugoabernier",
"tejasthite"
],
"repo": "SharePoint/sp-dev-fx-webparts",
"url": "https://github.com/SharePoint/sp-dev-fx-webparts/issues/909",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
489527291 | Icons does not display in peoples directory
I got the latest code and solution and added some properties to view in card in peoples directory. and bundled it and deployed. But i cannot view any of the fabric icons ?
Do u have any idea why>]?
Screenshot
I am also facing this issue.
Hi, are you still facing this issue? Are you able to share the code you modified?
| gharchive/issue | 2019-09-05T04:54:43 | 2025-04-01T04:33:04.753498 | {
"authors": [
"bcameron1231",
"ddmarsden",
"shwanthi"
],
"repo": "SharePoint/sp-starter-kit",
"url": "https://github.com/SharePoint/sp-starter-kit/issues/317",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
979845839 | xlsx.read not work in chrome
use In-Browser demo https://oss.sheetjs.com/
use same csv file , firefox show right result however chrome didn`t work fine.
firefox result:
chrome result:
click test.csv to download test file test.csv
thanks
| gharchive/issue | 2021-08-26T05:18:44 | 2025-04-01T04:33:04.767273 | {
"authors": [
"mario-mui"
],
"repo": "SheetJS/sheetjs",
"url": "https://github.com/SheetJS/sheetjs/issues/2368",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
676099918 | Adda option for creating new user data
action need to be performed with js
What acton do you need to be performed with js
Want to create user and store their data in local json file.
All from front end only.
User fills the form data and that data is submitted to json file.
addUser.html is the form which want to convert the details into json file
yes @kunalprakash1309
Assign this task to me. I want to try on it
hey @kunalprakash1309 are you still working on this?
| gharchive/issue | 2020-08-10T12:31:19 | 2025-04-01T04:33:04.898683 | {
"authors": [
"ShivSt",
"kunalprakash1309"
],
"repo": "ShivSt/demoWebsite",
"url": "https://github.com/ShivSt/demoWebsite/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
157872515 | How do you get the XML's response & request
Can you please explain what the easiest way to get a response/request XML/JSON strings without debugging the code?
any logs for that? couldn't find it.
Logging is the responsibility of the given adapters. You can look at the carrier you care about to see how it is logged. Or debug the request.
In general, reporting an issue should follow our guideline for contributing and should be well described problem with the code. Not simply a question.
| gharchive/issue | 2016-06-01T10:21:58 | 2025-04-01T04:33:04.931567 | {
"authors": [
"kmcphillips",
"megetron"
],
"repo": "Shopify/active_shipping",
"url": "https://github.com/Shopify/active_shipping/issues/377",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1282815735 | [Bug]: Deploy App (without running dev server) > App Url in Partners links to 404 page
Please confirm that you have:
[X] Searched existing issues to see if your issue is a duplicate. (If you’ve found a duplicate issue, feel free to add additional information in a comment on it.)
[X] Reproduced the issue in the latest CLI version.
In which of these areas are you experiencing a problem?
Extension
Expected behavior
Typically, the App Url would be set to an ngrok url if you ran npm run dev or yarn dev before deploying and creating the app through the CLI. In this case, if you don't ever run an app locally, a default App Url appears to be set. When a defauIt URL is used, it should resolve to a page that does not 404.
In addition to this, it would be helpful if this linked to a page to tell me that I need to either deploy the App to a public URL or run the app locally in order to generate a functional merchant install link.
Actual behavior
The link for the App Url in the Partners app is https://shopify.github.io/shopify-cli/help/start-app/ which goes to a 404 page.
Stack trace
No response
Reproduction steps
Create a new app using yarn create @shopify/app --template node.
Scaffold a new extension using yarn scaffold extension.
Deploy WITHOUT running yarn dev. Use yarn deploy to deploy and create a new app.
Navigate to the partners link and you'll see the App URL field is set to https://shopify.github.io/shopify-cli/help/start-app/ which is a 404 page.
Operating System
Mac OS Monterey
Shopify CLI version (check your project's package.json if you're not sure)
3.0.24
Shell
zsh
Node version (run node -v if you're not sure)
v18.0.0
What language and version are you using in your application?
Node
Hi @jamesvidler -- I believe this issue is resolved via https://github.com/Shopify/cli/pull/153. Do let us know if I'm missing something and your need is a different one.
| gharchive/issue | 2022-06-23T18:57:32 | 2025-04-01T04:33:04.939067 | {
"authors": [
"jamesvidler",
"shauns"
],
"repo": "Shopify/cli",
"url": "https://github.com/Shopify/cli/issues/47",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
813542930 | Issues with indexes
While generating a new db migration, @sambostock and I noticed that the descending order was dropped from the ["task_name", "created_at"] index when the migration ran.
Some of our adopters also seem to not have the order on the index present:
https://github.com/Shopify/arrive-server/blob/master/db/schema.rb#L393
https://github.com/Shopify/accio/blob/master/db/schema.rb#L192
@etiennebarrie had also flagged that the index on just task_name is potentially redundant - we should double check whether we need it or not.
Some updates on this:
Confirmed that our task_name index is redundant and can be dropped. Will open a PR
Adopters not having the sort order present on the index is actually not a bug, but instead due to the fact that most of our apps are using MySQL 5.7, and ordering is only present on MySQL 8.0+ 🤦♀️
The sort order disappearing from SQLite3 when running a new migration is actually the result of a bug in the SQLite3 Adapter in Rails.
| gharchive/issue | 2021-02-22T14:17:55 | 2025-04-01T04:33:04.947210 | {
"authors": [
"adrianna-chang-shopify"
],
"repo": "Shopify/maintenance_tasks",
"url": "https://github.com/Shopify/maintenance_tasks/issues/342",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
57246928 | Producer: implement AsyncClose
@wvanbergen
:+1: tested this on a local app.
| gharchive/pull-request | 2015-02-10T22:21:39 | 2025-04-01T04:33:04.986536 | {
"authors": [
"eapache",
"wvanbergen"
],
"repo": "Shopify/sarama",
"url": "https://github.com/Shopify/sarama/pull/258",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1836967262 | upgrade prettier plugins
Description
This PR upgrades prettier plugin versions so we can upgrade web's version to ^3
Also if you're in the process of making breaking changes - @typescript/eslint-plugin v6 recently got released and we'll need to update to use that if we want to support Typescript 5.2 or later.
| gharchive/pull-request | 2023-08-04T15:36:11 | 2025-04-01T04:33:05.002314 | {
"authors": [
"BPScott",
"m4thieulavoie"
],
"repo": "Shopify/web-configs",
"url": "https://github.com/Shopify/web-configs/pull/397",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
379308808 | Add Croissant Makers
The concept of Croissant Makers:
Give $20/10s. That's fun to implement into my quarter-second clock! :(
Costs $75 to hire
Has some sort of funny tooltip, just because.
Made this in Commit 49.
| gharchive/issue | 2018-11-09T20:22:07 | 2025-04-01T04:33:05.024907 | {
"authors": [
"Shrubhog"
],
"repo": "Shrubhog/expresso",
"url": "https://github.com/Shrubhog/expresso/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
124087696 | Network AVROTROS missing
Network AVROTROS missing in network_timezones.txt and also uploaded a logo for this network.
2015-12-28 10:58:19 INFO Thread-51 :: Network was not found in the network time zones: AVROTROS
Thanks for the pull request! Before a real human comes by, please make sure your PR has all of the below criteria checked
[x] Give a description on what the PR is for.
[x] Make sure your PR is based on the DEVELOP branch
[x] Don't send big changes all at once. Split up big PRs into multiple smaller PRs that are easier to manage and review
Please make sure you also read contribution guide and followed all of the steps.
Thanks!
The logo for the AVROTROS network. Do you want me to create a pr at https://github.com/SickRage/SickRage/tree/master/gui/slick/images/network/ for this??
@aboogerd the logo goes on the main repo like you had it
| gharchive/pull-request | 2015-12-28T16:25:35 | 2025-04-01T04:33:05.065882 | {
"authors": [
"SickRage2",
"aboogerd",
"duramato"
],
"repo": "SickRage/sickrage.github.io",
"url": "https://github.com/SickRage/sickrage.github.io/pull/22",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2687957887 | Collecting the sustainable materials for building
Below are the acceptance criteria for this issue which is an enhancement:
The game should identifies the specific locations where players can get sustainable materials.
Only sustainable materials should be able to collect for this quest.
Points will be rewarded for collecting materials without damaging the environment.
Done
| gharchive/issue | 2024-11-24T16:10:57 | 2025-04-01T04:33:05.069283 | {
"authors": [
"Siddhartha-atu"
],
"repo": "Siddhartha-atu/Zelda_Githubactions",
"url": "https://github.com/Siddhartha-atu/Zelda_Githubactions/issues/6",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1780904602 | Multi browser support
Since the extension has turned into a glorified script and style injector, adapting LiChess Tools for other browsers might be trivial. I would still need help from people using said browsers.
I use Brave, which is a version of Chrome. Edge is a version of Chrome. What else is there? Firefox? Safari? I don't like them, so how could I test them properly?
I don't have much affinity in the programming part of extensions, but I can try to help test in firefox
By the way, if you want to launch the .xpi file first before publishing https://addons.mozilla.org also works
The problem is with the content_scripts manifest property. While Firefox seems to support Manifest V3, it doesn't support the "world" setting which allows a script to be loaded in the context of the web page. I will have to research how to manage this in FF.
adding dubious tag because Firefox has 1.5% of browser share as opposed to Chrome+Edge which is 65%. I had no idea Firefox is so lame. :)
I will not do it. Firefox is small and stupid and Safari asks me for money.
For future Mac users who come across this issue and are looking into Safari compatibility: you should know, you can convert Chrome extensions to Safari extensions using Xcode on your Mac (no developer account needed). Just Google it and follow the instructions. Also, remember to enable unsigned extensions in Safari.
New developments make this a possibility. Firefox version implemented in v2.3.141
| gharchive/issue | 2023-06-29T14:07:52 | 2025-04-01T04:33:05.072875 | {
"authors": [
"KeatsPeeks",
"SergioGlorias",
"Siderite"
],
"repo": "Siderite/lichessTools",
"url": "https://github.com/Siderite/lichessTools/issues/34",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1563031201 | possible overflow bugs
https://github.com/SierraWireless/octave-orp/blob/3ccf4406a2c0cd9f5f8c081b2bf340e575158806/clients/c/src/orpUtils.c#L141-L150
At Line 144, there is no guarantee that message->status * -1 < sizeof(statusStr), which may lead to the following bug:
==257646==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000005 (pc 0x0000005371b0 bp 0x7fff256688e0 sp 0x7fff25668058 T0)
==257646==The signal is caused by a READ memory access.
==257646==Hint: address points to the zero page.
#0 0x5371b0 in __sanitizer::internal_strlen(char const*) (/home/parallels/octave-orp/clients/c/bin/orp+0x5371b0)
#1 0x4c57d6 in printf_common(void*, char const*, __va_list_tag*) (/home/parallels/octave-orp/clients/c/bin/orp+0x4c57d6)
#2 0x4c6c4e in printf (/home/parallels/octave-orp/clients/c/bin/orp+0x4c6c4e)
#3 0x56730c in orp_MessagePrint /home/parallels/octave-orp/clients/c/src/orpUtils.c:144:17
#4 0x5630f7 in orp_HdlcDeframe /home/parallels/octave-orp/clients/c/src/orpClient.c:446:9
https://github.com/SierraWireless/octave-orp/blob/3ccf4406a2c0cd9f5f8c081b2bf340e575158806/clients/c/src/orpProtocol.c#L1014-L1018
At Line 1017, offset may equal pktLen - 1. Thus, pktBuf[offset + 1] equals pktBuf[pktLen], which leads to the following overflow bug when calling strtoul:
==261454==ERROR: AddressSanitizer: SEGV on unknown address 0x6060002d0000 (pc 0x7efd0d5ef17b bp 0x000000000000 sp 0x7ffe53ff0c80 T0)
==261454==The signal is caused by a READ memory access.
#0 0x7efd0d5ef17b in __GI_____strtoul_l_internal /build/glibc-sMfBJT/glibc-2.31/stdlib/../stdlib/strtol_l.c:432:28
#1 0x55baad in orp_ProtocolDecode_v1 /home/parallels/octave-orp/clients/c/src/orpProtocol.c:1017:42
Similar issues may happen at Line 995, Line 1006, and other places where pktBuf[offset + 1] is used.
https://github.com/SierraWireless/octave-orp/blob/3ccf4406a2c0cd9f5f8c081b2bf340e575158806/clients/c/src/orpProtocol.c#L1054-L1059
At Line 1058, pktBuf + ORP_OFFSET_VARLENGTH may overrun the buffer.
==261221==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x6020000000d4 at pc 0x0000004c5908 bp 0x7ffd3ad375a0 sp 0x7ffd3ad36d20
READ of size 1 at 0x6020000000d4 thread T0
#0 0x4c5907 in printf_common(void*, char const*, __va_list_tag*) (/home/parallels/octave-orp/clients/c/bin/orp+0x4c5907)
#1 0x4c6c4e in printf (/home/parallels/octave-orp/clients/c/bin/orp+0x4c6c4e)
#2 0x55bec8 in orp_ProtocolDecode_v1 /home/parallels/octave-orp/clients/c/src/orpProtocol.c:1056:9
| gharchive/issue | 2023-01-30T19:04:47 | 2025-04-01T04:33:05.080803 | {
"authors": [
"harryreps"
],
"repo": "SierraWireless/octave-orp",
"url": "https://github.com/SierraWireless/octave-orp/issues/17",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
134555747 | Create a tab-separated text format for creating translations that are in the .po file
The .po file is a bit impenetrable to the average translator. Create a simple table view with the English source text in the first column, and the other languages in the second-third-etc. columns
One such file per foreign language would in principle also do the trick, but as the number of terms to be translated grows, it may be easier to keep things consistent if we would see all languages in a single table
Create a script for converting the tab-separated text to new/updated .po files for each language
There are programs for editing '.po' files that supposedly make translating easier, you might want to look into them. I have no experience with them myself, so I am not sure if they work with djangos .po files.
I have not used these apps myself, but here are a few examples:
https://poedit.net/
https://localise.biz/free/poeditor
Thanks @henrinie , I'll check it out.
@ocrasborn, are the suggestions Henri made sufficient? If so, can we close this issue?
Yes, let's close this for now, I think Poedit will do the job fine.
| gharchive/issue | 2016-02-18T11:33:54 | 2025-04-01T04:33:05.107326 | {
"authors": [
"henrinie",
"ocrasborn",
"vanlummelhuizen"
],
"repo": "Signbank/NGT-signbank",
"url": "https://github.com/Signbank/NGT-signbank/issues/132",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
523298537 | BREAKING CHANGE: Rename AppendToStream to camelCase from PascalCase
Seriously...sadface.
:tada: This issue has been resolved in version 1.0.0 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
| gharchive/issue | 2019-11-15T07:16:07 | 2025-04-01T04:33:05.196360 | {
"authors": [
"YannickMeeus"
],
"repo": "Silly-Goose-Software/event-sauced-ts",
"url": "https://github.com/Silly-Goose-Software/event-sauced-ts/issues/255",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2535413632 | 一直在加载 load project
已经无法使用了
用v0.2.4的版本尝试一下
麻烦了,还有一个问题,就是每一次打开项目,有概率无法加载, 其他插件也停止了,没有任何信息,需要启动,触发进度才可以继续操作
用了最新的版本还会吗,有复现的gif吗
可以查看一下 vscode 的 output 日志有没有什么报错之类的
目前没发现错误,正在load project
这是正常情况, 非正常情况就是一直 load project
你这个不是最新版本吧的截图吧
这是最新的
你本地开代理了吗,因为 unpkg 和 jsdelivr需要梯子才能拉下来
有代理,但不是全局的
duplicated: #89
| gharchive/issue | 2024-09-19T06:33:32 | 2025-04-01T04:33:05.201658 | {
"authors": [
"Simon-He95",
"xiaojia21190"
],
"repo": "Simon-He95/common-intellisense",
"url": "https://github.com/Simon-He95/common-intellisense/issues/85",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
173375171 | Hosting a second demo?
Hi there,
I would be happy to host a second demo if you want.
Thank you,
Timo
The demo has data donated by the admins of two separate servers. If you want your server to be included there, feel free to shoot me an email. My email is in my profile.
If you want to run it in your server and demo a different language, you can include the link to there in the email.
~Simon
I mailed you, if you like we can discuss the details via email?
| gharchive/issue | 2016-08-26T05:50:47 | 2025-04-01T04:33:05.219838 | {
"authors": [
"SimonOrJ",
"ntimo"
],
"repo": "SimonOrJ/CoreProtect-Lookup-Web-Interface",
"url": "https://github.com/SimonOrJ/CoreProtect-Lookup-Web-Interface/issues/8",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
928426957 | When running on Docker I just get 502
I've deployed the docker image to k8s for the ui and when I try to access the contianer on port 9000 I just get 502. I have tried running the docker image locally but to the same result.
Any logs that you can share to try and debug? docker-compose logs ui
Was this solved by the new version of the API server?
| gharchive/issue | 2021-06-23T16:12:24 | 2025-04-01T04:33:05.250498 | {
"authors": [
"dillu24",
"tomjohnburton"
],
"repo": "SimplyVC/panic_polkadot",
"url": "https://github.com/SimplyVC/panic_polkadot/issues/36",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1871505562 | add the docker file
The docker file has been added with a system to document what was simulated and the results.
Rejected as we can't remove the check for complex arithmetic. The code needs to be run in an environment where the complex variables are setup correctly.
| gharchive/pull-request | 2023-08-29T11:51:02 | 2025-04-01T04:33:05.251379 | {
"authors": [
"njsaugat",
"pmanandhar1452"
],
"repo": "SimulateAnything/wentworth-simulate-anything-acoustic-levitation-v0",
"url": "https://github.com/SimulateAnything/wentworth-simulate-anything-acoustic-levitation-v0/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2448215183 | Documentation of WG meeting Mon 5th Aug 2024 #37
Document the WG meeting on 5th August 2024, and add it to the Archives via the meeting summary tool https://archive-oracle.netlify.app/
Done https://snet-ambassadors.gitbook.io/singularitynet-archive/timeline/2024/august-2024/week-32#ai-ethics-wg
Transferred to Vani and sent for payment
| gharchive/issue | 2024-08-05T10:36:31 | 2025-04-01T04:33:05.253131 | {
"authors": [
"VanCardui"
],
"repo": "SingularityNet-Ambassador-Program/AI-Ethics-Workgroup",
"url": "https://github.com/SingularityNet-Ambassador-Program/AI-Ethics-Workgroup/issues/39",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
216999375 | Add Julia language definition
This PR adds a definition file for the Julia language. It uses the Julia standard of 4 spaces for indentation. With this, the Julia files in Julia's standard library seem to be highlighted correctly.
The only issue is names that contain numbers: the numbers are highlighted within the name. For example, in the type Int64, Int will appear normally (i.e. not highlighted) but 64 is highlighted as a number. It's part of the identifier, so it would be nice if Int64 (as just one of many examples) could be homogeneously colored. Is there a way I can specify that in the configuration file?
Oddly enough, despite the aforementioned issue, the unsigned literal0x0000, for example, has the leading 0 highlighted but not the trailing zeros.
Unfortunately not in a generic way. wsedit's highlighter works like this:
Check all active -l* rules, take the longest match.
If no rule matches, highlight on character class.
You could of course mark Int64 as a keyword, but I don't think that's what you meant. I was thinking about adding an option to disable character class highlighting altogether, but I didn't have any good reason to do so yet.
... Now that I think about it, maybe switching from character-based highlighting to a style where the first character of a "word" determines its colour would be a good idea, but it doesn't look like a quick 5 minute fix. (Hell, it might even boost performance in some cases)
Cool, thanks for the explanation.
An advantage for Julia of the current behavior is that Julia allows numeric literal juxtaposition for multiplication. That is, 2x means 2*x.
Yeah, I think I've seen a few languages do that kind of thing. It's definitely cool to write, but it doesn't make my job any easier =)
So maybe "try to match the longest identifier ([A-Za-z_][A-Za-z0-9_]*) and make sure to paint it all white"?
That would at least work for Julia.
| gharchive/pull-request | 2017-03-25T19:53:44 | 2025-04-01T04:33:05.259066 | {
"authors": [
"SirBoonami",
"ararslan"
],
"repo": "SirBoonami/wsedit",
"url": "https://github.com/SirBoonami/wsedit/pull/17",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
230898476 | [Linux] Failure to Launch
I'm running into a bit of a problem, but perhaps it's just user error and I'm just overlooking something.
I'm using Ubuntu 17.04, and I've installed the following packages which I believe are the same as the dependencies listed on the download page of the TrueCraft website:
mono-complete
gtk-sharp2
libwebkitgtk-3.0-0
libsdl-mixer1.2
I'm able to use the launcher just fine; I was able to sign in, select single player, and generate a world.
The problem started once I attempted to launch the game, it simply returned me to the launcher. I checked around the issues for this problem, but could only find it for people using Windows, and your recommendation was to use Linux.
Here's the output from mono:
Image format: http://i.imgur.com/VeVlNTJ.png
What I assume is the important part:
Cannot connect to server socket err = No such file or directory
Cannot connect to server request channel
jack server is not running or cannot be started
JackShmReadWritePtr::~JackShmReadWritePtr - Init not done for 4294967295, skipping unlock
JackShmReadWritePtr::~JackShmReadWritePtr - Init not done for 4294967295, skipping unlock
X Error of failed request: BadValue (integer parameter out of range for operation)
Major opcode of failed request: 154 (GLX)
Minor opcode of failed request: 3 (X_GLXCreateContext)
Value in failed request: 0x0
Serial number of failed request: 78
Current serial number in output stream: 79
AL lib: (EE) alc_cleanup: 2 devices not closed
Any help is appreciated :)
can you try running the server and client separately? Run the server normally, and run the client with mono TrueCraft.Client.exe localhost LakenH.
Appreciate the quick response :)
I was able to run the server fine, but get the identical error when running the client with that command.
What graphics driver are you using?
Using a NVIDIA one, version 375.66, but I see that there are more recent versions; should I try updating it?
Try nouveau.
Only slightly better, lol. This time it at least launches, but still throws a similar error. However, after the first key press, the game crashes and gives quite the error.
Output:
https://gist.github.com/LakenH/39f7aa3b966bbdc67633031ae0bdc0c9
Ouch. I'm not sure what the issue is, but seeing as the client is going to be overhauled soon it's not worth looking into.
Understandable, thanks for your time!
After some further testing, it seems to only throw that error and crash once I open the inventory. I'm still able to maneuver with WASD and such. So this might be two separate issues.
| gharchive/issue | 2017-05-24T01:55:45 | 2025-04-01T04:33:05.265827 | {
"authors": [
"LakenH",
"SirCmpwn"
],
"repo": "SirCmpwn/TrueCraft",
"url": "https://github.com/SirCmpwn/TrueCraft/issues/267",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
180485384 | Add bind mapper
This pull request creates a bind mapping subsystem (with unit tests), and integrates it into aerc. Some default binds are hardcoded, but can be overwritten in aerc.conf.
I'm a little uncertain about how I've named the bind mapping system, so if you have any better ideas, or other suggestions for improvements, I'm open to them.
Potential way to provide vim-like behaviour:
At ui.c:187 Create a queue of pending input events.
Add events from tb_peek_event to input queue.
Pop events from the queue and process them one at a time.
When handling a command generated from bind_handle_key_event, determine which
input events it would generate, and add them to the input queue. Generating basic
key presses will be easy. Support for <CR> etc. will need work. Perhaps existing
logic from bind.c could be leveraged.
Processing continues until there are no more events.
There's a gotcha with this though: a bad user configuration could generate an infinite loop. This needs to be caught somehow.
An additional optimisation: Add a rerender_required flag to state, and make rerender simply set it. Then at the end of the ui_tick, if it's been set, unset it and actually rerender the ui. This will prevent spurious rerenders.
That sounds perfect.
This now handles key events as intended, however the new functionality (bind_translate_key_name) doesn't have any unit tests yet, and there's nothing to prevent an infinite loop.
How do you make a binding that i.e. types in Ctrl+C?
In the [input] section of aerc.conf:
f = <Ctrl+c>
Cool. LGTM. Do you want this merged now or do you want to add tests and loop checking first?
I'll add another unit test or two, then merge. I think loop checking can wait, as it's not that likely to happen by accident, and I'd rather give it some thought and put in an elegant safeguard, than hack something in now and never bother to improve it.
Sounds good.
A new unit test uncovered that I'd missed support for <Meta+key>. That's now been resolved, and this is ready for merging.
Thanks!
| gharchive/pull-request | 2016-10-01T23:11:10 | 2025-04-01T04:33:05.273426 | {
"authors": [
"SirCmpwn",
"eXeC64"
],
"repo": "SirCmpwn/aerc",
"url": "https://github.com/SirCmpwn/aerc/pull/11",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1671242751 | Supporting knowledge of other participants
In some of OFMC's examples, it is specified when a participant has knowledge of another participant. Is this something that should be supported?
Seems to be kinda supported already
| gharchive/issue | 2023-04-17T13:55:18 | 2025-04-01T04:33:05.278479 | {
"authors": [
"PhilipFlyvholm"
],
"repo": "SirSorensen/Seculizer",
"url": "https://github.com/SirSorensen/Seculizer/issues/49",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1004704132 | Downloads links in menu and content
I've added links to downloads in the menu and in relevant MD files.
Fixes #53
Looking at it I do believe we need to add some sort of indicator to point out its a different link then the others. But we can add that later
@markvanaalst : I agree. Also, since they are on another site, I was thinking perhaps the visualization used in the waffle menu that shows they go somewhere else? That would draw attention, but also be informative. I'm going to take a look into that today when I get a chance.
| gharchive/pull-request | 2021-09-22T19:42:12 | 2025-04-01T04:33:05.285165 | {
"authors": [
"jst-cyr"
],
"repo": "Sitecore/developer-portal",
"url": "https://github.com/Sitecore/developer-portal/pull/64",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
838625182 | Support import and export declarations
Hope to support ES Module.
acorn supported sourceType option. Can Sval supports this option or enhance the extensibility of options?
Any progress on this?
@curly210102 @TheShermanTanker
Supported from 0.5.1
| gharchive/issue | 2021-03-23T11:27:12 | 2025-04-01T04:33:05.302661 | {
"authors": [
"Siubaak",
"TheShermanTanker",
"curly210102"
],
"repo": "Siubaak/sval",
"url": "https://github.com/Siubaak/sval/issues/68",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1624957888 | Support to pull URLs from CSS
This PR fixes #68 , #212 , #142
I did not add a test for this new code. I tested this by pulling scraping resources from https://www.purdue.edu/itap/iamo/p/login.css with a depth of 1. This can be tried by executing:
cargo run https://www.purdue.edu/itap/iamo/p/login.css --depth 1
I have added a new struct named FileType which handles html, css, or other. This is a drop in replacement for the previously used is_html function in parser.rs.
Edit: There is a slight issue, where if the request for the CSS file is 302, and that CSS file has a relative URL path, then the parser will not update this accordingly. Looking to fix.
Edit 2: Auditing this bug, it appears that other files will suffer from not being to download relative dependencies on a 302 redirect. Perhaps this should go into another PR? (e.g. request /dir1/dir2/hello.html 302 -> /dir1/hello.html, hello.html contains dependency mystuff.css, where its real path is at /dir1/mystuff.css)
Thanks you very much!
I won't have time to review this in the next couple of days, but be sure that I'm really excited about this and I will give it a spin and a review as soon as I can!
Another item is the use of __querystring__ in line 13 of url_helper.rs. This may panic on an instance of a request where the parameter utilizes special characters like = or ;...maybe I should adjust for url encoding?
maybe I should adjust for url encoding?
That's a good solution
Unfortunately I've become way too busy to get back to this. I hope to revisit this in the future! Such a useful library...if I get a chance, I'll be sure to open another pr.
No problem, I totally get it. You went way further than I never did with this.
Thanks you very much :)
I gave this a spin again. I managed to fix an old issue where books.toscrape.com did not have the star ratings because we where missing a font so I'm quite happy with that.
I'm trying to fix this:
Another item is the use of querystring in line 13 of url_helper.rs. This may panic on an instance of a request where the parameter utilizes special characters like = or ;...maybe I should adjust for url encoding?
e.g.: thread '' panicked at 'Couldn't create folder ...some directory\fonts.googleapis.com\css__querystring__family=Archivo+Narrow:400,700: The directory name is invalid. (os error 267)', logger.rs:42:9
but cargo run https://www.purdue.edu/itap/iamo/p/login.css --depth 1 does not seems to provide this panick anymore
| gharchive/pull-request | 2023-03-15T07:59:44 | 2025-04-01T04:33:05.320520 | {
"authors": [
"Skallwar",
"dsgallups"
],
"repo": "Skallwar/suckit",
"url": "https://github.com/Skallwar/suckit/pull/213",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
354235773 | Build failed
Hello, there is a question when I build the QtMvvm with Qt Creator 5.9
'qpmx' 不是内部或外部命令,也不是可运行的程序
Project ERROR: qpmx initialization failed. Check the compilation log for details.
Did you install all the build dependencies found in the readme? Especially qpmx? If yes, ensure that qpmx and qpm are in the PATH
Also Qt 5.9 is not supported, as the QML-APIs used are only available from Qt 5.10 on.
Can I close this issue?
| gharchive/issue | 2018-08-27T08:55:51 | 2025-04-01T04:33:05.370009 | {
"authors": [
"Skycoder42",
"wxddegithub"
],
"repo": "Skycoder42/QtMvvm",
"url": "https://github.com/Skycoder42/QtMvvm/issues/3",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
218059637 | Fix plugin timer request not properly initialized
connects #333
Don’t know how to write unit test to guard against this particular kind of bug without refactoring the transport interface. @rickmak New issue #338 opened for that.
Merge via cb71a30
| gharchive/pull-request | 2017-03-30T02:01:03 | 2025-04-01T04:33:05.376754 | {
"authors": [
"cheungpat",
"rickmak"
],
"repo": "SkygearIO/skygear-server",
"url": "https://github.com/SkygearIO/skygear-server/pull/337",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
855802956 | Unable to start controller
The latest version doesn't start because of the following error
argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout 2021-04-12T10:06:16.393Z INFO controller-runtime.metricsmetrics server is starting to listen {"addr": ":8080"}
argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout 2021-04-12T10:06:16.393Z ERROR setup unable to read configuration {"error": "strconv.ParseBool: parsing \"\\\"true\\\"\": invalid syntax"}
argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout github.com/go-logr/zapr.(*zapLogger).Error
argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout /go/pkg/mod/github.com/go-logr/zapr@v0.2.0/zapr.go:132
argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout main.main
argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout /workspace/main.go:74
argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout runtime.main
argocd-progressive-rollout-5dc46f5d86-x5qb8 argocd-progressive-rollout /usr/local/go/src/runtime/proc.go:204
After some investigation this seems related to the quote clause in the Secret.yaml. We should also understand why the installation fails as per https://github.com/Skyscanner/argocd-progressive-rollout/runs/2322327460?check_suite_focus=true but the action results in a pass.
| gharchive/issue | 2021-04-12T10:06:54 | 2025-04-01T04:33:05.409477 | {
"authors": [
"maruina"
],
"repo": "Skyscanner/argocd-progressive-rollout",
"url": "https://github.com/Skyscanner/argocd-progressive-rollout/issues/51",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
141790850 | Move Routes out of Providers
Move the Routes out from regular Providers, and into a separate directory Routes. The Router component has to provide a helper provider abstract class to simplify and unify adding of routes to the Route Container.
This helper must define a register method and implement the Pimple\ServiceProviderInterface, so the user doesn't need to deal with this, but just extends this helper when adding new Route definition loader classes.
The register method must call the define method, which must be declared as abstract in the helper abstract class. The define method must add route info to a local property as array. The array must contain:
uri - can be regex, all rules from Router component apply
method - HTTP method in upper case
action - the route definition, callable
After define method call the register method must add all new definitions to the Route Container.
Resolved.
| gharchive/issue | 2016-03-18T07:02:08 | 2025-04-01T04:33:05.425767 | {
"authors": [
"slax0rr"
],
"repo": "SlaxWeb/Framework",
"url": "https://github.com/SlaxWeb/Framework/issues/47",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1316005525 | [Betterdiscord] Submit the theme to the official website
Suggestion
Title^
Benefit
Easier access
This suggestion is unique
[X] I have searched the issue tracker and did not find an issue describing my suggestion, especially not one that has been rejected.
You may use the editor below to elaborate further.
No response
Please read the README. You will see why this theme is not uploaded there.
might I ask if it's the ¨Original Work¨ requirement that's the problem?
| gharchive/issue | 2022-07-24T20:49:57 | 2025-04-01T04:33:05.427879 | {
"authors": [
"Slddev",
"isaktexas"
],
"repo": "Slddev/ModernX-Theme",
"url": "https://github.com/Slddev/ModernX-Theme/issues/30",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
85354705 | Exception while updating neighbours
Minecraft - 1.7.10
--------Involved Mods------
Ender IO - 2.2.8.381
Sanguimancy - 1.1.9-29
Mekinism - 8.1.3.231
--------Crash Report-------
Pastebin - http://pastebin.com/menKawd7
I can't tell you what was going on before this crash happened as this is from one of the servers using the Resonant Rise pack. I am still trying to recreate this issue, just thought you might have some insight on this one.
This issue is between Sanguimancy and Mekanism...EIO is just triggering it.
cool just checking.
| gharchive/issue | 2015-06-05T00:56:41 | 2025-04-01T04:33:05.430358 | {
"authors": [
"GodNero",
"tterrag1098"
],
"repo": "SleepyTrousers/EnderIO",
"url": "https://github.com/SleepyTrousers/EnderIO/issues/2543",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
326265050 | [Request] Processing (killing) of Soul Vial'ed mobs
A Machine that when powered and supplied with soul vials with mobs within, kills the creature as if it had been spawned in to the world and slain by any other means of killing. Could also have different toggle able settings to allow for player only drops, experience and adding looting effect somehow, all theses effect would make it use more energy or time of course
Thank you for your suggestion, but we manage suggestions on http://ideas.theideawall.com/EnderIO
This ticket will be closed because we want the list of open tickets to reflect known issues.
Ok i will create a copy of the suggestion, on the mention site.
| gharchive/issue | 2018-05-24T19:46:48 | 2025-04-01T04:33:05.432648 | {
"authors": [
"HenryLoenwind",
"andr9528"
],
"repo": "SleepyTrousers/EnderIO",
"url": "https://github.com/SleepyTrousers/EnderIO/issues/4659",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
152794093 | Show Block Name in GuiExternalConnectionSelector in 1.8.9.
Implemens #1984
See #3287 for 1.7.10 PR.
Oh, and something I noticed on your screenshot: Your text is touching the buttons, it needs to move down 1 pixel.
Text is now 1 pixel lower and rendered behind buttons. Didn't look like i had to do some GL-stuff.
Is it just that image or is the background a solid black now?
Its just the background. if you look VERY closely you can see a coal ore on the right. I think i just took the screenshot at night.
ok, that's fine. It may also be that the monitors at work are not really configured for anything. Actually, I know they aren't---I set them up. ;)
2 things: First, Block.getUnlocalizedName() is notoriously unreliable. It doesn't even work for all vanilla blocks, let alone modded blocks. Second, I expanded it a bit, I hope you don't mind:
It's great. Thanks a lot for improving it! Especially that late at night.
It's now even better, than I could have imagined.
That's the wonder of collaboration; the result is better than what any participant could have accomplished on their own.
But I'm still not sure if we are making clear what each button represents. So far we only have:
Its position
The name of the direction
The name of the block
The picture of the block
Maybe we should add a couple more hints....lol, just kidding ;)
I think it looks best with normal sized text in the buttons, and unicode text below.
| gharchive/pull-request | 2016-05-03T14:53:32 | 2025-04-01T04:33:05.437826 | {
"authors": [
"HenryLoenwind",
"MaPePeR",
"tterrag1098"
],
"repo": "SleepyTrousers/EnderIO",
"url": "https://github.com/SleepyTrousers/EnderIO/pull/3288",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
1713829448 | Update English documentation and screenshot to match the latest version of the software
Update this page, including screenshots to match the current software version.
Got it, thanks!
Le mer. 17 mai 2023 à 12:45, Andras Lasso @.***> a
écrit :
Assigned #27 https://github.com/Slicer/SlicerLanguagePacks/issues/27 to
@MohamedAllali https://github.com/MohamedAllali.
—
Reply to this email directly, view it on GitHub
https://github.com/Slicer/SlicerLanguagePacks/issues/27#event-9268934383,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AWFEO4M4IU5EML6EA5VGXMTXGTB4RANCNFSM6AAAAAAYFBFQIM
.
You are receiving this because you were assigned.Message ID:
@.***>
Good evening, Mr. Andras
I hope everything is going well. I want to ask you a question related to
SlicerLanguagePacks update, isn't it that Qt toolkit is now integrated with
SlicerLanguagePacks, so there is no need to download and install, my
question to you is it better to delete the download and install Qt toolkit
part of HowToUse
https://github.com/Slicer/SlicerLanguagePacks/blob/main/HowToUse.md#:~:text=this language selector.-,Download and install Qt toolkit,is done%2C the lrelease file will be located in "/usr/bin/" .,-SlicerLanguagePacks/HowToUse.md
.
Le mer. 17 mai 2023 à 13:48, Mohamed Alalli BILAL @.***>
a écrit :
Got it, thanks!
Le mer. 17 mai 2023 à 12:45, Andras Lasso @.***> a
écrit :
Assigned #27 https://github.com/Slicer/SlicerLanguagePacks/issues/27
to @MohamedAllali https://github.com/MohamedAllali.
—
Reply to this email directly, view it on GitHub
https://github.com/Slicer/SlicerLanguagePacks/issues/27#event-9268934383,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AWFEO4M4IU5EML6EA5VGXMTXGTB4RANCNFSM6AAAAAAYFBFQIM
.
You are receiving this because you were assigned.Message ID:
@.***>
Ok, thanks for the clarification.
Happy Weekend.
Mohamed.
Le sam. 20 mai 2023 à 14:15, Andras Lasso @.***> a
écrit :
Correct. Manual installation of Qt is not needed anymore (at least in
recent Slicer Preview Releases), so it should be removed from the
instructions.
—
Reply to this email directly, view it on GitHub
https://github.com/Slicer/SlicerLanguagePacks/issues/27#issuecomment-1555921185,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AWFEO4KHCXSZLVPEPIDJ7RLXHDGZBANCNFSM6AAAAAAYFBFQIM
.
You are receiving this because you were mentioned.Message ID:
@.***>
@lassoan I've updated How to use, you can see from here, wait for your comment if there's anything that needs to be modified.
done
| gharchive/issue | 2023-05-17T12:44:45 | 2025-04-01T04:33:05.490658 | {
"authors": [
"MohamedAllali",
"lassoan"
],
"repo": "Slicer/SlicerLanguagePacks",
"url": "https://github.com/Slicer/SlicerLanguagePacks/issues/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
193979240 | Support for Symfony 3
Will this support for symfony 3 version in the future?
Is this repo even still maintained ?
| gharchive/issue | 2016-12-07T07:44:19 | 2025-04-01T04:33:05.491637 | {
"authors": [
"SlimenTN",
"rainercedric23"
],
"repo": "SlikNL/DompdfBundle",
"url": "https://github.com/SlikNL/DompdfBundle/issues/15",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1167237672 | Use POWER_SAVING_LEGACY by default
BNO users experience strange delays with new power saving options. Revert to legacy to mitigate the issue and for further investigation.
Closing my pull request in favor of #125.
| gharchive/pull-request | 2022-03-12T09:39:10 | 2025-04-01T04:33:05.494351 | {
"authors": [
"deiteris"
],
"repo": "SlimeVR/SlimeVR-Tracker-ESP",
"url": "https://github.com/SlimeVR/SlimeVR-Tracker-ESP/pull/126",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1057619752 | 🛑 My blog(github.io) is down
In 35d3d9c, My blog(github.io) (https://smartbrave.github.io/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: My blog(github.io) is back up in a0eaf29.
| gharchive/issue | 2021-11-18T17:33:27 | 2025-04-01T04:33:05.514686 | {
"authors": [
"SmartBrave"
],
"repo": "SmartBrave/uptime",
"url": "https://github.com/SmartBrave/uptime/issues/1155",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
999590635 | 🛑 My blog(CDN) is down
In 7e45363, My blog(CDN) (https://sbrave.cn) was down:
HTTP code: 0
Response time: 0 ms
Resolved: My blog(CDN) is back up in 7f2e277.
| gharchive/issue | 2021-09-17T17:49:39 | 2025-04-01T04:33:05.517134 | {
"authors": [
"SmartBrave"
],
"repo": "SmartBrave/uptime",
"url": "https://github.com/SmartBrave/uptime/issues/331",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1751262187 | 🛑 My blog(CDN) is down
In 9b72292, My blog(CDN) (https://sbrave.cn) was down:
HTTP code: 0
Response time: 0 ms
Resolved: My blog(CDN) is back up in 3f6ded8.
| gharchive/issue | 2023-06-11T05:58:29 | 2025-04-01T04:33:05.519577 | {
"authors": [
"SmartBrave"
],
"repo": "SmartBrave/uptime",
"url": "https://github.com/SmartBrave/uptime/issues/3969",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2656666690 | Avoid native handling for zwave devices with odd endpoint layouts
Check all that apply
Type of Change
[ ] WWST Certification Request
If this is your first time contributing code:
[ ] I have reviewed the README.md file
[ ] I have reviewed the CODE_OF_CONDUCT.md file
[ ] I have signed the CLA
[ ] I plan on entering a WWST Certification Request or have entered a request through the WWST Certification console at developer.smartthings.com
[x] Bug fix
[ ] New feature
[ ] Refactor
Checklist
[x] I have performed a self-review of my code
[x] I have commented my code in hard-to-understand areas
[x] I have verified my changes with unit tests
Description of Change
The native handlers for switch on/off and setLevel commands only send zwave commands to the root endpoint, 0, and these devices are overriding the base drivers component to endpoint mapping.
Summary of Completed Tests
@lelandblue this may be something that we want to skip the beta soak time to help users experiencing the issue sooner rather than later.
| gharchive/pull-request | 2024-11-13T20:20:23 | 2025-04-01T04:33:05.530740 | {
"authors": [
"cjswedes"
],
"repo": "SmartThingsCommunity/SmartThingsEdgeDrivers",
"url": "https://github.com/SmartThingsCommunity/SmartThingsEdgeDrivers/pull/1752",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2426447323 | 🛑 MPC Homepage is down
In 37075ed, MPC Homepage (https://www.minorplanetcenter.net) was down:
HTTP code: 0
Response time: 0 ms
Resolved: MPC Homepage is back up in e6182a2 after 4 minutes.
| gharchive/issue | 2024-07-24T02:20:20 | 2025-04-01T04:33:05.568990 | {
"authors": [
"ChrisMoriarty"
],
"repo": "Smithsonian/upptime",
"url": "https://github.com/Smithsonian/upptime/issues/2854",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2435638193 | 🛑 NEOCP is down
In a1563a1, NEOCP (https://minorplanetcenter.net/iau/NEO/toconfirm_tabular.html) was down:
HTTP code: 0
Response time: 0 ms
Resolved: NEOCP is back up in bea5b90 after 4 minutes.
| gharchive/issue | 2024-07-29T14:50:14 | 2025-04-01T04:33:05.571544 | {
"authors": [
"ChrisMoriarty"
],
"repo": "Smithsonian/upptime",
"url": "https://github.com/Smithsonian/upptime/issues/3883",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2440153319 | 🛑 MPC Homepage is down
In 7a84076, MPC Homepage (https://www.minorplanetcenter.net) was down:
HTTP code: 0
Response time: 0 ms
Resolved: MPC Homepage is back up in 7726613 after 4 minutes.
| gharchive/issue | 2024-07-31T14:06:39 | 2025-04-01T04:33:05.574041 | {
"authors": [
"ChrisMoriarty"
],
"repo": "Smithsonian/upptime",
"url": "https://github.com/Smithsonian/upptime/issues/4460",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
127416809 | Fixed a bug where string couldn't be app arguments
lib.load_string failed to properly call Lua's loadstring built-in.
Merged into max-next.
| gharchive/pull-request | 2016-01-19T10:39:19 | 2025-04-01T04:33:05.591949 | {
"authors": [
"aequabit",
"eugeneia"
],
"repo": "SnabbCo/snabbswitch",
"url": "https://github.com/SnabbCo/snabbswitch/pull/717",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
312431787 | UIViewController in TabBarViewController hidden tabbar item.snp.bottom.equalToSuperview() invalid
Bottom
let logoView = OWSeriesLogoView()
view.addSubview(logoView)
logoView.snp.makeConstraints { (maker) in
maker.top.equalTo(menuView.snp.bottom)
maker.left.right.equalToSuperview()
maker.bottom.equalToSuperview()
}
Issue Info
Info
Value
Platform
e.g. ios
Platform Version
e.g. 9.0
SnapKit Version
e.g. 4.0.0
Issue Description
UIViewController in TabBarViewController hidden tabbar item.snp.bottom.equalToSuperview() invalid
There are 44 blanks from the bottom
Not really sure I can help here your sample code is rather incomplete and the description lacking some details.
Are you trying to create auto layout pinned to items in the tab bar? I'm not sure if that is legal in auto layout.
| gharchive/issue | 2018-04-09T08:02:43 | 2025-04-01T04:33:05.595305 | {
"authors": [
"ljs19850321",
"robertjpayne"
],
"repo": "SnapKit/SnapKit",
"url": "https://github.com/SnapKit/SnapKit/issues/505",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2058307562 | Change geographic location in opera proxy for android
Hello, thank you again for this VPN :)
I have a question please, why opera does not protect its VPN i.e. why opera does not put a protection mechanism to prevent the use of its proxy outside of opera browser? or maybe there is ads when using this proxy so opera can finance its vpn ?
I have another question please: when using the Opera browser, the Opera browser automatically chooses the best VPN location for faster browsing, can Standalone Opera VPN client do this by choosing the default location? fastest geographic location automatically?
and also how we can choose the geographical location when we use opera-proxy for android because we cannot add command line arguments to the shortcut like in windows 10 os? Thanks !
Another question please : we know that opera was bought by a Chinese consortium in 2016, I do not accuse this company which is certainly honest but we know that the Chinese communist party likes to control and monitor people and likes to spy on people and imposes on Chinese companies to follow their orders, so is there a security risk when using opera proxy when for example the Chinese communist party orders opera to give out user passwords?
Another quetion please : hola -proxy has option " peer proxies on residental IPs (consumer ISP) ",
I read that the tuxler vpn
https://www.tuxlervpn.com/
allows us to use the real IP addresses of users and if a malicious user takes our IP address and does illegal activity, we can be punished and go to prison....is it the same thing with hola proxy when we use hola -proxy has option " peer proxies on residential IPs" ? hola proxy allows people to use our real ip address ?
I have a question please, why opera does not protect its VPN i.e. why opera does not put a protection mechanism to prevent the use of its proxy outside of opera browser? or maybe there is ads when using this proxy so opera can finance its vpn ?
They actually do protect it from unauthorized usage. There is a login and password required for interaction with their API, and sophisticated digest authentication scheme required to sign requests. But since these credentials are baked into Opera browser, it is possible to analyze compiled machine code and figure out how browser interacts with API to get access. But this is as much as they can do to provide public service in their software without user registration.
I have another question please: when using the Opera browser, the Opera browser automatically chooses the best VPN location for faster browsing, can Standalone Opera VPN client do this by choosing the default location? fastest geographic location automatically?
If someone will implement this, I'll be glad to accept pull request. But in general they offer just three continents (America, Europe, Asia) and it's quite evident which is closer to you.
and also how we can choose the geographical location when we use opera-proxy for android because we cannot add command line arguments to the shortcut like in windows 10 os? Thanks !
If you're running opera-proxy on Android using something like Termux, you can add command line parameters the same way after name of executable.
Another question please : we know that opera was bought by a Chinese consortium in 2016, I do not accuse this company which is certainly honest but we know that the Chinese communist party likes to control and monitor people and likes to spy on people and imposes on Chinese companies to follow their orders, so is there a security risk when using opera proxy when for example the Chinese communist party orders opera to give out user passwords?
Any VPN company, Internet service provider or hosting provider may listen users traffic, regardless Chinese or not. Even European hosting providers follow such orders and even actively interfere with service's traffic to perform man-in-the-middle attack. However, in Opera case such traffic interception is kind of fruitless. Almost all sites are HTTPS now and they won't be able to perform decryption (unless they have TLS certs of that site or cooperating with them certificate authority). Without decryption all they can do is to learn visited domains and your connecting IP address. So I would say you're not risking your passwords more than usual.
Another quetion please : hola -proxy has option " peer proxies on residental IPs (consumer ISP) ",
I read that the tuxler vpn
https://www.tuxlervpn.com/
allows us to use the real IP addresses of users and if a malicious user takes our IP address and does illegal activity, we can be punished and go to prison....is it the same thing with hola proxy when we use hola -proxy has option " peer proxies on residential IPs" ? hola proxy allows people to use our real ip address ?
Right now Hola forwards traffic of free users mostly through datacenter IP addresses because residential traffic is quite expensive. However in general Hola allows the use of residential IP addresses for some cases. Otherwise it will not be possible to access services with geo-restricted video content like Netflix or Hulu, which do not expect user to have datacenter IP address. Such IP sharing was a public concern and caused many complaints: http://adios-hola.org/
hola-proxy does not allow your IP to be used by others. It was specifically one of goals of such alternative implementation. Such function does not exists in hola-proxy and never will be.
Ohhh very very interesting and useful information! Thanks a lot for your precious help and for all efforts to make many great awesome softwares 🥇 💯 !
We suggest that you put a pop-up window that launches automatically when we use one of your softwares to make easily a donation via paypal or buy me a coffee or bitcoin.....when we use one of your magnificent softwares so it will allow people who can makes donations to support these fantastic projects and also display messages to encourage people to share your softwares on social networks like Twitter, Facebook, Snap....it will give even more visibility to this amazing softwares....
I also suggest making a discussion section so you can include these important information and revelations that you so kindly made to me around the internal functioning of opera proxy, hola proxy.....in order to definitively answer any future questions and curiosity of users....thank you again very much for all your efforts to invent such useful softwares 🥇 !
| gharchive/issue | 2023-12-28T11:09:33 | 2025-04-01T04:33:05.611874 | {
"authors": [
"Snawoot",
"trimechee"
],
"repo": "Snawoot/opera-proxy",
"url": "https://github.com/Snawoot/opera-proxy/issues/42",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
599057421 | Replace programming language icons
Used icons of programming languages are outdated in my opinion. I propose replacing them with Material, it will be much nicer to look aesthetically.
If you can send me some examples of icons that could look better than the current icons I can consider replacing them.
What resolution should the icons be in? .png (What is the resolution? 512x512, 256x256, 64x54) ot .svg (vector)? @Snazzah
@dmsass 512x512
For reference, here are the current icons on the application.
I do agree that there lacks any consistency in style, I am open to any changes.
My suggestions
javascript: https://upload.wikimedia.org/wikipedia/commons/thumb/9/99/Unofficial_JavaScript_logo_2.svg/512px-Unofficial_JavaScript_logo_2.svg.png
python: https://upload.wikimedia.org/wikipedia/commons/thumb/c/c3/Python-logo-notext.svg/512px-Python-logo-notext.svg.png
rust: https://www.rust-lang.org/logos/rust-logo-512x512-blk.png
text: https://upload.wikimedia.org/wikipedia/commons/thumb/a/ae/Icon-txt.svg/512px-Icon-txt.svg.png
typescript: https://www.typescriptlang.org/icons/icon-512x512.png
unknown: lets use twemoji cause why not https://discord.com/assets/aef26397c9a6a3afee9c857c5e6f3317.svg
v: https://vlang.io/img/v-logo.png
sass because cool
sass: https://sass-lang.com/assets/img/logos/logo-b6e1ef6e.svg
@botatooo I replaced the V logo and Python logo
The JS and TS logos you provided are a bit less consistent and might clip the text when using the big_icon setting.
I'm considering replacing the ? unknown logo with a blank file icon
Rust is more or less the same as it is already, maybe add a color background like for JS and TS? (Same for Sass)
i’ve found this for javascript, seems to fit with html and css’s design
for typescript idk if we could fix that because that’s the logo they use in their marketing and stuffs
From what I've seen, the first suggestion for the JS logo is indeed the most commonly used icon, but I see no benefit of using it over the better suited logo we currently have that allows it to be cut off a bit at the edges.
my fork has custom consistent icons that i think are really nice :l
if you want i can just zip em up and upload
@TorchedSammy Those icons look great! Could you upload them in a zip?
Here they are, ill probably make a separate repo for convenience
@TorchedSammy we could just make a folder in the repo to add the icons into for anyone that wants to use their own application ID
@Snazzah oh yea, feel free to and for any new icons ya can just ask me if you'd like
@TorchedSammy Also, speaking of your fork, would it be alright if you pushed some changes here? I've been meaning to get this package updated a bit anyways.
@Snazzah yep, I can open another PR with the changes in a while
| gharchive/issue | 2020-04-13T18:35:05 | 2025-04-01T04:33:05.622812 | {
"authors": [
"Calvin-rgb",
"FichteFoll",
"Snazzah",
"TorchedSammy",
"botatooo",
"dmsass"
],
"repo": "Snazzah/SublimeDiscordRP",
"url": "https://github.com/Snazzah/SublimeDiscordRP/issues/49",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2640987615 | ss_mkt Parameter Causing Anomaly Detection in DuckDuckGo Results
Description
When using DuckDuckGo's API for querying, adding the ss_mkt parameter seems to trigger anomaly detection. Specifically, if ss_mkt=FR (or any other country code variant) is appended to the request URL, the response includes a script with window.execDeep = function() {DDG.deep.anomalyDetectionBlock({...})}, which appears to be blocking or altering the query response.
Steps to Reproduce
Use the following working URL to perform a query (e.g., what is weather today):
https://links.duckduckgo.com/d.js?q=what is weather today&kl=wt-wt&l=us-en&p=1&s=0&a=h_&dl=en&ct=FR&df=m&vqd=4-130790525484318490675696036619647809147&bing_market=en-US&p_ent=&dp=a2VtvYQbwBUAnJjgW_GROFNV_mc6SfdjMK9PLqH42pqZ3fsCcOsNuDJagf0YUf4NazjDeNsiE8uQBCPRHf0rDV9gnMMek-we5JwQZ87pMeFh4-7e5KVEkLAYafYyGqW7.FOFBUBx1to6nRs7UUN6QSA&sjs=forecast&perf_id=2d03f98f2d7a55da&parent_perf_id=7fe2f5d308add94a&host_region=eun&sp=1&dfrsp=1&bpa=1&baa=1&bcca=1&btaa=1&wrap=1&aps=0&bccaexp=b&biaexp=b&btaaexp=b&eclsexp=b&litexp=b&msvrtexp=b
Add the ss_mkt parameter, e.g., ss_mkt=FR, to the URL:
https://links.duckduckgo.com/d.js?q=what is weather today&kl=wt-wt&l=us-en&p=1&s=0&a=h_&dl=en&ct=FR&df=m&vqd=4-130790525484318490675696036619647809147&bing_market=en-US&p_ent=&dp=a2VtvYQbwBUAnJjgW_GROFNV_mc6SfdjMK9PLqH42pqZ3fsCcOsNuDJagf0YUf4NazjDeNsiE8uQBCPRHf0rDV9gnMMek-we5JwQZ87pMeFh4-7e5KVEkLAYafYyGqW7.FOFBUBx1to6nRs7UUN6QSA&sjs=forecast&perf_id=2d03f98f2d7a55da&parent_perf_id=7fe2f5d308add94a&host_region=eun&sp=1&dfrsp=1&bpa=1&baa=1&bcca=1&btaa=1&wrap=1&aps=0&bccaexp=b&biaexp=b&btaaexp=b&eclsexp=b&litexp=b&msvrtexp=b&ss_mkt=FR
3. Observe the response which includes window.execDeep = function() {DDG.deep.anomalyDetectionBlock({...})} instead of the expected data.
Expected Behavior
Adding the ss_mkt parameter should not trigger anomaly detection, and the query should return results as expected.
Actual Behavior
Adding ss_mkt=FR results in the DuckDuckGo response containing a script with DDG.deep.anomalyDetectionBlock, which seems to block or modify the response.
Environment
Query Tool: DuckDuckGo API
Affected Parameter: ss_mkt
Additional Information
This anomaly is observed specifically when setting the ss_mkt parameter to a value such as FR. Other values may produce similar results.
Request
If the current DuckDuckGo API does not support the ss_mkt parameter, please consider removing or disabling it to avoid triggering anomaly detection blocks.
I have the same issue using langchain and the duck duck search library, please consider removing the parameter from the code.
Temporary fix
import { search, SafeSearchType } from 'duck-duck-scrape';
// import * as DDG from 'duck-duck-scrape';
const searchResults = await search('node.js', {
safeSearch: SafeSearchType.STRICT
}, {
uri_modifier: (rawUrl) => {
const url = new URL(rawUrl);
url.searchParams.delete("ss_mkt");
return url.toString();
}
);
Hello,
Is there any reference where the have mentioned about replacing ss_mkt by bing_market?.
For me replacing ss_mkt by bing_market flag does not give any answer.
Thank you.
Looking at doing a DDG request now does seem the query parameter was replaced with bing_market, also assuming the values were somewhat changed since the default is "us" with what im receiving being "en-US". Don't know of other changes but that seems to be the most obvious one.
@flaviogoncalves If you just make a PR and summarize your changes I can review and merge
It seems that new version is not fixing the issue. I can do one query per like ~15-30 minutes
Not sure how else to fix this other than adding common browser headers to each request, I don't know what else I really can do to solve this. You might just need to deal with proxies whenever you hit an anomaly.
| gharchive/issue | 2024-11-07T13:07:34 | 2025-04-01T04:33:05.632406 | {
"authors": [
"Snazzah",
"Tomas2D",
"dzek69",
"flaviogoncalves",
"sarfudheen"
],
"repo": "Snazzah/duck-duck-scrape",
"url": "https://github.com/Snazzah/duck-duck-scrape/issues/140",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
376992691 | Update kubernetes to newer version
apiVersion no longer beta
update to using ClusterIP for newer networking
@jondubois Hey John,
I'll make those changes shortly. I only add the server and deployment to the name to make it easier for me to remember what's going on in our production environment. There is no need to do it here. Sorry about that. As far as ClusterIP is concerned, it's the way you're supposed to do internal networking in Kubernetes. The other kind is going to be deprecated like loadbalancer is. That's at least my understanding. Thanks!
oops... wrong account.. sorry lol
@jondubois I fixed those changes you asked about. Thanks!
@jondubois hey man I was thinking if the docker images shared the same persistent storage claim could you replicate state servers and remove the single point of failure? You could use ClusterIP to handle the communication. What do you think?
scc-state doesn't use persistent storage; it's all in memory. scc-state is not really a single point of failure because SCC can keep working without it. There will only be partial disruption of the service if both scc-state and an scc-broker fail at the exact same time; and even if this unlikely scenario were to happen, the disruption should only last a few seconds.
I will update the scc-guide to remove the phrase 'single point of failure' because it sound like it's a problem but it's not. You always need to make a trade-off between consistency or availability and SCC leans slightly more towards consistency but it's still very resilient in terms of availability.
@happilymarrieddad Your PR looks good now. It might take a bit of time to merge because I want to test it first.
@jondubois no worries
@happilymarrieddad Thanks for this contribution :)
Sorry it took so long; I was busy and didn't get the time to test it until recently.
After testing it on GKE, I changed it to expose the socketcluster (worker) service using NodePort otherwise it doesn't work on GKE. This could also be an issue on other platforms.
@jondubois cool thanks! Yea, I don't know.. I use kops and AWS for production. I played around a little bit on GKE but it was too much magic for me. Thanks!
| gharchive/pull-request | 2018-11-02T22:57:53 | 2025-04-01T04:33:05.656541 | {
"authors": [
"0xsegfault",
"happilymarrieddad",
"jondubois"
],
"repo": "SocketCluster/socketcluster",
"url": "https://github.com/SocketCluster/socketcluster/pull/453",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2152763903 | Add option for skipping instrument rescaling when not available
As described in #47.
This PR adds an option error_on_missing to read_nda (and read) that allows the parser to continue even when instrument scaling is not possible due to to a missing multiplier. The current behaviour is preserved by the default value of error_on_missing = True.
This allows the package to be used more easily downstream without needing to wait for new multipliers to be added and released.
Before merging this, there is no test data available to contributors (presumably you have some somewhere!) so consider this PR as untested! (It works for me with my own data, of course...) -- see #48 for my thoughts on how best to do this.
See #47
| gharchive/pull-request | 2024-02-25T13:22:15 | 2025-04-01T04:33:05.700249 | {
"authors": [
"ml-evs"
],
"repo": "Solid-Energy-Systems/NewareNDA",
"url": "https://github.com/Solid-Energy-Systems/NewareNDA/pull/49",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
293812820 | Microsoft Core Dependency Injection as a DI for SolrNet.Cloud
Hi, I have asked question in SO https://stackoverflow.com/questions/48564227/can-i-use-microsoft-core-dependency-injection-with-solrnet-cloud-core about ability to use Microsoft Core Dependency Injection as a DI for SolrNet.Cloud
But I can assume that there is no support of Microsoft DI in Solr.Net.Cloud.
Are there any plans to introduce support of Microsoft Core DI in Solr.Net.Cloud in future?
@Fatalityap Sure, we want to add also Microsoft DI support for SolrNet.Cloud . Please feel free to start the work on it and help us getting it supported.
Ok, I have talked with my management. Maybe we can write support for it. At least my management do not mind for now.
Did this ever get progressed? I am facing the same issue.
Also looking for Microsoft Core DI support. Any news on this?
| gharchive/issue | 2018-02-02T08:44:27 | 2025-04-01T04:33:05.708848 | {
"authors": [
"Fatalityap",
"Wintermute79",
"gjunge",
"technicaldogsbody"
],
"repo": "SolrNet/SolrNet",
"url": "https://github.com/SolrNet/SolrNet/issues/381",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1272463521 | Non-recoverable error on launch
Mod list: {big ones K2+SE}
mod-list.zip
First save: {file was too big to upload on github}
https://drive.google.com/file/d/1eyUhUnJWvyoELBNjYKQ-s0_fZ1OVQ6os/view?usp=sharing
Error message:
Second save: {same mods}
FILIP MOD 002.zip
Error message:
Error is shown on both saves seconds after loading of map , only new mod that was added to modpack during playthrough , nothing removed ...
Edit: Only thing i can think of that only mod that is somewhat modying something about trains is LTN but that dont do anything with signals only with train stations
Fixed by commits 61ba57df22043e855c62db69909f7dd5b5e960af and c2000a77fc919b98ae3aee9961c2c5b84d4444a0 (version 1.1.1). This rail network is crazy, and there's a lot of weird things going on. (The mod also identified plenty of issues :))
The way this railway network is made causes a lot of exploration to get enough rails in the graph. This makes the update take a few seconds (!). Makes it also a good test-case for potential optimization. Al;so makes me think that a non-realtime (single use) annotation (through some button) could be useful for such large and interconnected networks.
Very nice test case for the mod overall, I'll be using it a lot. Thank you for reporting this.
| gharchive/issue | 2022-06-15T16:18:22 | 2025-04-01T04:33:05.796901 | {
"authors": [
"Sopel97",
"TheAsKo"
],
"repo": "Sopel97/railway_signalling_overseer",
"url": "https://github.com/Sopel97/railway_signalling_overseer/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
951014372 | Reading log guidance
Hi I wanted to test various mitigation settings to compare performance and was wondering what was the best way going about to do that from inspecting logs, given a line sequence like:
2021-07-22 20:29:10.780 AnimationLockLatencyHandler 16f4: C2S_ActionRequest(0175): actionId=08cf sequence=0e22 delay=1358ms
2021-07-22 20:29:10.963 AnimationLockLatencyHandler 16f4: S2C_ActionEffect(0102): actionId=08cf sourceSequence=0e22 rtt=188ms pingLatency=33ms delay=75ms wait=600ms->487ms next=19:29:11.4507252
2021-07-22 20:29:11.451 AnimationLockLatencyHandler 16f4: S2C_ActionEffect(0102): actionId=0007 sourceSequence=0000 serverOriginated wait=108ms->100ms next=19:29:11.5517470
2021-07-22 20:29:12.214 AnimationLockLatencyHandler 16f4: C2S_ActionRequest(0175): actionId=1e8e8c sequence=0e23 delay=666ms
Also sometimes the wait value is not 600, even for non-single digit actionids (which I read on a comment somewhere to ignore)
2021-07-22 20:28:30.228 AnimationLockLatencyHandler 16f4: S2C_ActionEffect(0102): actionId=4976 sourceSequence=0e01 rtt=204ms pingLatency=34ms delay=75ms wait=349ms->220ms next=19:28:30.4481509
2021-07-22 20:28:30.510 AnimationLockLatencyHandler 16f4: C2S_ActionRequest(0175): actionId=4975 sequence=0e02 delay=76ms
What values are we trying to minimize here? is delay=X what the program "adjusts" our latency to be in terms of performing the actions locally?
actionId states which action you used, including auto attacks, weaponskills, spells, and abilities. When auto attack happens, wait value may be any value between previous action's animation lock duration and 100ms. delay is the simulated server response time. Note that it's not ping; the server may take longer than usual even when your ping itself is stable, like when you were moving around.
Thanks for responding. So in essence if we want to see whether the options such as Use Delay Detection and Use Latency Correction are better or worse, we want to minimize the wait value? (to ensure the lowest amount of time between animation lock and next action)?
Yes, that's correct.
| gharchive/issue | 2021-07-22T20:13:51 | 2025-04-01T04:33:05.800275 | {
"authors": [
"Soreepeong",
"perks"
],
"repo": "Soreepeong/XivAlexander",
"url": "https://github.com/Soreepeong/XivAlexander/issues/117",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2554103335 | 🛑 Marcel Web is down
In ecc1b3b, Marcel Web ($MARCEL_WEB) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Marcel Web is back up in 82c42db after 3 hours, 1 minute.
| gharchive/issue | 2024-09-28T08:14:05 | 2025-04-01T04:33:05.844339 | {
"authors": [
"Sundypha"
],
"repo": "Source-Graphics-GmbH/upptime",
"url": "https://github.com/Source-Graphics-GmbH/upptime/issues/1122",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1138130860 | register
@JuliaRegistrator register
Error while trying to register: "Tag with name v0.4.0 already exists and points to a different commit"
| gharchive/issue | 2022-02-15T03:28:51 | 2025-04-01T04:33:05.861640 | {
"authors": [
"JuliaRegistrator",
"shinaoka"
],
"repo": "SpM-lab/SparseIR.jl",
"url": "https://github.com/SpM-lab/SparseIR.jl/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1887268758 | 🛑 Yiffed Main is down
In 75f4c9a, Yiffed Main (https://yiffed.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Yiffed Main is back up in 2d7b0a8 after 10 minutes.
| gharchive/issue | 2023-09-08T09:08:45 | 2025-04-01T04:33:05.927751 | {
"authors": [
"SparksTheFolf"
],
"repo": "SparksTheFolf/STF-Uptime-Status",
"url": "https://github.com/SparksTheFolf/STF-Uptime-Status/issues/246",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1901358194 | 🛑 Yiffed Main is down
In f29b1ae, Yiffed Main (https://yiffed.net/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Yiffed Main is back up in 9ea6436 after 36 minutes.
| gharchive/issue | 2023-09-18T16:49:50 | 2025-04-01T04:33:05.930091 | {
"authors": [
"SparksTheFolf"
],
"repo": "SparksTheFolf/STF-Uptime-Status",
"url": "https://github.com/SparksTheFolf/STF-Uptime-Status/issues/897",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2177601144 | Using Flags
Hi
Is there a way to use Enum Lookup Tables with Flags? I tried (of course) already and even though it works with an individual Flag, it fails when using multiple Flags as the foreign key reference doesn't handle bitwise added values.
So is there any way to make this work?
Possibly you would create overload for this method of "Flagged" case: EnumLookupExtension.cs#L129
Now it creates row per enum. You should create row per each combination of flags. Then any bitwise would be found in foreign table.
I will try to go that way. Just didn't want to "invent the wheel" if it was already there and I just missed it.
I don't claim credit for the GetCombinations method (google was my friend), but in case someone else is searching this then here's my change to GetEnumData and the added GetCombinations method:
private static object[] GetEnumData(Type enumType, Type concreteType, bool useNumberLookup, bool usesDescription,
Dictionary<int, string> enumValueDescriptions)
{
if (enumType.IsDefined(typeof(FlagsAttribute), false))
{
return GetCombinations((Enum)Activator.CreateInstance(enumType))
.Select(x =>
{
var instance = Activator.CreateInstance(concreteType);
concreteType
.GetProperty("Id")
?.SetValue(instance, x.Key);
if (useNumberLookup)
{
concreteType
.GetProperty("Name")
?.SetValue(instance, x.Value.ToString());
}
if (usesDescription)
{
concreteType
.GetProperty("Description")
?.SetValue(instance, enumValueDescriptions[x.Key]);
}
return instance;
})
.ToArray();
}
return Enum
.GetValues(enumType)
.OfType<object>()
.Select(x =>
{
var instance = Activator.CreateInstance(concreteType);
concreteType
.GetProperty("Id")
?.SetValue(instance, x);
if (useNumberLookup)
{
concreteType
.GetProperty("Name")
?.SetValue(instance, x.ToString());
}
if (usesDescription)
{
concreteType
.GetProperty("Description")
?.SetValue(instance, enumValueDescriptions[(int)x]);
}
return instance;
})
.ToArray();
}
private static Dictionary<int, string> GetCombinations(this Enum concreteEnum)
{
var fields = concreteEnum
.GetType()
.GetFields()
.Where(f => f.Name != "value__")
.DistinctBy(f => Convert.ToInt32(f.GetRawConstantValue()));
var result = fields.ToDictionary(f => Convert.ToInt32(f.GetRawConstantValue()), f => f.Name);
var max = Enum
.GetValues(concreteEnum.GetType())
.Cast<int>()
.Max();
var upperBound = max * 2;
for (var i = 0; i <= upperBound; i += 2)
{
var s = Convert
.ToString(i, 2)
.PadLeft(Math.Abs(i - max), '0');
var bits = s
.Select(chs => chs == '1')
.Reverse()
.ToArray();
if (!result.ContainsKey(i))
{
var newComb = string.Empty;
for (var j = 1; j < bits.Length; j++)
{
var idx = 1 << j;
if (bits[j] && result.TryGetValue(idx, out var value))
{
newComb = newComb + value + " | ";
}
}
newComb = newComb.Trim([' ', '|']);
if (!result.ContainsValue(newComb) && !string.IsNullOrEmpty(newComb))
{
result.Add(i, newComb);
}
}
}
return result;
}
| gharchive/issue | 2024-03-10T08:28:18 | 2025-04-01T04:33:05.934355 | {
"authors": [
"AllanMichaelsen",
"john-bartu"
],
"repo": "SpatialFocus/EntityFrameworkCore.Extensions",
"url": "https://github.com/SpatialFocus/EntityFrameworkCore.Extensions/issues/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
522726953 | Fail during make install
Environment
System: window10 WSL Centos7
scipy==1.1.0
install pillow==5.1.0
Problem
I have already chanege the 'CV_BGR2GRAY' to 'COLOR_BGR2GRAY' in file st_tissue_recognition/src/util.cpp
then cmake and make install again
but still get Error in step make install
Error Information:
[ 87%] Linking CXX executable recognize-tissue
/mnt/e/BioInfo/APP/Anaconda3/bin/../lib/gcc/x86_64-conda_cos6-linux-gnu/7.3.0/../../../../x86_64-conda_cos6-linux-gnu/bin/ld: warning: libz.so.1, needed by /usr/local/lib64/libopencv_core.so.4.1.2, not found (try using -rpath or -rpath-link)
/mnt/e/BioInfo/APP/Anaconda3/bin/../lib/gcc/x86_64-conda_cos6-linux-gnu/7.3.0/../../../../x86_64-conda_cos6-linux-gnu/bin/ld: warning: libjpeg.so.9, needed by /usr/local/lib64/libopencv_imgcodecs.so.4.1.2, not found (try using -rpath or -rpath-link)
/mnt/e/BioInfo/APP/Anaconda3/bin/../lib/gcc/x86_64-conda_cos6-linux-gnu/7.3.0/../../../../x86_64-conda_cos6-linux-gnu/bin/ld: warning: libpng15.so.15, needed by /usr/local/lib64/libopencv_imgcodecs.so.4.1.2, not found (try using -rpath or -rpath-link)
/mnt/e/BioInfo/APP/Anaconda3/bin/../lib/gcc/x86_64-conda_cos6-linux-gnu/7.3.0/../../../../x86_64-conda_cos6-linux-gnu/bin/ld: CMakeFiles/recognize-tissue.dir/main.cpp.o: in function `main':
main.cpp:(.text+0xfd): undefined reference to `cv::imread(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, int)'
.....
/mnt/e/BioInfo/APP/Anaconda3/bin/../lib/gcc/x86_64-conda_cos6-linux-gnu/7.3.0/../../../../x86_64-conda_cos6-linux-gnu/bin/ld: /usr/local/lib64/libopencv_imgcodecs.so.4.1.2: undefined reference to `png_set_filter@PNG15_0'
collect2: error: ld returned 1 exit status
make[2]: *** [src/CMakeFiles/recognize-tissue.dir/build.make:100: src/recognize-tissue] Error 1
make[1]: *** [CMakeFiles/Makefile2:169: src/CMakeFiles/recognize-tissue.dir/all] Error 2
I think it may cause by cross-compiled which I'm not so familiar with。
The only one thing that I'm sure is the openCV has install successfuly, because:
(base) [Robin@SC-201910280935 python-module]$ python
Python 3.7.4 (default, Aug 13 2019, 20:35:49)
[GCC 7.3.0] :: Anaconda, Inc. on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import cv2
>>> cv2.__version__
'4.1.1'
>>>
This is the first time for me to use issues in Github,and I am extremely thristy to your reply to my message~Thx~
I find a solution:Change the the 'CV_BGR2GRAY' to 'COLOR_BGR2GRAY' in file st_tissue_recognition/src/util.cpp, cd into the /python-mudule/
then
python setup.py build
python setup.py install
it works for me
Thanks for reporting back the solution!
| gharchive/issue | 2019-11-14T09:24:03 | 2025-04-01T04:33:05.939183 | {
"authors": [
"BioAIEvolu",
"ludvb"
],
"repo": "SpatialTranscriptomicsResearch/st_tissue_recognition",
"url": "https://github.com/SpatialTranscriptomicsResearch/st_tissue_recognition/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2423920112 | BED-4598: added group entities to open api spec
Description
Added endpoint definitions for group entity endpoints to the open api spec
Motivation and Context
This PR addresses: BED-4141 / BED-4598
Types of changes
Chore (a change that does not modify the application functionality)
Checklist:
[x] I have met the contributing prerequisites
Assigned myself to this PR
Added the appropriate labels
Associated an issue: https://github.com/SpecterOps/BloodHound/issues/672
Read the Contributing guide: https://github.com/SpecterOps/BloodHound/wiki/Contributing
[x] I have ensured that related documentation is up-to-date
Open API docs
Code comments (GoDocs / JSDocs)
[x] I have followed proper test practices
Added/updated tests to cover my changes
All new and existing tests passed
I also added a README with some helpful information about paths, and removed a couple duplicate files
| gharchive/pull-request | 2024-07-22T23:17:54 | 2025-04-01T04:33:05.945519 | {
"authors": [
"sircodemane"
],
"repo": "SpecterOps/BloodHound",
"url": "https://github.com/SpecterOps/BloodHound/pull/724",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
210292583 | Poor performance with modsecurity enabled
As suggested within issues https://github.com/SpiderLabs/ModSecurity/issues/1318 I'll open a new one here. With MS enabled with the current codebase the performance test with "ab" from apache2-utils is not going over 300req/s.
With a virtual machine on Virtualbox with 4 cores and 4gb ram:
root@nginx:~# ab -n 10000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 1000 requests
Completed 2000 requests
Completed 3000 requests
Completed 4000 requests
Completed 5000 requests
Completed 6000 requests
Completed 7000 requests
Completed 8000 requests
Completed 9000 requests
Completed 10000 requests
Finished 10000 requests
Server Software: nginx/1.11.9
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 612 bytes
Concurrency Level: 20
Time taken for tests: 73.444 seconds
Complete requests: 10000
Failed requests: 0
Keep-Alive requests: 9909
Total transferred: 8499545 bytes
HTML transferred: 6120000 bytes
Requests per second: 136.16 [#/sec] (mean)
Time per request: 146.889 [ms] (mean)
Time per request: 7.344 [ms] (mean, across all concurrent requests)
Transfer rate: 113.02 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.0 0 0
Processing: 14 147 31.7 141 1308
Waiting: 14 147 31.7 141 1308
Total: 14 147 31.7 141 1308
Percentage of the requests served within a certain time (ms)
50% 141
66% 147
75% 151
80% 154
90% 166
95% 188
98% 207
99% 232
100% 1308 (longest request)
And without:
root@nginx:~# /opt/nginx/sbin/nginx
root@nginx:~# ab -n 10000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 1000 requests
Completed 2000 requests
Completed 3000 requests
Completed 4000 requests
Completed 5000 requests
Completed 6000 requests
Completed 7000 requests
Completed 8000 requests
Completed 9000 requests
Completed 10000 requests
Finished 10000 requests
Server Software: nginx/1.11.9
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 612 bytes
Concurrency Level: 20
Time taken for tests: 0.197 seconds
Complete requests: 10000
Failed requests: 0
Keep-Alive requests: 9909
Total transferred: 8499545 bytes
HTML transferred: 6120000 bytes
Requests per second: 50755.75 [#/sec] (mean)
Time per request: 0.394 [ms] (mean)
Time per request: 0.020 [ms] (mean, across all concurrent requests)
Transfer rate: 42128.99 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.0 0 0
Processing: 0 0 0.3 0 5
Waiting: 0 0 0.3 0 5
Total: 0 0 0.3 0 5
Percentage of the requests served within a certain time (ms)
50% 0
66% 0
75% 0
80% 0
90% 0
95% 1
98% 1
99% 2
100% 5 (longest request)
I tested this also with Nginx+ but then the tool stucks a 90%, no mather how many requests (20 cores and 64GB ram):
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 100 requests
Completed 200 requests
Completed 300 requests
Completed 400 requests
Completed 500 requests
Completed 600 requests
Completed 700 requests
Completed 800 requests
Completed 900 requests
apr_pollset_poll: The timeout specified has expired (70007)
Total of 999 requests completed
But with just 100 it's ok:
Server Software: nginx
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 162 bytes
Concurrency Level: 20
Time taken for tests: 0.243 seconds
Complete requests: 100
Failed requests: 0
Non-2xx responses: 100
Keep-Alive requests: 100
Total transferred: 31000 bytes
HTML transferred: 16200 bytes
Requests per second: 410.71 [#/sec] (mean)
Time per request: 48.696 [ms] (mean)
Time per request: 2.435 [ms] (mean, across all concurrent requests)
Transfer rate: 124.34 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.7 0 3
Processing: 3 42 34.6 58 99
Waiting: 3 42 34.6 58 99
Total: 3 42 34.6 60 99
Percentage of the requests served within a certain time (ms)
50% 60
66% 66
75% 70
80% 74
90% 86
95% 92
98% 97
99% 99
100% 99 (longest request)
Perhaps others can post their results too.
Ah, I also tried to disable logging but it doens't affect the performance at all.
Ok, I disabled modsec_audit.log and now it seems better the more requests I send :)
`root@waf-1-a-02:~# ab -n 100 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient).....done
Server Software: nginx
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 162 bytes
Concurrency Level: 20
Time taken for tests: 0.175 seconds
Complete requests: 100
Failed requests: 0
Non-2xx responses: 100
Keep-Alive requests: 100
Total transferred: 31000 bytes
HTML transferred: 16200 bytes
Requests per second: 570.71 [#/sec] (mean)
Time per request: 35.044 [ms] (mean)
Time per request: 1.752 [ms] (mean, across all concurrent requests)
Transfer rate: 172.77 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.7 0 3
Processing: 4 29 20.5 25 76
Waiting: 3 29 20.5 25 76
Total: 4 30 20.6 26 77
Percentage of the requests served within a certain time (ms)
50% 26
66% 44
75% 48
80% 51
90% 59
95% 65
98% 72
99% 77
100% 77 (longest request)
root@waf-1-a-02:~# ab -n 1000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 100 requests
Completed 200 requests
Completed 300 requests
Completed 400 requests
Completed 500 requests
Completed 600 requests
Completed 700 requests
Completed 800 requests
Completed 900 requests
Completed 1000 requests
Finished 1000 requests
Server Software: nginx
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 162 bytes
Concurrency Level: 20
Time taken for tests: 1.458 seconds
Complete requests: 1000
Failed requests: 0
Non-2xx responses: 1000
Keep-Alive requests: 996
Total transferred: 309980 bytes
HTML transferred: 162000 bytes
Requests per second: 685.68 [#/sec] (mean)
Time per request: 29.168 [ms] (mean)
Time per request: 1.458 [ms] (mean, across all concurrent requests)
Transfer rate: 207.57 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.2 0 3
Processing: 2 29 18.6 31 84
Waiting: 2 29 18.6 31 84
Total: 2 29 18.6 33 85
Percentage of the requests served within a certain time (ms)
50% 33
66% 39
75% 41
80% 51
90% 54
95% 54
98% 62
99% 72
100% 85 (longest request)
root@waf-1-a-02:~# ab -n 2000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 200 requests
Completed 400 requests
Completed 600 requests
Completed 800 requests
Completed 1000 requests
Completed 1200 requests
Completed 1400 requests
Completed 1600 requests
Completed 1800 requests
Completed 2000 requests
Finished 2000 requests
Server Software: nginx
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 162 bytes
Concurrency Level: 20
Time taken for tests: 2.007 seconds
Complete requests: 2000
Failed requests: 0
Non-2xx responses: 2000
Keep-Alive requests: 1988
Total transferred: 619940 bytes
HTML transferred: 324000 bytes
Requests per second: 996.74 [#/sec] (mean)
Time per request: 20.065 [ms] (mean)
Time per request: 1.003 [ms] (mean, across all concurrent requests)
Transfer rate: 301.72 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.2 0 3
Processing: 2 20 23.0 3 94
Waiting: 2 20 23.0 3 94
Total: 3 20 23.0 3 94
Percentage of the requests served within a certain time (ms)
50% 3
66% 43
75% 44
80% 44
90% 58
95% 60
98% 61
99% 70
100% 94 (longest request)
root@waf-1-a-02:~# ab -n 3000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 300 requests
Completed 600 requests
Completed 900 requests
Completed 1200 requests
Completed 1500 requests
Completed 1800 requests
Completed 2100 requests
Completed 2400 requests
Completed 2700 requests
Completed 3000 requests
Finished 3000 requests
Server Software: nginx
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 162 bytes
Concurrency Level: 20
Time taken for tests: 2.702 seconds
Complete requests: 3000
Failed requests: 0
Non-2xx responses: 3000
Keep-Alive requests: 2981
Total transferred: 929905 bytes
HTML transferred: 486000 bytes
Requests per second: 1110.42 [#/sec] (mean)
Time per request: 18.011 [ms] (mean)
Time per request: 0.901 [ms] (mean, across all concurrent requests)
Transfer rate: 336.13 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.1 0 3
Processing: 2 18 20.5 3 90
Waiting: 2 18 20.5 3 90
Total: 2 18 20.5 3 91
Percentage of the requests served within a certain time (ms)
50% 3
66% 13
75% 42
80% 43
90% 44
95% 58
98% 59
99% 60
100% 91 (longest request)
root@waf-1-a-02:~# ab -n 10000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 1000 requests
Completed 2000 requests
Completed 3000 requests
Completed 4000 requests
Completed 5000 requests
Completed 6000 requests
Completed 7000 requests
Completed 8000 requests
Completed 9000 requests
Completed 10000 requests
Finished 10000 requests
Server Software: nginx
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 162 bytes
Concurrency Level: 20
Time taken for tests: 5.262 seconds
Complete requests: 10000
Failed requests: 0
Non-2xx responses: 10000
Keep-Alive requests: 9913
Total transferred: 3099565 bytes
HTML transferred: 1620000 bytes
Requests per second: 1900.24 [#/sec] (mean)
Time per request: 10.525 [ms] (mean)
Time per request: 0.526 [ms] (mean, across all concurrent requests)
Transfer rate: 575.19 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.1 0 3
Processing: 2 11 16.2 3 91
Waiting: 2 10 16.2 3 91
Total: 2 11 16.2 3 91
Percentage of the requests served within a certain time (ms)
50% 3
66% 3
75% 5
80% 6
90% 43
95% 45
98% 57
99% 59
100% 91 (longest request)`
Hi @mimugmail, @zimmerle,
I have created a separate repository that can be used to easily spin up reproducible environments (thanks goes to Vagrant and SaltStack) suitable for running benchmarks against nginx with libmodsecurity connector module:
https://github.com/defanator/modsecurity-performance
It also can be extended to include more sophisticated things, if anyone would like to.
For now I'm seeing more than 4x performance drop when libmodsecurity is turned on without any actual rules configured, and more than 170x performance drop when libmodsecurity is turned on with OWASP CRS v3.0.0 loaded (I'm not including any absolute numbers here for obvious reasons; you can find my own results in the repo - I'll add a bit more details about the hardware used as well).
Main goal is to give an ability for anyone to easily reproduce the tests against particular combination of libmodsecurity, ModSecurity-nginx connector module, and nginx.
@mimugmail - can I ask you to share your modsecurity.conf you were using while testing?
(if it differs from the default one here - https://github.com/SpiderLabs/ModSecurity/blob/master/modsecurity.conf-recommended)
`# -- Rule engine initialization ----------------------------------------------
Enable ModSecurity, attaching it to every transaction. Use detection
only to start with, because that minimises the chances of post-installation
disruption.
SecRuleEngine On
-- Request body handling ---------------------------------------------------
Allow ModSecurity to access request bodies. If you don't, ModSecurity
won't be able to see any POST parameters, which opens a large security
hole for attackers to exploit.
SecRequestBodyAccess On
Enable XML request body parser.
Initiate XML Processor in case of xml content-type
SecRule REQUEST_HEADERS:Content-Type "(?:text|application)/xml"
"id:'200000',phase:1,t:none,t:lowercase,pass,nolog,ctl:requestBodyProcessor=XML"
Enable JSON request body parser.
Initiate JSON Processor in case of JSON content-type; change accordingly
if your application does not use 'application/json'
SecRule REQUEST_HEADERS:Content-Type "application/json"
"id:'200001',phase:1,t:none,t:lowercase,pass,nolog,ctl:requestBodyProcessor=JSON"
Maximum request body size we will accept for buffering. If you support
file uploads then the value given on the first line has to be as large
as the largest file you are willing to accept. The second value refers
to the size of data, with files excluded. You want to keep that value as
low as practical.
SecRequestBodyLimit 536870912
SecRequestBodyNoFilesLimit 536870912
Store up to 128 KB of request body data in memory. When the multipart
parser reachers this limit, it will start using your hard disk for
storage. That is slow, but unavoidable.
SecRequestBodyInMemoryLimit 134217728
What do do if the request body size is above our configured limit.
Keep in mind that this setting will automatically be set to ProcessPartial
when SecRuleEngine is set to DetectionOnly mode in order to minimize
disruptions when initially deploying ModSecurity.
SecRequestBodyLimitAction ProcessPartial
Verify that we've correctly processed the request body.
As a rule of thumb, when failing to process a request body
you should reject the request (when deployed in blocking mode)
or log a high-severity alert (when deployed in detection-only mode).
SecRule REQBODY_ERROR "!@eq 0"
"id:'200002', phase:2,t:none,log,deny,status:400,msg:'Failed to parse request body.',logdata:'%{reqbody_error_msg}',severity:2"
By default be strict with what we accept in the multipart/form-data
request body. If the rule below proves to be too strict for your
environment consider changing it to detection-only. You are encouraged
not to remove it altogether.
SecRule MULTIPART_STRICT_ERROR "!@eq 0"
"id:'200003',phase:2,t:none,log,deny,status:400,
msg:'Multipart request body failed strict validation:
PE %{REQBODY_PROCESSOR_ERROR},
BQ %{MULTIPART_BOUNDARY_QUOTED},
BW %{MULTIPART_BOUNDARY_WHITESPACE},
DB %{MULTIPART_DATA_BEFORE},
DA %{MULTIPART_DATA_AFTER},
HF %{MULTIPART_HEADER_FOLDING},
LF %{MULTIPART_LF_LINE},
SM %{MULTIPART_MISSING_SEMICOLON},
IQ %{MULTIPART_INVALID_QUOTING},
IP %{MULTIPART_INVALID_PART},
IH %{MULTIPART_INVALID_HEADER_FOLDING},
FL %{MULTIPART_FILE_LIMIT_EXCEEDED}'"
Did we see anything that might be a boundary?
SecRule MULTIPART_UNMATCHED_BOUNDARY "!@eq 0"
"id:'200004',phase:2,t:none,log,deny,msg:'Multipart parser detected a possible unmatched boundary.'"
PCRE Tuning
We want to avoid a potential RegEx DoS condition
SecPcreMatchLimit 1000
SecPcreMatchLimitRecursion 1000
Some internal errors will set flags in TX and we will need to look for these.
All of these are prefixed with "MSC_". The following flags currently exist:
MSC_PCRE_LIMITS_EXCEEDED: PCRE match limits were exceeded.
SecRule TX:/^MSC_/ "!@streq 0"
"id:'200005',phase:2,t:none,deny,msg:'ModSecurity internal error flagged: %{MATCHED_VAR_NAME}'"
-- Response body handling --------------------------------------------------
Allow ModSecurity to access response bodies.
You should have this directive enabled in order to identify errors
and data leakage issues.
Do keep in mind that enabling this directive does increases both
memory consumption and response latency.
SecResponseBodyAccess On
Which response MIME types do you want to inspect? You should adjust the
configuration below to catch documents but avoid static files
(e.g., images and archives).
SecResponseBodyMimeType text/plain text/html text/xml
Buffer response bodies of up to 512 KB in length.
SecResponseBodyLimit 524288
What happens when we encounter a response body larger than the configured
limit? By default, we process what we have and let the rest through.
That's somewhat less secure, but does not break any legitimate pages.
SecResponseBodyLimitAction ProcessPartial
-- Filesystem configuration ------------------------------------------------
The location where ModSecurity stores temporary files (for example, when
it needs to handle a file upload that is larger than the configured limit).
This default setting is chosen due to all systems have /tmp available however,
this is less than ideal. It is recommended that you specify a location that's private.
SecTmpDir /tmp/
The location where ModSecurity will keep its persistent data. This default setting
is chosen due to all systems have /tmp available however, it
too should be updated to a place that other users can't access.
SecDataDir /tmp/
-- File uploads handling configuration -------------------------------------
The location where ModSecurity stores intercepted uploaded files. This
location must be private to ModSecurity. You don't want other users on
the server to access the files, do you?
#SecUploadDir /opt/modsecurity/var/upload/
By default, only keep the files that were determined to be unusual
in some way (by an external inspection script). For this to work you
will also need at least one file inspection rule.
#SecUploadKeepFiles RelevantOnly
Uploaded files are by default created with permissions that do not allow
any other user to access them. You may need to relax that if you want to
interface ModSecurity to an external program (e.g., an anti-virus).
#SecUploadFileMode 0600
-- Debug log configuration -------------------------------------------------
The default debug log configuration is to duplicate the error, warning
and notice messages from the error log.
#SecDebugLog /var/log/modsec_debug.log
#SecDebugLogLevel 4
-- Audit log configuration -------------------------------------------------
Log the transactions that are marked by a rule, as well as those that
trigger a server error (determined by a 5xx or 4xx, excluding 404,
level response status codes).
#SecAuditEngine Off
SecAuditEngine RelevantOnly
#SecAuditLogRelevantStatus "^(?:5|4(?!04))"
SecAuditLogRelevantStatus "403"
Log everything we know about a transaction.
#SecAuditLogParts ABIJDEFHZ
SecAuditLogParts ABCDEFHIJKZ
Use a single file for logging. This is much easier to look at, but
assumes that you will use the audit log only ocassionally.
SecAuditLogType Serial
SecAuditLog /var/log/modsec_audit.log
Specify the path for concurrent audit logging.
#SecAuditLogStorageDir /opt/modsecurity/var/audit/
-- Miscellaneous -----------------------------------------------------------
Use the most commonly used application/x-www-form-urlencoded parameter
separator. There's probably only one application somewhere that uses
something else so don't expect to change this value.
SecArgumentSeparator &
Settle on version 0 (zero) cookies, as that is what most applications
use. Using an incorrect cookie version may open your installation to
evasion attacks (against the rules that examine named cookies).
SecCookieFormat 0
Specify your Unicode Code Point.
This mapping is used by the t:urlDecodeUni transformation function
to properly map encoded data to your language. Properly setting
these directives helps to reduce false positives and negatives.
SecUnicodeMapFile unicode.mapping 20127
Improve the quality of ModSecurity by sharing information about your
current ModSecurity version and dependencies versions.
The following information will be shared: ModSecurity version,
Web Server version, APR version, PCRE version, Lua version, Libxml2
version, Anonymous unique id for host.
SecStatusEngine On`
@mimugmail, @zimmerle, I've spent some time running simple benchmarks on multi-core VM, and found something interesting about SecAuditEngine:
https://github.com/defanator/modsecurity-performance/wiki#2017-03-05
@mimugmail, I've just updated your comments to get proper formatting for preformatted text (commands output) - JFYI.
@defanator Thanks :) Just did a copy+paste.
This behavior fits perfectly with my measurements. Have you already tried concurrent logging?
I changed to:
SecAuditLogType Concurrent
SecAuditLog /var/log/modsec_audit.log
SecAuditLogStorageDir /opt/nginx/logs/audit
But performance is still around 144req/s
Hi @mimugmail and @defanator,
Do you mind to test the branch v3/dev/speedup?
I did some performance improvements on v3/dev/speedup. There are space for more, but it will be good to count in your tests to confirm that I am on the right track.
Oh btw, please give it a try with the SecAuditLog disabled, going to look into the audit log generation tomorrow.
@zimmerle, please check https://github.com/defanator/modsecurity-performance/wiki#2017-03-10
I ran a few cycles with SecAuditEngine RelevantOnly, and then switched to SecAuditEngine Off - numbers were nearly the same, in the margin of error.
However, in the worst case (full OWASP CRS v3.0.0 is loaded) average RPS is more than 10% greater, and latency is more than 10% lower than in previous run (v3/master as of 20170307).
SecAuditEngine RelevantOnly seems to affect only multi-worker configurations (probably some locking issues - I did not have time to test further).
@zimmerle
With SecAuditEngine On
`root@nginx:/opt# ab -n 10000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 1000 requests
Completed 2000 requests
Completed 3000 requests
Completed 4000 requests
Completed 5000 requests
Completed 6000 requests
Completed 7000 requests
Completed 8000 requests
Completed 9000 requests
Completed 10000 requests
Finished 10000 requests
Server Software: nginx/1.11.9
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 612 bytes
Concurrency Level: 20
Time taken for tests: 80.526 seconds
Complete requests: 10000
Failed requests: 0
Keep-Alive requests: 9909
Total transferred: 8499545 bytes
HTML transferred: 6120000 bytes
Requests per second: 124.18 [#/sec] (mean)
Time per request: 161.052 [ms] (mean)
Time per request: 8.053 [ms] (mean, across all concurrent requests)
Transfer rate: 103.08 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.0 0 0
Processing: 15 161 32.7 159 1517
Waiting: 15 161 32.7 159 1517
Total: 15 161 32.7 159 1517
Percentage of the requests served within a certain time (ms)
50% 159
66% 161
75% 163
80% 165
90% 169
95% 175
98% 184
99% 215
100% 1517 (longest request)
root@nginx:/opt# ab -n 1000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 100 requests
Completed 200 requests
Completed 300 requests
Completed 400 requests
Completed 500 requests
Completed 600 requests
Completed 700 requests
Completed 800 requests
Completed 900 requests
Completed 1000 requests
Finished 1000 requests
Server Software: nginx/1.11.9
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 612 bytes
Concurrency Level: 20
Time taken for tests: 8.606 seconds
Complete requests: 1000
Failed requests: 0
Keep-Alive requests: 1000
Total transferred: 850000 bytes
HTML transferred: 612000 bytes
Requests per second: 116.19 [#/sec] (mean)
Time per request: 172.127 [ms] (mean)
Time per request: 8.606 [ms] (mean, across all concurrent requests)
Transfer rate: 96.45 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.0 0 0
Processing: 7 171 18.3 169 243
Waiting: 7 171 18.3 169 243
Total: 8 171 18.2 169 243
Percentage of the requests served within a certain time (ms)
50% 169
66% 173
75% 176
80% 177
90% 183
95% 199
98% 211
99% 226
100% 243 (longest request)`
WIth SecAuditEngine Off results are the same ...
@mimugmail @defanator thanks!
@mimugmail how many workers do you have?
@zimmerle meeep you got me, now it's 100 (was 1) 👍
SecAuditEngine Off:
`root@nginx:/opt# ab -n 10000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 1000 requests
Completed 2000 requests
Completed 3000 requests
Completed 4000 requests
Completed 5000 requests
Completed 6000 requests
Completed 7000 requests
Completed 8000 requests
Completed 9000 requests
Completed 10000 requests
Finished 10000 requests
Server Software: nginx/1.11.9
Server Hostname: 127.0.0.1
Server Port: 80
Document Path: /
Document Length: 612 bytes
Concurrency Level: 20
Time taken for tests: 25.751 seconds
Complete requests: 10000
Failed requests: 0
Keep-Alive requests: 9909
Total transferred: 8499545 bytes
HTML transferred: 6120000 bytes
Requests per second: 388.34 [#/sec] (mean)
Time per request: 51.501 [ms] (mean)
Time per request: 2.575 [ms] (mean, across all concurrent requests)
Transfer rate: 322.34 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 0 0.8 0 28
Processing: 7 51 71.6 33 1140
Waiting: 0 50 71.5 32 1140
Total: 7 51 72.0 33 1151
Percentage of the requests served within a certain time (ms)
50% 33
66% 42
75% 49
80% 56
90% 92
95% 148
98% 235
99% 352
100% 1151 (longest request)
`SecAuditEngine On:
`root@nginx:/opt# ab -n 10000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 1000 requests
Completed 2000 requests
Completed 3000 requests
Completed 4000 requests
Completed 5000 requests
Completed 6000 requests
Completed 7000 requests
Completed 8000 requests
Completed 9000 requests
apr_pollset_poll: The timeout specified has expired (70007)
Total of 9981 requests completed`
Nginx 1.11.9, 4GB RAM, 4 CPUs
Also with just 1000:
`root@nginx:/opt# ab -n 1000 -c 20 -k http://127.0.0.1:80/
This is ApacheBench, Version 2.3 <$Revision: 1604373 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking 127.0.0.1 (be patient)
Completed 100 requests
Completed 200 requests
Completed 300 requests
Completed 400 requests
Completed 500 requests
Completed 600 requests
Completed 700 requests
Completed 800 requests
Completed 900 requests
apr_pollset_poll: The timeout specified has expired (70007)
Total of 981 requests completed`
@mimugmail Great! Thanks ;)
@mimugmail, running 100 workers on 4-core machine serving CPU-heavy load (like ModSecurity) won't lead to any benefits. I don't see much sense in setting worker_processes to a number greater than total number of available cores (for this particular pattern).
Also, your latest results do not include any summary for SecAuditEngine On case due to timeouts detected by ab tool.
@defanator Yep, I set it to 4, no difference. Funny thing is that between 1000 and 10000 checks it always stucks at the last 105.
With just 100 I get:
Requests per second: 206.24 [#/sec] (mean)
the branch v3/dev/speedup was merged into v3/master.
2G KVM VPS 1vcore
ab -n 10000 -c 20 -k http://theo.in/
This is ApacheBench, Version 2.3 <$Revision: 1706008 $>
Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/
Licensed to The Apache Software Foundation, http://www.apache.org/
Benchmarking theo.in (be patient)
Completed 1000 requests
Completed 2000 requests
Completed 3000 requests
Completed 4000 requests
Completed 5000 requests
Completed 6000 requests
Completed 7000 requests
Completed 8000 requests
Completed 9000 requests
Completed 10000 requests
Finished 10000 requests
Server Software: XtendWeb-nginx
Server Hostname: theo.in
Server Port: 80
Document Path: /
Document Length: 623 bytes
Concurrency Level: 20
Time taken for tests: 156.437 seconds
Complete requests: 10000
Failed requests: 0
Keep-Alive requests: 9908
Total transferred: 8149540 bytes
HTML transferred: 6230000 bytes
Requests per second: 63.92 [#/sec] (mean)
Time per request: 312.873 [ms] (mean)
Time per request: 15.644 [ms] (mean, across all concurrent requests)
Transfer rate: 50.87 [Kbytes/sec] received
Connection Times (ms)
min mean[+/-sd] median max
Connect: 0 3 28.2 0 465
Processing: 190 310 74.8 297 688
Waiting: 190 310 74.8 297 688
Total: 190 313 80.7 298 959
Percentage of the requests served within a certain time (ms)
50% 298
66% 327
75% 346
80% 359
90% 401
95% 468
98% 556
99% 599
100% 959 (longest request)
# nginx -V
nginx version: nginx/1.11.12
built by gcc 4.8.5 20150623 (Red Hat 4.8.5-11) (GCC)
built with LibreSSL 2.5.1
TLS SNI support enabled
configure arguments: --prefix=/etc/nginx --sbin-path=/usr/sbin/nginx --modules-path=/etc/nginx/modules --with-pcre=./pcre-8.40 --with-pcre-jit --with-zlib=./zlib-1.2.11 --with-openssl=./libressl-2.5.1 --conf-path=/etc/nginx/nginx.conf --error-log-path=/var/log/nginx/error_log --http-log-path=/var/log/nginx/access_log --pid-path=/var/run/nginx.pid --lock-path=/var/run/nginx.lock --http-client-body-temp-path=/var/cache/nginx/client_temp --http-proxy-temp-path=/var/cache/nginx/proxy_temp --http-fastcgi-temp-path=/var/cache/nginx/fastcgi_temp --http-uwsgi-temp-path=/var/cache/nginx/uwsgi_temp --http-scgi-temp-path=/var/cache/nginx/scgi_temp --user=nobody --group=nobody --with-http_ssl_module --with-http_realip_module --with-http_addition_module --with-http_sub_module --with-http_dav_module --with-http_flv_module --with-http_mp4_module --with-http_gunzip_module --with-http_gzip_static_module --with-http_random_index_module --with-http_secure_link_module --with-http_stub_status_module --with-http_auth_request_module --add-dynamic-module=naxsi-http2/naxsi_src --with-file-aio --with-threads --with-stream --with-stream_ssl_module --with-http_slice_module --with-compat --with-http_v2_module --with-http_geoip_module=dynamic --add-dynamic-module=ngx_pagespeed-release-1.11.33.4-beta --add-dynamic-module=/usr/local/rvm/gems/ruby-2.3.1/gems/passenger-5.1.2/src/nginx_module --add-dynamic-module=ngx_brotli --add-dynamic-module=echo-nginx-module-0.60 --add-dynamic-module=headers-more-nginx-module-0.32 --add-dynamic-module=ngx_http_redis-0.3.8 --add-dynamic-module=redis2-nginx-module --add-dynamic-module=srcache-nginx-module-0.31 --add-dynamic-module=ngx_devel_kit-0.3.0 --add-dynamic-module=set-misc-nginx-module-0.31 --add-dynamic-module=ModSecurity-nginx --with-cc-opt='-O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector-strong --param=ssp-buffer-size=4 -grecord-gcc-switches -m64 -mtune=generic' --with-ld-opt=-Wl,-E
nginx_conf.txt
A couple of benchmarks with recent libmodsecurity sources: https://github.com/defanator/modsecurity-performance/wiki#2017-04-11-1 (single worker), https://github.com/defanator/modsecurity-performance/wiki#2017-04-11-2 (multi worker).
Noticeable improvement is that multi-worker setup does not lock all the processing inside single worker (as it was before [1]), so I was able to reach almost 1k RPS with the full OWASP CRS v3.0.0 loaded, with 6 workers.
[1] https://github.com/defanator/modsecurity-performance/wiki#2017-03-05
I am closing this issue in favor of @defanator wiki. I would encourage you guys to keep the work monitoring the performance. whenever you find a problem with it, please re-open this issue.
@defanator did you get a chance to re-run the benchmark again?
@luengnat I run it periodically, just updated the results page with most recent build here: https://github.com/defanator/modsecurity-performance/wiki#2017-12-11
| gharchive/issue | 2017-02-26T06:51:17 | 2025-04-01T04:33:06.083986 | {
"authors": [
"AnoopAlias",
"defanator",
"luengnat",
"mimugmail",
"zimmerle"
],
"repo": "SpiderLabs/ModSecurity-nginx",
"url": "https://github.com/SpiderLabs/ModSecurity-nginx/issues/35",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
21181969 | Adding parameters to transformation functions
MODSEC-210: It would be very useful to have the possibility to send (string) parameters to transformation functions, like
t:replace(a,b)
The syntax is only an example, it could be completely different.
As we have a parser on v3, we could easily have that implemented per-transformation. Said that, we no longer need to have such issue.
| gharchive/issue | 2013-10-17T20:18:39 | 2025-04-01T04:33:06.087523 | {
"authors": [
"rcbarnett-zz",
"zimmerle"
],
"repo": "SpiderLabs/ModSecurity",
"url": "https://github.com/SpiderLabs/ModSecurity/issues/364",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
604422252 | what version of discord.js is this?
Also, you dont mention anywhere in the readme that you need to npm install ytdl and yt-search and all the other things
V12 Discord.js and also it's common sense to download the npms unless you just wanna fork without any knowledge or experience of a Discord bot.
@Ravenwulf running npm install installs every packages needed for the bot since package.json contains all the needed information. It'll also install the correct version of said package. Simply rum npm install as stated in the README file.
| gharchive/issue | 2020-04-22T03:19:07 | 2025-04-01T04:33:06.114757 | {
"authors": [
"ItsAuric",
"Ravenwulf",
"Spimy"
],
"repo": "Spimy/muse",
"url": "https://github.com/Spimy/muse/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
624229604 | Synchronize Gradle wrappers among builds
Now that we have custom pull scripts (see #534), we can sync the Gradle wrapper configs and JARs automatically upon pull instead of manually updating ./gradle/wrapper, ./tools/smoke-tests/gradle/wrapper, and ./base-validating-builders/gradle/wrapper.
This issue relates to #578.
No longer relevant to this repo and is addressed in mc-java.
| gharchive/issue | 2020-05-25T11:09:00 | 2025-04-01T04:33:06.134791 | {
"authors": [
"alexander-yevsyukov",
"dmdashenkov"
],
"repo": "SpineEventEngine/base",
"url": "https://github.com/SpineEventEngine/base/issues/538",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
489797376 | Publish 1.0.8-SNAPSHOT
This PR advances the version of the library to 1.0.8-SNAPSHOT.
@alexander-yevsyukov, PTAL.
| gharchive/pull-request | 2019-09-05T14:26:09 | 2025-04-01T04:33:06.135787 | {
"authors": [
"dmdashenkov"
],
"repo": "SpineEventEngine/time",
"url": "https://github.com/SpineEventEngine/time/pull/71",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
270055264 | [beta] Transpiled error class is not working as expected.
Catching errors is broken with the message: TypeError: _this4.addToMessage is not a function
Even though babel is trying to create the error classes here: https://github.com/SpireTeam/whitelister/blob/feature/bundle/dist/whitelister.js#L206
See codepen: https://codepen.io/anon/pen/eeNMWY?editors=1111#anon-login
Turns out this had to do with extending a global when using babel, specifically Error.
Fixed in 0.1.0-beta2 using the babel plugin referenced on MDN here.
| gharchive/issue | 2017-10-31T17:48:52 | 2025-04-01T04:33:06.138378 | {
"authors": [
"davidpaulhunt"
],
"repo": "SpireTeam/whitelister",
"url": "https://github.com/SpireTeam/whitelister/issues/10",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1780996587 | Added top contributors in ABOUTUS page
Fixes Issue
Closes #1656
Changes proposed
Screenshots
Checklist
[x ] You've linked this PR to the correct issue.
[x ] You have checked that the code is working correctly.
[ ] You ⭐️ the repository!
Note to reviewers
@Spyware007 do add level in this PR ,it will be kind of you : |)
| gharchive/pull-request | 2023-06-29T14:52:40 | 2025-04-01T04:33:06.176093 | {
"authors": [
"ManishaSwain8"
],
"repo": "Spyware007/Animating-Buttons",
"url": "https://github.com/Spyware007/Animating-Buttons/pull/1674",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1954191412 | Incorporar métodos de eliminación a EntityPersist
Incorporar métodos de eliminación a EntityPersist.
se incorporo primer metodo de eliminación DeleteIds
| gharchive/issue | 2023-10-20T12:29:24 | 2025-04-01T04:33:06.176914 | {
"authors": [
"ivancas84"
],
"repo": "SqlOrganize/SqlOrganizeCs",
"url": "https://github.com/SqlOrganize/SqlOrganizeCs/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1659431107 | Testing, please ignore. (qdocvn)
Bug bounty test - please ignore.... Please DO NOT APPROVE THIS! (qdocvn)
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.alveranuno seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
| gharchive/pull-request | 2023-04-08T07:30:54 | 2025-04-01T04:33:06.181454 | {
"authors": [
"CLAassistant",
"alveranuno"
],
"repo": "Squarespace/cldr-engine-java",
"url": "https://github.com/Squarespace/cldr-engine-java/pull/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
70929627 | Disconnect cables with wrench
Add the ability to disconnect multipart cables with buildwraft wrenches. We could store this in another flag and merge the two when checking canBeVisited and canVisitTo.
I like the idea, but then the cable has to be a multipart cable to use this feature, which can be frustrating when you want to use it on a non-multipart cable.
I know, we could add a custom handler to the activate method to convert it to a multipart block - but this seems less than ideal. The alternative would be an 'advanced cable' which has more functionality but I feel we would have to make it really worthwhile for it to exist.
That's an interesting idea. I dunno though...
An alternative would be to dye cables instead - so different coloured cables cannot connect to each other.
The most sensible thing to do seems to be to add a IColoredNode or something which controls if a connection should occur. That can probably wait until 0.2.1 though.
Yea I'd say we hold onto this one until 0.2.1.
I've thought about this a lot and I'm not sure I see the point in this: I can't see any benefit of being able to separate networks. It would be a really cool feature to add but I don't want to add a load of unneeded complexity.
I might open if someone can think of an application of this: separating cables without FMP installed would be good.
| gharchive/issue | 2015-04-25T15:46:22 | 2025-04-01T04:33:06.190657 | {
"authors": [
"ElvishJerricco",
"SquidDev"
],
"repo": "SquidDev-CC/CC-Tweaks",
"url": "https://github.com/SquidDev-CC/CC-Tweaks/issues/19",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1126739843 | Add Delete to context menu
I hadn't any issues with the dirty indicator for many months. I think it's already safe to have Delete in the Context Menu.
If the repo is dirty, show ~ "The repository has local changes. Do you want to move it to the trash?"
VS Code has this windows when deleting a file, we can have a similar text / look and feel:
It's possible to use the trash pkg.
Added in 1.6.0. Note however that we send it to trash as there is no reliable package for it in npm. Maybe copy VSCode code for it?
| gharchive/issue | 2022-02-08T03:13:04 | 2025-04-01T04:33:06.193002 | {
"authors": [
"SrBrahma"
],
"repo": "SrBrahma/GitHub-Repository-Manager",
"url": "https://github.com/SrBrahma/GitHub-Repository-Manager/issues/34",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
257754074 | can't use that module with angular-cli optimized builds
Hi
I am using angular/cli=1.4.2, typescript=2.5.2 and latest version of Angular2-Toaster
I am trying to compile my app like ng build -e=prod -aot --build-optimizer, but I have these errors:
my component
problematic line in toaster.js
It looks like easy issues tp fix, but I just wodering why BodyOutputType.Default expression fails
I think I just need to override configs, but maybe you want to fix that issue
This is very curious...the BodyOutputType IS defaulted intentionally. I suspect this is due to the optimize flag attempting dead code elimination and it may believe the property isn't needed. I will try to reproduce.
This is currently fixed in 5.0.0-beta.1 and will be released in the full 5.0.0 release.
Released in 5.0.0.
| gharchive/issue | 2017-09-14T15:07:02 | 2025-04-01T04:33:06.206824 | {
"authors": [
"Stabzs",
"ToGoBananas"
],
"repo": "Stabzs/Angular2-Toaster",
"url": "https://github.com/Stabzs/Angular2-Toaster/issues/128",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
234941034 | Add PUT method to action executions API
Allow action executions to be updated over API. The API will only allow status and result to be updated for incomplete action executions. This patch is required for 1) mistral->st2 callback on workflow completion and 2) st2 response to update st2.ask action execution.
LGTM.
| gharchive/pull-request | 2017-06-09T21:17:16 | 2025-04-01T04:33:06.213285 | {
"authors": [
"Kami",
"m4dcoder"
],
"repo": "StackStorm/st2",
"url": "https://github.com/StackStorm/st2/pull/3461",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
616718281 | Do not lookup release from package cloud if fork in circle config.yml
if package cloud token is not set use hard coded value of 1
fixes #752
PR from forks are not able to be run through unit testing. This is because the package cloud token is not set. Here we are not going to lookup the release number in package cloud if it is a fork.
This was stolen from https://github.com/StackStorm/st2-packages/blob/29a493e33de928e6df452a9102e80aaeadc5ec53/.circle/buildenv_st2.sh#L45-L51
Looks like it passed.
| gharchive/pull-request | 2020-05-12T14:48:55 | 2025-04-01T04:33:06.217044 | {
"authors": [
"guzzijones"
],
"repo": "StackStorm/st2web",
"url": "https://github.com/StackStorm/st2web/pull/753",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2756434625 | node22.x not working
There was a typo in the previous release. There is the node18.x twice and no node22.x
Please fix support for node22.x
Thanks!
Please upgrade to 0.15.1 as it contains a fix.
| gharchive/issue | 2024-12-23T17:04:56 | 2025-04-01T04:33:06.218151 | {
"authors": [
"buggy",
"xpert4"
],
"repo": "StackToolbox/aws-sam-webpack-plugin",
"url": "https://github.com/StackToolbox/aws-sam-webpack-plugin/issues/101",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
182804424 | Bad require in stack bin file.
In stack/bin/stack file the require in the following line fails:
https://github.com/Stackvana/stack/blob/master/bin/stack#L39
I think it should be
var viewPresenter = #require('../lib/viewPresenter');
Ahh! Sorry about that, will patch now and publish.
| gharchive/issue | 2016-10-13T14:24:48 | 2025-04-01T04:33:06.219583 | {
"authors": [
"Marak",
"labajo"
],
"repo": "Stackvana/stack",
"url": "https://github.com/Stackvana/stack/issues/10",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
361095748 | LND won't start after updating to v0.5-beta
I've tried updating to v0.5-beta following the guide but LND fails to start:
Sep 15 11:24:49 pideypideypideypi systemd[1]: Started LND Lightning Daemon.
Sep 15 11:24:49 pideypideypideypi systemd[1]: lnd.service: Main process exited, code=exited, status=1/FAILURE
Sep 15 11:24:49 pideypideypideypi systemd[1]: lnd.service: Unit entered failed s tate.
Sep 15 11:24:49 pideypideypideypi systemd[1]: lnd.service: Failed with result 'e xit-code'.
No problems after reverting to v0.4.2-beta. I'm not sure how to pull more detailed logs.
I have the same issue, fwiw. Although I'm on "0.5 release candidate" (can't update to "proper" 0.5 due to the missing link in the LND repository, as is also described here:
https://github.com/Stadicus/guides/issues/210).
Also, the "logs" folder under /home/bitcoin/.lnd is showing as empty (0 bytes), so apparently no log is being written after I try to restart the LND client and it fails repeatedly.
Please try to run lnd manually with the bitcoin user to see the terminal output.
Can you start lnd from the root account, just by typing lnd?
Seems to be possible in general, although it gives out loadConfig: either bitcoin.active or litecoin.active must be set to 1 (true).
It's probably not relevant to the issue in the OP though (except if they followed the same older version of the guide, and it's somehow connected to the issue), so hopefully this doesn't derail the thread too much :)
That might be the issue. Do you have the option bitcoin.active=1 set in your lnd.config?
My lnd.conf (no reference, might have some stupid shit in it)
# RaspiBolt LND Mainnet: lnd configuration
# /home/bitcoin/.lnd/lnd.conf
[Application Options]
debuglevel=debug
maxpendingchannels=5
alias=Stadicus RaspiBolt
color=#68F442
restlisten=localhost:8080
rpclisten=localhost:10009
[Bitcoin]
bitcoin.active=1
bitcoin.mainnet=1
bitcoin.node=bitcoind
[autopilot]
autopilot.active=0
I think I found a fix, when running 0.5 rc2.
In the release notes, changes are needed the .conf for lnd and bitcoind. After I made these changes and restarted both, lnd would start.
Users running a bitcoind backend will need to modify their existing ZMQ-related settings. Previously, lnd listened on one shared port for both blocks and transactions. 0.5 requires two distinct ports to be set, such that they are isolated. Users should remove the old bitcoind.zmqpath entry in lnd.conf, and, for example, replace it with:
bitcoind.zmqpubrawblock=tcp://127.0.0.1:28332
bitcoind.zmqpubrawtx=tcp://127.0.0.1:28333
Also ensure that bitcoind is exposing distinct ZMQ ports in bitcoin.conf:
zmqpubrawblock=tcp://127.0.0.1:28332
zmqpubrawtx=tcp://127.0.0.1:28333
Was there a breaking change that I did not cover in the FAQ update section?
Nope, I upgraded using the updated FAQ and everything went fine.
| gharchive/issue | 2018-09-18T00:52:39 | 2025-04-01T04:33:06.226893 | {
"authors": [
"Stadicus",
"pardus79",
"raindogdance",
"whatthemoses"
],
"repo": "Stadicus/guides",
"url": "https://github.com/Stadicus/guides/issues/216",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.