added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:36:46.623247
2024-08-17T22:09:39
2471687705
{ "authors": [ "Benjamin-Loison" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:430", "repo": "Benjamin-Loison/organicmaps", "url": "https://github.com/Benjamin-Loison/organicmaps/issues/50" }
gharchive/issue
Anti-Features seem to hide by default in F-Droid the app when search it Related to Benjamin_Loison/fdroiddata/issues/8. Maybe it is because Organic Maps seems to have intermediary servers in comparison with OpenStreetMap. https://f-droid.org/en/docs/Anti-Features/#NonFreeNet https://f-droid.org/en/docs/Anti-Features/#TetheredNet See fdroid/fdroiddata/issues/3442.
2025-04-01T06:36:46.624148
2019-10-14T23:43:00
506925391
{ "authors": [ "BennyCarlsson", "orangegrove1955" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:431", "repo": "BennyCarlsson/MyPortfolio-Hacktoberfest2019", "url": "https://github.com/BennyCarlsson/MyPortfolio-Hacktoberfest2019/pull/240" }
gharchive/pull-request
Added github corner to profile Using tholman/github-corner, added a small graphic to top right of index.html that will take users to the GitHub repo for the project great 👍
2025-04-01T06:36:46.650065
2022-06-09T07:20:11
1265705498
{ "authors": [ "ULIFTWHITE" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:432", "repo": "BertrandBev/code_field", "url": "https://github.com/BertrandBev/code_field/issues/40" }
gharchive/issue
[code_text_field] Hello, I want to get changing value of code snippet. Is there any way to get it? For example, the initial code snippet is 'print("hello python")', and I add 'print("HWY")'. Then, How can I get the value of 'print("hello python") print("HWY")'? I figured it out. Thanks :)
2025-04-01T06:36:46.729025
2024-11-02T18:29:35
2630689781
{ "authors": [ "BiagioFesta", "tetter27" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:433", "repo": "BiagioFesta/wtransport", "url": "https://github.com/BiagioFesta/wtransport/issues/234" }
gharchive/issue
How to solve the SNI setting on a client I found that the error occurs when a client with SNI configured connects. 2024-11-02T17:03:59.352264Z WARN rustls::msgs::handshake: Illegal SNI hostname received "xxx.xxx.xxx.xxx" 2024-11-02T17:03:59.352379Z DEBUG quinn_proto::endpoint: handshake failed: the cryptographic handshake failed: error 50: received corrupt message of type InvalidServerName xxx.xxx.xxx.xxx: The IP address of the actual server. If this hostname was localhost, it could be solved. Can this problem be solved with existing options? Thank you! Can you please provide more details? How client SNI is set? (E.g., how [wtransport::Endpoint::connect](https://docs.rs/wtransport/latest/wtransport/ struct.Endpoint.html#method.connect) is called?) Are those logs from server?
2025-04-01T06:36:46.731513
2024-03-15T10:27:03
2188192070
{ "authors": [ "aymericdelab", "r0man1337", "svetaet24" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:434", "repo": "BibliothecaDAO/eternum", "url": "https://github.com/BibliothecaDAO/eternum/issues/400" }
gharchive/issue
Add sounds add sounds to following actions: [ ] exploration [ ] hex travel [ ] Lords sound (like resources sound) [ ] Level UP sound [ ] running sound fixed in #485
2025-04-01T06:36:46.766258
2024-09-01T17:22:27
2499591825
{ "authors": [ "Evanalfredd", "hufman" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:435", "repo": "BimmerGestalt/AAIdrive", "url": "https://github.com/BimmerGestalt/AAIdrive/issues/816" }
gharchive/issue
Id6 Samsung s23 ultra connection issues Hello! I am having issues connecting my phone to the car, and app. I have the connected drive subscription. Everything connected except the last options which is usb not in transfer mode. But it IS in transfer mode. I attached pictures. I'm aware of the app protection that the s23 ultra has. I have done all the app permission settings I've seen in the other chats. I would like to thank the creators as well.. lol first time user and I think this is the best thing that has happened to bmw. I just wish it would work for me. I paid $150 for the bmw connect to drive subscription just so I can take advantage of this app.. 😭 USB mode is flaky with MyBMW, if you want to use it you should use BMW Connected 6.4. For Bluetooth, make sure the car shows the Apps option when pairing a new mobile device. If the Apps option is missing, then your ConnectedDrive subscription isn't active yet or maybe the functionality is under a different subscription level. I tried downloading the apk file for the connect drive app 6.4. It does nothing Yes that is correct. It just needs to be installed, and it will attempt any car connection in the background. If your car supports the connection tho. Have you verified that the Apps option shows when you go to Pair a new device? Yes that is correct. It just needs to be installed, and it will attempt any car connection in the background. If your car supports the connection tho. Have you verified that the Apps option shows when you go to Pair a new device? https://github.com/user-attachments/assets/3595c933-2105-4214-aec0-d4b546a02155 Idk if you can upload videos here. But if you can see it.. you can see the apps option is flickering.. should I uninstall the my bmw app? Thank you for your help, I hope I get this solved 😭 Tha flickering is what I see when MyBMW tries to run over USB. I don't know why your MyBMW is acting weird over Bluetooth, perhaps it is confused by also having USB plugged in, and you should eliminate extra variables by trying one connection at a time.
2025-04-01T06:36:46.818322
2018-11-14T17:46:47
380814314
{ "authors": [ "schuhschuh" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:436", "repo": "BioMedIA/MIRTK", "url": "https://github.com/BioMedIA/MIRTK/issues/668" }
gharchive/issue
Incorrect resample-image -size option help The default is the input voxel size, not constant 1 1 1. See https://github.com/BioMedIA/MIRTK/blob/463d90ac6145be7627bb240dae5649eddf1eebb6/Applications/src/resample-image.cc#L60. Fixed by 9482b884c8022bbcc1fd4f85dd702a48dfad45d1.
2025-04-01T06:36:46.855581
2019-02-20T13:11:31
412417634
{ "authors": [ "Peteck", "bitbager" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:437", "repo": "BitBagCommerce/SyliusCmsPlugin", "url": "https://github.com/BitBagCommerce/SyliusCmsPlugin/pull/239" }
gharchive/pull-request
Related taxons Q A Bug fix? yes New feature? yes BC breaks? no Deprecations? no Related tickets fixes #X, partially #Y, mentioned in #Z License MIT Related taxons to block Added the feature for adding related taxons to block. This is usefull for some like if you wan't to create sliders for selected taxons on the homepage. {% for taxon in block.taxons %} <h3 class="section-headline">{{ block.name }}</h3> <div class="sub">{{ block.content|raw }}</div> {{ render(url('sylius_shop_partial_product_index_by_taxon_code', {'code': taxon.code, 'count': 16, 'template': '@SyliusShop/Product/Slider/_productSlider.html.twig'})) }} {% endfor %} Removed constraint for Sylius v1.4 I could not install the bitbag cms plugin for my v1.3 app anymore, and I believe there was no reason to constraint it to v1.4, as the plugin seem to work fine with my 1.3 app. Fixed a mistake I made I removed the parent from the repositories as it seemed to not break anything, it did tho, so I reverted it. Hello Peteck! Do you think you might be able to provide Behat scenarios & Specs to cover these features? Yes. I will make some :-) I've only made specs for the reason that there only exist specs for associated products, channels and sections. Hi there! By any chance - could you upgrade the PR?
2025-04-01T06:36:46.876347
2019-05-24T15:18:06
448224934
{ "authors": [ "christroutner", "coveralls" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:438", "repo": "Bitcoin-com/rest.bitcoin.com", "url": "https://github.com/Bitcoin-com/rest.bitcoin.com/pull/417" }
gharchive/pull-request
Unit test mocked for SLP txDetails route Gabriel asked me to go back and add unit test mocking for the failing txDetails route. As laid out in Issue #415, this required a different approach to the mocks because slpjs is now a dependency-of-a-dependency and can not be mocked out using proxyquire as done before. Here is the new code path relationship between rest, SLP-SDK, and slpjs: The slp.ts route requires slp-sdk, which is a class. slp-sdk is instantiated. slp.js is a property of the instantiated slp-sdk object. Inside the txDetails route, the slpjs.BitboxNetwork class is then instantiated as tmpbitboxNetwork (and slp-sdk class is passed to it as an instance of BITBOX class) tmpbitboxNetwork.getTransactionDetails() is then called That's a very complex code path, and as a result is very difficult to mock for a unit test. Therefore, I cheated. The final call (which is where the error originates) is this line, which calls the logic in the slpjs library: https://github.com/Bitcoin-com/rest.bitcoin.com/blob/5a66d205edba3439151ae4a3477d3657cdd1e401/src/routes/v2/slp.ts#L1327 Therefore, I wrapped that line in a new function called getSlpjsTxDetails(). By wrapping it in a function, I could then use Sinon to stub out the function and replace the returned data with mocked data. This allows the unit test (the test for the code in rest) to pass. It does not fix the error in slpjs. I'm still not sure what is causing that error, but I know it's not in the rest code base. Pull Request Test Coverage Report for Build 2214 0 of 0 changed or added relevant lines in 0 files are covered. 120 unchanged lines in 1 file lost coverage. Overall coverage increased (+0.08%) to 70.227% Files with Coverage Reduction New Missed Lines % dist/routes/v2/slp.js 120 49.79% Totals Change from base Build 2200: 0.08% Covered Lines: 2053 Relevant Lines: 2727 💛 - Coveralls
2025-04-01T06:36:46.877874
2018-08-01T07:15:55
346476501
{ "authors": [ "SunLn", "usatie" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:439", "repo": "BitcoinCashKit/BitcoinKit", "url": "https://github.com/BitcoinCashKit/BitcoinKit/pull/67" }
gharchive/pull-request
fix issue #49 fix issue #49 Hi @SunLn ! Thanks for dropping us the PR! It was really helpful but the diff is too much and Carthage files should not be pushed to this repo. So we can't merge this PR, sorry! But we'll soon fix the problem you raised!
2025-04-01T06:36:46.879234
2024-09-26T01:10:12
2549262534
{ "authors": [ "GBKS", "yashrajd" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:440", "repo": "BitcoinDesign/Bitcoin-Core-App", "url": "https://github.com/BitcoinDesign/Bitcoin-Core-App/issues/120" }
gharchive/issue
Add testing steps & info for assumeutxo As discussed in the Bitcoin Core App today, we should provide directions to test assumeutxo. This issue is meant to track that. Please assign it to me, so I can figure this out and open a PR to add these steps to the Snapshot page of the website. This info could go on the Snapshot page.
2025-04-01T06:36:46.885346
2022-07-05T23:58:35
1294912199
{ "authors": [ "shashank-reddy-code" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:441", "repo": "BitskiCo/ethereum-etl", "url": "https://github.com/BitskiCo/ethereum-etl/pull/31" }
gharchive/pull-request
fix duplicate token_ids in batch transfer bug Batch transfers could have duplicate token_ids as part of same log. This will break the primary key constraint in database (log_index, token_id, transaction_hash). This PR fixes the bug by adding up the amounts per token_id Thanks @andschneider who brought up this bug!
2025-04-01T06:36:46.929000
2019-07-11T04:19:31
466653980
{ "authors": [ "BlackGlory", "leesei" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:442", "repo": "BlackGlory/copycat", "url": "https://github.com/BlackGlory/copycat/issues/9" }
gharchive/issue
Percent encode parenthesis in URL when creating Markdown link URL: https://www.wikiwand.com/en/Ring_(mathematics) "Tab link -> Markdown": [Ring (mathematics) - Wikiwand](https://www.wikiwand.com/en/Ring_(mathematics)) which caused problem in some Markdown parser I recommends outputting this: [Field (mathematics) - Wikiwand](https://www.wikiwand.com/en/Field_%28mathematics%29) Also encode space as %20 This issue is more complicated than I thought. I found that brackets are explicitly listed as reserved characters in the URL encoding. A Markdown parser should have the ability to distinguish between URLs and Markdown syntax, and if it can't, then it needs to be fixed. This extension should not change the output for this particular case.
2025-04-01T06:36:46.955459
2023-07-27T18:18:39
1824935420
{ "authors": [ "BlackSheepBoy69", "billabongbruno" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:443", "repo": "BlackSheepBoy69/HexFlow-Launcher-Unofficial-Custom", "url": "https://github.com/BlackSheepBoy69/HexFlow-Launcher-Unofficial-Custom/issues/11" }
gharchive/issue
PSVita Android Game Ports with two START buttons Hey there. Sorry for being a hassle again. I figured I would open a new issue, seeing as this is unrelated to the enhancement we were discussing. Today I installed This War of Mine with the DLC "Stories" and realized that it has two different START buttons (each redirecting to a different portion of the game). By default, HexFlow Custom launches the top one and even if I press the bottom one, it has no effect in changing the game. Is there any hidden option to enable a non-automatic-start on a per-game basis or an option to stop auto-loading when there is more than one START instance that I'm missing? Sorry for bothering you again. Thank you in advance. Best regards, Bruno. You are no hassle! I'll check this out on the weekend- You are no hassle! I'll check this out on the weekend- Well, I suppose I am, seeing as this is the second thing I ask of you. lol Thank you for checking this out. See you soon. Bruno. The 2nd 'start button' is actually just a banner. I might say split the app into 2 apps, but unfortunately the game only has one binary (the .bin file inside ux0:app/TWOM00000 ). The ability for HexFlow to use banners would be really complicated and I wouldn't really know a clean way to implement it (it would get messy for Japanese games which often have sooo many banners) The 2nd 'start button' is actually just a banner. I might say split the app into 2 apps, but unfortunately the game only has one binary (the .bin file inside ux0:app/TWOM00000 ) so it would be a lot more entailed, for example, having to mess with all Rinne's code and compile to 2 binaries. The ability for HexFlow to use banners would be really complicated and I wouldn't really know a clean way to implement it. It might be better to post it as an issue on TWOM to say like "Hey, if I have the game 1 open and hit the banner, make it close the game and run game 2" instead of keep playing game 1 which it seems to do" Just closing the issue for now because it seems more of an issue with TWOM (I don't know any other games that have this issue, standard procedure is to have a little in-game mini-menu inside that lets you pick the game) I see. That's unfortunate, but not a big issue. I have come across more Android ports that have multiple banners, but they are all "Configurator" banners rather than other portions of the games themselves, so nothing really relevant. The only one that has another game baked into it seems to be TWOM. Thank you for your reply, nevertheless. Unrelated to HexFlow, do you know if there is anyway I can create a custom PSVita Bubble that sets its main banner to the content of TWOM's secondary banner, as sort of a workaround? (Kind of like creating a clone bubble, but redirecting its shorcuts, which I believe pointed to something like "psla:stories" or something) Is this possible or would it require editing the eboot.bin? If not at all possible, I will probably keep TWOM's bubble visible, along with HexFlow Custom. Thank you once more for all of your help and the time you took to reply. Best regards, Bruno. Actually I know a really easy way to solve this... 1 sec Nope... For whatever reason it seems to refuse to launch directly by a separate bubble. You can find my attempt here: https://www.mediafire.com/file/ylga5a2om8fv0zt/TWOM_Stories_Redirector_FAILED.vpk/file Note: The apptitle is "Wordle" because I was too lazy to change it. I'm not super passionate about getting it fixed, if you are, please go to the This War Of Mine Github page and open an issue: Title: LUA Direct Launch Text something like: I was trying to make a little separate bubble that could launch straight to TWOM:Stories (so it could easily be accessed by HexFlow Launcher) using your Lua Player Plus. I tried this: System.executeUri("psgm:play?titleid=TWOM00000&param=stories") but it doesn't seem to work. Do you know what would be able to launch into TWOM:Stories? Not part of the text to put in the issue, this is just me talking: I really did try that but it didn't work. For the wordle thing I just made a copy of the TWOM game, stuffed it into a Wordle VPK (since that's the only homebrew I know that primarily uses PSLA: launch. If you want to try out LUA code, just put it in HexFlow Launcher Custom right below the area that says: elseif (Controls.check(pad, SCE_CTRL_SELECT) and not Controls.check(oldpad, SCE_CTRL_SELECT)) then (you may have to uncomment it, ex: remove the "--" at the start of the "elseif (Contro........" since it's probably commented out in the public version. Hey there. First of all, thank you for your help, input and attempts at solving this issue. I do realize it's a TWOM issue and will follow your advice and open an issue there, to see what the dev can do about it, if anything. I have no idea on how to launch TWOM:Stories other than the psla:stories redirection, unfortunately. I am not familiar with LUA coding, but I guess I'll dive into it a bit and see if I can come up with something. If I am able to boot it at all, even though I'm a novice to LUA and the coding might be rough, I'll be sure to share my findings with you, should anyone have a similar issue, thus allowing HexFlow Custom to have a workaround / fix for all thes multiple-banner apps. Once again, I cannot thank you enough for the time spent on this. See you soon. Best regards, Bruno. I opened the issue here and will keep you posted if and when an answer comes along and if my findings come to fruition. Thank you once again. Best regards, Bruno. Technically, I wasn't planning to have HexFlow be able to launch the banner, it's just I could make a whole separate vpk (as you saw in the FAILED vpk I sent you) and I could just make it like HexFlow Launcher where it takes an index.lua, but the index.lua for that would be only like 2 lines of code, the system.launch(TWOM:stories)... and system.exit to close the redirector. Ex: The redirector would run on LUA like HexFlow does so it could easily be fixed if there's any further issues... I hope that would be a good solution? I mean, anything works, really. Your solution is perfectly reasonable and seems like a good workaround. Hopefully Rinne replies soon enough, so that some light can be shed on the matter. Nope... For whatever reason it seems to refuse to launch directly by a separate bubble. The only hope might be indirectly (with LUA). You can find my direct attempt here: https://www.mediafire.com/file/ylga5a2om8fv0zt/TWOM_Stories_Redirector_FAILED.vpk/file Note: The apptitle is "Wordle" because I was too lazy to change it. I'm not super passionate about getting it fixed, if you are, please go to the This War Of Mine Github page and open an issue: Title: LUA Direct Launch Text something like: I was trying to make a little separate bubble that could launch straight to TWOM:Stories (so it could easily be accessed by HexFlow Launcher) using your Lua Player Plus. I tried this: System.executeUri("psgm:play?titleid=TWOM00000&param=psla:stories") but it doesn't seem to work. Do you know what would be able to launch into TWOM:Stories? Not part of the text to put in the issue, this is just me talking: I really did try that but it didn't work. For the wordle thing I just made a copy of the TWOM game, stuffed it into a Wordle VPK (since that's the only homebrew I know that primarily uses PSLA: launch. If you want to try out LUA code, just put it in HexFlow Launcher Custom right below the area that says: elseif (Controls.check(pad, SCE_CTRL_SELECT) and not Controls.check(oldpad, SCE_CTRL_SELECT)) then (Example usage: remove the "--" at the start of the "elseif (Contro........", since it's meant for bugtesting and is commented out in the public version of HexFlow Custom, then in the next line put the "System.executeUri("psgm:play?titleid=TW.........." and you can try editting it in some ways to find a way that might work?) Closing the issue for now as it has more to do with the weird way TWOM launches and would be more of an issue for their GitHub. If they say anything back, I'll see what I can do Hey there again. So, I am away from home right now, so I can't really check in my PSVita. I just noted something in the VPK you compiled, while trying to come up with a solution or workaround. Under settings.cfg, the enable_dlc is set to 0, rather than 1. As far as I can tell, Stories is a DLC itself. I know that in my "normal" installation, I did change enable_dlc to 1 and while have the TWOM bubble, both banners launched each version of the game. Could it be possible that this is the issue? (I have no way of testing now, only in about 15 hours, once I get home). You definitely aren't being a bother, I have probably spent like an hour max total working on the issue, it's just I was mostly waiting on Rinne to say how to make it work. I tried enabling DLC's, but it still says "Error could not load ux0:data/twom/libAndroidGame.so" I think it might be related that kubridge does special actions based on the app's ID and this app ID is not the same as TWOM's. Waiting on Rinne to reply back because I tested all I could think to try
2025-04-01T06:36:46.968950
2016-10-30T06:00:28
186119162
{ "authors": [ "ochompsky", "sinfuljosh" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:444", "repo": "BlackrockDigital/startbootstrap-sb-admin-2", "url": "https://github.com/BlackrockDigital/startbootstrap-sb-admin-2/issues/156" }
gharchive/issue
How to remove left column? How do I remove the left column and just have the dashboard and header? I've pretty much strippedout all the code but it's still there. I copied the "blank.html" from pages and removed the left side bar nav code. https://gist.github.com/sinfuljosh/0d1e6436a8504598bd30aa9ec17a0e66 Have not made any changes to css or js. Its to show you the section of html that is generating the the side bar. Thanks @sinfuljosh Looks like the 'page-wrapper' div is the one with the css to modify. Removing it took care of things, thanks!
2025-04-01T06:36:47.072607
2014-08-09T13:23:59
39885349
{ "authors": [ "amacgregor", "tvbeek" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:445", "repo": "Block8/PHPCI", "url": "https://github.com/Block8/PHPCI/issues/555" }
gharchive/issue
CodeSniffer not showing any results The codesniffer table always seems to be empty no matter if the plugin fails or succeeds. If I run phpcs manually against the same code base I'm getting errors. Is this a bug on just the view? or could it be that PHPCS is failing and never properly running? How can I verify that ? Fixed with #540 so I close this one
2025-04-01T06:36:47.129594
2022-07-11T18:57:17
1301097584
{ "authors": [ "jsgersing", "miguelaledesma", "sspradling78" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:446", "repo": "BloomTech-Labs/underdog-devs-ds-a", "url": "https://github.com/BloomTech-Labs/underdog-devs-ds-a/pull/169" }
gharchive/pull-request
Create Feedback_Visualization_For__Mentors_Sample.ipynb Description This is a notebook that shows some visualizations for Mentor feedback by Mentees. Fixes # BL-136, BL-393 Type of change Notebook addition Please delete options that are not relevant. [x] New feature Checklist: [x] My code follows PEP8 style guide [x] I have removed unnecessary print statements from my code [x] I have made corresponding changes to the documentation if necessary [x] My changes generate no errors [x] No commented-out code [x] Size of pull request kept to a minimum [x] Pull request description clearly describes changes made & motivations for said changes Loom https://www.loom.com/share/25bd4421e8d344658b9c2a19b6445f5f Great work on the visualization you made it really easy to understand and follow along in the video. This flexible visualization tool provides admins with the ability to get a quick overview, a detailed look at individual mentors' feedback outcome and vader score, and even look at sentiment changes over certain time ranges. Nice addition! @ErinNC, I figured out those changes and did some more fine detail work. It looks better than before. Thanks for your input.
2025-04-01T06:36:47.141331
2024-04-22T12:03:42
2256361963
{ "authors": [ "BluWizard10", "rhaamo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:447", "repo": "BluWizard10/Blu-Hierarchy", "url": "https://github.com/BluWizard10/Blu-Hierarchy/issues/1" }
gharchive/issue
Copy Path menu item in hierarchy I had the need for a Copy Path in the hierarchy context menu, the following code adapted a bit from https://forum.unity.com/threads/please-include-a-copy-path-when-right-clicking-a-game-object.429480/#post-2777071 do works: public static class CopyPathMenuItem { [MenuItem("GameObject/Copy Path")] private static void CopyPath() { var go = Selection.activeGameObject; if (go == null) { return; } var path = go.name; while (go.transform.parent != null) { go = go.transform.parent.gameObject; path = string.Format("{0}/{1}", go.name, path); } EditorGUIUtility.systemCopyBuffer = path; } [MenuItem("GameObject/Copy Path", true)] private static bool CopyPathValidation() { // We can only copy the path in case 1 object is selected return Selection.gameObjects.Length == 1; } } I have put it right before L113 in BluHierarchy.cs. I think it could be useful to add, and maybe check for a VRCDescriptor to get the path under the "Avatar name in hierarchy" like "Armature/Foo/Bar" instead of "My Cute Avatar/Armature/Foo/Bar" I can see the usefulness of this. I'm not sure how I can check for a VRC Avatar Descriptor to get the path under the avatar, but I'll look into it. Thanks for the suggestion!
2025-04-01T06:36:47.147695
2024-11-16T21:41:23
2664980370
{ "authors": [ "EmmaG2020", "djankies" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:448", "repo": "Blue-Ocean-Group-1/flalingo", "url": "https://github.com/Blue-Ocean-Group-1/flalingo/pull/8" }
gharchive/pull-request
completed upload picture and personal info -upload picture with cloudinary without backend integration -complete personal info section without backend integration -need to send out updated .env after merge -lint error occurred in some files This pull request includes several changes aimed at adding new features and improving existing functionalities. The key updates include enabling auto-save in VSCode settings, adding Cloudinary API integration, and enhancing the ProfilePage component to support profile picture uploads and editing user details. New Features: Profile Page Enhancements: Added image upload functionality using Cloudinary API. (client/src/pages/ProfilePage.jsx) Enabled editing of profile details such as name, username, country, phone number, gender, and email. (client/src/pages/ProfilePage.jsx) Environment Configuration: Added CLOUDINARY_API_LINK to environment variables to support Cloudinary integration. (client/config/env.js, client/example.env) [1] [2] Configuration Updates: VSCode Settings: Enabled auto-save on focus change to improve developer workflow. (.vscode/settings.json) Minor Changes: Code Cleanup: Added blank lines for better readability in route files. (server/src/routes/deck.routes.js, server/src/routes/user.routes.js) [1] [2]
2025-04-01T06:36:47.155264
2024-09-23T09:30:06
2542063780
{ "authors": [ "SoRadGaming", "zlshames" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:449", "repo": "BlueBubblesApp/bluebubbles-app", "url": "https://github.com/BlueBubblesApp/bluebubbles-app/issues/2812" }
gharchive/issue
Sending Gallery of Images Does not Work When Uploading multiple photos, instead of bunching them up into a gallery, it sends each one one at a time. This might be a server issue based on how it sends, but this feature is super useful, and for a while was supported by Beeper Mini. Here is an image of what it looks like on iPhone. This is not supported by the app at the moment, but should be supported by the API. It will be a matter of implementing the functionality client side
2025-04-01T06:36:47.158813
2024-12-27T07:23:05
2760513885
{ "authors": [ "jjoelj", "nicotriballier" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:450", "repo": "BlueBubblesApp/bluebubbles-server", "url": "https://github.com/BlueBubblesApp/bluebubbles-server/issues/721" }
gharchive/issue
"Accounts managed by Family Link are not allowed to sign in here." error when trying to sync google contacts with a supervised account Hi team, I've configure a BB server for my daughter on Android (i'm supervising her google account via Google Family Link). Receiving/sending messages work fine. The only issue at this moment is synching contacts. I'm running the BB server on a mac mini - on that mac mini in iMessage i can see that the name of each recipient is properly displayed, as it's coming from the contact app (there are contacts from icloud and contacts from google [aka 'internet account']. But in the BB client on Android all recipients shows as their phone numbers and not their names. I've been trying to restart the BB server, played around with the BB server 'refresh contact' options - no progress. Then I saw that BB server offers to signin using google to sync the google contacts directly to make them available to the BB clients. I tried to signin with my daughter's account, got prompted to sign in using the app's web view but I got the error "Accounts managed by Family Link are not allowed to sign in here.". Note that I dont get that error when I signin in Chrome on the same device. It looks like something is off with the app's webview. Question - could u update the app so it doesn't show a web view and instead prompt to signin in Chrome directly? (like Slack does) Thanks in advance! PS - awesome work building BB!!! The BB client doesn't get contacts from the server. It should show contact names provided you've granted the contacts permission to the android app and the contacts exist on the phone. Hi Joel, thank you for the super quick response! I just manually granted the BB client access to the contact app and it immediately solved the problem, thanks a lot!!
2025-04-01T06:36:47.165254
2023-10-12T01:45:12
1939028573
{ "authors": [ "Pressacco" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:451", "repo": "BlueDotBrigade/weevil", "url": "https://github.com/BlueDotBrigade/weevil/issues/405" }
gharchive/issue
The focused filter will have the most visual real-estate Tests #404 Test Steps Using Weevil, open a log file. Click on the Inclusive filter TextBox The Inclusive filter width should expand to take up more visual real-estate than the Exclude filter. It is not worth the investment in time to automate this test.
2025-04-01T06:36:47.219805
2022-04-27T07:59:57
1216958433
{ "authors": [ "Ferruck", "jkloetzke" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:454", "repo": "BobBuildTool/basement", "url": "https://github.com/BobBuildTool/basement/pull/133" }
gharchive/pull-request
utils/xz-utils: Provide a dev package Upgrade to 5.2.5 whilst we're at it. This probably needs some fixes in the example repos. I'll provide those if this MR is considered fine for inclusion. LGTM. I still need to update the example repos to the new split layout... :see_no_evil: I still need to update the example repos to the new split layout... :see_no_evil: I'll do this too. I'll do this too. I'll postpone the update of the examples until BobBuildTool/basement-gnu-linux#1 is merged. I just want to follow up on this: is there anything I've missed? AFAICS, when https://github.com/BobBuildTool/basement-gnu-linux/pull/1 is merged, this and #134 can be merged, too. Afterwards, I'll update the example repos to use the latest basement including the split. I just want to follow up on this: is there anything I've missed? No. I just have been lazy... :see_no_evil:
2025-04-01T06:36:47.223018
2022-12-07T19:31:56
1482683991
{ "authors": [ "rebeccahongsf" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:455", "repo": "BobaTalks/bobatalks.github.io", "url": "https://github.com/BobaTalks/bobatalks.github.io/pull/10" }
gharchive/pull-request
issue#4 | @rebeccahongsf | update ordering to work on multiple browsers Summary update to ascending order which works on multiple browsers Issue ticket number and link #4 Checklist before requesting a review [x] I have performed a self-review of my code Screenshots (if applicable) Firefox: Chrome: @jwu910
2025-04-01T06:36:47.226004
2019-11-29T04:30:44
530150700
{ "authors": [ "Bobgy", "gaoning777" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:456", "repo": "Bobgy/manifests", "url": "https://github.com/Bobgy/manifests/pull/1" }
gharchive/pull-request
Multi user deployment changes @gaoning777 Here are our deployment changes. Thanks Yuan Superseded by https://github.com/Bobgy/manifests/pull/2
2025-04-01T06:36:47.244972
2017-11-22T14:04:58
276079094
{ "authors": [ "Bogdan-Lyashenko", "j2l" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:457", "repo": "Bogdan-Lyashenko/js-code-to-svg-flowchart", "url": "https://github.com/Bogdan-Lyashenko/js-code-to-svg-flowchart/issues/14" }
gharchive/issue
CLI ETA? Hey, Thanks for this one! I'd like to test on a React Native using CLI. Is there an ETA? Pasting in a code editor didn't work :( @j2l , hi, thanks. Can you please show which code doesn't work? Or what error are you getting in dev-tools. Thanks. Thanks for replying quickly! First I tried to paste some react native code in a online editor (from Under the hood ReactJS) and svg wasn't produced, no error seen. Then I tried CLI: npm i -g js2flowchart installed it but jsflowchart is a unknown command. Hey, it wasn’t implemented yet, that’s why it didn’t work like global install. I pushed a CLI feature yesterday, please update js2flowchart from npm and try again. Thanks. Let me know if still have issues. Cool! But it doesn't like some react native: js2flowchart App.js Error at parseCodeToAST: Unexpected token (51:8) C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:18131 throw e; ^ SyntaxError: Unexpected token (51:8) at Parser.pp$5.raise (C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:13702:13) at Parser.pp.unexpected (C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:11009:8) at Parser.pp$1.parseClassProperty (C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:11819:50) at Parser.pp$1.parseClassBody (C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:11764:34) at Parser.pp$1.parseClass (C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:11654:8) at Parser.pp$1.parseExport (C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:11890:19) at Parser.pp$1.parseStatement (C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:11132:74) at Parser.pp$1.parseBlockBody (C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:11516:21) at Parser.pp$1.parseTopLevel (C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:11026:8) at Parser.parse (C:\Users\pm\AppData\Roaming\npm\node_modules\js2flowchart\dist\js2flowchart.js:10921:17) @j2l, there is a code state = { appIsReady: false }; inside of class definition. It's not like es6 way of doing that and I use es6-only parser which doesn't like it (it's not valid actually from that point of view), so it breaks there. So if you remove that 3 lines and try without it, you'll see it works fine. Indeed :) Fixed for me. Cheers!
2025-04-01T06:36:47.246966
2019-09-19T10:18:04
495707075
{ "authors": [ "BohdanNikoletti", "Harishreddy122" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:458", "repo": "BohdanNikoletti/SFaceCompare", "url": "https://github.com/BohdanNikoletti/SFaceCompare/issues/15" }
gharchive/issue
Not Working This is not working for me giving results as matching for all the faces This lib is build on top of OpenFace ML. I can't improve it for now and in nearest feature.
2025-04-01T06:36:47.275035
2023-05-23T13:03:16
1722043563
{ "authors": [ "ssddanbrown", "techauthoruk" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:459", "repo": "BookStackApp/BookStack", "url": "https://github.com/BookStackApp/BookStack/issues/4257" }
gharchive/issue
Install BookStack on Synology NAS Attempted Debugging [X] I have read the debugging page Searched GitHub Issues [X] I have searched GitHub for the issue. Describe the Scenario Hi I am trying to create a backup install of my main BookStack install on a Synology DS414j NAS. I have two real questions: How to install BookStack on the NAS How to 'mirror' my server install to the NAS so they are synchronised Question 1: I have updated the NAS to the latest allowed DSM version (DSM 7.1.1-42962 Update 5), and updated Web Station. I have also installed PHP 8.0. However, all the tutorials I have found say to install Docker or Portainer to get a working BookStack install. My problem is that neither package are available (as far as I can see), so I can't even set up a container to install BookStack to! Has anyone recently managed to create a BookStack install on this NAS? Question 2: Regardless of whether I get the NAs install working, what is the best way of synchronising two installs? Cani I rsync the install from one device to another? Can I restore an sql dump and restore the images, etc? I am mindful of the fact that the .env settings for the db may be different. I am no expert on this stuff, so would welcome some pointers. Mark Exact BookStack Version Latest version Log Content No response PHP Version 8.x Hosting Environment Server - Ubuntu 22.04LTS NAS....? Hi @techauthoruk My problem is that neither package are available (as far as I can see), so I can't even set up a container to install BookStack to! Yeah, the lower value range of Synology boxes (Which I think are those with j at the end of the model name) won't have certain packages due to system limitations. Has anyone recently managed to create a BookStack install on this NAS? It likely is possible to get it running, but it might be messy to do outside of docker due to the requirements needed. It's not really what a Synology NAS is suited for tbh. Regardless of whether I get the NAs install working, what is the best way of synchronising two installs? Cani I rsync the install from one device to another? What is your actual goal in this? There are methods but no easy way for a bi-directional sync. For a single-direction sync you can indeed dump the database and copy the files over. Our backup and restore docs detail the mechanics of this. As an easier potential option, The latest release includes a system CLI that allows potentially easier backup/restore. It's in early alpha and can still be subject to bugs/issues, but it automates and standardises a lot of the process in the docs linked above. I'm not confident it'd run without some pains on a Synology system though. Personally, through experience I've found it's more hassle than it's worth to stretch the use-case of something like a Synology box, which ideally you'd want to remain reliable in it's primary purpose. You'd be going against the grain and there'd be little existing guidance/docs/experience to help when you run into issues (as you're finding now). I have a separate little intel NUC box to run apps, running proxmox to separate/compartmentalise things via VMs or LXC containers. I have a little overview of my setup here. If you just need to have a backup of your system though, you could maybe just look to use the new BookStack system CLI to create backups then store the resulting ZIP on your NAS like any other file. @ssddanbrown thank you for all your comments. I have abandoned the idea of using the Synology now - there is just no reliable way to get Docker installed. My goal was actually to have a replica of my main bookstack instance installed elsewhere for reference. The sync would only ever be one way (main instance to backup). The reason behind this - I keep all my notes and instructions for the management of my software apps in bookstack. My server went down a few days ago, and of course I couldn't access my notes to do reinstalls, etc! I wanted a backup on another machine/device so that I could see my notes while I sorted the issues. I do routinely back up the bookstack instance in line with the comments on your site, so it's not a problem restoring bookstack if needed. I tried this morning restoring the sql backup from my main instance into a bookstack install on another machine, but this didn't work - I get a 419 error when trying to log in. I'm going to play around a little more to see if I can find out why. Thank you again for your comments. OK, so resolved now. My 419 issues seemed to be due to an incorrect database pasword. Changed the password, and everything is working as it should, so closing this.
2025-04-01T06:36:47.279956
2023-11-06T21:02:21
1980053033
{ "authors": [ "slimninja" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:460", "repo": "BookStackApp/BookStack", "url": "https://github.com/BookStackApp/BookStack/issues/4654" }
gharchive/issue
403 error only when inserting specific code block in editor Describe the Bug I've found an odd bug that occurs specifically with one piece of code within my instance. Narrowing this down took quite a bit. The page I'm using is a reference page where I have approx 15 different SQL queries for reference. When inserting the following text as a code block in any page, "sp__dbutilisation tempdb;" the save button leads immediately to a 403 error page. My bookstack instance immediately slows down for the next few minutes. When I remove this code block, the page saves fine. I'm assuming that that the db command is getting executed on save? I updated to the latest instance and see the same issue. DB is MySQL Steps to Reproduce Create new page named "Queries" or anything else Create a code block with the following code (code block, not in-line code) sp__dbutilisation tempdb; Try saving page Immediately see 403 error Expected Behaviour Page saving without 403 error Screenshots or Additional Context No response Browser Details Chrome Exact BookStack Version 2023.10.1 Searched a bit more. Met me try #555 and #1792 and will update here if the issue persists. Can not recreate the issue on the demo site. Solved by disabling apache mod_security m
2025-04-01T06:36:47.288496
2024-07-08T20:51:44
2396546387
{ "authors": [ "ssddanbrown", "thickconfusion" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:461", "repo": "BookStackApp/BookStack", "url": "https://github.com/BookStackApp/BookStack/issues/5107" }
gharchive/issue
Internally Hosted Draw.IO is not usable, "This content is blocked. Contact the site owner to fix the issue." Describe the Bug Similar to #2285 , I am getting a gray page in Chrome that says "This content is blocked. Contact the site owner to fix the issue." I have the following environment variables set for the container: DRAWIO=http://<IP_ADDRESS>:8080/?embed=1&proto=json&spin=1&configure=1&stealth=1 I have also attempted to modify this environment variable: ALLOWED_IFRAME_SOURCES= I've tried: http://<IP_ADDRESS>:8080 https://<IP_ADDRESS>:8443 http://<IP_ADDRESS>* The only one that "works" is if I make it ALLOWED_IFRAME_SOURCES="*", which seems like a security vulnerability even if I'm running this on a LAN. Note: I can access the plain old Draw.IO interface just fine: http://<IP_ADDRESS>:8080, and it loads. Steps to Reproduce Edit a page, click the icon to work on a Draw.io image. Expected Behaviour I expect to load into a Draw.IO instance. Screenshots or Additional Context No response Browser Details Chrome and Edge on Windows 11 Exact BookStack Version v24.05.2 Hi @thickconfusion, You shouldn't need to adjust the iframe sources since BookStack will look to automatically add any custom drawio URL, where set, to the CSP rules. Maybe our custom handling is tripping up any additional rules you're adding. It does look though like we are not currently handling scenarios where non-protocol-standard ports are used. I've marked this to be tested for next patch, against a custom-ported drawio instance. Dev reference https://github.com/BookStackApp/BookStack/blob/78ebcb6f38ee7a984b26cd56dff882ae9d7e9f95/app/Util/CspService.php#L144 Sure, I was just saying that we attempt to handle this so you shouldn't have to set the iframe sources, but we currently don't handle custom defined ports. I've now fixed port handling via 897bb338f956245e2c86bda6cd5c6a67711f9448, with testing to cover, which will be part of the next patch release so I'll therefore close this off. Not sure why your custom ALLOWED_IFRAME_SOURCES additions did not work, since I could work around this on my dev instance via this method, but could be down to browser specifics or configuration changes not take place when expected. If you still have issues after the next patch release feel free to still comment here for further investigation.
2025-04-01T06:36:47.310244
2024-07-08T19:12:58
2474287476
{ "authors": [ "gabitoesmiapodo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:462", "repo": "BootNodeDev/dAppBoosterLandingPage", "url": "https://github.com/BootNodeDev/dAppBoosterLandingPage/issues/2" }
gharchive/issue
[UI] Token input (add missing part) The bottom part of this is still missing. This is kind of unnecessary, so it won't be implemented.
2025-04-01T06:36:47.314421
2024-01-12T15:37:44
2079122859
{ "authors": [ "BorisNA" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:463", "repo": "BorisNA/wordforms", "url": "https://github.com/BorisNA/wordforms/issues/3" }
gharchive/issue
FR: short script to (auto)test wordforms? A script should be created to (possibly auto-) test a wordform file. this script should be extensible need to run it on all existing wordform files Common error: consecutive colons (":") in a line. Possibly due to errors when converting from hunspell For example forms_TR.txt:18516 Script is at 3c019616510f585dff69e00dbcab23bce6df1274 Fix is at 0f9abb57b65d2e05e518fa0df75344105fd49e1a The script is created, all errors are fixed (PL - with #1). Warning possibly could be fixed with #4
2025-04-01T06:36:47.317011
2017-04-21T08:08:06
223309447
{ "authors": [ "devsli" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:464", "repo": "BosNaufal/react-scrollbar", "url": "https://github.com/BosNaufal/react-scrollbar/issues/20" }
gharchive/issue
Links are not working on mobile devices Hi Naufal, thanks for magnificent component! I experienced little bug recently, links are not clickable inside of <ReactScrollbar /> on mobile devices Example: https://spreecode.github.io/react-scrollbar/ Safari iOS 9.3.5: Not working Safari iOS 10.2: Not working Opera Mini: Not working Mobile Chrome: Tricky (sometimes working) Mobile Firefox: Not working I think the problem is here, when you comment out these lines then links are clickable. Obviously when commented out the scrolling behavior is not working as expected.
2025-04-01T06:36:47.339190
2023-09-22T10:48:39
1908671597
{ "authors": [ "BottlecapDave", "Jonaz80" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:465", "repo": "BottlecapDave/HomeAssistant-OctopusEnergy", "url": "https://github.com/BottlecapDave/HomeAssistant-OctopusEnergy/issues/416" }
gharchive/issue
Binary Sensor to send notification in advance of operation. Describe the feature Send a notification when new rates published of when a binary sensor will start/stop. Use case: dishwasher on/off is momentary switch but start can be delayed by the user. This advanced/ 'heads-up' notification message would allow recipients to know the correct time to set delay to get lowest overnight price for the dishwasher cycle. Expected behavior Trigger: Tomorrow's rates published binary 'notification' sensor run time calculated notification sent via mobile app containing start/stop times that meet the conditions of the sensor, or a use configurable message with these times as values that can be added to the message with whatever formatting the user chooses. Hello. If I understand your scenario correctly, a new sensor isn't needed as you can do this with existing HA functionality. The times for when the sensor is due can be found as attributes on the sensor. I for instance have a template sensor which extracts this time out vacuum_time: friendly_name: Vacuum time icon_template: mdi:clock value_template: >- {{ as_local(state_attr("binary_sensor.octopus_energy_target_vacuum", "next_time")).strftime('%H:%M') }} and then I have an automation that sends an alert at a specific time to tell me when my vacuum is due on - alias: Cleaning - Clean house notice trigger: - platform: event event_type: morning_reminder condition: - condition: state entity_id: input_boolean.is_cleaning_scheduled state: 'on' action: - event: notify_channels event_data_template: mode: speaker title: Cleaning occurring message: > {{ [ "I'm due to clean today at " + states("sensor.vacuum_time") + ". If you don't want this, ask me to 'turn off cleaning'.", "I'll be cleaning today at " + states("sensor.vacuum_time") + ". If it's not convenient, ask me to 'turn off cleaning'.", ] | random }} If you're wanting to send an alert x minutes before then I have a template sensor which stores the datetime the notification should go off vacuum_warning_one: friendly_name: "Vacuum warning one" device_class : timestamp value_template: >- {% set date = state_attr("binary_sensor.octopus_energy_target_vacuum_working_hours", "next_time") %} {% if date != None %} {{ date - timedelta( minutes = 30 ) }} {% else %} {{ as_datetime("2021-01-01T00:00Z") }} {% endif %} and then an automation looks like - alias: Cleaning - Clean house warning trigger: - platform: time at: sensor.vacuum_warning_one condition: - condition: state entity_id: input_boolean.is_cleaning_scheduled state: 'on' action: - event: notify_channels event_data_template: mode: speaker title: Cleaning occurring message: > {% set target_time = as_local(state_attr("binary_sensor.octopus_energy_target_vacuum", "next_time")) %} {% set minutes = ((as_timestamp(target_time) - as_timestamp(now())) / 60) | round(0) %} {{ [ "I'm planning to clean the house in " + minutes|string + " minutes. If you don't want this, ask me to 'turn off cleaning'.", "I'll be cleaning the house in " + minutes|string + " minutes. If it's not convenient, ask me to 'turn off cleaning'.", ] | random }} As you can see I use an input boolean to determine if the vacuum should go on or not, which I use as a condition for an automation when triggering the vacuum - alias: Cleaning - Start Automatic While Everyone Out (Non Work Day) id: 5418B9B8-3B54-49A2-8DCA-C9846B67F751 trigger: - platform: state entity_id: binary_sensor.octopus_energy_target_vacuum to: 'on' condition: - condition: state entity_id: input_boolean.is_cleaning_scheduled state: 'on' action: - event: start_cleaning Let me know if I've misunderstood your scenario. Thank you, this looks exactly what I am looking for... and more ! Much appreciated 👍 Excellent. I'll close this feature request then :)
2025-04-01T06:36:47.343940
2017-03-03T05:19:14
211597987
{ "authors": [ "Bouke", "vikram-lapenatech" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:466", "repo": "Bouke/django-two-factor-auth", "url": "https://github.com/Bouke/django-two-factor-auth/issues/196" }
gharchive/issue
Incorrect gateway used? Even if i am giving valid twilio credentials , the message "your token is 1234 " is diplaying. TWO_FACTOR_CALL_GATEWAY = 'two_factor.gateways.twilio.gateway.Twilio' TWO_FACTOR_SMS_GATEWAY = 'two_factor.gateways.twilio.gateway.Twilio' TWILIO_ACCOUNT_SID = 'AC86a1663acf************affba5d3' TWILIO_AUTH_TOKEN = '029be37d7**********1c0257c9b' TWILIO_CALLER_ID = '+919781539134' #verified caller ID It appears as if you're using the example project, as it would be the only explanation for that message appearing on your screen. Please follow the installation/configuration instructions, somewhere it tells you to set the correct TWO_FACTOR_CALL_GATEWAY and TWO_FACTOR_SMS_GATEWAY.
2025-04-01T06:36:47.389272
2022-08-22T01:40:19
1345682627
{ "authors": [ "IvanJRCH", "JoelBonetR" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:467", "repo": "BrainJS/brain.js", "url": "https://github.com/BrainJS/brain.js/issues/836" }
gharchive/issue
Chatbot ? Hola cómo están 👍, alguien a usado brain.js? estoy programando un chatbot pero la manera en que lo hago no me gusta mucho, a si que quería saber si alguien que lo a ya hecho antes me puede dar un ejemplo de código con brain.js para enseña al bot a escribir y a responder al texto del usuario de una manera coherente luego solo faltaria clasificar algunas palabras para ejecutar comandos, ví este ejemplo usando redes neuronales recurrentes: const trainingData = [ 'Jane saw Doug.', 'Spot saw himself.', 'Doug saw Jane.' ]; const lstm = new brain.recurrent.LSTM(); const result = lstm.train( trainingData, { iterations: 1000 } ); const run1 = lstm.run('Jane'); const run2 = lstm.run('Spot'); const run3 = lstm.run('Doug'); console.log('run 1: Jane' + run1); console.log('run 2: Spot' + run2); console.log('run 3: Doug' + run3); Se supone que esto le enseña a la red a crear fraces lo probé y le agregue mis propios datos de entrenamiento pero tengo como resultado puras cosas sin sentido, no quiero usar servicios externos, quiero crear mi propio chatbot, alguien ha usado brain.js y hecho esto? Pero de una manera que no se tenga que clasificar el texto? Quiero que la IA aprenda a escribir primero español pasándole un archivo grande lleno de texto, luego como la red ya sabe escribir quiero procesar el texto del usuario input y como salida output solo una sugerencia de como debería responder la red para que la conversación se sienta natural y he evitarme tener que clasificar a mano un montón de texto Hola IvanJRCH, escribo la respuesta en inglés para mantener una cohesión lingüística en el repositorio: I understand that you want to train your AI, please see the training reference. You'll need to work out the training before the chatbot answers make sense at all, I suggest you to define the use-cases, make sure your intents are distinct and that each intent contains many utterances. Also note that your job isn’t done after your chatbot has been deployed. Continuous improvement is important for a successful result and identifying situations where your chatbot needs more training will give you important insights about it. I don't spot any bug or problem related with Brain.js here so I'd recommend to close this issue. hello, thank you very much for your answer, I have created a chatbot, but the chatbot only responds with predefined responses that are already in training, what I would like is for the AI to learn to write, and to respond to the user with consistency only by passing examples of a normal conversation that I will take from a plain txt file, but that the AI does not respond with predefined text but creates its own dialog, then I plan to use another neural network to classify the intention of the text of the AI not of the user, in order to execute commands, I saw the example of using recursive.LSTM(); but notice that it only repeats what is already in the training, how can I make the AI create its own conversations but not with predefined texts, is using LSTM the right way? Could you give me an example? I don't understand how to put that in the code. Also, I don't understand the neural network well either. Thank you very much for answering. Hi IvanJRCH, I'm not an expert on ML, just reached this project as a way to learn more in my free time. Also got this resource that may help (didn't read it yet) https://www.deeplearningbook.org/ I suggest you to interact with the community through other sites more prone to the conversation such stackoverflow, dev.to or any other forum that's specific to have conversations around the niche of ML. This way we don't convert this issue tracker into something else. Have a great day and hope you find the way to solve your issue :) Thank you very much, I'll take a look at the link.
2025-04-01T06:36:47.390889
2023-11-15T15:04:07
1994959670
{ "authors": [ "neuronflow" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:468", "repo": "BrainLesion/panoptica", "url": "https://github.com/BrainLesion/panoptica/issues/38" }
gharchive/issue
rename to SQ_ASSD instance_ASSD should be renamed to SQ_ASSD to be consistent with the other metrics. same goes for the respective SD. should we also introduce PQ_ASSD?
2025-04-01T06:36:47.394636
2017-11-21T21:19:33
275863193
{ "authors": [ "EvangelosG", "sojanpr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:470", "repo": "BranchMetrics/android-branch-deep-linking", "url": "https://github.com/BranchMetrics/android-branch-deep-linking/pull/508" }
gharchive/pull-request
Ensuring backward compatibility in case using a deprecated BUO method for latest SDK BUO# addContentMetadata() is deprecated and new integration should use BUO#setContentMetadata(). This fix ensure backward compatibility in case the app is still using the deprecated methods @aaustin @EvangelosG @derrickstaten 👍
2025-04-01T06:36:47.424950
2024-02-21T08:08:58
2146073933
{ "authors": [ "Breakthrough", "babyta" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:471", "repo": "Breakthrough/PySceneDetect", "url": "https://github.com/Breakthrough/PySceneDetect/issues/381" }
gharchive/issue
url input ,1080p video much slower as title,when test,i found that problem。Looking forward to your reply I'm not sure this is enough information to go by, could you please provide: Description: Describe what the bug or issue is (e.g. crashes when setting X) and how it can be reproduced. Command: Place a full copy of the command line options you are using here, for example: scenedetect -i some_video.mp4 -s some_video.stats.csv -o outdir detect-content --threshold 28 list-scenes save-images Output: Copy the output of running the application here. Where possible, generate a debug log by adding -v debug -l BUG_REPORT.txt to the beginning of your command, and attach BUG_REPORT.txt to your issue. Environment: The operating system and how you installed PySceneDetect may be relevant to the issue. Please run scenedetect version --all and copy the output here, or provide other details on how PySceneDetect was installed. Media/Files: Attach or link to any files relevant to the issue, including videos (or YouTube links), scene files, stats files, and log output. 这是来自QQ邮箱的假期自动回复邮件。您好,我最近正在休假中,无法亲自回复您的邮件。我将在假期结束后,尽快给您回复。 https://drive.google.com/file/d/1h5tztZkpV6ziJzgtVieJqcnhF4Bo82f_/view?usp=sharing =================This is the url =================This is a speed comparison. The URL input is very slow, but the same video is downloaded locally as input and the strips are split very quickly. from scenedetect import open_video, SceneManager, split_video_ffmpeg from scenedetect.detectors import ContentDetector from scenedetect.video_splitter import split_video_ffmpeg video_path = '1080pvdieo.mp4' threshold=27.0 Open our video, create a scene manager, and add a detector. video = open_video(video_path) scene_manager = SceneManager() scene_manager.add_detector( ContentDetector(threshold=threshold)) scene_manager.detect_scenes(video, show_progress=True) scene_list = scene_manager.get_scene_list() ================= this is code URL input is handled by OpenCV if you don't change anything, have you tried using a different backend? https://www.scenedetect.com/docs/0.6.2/api/backends.html Traceback (most recent call last): File "C:\Users\Administrator\Desktop\test.py", line 9, in video = open_video(video_path, backend='pyav') File "D:\ProgramData\miniconda3\envs\py39pt1121\lib\site-packages\scenedetect_init_.py", line 143, in open_video return backend_type(path, framerate, **kwargs) File "D:\ProgramData\miniconda3\envs\py39pt1121\lib\site-packages\scenedetect\backends\pyav.py", line 107, in init self._io = open(path_or_io, 'rb') OSError: [Errno 22] Invalid argument When I use pyav as backends, I get an error directly. cv2.VideoCapture When reading a URL, if the video resolution corresponding to the URL is very high and the data rate is very high, it will be very slow. Thank you for following up on this, I appreciate it. If you suspect this is due to OpenCV and isn't just an issue with using ffmpeg (or which ever backend OpenCV itself is using to process the stream), you might want to reach out in their repo. Best regards!
2025-04-01T06:36:47.449431
2019-08-20T17:52:11
482998457
{ "authors": [ "BretFisher", "syntaqx" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:472", "repo": "BretFisher/dogvscat", "url": "https://github.com/BretFisher/dogvscat/issues/24" }
gharchive/issue
RexRay doesn't work on Docker latest (19) https://github.com/BretFisher/dogvscat/blob/master/stack-rexray.yml Using docker stack deploy -c stack-rexray.yml rexray gives back the following error: network "bridge" is declared as external, but it is not in the right scope: "local" instead of "swarm" Would love to see what those TODO notes resolve into as well as this has been very useful for my swarm. Did that work for you?
2025-04-01T06:36:47.460808
2017-01-31T15:32:47
204332218
{ "authors": [ "JoshuaKlassen", "ctwomey1" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:473", "repo": "Brightspace/d2l-my-courses-ui", "url": "https://github.com/Brightspace/d2l-my-courses-ui/pull/289" }
gharchive/pull-request
US81659 - Update lang terms Replace English default terms in non English locales with the proper translated term Add NL locale Tested and looks fine: Test cases: Cycle through all lms languages in chrome and ensure no exceptions are thrown in the console Ensure new lang terms are reflected in the app by setting updates on, and setting courses into inactive, inactive started, inactive ended, starting later, already ended.
2025-04-01T06:36:47.467164
2024-03-11T20:17:22
2180144061
{ "authors": [ "Dan-DeAraujo", "devpow112" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:474", "repo": "Brightspace/test-reporting-node", "url": "https://github.com/Brightspace/test-reporting-node/issues/149" }
gharchive/issue
QE-147 Include error message with failed (flaky?) tests Would help understand test failure reason trends, and we can focus de-flaking efforts on areas of higher impact. EG: This run failed with Error: locator.fill: Target page, context or browser has been closed; which is clearly not something we can resolve directly, but if we notice an uptick post-Playwright update, we can complain to them about it OTOH, if we notice alot of tests failing to find locators or whatever, maybe we need better training on how to pick good ones Not sure if we can get it for flaky tests (ie: the first run of a retried set) too, but if we can that'd be even better Duplicate of https://github.com/Brightspace/test-reporting-node/issues/87. Gonna try and send error message, file, line and column values for each test when an error occurs. For retired tests I'm just gonna send the last error for now. If we find we need all error information or something else we can explore that separately. Moved to Jira
2025-04-01T06:36:47.472788
2024-04-25T00:36:38
2262402896
{ "authors": [ "BlameFelix", "TaylorT52", "syndiate" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:475", "repo": "Brikwerk/nxbt", "url": "https://github.com/Brikwerk/nxbt/issues/152" }
gharchive/issue
Ability to hold down one button throughout runtime of NXBT macro It seems that the NXBT macro scripting "language" only supports subsequent inputs, and not continuous inputs that may continue to be sent simultaneously alongside other inputs. ^ also looking for this feature Hi, idk your exact usecase but checkout this: https://github.com/Brikwerk/nxbt/issues/151#issuecomment-2124226564 @BlameFelix Thanks! I'll look into it soon and report the results here after.
2025-04-01T06:36:47.481047
2016-09-05T15:32:44
175094007
{ "authors": [ "Cheesebaron" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:476", "repo": "BruelAndKjaer/Chafu", "url": "https://github.com/BruelAndKjaer/Chafu/issues/8" }
gharchive/issue
Picking video or image externally crashes app After having added Video preview support, the observer does not correctly take in account videos when reordering the collection, when a picture/video is taken or deleted outside of the app and app is brought into foreground again. Fixed in https://github.com/BruelAndKjaer/Chafu/commit/b291db7a3330f2afc3069428a49fb0a708f38df9 Using ObservableCollection instead and updating indexes using CollectionChangedEvent
2025-04-01T06:36:47.511860
2024-03-07T10:21:45
2173511026
{ "authors": [ "Bugs5382", "krisc-informatica" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:477", "repo": "Bugs5382/node-hl7-server", "url": "https://github.com/Bugs5382/node-hl7-server/issues/61" }
gharchive/issue
res.getAckMessage() is not defined In the example the code uses getAckMessage() but that method is not defined for sendResponse. const messageRes = res.getAckMessage() I will review tonight and have a fix out maybe by the end of the week. This is an error just in my documentation. The "sender" handler: const listener = server.createInbound({port: 3000}, async (req, res) => { const messageReq = req.getMessage() expect(messageReq.get('MSH.12').toString()).toBe('2.7') await res.sendResponse('AA') //Here, we are sending back to the client the Ack or Failure. This is where the AA or AF message gets generated. }) and in the client: const outbound = client.createConnection({ port: 3000 }, async (res) => { const messageRes = res.getMessage() expect(messageRes.get('MSA.1').toString()).toBe('AA') // this is where we are confirming the message response sent by the server. dfd.resolve() }) So I can include: getAckMessage(): Message | undefined { return this._ack } ...but it might be undefined if called before sending the AA or AF. :tada: This issue has been resolved in version 2.1.0-beta.3 :tada: The release is available on: npm package (@develop dist-tag) GitHub release Your semantic-release bot :package::rocket: :tada: This issue has been resolved in version 2.2.0-beta.1 :tada: The release is available on: npm package (@develop dist-tag) GitHub release Your semantic-release bot :package::rocket:
2025-04-01T06:36:47.522970
2024-09-17T22:18:57
2532274870
{ "authors": [ "Iwaslazkis", "NicoleOkamoto", "Sahib4", "ViniOkamoto", "hzz4343", "khnatiuk", "namanparashar123" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:478", "repo": "Builder-s-League/BuildersLeague-Edition1", "url": "https://github.com/Builder-s-League/BuildersLeague-Edition1/issues/4" }
gharchive/issue
[1|Topic Detail Page] Breadboard 🍰 Appetite 🍰 Property Value Points 3 points Difficulty rating ★★☆☆☆ Maximum Assignees 2 person 🤔 Problem Statement Make a breadboard for each screen in the fat-marker sketch. Every place and affordance should be accounted for, but don't focus on making it look pretty. This should be essentially a bare HTML file, with no JS, and only CSS if required for basic layouts. Implement fake navigation and interactions using normal <a> tags. This purely for getting an interactive "skeleton" up, and seeing in real life whether this design would work, or has any major problems. Please see this page for more info on what your breadboard should be like (and, if you want understand the fat-marker sketch and places affordances drawing better, please see this page). 🧪 Required Tests None (but the Code Wizard will review whether or not your code warrants some tests! Review the Code & Testing Guidelines for more info) ⚠️ Careful about: Don't make a separate desktop design. There should be one design, that works both in landscape and portrait orientations. 🤖 Technologies focused on in this feature React Tailwind Shadcn I can work on this one @puma @Sahib4 it's yours! You're now assigned. I can work on this as well @puma @khnatiuk I added in @namanparashar123 , but he is talking with @Sahib4 right now. @namanparashar123 @Sahib4 can you come see me? There are folks ready to work on Phase 2 but we're unsure whether you have completed Phase 1 yet. @ViniOkamoto please let a PuMA know if you want to be added to this issue. On it already! @khnatiuk Please add me to it @puma Yay!
2025-04-01T06:36:47.523759
2024-01-26T17:14:26
2102567325
{ "authors": [ "samijaber" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:479", "repo": "BuilderIO/mitosis", "url": "https://github.com/BuilderIO/mitosis/pull/1345" }
gharchive/pull-request
Feature/pnpm Description to fix changesets publishing Closing as there are way too many "@babel/" conflicts to sift through
2025-04-01T06:36:47.524867
2023-03-03T18:16:11
1609067660
{ "authors": [ "nnelgxorz", "qq99" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:480", "repo": "BuilderIO/qwik", "url": "https://github.com/BuilderIO/qwik/issues/3250" }
gharchive/issue
[📖] Unclear instructions when it comes to testing/running your SSG'd site Suggestion node server/entry.ssr.js just returns instantly (verified I ran yarn build.server before), so it's not clear if this is old documentation for an older version I think you'd need something like npx http-server dist/ after running build, but that's npm specific. What's the cool kid way to run a simple http server locally these days?
2025-04-01T06:36:47.540871
2018-08-13T06:46:19
349905924
{ "authors": [ "Crecket", "L00Cyph3r" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:481", "repo": "BunqCommunity/bunqDesktop", "url": "https://github.com/BunqCommunity/bunqDesktop/issues/256" }
gharchive/issue
Missing linux-build from latest release and hashes The latest release (0.8.11) does not contain a 'linux-unpacked.tar.gz' file or anything similiar. Since I'm currently trying to maintain the Arch AUR package it would be nice if we could make sure this file exists in coming updates in an expectable filename. And if it's not too much of a hassle, it might be nice if the build-process creates SHA-hashes for the generated distributables (since it's somewhat bank/privacy-related). Of course I could generate hashes myself, but 'original' ones are probably trustworthier. I'll check why the unpacked file is suddenly no longer uploaded since 0.8.11, it might have to do with the hotfix I ran a while. For the hashes you can always check the build for that specific release. At the bottom of every travis build a list of hashes are created. I'll reply here for the correct ones for the current build when I fix it 👍 Added the download file and the checksum for it is 8f36b98ea79b1323ffe3968c2105a099073700c536ed005868f6c702de5cb68c. I'll make a habit out of adding a link to the travis build as well so you can find the hashes more easily next time and to keep things a bit more transparent. Thanks for reporting this btw, I completely missed it 👍
2025-04-01T06:36:47.552419
2018-12-01T08:15:15
386436429
{ "authors": [ "Burry", "JohanSF" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:482", "repo": "Burry/organizr-v2-plex-theme", "url": "https://github.com/Burry/organizr-v2-plex-theme/issues/13" }
gharchive/issue
Refresh button still not working in 2.7.0 I have 2.7.0 installed via the marketplace, running the organizr docker container with :latest tag. Clicking the refresh button on a card opens opens the info instead of refreshing the cover: I am currently using "Organizr V2 On-top of Burry" from https://github.com/Archmonger/Blackberry-Flat, but the issue persists even with that turned off. I have also tried clearing cache in my browser and in CloudFlare. Fixed in 2.8.0 Will test when update is available through the marketplace. Confirmed that it now works :D thank you!
2025-04-01T06:36:47.558368
2022-07-08T01:05:15
1298334102
{ "authors": [ "BrendanJM", "Kfelts", "akcode47", "chhopsky", "kevingigiano" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:483", "repo": "Buuntu/fastapi-react", "url": "https://github.com/Buuntu/fastapi-react/issues/192" }
gharchive/issue
Fresh build- front end errors After a clean install and trying to run this, the frontend locks up and I get the following errors: Warning: React does not recognize the `computedMatch` prop on a DOM element. If you intentionally want it to appear in the DOM as a custom attribute, spell it as lowercase `computedmatch` instead. If you accidentally passed it from a parent component, remove it from the DOM element. ReferenceError: Can't find variable: process TypeScript error in /app/src/index.tsx(8,4): 'Router' cannot be used as a JSX component. Its instance type 'BrowserRouter' is not a valid JSX element. The types returned by 'render()' are incompatible between these types. Type 'React.ReactNode' is not assignable to type 'import("/node_modules/@types/react-transition-group/node_modules/@types/react/index").ReactNode'. TS2786 6 | 7 | ReactDOM.render( > 8 | <Router> | ^ 9 | <App /> 10 | </Router>, 11 | document.getElementById('root') TypeScript error in /app/src/Routes.tsx(31,6): 'Switch' cannot be used as a JSX component. Its instance type 'Switch' is not a valid JSX element. The types returned by 'render()' are incompatible between these types. Type 'React.ReactNode' is not assignable to type 'import("/node_modules/@types/react-transition-group/node_modules/@types/react/index").ReactNode'. TS2786 TypeScript error in /app/src/views/SignUp.tsx(58,6): 'Redirect' cannot be used as a JSX component. Its instance type 'Redirect' is not a valid JSX element. TS2786 Any idea how to troubleshoot this? Which version of React and React Router are you using? React Router v6 made a lot of changes. Warning: React does not recognize the computedMatch prop on a DOM element. If you intentionally want it to appear in the DOM as a custom attribute, spell it as lowercase computedmatch instead. If you accidentally passed it from a parent component, remove it from the DOM element. ReferenceError: Can't find variable: process ``` TypeScript error in /app/src/index.tsx(8,4): 'Router' cannot be used as a JSX component. Its instance type 'BrowserRouter' is not a valid JSX element. The types returned by 'render()' are incompatible between these types. Type 'React.ReactNode' is not assignable to type 'import("/node_modules/@types/react-transition group/node_modules/@types/react/index").ReactNode'. TS2786 6 | 7 | ReactDOM.render( > 8 | <Router> | ^ 9 | <App /> 10 | </Router>, 11 | document.getElementById('root') ReactDOM.render has been changed to createRoot. 1 | Import { createRoot} from 'react-dom/client' 6 | const container = document.getElementById('root'); 7 | const root = createRoot(container!); 8 | root.render( 9 | <Router > 10 | <App /> 11 | <Router /> 12 | ); 'Switch' cannot be used as a JSX component. Its instance type 'Switch' is not a valid JSX element.``` In v6 'Switch' was removed entirely. 'Switch' is now 'Routes' The types returned by 'render()' are incompatible between these types. Type 'React.ReactNode' is not assignable to type 'import("/node_modules/@types/react-transition-group/node_modules/@types/react/index").ReactNode'. TS2786 There have also been changes made in the way that routes are rendered. This is what it looked like prior: <header className={classes.header}> <Route path="/login" component={Login} /> <Login /> <Route path="/signup" component={SignUp} /> <Route path="/logout" render={() => { logout(); history.push('/'); return null; }} /> <PrivateRoute path="/protected" component={Protected} /> <Route exact path="/" component={Home} /> </header> </div> </Routes> This requires some refactoring, so I'd recommend taking a look at what upgrades are needed for React 18 and react hooks 'Redirect' cannot be used as a JSX component. In v6 'Redirect' has been changed to 'Navigate'. Change 'useHistory' to 'useNavigate' and then 'Redirect' to 'Navigate' in the jsx component. I would appreciate it if someone could push a PR with the latest react and react-router. As it stands now, the frontend build fails, leading to the above error. They're using the version that's in the clean install, as described in the readme.md I see the same behavior following the instructions provided, and given I was installing this to start playing with and learning react, makes the whole thing pretty useless. Confirmed this is still a problem. I was hoping to use this as a basis for a pet project/learning. Unfortunately, standing up from scratch using README results in a broken state as mentioned by @Kfelts and @chhopsky
2025-04-01T06:36:47.577346
2021-06-21T07:41:28
925924323
{ "authors": [ "C0MM4ND", "tejainece" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:484", "repo": "C0MM4ND/dart-secp256k1", "url": "https://github.com/C0MM4ND/dart-secp256k1/issues/7" }
gharchive/issue
OOP rewrite Hi, Thanks for the great work. I intend to use this package for my Cryptocurrency project and rope this package in as dependency for my ninja-dart project. Right now, the library is very golang-like. Would you mind if I rewrite this package to be more OOP and Darty and send you a pull request? Thanks again! Of course you can and it is welcomed. But I will rewrite this with my new packages ecdsa and elliptic. Maybe it would be better that starting your contribution after reviewing my new code?
2025-04-01T06:36:47.703391
2017-10-18T19:19:32
266611402
{ "authors": [ "AABoyles" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:487", "repo": "CDCgov/MicrobeTRACE", "url": "https://github.com/CDCgov/MicrobeTRACE/issues/90" }
gharchive/issue
Video-style controls Add a UI Component to "Play" the network like a video. Also "Pause", "Speed up", "Slow Down" Added in 9acc6667bb0d0a68428fcffc6c440922a177c04e
2025-04-01T06:36:47.709120
2024-02-01T16:14:33
2112891415
{ "authors": [ "adamdoe", "mpallansch" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:488", "repo": "CDCgov/cdc-open-viz", "url": "https://github.com/CDCgov/cdc-open-viz/pull/1062" }
gharchive/pull-request
DEV 6811 Fixes issues from testing: If you remove the Latitude/Longitude configuration after selecting it, you get an unrecoverable error Can we have the cities we support by default be placed, even if others are placed using lat/long? Example, "New York City" in the data with no lat/long If you switch to pin, then switch back, the dots disappear because the "geoCodeCircleSize" or some setting is now gone Need to expose the "Geocode Circle Size" setting when lat/long is used. Could that default to the size of the normal city circles and allow folks to adjust it whether lat/long is used or not? If you filter the map to only show some locations, the custom locations still appear @mpallansch - I tested with the world geo code example file. I'm still seeing the circle appear for NYC when I have the location set to United States. Same with having the location set to Alaska. @mpallansch - I tested with the world geo code example file. I'm still seeing the circle appear for NYC when I have the location set to United States. Same with having the location set to Alaska. Looks like that file has erroneous lat/long data for each point (US, Alaska, and NYC all have NYC lat/long data): { "Country": "New York City", "Cases": 300, "Category": "Has not historically reported monkeypox", "AsOf": "11 Jul 2022 5:00 PM EDT", "longitude": "-74.006", "latitude": "40.712" }, { "Country": "United States of America", "Cases": 10, "Category": "Has not historically reported monkeypox", "AsOf": "11 Jul 2022 5:00 PM EDT", "longitude": "-74.006", "latitude": "40.712" }, { "Country": "Alaska", "Cases": 500, "Category": "Has not historically reported monkeypox", "AsOf": "11 Jul 2022 5:00 PM EDT", "longitude": "-74.006", "latitude": "40.712" } I don't think this will be an issue for valid use cases.
2025-04-01T06:36:47.718501
2023-11-07T03:15:16
1980461808
{ "authors": [ "audreykwr" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:489", "repo": "CDCgov/prime-reportstream", "url": "https://github.com/CDCgov/prime-reportstream/issues/12108" }
gharchive/issue
Finalize messaging recommendations/application User story As a ReportStream user, I want to understand if ReportStream can help me so I can decide what to do. Background & context User tests revealed some need for further clarity in our messaging. Last sprint reviewed the current messaging in light of user test feedback and other current realities. This ticket builds on that to provide specific recommendations for alignment/approval. Open questions Working links Acceptance criteria [x] Recommendations documented and reviewed with necessary stakeholders [ ] Next steps planned (help create engr ticket) Full deck of updates here. Will create a separate doc specific to home page changes for the eng ticket.
2025-04-01T06:36:47.729932
2024-05-14T03:50:01
2294323015
{ "authors": [ "JohnNKing" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:490", "repo": "CDCgov/trusted-intermediary", "url": "https://github.com/CDCgov/trusted-intermediary/issues/1090" }
gharchive/issue
AL Gap Analysis - Test Orders Story As AL HCO, to ensure we'll receive all needed results data to an EHR, we need to relate the ReportStream results information to what will be received by the EHR. Pre-conditions [ ] Assumptions of prior or future work that's out of scope for this story Acceptance Criteria [ ] Message profiles are now finalized [ ] Publish list of transformation requirements for Intermediary [ ] Get approvals [ ] Create explicit stories for Intermediary transformation requirements Tasks [x] Draft side-by-side comparison of sender & receiver profiles [ ] Finalize side-by-side comparison of sender & receiver profiles [ ] Identify/highlight differences (gaps) [ ] Confirm gap solutions (3 possible actions) [ ] 1. Sender will update configurations/message mappings to align with receiver profile [ ] 2. Receiver will update message processing to align with sender profile [ ] 3. Intermediary will transform/translate messages mid-stream to alleviate burden on sender & receiver [ ] Update message profiles (if sender or receiver are making changes to accommodate) Definition of Done [ ] Documentation tasks completed [ ] Documentation and diagrams created or updated [ ] Implementation guide (/ig folder) [ ] ADRs (/adr folder) [ ] Main README.md [ ] Other READMEs in the repo [ ] If applicable, update the ReportStream Setup section in README.md [ ] Threat model updated [ ] API documentation updated [ ] Code quality tasks completed [ ] Code refactored for clarity and no design/technical debt [ ] Adhere to separation of concerns; code is not tightly coupled, especially to 3rd party dependencies [ ] Code is reviewed or developed by pair; 1 approval is needed but consider requiring an outside-the-pair reviewer [ ] Code quality checks passed [ ] Security & Privacy tasks completed [ ] Security & privacy gates passed [ ] Testing tasks completed [ ] Load tests passed [ ] Unit test coverage of our code >= 90% [ ] Build & Deploy tasks completed [ ] Build process updated [ ] API(s) are versioned [ ] Feature toggles created and/or deleted. Document the feature toggle [ ] Source code is merged to the main branch Research Questions Optional: Any initial questions for research Decisions Optional: Any decisions we've made while working on this story Notes Profile stories: #992 #618 Also pending verification of order profile by Cerner
2025-04-01T06:36:47.758917
2021-07-06T11:49:28
937813656
{ "authors": [ "j1y3p4rk" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:491", "repo": "CESNET/netopeer2", "url": "https://github.com/CESNET/netopeer2/issues/947" }
gharchive/issue
Invalid Base64 charater 0xa I am using netopeer2-cli and netopeer2-server. I am trying to merge tls_keystore.xml to ietf-keystore as written in README, to try to enable TLS. When I type the command as below, I see the sysrepocfg error about data parsing failed.. Could you help me to solve this error? ubuntu@:~/netopeer2/build$ sudo sysrepocfg -f xml -d startup --edit=../example_configuration/tls_keystore.xml -m ietf-keystore -v3 [INF]: Scheduled changes not applied because of other existing connections. [INF]: Connection 19 created. [INF]: Session 117 (user "root", CID 19) created. sysrepocfg error: libyang: Invalid Base64 character 0xa. sysrepocfg error: Data parsing failed [INF]: No datastore changes to apply. removed unnecessary empty spaces.
2025-04-01T06:36:47.763324
2020-02-16T18:05:00
565944205
{ "authors": [ "pajavyskocil" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:492", "repo": "CESNET/proxyidp-nagios-scripts", "url": "https://github.com/CESNET/proxyidp-nagios-scripts/pull/8" }
gharchive/pull-request
Rewrited auth scripts All script now timeouts if the script runs too long Login functions is now moved to separate helper scripts Added missing newlines Merging after discussion with Dominik Bučík
2025-04-01T06:36:47.967380
2018-02-06T14:41:15
294783323
{ "authors": [ "kevinrue", "sebastian-luna-valero" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:493", "repo": "CGATOxford/CGATPipelines", "url": "https://github.com/CGATOxford/CGATPipelines/pull/396" }
gharchive/pull-request
delay csvdb load tasks by a random offset to avoid database lock issues This resolved my issues when processing a lot of FASTQ files in parallel. (There may be better ways to address the issue, though) Thanks for the input, Kevin. It looks like a good workaround. Please bear in mind that SQLite is not a proper Relational Database Management System, and actually its performance is dependent on the underlying filesystem. I tried to fix SQLite operational errors with the following changes but more thought might be required: Pipeline changes: https://github.com/CGATOxford/CGATPipelines/pull/380 scripts changes: https://github.com/CGATOxford/cgat/pull/377 Those changes were added to the master in the v0.3.2 release, could you please let me know what version are you running with the following commands: pip list | grep -i cgat Best regards, Sebastian Hi Sebastian, I think I should be up to date. I pulled from the master branch this morning. $ pip list | grep -i cgat DEPRECATION: The default format will switch to columns in the future. You can use --format=(legacy|columns) (or define a format=(legacy|columns) in your pip.conf under the [list] section) to disable this warning. CGAT (0.3.2, /gfs/devel/kralbrecht/cgat) CGATPipelines (0.3.2, /gfs/devel/kralbrecht/CGATPipelines) CGATReport (<IP_ADDRESS>) Let me know if there's any command I should run to update the CGAT installation, or something. For the record, I was systematically getting the 'database locked' error message for pipeline_readqc.py with 384 FASTQ files. I don't know exactly where this falls in the range of input files that you usually process in a single run. As I said, the random delay of 1-30 seconds worked in my case, but might need to be increased or even parameterised if the error pops out again for larger amounts of input files. Best Kevin Thanks, Kevin. We try to solve this by asking ruffus to limit the number of concurrent jobs to 1`: @jobs_limit(PARAMS.get("jobs_limit_db", 1), "db") However, it does not seem to work properly. We are about to perform a major refactoring on the code, and we'll try to solve this problem then. The idea you propose in this PR is a potential solution. I will leave this open for our reference and I will decide what to do during the code refactoring. Many thanks! Sebastian Hi @sebastian-luna-valero Yes I saw the jobs_limit instruction in the code, and it did puzzle me as to why it does not seem to work. However, I was short on time and instead of investigating the issue, implemented the quick fix offered in this PR. Obviously, the ideal fix here would be to somehow have this jobs_limit instruction work as expected. I 100% agree that this PR is not the ideal implementation, and I would not take it personally if it is closed without being merged. Happy to see it used as a reference. As a user, I'll say that I'm happy enough to have this random delay even increased to a minute or two if needed, which would still represent a fraction of the total pipeline run, in compensation for reducing the chance of database block virtually to nil. Thanks for your work on the refactoring, I look forward to the result!
2025-04-01T06:36:47.972461
2021-06-07T05:28:54
913109516
{ "authors": [ "sgarg15" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:494", "repo": "CHB61/multi_select_flutter", "url": "https://github.com/CHB61/multi_select_flutter/issues/45" }
gharchive/issue
Colors for Cancel and Ok Button and Select Chip when selected Hey, I was trying to implement this package into my app and I wanted to change the colors for the buttons when you open the Bottom Sheet and when you select the chips inside the Sheet. I tried looking online and within the actual code but couldn't find anything that worked. If theres a solution for this please let me know. Thank you. My code for the MultiSelectBottomShieldField looks like the following: Container( width: 335, decoration: BoxDecoration( color: Color(0xFFF0F0F0), border: Border.all( color: Color(0xFFE8E8E8), width: 2, ), borderRadius: BorderRadius.circular(8), ), child: Column( children: <Widget>[ MultiSelectBottomSheetField<Software>( initialChildSize: 0.4, decoration: BoxDecoration(), listType: MultiSelectListType.CHIP, initialValue: _selectedSoftware, searchable: true, items: _items, buttonText: Text("Select Pharmacy Software...", style: GoogleFonts.inter( color: Color(0xFFBDBDBD), fontSize: 16)), onConfirm: (values) { _selectedSoftware = values; }, chipDisplay: MultiSelectChipDisplay( items: _selectedSoftware .map((e) => MultiSelectItem(e, e.toString())) .toList(), chipColor: Color(0xFF5DB075), onTap: (value) { _selectedSoftware.remove(value); return _selectedSoftware; }, textStyle: TextStyle(color: Colors.white), ), ), ], ), ), Never mind! I missed the selectedColor and selectedItemsTextStyle fields.
2025-04-01T06:36:47.973389
2023-03-20T10:34:49
1631825225
{ "authors": [ "Sarbojit2019" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:495", "repo": "CHIP-SPV/H4I-MKLShim", "url": "https://github.com/CHIP-SPV/H4I-MKLShim/pull/4" }
gharchive/pull-request
Added openCL support HIP supports openCL as backend hence enabled for hipblas as well. Closing it because new PR has included the change.
2025-04-01T06:36:47.990817
2024-09-16T15:01:15
2528727292
{ "authors": [ "CIOCOLATA47", "NthnH" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:496", "repo": "CIOCOLATA47/AntiAfk", "url": "https://github.com/CIOCOLATA47/AntiAfk/issues/11" }
gharchive/issue
improve mousemovement The mousemovement could use a multiplier parameter (more extreme movement) and also a vertical movement The server I play on requires both horizontal and vertical mouse movement (and it could be that a few pixels is not enough), which I already achieved with a macro but I dont want to keep my MC window active all the time Aight i added it use /antiafk mousemovement and /antiafk mousemovement to disable/enable it
2025-04-01T06:36:48.011695
2017-04-07T12:09:28
220195706
{ "authors": [ "RinkeHoekstra" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:497", "repo": "CLARIAH/wp4-converters", "url": "https://github.com/CLARIAH/wp4-converters/issues/20" }
gharchive/issue
Add an option to generate graph URIs without hashes Both in the nano publication code, but also from the command line. Fixed in 6eae1be
2025-04-01T06:36:48.024812
2024-11-29T12:55:21
2705040787
{ "authors": [ "ollylucl" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:498", "repo": "CLOSER-Cohorts/dashboard", "url": "https://github.com/CLOSER-Cohorts/dashboard/pull/166" }
gharchive/pull-request
issue-147-add-dashboard-for-showing-question-variable-topic-mismatches Add dashboard for displaying question/variable pair where their assigned topics do not match. Addresses issue https://github.com/CLOSER-Cohorts/dashboard/issues/161 @spuddybike just a reminder about this open PR, from before the EDDI conference.
2025-04-01T06:36:48.107792
2022-09-02T20:22:35
1360582346
{ "authors": [ "daisukes", "tatsuya-ishihara" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:500", "repo": "CMU-cabot/cabot", "url": "https://github.com/CMU-cabot/cabot/pull/63" }
gharchive/pull-request
use ROS timer instead of std::thread and use multi thread spinner @tatsuya-ishihara Could you review the change? I didn't know the multi-thread spinning APIs and tried to use my own thread to keep the FPS up. However, it looks like a cause of a bug in a certain situation. So, now I reuse ROS timer and use multi-thread spinning both in node and nodelet implementation. I could not reproduce the bug situation, so I just tested if it was working in a test condition (a few people) without navigation. Thank you. It looks good to me. @tatsuya-ishihara, could you "review" as github function to change the status?
2025-04-01T06:36:48.122073
2023-04-06T09:47:48
1657061895
{ "authors": [ "HarshaMM", "audoan99", "efriedrich", "khangtb1" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:501", "repo": "COVESA/wayland-ivi-extension", "url": "https://github.com/COVESA/wayland-ivi-extension/pull/155" }
gharchive/pull-request
ivi-controller: update data type of member in struct ivishell From weston version 8.0.0, data type of param in weston_config_section_get_bool() changed from int to bool. So, we need to update in wayland-ivi-extension source. (< version 8.0.0 ) (>= version 8.0.0) Tested-by: Au Doan Ngoc<EMAIL_ADDRESS> reviewed-by: Tran Ba Khang<EMAIL_ADDRESS> Reviewed-by: Harsha M M<EMAIL_ADDRESS> thanks for the hint we would consider this, in case you have already prepared something just push it ;-)
2025-04-01T06:36:48.146305
2017-08-11T22:57:27
249766487
{ "authors": [ "amc-nu", "gbiggs" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:502", "repo": "CPFL/Autoware", "url": "https://github.com/CPFL/Autoware/issues/776" }
gharchive/issue
Feature ParticleFilter based tracker New Feature Vscan based Rao-Blackwellized Particle Filter tracker adapted to work on Autoware Expected Result https://drive.google.com/open?id=0BzYuVrO9pnh6VUhQd0N2TjBsM3c Documentation Based on the work described in: http://ieeexplore.ieee.org/document/7759043/ Constraints Requires hdl-64 to work reliably Uses 2d scan to match between frames Can only track up to 5 objects in realtime, using a GPU Needs the correct pose, for that reason in some cases, the tracking will fail @amc-nu What's the status on this PR?
2025-04-01T06:36:48.190814
2023-11-08T14:28:08
1983739383
{ "authors": [ "jlahovnik" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:503", "repo": "CS-SI/eodag", "url": "https://github.com/CS-SI/eodag/pull/916" }
gharchive/pull-request
850 avoid search by id before every server mode download download info is stored in instance of EODataAccessGateway with key composed of product type, provider and id if download is attempted for product where id is not in download info and search by id is available for provider and product type, it will be executed; if search by id is not possible an error will be raised to do a search updated based on input from @alambare-csgroup closed because change requests were implemented on new branch -> new PR: https://github.com/CS-SI/eodag/pull/1012
2025-04-01T06:36:48.198281
2018-09-19T16:36:41
361832003
{ "authors": [ "arsalanc-v2", "jjlee050" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:504", "repo": "CS2103-AY1819S1-W14-1/main", "url": "https://github.com/CS2103-AY1819S1-W14-1/main/issues/41" }
gharchive/issue
As a medical clinic receptionist, I want to request for a specific doctor for consultation So that I can satisfy patient need. what's the difference between this and #36 ?
2025-04-01T06:36:48.206890
2016-10-06T05:40:51
181328993
{ "authors": [ "sunset1215" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:505", "repo": "CS2103AUG2016-T13-C1/main", "url": "https://github.com/CS2103AUG2016-T13-C1/main/issues/2" }
gharchive/issue
As a user, I can tag my tasks based on priority. so that I can prioritize my goals this feature is not implemented
2025-04-01T06:36:48.257339
2016-03-02T01:52:08
137747628
{ "authors": [ "SteveBarnett", "gregbenner" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:506", "repo": "CTFEDs/ctfeds.org", "url": "https://github.com/CTFEDs/ctfeds.org/issues/17" }
gharchive/issue
"Attendees must read and follow our Code of Conduct" Broken link on meetup and ctfeds.org ... Fixed! Thank you for spotting it and letting us know. :)
2025-04-01T06:36:48.278241
2021-10-29T01:23:25
1039078417
{ "authors": [ "Feissit" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:507", "repo": "CUCentralAdvancement/essential-cu", "url": "https://github.com/CUCentralAdvancement/essential-cu/issues/459" }
gharchive/issue
Full Financials PDF Get financials PDF from Patty Leslie. Latest update should be ready by tomorrow (Fri Oct 29) This looks good to me Get Outlook for Androidhttps://aka.ms/AAb9ysg From: Alex Finnarn @.> Sent: Friday, October 29, 2021 1:33:45 PM To: CUCentralAdvancement/essential-cu @.> Cc: Yifei Wu @.>; Assign @.> Subject: Re: [CUCentralAdvancement/essential-cu] Full Financials PDF (Issue #459) PDF link added to https://essential-staging-cu.herokuapp.com/impact-reports/joy/financials — You are receiving this because you were assigned. Reply to this email directly, view it on GitHubhttps://github.com/CUCentralAdvancement/essential-cu/issues/459#issuecomment-955000045, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AVYBARVQMIPCI2WYNGOCDEDUJLZJTANCNFSM5G6F2UQA. Triage notifications on the go with GitHub Mobile for iOShttps://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675 or Androidhttps://play.google.com/store/apps/details?id=com.github.android&referrer=utm_campaign%3Dnotification-email%26utm_medium%3Demail%26utm_source%3Dgithub.
2025-04-01T06:36:48.286727
2022-12-09T17:37:13
1487145926
{ "authors": [ "kylebgorman" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:508", "repo": "CUNY-CL/yoyodyne", "url": "https://github.com/CUNY-CL/yoyodyne/issues/5" }
gharchive/issue
Testing [copied from CUNY-CL/abstractness/issues/87] We should add integration tests (I hesitate to call these unit tests), simply limiting ourselves to the model sizes and data quantities we can run on CircleCI's free tier. We get 6,000 compute-minutes per month...all of this is pretty generous except that I am unclear whether we can use their GPU images or are stuck on CPU (ideally we'd parameterize tests on both). I think it ought to be possible to do actual training of the major models using, say, 1,000 examples. Unit tests could include g2p (for feature-less) and inflection (for feature-full) from SIGMORPHON. The current training and prediction functions are structured to read and write directly to the file system. They should be modularized to take ordinary arguments and return the results: for training, a function could simply return the best model (or its path) with metadata (wall clock time, training accuracy, development accuracy), and then the command-line enabled version of that loop could invoke this for prediction, a function could simply return the accuracy. These functions can then be called by the existing (null return type) training and prediction functions, the ones parameterized with click flags. This will also support two other projects (issues coming soon): benchmarking W&B-enabled hyperparameter sweeping This is a blocker for a post-beta release candidate. Yoyodyne test strategy.pdf The above describes my current thinking about the test strategy.
2025-04-01T06:36:48.316368
2024-10-15T12:40:47
2588663599
{ "authors": [ "SolidProgramming", "tpurschke" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:510", "repo": "CactuseSecurity/firewall-orchestrator", "url": "https://github.com/CactuseSecurity/firewall-orchestrator/pull/2593" }
gharchive/pull-request
Add missing FK connection Closes #2591 did you close this on purpose? the missing FW connection changes are not merged yet. yes. i renamed the branch and it auto closed this PR. i remade antother
2025-04-01T06:36:48.383170
2024-06-30T17:46:57
2382357650
{ "authors": [ "Cale-Torino", "jusromaine" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:511", "repo": "Cale-Torino/FreeTAKServer_Manager", "url": "https://github.com/Cale-Torino/FreeTAKServer_Manager/issues/9" }
gharchive/issue
Issue with using the FreeTAKServer Manager. Followed the youtube video and got it too install no problem but when I start the server I get two cmd screens like this, I have added screen shots too this. Thanks also tried this of no and still same issues Hi there, Looks like some dependencies are conflicting. You can check them here. for verification. The manager was made for the older version 1 of FTS and there is a Winforms and WPF version both should work the same but sometimes the different API calls can cause a strange error/bug however in this case it's more likely a Python thing since all the software is using python for the FTS stuff the C# part is just managing the interface an minor things like logging etc. Are you using the version in the releases tab? and did you install via the MSI installer? Also I tested the setup using Python 3.11.3 and Python 3.8.10 only, so others may have made changes to modules. Let me know. If you're still having issues you can try setup a VM in windows to install the latest version 2 stuff via Linux there's been many improvements to the Linux install. I'm not sure when I'll take a look at updating this software again been very busy at work recently
2025-04-01T06:36:48.391650
2021-10-15T11:44:47
1027361265
{ "authors": [ "gibrankasif" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:512", "repo": "Call-for-Code-for-Racial-Justice/Incident-Accuracy-Reporting-System", "url": "https://github.com/Call-for-Code-for-Racial-Justice/Incident-Accuracy-Reporting-System/pull/71" }
gharchive/pull-request
removed base package.json files #67 Discarded package.json and package.lock.json files Hi, @upkarlidder I've added the PR.
2025-04-01T06:36:48.436418
2024-04-30T11:28:40
2271205119
{ "authors": [ "ArtIntAI", "Chuancysun", "duzekunKTH" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:513", "repo": "Cambricon/mlu-ops", "url": "https://github.com/Cambricon/mlu-ops/pull/1019" }
gharchive/pull-request
[WIP] add new operator LU factorization Thanks for your contribution and we appreciate it a lot. :rocket::rocket: 1. Motivation add floating point operator LU factorization 2. Modification add implementation of floating point LU factorization 3. Test Report 3.1 Modification Details 3.1.1 Accuracy Acceptance Standard For static threshold standard details, see: MLU-OPS™ Accuracy Acceptance Standard. static threshold diff1 [ ] float32 mlu diff1 <= 1e-5 [*] float32 mlu diff1 <= 3e-3 [ ] float16 mlu diff1 <= 3e-3 diff2 [ ] float32 mlu diff2 <= 1e-5 [* ] float32 mlu diff2 <= 3e-3 [ ] float16 mlu diff2 <= 3e-3 diff3 [ ] mlu diff3 == 0 [ ] mlu diff3_1 == 0 [ ] mlu diff3_2 == 0 dynamic threshold [ ] diff1: mlu diff1 <= max(baseline diff1 * 10, static threshold) [ ] diff2: mlu diff2 <= max(baseline diff2 * 10, static threshold) [ ] diff3: mlu diff3 <= max(baseline diff3 * 10, static threshold) float32, threshold = 1e-5 float16, threshold = 1e-3 3.1.2 Operator Scheme checklist Supported hardware [* ] MLU370 [ ] MLU590 Job types [ ] BLOCK [ ] UNION1 [ ] UNION2 [ ] UNION4 [* ] The operator will dynamically select the most suitable task type, for example, UNION8 3.2 Accuracy Test 3.2.1 Accuracy Test If you have checked the following items, please tick the relevant box. [ ] Data type test (e.g. float32/int8) [ ] Multi-dimensional tensor test [ ] Layout test [ ] Different size/integer remainder end segment/alignment misalignment test [ ] Zero dimensional tensor test/zero element test [ ] stability test [ ] Multiple platform test [ ] Gen_case module test, see: Gencase-User-Guide-zh [ ] Nan/INF tests [ ] Bug fix tests [ ] For memory leak check details, see: GTest-User-Guide-zh [ ] For code coverage check details, see: GTest-User-Guide-zh [ ] For I/O calculation efficiency check details, see: MLU-OPS™-Performance-Acceptance-Standard 3.3 Performance Test Platform:MLU370 ----------- case0 ----------- case0 [Op name ]: sgetrf [Shape ]: input.shape=[256,256], output.shape=[256,256] [Data type] ]: float32 [MLU Hardware Time ]: 6460 (us) [MLU Interface Time ]: 15336.7 (us) [MLU IO Efficiency ]: 0.00026419 [MLU Compute Efficiency ]: 9.90712e-06 [MLU Workspace Size ]: -1 (Bytes) [MLU Kernel Name(s) ]: {} [MLU TheoryOps ]: 65536 (Ops) [MLU TheoryIOs ]: 524288 (Bytes) [MLU ComputeForce ]: 1.024e+12 (op/s) [MLU IoBandWidth ]: 307.2 (GB/s) [GPU Hardware Time ]: -1 (us) [GPU IO Efficiency ]: -1 [GPU Compute Efficiency ]: -1 [GPU Workspace Size ]: -1 (Bytes) [Diffs]: [output] DIFF1: 1.798500e-04 DIFF2: 7.016698e-04 [^ OK ] ../../test/mlu_op_gtest/pb_gtest/src/zoo/sgetrf/test_case/case0.prototxt [ OK ] sgetrf/TestSuite.mluOp/0 (36 ms) [----------] 1 test from sgetrf/TestSuite (36 ms total) [----------] Global test environment tear-down [ SUMMARY ] Total 1 cases of 1 op(s). ALL PASSED. [==========] 1 test case from 1 test suite ran. (3727 ms total) [ PASSED ] 1 test case. 3.4 Summary Analysis Please give a brief overview here, if you want to note and summarize the content. 当前文件有冲突,请rebase master 的同时,解除文件冲突。 Conflicting files docs/user_guide/9_operators/index.rst mlu_op.h 当前文件有冲突,请rebase master 的同时,解除文件冲突。 Conflicting files docs/user_guide/9_operators/index.rst mlu_op.h Feature: Refactor roi align rotated forward 参考这个格式修改下commit信息
2025-04-01T06:36:48.440864
2015-01-28T04:51:09
55714280
{ "authors": [ "GeorgeDewar", "Senjai" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:514", "repo": "CanCanCommunity/cancancan", "url": "https://github.com/CanCanCommunity/cancancan/issues/174" }
gharchive/issue
Cannot use Hash as a resource I realise this is an unusual way to use CanCan, but I have noticed that the following fails unpredictably: can :some_action, Hash do |params| # actual logic omitted puts params true end If I try can? :some_action, {a: 1}, it will return false and not run the block above at all. If I try can? :some_action, {a: {a: 1}}, it will behave correctly, printing {a: {a: 1}}. This appears to be due to rule.rb's relevant method, which contains the following code: 24: def relevant?(action, subject) 25: subject = subject.values.first if subject.class == Hash 26: @match_all || (matches_action?(action) && matches_subject?(subject)) 27: end I'm curious what that is meant to do? Any thoughts? Is the use case so mad that it just should never be done? The use case is basically where the subject is just one of several arbitrary parameters. For example, rather than "view a client", "view a client's financial projections" or "view a client's contact details". Any tips on how to achieve logic that goes beyond action and subject would be helpful also. I see that this behaviour doesn't occur if the subject is a subclass of Hash, such as ActionController::Parameters. @GeorgeDewar I'll look into this when I can, but I agree this case is not conventional. Ideally, you should never be passing in a Hash here. An object can have a hash for its internal data structure, but I would recommend actually modeling your application with classes representative of the data they contain and operate on. I haven't dug into that yet, but I'm betting strongly that this won't be changed. Parameters doesn't get caught by this because it checks class, not superclass, nor does it use is_a?. When you pass a hash, I believe it is expecting a list of conditions.
2025-04-01T06:36:48.452729
2021-03-10T12:56:44
827696440
{ "authors": [ "AbdulmalekAlshugaa", "Canato", "RanjitPati", "TomasMVazquez" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:515", "repo": "CanHub/Android-Image-Cropper", "url": "https://github.com/CanHub/Android-Image-Cropper/issues/84" }
gharchive/issue
[BUG] -Getting File Not Found Exception when creating new File with cropped uri.getPath() Lib Version 2.2.1 Describe the bug I am able to crop the image and put in ImageView successfully. But when I am using the same URI path to create a file it's throwing FileNotFoundException. The URI path is starting like content://..something/package name/ myfiles/... something like this. When I debugged I am able to see some data like NO CACHE inside the URI value. I am pasting my code below: Calling Method to crop: public void onSelectImageClick() { CropImage .activity(null) .setOutputCompressFormat(Bitmap.CompressFormat.JPEG) .setGuidelines(CropImageView.Guidelines.ON) .setFixAspectRatio(true) .start(this); } Getting cropped result in onActivityResuly() @Override protected void onActivityResult(int requestCode, int resultCode, @Nullable Intent data) { super.onActivityResult(requestCode, resultCode, data); if (requestCode == CropImage.CROP_IMAGE_ACTIVITY_REQUEST_CODE) { CropImage.ActivityResult result = CropImage.getActivityResult(data); if (resultCode == RESULT_OK) { Glide.with(this) .load(result.getUri()) .into(actorPic); isImageViewAdded = true; //handleCropResult(CropImage.getActivityResult(data)) imagePath = result.getUri(); //imagePath = CropImage.getPickImageResultUri(this, data); //insertSingleItem(result.getUri()); } else if (resultCode == CropImage.CROP_IMAGE_ACTIVITY_RESULT_ERROR_CODE) { Toast.makeText(this, "Cropping failed: " + result.getError(), Toast.LENGTH_LONG).show(); } } if(requestCode == CropImage.PICK_IMAGE_CHOOSER_REQUEST_CODE){ Uri imageUri = CropImage.getPickImageResultUri(AddActorsActivity.this, data); imagePath = imageUri; } } But the imageview is able to find the path and successfully update the cropped image inside it through Glide. Expected behavior I think the cropped Image is not caching and so it's happening like this. Hi, In my case I solved the problem following this: https://developer.android.com/training/data-storage/shared/documents-files?hl=es-419 Hope it helps you too! Regards, I am creating the file like new File(uri.getPath()) . The same code was working before. @TomasMVazquez Can you please share the piece of code which you have used to create the file from the uri. I am creating the file like new File(uri.getPath()) . The same code was working before. @TomasMVazquez Can you please share the piece of code which you have used to create the file from the uri. @RanjitPati maybe this is happening because of Android OS permission changes. Now using scope storage we don't get a file path anymore. Maybe you can change the URI string "content" to "file" But is not the real fix, the library will return the URI for the image using the scope storage like google force us now. If you plan to create a file you need to get write storage permission and get the path where you put the image. Make sense? com.theartofdev.edmodo:android-image-cropper:2.8.0 It works here pretty fine I am creating the file like new File(uri.getPath()) . The same code was working before. @TomasMVazquez Can you please share the piece of code which you have used to create the file from the uri. What I needed was to get Base64 from Uri/Path, originaly I was getting the path but with the change I'm using directly the Uri: Original Code: fun imageFileToBase64(imageFile: File): String { return FileInputStream(imageFile).use { inputStream -> ByteArrayOutputStream().use { outputStream -> Base64OutputStream(outputStream, Base64.DEFAULT).use { base64FilterStream -> inputStream.copyTo(base64FilterStream) base64FilterStream.flush() outputStream.toString() } } } } New Code: fun imageUriToBase64(context: Context, uri: Uri): String? { val contentResolver = context.contentResolver return contentResolver.openInputStream(uri)?.use { inputStream -> ByteArrayOutputStream().use { outputStream -> Base64OutputStream(outputStream, Base64.DEFAULT).use { base64FilterStream -> inputStream.copyTo(base64FilterStream) base64FilterStream.flush() outputStream.toString() } } } } com.theartofdev.edmodo:android-image-cropper:2.8.0 It works here pretty fine, you can check this library for you reference. This was the main reason I fork from the old library @AbdulmalekAlshugaa , this was the way it work before OS 11. Now with OS 11, this behaviour changed so the library was updated, not we need to update the usage of it too. Please let's put all discussion about this on the same place: https://github.com/CanHub/Android-Image-Cropper/discussions/87 Sadly this is an Android OS change, but if anyone know a better fix for the latest OS, using scope storage and keeping it file drop a PR cause will help everyone =) I will close so we focus the discussion on the same place https://github.com/CanHub/Android-Image-Cropper/discussions/87 Please try the latest 3.0.0 release and let me know
2025-04-01T06:36:48.454225
2024-05-20T21:28:17
2306826181
{ "authors": [ "CanadaHonk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:516", "repo": "CanadaHonk/porffor", "url": "https://github.com/CanadaHonk/porffor/issues/30" }
gharchive/issue
Profile-guided optimization Add a pgo mode to the compiler, which compiles with logging for runtime types (or more in the future) which is dumped to file. Then, add an argument to regular compile which takes in such file as input and uses it for optimization. Done in 97bb4f33be95a23eaebd017cea80a90a1d49abef!
2025-04-01T06:36:48.484802
2023-07-07T10:49:18
1793305290
{ "authors": [ "codecov-commenter", "onlyjackfrost" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:517", "repo": "Canner/vulcan-sql", "url": "https://github.com/Canner/vulcan-sql/pull/225" }
gharchive/pull-request
Chore: bump to 0.6.0 Description Bump to 0.6.0 Codecov Report Patch coverage: 88.57% and project coverage change: -0.05 :warning: Comparison is base (d121a7a) 90.52% compared to head (4e99b6a) 90.47%. :exclamation: Your organization is not using the GitHub App Integration. As a result you may experience degraded service beginning May 15th. Please install the Github App Integration for your organization. Read more. Additional details and impacted files @@ Coverage Diff @@ ## develop #225 +/- ## =========================================== - Coverage 90.52% 90.47% -0.05% =========================================== Files 331 335 +4 Lines 5477 5596 +119 Branches 732 742 +10 =========================================== + Hits 4958 5063 +105 - Misses 374 384 +10 - Partials 145 149 +4 Flag Coverage Δ build 90.55% <ø> (ø) catalog-server 100.00% <ø> (ø) cli 75.85% <ø> (ø) extension-authenticator-canner 78.37% <100.00%> (-2.11%) :arrow_down: extension-dbt 97.43% <ø> (ø) extension-driver-canner 84.65% <ø> (ø) extension-driver-clickhouse 88.09% <88.09%> (?) extension-driver-pg 96.11% <ø> (ø) extension-driver-snowflake 96.26% <ø> (ø) extension-store-canner 98.30% <ø> (ø) integration-testing 90.27% <ø> (ø) serve 87.17% <92.30%> (+0.05%) :arrow_up: Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage Δ ...-driver-clickhouse/src/lib/clickHouseDataSource.ts 81.66% <81.66%> (ø) ...c/lib/middleware/auth/authCredentialsMiddleware.ts 90.00% <85.71%> (-1.67%) :arrow_down: .../extension-driver-clickhouse/src/lib/typeMapper.ts 90.24% <90.24%> (ø) ...-authenticator-canner/src/lib/authenticator/pat.ts 75.75% <100.00%> (-2.63%) :arrow_down: packages/extension-driver-clickhouse/src/index.ts 100.00% <100.00%> (ø) .../extension-driver-clickhouse/src/lib/sqlBuilder.ts 100.00% <100.00%> (ø) ...kages/serve/src/lib/auth/httpBasicAuthenticator.ts 90.90% <100.00%> (+3.67%) :arrow_up: ...es/serve/src/lib/auth/passwordFileAuthenticator.ts 80.48% <100.00%> (-1.34%) :arrow_down: ...ges/serve/src/lib/auth/simpleTokenAuthenticator.ts 100.00% <100.00%> (ø) :umbrella: View full report in Codecov by Sentry. :loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
2025-04-01T06:36:48.500954
2022-09-17T23:52:20
1376898585
{ "authors": [ "fmaclen", "gokullan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:518", "repo": "Canutin/svelte-currency-input", "url": "https://github.com/Canutin/svelte-currency-input/issues/30" }
gharchive/issue
Adding 0's after the "fractionDigits" limit doesn't trigger formatting $1.2211111111 > $1.22 ❤️ $1.2200000000 > $1.2200000000 💔 Hi, there! I would like to take up this issue. Can it be assigned to me? @gokullan sounds good, submit a PR when you are ready. Thanks!
2025-04-01T06:36:48.502913
2021-10-02T14:18:12
1014063974
{ "authors": [ "goliakshay357", "infinityover" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:519", "repo": "Canvasbird/canvasboard", "url": "https://github.com/Canvasbird/canvasboard/issues/420" }
gharchive/issue
Lazy Loading Is your feature request related to a problem? Please describe. To load the welcome page faster Describe the solution you'd like 💡 Use lazy loading module to detach the welcome part so the initial page can load faster Can i get this task? Hello @infinityover, Sure!
2025-04-01T06:36:48.509966
2016-11-24T08:06:18
191455591
{ "authors": [ "jmlambert78", "meken", "tayzlor" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:520", "repo": "Capgemini-AIE/ethereum-docker", "url": "https://github.com/Capgemini-AIE/ethereum-docker/issues/17" }
gharchive/issue
While deploying the docker-compose there are errors on the primus module as strict & const seem not compatible... I just launched the docker compose to run a cluster, and launching the eth-netstats is broken. JMarc /eth-netstats/node_modules/primus/index.js:177 const sandbox = Object.keys(global).reduce((acc, key) => { ^^^^^ SyntaxError: Use of const in strict mode. at exports.runInThisContext (vm.js:73:16) at Module._compile (module.js:443:25) at Object.Module._extensions..js (module.js:478:10) at Module.load (module.js:355:32) at Function.Module._load (module.js:310:12) at Module.require (module.js:365:17) at require (module.js:384:17) at Object. (/eth-netstats/app.js:44:14) at Module._compile (module.js:460:26) at Object.Module._extensions..js (module.js:478:10) npm ERR! Linux 3.13.0-52-generic npm ERR! argv "/usr/local/bin/node" "/usr/local/bin/npm" "start" npm ERR! node v0.12.17 npm ERR! npm v2.15.1 npm ERR! code ELIFECYCLE npm ERR<EMAIL_ADDRESS>start: node ./bin/www npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the<EMAIL_ADDRESS>start script 'node ./bin/www'. npm ERR! This is most likely a problem with the eth-netstats package, npm ERR! not with npm itself. npm ERR! Tell the author that this fails on your system: npm ERR! node ./bin/www npm ERR! You can g I had a similar issue, and fixed that by upgrading the node version for the eth-netstats container; updating the first line of the eth-netstats/Dockerfile from FROM node:0.12 to FROM node:7 did it for me. yes thx its fixed in my case as well, we should update the file and release... Closing via https://github.com/Capgemini-AIE/ethereum-docker/pull/18
2025-04-01T06:36:48.540746
2018-01-28T06:46:41
292183097
{ "authors": [ "ftrotter" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:522", "repo": "CareSet/DURC", "url": "https://github.com/CareSet/DURC/issues/2" }
gharchive/issue
URL calculations when laravel is run under a sub-directory are broken Assuming you are running laravel under the /project/ subdirectory for a domain. you usually use a custom url-rewrite to ensure that http://example.com/project/DURC/author/ is rewritten to http://example.com/project/index.php/DURC/Author/ But in that case the routes need to be build in terms of /DURC/author but the links in all of the templates need to be for /project/DURC/author/ perhaps this is an indication that the whole url calculation method is simplistic and needs to be improved. No one hosts Laravel this way. It is a bother to code this and it is not worth it.
2025-04-01T06:36:48.544224
2016-01-27T01:52:25
129001652
{ "authors": [ "benwilson512", "icn" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:524", "repo": "CargoSense/ex_aws", "url": "https://github.com/CargoSense/ex_aws/issues/106" }
gharchive/issue
The issue has been resolved Feel free to just close the issue next time instead of renaming the title :)
2025-04-01T06:36:48.554984
2023-01-01T23:37:22
1515898638
{ "authors": [ "CarlosEsco", "yapudjus" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:525", "repo": "CarlosEsco/Neko", "url": "https://github.com/CarlosEsco/Neko/issues/1288" }
gharchive/issue
consistency of scanlation team when reading Describe your suggested feature when reading, if you start with, say asura scan but there's also entries for the same chapter from, say, flame scan, if i start reading from asura scan, it sometimes switch to flamescan in the next chapter, even if asura scan is available, it's not a problem if it's on the same language, but it's infuriating when it switch from English to French when going to the next chapter it should also be noted that when a chapter gets marked as read when i finish it, it doesn't mark the same chapter from another team as read too Other details No response Acknowledgements [X] I have searched the existing issues and this is a new ticket, NOT a duplicate or related to another open issue. [X] I have written a short but informative title. [X] I have updated the app to the newest version Latest. [X] I have checked through the app settings for my feature. [X] I will fill out all of the requested information in this form. Are you talking about when skip duplicates is enabled? i do have skip duplicate enabled the thing I'm talking about is consistency across chapters when skipping duplicates, and the fact that duplicate aren't marked as read either Well duplicates marked as read is not what the option provides. As for skip duplicates it doesn't take into account the scanlator cause it's a pain to try to get something like that working. Just filter out the scanlator you want then make the others as read if it's that big of issue. Then unfilter the scanlator This should be slightly more consistent now
2025-04-01T06:36:48.624395
2017-05-11T20:39:34
228114265
{ "authors": [ "aaroncrespo", "mdiep" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:526", "repo": "Carthage/Commandant", "url": "https://github.com/Carthage/Commandant/issues/104" }
gharchive/issue
Dont mangle key typos. https://github.com/Carthage/Carthage/issues/1926 I think for the sake of better error messages Commandant should not try and convert flags to a set https://github.com/Carthage/Commandant/blob/903eaec12f3a68782496065dcc68393b8ab1b4bd/Sources/Commandant/ArgumentParser.swift#L75 It will mangle a --key typo and produce confusing errors. If flags are passed as used ex: $ carthage update -platform Unrecognized arguments: -armpfolt Should become $ carthage update -platform Unrecognized arguments: -platform That's only done if there's a single -, which denotes a single-letter flag/argument. That's how basically all *nix tools work. do all nix tools randomly the input in the error message? This issue isn't about the fact that there is an error message. this issue is about the formatting and content of the error message. It looks like most tools bail on the first unknown option: $ python -asenxtaboe Unknown option: -a I'm not sure if that's better as it gives you less information.
2025-04-01T06:36:48.625543
2016-04-01T08:38:08
145121399
{ "authors": [ "dgaubert", "jgoizueta" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:527", "repo": "CartoDB/Windshaft-cartodb", "url": "https://github.com/CartoDB/Windshaft-cartodb/pull/407" }
gharchive/pull-request
Fix overviews integration for named layers Fix #400 @dgaubert @rtorre please take a look :+1:
2025-04-01T06:36:48.627903
2017-10-31T14:58:45
269991292
{ "authors": [ "ivanmalagon" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:528", "repo": "CartoDB/cartodb.js", "url": "https://github.com/CartoDB/cartodb.js/pull/1858" }
gharchive/pull-request
Formula dataview https://github.com/CartoDB/cartodb.js/issues/1851 @alonsogarciapablo I don't feel like releasing only the bboxFilter changes. Let's do it after the integration phase of all dataview changes.
2025-04-01T06:36:48.630772
2015-07-15T22:09:24
95301239
{ "authors": [ "ethervoid", "saleiva", "stevage" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:529", "repo": "CartoDB/cartodb", "url": "https://github.com/CartoDB/cartodb/issues/4483" }
gharchive/issue
Broken date format in CSV causes whole file to fail Just had a CSV fail with error 9999 due to a column like this: date 2000-01-01 2000-1-NA IMHO it should never fail to import just because of some data values which don't fit the type that CartoDB has assigned. @stevage what do you propose for the failing cells? Set them to null? Either that or convert the whole column back to string. (I don't know anything about how this is implemented.) If we set the column back to string, we will just have the problem when converting later. So better to do at import time IMO If by "later" you mean, when the user explicitly converts the column type to date, that's better in two ways: You know that they really wanted a date. (Currently you're not that sure - you see one value that looks like a date, and another one that definitely isn't a date.) You're in a better position to give specific error feedback. (As opposed to just aborting with a generic error message.) Deployed a change that fixes this
2025-04-01T06:36:48.633886
2016-08-29T10:04:42
173740309
{ "authors": [ "iriberri", "rafatower" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:530", "repo": "CartoDB/cartodb", "url": "https://github.com/CartoDB/cartodb/issues/9588" }
gharchive/issue
Import timeouts are happening from the import_cleanup process at Table model This part of the code is using connections to the database "as superuser". Perhaps we should switch it by the new methods (direct connections or transactions with timeout) we've been using lately. https://github.com/CartoDB/cartodb/blob/6f060448fb68288d7f9695906df4b556c213b6fc/app/models/table.rb#L302 cc @oriolbx Started to check this. The import_cleanup process is actually flagged as potential code to be removed because the CartoDBfication already manages similar issues at _CDB_Has_Usable_Primary_ID. I'm gonna study it to see if it can be completely dropped, or otherwise treat better the timeout. There's a difference between this code and CartoDBfy: the ruby part gets rid of other known columns, generated usually by Ogr2ogr which are used as PK, but CartoDBfication doesn't. Removing the code without changing the PK name in ogr2ogr would provoke that exporting and importing a table (with cartodb_id) would add an extra gid column on each import, because this extra column wouldn't be dropped. Already solved in the mentioned PR, thus closing this.
2025-04-01T06:36:48.687415
2020-10-08T17:36:13
717536018
{ "authors": [ "goral09", "marc-casperlabs" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:531", "repo": "CasperLabs/casper-node", "url": "https://github.com/CasperLabs/casper-node/pull/343" }
gharchive/pull-request
Queue snapshots This is a working version that allows dumping snapshots of the queue of a node by running $ kill -USR1 $NODE_PID where $NODE_PID should actually the PID of the node that should dump its queue. It will create a queue_dump.json file in the working directoy of the node. Example, dumping a node with a single event in it: $ jq < queue_dump.json { "NetworkIncoming": [], "Network": [], "Regular": [ "AddressGossiper" ], "Api": [] } (the jq is not necessary here, but allows filtering. I use it because it makes the output pretty :)) Currently only the name of the event is dumped, as all the inner fields are skipped. If more output it desired, Serialize should be derived on inner events, possibly with skip_serialize on problematic fields. Some jq examples Counts of all event types: jq 'map_values( map(keys[0] | {"type": ., weight: 1})| group_by(.type) | map ([.[0].type,(.|length)]) | map({(.[0]): .[1]}) )' queue_dump.json Count each queue jq 'map_values(map(keys[0]))' queue_dump.json @marc-casperlabs How about we get this to compile and then squash all commits into one? If people are expected to keep it on the side (without merging to master) and use when necessary, it will be much easier to have it as a single commit that can be cherry-picked easily. Thinking about it, I think it's not harmful to merge this into master - the cost of having a few extra Serialize / Deserialize derives floating around. Actual runtime cost is is a single, never-taken branch on an atomic bool. I'd prefer to feature-gate the functionality (really just the signal-handling code so that we don't query a bool on every event, not suggesting we feature-gate the bulk of the changes in this PR). But that's not a blocker - leave as is if you prefer. I think that's absolutely unnecessary. We are talking about what I believe amounts to one or two CPU instructions, essentially never branching! That's not even remotely close to the threshold I would consider required to make feature gating worthwhile. I can remove a log message somewhere to compensate :) @marc-casperlabs How about we get this to compile and then squash all commits into one? If people are expected to keep it on the side (without merging to master) and use when necessary, it will be much easier to have it as a single commit that can be cherry-picked easily. This will no longer work, as we have more and more events and such implement Serialize to be dumpable. I'd argue that it is more trouble than it is worth at this point. @marc-casperlabs How about we get this to compile and then squash all commits into one? If people are expected to keep it on the side (without merging to master) and use when necessary, it will be much easier to have it as a single commit that can be cherry-picked easily. This will no longer work, as we have more and more events and such implement Serialize to be dumpable. I'd argue that it is more trouble than it is worth at this point. Sure, that comment is over month old. bors r+
2025-04-01T06:36:48.688830
2023-07-04T12:07:30
1787804871
{ "authors": [ "hemidactylus" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:532", "repo": "CassioML/cassio-website", "url": "https://github.com/CassioML/cassio-website/issues/54" }
gharchive/issue
Add token role info in each notebook for clarity Pending a better autogen token, make it clear in each colab that the Token should be a manually-generated "DB Admin" token. in prod now
2025-04-01T06:36:48.702405
2022-05-02T15:54:29
1223052440
{ "authors": [ "Andre-Diamond" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:533", "repo": "Catalyst-Circle/Catalyst-Circle-Admin-Coordination", "url": "https://github.com/Catalyst-Circle/Catalyst-Circle-Admin-Coordination/issues/120" }
gharchive/issue
Town Hall Slides - Monday, 09th May, 2022 Townhall slides - Monday, 09th May, 2022 Previous Town Hall - https://github.com//Catalyst-Circle/Catalyst-Circle-Coordination/issues/108 CC Admin meeting - [ ] Prepare Town Hall slides for Wednesday Previous in series: #108 Next in series: #121
2025-04-01T06:36:48.705815
2017-06-19T18:53:35
236986599
{ "authors": [ "aribornstein", "whaozl" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:534", "repo": "CatalystCode/VOTT", "url": "https://github.com/CatalystCode/VOTT/issues/125" }
gharchive/issue
Export Format Yolo, the bbox txt is all Infinity Infinity Infinity Infinity Export Format Yolo, the bbox txt is all Infinity Infinity Infinity Infinity I will take a look can you send me example output.. @aribornstein Now, I am anxious to turn to yolo, please help me solve, thank you. It looks like there is a scaling error in the image support version of the code. I will hopefully have some time to resolve in the upcoming week. In the meantime it might be worth while to look at the CNTK FastRCNN option. https://github.com/Microsoft/CNTK/wiki/Object-Detection-using-Fast-R-CNN I've identified whats causing the problem and am working to resolve it. It's caused by the fact that while video frames all have the same dimensions images don't. I will update you once this is fixed. Should be resolved now try the new version :)
2025-04-01T06:36:49.058556
2023-10-02T15:00:12
1922095591
{ "authors": [ "joshsh" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:535", "repo": "CategoricalData/hydra", "url": "https://github.com/CategoricalData/hydra/issues/106" }
gharchive/issue
Create bidirectional "merged vertex" and "merged edge" coders In some applications, it is necessary to de-type property graph vertices and edges for the sake of interoperability with type-unaware components, merging them all into a single undifferentiated set of vertices and/or edges -- every vertex or edge with the same type as every other vertex or edge. However, if we simply add a "label" property for each element, and associate labels with the original types of the elements, we can reconstitute them when mapping data in the other direction -- from the merged view back to the typed view. These coders will be needed first in Java, so will be prototyped in that language. Done and tested.
2025-04-01T06:36:49.152571
2024-10-21T09:27:14
2601919457
{ "authors": [ "Propo41", "ali-ahnaf", "talhawebguru" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:536", "repo": "Cefalo/quick-meet", "url": "https://github.com/Cefalo/quick-meet/issues/75" }
gharchive/issue
[BUG] Google icon is not vertically centered in the login page Describe the bug The google icon is not vertically centered To Reproduce Steps to reproduce the behavior: Go to 'Login page' Expected behavior The google icon should be vertically centered Screenshots @jinxsaber hi, are you working on this? @ali-ahnaf Can i work on this issue can you assign me ? @ali-ahnaf Can i work on this issue can you assign me ? Sure, go ahead. Let me know if you need help installing the development environment.
2025-04-01T06:36:49.190613
2020-09-08T20:52:45
696169942
{ "authors": [ "Sindica" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:539", "repo": "CentaurusInfra/arktos", "url": "https://github.com/CentaurusInfra/arktos/issues/693" }
gharchive/issue
Solve lease renewal issue in Kube-Scheduler/KCM/WCM What would you like to be added: In performance test, we saw master components starting to have issue when it cannot renew its lease. Why is this needed: Resolved by bug fixing client sharing same rate limiter
2025-04-01T06:36:49.195413
2021-03-21T03:18:49
836972685
{ "authors": [ "XiaoningDing", "yb01" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:540", "repo": "CentaurusInfra/arktos", "url": "https://github.com/CentaurusInfra/arktos/pull/1040" }
gharchive/pull-request
fix issue 1038. kubeconfig for kcm on TP to use the proxy till the KCM supports direct talk to RPs What type of PR is this? /kind bug What this PR does / why we need it: fix bug in KCM on TP cluster to talk to the proxy instead of pointing to the local cluster --- the RP directly connect from KCM on TP is not yet supported in main -- it is currently tested in 430 branch Which issue(s) this PR fixes: Fixes #1038 Special notes for your reviewer: unblock perf testing in master branch. essential changes is if [[ "${KUBERNETES_TENANT_PARTITION:-false}" == "true" ]]; then create-kubeconfig "kube-controller-manager" ${KUBE_CONTROLLER_MANAGER_TOKEN} "${PROXY_RESERVED_IP}" "443" "https" Does this PR introduce a user-facing change?: none **Testing: 1 TP, 1 RP scaleout scaleup test verified bug fix: ybai2016@yb01-multiple-rp-kubemark-tp-1-master /var/log $ grep -i "Found orphaned Pod" kube-controller-manager.log ybai2016@yb01-multiple-rp-kubemark-tp-1-master /var/log $ we are verifying to integrate with bug fix 1039. so for now, please do not merge this PR. /lgtm /lgtm /approve