added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:55:04.213544
2021-04-22T08:17:59
864676482
{ "authors": [ "OSBotify", "isagoico", "jasperhuangg", "roryabraham", "stitesExpensify" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13910", "repo": "Expensify/Expensify.cash", "url": "https://github.com/Expensify/Expensify.cash/pull/2532" }
gharchive/pull-request
Emoji Picker Menu Navigation Details Added the ability to navigate through the emoji picker menu with arrow keys, and to send the highlighted emoji with [Enter]. It should function the same as the emoji picker in Slack. the highlighted emoji should never be cutoff by the window the first emoji is highlighted upon searching emojis hovering over an emoji with your cursor will also change the currently highlighted emoji if no emoji is highlighted, pressing an arrow key will highlight the first emoji the cursor in the search input is not affected by arrow key presses that change the highlighted emoji Fixed Issues Fixes https://github.com/Expensify/Expensify.cash/issues/2450 Tests I tested this feature on Web and Desktop since those are the only two places it can actually be used, but I also tested it on Mobile Web to ensure that the touchscreen check works and that the Emoji Picker Menu doesn't crash. Opened the emoji picker. Scroll down till you can see the end of on category and the start of another. Use your mouse to hover over an emoji near this boundary. Use the up/down arrow keys to move the highlight around this border. Verified that it skips over the header. Move the highlight at the top and bottom edges of the window with the arrow keys. Verified that the menu scrolls to keep the highlighted emoji in the window. Type in search terms to filter by. Verified that the first search result is highlighted. Verified that nothing breaks when you try to go out of bounds. Verified that you can move the cursor using the appropriate arrow keys when they would move the highlight out of bounds. Pressed [Enter]. Verified that the highlighted emoji was inputted into the compose box. Repeat steps 1-6 two more times to verify that event listeners are removed and added correctly. QA Steps Identical to the tests above. See videos for examples. Tested On [x] Web [ ] Mobile Web [x] Desktop [ ] iOS [ ] Android Screenshots Web https://user-images.githubusercontent.com/31285285/115820241-0048f200-a433-11eb-9dd8-d6c274158365.mp4 Desktop https://user-images.githubusercontent.com/31285285/115819992-8ca6e500-a432-11eb-9322-4cabf9f107f3.mp4 cc @stitesExpensify @pecanoro @stitesExpensify Thanks for clarifying my questions! Should be good for another review. Tested and it works great! Just some code organization comments and then we should be good to go! Hey @stitesExpensify thanks so much for the review! I've addressed most of your concerns but have a few clarifications of my own. Let me know what you think! @stitesExpensify Ready for another review, thanks for your feedback! 🚀 Deployed to staging in version: 1.0.35-2🚀 platform result 🤖 android 🤖 success ✅ 🖥 desktop 🖥 success ✅ 🍎 iOS 🍎 success ✅ 🕸 web 🕸 success ✅ Emoji Picker - Moving by arrow keys will eventually lose selection Expected result Focus isn't lost even after reaching the last emoji on the list. Actual result After reaching the last emoji on the list, the focus is not visible anymore. Action Performed Log in to expensify.cash Navigate to a conversation Click on the emoji icon Use the arrow keys to reach the bottom of the list Platform Web ✔️ Build: 1.0.36-0 Notes/Images/Video https://user-images.githubusercontent.com/44479856/116945606-15e5c500-ac46-11eb-9e65-ce28fcc1e469.mp4 @jasperhuangg Not sure if the issue above should be a deploy blocker, let me know if I should open as a separate issue. No need for this to be a deploy blocker, @isagoico, can you create a separate issue and tag @jasperhuangg?
2025-04-01T04:55:04.218596
2021-05-26T02:03:03
901628615
{ "authors": [ "Jag96", "OSBotify", "roryabraham" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13911", "repo": "Expensify/Expensify.cash", "url": "https://github.com/Expensify/Expensify.cash/pull/3149" }
gharchive/pull-request
Make easily noticeable temporary change Details This PR is being used to test the CP workflow. It is "PR A", referenced in Step 5 of the testing plan here. This change will be reverted soon Fixed Issues Part of testing https://github.com/Expensify/Expensify.cash/pull/2978 Tests Merge this PR Verify that it gets the ✋ This PR will not be deployed ... comment Verify that a staging deploy does not occur. Verify that the red FAB is not present on staging. Later complete steps 10-12 of the testing plan QA Steps Verify that the FAB is not red 😂 Skipping tests/actions to test CP PR flow :hand: This PR was not deployed to staging yet because QA is ongoing. It will be automatically deployed to staging after the next production release. Test steps 1-4 passed 👍
2025-04-01T04:55:04.221072
2024-03-29T13:43:44
2215394197
{ "authors": [ "BartoszGrajdek" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13912", "repo": "Expensify/expensify-common", "url": "https://github.com/Expensify/expensify-common/pull/674" }
gharchive/pull-request
feat: remove emoji tag from data-code-raw in pre Fixed Issues $ GH_LINK Tests What unit/integration tests cover your change? What autoQA tests cover your change? What tests did you perform that validates your changed worked? QA What does QA need to do to validate your changes? What areas to they need to test for regressions? I have read the CLA Document and I hereby sign the CLA recheck
2025-04-01T04:55:04.235043
2020-10-30T10:25:53
733058692
{ "authors": [ "AlsoInteresting", "Morph21" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13913", "repo": "Exslims/MercuryTrade", "url": "https://github.com/Exslims/MercuryTrade/issues/415" }
gharchive/issue
sound source How do I change sound source? Whan I change headset in POE, this doesn't copy to Mercury Trade. Even after restarting the pc. When I put the old headset in the pc, it works. MercuryTrade here is discontinued. Feel free to try a community fork version here: https://github.com/Morph21/MercuryTrade-Community-Fork It's possible to change sounds. Also different headset is not an issue on POE or MT side, check your system settings for solving sound issues MercuryTrade here is discontinued. Feel free to try a community fork version here: https://github.com/Morph21/MercuryTrade-Community-Fork It's possible to change sounds. Also different headset is not an issue on POE or MT side, check your system settings for solving sound issues
2025-04-01T04:55:04.324994
2019-01-24T08:38:11
402596173
{ "authors": [ "Leylan24", "Sonic0", "lgr9603", "tdh8316" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13914", "repo": "F-Army/arduino-dw1000-ng", "url": "https://github.com/F-Army/arduino-dw1000-ng/issues/139" }
gharchive/issue
Example TwoWayRanging is not work. Step 1: What do you want? TwoWayRangingInitiator and TwoWayRangingResponder do not work. therefore I want them to work. Step 2: Describe your environment Arduino IDE version or Toolchain: Arduino IDE 1.8.7 Arduino device: Arduino Uno R3 Used adapter board: None Step 3: Describe the problem Affected file(s) or example(s): TwoWayRangingResponder.ino TwoWayRangingInitiator.ino Observed Results: It seems not to work! Thus I tried Basic Connectivity Test. I'm not sure it's correct output because I'm a beginner of Arduino and DWM1000. But I don't think there is any noticeable problem. And then I uploaded Basic Sender/Receiver to Arduino. I still have no clue about this, but it seems to work fine. EDIT: thotro/arduino-dw1000 I tried also thotro's library. The device is removed immediately after it is added. Why are BasicSender and Receiver working but TwoWayRanging not working? Expected Results: Jaja, I expected they work. Wiring Where did you put the 10K resistor ? I am having the same issue my Basic Receiver and Basic Sender is working but the Two Way Ranging example doesnt work @lgr9603 Where did you put the 10K resistor ? I am having the same issue my Basic Receiver and Basic Sender is working but the Two Way Ranging example doesnt work I think that the @tdh8316 's solution is relative to section 5.2.2 inside DWM1000 datasheet @Leylan24 I'm a coworker of @tdh8316 . Well, we put the 10K resistor on pin 2 to remove noises flowing through it. And drop the voltage of digital pins from 5V to 3.3V. Dwm1000 can work between 2.4V and 3.75V. (Along to datasheet on decawaver.com) As I said, arduino uno digital pins use 5V, bout DWM1000 sends output signals in 3.3V. So we should increase voltage of DWM1000 signal from 3.3V to 5V. Thanks for you reply. But does the ranges only print when the IRQ pin gets pulled high? But does the ranges only print when the IRQ pin gets pulled high? With this library, the IRQ pin is not required. There are two different init processes, with or without Interrupt pin.
2025-04-01T04:55:04.331748
2020-10-06T22:10:44
716055084
{ "authors": [ "dstokesf5", "jmcalalang" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13915", "repo": "F5Networks/f5-appsvcs-extension", "url": "https://github.com/F5Networks/f5-appsvcs-extension/issues/357" }
gharchive/issue
Ability to create DNS_Logging_Profile from AS3 Is your feature request related to a problem? Please describe. Feature Request Describe the solution you'd like AS3 has the ability to reference DNS_Logging_Profile as a pre-created item, we would like the ability to create it. I don't see an option in the schema to create one. "description": "Reference to a DNS logging profile", "type": "object", "properties": { "bigip": { "description": "Pathname of existing BIG-IP DNS logging profile", "type": "string", "format": "f5bigip" } }, Describe alternatives you've considered There isn't an Ansible module for this either, it would require imperative calls to create the object. REST endpoint: /mgmt/tm/ltm/profile/dns-logging Options: appService: description: enableQueryLogging: enableResponseLogging: includeCompleteAnswer: includeQueryId: includeSource: includeTimestamp: includeView: logPublisher: name: Additional context Object in question: Thank you for your feedback. Due to the age of this feature request, I would like to know if this is still something you need or are interested in.
2025-04-01T04:55:04.336913
2024-01-02T04:32:21
2061931002
{ "authors": [ "dstokesf5", "mdditt2000", "normelton" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13916", "repo": "F5Networks/f5-appsvcs-extension", "url": "https://github.com/F5Networks/f5-appsvcs-extension/issues/790" }
gharchive/issue
Unable to compare SSL SNI/hostname against data group Environment Application Services Version: 3.48.0 BIG-IP Version: <IP_ADDRESS> Summary When building a policy, I can compare the HTTP hostname against a data group, but I cannot compare the SSL SNI hostname against the same data group. I get the following error Steps To Reproduce Submit the following policy: "http-redirect": { "class": "Endpoint_Policy", "strategy": "first-match", "rules": [ { "name": "permit", "conditions": [{ "type": "httpHost", "event": "request", "host": { "operand": "ends-with", "datagroup": { "use": "permitted-domains" } } }] } ] } It works fine. Submit this policy instead: "https-block": { "class": "Endpoint_Policy", "strategy": "first-match", "rules": [ { "name": "permit", "conditions": [{ "type": "sslExtension", "event": "ssl-client-hello", "serverName": { "operand": "ends-with", "datagroup": { "use": "permitted-domains" } } }] } ] } It fails to post, I get the following error: { "code": 422, "message": "declaration failed", "response": "01020036:3: The requested datagroup lookup failure: (use) was not found.", "host": "localhost", "tenant": "init-policy", "runTime": 2365 }, it looks like two things are changing from your working and not working examples. If the permitted-domains data group is managed by AS3, can you try updating your failing example to use a use pointer instead of a bigip pointer? Good observation. In all my testing, I copied/pasted the wrong example. The behavior is the same when with a use pointer. I updated my post above. Add Jira AUTOTOOL-4132 to next sprint for prioritization discussion
2025-04-01T04:55:04.341988
2021-08-24T20:54:45
978478155
{ "authors": [ "amolari", "dstokesf5" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13917", "repo": "F5Networks/f5-telemetry-streaming", "url": "https://github.com/F5Networks/f5-telemetry-streaming/issues/161" }
gharchive/issue
Datadog: add customizable service field option Is your feature request related to a problem? Please describe. The service field option, used by Datadog logging, is not customizable but statically defined in the code. const DATA_DOG_SERVICE_FIELD = 'f5-telemetry' Describe the solution you'd like Be able to set a custom value in the Datadog consumer declaration. Thank you for your feedback. I have added your request to our internal product backlog as AUTOTOOL-2753. I guess it had been fixed in v1.24.0? This has been resolved in the 1.24.0 release.
2025-04-01T04:55:04.374534
2023-09-12T12:56:42
1892466057
{ "authors": [ "coveralls", "domna", "sherjeelshabih" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13918", "repo": "FAIRmat-NFDI/pynxtools", "url": "https://github.com/FAIRmat-NFDI/pynxtools/pull/157" }
gharchive/pull-request
Adds passing of objects to transmission reader This adds the functionality of passing a dict to the transmission readers convert function. @hampusnasstrom So you pass Dict type objects and append them to the template? No other object types are accepted currently, right? Pull Request Test Coverage Report for Build<PHONE_NUMBER> 10 of 17 (58.82%) changed or added relevant lines in 3 files are covered. 4 unchanged lines in 3 files lost coverage. Overall coverage decreased (-0.005%) to 51.087% Changes Missing Coverage Covered Lines Changed/Added Lines % pynxtools/dataconverter/readers/utils.py 6 13 46.15% Files with Coverage Reduction New Missed Lines % pynxtools/dataconverter/readers/shared/map_concepts/mapping_functors.py 1 11.86% pynxtools/nyaml2nxdl/nyaml2nxdl_forward_tools.py 1 79.42% pynxtools/nexus/nxdl_utils.py 2 74.79% Totals Change from base Build<PHONE_NUMBER>: -0.005% Covered Lines: 5993 Relevant Lines: 11731 💛 - Coveralls So you pass Dict type objects and append them to the template? No other object types are accepted currently, right? Exactly, I expect an iterable of dicts. Every other element is simply warned about and ignored
2025-04-01T04:55:04.435840
2019-11-07T21:43:25
519530317
{ "authors": [ "Robadob" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13919", "repo": "FLAMEGPU/FLAMEGPU2_dev", "url": "https://github.com/FLAMEGPU/FLAMEGPU2_dev/pull/130" }
gharchive/pull-request
Feature: Host Functions This pull request/branch tracks the development of host functions. Example documentation: https://gist.github.com/Robadob/081d04d2d515ce3b090bf389dc9195dc This addresses many issues, as such the branch may be long running and merged incrementally: PR #135 [x] #46 - Host functions init/step/exit bd5097c [x] #54 - Exit early from step function bd5097c [x] #47 - Host layer functions (step function in any layer) 7c0c204 [x] #53 - CUB reductions (min, max, sum, count) 046fa07 [x] #80 - Histogram reduction 09ac715 [x] #83 - Custom reduction functions (cub reduce, thrust transformReduce) ffa28c9 PR: #138 [x] #90 - Constant/Immutable Environmental variables PR: #196 [x] #61 - Agent creation on the host Misc/Low Priority [ ] #162 - Large environment properties (held in global memory) [ ] #51 - 'Access agent data on the host'???? TODO: [x] Rename FLAMEGPU_API->FLAMEGPU_DEVICE_API Host function definition [x] Host Fn [x] Host Condition Add Host fn/condition to model [x] Init [x] Step [x] Exit [x] Layer [x] ExitCondition gpu/CUDAAgentModel (Controls execution of model) step(): Update to support host functions [x] Init [x] Step [x] Exit [x] HostLayer [x] ExitCondition Environmental vars [x] Define [x] Read in device fn Runtime/flamegpu_host_api.h Environmental vars [x] Read on host fn [x] Write in host fn [x] Wrapper for <random> Reduction [x] Sum [x] Min [x] Max [x] Reduce (user defined template binary op a+b->c) Others from thrust? [x] Count [x] TransformReduce Message reductions? Histogram [x] Even Custom bounds? (This would require having to cache their bounds on device somehow) [x] Agent Creation [x] Tests [x] Example usage in model TODO Monday: Custom reduction fn \<random> wrapper using runtime seed. This PR has served it's purpose all parity tasks complete.
2025-04-01T04:55:04.441736
2024-03-20T21:10:03
2198574482
{ "authors": [ "FMaz008", "crysallis", "runew0lf" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13920", "repo": "FMaz008/VineHelper", "url": "https://github.com/FMaz008/VineHelper/issues/58" }
gharchive/issue
Request: Option to toggle the "green styling" It makes it very hard to read with darkmode on As an "older gentleman" it kinda all blurs into one colour! The main problem is the text color, which we will improve upon. But we have been working toward centralizing styling colors so it might be possible in the future (not yet) to have themes. Baby steps... ... but for now we'll fix the text which is badly colored as you showed in your screenshots. Thank you for giving feedback, some of those area haf not occured to me, so this is definitely helpful. I'll address this if you want Fixed.
2025-04-01T04:55:04.557982
2020-06-12T16:58:00
637888270
{ "authors": [ "JamesWP", "eurocoinsalbum" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13921", "repo": "FRUK-Simulator/Simulator", "url": "https://github.com/FRUK-Simulator/Simulator/issues/37" }
gharchive/issue
create robot control react element to show the current state of the robot sensor data, motor output etc..., position... I implemented this already. Will add a PR today #46 has an extremely simple version of this, only contains the motor power for now, but ill close this issue.
2025-04-01T04:55:04.635509
2022-05-29T07:57:55
1251818321
{ "authors": [ "Fabian42" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13922", "repo": "Fabian42/bash_scripts", "url": "https://github.com/Fabian42/bash_scripts/issues/44" }
gharchive/issue
Automatically download new SoundCloud tracks The text in the followthatpage mails might be enough to automatically download new SoundCloud tracks from artists I follow with that tool. The hardest part of that might be reading the mail using Bash, but surely there's a tool for that. Test cases would have to include tracks with the same name as something that that artist has already uploaded before, because then they get an attached "-2" or similar in the URL, but it might also be slightly different. Checking the upload date and comparing that to the mail should definitely be done. Maybe that's also enough and the actual content of the mail can be ignored, just knowing that someone uploaded and then downloading the last week of their tracks would be OK. Filtering out false positives like in the mail macro would also be necessary for as long as that followthatpage bug exists. Alternatively, there's something with RSS on SoundCloud, which would be a less hacky solution (RSS→dl instead of website→follow→mail→guess link→verify date→dl). But I never got any of those RSS feeds to work so far. alternative: query pages myself, can also apply to more than SC auto-download WL and temp_music similar: webcomics, Lateral
2025-04-01T04:55:04.777566
2023-06-07T03:55:09
1744981636
{ "authors": [ "chris-t-li" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13923", "repo": "Fantasy-Fit/fantasy-fit-web", "url": "https://github.com/Fantasy-Fit/fantasy-fit-web/pull/173" }
gharchive/pull-request
refactored dockerfiles to reduce image sizes Update Ruby to 3.2.2; Add redis and helper services to docker-compose; Refactored dockerfile for client - use node:18-alpine and multi stage build to reduce image size Refactored dockerfile for api - use ruby:3.2.2-slim-bullseye and update dependencies to reduce image size Time to build ~ 122s Image sizes: client reduced from 1.27GB to 309MB. rails api reduced from 1.07GB to 678MB This addresses issue #172 Significantly reduced time to stop containers: from ~10+ seconds for client to 0.4s
2025-04-01T04:55:04.859085
2020-07-28T13:00:52
667053734
{ "authors": [ "NareshNama", "bell-steven" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13924", "repo": "FaridSafi/react-native-google-places-autocomplete", "url": "https://github.com/FaridSafi/react-native-google-places-autocomplete/issues/595" }
gharchive/issue
onChangeText is not working <GoogleAutoCompleteAddress onPress={(data, details = null) => { this.fetchGoogleLocationDetails(details); }} textInputProps={{ value:lossAddress, onChangeText:{_onChangeTextValue()} }} editable={true} autoFocus={false} /> onChangeText is not working and showing syntax error Please provide a reproducible example.
2025-04-01T04:55:04.862074
2017-04-03T16:26:44
218989143
{ "authors": [ "ConnorRigby", "Daniil145", "RickCarlino" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13925", "repo": "FarmBot/farmbot_os", "url": "https://github.com/FarmBot/farmbot_os/issues/282" }
gharchive/issue
End stops don't work on after new release my.farmbot.io You disable endstops? @Daniil145 sorry about that, they werent disabled, but there was a new version of teh Arduino Firmware released. Could you try disabling and re enabling end stops? @ConnorRigby I dont understand "disabling and re enabling end stops" on interface dont have toggle off and on Endpoints. What can I do to activate them? With the loss of the Internet (especially when I use mobile Internet) the system behaves unpredictably. Especially this is critical for the Z axis. To do this, I need limit switches. Oh sorry i didnt realize they werent on the interface any more. @RickCarlino any thoughts on this one? @Daniil145 I've recently added endstop toggle buttons to the device panel. They will be available in the next production release (probably this week). @Daniil145 We just did a production deploy. Please let us know if you find any issues.
2025-04-01T04:55:04.972977
2023-03-06T17:59:35
1611946418
{ "authors": [ "rachithrr", "vkrishnanfb" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13926", "repo": "FeatureBaseDB/featurebase", "url": "https://github.com/FeatureBaseDB/featurebase/pull/2296" }
gharchive/pull-request
FB-1895: Implement DateTimeFromParts this uses the tiem.date() function Well, I have to add limits to the time ranges. For example, dax expects the year range [0 , 9999]. I'll update the code and re request to review. Looks good to me.
2025-04-01T04:55:04.997579
2022-06-23T18:18:55
1282777484
{ "authors": [ "FeiyouG", "bugabinga" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13928", "repo": "FeiyouG/command_center.nvim", "url": "https://github.com/FeiyouG/command_center.nvim/issues/15" }
gharchive/issue
Error if nothing was selected Open Command Center Type some gibberish until the filtered list list empty Hit <ENTER> Alternatively, open Command Center without and bindings added and hit <ENTER>. E5108: Error executing lua ...mand_center/lua/telescope/_extensions/command_center.lua:151: attempt to index local 'selection' (a nil value) stack traceback: ...mand_center/lua/telescope/_extensions/command_center.lua:151: in function 'run_replace_or_original' ...ite/pack/paqs/opt/telescope/lua/telescope/actions/mt.lua:65: in function 'key_func' .../site/pack/paqs/opt/telescope/lua/telescope/mappings.lua:242: in function 'execute_keymap' [string ":lua"]:1: in main chunk Telescope + CC config: telescope.setup { extensions = { command_center = { -- Change what to show on telescope prompt and in which order -- Currently support the following three components -- Components may repeat components = { command_center.component.DESCRIPTION, command_center.component.KEYBINDINGS, -- command_center.component.COMMAND, }, -- Change the separator used to separate each component separator = " ", -- When set to false, -- The description compoenent will be empty if it is not specified auto_replace_desc_with_cmd = false, } } } -- Plug Command Center into Telescope telescope.load_extension('command_center') Fixed in e3b32c7a2b8ffbb92bd4b7158fcb44eb40ba5d20
2025-04-01T04:55:05.009584
2021-01-29T17:09:32
797044762
{ "authors": [ "Femme-js", "Swastik1710" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13931", "repo": "Femme-js/Hoaxify", "url": "https://github.com/Femme-js/Hoaxify/issues/34" }
gharchive/issue
Adding a seperate line for admin in readme.md file and creating a Contributors table which updates automatically. As of now , the Contributors table in the readme.md file is not up to date. Moreover one needs to update it manually each and every time. I wanna automate this by adding a new Contributors table which will update by itself. Moreover , I wanna add a separate line to showcase the Project Admin if you want. Can you please assign this to me ? Thanks. Can you please assign this to me ? Thanks. Great! Hey @Femme-js , I've made a PR. Please have a look,
2025-04-01T04:55:05.026214
2024-12-27T22:20:46
2761313287
{ "authors": [ "lesniakt", "smturro2" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13932", "repo": "Fernando-Urbano/portfolio-management", "url": "https://github.com/Fernando-Urbano/portfolio-management/issues/36" }
gharchive/issue
Considering a Environment Manager @lesniakt Do you have any opinions/thoughts on our environment management strategy? Now would be a good time to switch rather then later. I would like to stick with the anaconda environment and rename the environment so it doesn't overlap with the standard FINM environment used in classes and august review. This approach would be very familiar to new members coming from the FINM program. When we have tests we would build the package locally using pip install . -e and test would be run through a build instead of importing it from the folder over. Using pipenv will make it easier to upgrade package versions. We can also set this up so it automatically builds and installs the local package. I've worked with this a ton and is great. The downside being that the setup is different making it less friendly to new contributers in the FINM department. If possible we should prioritize making ease of development for new contributers. I don't have a ton of experience upgrading an anaconda env so please add some thoughts if you have. I imagine we would simply reinstall the whole anaconda env (without version locks) and then update the requirements files. I've personally not used pipenv. Typically for quick and dirty projects venv is sufficient, and for larger scale projects, especially production projects, I tend to see Conda used as usually there are non-python dependencies involved. From the literature I've read, it seems like non-python dependencies is where pipenv would struggle. Numpy and scipy come to mind as potentially requiring non-python dependencies -- but again I haven't played with it. I would be curious how reproducible pipenv is with those, as this package will presumable use those dependencies, I think the typical student in this program is just editing from requirements.txt files, so I share your concern on simplicity here. My lean would be to just use the existing requirements.txt workflow until we have some level of code coverage. I think that is a reasonable prerequisite to have before changing environment management workflows, as we'll want reasonable certainty that the changes don't break workflows accidentally. Regarding "ease of use" for upgrading packages, I would suggest looking at integration with RenovateBot. A quick looks says both conda/pipenv are supported. Again, we'd want code coverage in place before doing this.
2025-04-01T04:55:05.029201
2023-03-27T20:20:35
1642758031
{ "authors": [ "AlekSi", "b1ron" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13933", "repo": "FerretDB/FerretDB", "url": "https://github.com/FerretDB/FerretDB/issues/2299" }
gharchive/issue
Projection with _id returns all fields Versions commit.txt: a06bd4bf1832bceed7189cff07d788cdb11e2155 version.txt: v0.9.4-1-ga06bd4bf What did you do? See repro in #2298. What did you expect to see? 2 documents returned with just the _id field. What did you see instead? All fields. Environment - OS: Ubuntu - Architecture: x86_64 - Version: 22.10 - Go: go version go1.20 linux/amd64 - Git: 2.37.2 - Deployment: Docker - Deployment details: 23.0.1 See repro in https://github.com/FerretDB/FerretDB/pull/2298. That PR should be re-opened and updated to check that this issue is indeed fixed
2025-04-01T04:55:05.033055
2023-03-28T08:31:42
1643459807
{ "authors": [ "AlekSi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13934", "repo": "FerretDB/FerretDB", "url": "https://github.com/FerretDB/FerretDB/pull/2301" }
gharchive/pull-request
Use iterators for projections After #2254. Description Closes #{issue_number}. Readiness checklist [ ] I added/updated unit tests. [ ] I added/updated integration/compatibility tests. [ ] I added/updated comments and checked rendering. [ ] I made spot refactorings. [ ] I updated user documentation. [ ] I ran task all, and it passed. [ ] I ensured that PR title is good enough for the changelog. [ ] (for maintainers only) I set Reviewers (@FerretDB/core), Assignee, Labels, Project and project's Sprint fields. [ ] I marked all done items in this checklist. Merged into #2254.
2025-04-01T04:55:05.036532
2023-05-30T13:27:15
1732279705
{ "authors": [ "rumyantseva" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13935", "repo": "FerretDB/FerretDB", "url": "https://github.com/FerretDB/FerretDB/pull/2739" }
gharchive/pull-request
Make testFindAndModifyCompat work with compatTestCaseResultType Description Closes #2734. Readiness checklist [ ] I added/updated unit tests. [x] I added/updated integration/compatibility tests. [ ] I added/updated comments and checked rendering. [x] I made spot refactorings. [ ] I updated user documentation. [ ] I ran task all, and it passed. [x] I ensured that PR title is good enough for the changelog. [x] (for maintainers only) I set Reviewers (@FerretDB/core), Labels, Project and project's Sprint fields. [x] I marked all done items in this checklist. @w84thesun ah, thanks for checking! I haven't pushed the code 😅
2025-04-01T04:55:05.043431
2020-06-16T13:38:23
639673110
{ "authors": [ "Sc0r", "klightspeed" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13936", "repo": "FezVrasta/ark-server-tools", "url": "https://github.com/FezVrasta/ark-server-tools/issues/1112" }
gharchive/issue
Please ensure the server is running the latest version of all mod content I get this issue whenever I try to join my server. "Please ensure the server you are trying to join is running the latest version of all mod content" Why? ~$ arkmanager checkmodupdate --updatemods @crystal Running command 'checkmodupdate' for instance 'crystal' All mods are up to date EDIT huh, now it works without anything changing. odd. It could have been that the client hadn't finished updating its mods. Since mods do not have any inherent version, there is no way to tell whether the server or client has more up-to-date mods.
2025-04-01T04:55:05.045342
2018-12-13T14:39:28
390708050
{ "authors": [ "juanjoDiaz" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13937", "repo": "FidelLimited/serverless-plugin-warmup", "url": "https://github.com/FidelLimited/serverless-plugin-warmup/pull/86" }
gharchive/pull-request
Minor improvements Just a bit of cleaning up. Avoid doing the same job (loading the config) twice and move all the checks to top-level hooks. I noticed that all the possible variable including the absolute path of the project where being added to the generated lambda so the last commit cleans that so only necessary variables are included.
2025-04-01T04:55:05.046372
2021-03-20T18:11:39
836877455
{ "authors": [ "AlexLanzano" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13938", "repo": "Field-Programmable-Smartwatch/firmware", "url": "https://github.com/Field-Programmable-Smartwatch/firmware/issues/7" }
gharchive/issue
Implement log levels Implement different log levels (Error, debug, info, etc). Have log messages be able to be hidden or shown based on build environment variable or changed via a serial user interface Implemented in a3bf35b2de76ca4b34aaf711bda04b4a4842dd4e
2025-04-01T04:55:05.058290
2020-07-31T01:14:08
669317228
{ "authors": [ "joshmorel", "jyclam", "kterry1", "manualzuru" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13939", "repo": "FightPandemics/FightPandemics", "url": "https://github.com/FightPandemics/FightPandemics/issues/1306" }
gharchive/issue
Refactor and remove OrgProfileComplete page, instead use the ProfileCompleted page The call to action button should say View Help Board unless the unlogged in user clicked on Create a Post and went through the sign up process, in which case it should say Continue posting Currently the pages look like this: OrgProfileComplete ProfileCompleted @jyclam I'll take this on, but I wanted to clarify the ask here. It says "Refactor and remove OrgProfileComplete page". Am I refactoring OrgProfileComplete, or completely deleting it and just using "ProfileCompleted" where "OrgProfileComplete" is used? Didn't mean to close. Let me know your thoughts on the above. I wanted to make sure I know what is wanted here before proceeding. Also, how can i differentiate whether the user went through "Create Post" and then signed up, or whether they signed up the usual way.? Hi @kterry1, from my understanding, there are 2 different pages (UI wise) that are trying to accomplish the same function. The main difference is the 2nd CTA changes depending on the path that you went through. With the ProfileCompleted page, we should be able to do everything that is needed. There are 3 paths with the last CTA that changes. 1- if you click in register -> Check the Help Board 2- if you try to make a comment + like and you aren't signed in. -> Check the Post (the post that you were liking or commenting) 3- if you click on Create Post and you aren't signed in. -> Continue Posting Let me know if there are other questions🙏🏽 Thanks @manualzuru! I'll dig into it today and see what I can do. @jyclam @manualzuru So I did some testing with this. Since the Sign In, Create a post, and Join Now buttons all take you to '/auth/signup', I'm not seeing a way to conditionally render based on the path the user took to get to '/profile-completed' or '/create-organisation-complete' and have it all happen in one page. If the wording of that button is really important, then I'd say we could stick with multiple pages, otherwise, here is what I had so far, but I don't think it really accomplishes what is asked for: https://share.getcloudapp.com/OAuqOwG0 Maybe @joshmorel or @idjevm have an idea? Side note: We need to change all instances of "organisation" in the codebase to "organization". You can probably do a conditional sessionStorage like here https://github.com/FightPandemics/FightPandemics/blob/487e856300292dc7e754be23134b4025d3dfdb82/client/src/pages/PostPage.js#L162 https://developer.mozilla.org/en-US/docs/Web/API/Window/sessionStorage I'll give that a try. Thanks! The fix is based on jyclam's original post as it made the most sense. @kterry1 @jyclam Sorry I didn't take a look at this early when you tagged me - since the PR is in. I'm really apprehensive about using the same file/component for different routes. It can result in a hard to maintain situation as I see with our "Login.js" file - see #1164 (that one is more complex, so not exactly comparable but similar principle). I think it would be better to have more re-usable components for similar pages but use one top-level file per route. You see this with a lot of filesystem based frameworks like vuex/next.js etc and I think it's a good approach. With this PR #1328 - we need to fix the issues - but we should decide: Make an exception here - and address the issues as described in my first comment in the PR Or come up with a rule and stick to it - one file per route Thoughts @robinv85 @idjevm ? Also I did try to figure out what the design's flow is supposed to be after create org complete in Figma but it wasn't clear (https://www.figma.com/file/kT6SJLL4v3w06yFF2cGIun/FightPandemics-MVP?node-id=1852%3A8716). Is "View my profile" always supposed to go to MY profile as opposed to the recently created organisation's profile (current state). The latter makes more sense but the design doesn't indicate which. But also in such a case I'd expect the label to be different "View organisation" instead. Maybe the design was incomplete after copy/paste from user profile? Any insight into this @manualzuru
2025-04-01T04:55:05.062251
2021-05-25T21:27:46
901390178
{ "authors": [ "adorapham", "hiran11", "xJamesScott" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13940", "repo": "FightPandemics/FightPandemics", "url": "https://github.com/FightPandemics/FightPandemics/issues/2398" }
gharchive/issue
User Store 9.1-Unable to set the role for a member after accepting the applicant Unable to set the role for a member after accepting the applicant. please see the record on , near the end starting from 1:30 https://app.slack.com/client/T010BHVS3EF/C010BHVT1MZ/files/F022X4NME85 @adorapham Bug fix issued @ChunPingYang @xJamesScott Looks like the issue is not fixed . As an individual user even after i join the Org "Join Us" CTA button is not disabled . Even when my application as an individual has been accepted "Join Us " CTA button is kept enabled. The individual name is listed as member in Members Tab. As a owner of the account , unable to set the role permission. The static data is populated in the Role description for Wiki editor and Administrator. https://user-images.githubusercontent.com/51660881/120036342-01d77e00-bfce-11eb-88e5-8639e007f2f9.mp4 https://user-images.githubusercontent.com/51660881/120036352-04d26e80-bfce-11eb-946f-2ca737365fdf.mp4 https://user-images.githubusercontent.com/51660881/120036373-0c921300-bfce-11eb-8856-7310b676ecfc.mp4 https://user-images.githubusercontent.com/51660881/120036393-1287f400-bfce-11eb-9b03-2fa09c872adb.mp4
2025-04-01T04:55:05.071341
2020-04-03T21:09:38
593642449
{ "authors": [ "CeesWang", "MohamedEl1", "T-Allen", "indirabhowmick", "manualzuru", "robinv85" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13941", "repo": "FightPandemics/FightPandemics", "url": "https://github.com/FightPandemics/FightPandemics/issues/50" }
gharchive/issue
User Profile - Front End [ ] frontend user profile Here is the Requirements for Create Individual Profile UI https://www.notion.so/fightpandemics/Create-Individual-Profile-UI-cf719e2292874303a2ce943c912bdbf3 Finished styling the user profile, created edit-profile / edit-account routes and components wip styling on edit-profile and edit-account components handle data flow for my activity in user profile @CeesWang - any updates on this issue? @indirabhowmick talking to T-Allen regarding further progress. @CeesWang Thanks for the update. Please keep us posted on the progress of the issue @CeesWang @T-Allen is this solved? @manualzuru is this the same as issue #176 ? It is the same mobile and desktop only styling changes depending on the device @T-Allen I'm still confused, because going by what's on Zeplin, the pages seem quite different: create-profile (desktop) create-profile (mobile) edit-profile (desktop) edit-account-info (desktop) @T-Allen by Mobile we meant mobile web not the native mobile app. Please reach out if there are any questions. @indirabhowmick Right, I am working on the web app. @T-Allen you are right. Those are 2 different screens, you are right. We need both for having profile. Can we update this ticket or add another ticket and maybe tags to have a better view on the user profile status? The user v2 data model is not implemented yet and I'm working on "create user" so I think both of these edit / view user aren't implemented with API yet? No they are not implemented with API yet On Tue, May 26, 2020 at 5:27 PM Robin Venneman<EMAIL_ADDRESS>wrote: Can we update this ticket or add another ticket and maybe tags to have a better view on the user profile status? The user v2 data model is not implemented yet and I'm working on "create user" so I think both of these edit / view user aren't implemented with API yet? — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/FightPandemics/FightPandemics/issues/50#issuecomment-634288870, or unsubscribe https://github.com/notifications/unsubscribe-auth/AIAFWLSDD5OCCSNYDMC6JBLRTQX5DANCNFSM4L4OHU2Q . @CeesWang i'm working on this to finish some details and integrate with API @robinv85 awesome let me know if theres anything I can do regarding this issue
2025-04-01T04:55:05.079230
2017-01-08T20:24:13
199443931
{ "authors": [ "FiloSottile", "zmanian" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13942", "repo": "FiloSottile/zcash-mini", "url": "https://github.com/FiloSottile/zcash-mini/issues/1" }
gharchive/issue
Should also export a viewing key The Zcash protocol also has support for view keys. Views keys can determine if a tx has been received but not spend them. I don't believe there is any support for view keys in the full node wallet yet but we could generate and export a view key that could immediately be imported by a user's full node. Ah wonderful, I thought they didn't exist and wished they did. Yep, let's add them, and add a mode to input the secret key instead of generating it, to get address/view key from a saved key.
2025-04-01T04:55:05.080272
2024-12-28T08:10:41
2761624014
{ "authors": [ "FinalForEach", "scoundri" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13943", "repo": "FinalForEach/Cosmic-Reach-Localization", "url": "https://github.com/FinalForEach/Cosmic-Reach-Localization/pull/273" }
gharchive/pull-request
Fixed Grammar Issues and made it clearer and easier to read. Turkish I hate Polite suffixes Thank you! Merging now.
2025-04-01T04:55:05.097419
2017-05-11T14:23:15
228009262
{ "authors": [ "JoyceBabu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13944", "repo": "Financial-Times/polyfill-service", "url": "https://github.com/Financial-Times/polyfill-service/issues/1200" }
gharchive/issue
Element polyfill fails on UCBrowser Mini The Element polyfill fails on UCBrowser Mini because document node cannot have more than one child The polyfill uses attachEvent. The event onpropertychange proprietary to IE. Irrespective of that, using addEventListener will prevent it from failing completely on other browsers that require Element polyfill. createElement('*') is used, which is not valid. Closing this as per https://github.com/Financial-Times/polyfill-service/pull/1201
2025-04-01T04:55:05.110659
2024-03-07T07:09:08
2173126675
{ "authors": [ "codecov-commenter", "zemyblue" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13945", "repo": "Finschia/finschia-proxy", "url": "https://github.com/Finschia/finschia-proxy/pull/35" }
gharchive/pull-request
Revert "feat: apply the changes of Finschia v3.0.0-rc1" Reverts Finschia/finschia-proxy#34 Codecov Report All modified and coverable lines are covered by tests :white_check_mark: Project coverage is 55.45%. Comparing base (2961479) to head (d2b5672). Additional details and impacted files @@ Coverage Diff @@ ## main #35 +/- ## ======================================= Coverage 55.45% 55.45% ======================================= Files 12 12 Lines 1502 1502 ======================================= Hits 833 833 Misses 616 616 Partials 53 53 Files Coverage Δ ante/ante.go 62.50% <ø> (ø) ante/txfilter.go 85.93% <ø> (ø) app/ante.go 0.00% <ø> (ø) app/app.go 83.33% <ø> (ø) app/encoding.go 100.00% <ø> (ø) cmd/fnsad-proxy/cmd/root.go 34.20% <ø> (ø)
2025-04-01T04:55:05.114209
2022-02-21T08:13:19
1145457713
{ "authors": [ "EmmanyTaskWorld", "Fintasys", "tzthantzin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13946", "repo": "Fintasys/emoji_picker_flutter", "url": "https://github.com/Fintasys/emoji_picker_flutter/issues/57" }
gharchive/issue
How can I get unify code from emoji? How can I get unify code from emoji? I click 'grinning face', its unified code is 1F600. How can I get that value from emoji picker? https://unicode.org/emoji/charts-13.0/full-emoji-list.html Thank you very much A emoji consists of runes. So in dart the runes are from type Integer and you can easily format them to hex. EmojiPicker( onEmojiSelected: (Category category, Emoji emoji) { emoji.emoji.runes.forEach((element) { print(element.toRadixString(16)); }); } ... ) I hope that helps you ! Thank you very much @Fintasys . It is working Thank you very much 🙏 When i select country flag..using emoji.emoji.runes.forEach((element) { print(element.toRadixString(16)); // 1F600 }); doesn't seem to work as country flag emoji unicode should be in this form (U+1F1E6-1F1EE)
2025-04-01T04:55:05.146857
2021-01-22T09:47:59
791858601
{ "authors": [ "YannicEl", "dackers86", "mbleigh" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13947", "repo": "FirebaseExtended/experimental-extensions", "url": "https://github.com/FirebaseExtended/experimental-extensions/issues/31" }
gharchive/issue
[firestore-scheduled-writes] Error logs appear even with successful writes Expected Behavior Logging should be successful and not contain error logs when a successful rite has been made. Actual Behavior An error log appears, stating the document cannot be found. The extensions appear to have worked as expected. Steps to Reproduce the Problem Follow standard directions for installation in README. Run the example code snippet const TEN_MINUTES_MS = 10 * 60 * 1000; firebase .firestore() .collection("queued_writes") .add({ state: "PENDING", data: { message: "Hello from the future!" }, deliverTime: firebase.firestore.Timestamp.fromMillis( Date.now() + TEN_MINUTES_MS ) }); What happened? How can we make the problem occur?  Specifications Extension name: firestore-schedule-writes Version: 0.0.1 Is this being worked on? If not I could issue a PR. PR would be great! I didn't notice this issue until just now On Fri, Jun 4, 2021, 6:43 AM YannicEl @.***> wrote: Is this being worked on? If not I could issue a PR. — You are receiving this because you were assigned. Reply to this email directly, view it on GitHub https://github.com/FirebaseExtended/experimental-extensions/issues/31#issuecomment-854733896, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAAAH7S4CDQ365E2STOJ7E3TRDKANANCNFSM4WOHEFWA . the PR is ready
2025-04-01T04:55:05.156136
2020-03-26T18:51:13
588650509
{ "authors": [ "TahaTesser", "aantelov87" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13948", "repo": "FirebaseExtended/flutterfire", "url": "https://github.com/FirebaseExtended/flutterfire/issues/2242" }
gharchive/issue
[firebase_ml_vision] PDF417 reader is not working for DL Trying to read a PDF417 barcode for a US DL using the new barcode library instead of the one from Google Mobile Vision but your lib is not properly reading and camera is unusable for Samsung S8 when resolution is set to ResolutionPreset.high. Any ideas how can we use this feature from the new API? I am testing with the example but I am not getting any good results. Thank you Hi @aantelov87 can you please provide your flutter doctor -v and flutter run --verbose? Are you using official firebase_ml_vision example? Thank you Hi @TahaTesser Yes, I am using the official firebase_ml_vision example but I just changed the following line https://github.com/FirebaseExtended/flutterfire/blob/master/packages/firebase_ml_vision/example/lib/material_barcode_scanner.dart#L156 with final ResolutionPreset preset = ResolutionPreset.high; because with the resolution used in the official example, the detector is not able to detect PDF417 barcode for DL. In the attachment flutter run --verbose when the resolution is set to high for PDF417 detection. Please, let me know if you need anything else. `flutter doctor -v [✓] Flutter (Channel stable, v1.12.13+hotfix.8, on Mac OS X 10.15.3 19D76, locale en-US) • Flutter version 1.12.13+hotfix.8 at /Users/aantelov/GPN/flutter • Framework revision 0b8abb4724 (6 weeks ago), 2020-02-11 11:44:36 -0800 • Engine revision e1e6ced81d • Dart version 2.7.0 [✓] Android toolchain - develop for Android devices (Android SDK version 29.0.3) • Android SDK at /Users/aantelov/Library/Android/sdk • Android NDK location not configured (optional; useful for native profiling support) • Platform android-29, build-tools 29.0.3 • ANDROID_HOME = /Users/aantelov/Library/Android/sdk • Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java • Java version OpenJDK Runtime Environment (build 1.8.0_212-release-1586-b4-5784211) • All Android licenses accepted. [!] Xcode - develop for iOS and macOS ✗ Xcode installation is incomplete; a full installation is necessary for iOS development. Download at: https://developer.apple.com/xcode/download/ Or install Xcode via the App Store. Once installed, run: sudo xcode-select --switch /Applications/Xcode.app/Contents/Developer sudo xcodebuild -runFirstLaunch • CocoaPods version 1.9.0 [✓] Android Studio (version 3.6) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin version 44.0.2 • Dart plugin version 192.7761 • Java version OpenJDK Runtime Environment (build 1.8.0_212-release-1586-b4-5784211) [✓] Connected device (1 available) • SM G950F • ce06182664a6821703 • android-arm64 • Android 9 (API 28) ` output-modified.txt Hello @TahaTesser Have I added the issue in the correct project? because it looks like nobody is looking into this since 6 days ago.
2025-04-01T04:55:05.177719
2022-01-27T03:08:55
1115719920
{ "authors": [ "AAkira", "darshankawar" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13949", "repo": "FirebaseExtended/flutterfire", "url": "https://github.com/FirebaseExtended/flutterfire/issues/7956" }
gharchive/issue
🐛 [firebase_auth] Cannot handle quota-exceeded error on Android Bug report Describe the bug We use the phone auth. We've tried to authenticate multiple times with the spark(free) plan. We were able to handle the quota-exceeded error on iOS, but we only got the too-many-requests error on Android. However, when we changed to the blaze(paid) plan, we only got too-many-request errors on iOS as well. Steps to reproduce Send SMS multiple times I called FirebaseAuth.instance.verifyPhoneNumber Handle verificationFailed on Android and iOS Expected behavior Handle the quota-exceeded error on Android Flutter doctor Click To Expand Doctor summary (to see all details, run flutter doctor -v): [✓] Flutter (Channel stable, 2.8.0, on macOS 11.6.1 20G224 darwin-x64, locale ja-JP) [!] Android toolchain - develop for Android devices (Android SDK version 31.0.0) ✗ cmdline-tools component is missing Run `path/to/sdkmanager --install "cmdline-tools;latest"` See https://developer.android.com/studio/command-line for more details. ✗ Android license status unknown. Run `flutter doctor --android-licenses` to accept the SDK licenses. See https://flutter.dev/docs/get-started/install/macos#android-setup for more details. [✓] Xcode - develop for iOS and macOS (Xcode 13.2.1) [✓] Chrome - develop for the web [✓] Android Studio (version 2020.3) [✓] Android Studio (version 2020.3) [✓] IntelliJ IDEA Ultimate Edition (version 2021.2.3) [✓] IntelliJ IDEA Ultimate Edition (version 2021.2.3) [✓] IntelliJ IDEA Ultimate Edition (version 2021.1.2) [✓] Connected device (3 available) Flutter dependencies Click To Expand Dart SDK 2.15.0 Flutter SDK 2.8.0 m3pay 1.14.0+24 dependencies: - cached_network_image 3.1.0+1 [flutter flutter_cache_manager octo_image cached_network_image_platform_interface cached_network_image_web] - camera 0.9.4+6 [camera_platform_interface camera_web flutter pedantic quiver flutter_plugin_android_lifecycle] - collection 1.15.0 - convex_bottom_bar 3.0.0 [flutter] - device_info_plus 3.1.1 [flutter device_info_plus_platform_interface device_info_plus_linux device_info_plus_macos device_info_plus_web device_info_plus_windows] - dio 4.0.4 [http_parser path] - dotted_border 2.0.0+2 [flutter path_drawing] - email_validator 2.0.1 - firebase_analytics 8.3.4 [firebase_analytics_platform_interface firebase_analytics_web firebase_core flutter meta] - firebase_auth 3.2.0 [firebase_auth_platform_interface firebase_auth_web firebase_core firebase_core_platform_interface flutter meta] - firebase_core 1.10.0 [firebase_core_platform_interface firebase_core_web flutter meta] - firebase_crashlytics 2.3.0 [firebase_core firebase_core_platform_interface firebase_crashlytics_platform_interface flutter stack_trace] - firebase_dynamic_links 3.0.1 [firebase_core flutter] - firebase_messaging 11.1.0 [firebase_core firebase_core_platform_interface firebase_messaging_platform_interface firebase_messaging_web flutter meta] - flutter 0.0.0 [characters collection meta typed_data vector_math sky_engine] - flutter_hooks 0.18.0 [flutter] - flutter_inappwebview 5.3.2 [flutter] - flutter_linkify 5.0.2 [flutter linkify] - flutter_local_notifications 9.0.3 [clock flutter flutter_local_notifications_linux flutter_local_notifications_platform_interface timezone] - flutter_localizations 0.0.0 [flutter intl characters clock collection meta path typed_data vector_math] - flutter_speed_dial 4.6.6 [flutter] - flutter_stripe 2.0.2 [flutter stripe_android stripe_ios stripe_platform_interface] - flutter_svg 1.0.0 [flutter meta path_drawing vector_math xml] - freezed_annotation 1.0.0 [collection json_annotation meta] - geolocator 7.7.1 [flutter geolocator_platform_interface geolocator_android geolocator_apple geolocator_web] - git_info_plus 0.0.4 [flutter] - google_maps_flutter 2.1.0 [flutter flutter_plugin_android_lifecycle google_maps_flutter_platform_interface] - hooks_riverpod 1.0.0 [collection flutter flutter_hooks flutter_riverpod riverpod state_notifier] - image 3.0.8 [archive meta xml] - intl_utils 2.5.1 [analyzer archive args dart_style http intl path petitparser yaml] - json_annotation 4.3.0 [meta] - kana_kit 2.0.0 [equatable] - logger 1.1.0 - openapi 1.0.0 [dio built_value built_collection] - package_info_plus 1.3.0 [flutter package_info_plus_platform_interface package_info_plus_linux package_info_plus_macos package_info_plus_windows package_info_plus_web] - path_provider 2.0.6 [flutter path_provider_linux path_provider_macos path_provider_platform_interface path_provider_windows] - permission_handler 8.2.6 [flutter meta permission_handler_platform_interface] - phone_number 0.12.0+1 [flutter meta] - qr_code_scanner 0.6.1 [js flutter flutter_web_plugins] - shared_preferences 2.0.8 [flutter meta shared_preferences_linux shared_preferences_macos shared_preferences_platform_interface shared_preferences_web shared_preferences_windows] - shimmer 2.0.0 [flutter] - sms_autofill 2.2.0 [pin_input_text_field flutter] - time_machine 0.9.17 [meta collection] - transparent_image 2.0.0 - tuple 2.0.0 [quiver] - url_launcher 6.0.12 [flutter meta url_launcher_linux url_launcher_macos url_launcher_platform_interface url_launcher_web url_launcher_windows] dev dependencies: - build_runner 2.1.5 [args async analyzer build build_config build_daemon build_resolvers build_runner_core code_builder collection crypto dart_style frontend_server_client glob graphs http_multi_server io js logging meta mime package_config path pool pub_semver pubspec_parse shelf shelf_web_socket stack_trace stream_transform timing watcher web_socket_channel yaml] - flutter_launcher_icons 0.9.2 [args image path yaml] - flutter_test 0.0.0 [flutter test_api path fake_async clock stack_trace vector_math async boolean_selector characters charcode collection matcher meta source_span stream_channel string_scanner term_glyph typed_data] - freezed 1.0.0 [analyzer build build_config collection meta source_gen freezed_annotation] - json_serializable 6.0.1 [analyzer async build build_config collection json_annotation meta path pub_semver pubspec_parse source_gen source_helper] - mockito 5.0.16 [analyzer build code_builder collection dart_style matcher meta path source_gen test_api] dependency overrides: - freezed_annotation 1.0.0 [collection json_annotation meta] transitive dependencies: - _fe_analyzer_shared 30.0.0 [meta] - analyzer 2.7.0 [_fe_analyzer_shared cli_util collection convert crypto glob meta package_config path pub_semver source_span watcher yaml] - archive 3.1.2 [crypto path] - args 2.3.0 - async 2.8.2 [collection meta] - boolean_selector 2.1.0 [source_span string_scanner] - build 2.1.1 [analyzer async convert crypto glob logging meta path] - build_config 1.0.0 [checked_yaml json_annotation path pubspec_parse yaml] - build_daemon 3.0.0 [built_collection built_value http_multi_server logging pedantic path pool shelf shelf_web_socket stream_transform watcher web_socket_channel] - build_resolvers 2.0.4 [analyzer async build crypto graphs logging path package_config pool pub_semver stream_transform yaml] - build_runner_core 7.2.2 [async build build_config build_resolvers collection convert crypto glob graphs json_annotation logging meta path package_config pool timing watcher yaml] - built_collection 5.1.1 - built_value 8.1.2 [built_collection collection fixnum meta] - cached_network_image_platform_interface 1.0.0 [flutter flutter_cache_manager] - cached_network_image_web 1.0.1 [flutter flutter_cache_manager cached_network_image_platform_interface] - camera_platform_interface 2.1.1 [cross_file flutter meta plugin_platform_interface stream_transform] - camera_web 0.2.1+1 [camera_platform_interface flutter flutter_web_plugins stream_transform] - characters 1.2.0 - charcode 1.3.1 - checked_yaml 2.0.1 [json_annotation source_span yaml] - cli_util 0.3.3 [meta path] - clock 1.1.0 - code_builder 4.1.0 [built_collection built_value collection matcher meta] - convert 3.0.1 [typed_data] - cross_file 0.3.1+5 [flutter meta] - crypto 3.0.1 [collection typed_data] - dart_style 2.1.0 [analyzer args path pub_semver source_span] - dbus 0.5.6 [args ffi meta pedantic xml] - device_info_plus_linux 2.1.0 [device_info_plus_platform_interface file flutter meta] - device_info_plus_macos 2.2.0 [device_info_plus_platform_interface flutter] - device_info_plus_platform_interface 2.2.1 [flutter meta plugin_platform_interface] - device_info_plus_web 2.1.0 [device_info_plus_platform_interface flutter_web_plugins flutter] - device_info_plus_windows 2.1.0 [device_info_plus_platform_interface ffi flutter win32] - equatable 2.0.3 [collection meta] - fake_async 1.2.0 [clock collection] - ffi 1.1.2 - file 6.1.2 [meta path] - firebase 9.0.2 [http http_parser js] - firebase_analytics_platform_interface 2.0.1 [flutter meta] - firebase_analytics_web 0.3.0+1 [firebase firebase_analytics_platform_interface flutter flutter_web_plugins meta] - firebase_auth_platform_interface 6.1.4 [firebase_core flutter meta plugin_platform_interface] - firebase_auth_web 3.2.0 [firebase_auth_platform_interface firebase_core firebase_core_web flutter flutter_web_plugins http_parser intl js meta] - firebase_core_platform_interface 4.1.0 [collection flutter meta plugin_platform_interface] - firebase_core_web 1.2.0 [firebase_core_platform_interface flutter flutter_web_plugins js meta] - firebase_crashlytics_platform_interface 3.1.6 [collection firebase_core flutter meta plugin_platform_interface] - firebase_messaging_platform_interface 3.0.9 [firebase_core flutter meta plugin_platform_interface] - firebase_messaging_web 2.1.0 [firebase_core firebase_core_web firebase_messaging_platform_interface flutter flutter_web_plugins js meta] - fixnum 1.0.0 - flutter_blurhash 0.6.0 [flutter meta pedantic] - flutter_cache_manager 3.1.2 [clock collection file flutter http path path_provider pedantic rxdart sqflite uuid] - flutter_local_notifications_linux 0.3.0 [flutter flutter_local_notifications_platform_interface dbus path xdg_directories] - flutter_local_notifications_platform_interface 5.0.0 [flutter plugin_platform_interface] - flutter_plugin_android_lifecycle 2.0.3 [flutter] - flutter_riverpod 1.0.0 [collection flutter meta riverpod state_notifier] - flutter_web_plugins 0.0.0 [flutter js characters collection meta typed_data vector_math] - frontend_server_client 2.1.2 [async path] - geolocator_android 2.1.0 [flutter geolocator_platform_interface] - geolocator_apple 1.2.0 [flutter geolocator_platform_interface] - geolocator_platform_interface 2.3.5 [flutter plugin_platform_interface vector_math meta] - geolocator_web 2.0.6 [flutter flutter_web_plugins geolocator_platform_interface] - glob 2.0.1 [async collection file path pedantic string_scanner] - google_maps_flutter_platform_interface 2.1.3 [collection flutter meta plugin_platform_interface stream_transform] - graphs 2.1.0 [collection] - http 0.13.3 [async http_parser meta path pedantic] - http_multi_server 3.0.1 [async] - http_parser 4.0.0 [charcode collection source_span string_scanner typed_data] - intl 0.17.0 [clock path] - io 1.0.3 [meta path string_scanner] - js 0.6.3 - linkify 4.1.0 - logging 1.0.2 - matcher 0.12.11 [stack_trace] - meta 1.7.0 - mime 1.0.0 - octo_image 1.0.0+1 [flutter flutter_blurhash] - package_config 2.0.2 [path] - package_info_plus_linux 1.0.3 [package_info_plus_platform_interface flutter path] - package_info_plus_macos 1.3.0 [flutter] - package_info_plus_platform_interface 1.0.2 [flutter meta plugin_platform_interface] - package_info_plus_web 1.0.4 [flutter flutter_web_plugins http meta package_info_plus_platform_interface] - package_info_plus_windows 1.0.4 [package_info_plus_platform_interface ffi flutter win32] - path 1.8.0 - path_drawing 1.0.0 [vector_math meta path_parsing flutter] - path_parsing 1.0.0 [vector_math meta] - path_provider_linux 2.1.0 [flutter path path_provider_platform_interface xdg_directories] - path_provider_macos 2.0.2 [flutter] - path_provider_platform_interface 2.0.1 [flutter meta platform plugin_platform_interface] - path_provider_windows 2.0.3 [ffi flutter meta path path_provider_platform_interface win32] - pedantic 1.11.1 - permission_handler_platform_interface 3.7.0 [flutter meta plugin_platform_interface] - petitparser 4.3.0 [meta] - pin_input_text_field 4.1.1 [flutter] - platform 3.0.2 - plugin_platform_interface 2.0.1 [meta] - pool 1.5.0 [async stack_trace] - process 4.2.3 [file path platform] - pub_semver 2.1.0 [collection meta] - pubspec_parse 1.0.0 [checked_yaml collection json_annotation pub_semver yaml] - quiver 3.0.1 [matcher] - riverpod 1.0.0 [collection meta state_notifier] - rxdart 0.27.2 - shared_preferences_linux 2.0.2 [file meta flutter path path_provider_linux shared_preferences_platform_interface] - shared_preferences_macos 2.0.2 [flutter shared_preferences_platform_interface] - shared_preferences_platform_interface 2.0.0 [flutter] - shared_preferences_web 2.0.2 [flutter flutter_web_plugins meta shared_preferences_platform_interface] - shared_preferences_windows 2.0.2 [flutter file meta path path_provider_platform_interface path_provider_windows shared_preferences_platform_interface] - shelf 1.2.0 [async collection http_parser path stack_trace stream_channel] - shelf_web_socket 1.0.1 [shelf stream_channel web_socket_channel] - sky_engine 0.0.99 - source_gen 1.1.1 [analyzer async build dart_style glob meta path source_span yaml] - source_helper 1.3.0 [analyzer collection source_gen] - source_span 1.8.1 [collection path term_glyph] - sqflite 2.0.0+4 [flutter sqflite_common path] - sqflite_common 2.0.1+1 [synchronized path meta] - stack_trace 1.10.0 [path] - state_notifier 0.7.1 [meta] - stream_channel 2.1.0 [async] - stream_transform 2.0.0 - string_scanner 1.1.0 [charcode source_span] - stripe_android 2.0.2 [flutter] - stripe_ios 2.0.2 [flutter] - stripe_platform_interface 2.0.2 [flutter freezed_annotation json_annotation meta plugin_platform_interface] - synchronized 3.0.0 - term_glyph 1.2.0 - test_api 0.4.3 [async boolean_selector collection meta source_span stack_trace stream_channel string_scanner term_glyph matcher] - timezone 0.8.0 [path] - timing 1.0.0 [json_annotation] - typed_data 1.3.0 [collection] - url_launcher_linux 2.0.2 [flutter] - url_launcher_macos 2.0.2 [flutter] - url_launcher_platform_interface 2.0.4 [flutter plugin_platform_interface] - url_launcher_web 2.0.4 [flutter flutter_web_plugins meta url_launcher_platform_interface] - url_launcher_windows 2.0.2 [flutter] - uuid 3.0.4 [crypto] - vector_math 2.1.1 - watcher 1.0.0 [async path pedantic] - web_socket_channel 2.1.0 [async crypto stream_channel] - win32 2.2.9 [ffi] - xdg_directories 0.2.0 [meta path process] - xml 5.3.0 [collection meta petitparser] - yaml 3.1.0 [collection source_span string_scanner] @AAkira There's similar open issue describing your case, https://github.com/FirebaseExtended/flutterfire/issues/6687, so please follow-up in it for further updates. Closing this as a duplicate. Also take a look at this related open PR: https://github.com/FirebaseExtended/flutterfire/pull/3402
2025-04-01T04:55:05.187488
2020-03-09T03:55:22
577642954
{ "authors": [ "chinkysight", "kroikie" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13950", "repo": "FirebaseExtended/flutterfire", "url": "https://github.com/FirebaseExtended/flutterfire/pull/2148" }
gharchive/pull-request
[firebase_auth plugin] README.md Updated Description I had just updated the README.md file. There isn't any deletion of any pre-existed info in README.md Related Issues There isn't any issue created for this, but I thought it would be very helpful to add some examples of doing Firebase Facebook and Firebase Twitter authentication. Checklist Before you create this PR confirm that it meets all requirements listed below by checking the relevant checkboxes ([x]). This will ensure a smooth and quick review process. [x] I read the Contributor Guide and followed the process outlined there for submitting PRs. [x] If the pull request affects only one plugin, the PR title starts with the name of the plugin in brackets (e.g. [cloud_firestore]) [x] My PR includes unit or integration tests for all changed/updated/fixed behaviors (See Contributor Guide). [x] All existing and new tests are passing. [x] I updated/added relevant documentation (doc comments with ///). [x] The analyzer (flutter analyze) does not report any problems on my PR. [x] I read and followed the Flutter Style Guide. [x] I updated pubspec.yaml with an appropriate new version according to the pub versioning philosophy. [x] I updated CHANGELOG.md to add a description of the change. [x] I signed the CLA. [x] I am willing to follow-up on review comments in a timely manner. Breaking Change Does your PR require plugin users to manually update their apps to accommodate your change? [ ] Yes, this is a breaking change (please indicate a breaking change in CHANGELOG.md and increment major revision). [x] No, this is not a breaking change. @chinkysight Thanks for the PR, would you mind rebasing this so that the tests can pass and I'll be happy to review.
2025-04-01T04:55:05.223696
2020-01-12T21:12:22
548627683
{ "authors": [ "Fizzadar", "mkinney" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13951", "repo": "Fizzadar/pyinfra", "url": "https://github.com/Fizzadar/pyinfra/issues/229" }
gharchive/issue
bug in apt.repo()? I'm running into an issue using apt.repo(). Am I missing something? $ cat mike.py from pyinfra.modules import apt SUDO = True apt.repo( {'Install VirtualBox repo'}, 'deb https://download.virtualbox.org/virtualbox/debian bionic contrib', ) $ rm pyinfra-debug.log rm: pyinfra-debug.log: No such file or directory $ pyinfra -vv --debug @docker/ubuntu:bionic mike.py [pyinfra_cli.main] Checking potential directory: [pyinfra_cli.main] Setting directory to: --> Loading config... --> Loading inventory... [pyinfra_cli.inventory] Creating fake inventory... [pyinfra.api.inventory] Expanding inventory connector: docker The @docker connector is in Alpha! [pyinfra_cli.inventory] Looking for group data in: group_data/all.py [pyinfra_cli.inventory] Looking for group data in: group_data/bsd.py [pyinfra.api.inventory] Expanding inventory connector: docker --> Connecting to hosts... localhost: >>> docker run -d ubuntu:bionic sleep 10000 localhost: a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 [pyinfra.api.state] Activating host: ubuntu:bionic --> Preparing operations... Loading: mike.py [pyinfra.api.operation] Adding operation, {'Install VirtualBox repo'}, called @ mike.py:7, opLines=(0, 7), opHash=97d8032458c50f8c46e458a4ca15c67a69be92b3 [pyinfra.api.facts] Getting fact: apt_sources (ensure_hosts: (ubuntu:bionic,)) [pyinfra.api.util] Building command (shell_executable: sh): docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "cat /etc/apt/sources.list /etc/apt/sources.list.d/*.list || true" [pyinfra.api.connectors.local] --> Running command on localhost: sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "cat /etc/apt/sources.list /etc/apt/sources.list.d/*.list || true"' [ubuntu:bionic] >>> sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "cat /etc/apt/sources.list /etc/apt/sources.list.d/*.list || true"' [ubuntu:bionic] # See http://help.ubuntu.com/community/UpgradeNotes for how to upgrade to [ubuntu:bionic] # newer versions of the distribution. [ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic main restricted [ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic main restricted [ubuntu:bionic] [ubuntu:bionic] ## Major bug fix updates produced after the final release of the [ubuntu:bionic] ## distribution. [ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic-updates main restricted [ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic-updates main restricted [ubuntu:bionic] [ubuntu:bionic] ## N.B. software from this repository is ENTIRELY UNSUPPORTED by the Ubuntu [ubuntu:bionic] ## team. Also, please note that software in universe WILL NOT receive any [ubuntu:bionic] ## review or updates from the Ubuntu security team. [ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic universe [ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic universe [ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic-updates universe [ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic-updates universe [ubuntu:bionic] [ubuntu:bionic] ## N.B. software from this repository is ENTIRELY UNSUPPORTED by the Ubuntu [ubuntu:bionic] ## team, and may not be under a free licence. Please satisfy yourself as to [ubuntu:bionic] ## your rights to use the software. Also, please note that software in [ubuntu:bionic] ## multiverse WILL NOT receive any review or updates from the Ubuntu [ubuntu:bionic] ## security team. [ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic multiverse [ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic multiverse [ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic-updates multiverse [ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic-updates multiverse [ubuntu:bionic] [ubuntu:bionic] ## N.B. software from this repository may not have been tested as [ubuntu:bionic] ## extensively as that contained in the main release, although it includes [ubuntu:bionic] ## newer versions of some applications which may provide useful features. [ubuntu:bionic] ## Also, please note that software in backports WILL NOT receive any review [ubuntu:bionic] ## or updates from the Ubuntu security team. [ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic-backports main restricted universe multiverse [ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic-backports main restricted universe multiverse [ubuntu:bionic] [ubuntu:bionic] ## Uncomment the following two lines to add software from Canonical's [ubuntu:bionic] ## 'partner' repository. [ubuntu:bionic] ## This software is not part of Ubuntu, but is offered by Canonical and the [ubuntu:bionic] ## respective vendors as a service to Ubuntu users. [ubuntu:bionic] # deb http://archive.canonical.com/ubuntu bionic partner [ubuntu:bionic] # deb-src http://archive.canonical.com/ubuntu bionic partner [ubuntu:bionic] [ubuntu:bionic] deb http://security.ubuntu.com/ubuntu/ bionic-security main restricted [ubuntu:bionic] # deb-src http://security.ubuntu.com/ubuntu/ bionic-security main restricted [ubuntu:bionic] deb http://security.ubuntu.com/ubuntu/ bionic-security universe [ubuntu:bionic] # deb-src http://security.ubuntu.com/ubuntu/ bionic-security universe [ubuntu:bionic] deb http://security.ubuntu.com/ubuntu/ bionic-security multiverse [ubuntu:bionic] # deb-src http://security.ubuntu.com/ubuntu/ bionic-security multiverse [ubuntu:bionic] cat: '/etc/apt/sources.list.d/*.list': No such file or directory [pyinfra.api.connectors.local] --> Waiting for exit status... [pyinfra.api.connectors.local] --> Command exit status: 0 Loaded fact apt_sources [pyinfra.api.facts] Getting fact: find_in_file (ensure_hosts: (ubuntu:bionic,)) [pyinfra.api.util] Building command (shell_executable: sh): docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" /etc/apt/sources.list || (find /etc/apt/sources.list -type f > /dev/null && echo "__pyinfra_exists_/etc/apt/sources.list")" [pyinfra.api.connectors.local] --> Running command on localhost: sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" /etc/apt/sources.list || (find /etc/apt/sources.list -type f > /dev/null && echo "__pyinfra_exists_/etc/apt/sources.list")"' [ubuntu:bionic] >>> sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" /etc/apt/sources.list || (find /etc/apt/sources.list -type f > /dev/null && echo "__pyinfra_exists_/etc/apt/sources.list")"' | fact: find_in_file('/etc/apt/sources.list', '^.*deb https://download.virtualbox.org/virtualbox/debian bi / fact: find_in_file('/etc/apt/sources.list', '^.*deb https://download.virtualbox.org/virtualbox/debian bi - fact: find_in_file('/etc/apt/sources.list', '^.*deb https://download.virtualbox.org/virtualbox/debian bi \ fact: find_in_file('/etc/apt/sources.list', '^.*deb https://download.virtualbox.org/virtualbox/debian bi[pyinfra.api.connectors.local] --> Waiting for exit status... [pyinfra.api.connectors.local] --> Command exit status: 1 Loaded fact find_in_file: ('/etc/apt/sources.list', '^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$') [ubuntu:bionic] Ready: mike.py --> Proposed changes: Groups: @docker [ubuntu:bionic] Operations: 1 Commands: 1 --> Beginning operation run... --> Starting operation: Install VirtualBox repo [pyinfra.api.operations] Starting operation Install VirtualBox repo on ubuntu:bionic [pyinfra.api.util] Building command (shell_executable: sh): docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "# If the file now exists if [ -f "/etc/apt/sources.list" ]; then # Grep for the line, sed if matches (grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" "/etc/apt/sources.list" && sed -i="" "s/^.*deb https:\/\/download.virtualbox.org\/virtualbox\/debian bionic contrib.*$//" /etc/apt/sources.list) || # Else echo echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list # No file, just echo else echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list fi" [pyinfra.api.connectors.local] --> Running command on localhost: sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "# If the file now exists if [ -f "/etc/apt/sources.list" ]; then # Grep for the line, sed if matches (grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" "/etc/apt/sources.list" && sed -i="" "s/^.*deb https:\/\/download.virtualbox.org\/virtualbox\/debian bionic contrib.*$//" /etc/apt/sources.list) || # Else echo echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list # No file, just echo else echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list fi"' [ubuntu:bionic] >>> sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "# If the file now exists if [ -f "/etc/apt/sources.list" ]; then # Grep for the line, sed if matches (grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" "/etc/apt/sources.list" && sed -i="" "s/^.*deb https:\/\/download.virtualbox.org\/virtualbox\/debian bionic contrib.*$//" /etc/apt/sources.list) || # Else echo echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list # No file, just echo else echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list fi"' [ubuntu:bionic] https://download.virtualbox.org/virtualbox/debian: 4: https://download.virtualbox.org/virtualbox/debian: Syntax error: end of file unexpected (expecting ")") [pyinfra.api.connectors.local] --> Waiting for exit status... [pyinfra.api.connectors.local] --> Command exit status: 2 [ubuntu:bionic] Error [pyinfra.api.state] Failing hosts: ubuntu:bionic localhost: >>> docker commit a426aa3919c7 localhost: sha256:db4497e361e507180b287c769f7777969102d1ffd5d009a6772ccfb9a4c218a6 localhost: >>> docker rm -f a426aa3919c7 localhost: a426aa3919c7 [ubuntu:bionic] docker build complete, image ID: db4497e361e5 --> pyinfra error: No hosts remaining! $ ls pyinfra-debug.log ls: pyinfra-debug.log: No such file or directory $ Turns out this is an issue with Docker specifically - not quoting commands correctly; this should be resolved by https://github.com/Fizzadar/pyinfra/commit/0cc6433f6569d92596d5570423cba571208b7e0c.
2025-04-01T04:55:05.227372
2017-09-03T12:23:27
254871365
{ "authors": [ "Fizzadar", "tobald" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13952", "repo": "Fizzadar/pyinfra", "url": "https://github.com/Fizzadar/pyinfra/pull/116" }
gharchive/pull-request
Hackish: show "[sudo] password for user:" message when using @local A naïve fix for letting sudo inform the user it waits for his password. Only tested with @local. The following ticket details a more advanced take on the issue: https://github.com/Fizzadar/pyinfra/issues/40 @tobald sorry for never getting back to you on this one! I'm still not sure how or whether pyinfra should handle input. I'd like to do it as it makes the user experience far better; but the current plan in #40 is pretty involved and would make pyinfra lot more complicated :/ @tobald sorry for never getting back to you on this one! I'm still not sure how or whether pyinfra should handle input. I'd like to do it as it makes the user experience far better; but the current plan in #40 is pretty involved and would make pyinfra lot more complicated :/ No worries, I do not have a use for this feature anymore. Proper input management is the way to go as stated in #40, and seems complicated. Meanwhile we could detect/display something to the user, that was the purpose of this commit.
2025-04-01T04:55:05.236585
2024-10-27T09:09:14
2616438623
{ "authors": [ "VERT1DAN" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13953", "repo": "FlafyDev/auto-nong-indexes", "url": "https://github.com/FlafyDev/auto-nong-indexes/issues/5478" }
gharchive/issue
add F-777 - Sonic Blaster (DeCody remix) by DeCody Song Name F-777 - Sonic Blaster (DeCody remix) Artist Name DeCody Source Youtube Youtube Link 9QV_T53XoEA Direct File Link No response Song ID 574484 Start Offset [ms] 0 Extra Created from Auto Nong. Created for level: 26681070 Downloaded: Yes accept вс, 27 окт. 2024 г., 12:09 auto-nong[bot] @.***>: Thanks for contributing! Someone will soon review your request. Here is the JSON result for the reviewer: { "name": "F-777 - Sonic Blaster (DeCody remix)", "artist": "DeCody", "source": "youtube", "startOffset": 0, "yt-id": "9QV_T53XoEA", "songs": [ 574484 ] } To accept the request, a reviewer needs to write "accept". — Reply to this email directly, view it on GitHub https://github.com/FlafyDev/auto-nong-indexes/issues/5478#issuecomment-2439918167, or unsubscribe https://github.com/notifications/unsubscribe-auth/BMOTCG7SL5N6IC3ORUOC5GLZ5SUVVAVCNFSM6AAAAABQVUAT62VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDIMZZHEYTQMJWG4 . You are receiving this because you authored the thread.Message ID: @.***>
2025-04-01T04:55:05.245203
2023-05-31T16:28:42
1734589387
{ "authors": [ "khvn26", "matthewelwell" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13954", "repo": "Flagsmith/flagsmith", "url": "https://github.com/Flagsmith/flagsmith/issues/2256" }
gharchive/issue
Swagger is broken Describe the bug SwaggerGenerationError your query_serializer contains fields that conflict with the filter_backend or paginator_class on the view - GET /api/v1/audit/ It's likely the type mismatch between filter params and the DRF serializer: Filter: { description: '', in: 'query', name: 'is_system_event', required: False, > type: 'string' < }, Serializer field: { in: 'query', name: 'is_system_event', required: False, > type: 'boolean' < }, The error is currently only reproducible for user id 14119. Potentially related issue: https://github.com/axnsan12/drf-yasg/issues/514 To Reproduce Steps to reproduce the behavior: Login as user 14119. Go to https://api.flagsmith.com/api/v1/docs/?format=openapi Observe the HTTP 500 response. Expected behavior The server should render the OpenAPI schema. A unit test should catch Swagger generation errors. How are you running Flagsmith? [ ] Self Hosted with Docker [ ] Self Hosted with Kubernetes [x] SaaS at flagsmith.com [ ] Some other way I have looked into this a bit further - I don't believe it is related to the error provided in the description since that was fixed here: https://github.com/Flagsmith/flagsmith/pull/2326. It seems to only happen when users are logged in as a user that also has access to the django admin in the same browser session that they access the API docs. The issue is still reproducible in prod. Tried logging out and in again to no avail. Sorry, yes - it's still reproducible for sure, but I'm not convinced it's related to the exception in the description. It's definitely related to being logged in as a staff user though.
2025-04-01T04:55:05.256737
2019-08-07T15:26:13
477996991
{ "authors": [ "bootstraponline", "doodla" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13955", "repo": "Flank/flank", "url": "https://github.com/Flank/flank/issues/588" }
gharchive/issue
Realtime streaming of test results via firebase function Prototype how to get firebase function streaming data to Flank. https://github.com/firebase/functions-samples/tree/master/quickstarts/test-complete This might not be possible given the function will have to push ( post ) data to some place, and most CI systems don't allow incoming traffic. We're building a comprehensive build and test analytics platform that integrates with Flank. 😄 This is a research ticket that will help prove out those efforts. Data will likely be pushed to a Firebase database. For test analytics, we'll parse the final JUnit XML artifact as attached to the CI job. In this approach, the firebase function doesn't add value.
2025-04-01T04:55:05.293895
2021-06-19T15:26:58
925416032
{ "authors": [ "MarceloWatson", "flax-dev" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13956", "repo": "Flax-Network/flax-blockchain", "url": "https://github.com/Flax-Network/flax-blockchain/issues/2" }
gharchive/issue
API Helo ! Is there any API or something similar so I can try to develop a site with information about Flax? Hey, our website is https://flaxnetwork.org/. This is a fork of https://github.com/Chia-Network/chia-blockchain. You can see there for more info, but there is currently no user-facing API.
2025-04-01T04:55:05.307400
2022-07-18T13:53:59
1307998159
{ "authors": [ "dcconner", "drosen20" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13957", "repo": "FlexBE/flexbe_app", "url": "https://github.com/FlexBE/flexbe_app/issues/80" }
gharchive/issue
ROS2 OCS does not launch behavior I am experimenting with flexBE in ROS2 Galactic and I ran into an issue with launching a behavior from the OCS GUI to run on a remote machine. What I did: followed the Basic Tutorials to create a hello world behavior that waits a few seconds and then prints to the screen launched the behavior from the GUI after running ros2 launch flexbe_app flexbe_full.launch.py and observed the expected behavior Copied the tutorials_behaviors repo (created in the tutorials) to the computer on my robot and ran colcon build to set up the workspace On remote, launched the OCS with ros2 launch flexbe_app flexbe_ocs.launch.py. On robot, launched the behavior engine with ros2 launch flexbe_onboard behavior_onboard.launch.py. Results: The OCS was able to connect to the behavior engine. After seeing --- Behavior Engine ready! --- on the robot the OCS indicated --- Behavior Mirror ready! ---. But after clicking Start Execution I get an error Failed to retrieve behavior from library: [-] 787353483 Dropped behavior start request because preparation failed. Also if I switch the launch order (engine on robot first, then OCS on remote), when I click Start Execution the OCS hangs on the Waiting For Behavior screen with the message Start command has been sent. Now waiting for confirmation from onboard behavior... This is an annoyance and something we plan to look at this summer. Likely it will involve a script to copy behaviors from the install folder to the source. Suggestions are appreciated.
2025-04-01T04:55:05.317759
2018-01-04T13:05:07
285979756
{ "authors": [ "LululuSir", "raphaelschaad", "reactxp" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13959", "repo": "Flipboard/FLAnimatedImage", "url": "https://github.com/Flipboard/FLAnimatedImage/pull/198" }
gharchive/pull-request
If a GIF image has no Netscape 2.0 loop extension, it is meant to play once and then stop. If a GIF image has no Netscape 2.0 loop extension, it is meant to play once and then stop. Is there anyone else? Is there anyone else? yes @LululuSir Is this the same as #102 ? @LululuSir Is this the same as #102 ? @LululuSir Is this the same as #102 ? YES This example GIF from this PR comment shows no Netscape 2.0 loop extension (when inspected with > gifsicle --extension-info) and thus shows each frame exactly once and then stops in Chrome/Safari and in FLAnimatedImage. This seems to be the correct behavior. Note that the definition of FLAnimatedImage's loopCount follows ImageIO's definition. If you find a case where FLAnimatedImage behaves differently than Chrome/Safari, please provide the respective GIF file as an example. Then we can have a look at whether something would need to change in FLAnimatedImage.
2025-04-01T04:55:05.345958
2023-09-14T21:19:22
1897321542
{ "authors": [ "RyanDavies19", "sanguinariojoe" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13960", "repo": "FloatingArrayDesign/MoorDyn", "url": "https://github.com/FloatingArrayDesign/MoorDyn/issues/121" }
gharchive/issue
Local docs compiling @sanguinariojoe do you have any idea what could be causing this error in compiling the documentation locally: Patching output file 56/58 Patching output file 57/58 Patching output file 58/58 type lookup cache used 10518/65536 hits=122962 misses=10670 symbol lookup cache used 6206/65536 hits=83658 misses=6206 finished... /Users/rdavies/Library/Python/3.9/lib/python/site-packages/urllib3/__init__.py:34: NotOpenSSLWarning: urllib3 v2.0 only supports OpenSSL 1.1.1+, currently the 'ssl' module is compiled with 'LibreSSL 2.8.3'. See: https://github.com/urllib3/urllib3/issues/3020 warnings.warn( /Users/rdavies/Library/Python/3.9/lib/python/site-packages/breathe/project.py:116: RemovedInSphinx80Warning: Sphinx 8 will drop support for representing paths as strings. Use "pathlib.Path" or "os.fspath" instead. self._default_build_dir = os.path.dirname(app.doctreedir.rstrip(os.sep)) making output directory... done Exception occurred: File "/Applications/Xcode.app/Contents/Developer/Library/Frameworks/Python3.framework/Versions/3.9/lib/python3.9/re.py", line 210, in sub return _compile(pattern, flags).sub(repl, string, count) TypeError: expected string or bytes-like object The full traceback has been saved in /var/folders/46/4_pcs77x31j8w360rc7pfv6chrj_sl/T/sphinx-err-athjs7ze.log, if you want to report the issue to the developers. Please also report this if it was a user error, so that a better error message can be provided next time. A bug report can be filed in the tracker at <https://github.com/sphinx-doc/sphinx/issues>. Thanks! make[2]: *** [docs/sphinx/index.html] Error 2 make[1]: *** [docs/CMakeFiles/Sphinx.dir/all] Error 2 make: *** [all] Error 2 The traceback that was saved shows the following: # Platform: darwin; (macOS-13.5.2-x86_64-i386-64bit) # Sphinx version: 7.2.6 # Python version: 3.9.6 (CPython) # Docutils version: 0.18.1 # Jinja2 version: 3.1.2 # Pygments version: 2.15.1 # Last messages: # Loaded extensions: # Traceback: Traceback (most recent call last): File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/cmd/build.py", line 293, in build_main app = Sphinx(args.sourcedir, args.confdir, args.outputdir, File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/application.py", line 272, in __init__ self._init_builder() File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/application.py", line 342, in _init_builder self.builder.init() File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/builders/html/__init__.py", line 219, in init self.init_templates() File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/builders/html/__init__.py", line 270, in init_templates self.theme = theme_factory.create(themename) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py", line 227, in create self.load_extra_theme(name) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py", line 181, in load_extra_theme self.load_external_theme(name) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py", line 194, in load_external_theme theme_entry_points = entry_points(group='sphinx.html_themes') File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/__init__.py", line 933, in entry_points return EntryPoints(eps).select(**params) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/__init__.py", line 930, in <genexpr> eps = itertools.chain.from_iterable( File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/_itertools.py", line 16, in unique_everseen k = key(element) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/_py39compat.py", line 18, in normalized_name return dist._normalized_name File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/__init__.py", line 855, in _normalized_name or super()._normalized_name File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/__init__.py", line 473, in _normalized_name return Prepared.normalize(self.name) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/__init__.py", line 773, in normalize return re.sub(r"[-_.]+", "-", name).lower().replace('-', '_') File "/Applications/Xcode.app/Contents/Developer/Library/Frameworks/Python3.framework/Versions/3.9/lib/python3.9/re.py", line 210, in sub return _compile(pattern, flags).sub(repl, string, count) TypeError: expected string or bytes-like object Additionally early on in the process it says /bin/sh: sphinx-build: command not found To get it to compile I am required to modify line 75 of build/docs/CMakeFiles/Sphinx.dir/build.make to remove the quotes around the path to python that are autogenerated. This is confusing me because last week it was working fine and I was able to compile the docs locally without error. I havent changed any of the code other than .rst files, and made no modifications to my system as far as I can tell. I've tried uninstalling all the required packages and reinstalling them, as well as deleting the build folder and remaking it with cmake. Pretty eclectic error, but it seems to me that you do not have the theme installed: sphinx_rtd_theme As far as I remember it shall be installed with pip On Thu, 14 Sept 2023, 23:19 RyanDavies19, @.***> wrote: @sanguinariojoe https://github.com/sanguinariojoe do you have any idea what could be causing this error in compiling the documentation locally: `Patching output file 56/58 Patching output file 57/58 Patching output file 58/58 type lookup cache used 10518/65536 hits=122962 misses=10670 symbol lookup cache used 6206/65536 hits=83658 misses=6206 finished... /Users/rdavies/Library/Python/3.9/lib/python/site-packages/urllib3/init.py:34: NotOpenSSLWarning: urllib3 v2.0 only supports OpenSSL 1.1.1+, currently the 'ssl' module is compiled with 'LibreSSL 2.8.3'. See: urllib3/urllib3#3020 https://github.com/urllib3/urllib3/issues/3020 warnings.warn( /Users/rdavies/Library/Python/3.9/lib/python/site-packages/breathe/project.py:116: RemovedInSphinx80Warning: Sphinx 8 will drop support for representing paths as strings. Use "pathlib.Path" or "os.fspath" instead. self._default_build_dir = os.path.dirname(app.doctreedir.rstrip(os.sep)) making output directory... done Exception occurred: File "/Applications/Xcode.app/Contents/Developer/Library/Frameworks/Python3.framework/Versions/3.9/lib/python3.9/re.py", line 210, in sub return _compile(pattern, flags).sub(repl, string, count) TypeError: expected string or bytes-like object The full traceback has been saved in /var/folders/46/4_pcs77x31j8w360rc7pfv6chrj_sl/T/sphinx-err-athjs7ze.log, if you want to report the issue to the developers. Please also report this if it was a user error, so that a better error message can be provided next time. A bug report can be filed in the tracker at https://github.com/sphinx-doc/sphinx/issues. Thanks! make[2]: *** [docs/sphinx/index.html] Error 2 make[1]: *** [docs/CMakeFiles/Sphinx.dir/all] Error 2 make: *** [all] Error 2` The traceback that was saved shows the following: `# Platform: darwin; (macOS-13.5.2-x86_64-i386-64bit) Sphinx version: 7.2.6 Python version: 3.9.6 (CPython) Docutils version: 0.18.1 Jinja2 version: 3.1.2 Pygments version: 2.15.1 Last messages: Loaded extensions: Traceback: Traceback (most recent call last): File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/cmd/build.py", line 293, in build_main app = Sphinx(args.sourcedir, args.confdir, args.outputdir, File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/application.py", line 272, in init self._init_builder() File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/application.py", line 342, in _init_builder self.builder.init() File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/builders/html/ init.py", line 219, in init self.init_templates() File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/builders/html/ init.py", line 270, in init_templates self.theme = theme_factory.create(themename) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py", line 227, in create self.load_extra_theme(name) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py", line 181, in load_extra_theme self.load_external_theme(name) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py", line 194, in load_external_theme theme_entry_points = entry_points(group='sphinx.html_themes') File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/ init.py", line 933, in entry_points return EntryPoints(eps).select(**params) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/ init.py", line 930, in eps = itertools.chain.from_iterable( File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/_itertools.py", line 16, in unique_everseen k = key(element) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/_py39compat.py", line 18, in normalized_name return dist._normalized_name File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/ init.py", line 855, in _normalized_name or super(). normalized_name File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/init.py", line 473, in normalized_name return Prepared.normalize(self.name http://self.name) File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/init.py", line 773, in normalize return re.sub(r"[-.]+", "-", name).lower().replace('-', '') File "/Applications/Xcode.app/Contents/Developer/Library/Frameworks/Python3.framework/Versions/3.9/lib/python3.9/re.py", line 210, in sub return _compile(pattern, flags).sub(repl, string, count) TypeError: expected string or bytes-like object` — Reply to this email directly, view it on GitHub https://github.com/FloatingArrayDesign/MoorDyn/issues/121, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAMXKKES3RWPYV5DIBTM6BDX2NYGLANCNFSM6AAAAAA4Y4B574 . You are receiving this because you were mentioned.Message ID: @.***> @sanguinariojoe thanks, I was curious if you had ideas. Turns out when I installed Xcode with admin rights for something else, it created a second python copy. I had to go in and delete all that but things are working well now.
2025-04-01T04:55:05.394825
2024-02-01T11:53:40
2112327493
{ "authors": [ "elenaviter", "hardillb" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13961", "repo": "FlowFuse/nr-launcher", "url": "https://github.com/FlowFuse/nr-launcher/issues/206" }
gharchive/issue
Enhancing Node-RED Logging: Introducing Host Output Logging Capability Description At present, the logging mechanism for Node-RED projects directs all logs to the 'driver', which can be accessed through the FlowFuse Project logs interface. In environments where Kubernetes clusters are deployed and are integrated with monitoring solutions (such as log forwarding to OTEL/ FluentBit for pods, and then to traces/metrics/etc. to backends like Prometheus/Jaeger/, etc.), there is a valuable opportunity to streamline monitoring practices. Enabling Node-RED projects to output their logs directly to the pod's console will facilitate a more unified approach to monitoring by leveraging existing log aggregation configurations. https://github.com/FlowFuse/flowfuse/issues/3324 Thank you @hardillb - closing this one
2025-04-01T04:55:05.405227
2023-08-25T15:39:32
1867268042
{ "authors": [ "amzamani", "chungyau97", "isikhi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13962", "repo": "FlowiseAI/Flowise", "url": "https://github.com/FlowiseAI/Flowise/issues/832" }
gharchive/issue
[BUG]yarn start not working Describe the bug Cloned the repo and followed the steps using yarn To Reproduce cloned and changed directory then=> yarn install && yarn build && yarn start install and build work fine with no error when start is executed, error is encountered Error ` Tasks: 3 successful, 3 total Cached: 0 cached, 3 total Time: 2m25.321s ✨ Done in 146.13s. yarn run v1.22.19 $ run-script-os $ cd packages/server/bin && ./run start (node:1212) [ERR_DLOPEN_FAILED] Error Plugin: flowise: dlopen(/Users/amzamani/Desktop/incoming/ai/Flowise/node_modules/canvas/build/Release/canvas.node, 1): Library not loaded: @loader_path/libpixman-<IP_ADDRESS>.dylib Referenced from: /Users/amzamani/Desktop/incoming/ai/Flowise/node_modules/canvas/build/Release/canvas.node Reason: no suitable image found. Did find: /Users/amzamani/Desktop/incoming/ai/Flowise/node_modules/canvas/build/Release/libpixman-<IP_ADDRESS>.dylib: cannot load 'libpixman-<IP_ADDRESS>.dylib' (load command 0x80000034 is unknown) /Users/amzamani/Desktop/incoming/ai/Flowise/node_modules/canvas/build/Release/libpixman-<IP_ADDRESS>.dylib: cannot load 'libpixman-<IP_ADDRESS>.dylib' (load command 0x80000034 is unknown) module<EMAIL_ADDRESS>task: toCached plugin: flowise root: /Users/amzamani/Desktop/incoming/ai/Flowise/packages/server See more details with DEBUG=* (Use node --trace-warnings ... to show where the warning was created) › Error: command start not found error Command failed with exit code 2. info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command. error Command failed with exit code 2. info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command. Abus-MacBook-Air:Flowise amzamani$ ` MacOS Mojave- Version 10.14.4 Node- Version v18.17.1 Hi @amzamani. I believe your issue is related https://github.com/FlowiseAI/Flowise/issues/556 but do try sudo yarn from install to start I guess it is related with node-canvas. I did not faced this problem but i may suggest to this path to solve it: Take a look: https://github.com/Automattic/node-canvas#compiling Install them: brew install pkg-config cairo pango libpng jpeg giflib librsvg pixman then; remove node_modules (remove build artifacts) yarn install yarn build yarn dev if it wont succeed i can also suggest to install this npm install canvas --build-from-source then reproduce steps starting with remove node_modules
2025-04-01T04:55:05.410262
2023-10-15T21:09:55
1944058658
{ "authors": [ "HenryHengZJ", "matthias" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13963", "repo": "FlowiseAI/Flowise", "url": "https://github.com/FlowiseAI/Flowise/pull/1066" }
gharchive/pull-request
Feature/PluginSystem In this PR I propose a Plugin & hook system for Flowise Plugins are packages, that currently live in /plugins - but they should also be installable via npm It borrows ideas from the WordPress plugin and hook system. One of the most powerful forces why WordPress became such a popular OpenSource framework (despite a lot of not so likable things about WordPress code) Within a plugin package we use package.json/main to point to the plugins main class which extends the FlowisePlugin class It automatically loads components and credentials from the plugins directory by specifying nodesPath and credentialsPath => Check the components in the UI within a new "Custom" section Under the hood it's using an event based hook system (based on EventEmitter) that adds the ability to collect and manipulate (filter) data, via emitting and listening to events (we call them hooks) p.s. Some changes where introduced by yarn lint-fix p.p.s. Sorry for the first broken commit/pr - pls ignore / delete hey @matthias thanks a lot for the initiative! Am wondering whats the difference between plugin vs users creating their own node in packages/components/nodes? hey @HenryHengZJ - thanks for looking into this. I think it's the much cleaner separation than working directly in the same folder / code structure as the core project. I thought about it since I'm working with Flowise. Coming from WordPress I was wondering why it's not more common to have a plugin/hook architecture in Node/JS projects. Then I investigated deeper into langchain recently - and discovered that they provide a mechanism to have custom nodes in a separate folder - which I liked very much. I elaborated this idea, because I could imagine that from a certain point you don't want to have 100th of nodes in the core / UI - but rather be able to install additional "packages" of functionality (not just nodes) Currently it needs a very deep dive into the code until you know where to add what - and there is the constant need of updating and merging code bases. Adding own features by forking is possible, but I would rather not do it in the long run as it constantly requires to keep up with all the codebase (as you never know what could change) - and solve merging conflicts. Actually adding / loading nodes from a plugin was just a demo use case of the hooks. They (hooks) are the important part, as they allow Flowise to add clear entry points how to add things. This could also be additional navigation points or more down the road. Imagine other use cases, as adding / managing / observing sources. Adding alternative chat UIs. Adding api endpoints (routes). Could be interesting for custom(er) projects based on Flowise. But could also be a mechanism to develop new features, before they become part of the core.
2025-04-01T04:55:05.546400
2023-05-13T00:07:16
1708355584
{ "authors": [ "corneliusyaovi", "pastramahodu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13964", "repo": "Flutterwave/Angular", "url": "https://github.com/Flutterwave/Angular/pull/21" }
gharchive/pull-request
Bug bounty test - PLEASE DO NOT APPROVE. ijoech Bug bounty test - PLEASE DO NOT APPROVE. ijoech Spam pull request
2025-04-01T04:55:05.560494
2023-04-13T18:11:50
1666905740
{ "authors": [ "gouttegd" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13965", "repo": "FlyBase/drosophila-anatomy-developmental-ontology", "url": "https://github.com/FlyBase/drosophila-anatomy-developmental-ontology/issues/1621" }
gharchive/issue
Obsolete neuroepidermoblast neuroepidermoblast is a subclass of neuroblast. The term has no definition and is not used anywhere within FBbt. In FlyBase, it is used in one record (P{E(spl)m8-HLH-2.61}). The corresponding paper never uses the term neuroepidermoblast. It seems the term is intended to refer to cells of the proneural clusters of the neurectoderm – cells that have not yet committed to either the neuroblast or the epidermoblast fate. The term does not seem used in the literature at all. PubMed returns zero result; Google returns a handful of results, which are all ultimately linked either to the FBbt term directly or to the equivalent term in CL (which is almost certainly a direct “copy” of the FBbt term). I think the term should be obsoleted. In the one record where it has been used, it can probably be replaced by proneural cluster. The term could also get a proper definition as “cell of a proneural cluster that will become either a neuroblast or an epidermoblast” or similar, rather than being obsoleted.
2025-04-01T04:55:05.637996
2017-07-12T18:41:56
242476859
{ "authors": [ "tas50" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13968", "repo": "Foodcritic/foodcritic", "url": "https://github.com/Foodcritic/foodcritic/pull/653" }
gharchive/pull-request
FC001 should not alert on node.run_state[:foo] Signed-off-by: Tim Smith<EMAIL_ADDRESS> Resolves https://github.com/Foodcritic/foodcritic/issues/652
2025-04-01T04:55:05.810401
2024-12-24T07:59:21
2757371335
{ "authors": [ "JeyesHan", "JinYue2015" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13969", "repo": "FoundationVision/Infinity", "url": "https://github.com/FoundationVision/Infinity/issues/10" }
gharchive/issue
What does this line work for? Thanks. https://github.com/FoundationVision/Infinity/blob/0ac8556fa2cc23c70c342785e53145c5d6b09d46/infinity/models/infinity.py#L433 This line is to quick jump to the implementation of [SelfAttnBlock.forward, CrossAttnBlock.forward]. It's for reference only and no other special usuage.
2025-04-01T04:55:05.817067
2024-08-07T03:20:39
2452193793
{ "authors": [ "Lucky-Light-Sun", "iFighting" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13970", "repo": "FoundationVision/VAR", "url": "https://github.com/FoundationVision/VAR/issues/80" }
gharchive/issue
Question about the training dataset - tokenzier Hi, I recently read your paper “Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction” published in CVPR 2024. I am very interested in your method and have great respect for your contribution to the academic community. We are currently doing some experiments and would like to ask you a question about the tokenizer part. Is the pre-training dataset you are using OpenImage V6 or another version? Current we are using OpenImage v6 to train tokenizer. So do you think it's fair to compare with VAR tokenizer? Thank you again for your time and effort, and I look forward to your reply. Hi, I recently read your paper “Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction” published in CVPR 2024. I am very interested in your method and have great respect for your contribution to the academic community. We are currently doing some experiments and would like to ask you a question about the tokenizer part. Is the pre-training dataset you are using OpenImage V6 or another version? Current we are using OpenImage v6 to train tokenizer. So do you think it's fair to compare with VAR tokenizer? Thank you again for your time and effort, and I look forward to your reply. Hi, I recently read your paper “Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction” published in CVPR 2024. I am very interested in your method and have great respect for your contribution to the academic community. We are currently doing some experiments and would like to ask you a question about the tokenizer part. Is the pre-training dataset you are using OpenImage V6 or another version? Current we are using OpenImage v6 to train tokenizer. So do you think it's fair to compare with VAR tokenizer? Thank you again for your time and effort, and I look forward to your reply. @Lucky-Light-Sun Sorry for late reply. The pre-training dataset we are using is OpenImage V6. BTW, VAR is accepted as NeurIPS oral, not CVPR
2025-04-01T04:55:05.861271
2022-07-05T16:49:48
1294536571
{ "authors": [ "Pedrome203", "andyowli" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13971", "repo": "FranckFreiburger/vue-pdf", "url": "https://github.com/FranckFreiburger/vue-pdf/issues/375" }
gharchive/issue
Call the browser printing function, and a blank page will appear in PDF. When invoking the browser print function, garbled code will appear at first, and then it will be modified according to pull requests #130, and the text will be displayed normally, but there will be an extra blank page between pages. Does anyone encounter the same problem? How to solve it? Hi, were you able to solve it?
2025-04-01T04:55:05.866425
2019-01-28T02:42:30
403630238
{ "authors": [ "FrangSierra", "KryptKode", "abdul-hasib", "gigiojr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13972", "repo": "FrangSierra/RxFirebase", "url": "https://github.com/FrangSierra/RxFirebase/issues/76" }
gharchive/issue
subscribe when query is empty I'm try to get an user profile with multiples nodes of a realtime database with this code: public static void getUserProfile(String storeId, String userId, final Callback callback){ DatabaseReference userRef = RealtimeDatabase.getUserRef(userId); DatabaseReference addressRef = RealtimeDatabase.getUserAddressRef(userId); DatabaseReference contactRef = RealtimeDatabase.getUserContactRef(userId); DatabaseReference storeUserRef = RealtimeDatabase.getStoreUserRef(storeId, userId); DatabaseReference storeNameRef = RealtimeDatabase.getStoreRef(storeId).child(Store.PARAM_NAME); RxFirebaseDatabase.observeSingleValueEvent(userRef, User.class) .zipWith(RxFirebaseDatabase.observeSingleValueEvent(addressRef, Address.class), zipUserAddress) .zipWith(RxFirebaseDatabase.observeSingleValueEvent(contactRef, UserContact.class), zipUserContact) .zipWith(RxFirebaseDatabase.observeSingleValueEvent(storeUserRef, StoreUser.class), zipUserStore) .zipWith(RxFirebaseDatabase.observeSingleValueEvent(storeNameRef, String.class), zipUserStoreName) .subscribe(new Consumer<User>() { @Override public void accept(User user) { callback.onSuccess(user); } }, new Consumer<Throwable>() { @Override public void accept(Throwable throwable) { throwable.printStackTrace(); callback.onError(throwable); } }); } This code works well to a complete profile, but when profile does not have address node or contact node, for example, the function in subscribe is never called and this is a problem for me now. Can you help me? I am into the same issue, is there any update on this error? I have created 3 flowable objects and then I called .zip to get the output of these 3 flowables. The issue is that it works well if all the 3 flowable returns value, if any of the query fails to find documents based on query .zip gets stuck and it never calls onNext or onComplete @abdul-hasib Use Flowable.merge instead of Flowable.zip. The zip operator waits for all the observables to emit before it triggers onNext @KryptKode is right. Merge instead of Zip should fix it. Sorry for the delay on the response.
2025-04-01T04:55:05.868964
2021-11-24T23:53:06
1063006629
{ "authors": [ "Frankkkkk", "demoze" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13973", "repo": "Frankkkkk/python-pylontech", "url": "https://github.com/Frankkkkk/python-pylontech/issues/8" }
gharchive/issue
wonderfull, wonderfull, Exactly what I'am looking for to translate JK-BMS rs485 protocole (or other BMS) to pylontech rs485 protocole , and then have my voltronic/mppsolar inverter talk to my DIY battery. Thank you. Thanks @demoze ! 😀
2025-04-01T04:55:05.889498
2024-12-17T00:14:48
2743659541
{ "authors": [ "sitapriyamoorthi", "tefirman" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13974", "repo": "FredHutch/wdl-unit-tests", "url": "https://github.com/FredHutch/wdl-unit-tests/pull/24" }
gharchive/pull-request
Adding contribution guidelines and code of conduct Adding Unit test workflow to contribution guide Moved contribution guidelines from README to .github/CONTRIBUTING.md.
2025-04-01T04:55:05.898108
2024-03-28T22:14:00
2214256068
{ "authors": [ "Fredx87", "anulman" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13975", "repo": "Fredx87/openapi-io-ts", "url": "https://github.com/Fredx87/openapi-io-ts/pull/27" }
gharchive/pull-request
Fix request body file imports Replace custom imports with the generic getImports() when writing request body files. I noticed that anything that referenced the DateFromISOString type was breaking. Thank you @anulman for the recent PRs! Unfortunately, I am not planning to maintain this library anymore, since I stopped using io-ts and fp-ts. I would like to rewrite the library for effectand@effect/schema`, but it is not going to happen soon. That's all right! We've been maintaining our own fork at nmbrco/openapi-io-ts; was mostly trying to contrib back to your work in appreciation :) Tbqh we were on zod before, but needed more out of generation than their ecosystem supported. Your work in comparison was much easier to hack on, even with the io-ts learning curve. Thanks for publishing :)
2025-04-01T04:55:05.968328
2016-11-13T16:32:58
188979378
{ "authors": [ "Manish-Giri", "bstonesifer" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13976", "repo": "FreeCodeCamp/FreeCodeCamp", "url": "https://github.com/FreeCodeCamp/FreeCodeCamp/issues/11643" }
gharchive/issue
help please Challenge Name Issue Description Browser Information cant get the my test to pass. i have compared my code to others and ask for help in chat.. Browser Name, Version: firefox newest... At the very least, you haven't applied the smaller-image class to the img element, which is why your image isn't resizing. Please visit the Help Room if you need assistance on this. been to the help room.. its not helpful.. @bstonesifer Feel free to PM me on gitter if you're still stuck.
2025-04-01T04:55:05.970656
2015-08-18T22:37:29
101764302
{ "authors": [ "AryanJ-NYC", "jmichb" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13977", "repo": "FreeCodeCamp/FreeCodeCamp", "url": "https://github.com/FreeCodeCamp/FreeCodeCamp/issues/2128" }
gharchive/issue
Error seems to be in error Challenge http://www.freecodecamp.com/challenges/waypoint-make-instances-of-objects-with-a-constructor-function has an issue. Please describe how to reproduce it, and include links to screenshots if possible. To reproduce it, run the correct code. Then you'll see that the error reported seems to be, well, in error: assert(typeof(myCar.engines) === 'number', 'The property engine of myCar should be a number');The property engine of myCar should be a number See the screenshot. engines, not engine. You used engine in your code.
2025-04-01T04:55:05.974472
2015-11-22T07:48:02
118249907
{ "authors": [ "3laa-Khalil-Alarabi", "raisedadead" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13978", "repo": "FreeCodeCamp/FreeCodeCamp", "url": "https://github.com/FreeCodeCamp/FreeCodeCamp/issues/4631" }
gharchive/issue
Arabic translation of Basic Algorithm Scripting [ ] Meet Bonfire [ ] Reverse a String [ ] Factorialize a Number [ ] Check for Palindromes [ ] Find the Longest Word in a String [ ] Title Case a Sentence [ ] Return Largest Numbers in Arrays [ ] Confirm the Ending [ ] Repeat a string repeat a string [ ] Truncate a string [ ] Chunky Monkey [ ] Slasher Flick [ ] Mutations [ ] Falsy Bouncer [ ] Seek and Destroy [ ] Where do I belong Closing in favor of https://github.com/FreeCodeCamp/TranslationExpansion/issues/24
2025-04-01T04:55:05.978810
2016-01-11T03:03:11
125865182
{ "authors": [ "JBuji", "KeaDC", "sludge256" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13979", "repo": "FreeCodeCamp/FreeCodeCamp", "url": "https://github.com/FreeCodeCamp/FreeCodeCamp/issues/6046" }
gharchive/issue
Waypoint: Add Placeholder Text to a Text Field ( bug in placeholder exercise) has an issue. User Agent is: Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.106 Safari/537.36. Please describe how to reproduce this issue, and include links to screenshots if possible. My code: <link href="http://fonts.googleapis.com/css?family=Lobster" rel="stylesheet" type="text/css"> <style> .red-text { color: red; } h2 { font-family: Lobster, Monospace; } p { font-size: 16px; font-family: Monospace; } .thick-green-border { border-color: green; border-width: 10px; border-style: solid; border-radius: 50%; } .smaller-image { width: 100px; } </style> <h2 class="red-text">CatPhotoApp</h2> <p>Click here for <a href="#">cat photos</a>.</p> <a href="#"><img class="smaller-image thick-green-border" alt="A cute orange cat lying on its back" src="https://bit.ly/fcc-relaxing-cat"></a> <p>Things cats love:</p> <ul> <li>cat nip</li> <li>laser pointers</li> <li>lasagna</li> </ul> <p>Top 3 things cats hate:</p> <ol> <li>flea treatment</li> <li>thunder</li> <li>other cats</li> </ol> <input type="text"> <input type="text" placeholder="cat photo URL"> (the above line is reading as incorrect ) Name of exercise: Waypoint: Add Placeholder Text to a Text Field You're supposed to add the placeholder text to the existing input not create a new input. I passd "Add Placeholder Text to a Text Field" challange puting TWO lines: But it's strange. I passed "Add Placeholder Text to a Text Field" challenge putting TWO lines: <input type="text" placeholder="cat photo URL"> <input type="text" placeholder="cat photo URL"> But it's strange. /FCC bug/ Sorry I'm new on FCC & I've lost this lines in my earlier Comment
2025-04-01T04:55:05.980969
2015-06-05T00:15:10
85341484
{ "authors": [ "MightyBoosh92", "QuincyLarson" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13980", "repo": "FreeCodeCamp/freecodecamp", "url": "https://github.com/FreeCodeCamp/freecodecamp/issues/841" }
gharchive/issue
Completed Challenge without creating "CSS blue" class Challenge http://www.freecodecamp.com/challenges/waypoint-override-styles-with-important has an issue. I was able to complete this challenge without doing the "Create the CSS class "blue-text"" step. Thank you for your detailed explanation and screen shots. One of our campers has submitted a fix for this.
2025-04-01T04:55:05.982289
2016-12-22T11:51:07
197158096
{ "authors": [ "QuincyLarson", "raisedadead" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13981", "repo": "FreeCodeCamp/open2017", "url": "https://github.com/FreeCodeCamp/open2017/pull/14" }
gharchive/pull-request
feat(guests) : Add links to social profiles Adds a social link from the twitter handles and falls back to Github handle. P.S Demo available at https://raisedadead.com/open2017/ @raisedadead awesome! Well done!
2025-04-01T04:55:06.048345
2020-04-10T22:13:59
598117701
{ "authors": [ "jacobopantoja" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13982", "repo": "FreeRDP/FreeRDP", "url": "https://github.com/FreeRDP/FreeRDP/issues/6058" }
gharchive/issue
USB redirection not working in non-Windows platforms #It seems that commit ebdaea6f3b5e410b9a3380368d6023c52bed67cc introduces a change targeted to Windows builds, without effectively checking it, therefore breaking the USB redirection functionality in non-Windows builds. When loading the urbdrc channel in Linux build, you receive the following error: [00:01:01:900] [2961:2962] [WARN][com.freerdp.channels.urbdrc.client] - LIBUSB_OPTION_USE_USBDK Operation not supported or unimplemented on this platform [-12] Simply adding the adequate ifdef fixes the problem Pull request issued #6057 Fixed in 0f8437d1d86b2798edd46033974eb9678b6eecf0. Closing
2025-04-01T04:55:06.150384
2024-09-10T15:18:28
2516763491
{ "authors": [ "Frezyx", "Luwirispok" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13983", "repo": "Frezyx/talker", "url": "https://github.com/Frezyx/talker/issues/260" }
gharchive/issue
TalkerDataCard does not expand when changing title Describe the bug If you change the title in TalkerSettings in TalkerFlutter.init, TalkerDataCard stops responding to keystrokes, and responds to settings. To Reproduce Steps to reproduce the behavior: Connect talker_flutter, talker_dio_logger packages Create TalkerFlutter.init In TalkerSettings, change the titles to custom headers (at least network requests) Add TalkerDioLogger to dio.interceptors Add TalkerScreen call to TalkerSettings Make the request in the working application Click on the network request logs in TalkerScreen Error. TalkerDataCard with custom title has not changed Expected behavior Ability to expand TalkerDataCard and see detailed request information Screenshots https://github.com/user-attachments/assets/f20941ca-79f9-4c53-97e9-034672f550f1 Desktop (please complete the following information): Smartphone (please complete the following information): Device: Xiaomi MI 9T. OS: Android 11. Device: Blackview Active 8 Pro. OS: Android 13. Additional context Spotted in log requests, responses, and network errors because it shows a lot of data. Probably an issue with any logs with detailed data Package version talker_dio_logger: ^4.4.1 talker_bloc_logger: ^4.4.1 talker_flutter: ^4.4.1 [✓] Flutter (Channel stable, 3.24.1, on macOS 14.5 23F79 darwin-arm64, locale ru-RU) • Flutter version 3.24.1 on channel stable at /Users/specialny/tools/flutter • Upstream repository https://github.com/flutter/flutter.git • Framework revision 5874a72aa4 (3 недели назад), 2024-08-20 16:46:00 -0500 • Engine revision c9b9d5780d • Dart version 3.5.1 • DevTools version 2.37.2 [✓] Android toolchain - develop for Android devices (Android SDK version 34.0.0) • Android SDK at /Users/specialny/Library/Android/sdk • Platform android-34, build-tools 34.0.0 • Java binary at: /Applications/Android Studio.app/Contents/jbr/Contents/Home/bin/java • Java version OpenJDK Runtime Environment (build 17.0.11+0-17.0.11b1207.24-11852314) • All Android licenses accepted. [✓] Xcode - develop for iOS and macOS (Xcode 15.4) • Xcode at /Applications/Xcode.app/Contents/Developer • Build 15F31d • CocoaPods version 1.14.3 [✓] Chrome - develop for the web • Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome [✓] Android Studio (version 2024.1) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/6351-dart • Java version OpenJDK Runtime Environment (build 17.0.11+0-17.0.11b1207.24-11852314) [✓] VS Code (version 1.92.0) • VS Code at /Applications/Visual Studio Code.app/Contents • Flutter extension version 3.94.0 [✓] Connected device (4 available) • Mi 9T (mobile) • e7b48232 • android-arm64 • Android 11 (API 30) • macOS (desktop) • macos • darwin-arm64 • macOS 14.5 23F79 darwin-arm64 • Mac Designed for iPad (desktop) • mac-designed-for-ipad • darwin • macOS 14.5 23F79 darwin-arm64 • Chrome (web) • chrome • web-javascript • Google Chrome 128.0.6613.120 [✓] Network resources • All expected network resources are available. • No issues found! Hello @Luwirispok ! Can you explain how request logs looks in your VSCode console ?
2025-04-01T04:55:06.160313
2024-03-01T14:17:50
2163511688
{ "authors": [ "jos0405", "pavanobbeli" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13985", "repo": "FriendlyDotCH/mautic-multi-domain", "url": "https://github.com/FriendlyDotCH/mautic-multi-domain/issues/25" }
gharchive/issue
Unable to install plugin Team, I have manually installed this plugin https://github.com/FriendlyDotCH/mautic-multi-domain?tab=readme-ov-file and I am receiving an error when attempting to access Mautic. Could someone help me to resolve this issue? Are you working with Mautic 4 or 5? This is not M5 compatible yet
2025-04-01T04:55:06.168369
2013-04-22T10:16:37
13473188
{ "authors": [ "bamarni", "lyrixx" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13986", "repo": "FriendsOfPHP/Sismo", "url": "https://github.com/FriendsOfPHP/Sismo/issues/100" }
gharchive/issue
Add author email to Commit object What would you think about adding author's email to commit objects? My use case is to add at runtime commit's author as a notifier recipient : <?php use Sismo\Commit; use Sismo\Notifier\MailNotifier; class Notifier extends MailNotifier { public function notify(Commit $commit) { /* $authorEmail = $commit->getAuthorEmail(); // ??? $recipients = (array) $this->recipients; if (!in_array($authorEmail, $recipients)) { $this->recipients[] = $authorEmail; } */ $previousCommit = $commit->getProject()->getLatestCommit(); if (!$commit->isSuccessful() || !$previousCommit || $previousCommit->getStatusCode() != $commit->getStatusCode() ) { return parent::notify($commit); } return false; } } ping @bamarni hum it's hanging since a while :) in the meanwhile I switched to a cloud based solution so I can't implement this at the moment unfortunately..
2025-04-01T04:55:06.175559
2014-07-04T18:09:21
37181630
{ "authors": [ "perk11", "stevro" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13987", "repo": "FriendsOfSymfony/FOSCommentBundle", "url": "https://github.com/FriendsOfSymfony/FOSCommentBundle/pull/441" }
gharchive/pull-request
Using Symfony\Component\Validator\Context\LegacyExecutionContext Quick fix so voting works in Symfony 2.5. Has anyone provided a fix for this issue? Thanks.
2025-04-01T04:55:06.242744
2019-06-24T19:43:45
460062695
{ "authors": [ "fubar-coder", "lfalck" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13988", "repo": "FubarDevelopment/FtpServer", "url": "https://github.com/FubarDevelopment/FtpServer/issues/72" }
gharchive/issue
Error with validateScopes = true using version 3.0rc4 Hi! When i build the service provider with validation of scopes like this: services.BuildServiceProvider(validateScopes: true) I get this error: Cannot resolve scoped service 'FubarDev.FtpServer.Commands.IFtpCommandHandlerProvider' from root provider. ValidateScopes is also true when using CreateDefaultBuilder in ASP.NET Core which is where i encountered the issue. Here is a minimal example of the issue, if i remove validateScopes i can connect but with it the application crashes. Great work with 3.0, looking foward to the release! /Ludvig Fixed in RC5 (should be available soon) Great, thanks!
2025-04-01T04:55:06.264936
2015-06-02T06:57:51
83877883
{ "authors": [ "gokulkrishh", "logeshpaul", "rgksugan" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13989", "repo": "FuelFrontend/generator-smacss", "url": "https://github.com/FuelFrontend/generator-smacss/issues/46" }
gharchive/issue
Not listed in yeoman.io This generator is not listed in the yeoman.io site Updated the package.json file based on the reply - https://twitter.com/Vaxilart/status/592434542503469056 @logeshpaul @rgksugan Our generator is know listed in yeoman.io site. Check the screenshot. @gokulkrishh cool! There was an issue with the 'repo:url' in 'package.json' file. I fixed it and pushed it along with other fix we I did yesterday. Great that it fixed and listed! Thanks for checking :+1: Posting the line number here for reference https://github.com/FuelFrontend/generator-smacss/blob/9cb1ea13733101594b05abfd9b8d4ba488a7f96d/package.json#L35 Cool!!
2025-04-01T04:55:06.268133
2022-04-14T08:27:43
1204212908
{ "authors": [ "adlerjohn", "binggh", "mitchmindtree" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13990", "repo": "FuelLabs/sway-vscode-plugin", "url": "https://github.com/FuelLabs/sway-vscode-plugin/issues/42" }
gharchive/issue
sway vscode plugin crashes on start up The sway vscode plugin immediately crashes on startup. This is probably due to the changes to the LSP feature becoming a plugin instead of a native command: https://github.com/FuelLabs/sway/pull/1178 Error: no such subcommand: `lsp` [Info - 4:21:53 PM] Connection to server got closed. Server will restart. Error: no such subcommand: `lsp` [Info - 4:21:53 PM] Connection to server got closed. Server will restart. Error: no such subcommand: `lsp` [Info - 4:21:53 PM] Connection to server got closed. Server will restart. Error: no such subcommand: `lsp` [Info - 4:21:53 PM] Connection to server got closed. Server will restart. Error: no such subcommand: `lsp` [Error - 4:21:53 PM] Connection to server got closed. Server will not be restarted. @binggh you might just need to install the forc-lsp plugin - could you try cargo install forc-lsp, then re-launch VS Code and see if that works? I haven't looked into how VS Code plugins work yet myself (I do my editing with Vim), but perhaps there's some way we can make sure forc-lsp is installed during the plugin installation process. I think this should mostly be resolved once we have binary distributions. In the meantime, if that is indeed the fix, then The Book can simply be update to include instruction to install the lsp and other important plugins. @mitchmindtree oops, my bad - that fixed it 😅 will close this issue and update the docs at https://fuellabs.github.io/sway/latest/introduction/sway-toolchain.html
2025-04-01T04:55:06.273105
2021-01-20T15:06:05
790062444
{ "authors": [ "lknegendorf" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13991", "repo": "Fuenfgeld/ATeamDatenmanagementUndArchivierung", "url": "https://github.com/Fuenfgeld/ATeamDatenmanagementUndArchivierung/issues/22" }
gharchive/issue
Identify dependencies in the data needed for our research Depending on the results of #21! For example code->description and reasoncode->reasondescription These "translations" should not be in the main table but in the satellite-tables of our star scheme CODE -> DESCRIPTION CODE -> DESCRIPTION
2025-04-01T04:55:06.277422
2017-05-17T21:32:31
229493635
{ "authors": [ "coveralls", "kangkyu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13992", "repo": "Fullscreen/yt-support", "url": "https://github.com/Fullscreen/yt-support/pull/6" }
gharchive/pull-request
Add as_curl I borrowed as_curl method from 'yt' gem. and think it could be useful for debugging projects which use 'yt-core' gem. Thank you Coverage decreased (-4.3%) to 95.726% when pulling caaf5c2f40f6936623044eb48692b20cfbdedfb9 on kangkyu:as_curl into 38c6892b8a9a853b50fc47ef748f7a64dfcbe882 on Fullscreen:master. Coverage remained the same at 100.0% when pulling 6133e0193fdf6d0327bf7b7d542e85c896b443ca on kangkyu:as_curl into 38c6892b8a9a853b50fc47ef748f7a64dfcbe882 on Fullscreen:master. closes in favor of #8
2025-04-01T04:55:06.313197
2024-11-08T07:33:16
2643146600
{ "authors": [ "jamesbraza" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13994", "repo": "Future-House/aviary", "url": "https://github.com/Future-House/aviary/pull/117" }
gharchive/pull-request
Fixing CI by downpinning uv<0.5 From this CI run, we get: × No solution found when resolving dependencies for split │ (python_full_version == '3.11.*' and platform_python_implementation == │ 'PyPy'): ╰─▶ Because only the following versions of paper-qa are available: paper-qa<=5.0.0 paper-qa==5.0.1 ... paper-qa==5.3.2 and paper-qa==5.3.2 depends on fhaviary, we can conclude that paper-qa>5.3.1 depends on fhaviary. And because paper-qa>=5.0.0,<=5.3.1 depends on fhaviary, we can conclude that paper-qa>=5.0.0 depends on fhaviary. And because fhaviary[paperqa] depends on paper-qa>=5 and your workspace requires fhaviary[paperqa], we can conclude that your workspace's requirements are unsatisfiable. hint: The package `paper-qa` depends on the package `fhaviary` but the name is shadowed by one of your workspace members. Consider changing the name of the workspace member. Something about uv==0.5.0 broke our "circular" pinning with paper-qa (where paper-qa depends on fhaviary, and we have a convenience paperqa extra here). This PR just temporarily down-pins uv for this Looks like test passed this time, but lint still fails. Need to dig deeper
2025-04-01T04:55:06.316430
2021-09-02T09:33:10
986449759
{ "authors": [ "Fyrd", "j-hoffmann" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13995", "repo": "Fyrd/caniuse", "url": "https://github.com/Fyrd/caniuse/issues/6009" }
gharchive/issue
Missing browser usage data for iOS in Switzerland Usage data for Safari / Chrome on iOS devices seems to be missing. My screenshot below shows an example for browser versions 14.5 - 14.7, but all iOS versions are affected and are showing 0% for Switzerland. These numbers should be higher than global numbers since market share of iOS is generally higher in Switzerland. Numbers for other browsers, including Safari on macOS do not seem to be affected. I cannot tell exactly when this started happening, but it must be at least a few days. I believe this was recently fixed. Note that you may need to delete the current data then re-import. Thank you. I thought I had done a reimport, but trying it now helped. Problem fixed.
2025-04-01T04:55:06.326466
2022-10-10T20:19:09
1403668214
{ "authors": [ "dave-gantenbein", "kannon92", "suprjinx" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13996", "repo": "G-Research/armada", "url": "https://github.com/G-Research/armada/issues/1608" }
gharchive/issue
ArmadaOperator: Add airflow task id to Armada annotations When a user submits an Airflow task, we should add the airflow task id to the Armada annotations. An example can be below: jobSetId: test priority: 100 namespace: batch-test annotations: armadaproject.io/taskRunId: 0a016adf-8325-1d6a-8183-268de7a625ea @iJanki-gr Is there any other items in the airflow context that you would want added as annotations? I believe everything is in place to add the annotations, we just need to read them from airflow context and create them in the job request. Hey team! Please add your planning poker estimate with Zenhub @ClifHouck @dejanzele @headphonejames @kannon92 @richscott @Sharpz7 @suprjinx @kannon92 it seems like airflow has a "task_id" and a "run_id" as separate identifiers -- which should be used for "taskRunId" in the description? Please hold on this. I need to meet with Danielle to discuss airflow user request but it will have to happen after Kubecon. So @iJanki-gr and I meant to sync up before I left for Kubecon. He has some ideas of what potential airflow users would want from this work but I think we need to see some DAGs and see what is possible. At this moment, we use dagrunId as our job-set name. There was a request to dynamically use task name from armada job ids but I don't know if that is possible. Airflow seems to want static task names but it seems possible to do dynamic generation of dags but it seems to be an advanced feature so I think more information is necessary. Feel free to take lead on this if you want with Daniele.
2025-04-01T04:55:06.336670
2023-01-01T05:07:11
1515319947
{ "authors": [ "G0maa" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13997", "repo": "G0maa/my-school-server", "url": "https://github.com/G0maa/my-school-server/issues/32" }
gharchive/issue
Use of middlewares for zod As suggested by @Amr2812 Reference: https://github.com/G0maa/my-school-server/blob/a8971d8341a0ff4440dc9c9abbc886de34d3aaae/src/controllers/student.ts#L23-L24 and all other controllers. Goal: Single middleware use, single schema to fulfill the above need. Revisiting. I've been trying to adapt this to the current code, but I haven't been able to. So far what I have got is const validateSchemas = (schemas: AnyZodObject[]) => (req: Request, _res: Response, next: NextFunction) => { // Error handling in errror-handler middleware // await/promise is needed only if you use **async** `refinements` or `transformers` for (const schema of schemas) { schema.parse({ body: req.body, query: req.query, params: req.params, }); } return next(); }; And export const ZStudentPost = z.object({ body: ZStudent, }); I cant use transform since it will invalidate the state of req.body: export const ZUserPost = z .object({ body: ZUser, }) .transform((obj) => { return { ...obj, user: obj.body }; }); Which means, either: Find a way to parse req.body twice populating it with the needed schema i.e. req.body.user & req.body.student Redesign schemas around routes themselves rather than entities. I used ChatGPT for 3-4 tries to refine the suggested code: function validateAndExtract(schemas, properties) { return (req, res, next) => { try { let parsed = {}; schemas.forEach(schema => { parsed = Object.assign(parsed, schema.parse(req.body)); }); properties.forEach(property => { req.body[property] = parsed[property]; }); next(); } catch (err) { return res.status(400).json({ message: err.message }); } } } i.e. to use validateAndExtract([teacherSchema, userSchema], ['teacher', 'user']) Revisited, tried & applied: Here it is used to verify entire request, but manually giving types to req, this hugely depends on client sending different objects, user, student & studentDetails in req https://github.com/G0maa/my-school-server/blob/576c72f15ca8e0440695c4eb724b1f794d07f6b9/src/controllers/teacher.ts#L55-L67 Here it is too much to use a middleware for a small task: https://github.com/G0maa/my-school-server/blob/576c72f15ca8e0440695c4eb724b1f794d07f6b9/src/controllers/teacher.ts#L92-L101 And here you can't use it, req.query can't be sent as objects. https://github.com/G0maa/my-school-server/blob/576c72f15ca8e0440695c4eb724b1f794d07f6b9/src/controllers/teacher.ts#L28-L47 3rd method https://github.com/G0maa/my-school-server/blob/359143fb8524864d11340294aaa43a5e75f07a9f/src/controllers/fee.ts#L38-L56 For Reference, old: https://github.com/G0maa/my-school-server/blob/def70826f7e5d43e8f683b0816e4e686d51e6192/src/validator/studyClass.validator.ts Solved with #45
2025-04-01T04:55:06.361381
2019-12-16T14:29:46
538447637
{ "authors": [ "nicolsonaitken" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13999", "repo": "GCTC-NTGC/TalentCloud", "url": "https://github.com/GCTC-NTGC/TalentCloud/issues/2363" }
gharchive/issue
Bug - Unable to save changes to Profile b/c Password is incorrect. Description While logged into the Talent Cloud site, making changes to my Twitter handle, LinkedIn URL and tagline I am unable to save as there is an error message stating that: Current password isn't correct. As I am already logged in and made no changes to my password, I don't know why or how my password could be incorrect. Technical Info Device: Laptop Operating System: MacOs Browser: Chrome, Version 79.0.3945.79 (Official Build) (64-bit) Steps to Reproduce Steps to reproduce the behaviour. Example: Go to Talent Cloud Click on About Me Scroll down to Account Information Fill in additional details and Save Changes Interaction is included within the video attached. Video Profile.mp4.zip [ ] Regression test written. Password is now within settings and the issue looks to be resolved.
2025-04-01T04:55:06.366602
2021-09-26T23:41:34
1007538455
{ "authors": [ "Suryanshomar7240", "mohit355" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14000", "repo": "GDSC-IIIT-Kalyani/Winter-of-Code-2.0-frontend", "url": "https://github.com/GDSC-IIIT-Kalyani/Winter-of-Code-2.0-frontend/pull/22" }
gharchive/pull-request
project Section The Project section made is fully responsive but still, it can be improved in terms of animations filter option also add to filter those projects according to the topics @Suryanshomar7240 Thanks for your contribution.
2025-04-01T04:55:06.374542
2021-07-12T08:54:38
941826515
{ "authors": [ "4ian", "Bouh" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14001", "repo": "GDevelopApp/GDevelop-extensions", "url": "https://github.com/GDevelopApp/GDevelop-extensions/issues/181" }
gharchive/issue
Remove The Kongregate Extension from The Extension Library Description From https://github.com/4ian/GDevelop/issues/2797 Kongregate does not accept games any more, which makes the extension useless. Probably add a gamejolt or newgrounds API extension to GDevelop 5 instead. [x] Verification on Kongregate website [x] Check how deleted extension act in GDevelop if we try to update it Check how deleted extension act in GDevelop if we try to update it Nothing bad should happen as existing games having it will continue to work, and it will be removed from the list of extensions, so it won't be listed and you can't "update" it. Close by https://github.com/GDevelopApp/GDevelop-extensions/commit/a7f8ed33c5050d5ae9776bf8e48cc9a2c912485e
2025-04-01T04:55:06.377063
2021-10-08T19:28:35
1021402754
{ "authors": [ "Saad-Mahamood", "tonysun9" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14002", "repo": "GEM-benchmark/NL-Augmenter", "url": "https://github.com/GEM-benchmark/NL-Augmenter/issues/316" }
gharchive/issue
gender_neutral_rewrite Unresolved references to spaCy and Unresolved List reference When running the gender_neutral_rewrite there are several unresolved references to the spacy_nlp variable. In particular on line: Line 27: self.nlp = spacy_nlp if spacy_nlp else spacy.load("en_core_web_sm") Please use from initialize import spacy_nlp to get a handle on the global spacy instance. There is also an unresolved reference on Line 495: def generate(self, sentence: str) -> List[str]. List[str] is not resolvable. Should this be lower case? e.g. list[str] Thanks for pointing this out. Made a PR here: https://github.com/GEM-benchmark/NL-Augmenter/pull/324
2025-04-01T04:55:06.379378
2018-01-23T16:48:16
290910256
{ "authors": [ "dbauszus-glx" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14003", "repo": "GEOLYTIX/xyz", "url": "https://github.com/GEOLYTIX/xyz/issues/19" }
gharchive/issue
Generate database array from environmental settings Currently the database array is hardcoded at the beginning of modules which use it. const databases = { xyz: pgp(process.env.POSTGRES), ghs: pgp(process.env.POSTGRES_GHS) }; This array should be generated from the launch settings. Split name at underscore with the second part being the name of the database connection which is used in this instance. This is now possible. Database connection strings begin with DBS_ the keys are split at the underscore and added to a DBS array. The second part of the key is the name of the database connection which can be referenced in queries to the framework. let pgp = require('pg-promise')({ promiseLib: require('bluebird'), noWarnings: true }); const DBS = {}; Object.keys(process.env).map(function (key) { if (key.split('_')[0] === 'DBS') DBS[key.split('_')[1]] = pgp(process.env[key]) });
2025-04-01T04:55:06.381478
2023-04-24T18:19:18
1681798348
{ "authors": [ "gmao-rreichle", "mathomp4" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14004", "repo": "GEOS-ESM/GEOS_Util", "url": "https://github.com/GEOS-ESM/GEOS_Util/issues/15" }
gharchive/issue
remap_upper.py should pass in stretch factors to interp_restarts After talking with @bena-nasa, the right way to pass in stretch grid factors in remap_upper.py is to do so through command line arguments to interp_restarts.x. If you do it this way, the restarts get global metadata for the stretch factor, lon, and lat. Oh dear. I tried this and for some reason the winds (and other things) are very different when you a namelist vs options. @bena-nasa is looking at this. Addressed in #19
2025-04-01T04:55:06.389968
2024-12-20T15:33:51
2752969089
{ "authors": [ "Dooruk" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14005", "repo": "GEOS-ESM/swell", "url": "https://github.com/GEOS-ESM/swell/issues/490" }
gharchive/issue
(SE) Utilizing more Cylc features One of the main reasons we use Cylc is its design which includes many useful features for DA cycling. We currently exercise some of them but we should take advantage of more. I will create this as an epic and as users think of more features they can add this can expand. Necessary ones: [ ] Retry certain tasks in case they fail: Most of the time the failure of Run.. tasks are caused by filesystem issues rather than configuration problems. Cylc should try running a certain task at least one more time before giving up. This is especially important for suites involving ensembles as even if 31 members succeed and 1 fails the workflow will stall. [ ] Hold before certain tasks: @rtodling mentioned this one. The ability for a suite to "hold" before a certain task, say RunJediVarExecutable so that different configurations could be tested swiftly without the need to swell create a whole new suite. JCSDA's https://github.com/JCSDA-internal/skylab uses this feature in a different workflow engine called EWOK. [ ] Restart a failed suite from a certain cycle & task: This one is tricky, I know it is possible with Cylc but not sure how would this play out with Swell. After a suite cycled certain amount of days, we would want to continue from a that point rather than restarting altogether. Optional but useful ones: [ ] Workflow sending email if suite stops/fails. I was able to do this with a sandbox Cylc setup. Case in point for retrying, the task had SLURM issues (pink square) but now running after two failed attempts and with zero changes. I happened to be monitoring:
2025-04-01T04:55:06.417554
2017-05-01T19:41:31
225508284
{ "authors": [ "dalcinl", "v-dobrev" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14006", "repo": "GLVis/glvis", "url": "https://github.com/GLVis/glvis/pull/21" }
gharchive/pull-request
Use a larger socketserver backlog [server-backlog] Increasing the backlog should help in servicing requests from clients running in parallel with large number of processes. While the value of 128 is still relatively small (and not the definitive cure for parallel clients flooding the listening socket), it is usually the default maximum in Linux and Darwin kernels: $ uname Linux $ sysctl net.core.somaxconn net.core.somaxconn = 128 $ uname Darwin $ sysctl kern.ipc.somaxconn kern.ipc.somaxconn: 128 @stefanozampini, GLVis does not have a config file for options. The backlog can be made a command line option, if you want. I'm fine either way.
2025-04-01T04:55:06.669335
2020-08-18T16:04:43
681159501
{ "authors": [ "pjsharpe07" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14007", "repo": "GSA/sdg-indicators-usa", "url": "https://github.com/GSA/sdg-indicators-usa/pull/1011" }
gharchive/pull-request
Release 20200818 Part of release 20200820 This is part of the 20200820 release Going to close and reopen. Just going to close it all together
2025-04-01T04:55:06.692507
2021-01-15T20:09:17
787163609
{ "authors": [ "Shaun-Regenbaum" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14008", "repo": "GT-Jewish-DH/daf-render-lib", "url": "https://github.com/GT-Jewish-DH/daf-render-lib/issues/10" }
gharchive/issue
Create Example for Calling from Sefaria API Create an example where you call from Sefaria's API. We need to find a better way to simplify the example, we don't want to confuse people with all the processing junk. For now, I am going to push this off as it is more a documentation problem than simply creating an example. We need to find a better way to simplify the example, we don't want to confuse people with all the processing junk. For now, I am going to push this off as it is more a documentation problem than simply creating an example.
2025-04-01T04:55:06.700548
2024-03-20T11:44:10
2197329375
{ "authors": [ "kyoi-the-fox" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14009", "repo": "GTNewHorizons/BloodArsenal", "url": "https://github.com/GTNewHorizons/BloodArsenal/pull/24" }
gharchive/pull-request
add null checks for all sendPlayerInformation cases,to avoid crashes … …if ritual owner is not online and Altar is missing Pedestals are missing Item to enchant is missing Enchantment books are missing Not enough LP in the network to enchant items Enchantment type can not be applied on item flew under the radar because they only happen under specific conditions and were not part of the issue because of it that would prevent it from working in it's entirety if the player owning the activation crystal is not on the server,this would still allow it to be used by other team members,but prevents the crash
2025-04-01T04:55:06.737852
2024-03-22T13:43:29
2202528804
{ "authors": [ "Mi-Peng", "mtfishman" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14010", "repo": "GTorlai/PastaQ.jl", "url": "https://github.com/GTorlai/PastaQ.jl/issues/307" }
gharchive/issue
How to calculate the entropy/entanglement metrics efficiently Thanks for this awesome work. I'm wondering if PastaQ provides API to calculate the entropy and other entanglement metrics efficiently, or they are calculated manually, e.g., using eigenvalues to calculate von Neumann entropy, which is considered time-consuming when q-bit is large. Thanks. I would take a look at the ITensor discussion forum (https://itensor.discourse.group/) and ask a question there if your question isn't answered in one of the posts already. I'm closing since I think this is more of a user support question, and that functionality is outside of the scope of this package (we may add more convenient functions for how to do that in ITensor, that functionality probably would not go here).
2025-04-01T04:55:06.763803
2015-08-17T13:19:17
101419215
{ "authors": [ "aaronlevin", "ocharles" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14012", "repo": "Gabriel439/post-rfc", "url": "https://github.com/Gabriel439/post-rfc/issues/5" }
gharchive/issue
ad - automatic differentiation This has always stood out as an interesting and uniquely-haskell library. I'm not sure where it should go. I'd be happy to submit a PR with a suggestion of a good category. Data science? https://hackage.haskell.org/package/ad ad is generally useful in the machine learning field (though I also used it recently to do some root-finding I needed for graphics animations. It probably belongs in either machine learning or just numerics.
2025-04-01T04:55:06.778114
2020-06-22T01:13:18
642680759
{ "authors": [ "CanRau", "arp242", "peixian" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14013", "repo": "GaiAma/Coding4GaiAma", "url": "https://github.com/GaiAma/Coding4GaiAma/pull/26" }
gharchive/pull-request
Add support for self-hosted goatcounter Adds a selfHostUrl attribute to the pluginOptions, where a user can specify a selfhosted goatcounter URL. Sorry for the delay and thanks for the PR 🙏 Personally I think it would be best if the plugin would always accept the full URL, rather than just a "code". In the documentation for the integration there's an "integrations" section now with: The reason I use a full URL is so that it'll work in any use case: goatcounter.com, custom domains, self-host. I found this because another customer trying to use this plugin emailed me with issues and the URL they ended up using is: https://https//xxx.goatcounter.com/count.goatcounter.com/ which is obvious not correct heh. Assuming you agree this would be a good idea, I don't know what the best way forward is here with regards to compatibility, but perhaps renaming selfHostUrl to just url or endpoint and keeping the code as a "hidden" undocumented option might be a good idea? I assume that at this point only @peixian is using this, so it can be changed pretty safely at this point? Hey @arp242 thanks for your tip & insides 🙏 sorry for the late reply I didn't receive a notification, probably because it's closed, maybe in such cases it's needed to directly @ mention the person? 🤔 not sure about that 🤷‍♂️ @peixian Sorry for the delay, again, I now uploaded 0.4.0 to NPM 🎉 including your PR and will soonish make the needed changes suggested by Martin and upload those as 0.5.0 I guess. Yea endpoint sounds appropriate as it's matching GoatCounter's terminology 👌
2025-04-01T04:55:06.813143
2015-11-27T11:12:43
119178724
{ "authors": [ "Galexrt", "deevant07" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14014", "repo": "Galexrt/docker-zulip", "url": "https://github.com/Galexrt/docker-zulip/issues/51" }
gharchive/issue
Zulip exited with code 1(stuck while waiting for DB connections) Hi, I am not sure if the problem is with config but i am newbie to zulip and docker. The docker-compose.yml is configured as below. database: image: "galexrt/zulip-postgresql-tsearchextras:latest" environment: DB_NAME: zulip DB_USER: zulip DB_PASS: zulip volumes: - "/opt/docker/zulip/postgresql/data:/var/lib/postgresql:rw" memcached: image: "sameersbn/memcached:latest" restart: always rabbitmq: image: "rabbitmq:3.5.5" hostname: zulip-rabbit restart: always environment: RABBITMQ_DEFAULT_USER: "zulip" RABBITMQ_DEFAULT_PASS: "zulip" redis: image: "sameersbn/redis:latest" volumes: - "/opt/docker/zulip/redis:/var/lib/redis:rw" zulip: image: "galexrt/zulip:1.3.7-2" ports: - "8081:8080" - "8443:443" links: - database - memcached - rabbitmq - redis environment: ZULIP_SETTINGS_EXTERNAL_HOST: "domain.com" ZULIP_SETTINGS_ZULIP_ADMINISTRATOR<EMAIL_ADDRESS>ZULIP_SETTINGS_ADMIN_DOMAIN: "domain.com" ZULIP_SETTINGS_NOREPLY_EMAIL_ADDRESS<EMAIL_ADDRESS>ZULIP_SETTINGS_DEFAULT_FROM_EMAIL: "Zulip<EMAIL_ADDRESS>ZULIP_SETTINGS_EMAIL_HOST: "<IP_ADDRESS>" ZULIP_SETTINGS_EMAIL_HOST_USER: "testuser" ZULIP_SETTINGS_EMAIL_HOST_PASSWORD: "xxxx" ZULIP_USER_EMAIL<EMAIL_ADDRESS>ZULIP_USER_PASS: "xxxx" ZULIP_USER_DOMAIN: "example.com" volumes: - "/opt/docker/zulip/zulip:/data:rw" The part of the log file which highlights the problem is as follows. zulip_1 | Setting database configuration ... zulip_1 | Setting key "DATABASES", type "array". zulip_1 | Database configuration succeeded. zulip_1 | Setting caches configuration ... zulip_1 | Setting key "CACHES", type "array". zulip_1 | Caches configuration succeeded. zulip_1 | Activating authentication backends ... zulip_1 | Setting key "AUTHENTICATION_BACKENDS", type "array". zulip_1 | Adding authentication backend "EmailAuthBackend". zulip_1 | Authentication backend activation succeeded. zulip_1 | Setting redis configuration ... zulip_1 | Setting key "RATE_LIMITING", type "bool". zulip_1 | Setting key "REDIS_HOST", type "string". zulip_1 | Setting key "REDIS_HOST_PORT", type "int". zulip_1 | Redis configuration succeeded. zulip_1 | Setting rabbitmq configuration ... zulip_1 | Setting key "RABBITMQ_HOST", type "string". zulip_1 | Setting key "RABBITMQ_USERNAME", type "string". zulip_1 | Rabbitmq configuration succeeded. zulip_1 | Executing Zulip configuration ... zulip_1 | Setting key "EMAIL_HOST_USER", type "string". zulip_1 | Setting key "NOREPLY_EMAIL_ADDRESS", type "string". zulip_1 | Setting key "EMAIL_HOST", type "string". zulip_1 | Setting key "EXTERNAL_HOST", type "string". zulip_1 | Setting key "EMAIL_HOST_PASSWORD", type "string". zulip_1 | Setting key "ADMIN_DOMAIN", type "string". zulip_1 | Setting key "DEFAULT_FROM_EMAIL", type "string". zulip_1 | Setting key "ZULIP_ADMINISTRATOR", type "string". zulip_1 | Zulip configuration succeeded. zulip_1 | Auto backup enabled. zulip_1 | === End Initial Configuration Phase === zulip_1 | === Begin Bootstrap Phase === zulip_1 | Waiting for database server to allow connections ... zulip_1 | ...........................................................openstack_zulip_1 exited with code 1 Detailed log file is attached. Kindly help in resolving the problem. I am using the command sudo docker-compose up docker-zulip.txt Try again with the latest version https://github.com/Galexrt/docker-zulip/releases/tag/1.3.9 Hi, I am not able to pull the image 1.3.9 version from quay.io/galexrt/zulip:v1.3.9 with following error: Error response from daemon: unable to ping registry endpoint https://quay.io/v0/ v2 ping attempt failed with error: Get https://quay.io/v2/: x509: certificate signed by unknown authority v1 ping attempt failed with error: Get https://quay.io/v1/_ping: x509: certificate signed by unknown authority Can the image be created on docker.io as well so that its easy to download and proceed? https://hub.docker.com/r/galexrt/zulip/tags/ The above link is showing all tags except 1.3.9 Thanks, Deevan @deevant07 I created the tags on the docker hub now. Should be built in about 30 minutes. Thank you @Galexrt Hi, I am still facing the issue with version 1.3.9 with below error: zulip_1 | Zulip configuration succeeded. zulip_1 | Auto backup enabled. zulip_1 | === End Initial Configuration Phase === zulip_1 | === Begin Bootstrap Phase === zulip_1 | Waiting for database server to allow connections ... rabbitmq_1 | rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:18 === rabbitmq_1 | Starting RabbitMQ 3.5.5 on Erlang 18.1 rabbitmq_1 | Copyright (C) 2007-2015 Pivotal Software, Inc. rabbitmq_1 | Licensed under the MPL. See http://www.rabbitmq.com/ rabbitmq_1 | rabbitmq_1 | RabbitMQ 3.5.5. Copyright (C) 2007-2015 Pivotal Software, Inc. rabbitmq_1 | ## ## Licensed under the MPL. See http://www.rabbitmq.com/ rabbitmq_1 | ## ## rabbitmq_1 | ########## Logs: tty rabbitmq_1 | ###### ## tty rabbitmq_1 | ########## rabbitmq_1 | Starting broker... rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:18 === rabbitmq_1 | node : rabbit@zulip-rabbit rabbitmq_1 | home dir : /var/lib/rabbitmq rabbitmq_1 | config file(s) : /etc/rabbitmq/rabbitmq.config rabbitmq_1 | cookie hash : qRoOkIXORPV3bIHgNaTeww== rabbitmq_1 | log : tty rabbitmq_1 | sasl log : tty rabbitmq_1 | database dir : /var/lib/rabbitmq/mnesia/rabbit@zulip-rabbit rabbitmq_1 | rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:18 === rabbitmq_1 | Memory limit set to 1581MB of 3953MB total. rabbitmq_1 | rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:18 === rabbitmq_1 | Disk free limit set to 50MB rabbitmq_1 | rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 === rabbitmq_1 | Limiting to approx 524188 file handles (471767 sockets) rabbitmq_1 | rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 === rabbitmq_1 | FHC read buffering: ON rabbitmq_1 | FHC write buffering: ON rabbitmq_1 | rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 === rabbitmq_1 | Priority queues enabled, real BQ is rabbit_variable_queue rabbitmq_1 | rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 === rabbitmq_1 | msg_store_transient: using rabbit_msg_store_ets_index to provide index rabbitmq_1 | rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 === rabbitmq_1 | msg_store_persistent: using rabbit_msg_store_ets_index to provide index rabbitmq_1 | rabbitmq_1 | =WARNING REPORT==== 1-Dec-2015::11:59:19 === rabbitmq_1 | msg_store_persistent: rebuilding indices from scratch rabbitmq_1 | rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 === rabbitmq_1 | started TCP Listener on [::]:5672 rabbitmq_1 | completed with 0 plugins. rabbitmq_1 | rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 === rabbitmq_1 | Server startup complete; 0 plugins started. zulip_1 | ...........................................................openstack_zulip_1 exited with code 1 @deevant07 Can you provide your current docker-compose.yml? Remove all security data from it (passwords, etc.) Hi, Please find below the docker-compose.yml. database: image: "galexrt/zulip-postgresql-tsearchextras:latest" environment: DB_NAME: zulip DB_USER: zulip DB_PASS: zulip volumes: - "/opt/docker/zulip/postgresql/data:/var/lib/postgresql:rw" memcached: image: "sameersbn/memcached:latest" restart: always rabbitmq: image: "rabbitmq:3.5.5" hostname: zulip-rabbit restart: always environment: RABBITMQ_DEFAULT_USER: "zulip" RABBITMQ_DEFAULT_PASS: "zulip" redis: image: "sameersbn/redis:latest" volumes: - "/opt/docker/zulip/redis:/var/lib/redis:rw" zulip: image: "galexrt/zulip:v1.3.9" ports: - "8081:8080" - "8443:443" links: - database - memcached - rabbitmq - redis environment: ZULIP_SETTINGS_EXTERNAL_HOST: "domain.com" ZULIP_SETTINGS_ZULIP_ADMINISTRATOR<EMAIL_ADDRESS> ZULIP_SETTINGS_ADMIN_DOMAIN: "domain.com" ZULIP_SETTINGS_NOREPLY_EMAIL_ADDRESS<EMAIL_ADDRESS> ZULIP_SETTINGS_DEFAULT_FROM_EMAIL: "Zulip " ZULIP_SETTINGS_EMAIL_HOST: "xxxxxx" ZULIP_SETTINGS_EMAIL_HOST_USER: "xxxxxx" ZULIP_SETTINGS_EMAIL_HOST_PASSWORD: "xxxx" ZULIP_USER_EMAIL: "xxxxxxx" ZULIP_USER_PASS: "xxxxxx" ZULIP_USER_DOMAIN: "example.com" volumes: - "/opt/docker/zulip/zulip:/data:rw" Thanks, Deevan @deevant07 You are missing some environment variables for the zulip container. On the first look, you are missing the DB_* env variables. Please add them or download the neweset version of the docker-compose.yml https://github.com/Galexrt/docker-zulip/blob/master/docker-compose.yml and try again please. Thanks @Galexrt Installation is done successfully but facing one issue on signup. Your e-mail does not match any existing open organization. Use a different e-mail address, or contact<EMAIL_ADDRESS>with questions. I have gone through some links with the solution to edit the manage.py file but as i am running from docker i am not able to find the file on host machine directory. @deevant07 You need to configure the email settings in the docker-compose.yml. See https://github.com/Galexrt/docker-zulip/wiki/Configuration and https://github.com/Galexrt/docker-zulip/wiki/Zulip-Email-SMTP-Configuration. Closing as of the "bug" is fixed.
2025-04-01T04:55:07.503940
2021-07-30T12:23:15
956698188
{ "authors": [ "EmeraldBlock", "Gaming32" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14015", "repo": "Gaming32/ArrayV-v4.0", "url": "https://github.com/Gaming32/ArrayV-v4.0/pull/70" }
gharchive/pull-request
Advanced Shuffle Editor This editor allows you to chain shuffles as they run. This PR still allows you to use the old system if you like. So is it ready then @EmeraldBlock? I haven't yet tested the new changes, but if you think it's all fine, yeah
2025-04-01T04:55:07.505485
2023-03-13T16:26:51
1621835825
{ "authors": [ "Gaming32", "Madis0" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14016", "repo": "Gaming32/world-host", "url": "https://github.com/Gaming32/world-host/issues/4" }
gharchive/issue
Do other parties need the mod? Those that want to access the hosted server, that is. Please clarify that in the description of the mod. Currently there is a proxy IP that exists, but no way to view it. I might add that later. This functionality is available in World Host 0.2.2.
2025-04-01T04:55:07.516858
2015-10-09T18:07:50
110709913
{ "authors": [ "angieskazka" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14017", "repo": "Gapminder/vizabi", "url": "https://github.com/Gapminder/vizabi/issues/644" }
gharchive/issue
"Show" control, similar to "Find" Need the same thing to be able to show just 2 country mountains (edit show field of state) done in e1cda784054775313fa1e9c27861f4d5ec910587
2025-04-01T04:55:07.518836
2018-10-29T18:02:58
375146244
{ "authors": [ "Areloch", "marauder2k7", "wcbx" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14018", "repo": "GarageGames/Torque3D", "url": "https://github.com/GarageGames/Torque3D/pull/2281" }
gharchive/pull-request
OpenALEffects Ground work, this commit enables reverb zones and opens the slots for other effects to be added later. Tested using the full template. Build your openal dll and your full.exe use one of the environment presets and it will work i should add this also gives the option for exponential falloff Seems to all work, did note a typo with "PresetFroest" presumably that's supposed to be PresetForest? Also, if you could correct the indentations on the SFXEnvironment definitions, and the typo, i think it's good to go :) yeah it was supposed to be preset forest lol sorry rushing to test it out was a bit excited. bump
2025-04-01T04:55:07.522764
2023-12-07T23:21:20
2031675187
{ "authors": [ "WillEngler" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:14019", "repo": "Garden-AI/garden", "url": "https://github.com/Garden-AI/garden/issues/362" }
gharchive/issue
Display and "save" the base image a notebook uses in a notebook cell We have a little first draft system for pairing a notebook with a base image that looks like this ... User says garden-ai notebook start foo.ipynb --base-image=3.10-torch We save a mapping from current/directory/foo.ipynb to 3.10-torch in the user's local data Then the user can say garden-ai notebook start foo.ipynb and we open up it up again in the right image. This is nice as a quick first pass but ofc it's pretty fragile. If the user mvs the notebook to a different directory they'll be prompted again to specify the base image. I suggest that when making a new notebook from a template we create a first cell that presents (in a nice user-readable format) what base image this notebook uses. If the user decides to change it later, that is well and good! They can just edit the cell. This has a few benefits We don't need a local DB mapping anymore The user can move the notebook around or even send it to another machine and the notebook keeps its own metadata. (Especially useful if we're debugging someone else's notebook) Assumptions: Acceptance Criteria Given..., when..., then... Now that we let users specify a --requirements file, that's something we can also think about in this ticket. The base image we provide and the requirements a user has specified feel pretty linked in the user's mind. So we'll want to think about those holistically.