added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T04:55:04.213544
| 2021-04-22T08:17:59
|
864676482
|
{
"authors": [
"OSBotify",
"isagoico",
"jasperhuangg",
"roryabraham",
"stitesExpensify"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13910",
"repo": "Expensify/Expensify.cash",
"url": "https://github.com/Expensify/Expensify.cash/pull/2532"
}
|
gharchive/pull-request
|
Emoji Picker Menu Navigation
Details
Added the ability to navigate through the emoji picker menu with arrow keys, and to send the highlighted emoji with [Enter]. It should function the same as the emoji picker in Slack.
the highlighted emoji should never be cutoff by the window
the first emoji is highlighted upon searching emojis
hovering over an emoji with your cursor will also change the currently highlighted emoji
if no emoji is highlighted, pressing an arrow key will highlight the first emoji
the cursor in the search input is not affected by arrow key presses that change the highlighted emoji
Fixed Issues
Fixes https://github.com/Expensify/Expensify.cash/issues/2450
Tests
I tested this feature on Web and Desktop since those are the only two places it can actually be used, but I also tested it on Mobile Web to ensure that the touchscreen check works and that the Emoji Picker Menu doesn't crash.
Opened the emoji picker. Scroll down till you can see the end of on category and the start of another.
Use your mouse to hover over an emoji near this boundary.
Use the up/down arrow keys to move the highlight around this border. Verified that it skips over the header.
Move the highlight at the top and bottom edges of the window with the arrow keys. Verified that the menu scrolls to keep the highlighted emoji in the window.
Type in search terms to filter by. Verified that the first search result is highlighted. Verified that nothing breaks when you try to go out of bounds. Verified that you can move the cursor using the appropriate arrow keys when they would move the highlight out of bounds.
Pressed [Enter]. Verified that the highlighted emoji was inputted into the compose box.
Repeat steps 1-6 two more times to verify that event listeners are removed and added correctly.
QA Steps
Identical to the tests above. See videos for examples.
Tested On
[x] Web
[ ] Mobile Web
[x] Desktop
[ ] iOS
[ ] Android
Screenshots
Web
https://user-images.githubusercontent.com/31285285/115820241-0048f200-a433-11eb-9dd8-d6c274158365.mp4
Desktop
https://user-images.githubusercontent.com/31285285/115819992-8ca6e500-a432-11eb-9322-4cabf9f107f3.mp4
cc @stitesExpensify
@pecanoro @stitesExpensify Thanks for clarifying my questions! Should be good for another review.
Tested and it works great! Just some code organization comments and then we should be good to go!
Hey @stitesExpensify thanks so much for the review! I've addressed most of your concerns but have a few clarifications of my own. Let me know what you think!
@stitesExpensify Ready for another review, thanks for your feedback!
🚀 Deployed to staging in version: 1.0.35-2🚀
platform
result
🤖 android 🤖
success ✅
🖥 desktop 🖥
success ✅
🍎 iOS 🍎
success ✅
🕸 web 🕸
success ✅
Emoji Picker - Moving by arrow keys will eventually lose selection
Expected result
Focus isn't lost even after reaching the last emoji on the list.
Actual result
After reaching the last emoji on the list, the focus is not visible anymore.
Action Performed
Log in to expensify.cash
Navigate to a conversation
Click on the emoji icon
Use the arrow keys to reach the bottom of the list
Platform
Web ✔️
Build: 1.0.36-0
Notes/Images/Video
https://user-images.githubusercontent.com/44479856/116945606-15e5c500-ac46-11eb-9e65-ce28fcc1e469.mp4
@jasperhuangg Not sure if the issue above should be a deploy blocker, let me know if I should open as a separate issue.
No need for this to be a deploy blocker, @isagoico, can you create a separate issue and tag @jasperhuangg?
|
2025-04-01T04:55:04.218596
| 2021-05-26T02:03:03
|
901628615
|
{
"authors": [
"Jag96",
"OSBotify",
"roryabraham"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13911",
"repo": "Expensify/Expensify.cash",
"url": "https://github.com/Expensify/Expensify.cash/pull/3149"
}
|
gharchive/pull-request
|
Make easily noticeable temporary change
Details
This PR is being used to test the CP workflow. It is "PR A", referenced in Step 5 of the testing plan here. This change will be reverted soon
Fixed Issues
Part of testing https://github.com/Expensify/Expensify.cash/pull/2978
Tests
Merge this PR
Verify that it gets the ✋ This PR will not be deployed ... comment
Verify that a staging deploy does not occur.
Verify that the red FAB is not present on staging.
Later complete steps 10-12 of the testing plan
QA Steps
Verify that the FAB is not red 😂
Skipping tests/actions to test CP PR flow
:hand: This PR was not deployed to staging yet because QA is ongoing. It will be automatically deployed to staging after the next production release.
Test steps 1-4 passed 👍
|
2025-04-01T04:55:04.221072
| 2024-03-29T13:43:44
|
2215394197
|
{
"authors": [
"BartoszGrajdek"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13912",
"repo": "Expensify/expensify-common",
"url": "https://github.com/Expensify/expensify-common/pull/674"
}
|
gharchive/pull-request
|
feat: remove emoji tag from data-code-raw in pre
Fixed Issues
$ GH_LINK
Tests
What unit/integration tests cover your change? What autoQA tests cover your change?
What tests did you perform that validates your changed worked?
QA
What does QA need to do to validate your changes?
What areas to they need to test for regressions?
I have read the CLA Document and I hereby sign the CLA
recheck
|
2025-04-01T04:55:04.235043
| 2020-10-30T10:25:53
|
733058692
|
{
"authors": [
"AlsoInteresting",
"Morph21"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13913",
"repo": "Exslims/MercuryTrade",
"url": "https://github.com/Exslims/MercuryTrade/issues/415"
}
|
gharchive/issue
|
sound source
How do I change sound source?
Whan I change headset in POE, this doesn't copy to Mercury Trade.
Even after restarting the pc.
When I put the old headset in the pc, it works.
MercuryTrade here is discontinued. Feel free to try a community fork version here: https://github.com/Morph21/MercuryTrade-Community-Fork
It's possible to change sounds.
Also different headset is not an issue on POE or MT side, check your system settings for solving sound issues
MercuryTrade here is discontinued. Feel free to try a community fork version here: https://github.com/Morph21/MercuryTrade-Community-Fork
It's possible to change sounds.
Also different headset is not an issue on POE or MT side, check your system settings for solving sound issues
|
2025-04-01T04:55:04.324994
| 2019-01-24T08:38:11
|
402596173
|
{
"authors": [
"Leylan24",
"Sonic0",
"lgr9603",
"tdh8316"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13914",
"repo": "F-Army/arduino-dw1000-ng",
"url": "https://github.com/F-Army/arduino-dw1000-ng/issues/139"
}
|
gharchive/issue
|
Example TwoWayRanging is not work.
Step 1: What do you want?
TwoWayRangingInitiator and TwoWayRangingResponder do not work. therefore I want them to work.
Step 2: Describe your environment
Arduino IDE version or Toolchain: Arduino IDE 1.8.7
Arduino device: Arduino Uno R3
Used adapter board: None
Step 3: Describe the problem
Affected file(s) or example(s):
TwoWayRangingResponder.ino
TwoWayRangingInitiator.ino
Observed Results:
It seems not to work!
Thus I tried Basic Connectivity Test.
I'm not sure it's correct output because I'm a beginner of Arduino and DWM1000.
But I don't think there is any noticeable problem.
And then I uploaded Basic Sender/Receiver to Arduino.
I still have no clue about this, but it seems to work fine.
EDIT: thotro/arduino-dw1000
I tried also thotro's library.
The device is removed immediately after it is added.
Why are BasicSender and Receiver working but TwoWayRanging not working?
Expected Results:
Jaja, I expected they work.
Wiring
Where did you put the 10K resistor ? I am having the same issue my Basic Receiver and Basic Sender is working but the Two Way Ranging example doesnt work
@lgr9603
Where did you put the 10K resistor ? I am having the same issue my Basic Receiver and Basic Sender is working but the Two Way Ranging example doesnt work
I think that the @tdh8316 's solution is relative to section 5.2.2 inside DWM1000 datasheet
@Leylan24 I'm a coworker of @tdh8316 .
Well, we put the 10K resistor on pin 2 to remove noises flowing through it.
And drop the voltage of digital pins from 5V to 3.3V. Dwm1000 can work between 2.4V and 3.75V. (Along to datasheet on decawaver.com)
As I said, arduino uno digital pins use 5V, bout DWM1000 sends output signals in 3.3V. So we should increase voltage of DWM1000 signal from 3.3V to 5V.
Thanks for you reply.
But does the ranges only print when the IRQ pin gets pulled high?
But does the ranges only print when the IRQ pin gets pulled high?
With this library, the IRQ pin is not required. There are two different init processes, with or without Interrupt pin.
|
2025-04-01T04:55:04.331748
| 2020-10-06T22:10:44
|
716055084
|
{
"authors": [
"dstokesf5",
"jmcalalang"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13915",
"repo": "F5Networks/f5-appsvcs-extension",
"url": "https://github.com/F5Networks/f5-appsvcs-extension/issues/357"
}
|
gharchive/issue
|
Ability to create DNS_Logging_Profile from AS3
Is your feature request related to a problem? Please describe.
Feature Request
Describe the solution you'd like
AS3 has the ability to reference DNS_Logging_Profile as a pre-created item, we would like the ability to create it. I don't see an option in the schema to create one.
"description": "Reference to a DNS logging profile",
"type": "object",
"properties": {
"bigip": {
"description": "Pathname of existing BIG-IP DNS logging profile",
"type": "string",
"format": "f5bigip"
}
},
Describe alternatives you've considered
There isn't an Ansible module for this either, it would require imperative calls to create the object.
REST endpoint: /mgmt/tm/ltm/profile/dns-logging
Options:
appService:
description:
enableQueryLogging:
enableResponseLogging:
includeCompleteAnswer:
includeQueryId:
includeSource:
includeTimestamp:
includeView:
logPublisher:
name:
Additional context
Object in question:
Thank you for your feedback. Due to the age of this feature request, I would like to know if this is still something you need or are interested in.
|
2025-04-01T04:55:04.336913
| 2024-01-02T04:32:21
|
2061931002
|
{
"authors": [
"dstokesf5",
"mdditt2000",
"normelton"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13916",
"repo": "F5Networks/f5-appsvcs-extension",
"url": "https://github.com/F5Networks/f5-appsvcs-extension/issues/790"
}
|
gharchive/issue
|
Unable to compare SSL SNI/hostname against data group
Environment
Application Services Version: 3.48.0
BIG-IP Version: <IP_ADDRESS>
Summary
When building a policy, I can compare the HTTP hostname against a data group, but I cannot compare the SSL SNI hostname against the same data group. I get the following error
Steps To Reproduce
Submit the following policy:
"http-redirect": {
"class": "Endpoint_Policy",
"strategy": "first-match",
"rules": [
{
"name": "permit",
"conditions": [{
"type": "httpHost",
"event": "request",
"host": {
"operand": "ends-with",
"datagroup": {
"use": "permitted-domains"
}
}
}]
}
]
}
It works fine.
Submit this policy instead:
"https-block": {
"class": "Endpoint_Policy",
"strategy": "first-match",
"rules": [
{
"name": "permit",
"conditions": [{
"type": "sslExtension",
"event": "ssl-client-hello",
"serverName": {
"operand": "ends-with",
"datagroup": {
"use": "permitted-domains"
}
}
}]
}
]
}
It fails to post, I get the following error:
{
"code": 422,
"message": "declaration failed",
"response": "01020036:3: The requested datagroup lookup failure: (use) was not found.",
"host": "localhost",
"tenant": "init-policy",
"runTime": 2365
},
it looks like two things are changing from your working and not working examples. If the permitted-domains data group is managed by AS3, can you try updating your failing example to use a use pointer instead of a bigip pointer?
Good observation. In all my testing, I copied/pasted the wrong example. The behavior is the same when with a use pointer. I updated my post above.
Add Jira AUTOTOOL-4132 to next sprint for prioritization discussion
|
2025-04-01T04:55:04.341988
| 2021-08-24T20:54:45
|
978478155
|
{
"authors": [
"amolari",
"dstokesf5"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13917",
"repo": "F5Networks/f5-telemetry-streaming",
"url": "https://github.com/F5Networks/f5-telemetry-streaming/issues/161"
}
|
gharchive/issue
|
Datadog: add customizable service field option
Is your feature request related to a problem? Please describe.
The service field option, used by Datadog logging, is not customizable but statically defined in the code.
const DATA_DOG_SERVICE_FIELD = 'f5-telemetry'
Describe the solution you'd like
Be able to set a custom value in the Datadog consumer declaration.
Thank you for your feedback. I have added your request to our internal product backlog as AUTOTOOL-2753.
I guess it had been fixed in v1.24.0?
This has been resolved in the 1.24.0 release.
|
2025-04-01T04:55:04.374534
| 2023-09-12T12:56:42
|
1892466057
|
{
"authors": [
"coveralls",
"domna",
"sherjeelshabih"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13918",
"repo": "FAIRmat-NFDI/pynxtools",
"url": "https://github.com/FAIRmat-NFDI/pynxtools/pull/157"
}
|
gharchive/pull-request
|
Adds passing of objects to transmission reader
This adds the functionality of passing a dict to the transmission readers convert function.
@hampusnasstrom
So you pass Dict type objects and append them to the template? No other object types are accepted currently, right?
Pull Request Test Coverage Report for Build<PHONE_NUMBER>
10 of 17 (58.82%) changed or added relevant lines in 3 files are covered.
4 unchanged lines in 3 files lost coverage.
Overall coverage decreased (-0.005%) to 51.087%
Changes Missing Coverage
Covered Lines
Changed/Added Lines
%
pynxtools/dataconverter/readers/utils.py
6
13
46.15%
Files with Coverage Reduction
New Missed Lines
%
pynxtools/dataconverter/readers/shared/map_concepts/mapping_functors.py
1
11.86%
pynxtools/nyaml2nxdl/nyaml2nxdl_forward_tools.py
1
79.42%
pynxtools/nexus/nxdl_utils.py
2
74.79%
Totals
Change from base Build<PHONE_NUMBER>:
-0.005%
Covered Lines:
5993
Relevant Lines:
11731
💛 - Coveralls
So you pass Dict type objects and append them to the template? No other object types are accepted currently, right?
Exactly, I expect an iterable of dicts. Every other element is simply warned about and ignored
|
2025-04-01T04:55:04.435840
| 2019-11-07T21:43:25
|
519530317
|
{
"authors": [
"Robadob"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13919",
"repo": "FLAMEGPU/FLAMEGPU2_dev",
"url": "https://github.com/FLAMEGPU/FLAMEGPU2_dev/pull/130"
}
|
gharchive/pull-request
|
Feature: Host Functions
This pull request/branch tracks the development of host functions.
Example documentation: https://gist.github.com/Robadob/081d04d2d515ce3b090bf389dc9195dc
This addresses many issues, as such the branch may be long running and merged incrementally:
PR #135
[x] #46 - Host functions init/step/exit bd5097c
[x] #54 - Exit early from step function bd5097c
[x] #47 - Host layer functions (step function in any layer) 7c0c204
[x] #53 - CUB reductions (min, max, sum, count) 046fa07
[x] #80 - Histogram reduction 09ac715
[x] #83 - Custom reduction functions (cub reduce, thrust transformReduce) ffa28c9
PR: #138
[x] #90 - Constant/Immutable Environmental variables
PR: #196
[x] #61 - Agent creation on the host
Misc/Low Priority
[ ] #162 - Large environment properties (held in global memory)
[ ] #51 - 'Access agent data on the host'????
TODO:
[x] Rename FLAMEGPU_API->FLAMEGPU_DEVICE_API
Host function definition
[x] Host Fn
[x] Host Condition
Add Host fn/condition to model
[x] Init
[x] Step
[x] Exit
[x] Layer
[x] ExitCondition
gpu/CUDAAgentModel (Controls execution of model)
step(): Update to support host functions
[x] Init
[x] Step
[x] Exit
[x] HostLayer
[x] ExitCondition
Environmental vars
[x] Define
[x] Read in device fn
Runtime/flamegpu_host_api.h
Environmental vars
[x] Read on host fn
[x] Write in host fn
[x] Wrapper for <random>
Reduction
[x] Sum
[x] Min
[x] Max
[x] Reduce (user defined template binary op a+b->c)
Others from thrust?
[x] Count
[x] TransformReduce
Message reductions?
Histogram
[x] Even
Custom bounds? (This would require having to cache their bounds on device somehow)
[x] Agent Creation
[x] Tests
[x] Example usage in model
TODO Monday:
Custom reduction fn
\<random> wrapper using runtime seed.
This PR has served it's purpose all parity tasks complete.
|
2025-04-01T04:55:04.441736
| 2024-03-20T21:10:03
|
2198574482
|
{
"authors": [
"FMaz008",
"crysallis",
"runew0lf"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13920",
"repo": "FMaz008/VineHelper",
"url": "https://github.com/FMaz008/VineHelper/issues/58"
}
|
gharchive/issue
|
Request: Option to toggle the "green styling"
It makes it very hard to read with darkmode on
As an "older gentleman" it kinda all blurs into one colour!
The main problem is the text color, which we will improve upon. But we have been working toward centralizing styling colors so it might be possible in the future (not yet) to have themes.
Baby steps...
... but for now we'll fix the text which is badly colored as you showed in your screenshots.
Thank you for giving feedback, some of those area haf not occured to me, so this is definitely helpful.
I'll address this if you want
Fixed.
|
2025-04-01T04:55:04.557982
| 2020-06-12T16:58:00
|
637888270
|
{
"authors": [
"JamesWP",
"eurocoinsalbum"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13921",
"repo": "FRUK-Simulator/Simulator",
"url": "https://github.com/FRUK-Simulator/Simulator/issues/37"
}
|
gharchive/issue
|
create robot control react element to show the current state of the robot
sensor data, motor output etc..., position...
I implemented this already. Will add a PR today
#46 has an extremely simple version of this, only contains the motor power for now, but ill close this issue.
|
2025-04-01T04:55:04.635509
| 2022-05-29T07:57:55
|
1251818321
|
{
"authors": [
"Fabian42"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13922",
"repo": "Fabian42/bash_scripts",
"url": "https://github.com/Fabian42/bash_scripts/issues/44"
}
|
gharchive/issue
|
Automatically download new SoundCloud tracks
The text in the followthatpage mails might be enough to automatically download new SoundCloud tracks from artists I follow with that tool. The hardest part of that might be reading the mail using Bash, but surely there's a tool for that.
Test cases would have to include tracks with the same name as something that that artist has already uploaded before, because then they get an attached "-2" or similar in the URL, but it might also be slightly different. Checking the upload date and comparing that to the mail should definitely be done. Maybe that's also enough and the actual content of the mail can be ignored, just knowing that someone uploaded and then downloading the last week of their tracks would be OK.
Filtering out false positives like in the mail macro would also be necessary for as long as that followthatpage bug exists.
Alternatively, there's something with RSS on SoundCloud, which would be a less hacky solution (RSS→dl instead of website→follow→mail→guess link→verify date→dl). But I never got any of those RSS feeds to work so far.
alternative: query pages myself, can also apply to more than SC
auto-download WL and temp_music
similar: webcomics, Lateral
|
2025-04-01T04:55:04.777566
| 2023-06-07T03:55:09
|
1744981636
|
{
"authors": [
"chris-t-li"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13923",
"repo": "Fantasy-Fit/fantasy-fit-web",
"url": "https://github.com/Fantasy-Fit/fantasy-fit-web/pull/173"
}
|
gharchive/pull-request
|
refactored dockerfiles to reduce image sizes
Update Ruby to 3.2.2;
Add redis and helper services to docker-compose;
Refactored dockerfile for client - use node:18-alpine and multi stage build to reduce image size
Refactored dockerfile for api - use ruby:3.2.2-slim-bullseye and update dependencies to reduce image size
Time to build ~ 122s
Image sizes: client reduced from 1.27GB to 309MB. rails api reduced from 1.07GB to 678MB
This addresses issue #172
Significantly reduced time to stop containers: from ~10+ seconds for client to 0.4s
|
2025-04-01T04:55:04.859085
| 2020-07-28T13:00:52
|
667053734
|
{
"authors": [
"NareshNama",
"bell-steven"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13924",
"repo": "FaridSafi/react-native-google-places-autocomplete",
"url": "https://github.com/FaridSafi/react-native-google-places-autocomplete/issues/595"
}
|
gharchive/issue
|
onChangeText is not working
<GoogleAutoCompleteAddress
onPress={(data, details = null) => {
this.fetchGoogleLocationDetails(details);
}}
textInputProps={{
value:lossAddress,
onChangeText:{_onChangeTextValue()}
}}
editable={true}
autoFocus={false} />
onChangeText is not working and showing syntax error
Please provide a reproducible example.
|
2025-04-01T04:55:04.862074
| 2017-04-03T16:26:44
|
218989143
|
{
"authors": [
"ConnorRigby",
"Daniil145",
"RickCarlino"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13925",
"repo": "FarmBot/farmbot_os",
"url": "https://github.com/FarmBot/farmbot_os/issues/282"
}
|
gharchive/issue
|
End stops don't work on after new release
my.farmbot.io
You disable endstops?
@Daniil145 sorry about that, they werent disabled, but there was a new version of teh Arduino Firmware released.
Could you try disabling and re enabling end stops?
@ConnorRigby I dont understand "disabling and re enabling end stops" on interface dont have toggle off and on Endpoints. What can I do to activate them? With the loss of the Internet (especially when I use mobile Internet) the system behaves unpredictably. Especially this is critical for the Z axis. To do this, I need limit switches.
Oh sorry i didnt realize they werent on the interface any more.
@RickCarlino any thoughts on this one?
@Daniil145 I've recently added endstop toggle buttons to the device panel. They will be available in the next production release (probably this week).
@Daniil145 We just did a production deploy. Please let us know if you find any issues.
|
2025-04-01T04:55:04.972977
| 2023-03-06T17:59:35
|
1611946418
|
{
"authors": [
"rachithrr",
"vkrishnanfb"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13926",
"repo": "FeatureBaseDB/featurebase",
"url": "https://github.com/FeatureBaseDB/featurebase/pull/2296"
}
|
gharchive/pull-request
|
FB-1895: Implement DateTimeFromParts
this uses the tiem.date() function
Well, I have to add limits to the time ranges. For example, dax expects the year range [0 , 9999]. I'll update the code and re request to review.
Looks good to me.
|
2025-04-01T04:55:04.997579
| 2022-06-23T18:18:55
|
1282777484
|
{
"authors": [
"FeiyouG",
"bugabinga"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13928",
"repo": "FeiyouG/command_center.nvim",
"url": "https://github.com/FeiyouG/command_center.nvim/issues/15"
}
|
gharchive/issue
|
Error if nothing was selected
Open Command Center
Type some gibberish until the filtered list list empty
Hit <ENTER>
Alternatively, open Command Center without and bindings added and hit <ENTER>.
E5108: Error executing lua ...mand_center/lua/telescope/_extensions/command_center.lua:151: attempt to index local 'selection' (a nil value)
stack traceback:
...mand_center/lua/telescope/_extensions/command_center.lua:151: in function 'run_replace_or_original'
...ite/pack/paqs/opt/telescope/lua/telescope/actions/mt.lua:65: in function 'key_func'
.../site/pack/paqs/opt/telescope/lua/telescope/mappings.lua:242: in function 'execute_keymap'
[string ":lua"]:1: in main chunk
Telescope + CC config:
telescope.setup {
extensions = {
command_center = {
-- Change what to show on telescope prompt and in which order
-- Currently support the following three components
-- Components may repeat
components = {
command_center.component.DESCRIPTION,
command_center.component.KEYBINDINGS,
-- command_center.component.COMMAND,
},
-- Change the separator used to separate each component
separator = " ",
-- When set to false,
-- The description compoenent will be empty if it is not specified
auto_replace_desc_with_cmd = false,
}
}
}
-- Plug Command Center into Telescope
telescope.load_extension('command_center')
Fixed in e3b32c7a2b8ffbb92bd4b7158fcb44eb40ba5d20
|
2025-04-01T04:55:05.009584
| 2021-01-29T17:09:32
|
797044762
|
{
"authors": [
"Femme-js",
"Swastik1710"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13931",
"repo": "Femme-js/Hoaxify",
"url": "https://github.com/Femme-js/Hoaxify/issues/34"
}
|
gharchive/issue
|
Adding a seperate line for admin in readme.md file and creating a Contributors table which updates automatically.
As of now , the Contributors table in the readme.md file is not up to date. Moreover one needs to update it manually each and every time. I wanna automate this by adding a new Contributors table which will update by itself.
Moreover , I wanna add a separate line to showcase the Project Admin if you want.
Can you please assign this to me ?
Thanks.
Can you please assign this to me ?
Thanks.
Great!
Hey @Femme-js , I've made a PR.
Please have a look,
|
2025-04-01T04:55:05.026214
| 2024-12-27T22:20:46
|
2761313287
|
{
"authors": [
"lesniakt",
"smturro2"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13932",
"repo": "Fernando-Urbano/portfolio-management",
"url": "https://github.com/Fernando-Urbano/portfolio-management/issues/36"
}
|
gharchive/issue
|
Considering a Environment Manager
@lesniakt Do you have any opinions/thoughts on our environment management strategy? Now would be a good time to switch rather then later.
I would like to stick with the anaconda environment and rename the environment so it doesn't overlap with the standard FINM environment used in classes and august review. This approach would be very familiar to new members coming from the FINM program. When we have tests we would build the package locally using pip install . -e and test would be run through a build instead of importing it from the folder over.
Using pipenv will make it easier to upgrade package versions. We can also set this up so it automatically builds and installs the local package. I've worked with this a ton and is great. The downside being that the setup is different making it less friendly to new contributers in the FINM department.
If possible we should prioritize making ease of development for new contributers. I don't have a ton of experience upgrading an anaconda env so please add some thoughts if you have. I imagine we would simply reinstall the whole anaconda env (without version locks) and then update the requirements files.
I've personally not used pipenv. Typically for quick and dirty projects venv is sufficient, and for larger scale projects, especially production projects, I tend to see Conda used as usually there are non-python dependencies involved.
From the literature I've read, it seems like non-python dependencies is where pipenv would struggle. Numpy and scipy come to mind as potentially requiring non-python dependencies -- but again I haven't played with it. I would be curious how reproducible pipenv is with those, as this package will presumable use those dependencies,
I think the typical student in this program is just editing from requirements.txt files, so I share your concern on simplicity here.
My lean would be to just use the existing requirements.txt workflow until we have some level of code coverage. I think that is a reasonable prerequisite to have before changing environment management workflows, as we'll want reasonable certainty that the changes don't break workflows accidentally.
Regarding "ease of use" for upgrading packages, I would suggest looking at integration with RenovateBot. A quick looks says both conda/pipenv are supported. Again, we'd want code coverage in place before doing this.
|
2025-04-01T04:55:05.029201
| 2023-03-27T20:20:35
|
1642758031
|
{
"authors": [
"AlekSi",
"b1ron"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13933",
"repo": "FerretDB/FerretDB",
"url": "https://github.com/FerretDB/FerretDB/issues/2299"
}
|
gharchive/issue
|
Projection with _id returns all fields
Versions
commit.txt: a06bd4bf1832bceed7189cff07d788cdb11e2155
version.txt: v0.9.4-1-ga06bd4bf
What did you do?
See repro in #2298.
What did you expect to see?
2 documents returned with just the _id field.
What did you see instead?
All fields.
Environment
- OS: Ubuntu
- Architecture: x86_64
- Version: 22.10
- Go: go version go1.20 linux/amd64
- Git: 2.37.2
- Deployment: Docker
- Deployment details: 23.0.1
See repro in https://github.com/FerretDB/FerretDB/pull/2298.
That PR should be re-opened and updated to check that this issue is indeed fixed
|
2025-04-01T04:55:05.033055
| 2023-03-28T08:31:42
|
1643459807
|
{
"authors": [
"AlekSi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13934",
"repo": "FerretDB/FerretDB",
"url": "https://github.com/FerretDB/FerretDB/pull/2301"
}
|
gharchive/pull-request
|
Use iterators for projections
After #2254.
Description
Closes #{issue_number}.
Readiness checklist
[ ] I added/updated unit tests.
[ ] I added/updated integration/compatibility tests.
[ ] I added/updated comments and checked rendering.
[ ] I made spot refactorings.
[ ] I updated user documentation.
[ ] I ran task all, and it passed.
[ ] I ensured that PR title is good enough for the changelog.
[ ] (for maintainers only) I set Reviewers (@FerretDB/core), Assignee, Labels, Project and project's Sprint fields.
[ ] I marked all done items in this checklist.
Merged into #2254.
|
2025-04-01T04:55:05.036532
| 2023-05-30T13:27:15
|
1732279705
|
{
"authors": [
"rumyantseva"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13935",
"repo": "FerretDB/FerretDB",
"url": "https://github.com/FerretDB/FerretDB/pull/2739"
}
|
gharchive/pull-request
|
Make testFindAndModifyCompat work with compatTestCaseResultType
Description
Closes #2734.
Readiness checklist
[ ] I added/updated unit tests.
[x] I added/updated integration/compatibility tests.
[ ] I added/updated comments and checked rendering.
[x] I made spot refactorings.
[ ] I updated user documentation.
[ ] I ran task all, and it passed.
[x] I ensured that PR title is good enough for the changelog.
[x] (for maintainers only) I set Reviewers (@FerretDB/core), Labels, Project and project's Sprint fields.
[x] I marked all done items in this checklist.
@w84thesun ah, thanks for checking! I haven't pushed the code 😅
|
2025-04-01T04:55:05.043431
| 2020-06-16T13:38:23
|
639673110
|
{
"authors": [
"Sc0r",
"klightspeed"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13936",
"repo": "FezVrasta/ark-server-tools",
"url": "https://github.com/FezVrasta/ark-server-tools/issues/1112"
}
|
gharchive/issue
|
Please ensure the server is running the latest version of all mod content
I get this issue whenever I try to join my server.
"Please ensure the server you are trying to join is running the latest version of all mod content"
Why?
~$ arkmanager checkmodupdate --updatemods @crystal
Running command 'checkmodupdate' for instance 'crystal'
All mods are up to date
EDIT
huh, now it works without anything changing. odd.
It could have been that the client hadn't finished updating its mods. Since mods do not have any inherent version, there is no way to tell whether the server or client has more up-to-date mods.
|
2025-04-01T04:55:05.045342
| 2018-12-13T14:39:28
|
390708050
|
{
"authors": [
"juanjoDiaz"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13937",
"repo": "FidelLimited/serverless-plugin-warmup",
"url": "https://github.com/FidelLimited/serverless-plugin-warmup/pull/86"
}
|
gharchive/pull-request
|
Minor improvements
Just a bit of cleaning up.
Avoid doing the same job (loading the config) twice and move all the checks to top-level hooks.
I noticed that all the possible variable including the absolute path of the project where being added to the generated lambda so the last commit cleans that so only necessary variables are included.
|
2025-04-01T04:55:05.046372
| 2021-03-20T18:11:39
|
836877455
|
{
"authors": [
"AlexLanzano"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13938",
"repo": "Field-Programmable-Smartwatch/firmware",
"url": "https://github.com/Field-Programmable-Smartwatch/firmware/issues/7"
}
|
gharchive/issue
|
Implement log levels
Implement different log levels (Error, debug, info, etc).
Have log messages be able to be hidden or shown based on build environment variable or changed via a serial user interface
Implemented in a3bf35b2de76ca4b34aaf711bda04b4a4842dd4e
|
2025-04-01T04:55:05.058290
| 2020-07-31T01:14:08
|
669317228
|
{
"authors": [
"joshmorel",
"jyclam",
"kterry1",
"manualzuru"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13939",
"repo": "FightPandemics/FightPandemics",
"url": "https://github.com/FightPandemics/FightPandemics/issues/1306"
}
|
gharchive/issue
|
Refactor and remove OrgProfileComplete page, instead use the ProfileCompleted page
The call to action button should say View Help Board unless the unlogged in user clicked on Create a Post and went through the sign up process, in which case it should say Continue posting
Currently the pages look like this:
OrgProfileComplete
ProfileCompleted
@jyclam I'll take this on, but I wanted to clarify the ask here. It says "Refactor and remove OrgProfileComplete page". Am I refactoring OrgProfileComplete, or completely deleting it and just using "ProfileCompleted" where "OrgProfileComplete" is used?
Didn't mean to close. Let me know your thoughts on the above. I wanted to make sure I know what is wanted here before proceeding.
Also, how can i differentiate whether the user went through "Create Post" and then signed up, or whether they signed up the usual way.?
Hi @kterry1, from my understanding, there are 2 different pages (UI wise) that are trying to accomplish the same function. The main difference is the 2nd CTA changes depending on the path that you went through. With the ProfileCompleted page, we should be able to do everything that is needed.
There are 3 paths with the last CTA that changes.
1- if you click in register -> Check the Help Board
2- if you try to make a comment + like and you aren't signed in. -> Check the Post (the post that you were liking or commenting)
3- if you click on Create Post and you aren't signed in. -> Continue Posting
Let me know if there are other questions🙏🏽
Thanks @manualzuru! I'll dig into it today and see what I can do.
@jyclam @manualzuru So I did some testing with this. Since the Sign In, Create a post, and Join Now buttons all take you to '/auth/signup', I'm not seeing a way to conditionally render based on the path the user took to get to '/profile-completed' or '/create-organisation-complete' and have it all happen in one page. If the wording of that button is really important, then I'd say we could stick with multiple pages, otherwise, here is what I had so far, but I don't think it really accomplishes what is asked for: https://share.getcloudapp.com/OAuqOwG0
Maybe @joshmorel or @idjevm have an idea?
Side note: We need to change all instances of "organisation" in the codebase to "organization".
You can probably do a conditional sessionStorage like here https://github.com/FightPandemics/FightPandemics/blob/487e856300292dc7e754be23134b4025d3dfdb82/client/src/pages/PostPage.js#L162
https://developer.mozilla.org/en-US/docs/Web/API/Window/sessionStorage
I'll give that a try. Thanks!
The fix is based on jyclam's original post as it made the most sense.
@kterry1 @jyclam Sorry I didn't take a look at this early when you tagged me - since the PR is in.
I'm really apprehensive about using the same file/component for different routes. It can result in a hard to maintain situation as I see with our "Login.js" file - see #1164 (that one is more complex, so not exactly comparable but similar principle).
I think it would be better to have more re-usable components for similar pages but use one top-level file per route. You see this with a lot of filesystem based frameworks like vuex/next.js etc and I think it's a good approach.
With this PR #1328 - we need to fix the issues - but we should decide:
Make an exception here - and address the issues as described in my first comment in the PR
Or come up with a rule and stick to it - one file per route
Thoughts @robinv85 @idjevm ?
Also I did try to figure out what the design's flow is supposed to be after create org complete in Figma but it wasn't clear (https://www.figma.com/file/kT6SJLL4v3w06yFF2cGIun/FightPandemics-MVP?node-id=1852%3A8716). Is "View my profile" always supposed to go to MY profile as opposed to the recently created organisation's profile (current state). The latter makes more sense but the design doesn't indicate which. But also in such a case I'd expect the label to be different "View organisation" instead.
Maybe the design was incomplete after copy/paste from user profile? Any insight into this @manualzuru
|
2025-04-01T04:55:05.062251
| 2021-05-25T21:27:46
|
901390178
|
{
"authors": [
"adorapham",
"hiran11",
"xJamesScott"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13940",
"repo": "FightPandemics/FightPandemics",
"url": "https://github.com/FightPandemics/FightPandemics/issues/2398"
}
|
gharchive/issue
|
User Store 9.1-Unable to set the role for a member after accepting the applicant
Unable to set the role for a member after accepting the applicant.
please see the record on , near the end starting from 1:30
https://app.slack.com/client/T010BHVS3EF/C010BHVT1MZ/files/F022X4NME85
@adorapham Bug fix issued
@ChunPingYang @xJamesScott Looks like the issue is not fixed .
As an individual user even after i join the Org "Join Us" CTA button is not disabled .
Even when my application as an individual has been accepted "Join Us " CTA button is kept enabled. The individual name is listed as member in Members Tab.
As a owner of the account , unable to set the role permission.
The static data is populated in the Role description for Wiki editor and Administrator.
https://user-images.githubusercontent.com/51660881/120036342-01d77e00-bfce-11eb-88e5-8639e007f2f9.mp4
https://user-images.githubusercontent.com/51660881/120036352-04d26e80-bfce-11eb-946f-2ca737365fdf.mp4
https://user-images.githubusercontent.com/51660881/120036373-0c921300-bfce-11eb-8856-7310b676ecfc.mp4
https://user-images.githubusercontent.com/51660881/120036393-1287f400-bfce-11eb-9b03-2fa09c872adb.mp4
|
2025-04-01T04:55:05.071341
| 2020-04-03T21:09:38
|
593642449
|
{
"authors": [
"CeesWang",
"MohamedEl1",
"T-Allen",
"indirabhowmick",
"manualzuru",
"robinv85"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13941",
"repo": "FightPandemics/FightPandemics",
"url": "https://github.com/FightPandemics/FightPandemics/issues/50"
}
|
gharchive/issue
|
User Profile - Front End
[ ] frontend user profile
Here is the Requirements for Create Individual Profile UI
https://www.notion.so/fightpandemics/Create-Individual-Profile-UI-cf719e2292874303a2ce943c912bdbf3
Finished styling the user profile,
created edit-profile / edit-account routes and components
wip
styling on edit-profile and edit-account components
handle data flow for my activity in user profile
@CeesWang - any updates on this issue?
@indirabhowmick talking to T-Allen regarding further progress.
@CeesWang Thanks for the update. Please keep us posted on the progress of the issue
@CeesWang @T-Allen is this solved?
@manualzuru is this the same as issue #176 ?
It is the same mobile and desktop only styling changes depending on the device @T-Allen
I'm still confused, because going by what's on Zeplin, the pages seem quite different:
create-profile (desktop)
create-profile (mobile)
edit-profile (desktop)
edit-account-info (desktop)
@T-Allen by Mobile we meant mobile web not the native mobile app. Please reach out if there are any questions.
@indirabhowmick Right, I am working on the web app.
@T-Allen you are right. Those are 2 different screens, you are right. We need both for having profile.
Can we update this ticket or add another ticket and maybe tags to have a better view on the user profile status?
The user v2 data model is not implemented yet and I'm working on "create user" so I think both of these edit / view user aren't implemented with API yet?
No they are not implemented with API yet
On Tue, May 26, 2020 at 5:27 PM Robin Venneman<EMAIL_ADDRESS>wrote:
Can we update this ticket or add another ticket and maybe tags to have a
better view on the user profile status?
The user v2 data model is not implemented yet and I'm working on "create
user" so I think both of these edit / view user aren't implemented with API
yet?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/FightPandemics/FightPandemics/issues/50#issuecomment-634288870,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AIAFWLSDD5OCCSNYDMC6JBLRTQX5DANCNFSM4L4OHU2Q
.
@CeesWang i'm working on this to finish some details and integrate with API
@robinv85 awesome let me know if theres anything I can do regarding this issue
|
2025-04-01T04:55:05.079230
| 2017-01-08T20:24:13
|
199443931
|
{
"authors": [
"FiloSottile",
"zmanian"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13942",
"repo": "FiloSottile/zcash-mini",
"url": "https://github.com/FiloSottile/zcash-mini/issues/1"
}
|
gharchive/issue
|
Should also export a viewing key
The Zcash protocol also has support for view keys. Views keys can determine if a tx has been received but not spend them.
I don't believe there is any support for view keys in the full node wallet yet but we could generate and export a view key that could immediately be imported by a user's full node.
Ah wonderful, I thought they didn't exist and wished they did.
Yep, let's add them, and add a mode to input the secret key instead of generating it, to get address/view key from a saved key.
|
2025-04-01T04:55:05.080272
| 2024-12-28T08:10:41
|
2761624014
|
{
"authors": [
"FinalForEach",
"scoundri"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13943",
"repo": "FinalForEach/Cosmic-Reach-Localization",
"url": "https://github.com/FinalForEach/Cosmic-Reach-Localization/pull/273"
}
|
gharchive/pull-request
|
Fixed Grammar Issues and made it clearer and easier to read. Turkish
I hate Polite suffixes
Thank you! Merging now.
|
2025-04-01T04:55:05.097419
| 2017-05-11T14:23:15
|
228009262
|
{
"authors": [
"JoyceBabu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13944",
"repo": "Financial-Times/polyfill-service",
"url": "https://github.com/Financial-Times/polyfill-service/issues/1200"
}
|
gharchive/issue
|
Element polyfill fails on UCBrowser Mini
The Element polyfill fails on UCBrowser Mini because
document node cannot have more than one child
The polyfill uses attachEvent. The event onpropertychange proprietary to IE. Irrespective of that, using addEventListener will prevent it from failing completely on other browsers that require Element polyfill.
createElement('*') is used, which is not valid.
Closing this as per https://github.com/Financial-Times/polyfill-service/pull/1201
|
2025-04-01T04:55:05.110659
| 2024-03-07T07:09:08
|
2173126675
|
{
"authors": [
"codecov-commenter",
"zemyblue"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13945",
"repo": "Finschia/finschia-proxy",
"url": "https://github.com/Finschia/finschia-proxy/pull/35"
}
|
gharchive/pull-request
|
Revert "feat: apply the changes of Finschia v3.0.0-rc1"
Reverts Finschia/finschia-proxy#34
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Project coverage is 55.45%. Comparing base (2961479) to head (d2b5672).
Additional details and impacted files
@@ Coverage Diff @@
## main #35 +/- ##
=======================================
Coverage 55.45% 55.45%
=======================================
Files 12 12
Lines 1502 1502
=======================================
Hits 833 833
Misses 616 616
Partials 53 53
Files
Coverage Δ
ante/ante.go
62.50% <ø> (ø)
ante/txfilter.go
85.93% <ø> (ø)
app/ante.go
0.00% <ø> (ø)
app/app.go
83.33% <ø> (ø)
app/encoding.go
100.00% <ø> (ø)
cmd/fnsad-proxy/cmd/root.go
34.20% <ø> (ø)
|
2025-04-01T04:55:05.114209
| 2022-02-21T08:13:19
|
1145457713
|
{
"authors": [
"EmmanyTaskWorld",
"Fintasys",
"tzthantzin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13946",
"repo": "Fintasys/emoji_picker_flutter",
"url": "https://github.com/Fintasys/emoji_picker_flutter/issues/57"
}
|
gharchive/issue
|
How can I get unify code from emoji?
How can I get unify code from emoji?
I click 'grinning face', its unified code is 1F600.
How can I get that value from emoji picker?
https://unicode.org/emoji/charts-13.0/full-emoji-list.html
Thank you very much
A emoji consists of runes. So in dart the runes are from type Integer and you can easily format them to hex.
EmojiPicker(
onEmojiSelected: (Category category, Emoji emoji) {
emoji.emoji.runes.forEach((element) {
print(element.toRadixString(16));
});
}
...
)
I hope that helps you !
Thank you very much @Fintasys .
It is working
Thank you very much 🙏
When i select country flag..using emoji.emoji.runes.forEach((element) { print(element.toRadixString(16)); // 1F600 }); doesn't seem to work as country flag emoji unicode should be in this form (U+1F1E6-1F1EE)
|
2025-04-01T04:55:05.146857
| 2021-01-22T09:47:59
|
791858601
|
{
"authors": [
"YannicEl",
"dackers86",
"mbleigh"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13947",
"repo": "FirebaseExtended/experimental-extensions",
"url": "https://github.com/FirebaseExtended/experimental-extensions/issues/31"
}
|
gharchive/issue
|
[firestore-scheduled-writes] Error logs appear even with successful writes
Expected Behavior
Logging should be successful and not contain error logs when a successful rite has been made.
Actual Behavior
An error log appears, stating the document cannot be found. The extensions appear to have worked as expected.
Steps to Reproduce the Problem
Follow standard directions for installation in README.
Run the example code snippet
const TEN_MINUTES_MS = 10 * 60 * 1000;
firebase
.firestore()
.collection("queued_writes")
.add({
state: "PENDING",
data: { message: "Hello from the future!" },
deliverTime: firebase.firestore.Timestamp.fromMillis(
Date.now() + TEN_MINUTES_MS
)
});
What happened? How can we make the problem occur?

Specifications
Extension name: firestore-schedule-writes
Version: 0.0.1
Is this being worked on?
If not I could issue a PR.
PR would be great! I didn't notice this issue until just now
On Fri, Jun 4, 2021, 6:43 AM YannicEl @.***> wrote:
Is this being worked on?
If not I could issue a PR.
—
You are receiving this because you were assigned.
Reply to this email directly, view it on GitHub
https://github.com/FirebaseExtended/experimental-extensions/issues/31#issuecomment-854733896,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AAAAH7S4CDQ365E2STOJ7E3TRDKANANCNFSM4WOHEFWA
.
the PR is ready
|
2025-04-01T04:55:05.156136
| 2020-03-26T18:51:13
|
588650509
|
{
"authors": [
"TahaTesser",
"aantelov87"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13948",
"repo": "FirebaseExtended/flutterfire",
"url": "https://github.com/FirebaseExtended/flutterfire/issues/2242"
}
|
gharchive/issue
|
[firebase_ml_vision] PDF417 reader is not working for DL
Trying to read a PDF417 barcode for a US DL using the new barcode library instead of the one from Google Mobile Vision but your lib is not properly reading and camera is unusable for Samsung S8 when resolution is set to ResolutionPreset.high.
Any ideas how can we use this feature from the new API?
I am testing with the example but I am not getting any good results.
Thank you
Hi @aantelov87
can you please provide your flutter doctor -v and flutter run --verbose?
Are you using official firebase_ml_vision example?
Thank you
Hi @TahaTesser
Yes, I am using the official firebase_ml_vision example but I just changed the following line
https://github.com/FirebaseExtended/flutterfire/blob/master/packages/firebase_ml_vision/example/lib/material_barcode_scanner.dart#L156
with
final ResolutionPreset preset = ResolutionPreset.high;
because with the resolution used in the official example, the detector is not able to detect PDF417 barcode for DL.
In the attachment flutter run --verbose when the resolution is set to high for PDF417 detection.
Please, let me know if you need anything else.
`flutter doctor -v
[✓] Flutter (Channel stable, v1.12.13+hotfix.8, on Mac OS X 10.15.3 19D76, locale en-US)
• Flutter version 1.12.13+hotfix.8 at /Users/aantelov/GPN/flutter
• Framework revision 0b8abb4724 (6 weeks ago), 2020-02-11 11:44:36 -0800
• Engine revision e1e6ced81d
• Dart version 2.7.0
[✓] Android toolchain - develop for Android devices (Android SDK version 29.0.3)
• Android SDK at /Users/aantelov/Library/Android/sdk
• Android NDK location not configured (optional; useful for native profiling support)
• Platform android-29, build-tools 29.0.3
• ANDROID_HOME = /Users/aantelov/Library/Android/sdk
• Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_212-release-1586-b4-5784211)
• All Android licenses accepted.
[!] Xcode - develop for iOS and macOS
✗ Xcode installation is incomplete; a full installation is necessary for iOS development.
Download at: https://developer.apple.com/xcode/download/
Or install Xcode via the App Store.
Once installed, run:
sudo xcode-select --switch /Applications/Xcode.app/Contents/Developer
sudo xcodebuild -runFirstLaunch
• CocoaPods version 1.9.0
[✓] Android Studio (version 3.6)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 44.0.2
• Dart plugin version 192.7761
• Java version OpenJDK Runtime Environment (build 1.8.0_212-release-1586-b4-5784211)
[✓] Connected device (1 available)
• SM G950F • ce06182664a6821703 • android-arm64 • Android 9 (API 28)
`
output-modified.txt
Hello @TahaTesser
Have I added the issue in the correct project? because it looks like nobody is looking into this since 6 days ago.
|
2025-04-01T04:55:05.177719
| 2022-01-27T03:08:55
|
1115719920
|
{
"authors": [
"AAkira",
"darshankawar"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13949",
"repo": "FirebaseExtended/flutterfire",
"url": "https://github.com/FirebaseExtended/flutterfire/issues/7956"
}
|
gharchive/issue
|
🐛 [firebase_auth] Cannot handle quota-exceeded error on Android
Bug report
Describe the bug
We use the phone auth.
We've tried to authenticate multiple times with the spark(free) plan.
We were able to handle the quota-exceeded error on iOS, but we only got the too-many-requests error on Android.
However, when we changed to the blaze(paid) plan, we only got too-many-request errors on iOS as well.
Steps to reproduce
Send SMS multiple times
I called FirebaseAuth.instance.verifyPhoneNumber
Handle verificationFailed on Android and iOS
Expected behavior
Handle the quota-exceeded error on Android
Flutter doctor
Click To Expand
Doctor summary (to see all details, run flutter doctor -v):
[✓] Flutter (Channel stable, 2.8.0, on macOS 11.6.1 20G224 darwin-x64, locale ja-JP)
[!] Android toolchain - develop for Android devices (Android SDK version 31.0.0)
✗ cmdline-tools component is missing
Run `path/to/sdkmanager --install "cmdline-tools;latest"`
See https://developer.android.com/studio/command-line for more details.
✗ Android license status unknown.
Run `flutter doctor --android-licenses` to accept the SDK licenses.
See https://flutter.dev/docs/get-started/install/macos#android-setup for more details.
[✓] Xcode - develop for iOS and macOS (Xcode 13.2.1)
[✓] Chrome - develop for the web
[✓] Android Studio (version 2020.3)
[✓] Android Studio (version 2020.3)
[✓] IntelliJ IDEA Ultimate Edition (version 2021.2.3)
[✓] IntelliJ IDEA Ultimate Edition (version 2021.2.3)
[✓] IntelliJ IDEA Ultimate Edition (version 2021.1.2)
[✓] Connected device (3 available)
Flutter dependencies
Click To Expand
Dart SDK 2.15.0
Flutter SDK 2.8.0
m3pay 1.14.0+24
dependencies:
- cached_network_image 3.1.0+1 [flutter flutter_cache_manager octo_image cached_network_image_platform_interface cached_network_image_web]
- camera 0.9.4+6 [camera_platform_interface camera_web flutter pedantic quiver flutter_plugin_android_lifecycle]
- collection 1.15.0
- convex_bottom_bar 3.0.0 [flutter]
- device_info_plus 3.1.1 [flutter device_info_plus_platform_interface device_info_plus_linux device_info_plus_macos device_info_plus_web device_info_plus_windows]
- dio 4.0.4 [http_parser path]
- dotted_border 2.0.0+2 [flutter path_drawing]
- email_validator 2.0.1
- firebase_analytics 8.3.4 [firebase_analytics_platform_interface firebase_analytics_web firebase_core flutter meta]
- firebase_auth 3.2.0 [firebase_auth_platform_interface firebase_auth_web firebase_core firebase_core_platform_interface flutter meta]
- firebase_core 1.10.0 [firebase_core_platform_interface firebase_core_web flutter meta]
- firebase_crashlytics 2.3.0 [firebase_core firebase_core_platform_interface firebase_crashlytics_platform_interface flutter stack_trace]
- firebase_dynamic_links 3.0.1 [firebase_core flutter]
- firebase_messaging 11.1.0 [firebase_core firebase_core_platform_interface firebase_messaging_platform_interface firebase_messaging_web flutter meta]
- flutter 0.0.0 [characters collection meta typed_data vector_math sky_engine]
- flutter_hooks 0.18.0 [flutter]
- flutter_inappwebview 5.3.2 [flutter]
- flutter_linkify 5.0.2 [flutter linkify]
- flutter_local_notifications 9.0.3 [clock flutter flutter_local_notifications_linux flutter_local_notifications_platform_interface timezone]
- flutter_localizations 0.0.0 [flutter intl characters clock collection meta path typed_data vector_math]
- flutter_speed_dial 4.6.6 [flutter]
- flutter_stripe 2.0.2 [flutter stripe_android stripe_ios stripe_platform_interface]
- flutter_svg 1.0.0 [flutter meta path_drawing vector_math xml]
- freezed_annotation 1.0.0 [collection json_annotation meta]
- geolocator 7.7.1 [flutter geolocator_platform_interface geolocator_android geolocator_apple geolocator_web]
- git_info_plus 0.0.4 [flutter]
- google_maps_flutter 2.1.0 [flutter flutter_plugin_android_lifecycle google_maps_flutter_platform_interface]
- hooks_riverpod 1.0.0 [collection flutter flutter_hooks flutter_riverpod riverpod state_notifier]
- image 3.0.8 [archive meta xml]
- intl_utils 2.5.1 [analyzer archive args dart_style http intl path petitparser yaml]
- json_annotation 4.3.0 [meta]
- kana_kit 2.0.0 [equatable]
- logger 1.1.0
- openapi 1.0.0 [dio built_value built_collection]
- package_info_plus 1.3.0 [flutter package_info_plus_platform_interface package_info_plus_linux package_info_plus_macos package_info_plus_windows package_info_plus_web]
- path_provider 2.0.6 [flutter path_provider_linux path_provider_macos path_provider_platform_interface path_provider_windows]
- permission_handler 8.2.6 [flutter meta permission_handler_platform_interface]
- phone_number 0.12.0+1 [flutter meta]
- qr_code_scanner 0.6.1 [js flutter flutter_web_plugins]
- shared_preferences 2.0.8 [flutter meta shared_preferences_linux shared_preferences_macos shared_preferences_platform_interface shared_preferences_web shared_preferences_windows]
- shimmer 2.0.0 [flutter]
- sms_autofill 2.2.0 [pin_input_text_field flutter]
- time_machine 0.9.17 [meta collection]
- transparent_image 2.0.0
- tuple 2.0.0 [quiver]
- url_launcher 6.0.12 [flutter meta url_launcher_linux url_launcher_macos url_launcher_platform_interface url_launcher_web url_launcher_windows]
dev dependencies:
- build_runner 2.1.5 [args async analyzer build build_config build_daemon build_resolvers build_runner_core code_builder collection crypto dart_style frontend_server_client glob graphs http_multi_server io js logging meta mime package_config path pool pub_semver pubspec_parse shelf shelf_web_socket stack_trace stream_transform timing watcher web_socket_channel yaml]
- flutter_launcher_icons 0.9.2 [args image path yaml]
- flutter_test 0.0.0 [flutter test_api path fake_async clock stack_trace vector_math async boolean_selector characters charcode collection matcher meta source_span stream_channel string_scanner term_glyph typed_data]
- freezed 1.0.0 [analyzer build build_config collection meta source_gen freezed_annotation]
- json_serializable 6.0.1 [analyzer async build build_config collection json_annotation meta path pub_semver pubspec_parse source_gen source_helper]
- mockito 5.0.16 [analyzer build code_builder collection dart_style matcher meta path source_gen test_api]
dependency overrides:
- freezed_annotation 1.0.0 [collection json_annotation meta]
transitive dependencies:
- _fe_analyzer_shared 30.0.0 [meta]
- analyzer 2.7.0 [_fe_analyzer_shared cli_util collection convert crypto glob meta package_config path pub_semver source_span watcher yaml]
- archive 3.1.2 [crypto path]
- args 2.3.0
- async 2.8.2 [collection meta]
- boolean_selector 2.1.0 [source_span string_scanner]
- build 2.1.1 [analyzer async convert crypto glob logging meta path]
- build_config 1.0.0 [checked_yaml json_annotation path pubspec_parse yaml]
- build_daemon 3.0.0 [built_collection built_value http_multi_server logging pedantic path pool shelf shelf_web_socket stream_transform watcher web_socket_channel]
- build_resolvers 2.0.4 [analyzer async build crypto graphs logging path package_config pool pub_semver stream_transform yaml]
- build_runner_core 7.2.2 [async build build_config build_resolvers collection convert crypto glob graphs json_annotation logging meta path package_config pool timing watcher yaml]
- built_collection 5.1.1
- built_value 8.1.2 [built_collection collection fixnum meta]
- cached_network_image_platform_interface 1.0.0 [flutter flutter_cache_manager]
- cached_network_image_web 1.0.1 [flutter flutter_cache_manager cached_network_image_platform_interface]
- camera_platform_interface 2.1.1 [cross_file flutter meta plugin_platform_interface stream_transform]
- camera_web 0.2.1+1 [camera_platform_interface flutter flutter_web_plugins stream_transform]
- characters 1.2.0
- charcode 1.3.1
- checked_yaml 2.0.1 [json_annotation source_span yaml]
- cli_util 0.3.3 [meta path]
- clock 1.1.0
- code_builder 4.1.0 [built_collection built_value collection matcher meta]
- convert 3.0.1 [typed_data]
- cross_file 0.3.1+5 [flutter meta]
- crypto 3.0.1 [collection typed_data]
- dart_style 2.1.0 [analyzer args path pub_semver source_span]
- dbus 0.5.6 [args ffi meta pedantic xml]
- device_info_plus_linux 2.1.0 [device_info_plus_platform_interface file flutter meta]
- device_info_plus_macos 2.2.0 [device_info_plus_platform_interface flutter]
- device_info_plus_platform_interface 2.2.1 [flutter meta plugin_platform_interface]
- device_info_plus_web 2.1.0 [device_info_plus_platform_interface flutter_web_plugins flutter]
- device_info_plus_windows 2.1.0 [device_info_plus_platform_interface ffi flutter win32]
- equatable 2.0.3 [collection meta]
- fake_async 1.2.0 [clock collection]
- ffi 1.1.2
- file 6.1.2 [meta path]
- firebase 9.0.2 [http http_parser js]
- firebase_analytics_platform_interface 2.0.1 [flutter meta]
- firebase_analytics_web 0.3.0+1 [firebase firebase_analytics_platform_interface flutter flutter_web_plugins meta]
- firebase_auth_platform_interface 6.1.4 [firebase_core flutter meta plugin_platform_interface]
- firebase_auth_web 3.2.0 [firebase_auth_platform_interface firebase_core firebase_core_web flutter flutter_web_plugins http_parser intl js meta]
- firebase_core_platform_interface 4.1.0 [collection flutter meta plugin_platform_interface]
- firebase_core_web 1.2.0 [firebase_core_platform_interface flutter flutter_web_plugins js meta]
- firebase_crashlytics_platform_interface 3.1.6 [collection firebase_core flutter meta plugin_platform_interface]
- firebase_messaging_platform_interface 3.0.9 [firebase_core flutter meta plugin_platform_interface]
- firebase_messaging_web 2.1.0 [firebase_core firebase_core_web firebase_messaging_platform_interface flutter flutter_web_plugins js meta]
- fixnum 1.0.0
- flutter_blurhash 0.6.0 [flutter meta pedantic]
- flutter_cache_manager 3.1.2 [clock collection file flutter http path path_provider pedantic rxdart sqflite uuid]
- flutter_local_notifications_linux 0.3.0 [flutter flutter_local_notifications_platform_interface dbus path xdg_directories]
- flutter_local_notifications_platform_interface 5.0.0 [flutter plugin_platform_interface]
- flutter_plugin_android_lifecycle 2.0.3 [flutter]
- flutter_riverpod 1.0.0 [collection flutter meta riverpod state_notifier]
- flutter_web_plugins 0.0.0 [flutter js characters collection meta typed_data vector_math]
- frontend_server_client 2.1.2 [async path]
- geolocator_android 2.1.0 [flutter geolocator_platform_interface]
- geolocator_apple 1.2.0 [flutter geolocator_platform_interface]
- geolocator_platform_interface 2.3.5 [flutter plugin_platform_interface vector_math meta]
- geolocator_web 2.0.6 [flutter flutter_web_plugins geolocator_platform_interface]
- glob 2.0.1 [async collection file path pedantic string_scanner]
- google_maps_flutter_platform_interface 2.1.3 [collection flutter meta plugin_platform_interface stream_transform]
- graphs 2.1.0 [collection]
- http 0.13.3 [async http_parser meta path pedantic]
- http_multi_server 3.0.1 [async]
- http_parser 4.0.0 [charcode collection source_span string_scanner typed_data]
- intl 0.17.0 [clock path]
- io 1.0.3 [meta path string_scanner]
- js 0.6.3
- linkify 4.1.0
- logging 1.0.2
- matcher 0.12.11 [stack_trace]
- meta 1.7.0
- mime 1.0.0
- octo_image 1.0.0+1 [flutter flutter_blurhash]
- package_config 2.0.2 [path]
- package_info_plus_linux 1.0.3 [package_info_plus_platform_interface flutter path]
- package_info_plus_macos 1.3.0 [flutter]
- package_info_plus_platform_interface 1.0.2 [flutter meta plugin_platform_interface]
- package_info_plus_web 1.0.4 [flutter flutter_web_plugins http meta package_info_plus_platform_interface]
- package_info_plus_windows 1.0.4 [package_info_plus_platform_interface ffi flutter win32]
- path 1.8.0
- path_drawing 1.0.0 [vector_math meta path_parsing flutter]
- path_parsing 1.0.0 [vector_math meta]
- path_provider_linux 2.1.0 [flutter path path_provider_platform_interface xdg_directories]
- path_provider_macos 2.0.2 [flutter]
- path_provider_platform_interface 2.0.1 [flutter meta platform plugin_platform_interface]
- path_provider_windows 2.0.3 [ffi flutter meta path path_provider_platform_interface win32]
- pedantic 1.11.1
- permission_handler_platform_interface 3.7.0 [flutter meta plugin_platform_interface]
- petitparser 4.3.0 [meta]
- pin_input_text_field 4.1.1 [flutter]
- platform 3.0.2
- plugin_platform_interface 2.0.1 [meta]
- pool 1.5.0 [async stack_trace]
- process 4.2.3 [file path platform]
- pub_semver 2.1.0 [collection meta]
- pubspec_parse 1.0.0 [checked_yaml collection json_annotation pub_semver yaml]
- quiver 3.0.1 [matcher]
- riverpod 1.0.0 [collection meta state_notifier]
- rxdart 0.27.2
- shared_preferences_linux 2.0.2 [file meta flutter path path_provider_linux shared_preferences_platform_interface]
- shared_preferences_macos 2.0.2 [flutter shared_preferences_platform_interface]
- shared_preferences_platform_interface 2.0.0 [flutter]
- shared_preferences_web 2.0.2 [flutter flutter_web_plugins meta shared_preferences_platform_interface]
- shared_preferences_windows 2.0.2 [flutter file meta path path_provider_platform_interface path_provider_windows shared_preferences_platform_interface]
- shelf 1.2.0 [async collection http_parser path stack_trace stream_channel]
- shelf_web_socket 1.0.1 [shelf stream_channel web_socket_channel]
- sky_engine 0.0.99
- source_gen 1.1.1 [analyzer async build dart_style glob meta path source_span yaml]
- source_helper 1.3.0 [analyzer collection source_gen]
- source_span 1.8.1 [collection path term_glyph]
- sqflite 2.0.0+4 [flutter sqflite_common path]
- sqflite_common 2.0.1+1 [synchronized path meta]
- stack_trace 1.10.0 [path]
- state_notifier 0.7.1 [meta]
- stream_channel 2.1.0 [async]
- stream_transform 2.0.0
- string_scanner 1.1.0 [charcode source_span]
- stripe_android 2.0.2 [flutter]
- stripe_ios 2.0.2 [flutter]
- stripe_platform_interface 2.0.2 [flutter freezed_annotation json_annotation meta plugin_platform_interface]
- synchronized 3.0.0
- term_glyph 1.2.0
- test_api 0.4.3 [async boolean_selector collection meta source_span stack_trace stream_channel string_scanner term_glyph matcher]
- timezone 0.8.0 [path]
- timing 1.0.0 [json_annotation]
- typed_data 1.3.0 [collection]
- url_launcher_linux 2.0.2 [flutter]
- url_launcher_macos 2.0.2 [flutter]
- url_launcher_platform_interface 2.0.4 [flutter plugin_platform_interface]
- url_launcher_web 2.0.4 [flutter flutter_web_plugins meta url_launcher_platform_interface]
- url_launcher_windows 2.0.2 [flutter]
- uuid 3.0.4 [crypto]
- vector_math 2.1.1
- watcher 1.0.0 [async path pedantic]
- web_socket_channel 2.1.0 [async crypto stream_channel]
- win32 2.2.9 [ffi]
- xdg_directories 0.2.0 [meta path process]
- xml 5.3.0 [collection meta petitparser]
- yaml 3.1.0 [collection source_span string_scanner]
@AAkira
There's similar open issue describing your case, https://github.com/FirebaseExtended/flutterfire/issues/6687, so please follow-up in it for further updates.
Closing this as a duplicate. Also take a look at this related open PR: https://github.com/FirebaseExtended/flutterfire/pull/3402
|
2025-04-01T04:55:05.187488
| 2020-03-09T03:55:22
|
577642954
|
{
"authors": [
"chinkysight",
"kroikie"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13950",
"repo": "FirebaseExtended/flutterfire",
"url": "https://github.com/FirebaseExtended/flutterfire/pull/2148"
}
|
gharchive/pull-request
|
[firebase_auth plugin] README.md Updated
Description
I had just updated the README.md file. There isn't any deletion of any pre-existed info in README.md
Related Issues
There isn't any issue created for this, but I thought it would be very helpful to add some examples of doing Firebase Facebook and Firebase Twitter authentication.
Checklist
Before you create this PR confirm that it meets all requirements listed below by checking the relevant checkboxes ([x]). This will ensure a smooth and quick review process.
[x] I read the Contributor Guide and followed the process outlined there for submitting PRs.
[x] If the pull request affects only one plugin, the PR title starts with the name of the plugin in brackets (e.g. [cloud_firestore])
[x] My PR includes unit or integration tests for all changed/updated/fixed behaviors (See Contributor Guide).
[x] All existing and new tests are passing.
[x] I updated/added relevant documentation (doc comments with ///).
[x] The analyzer (flutter analyze) does not report any problems on my PR.
[x] I read and followed the Flutter Style Guide.
[x] I updated pubspec.yaml with an appropriate new version according to the pub versioning philosophy.
[x] I updated CHANGELOG.md to add a description of the change.
[x] I signed the CLA.
[x] I am willing to follow-up on review comments in a timely manner.
Breaking Change
Does your PR require plugin users to manually update their apps to accommodate your change?
[ ] Yes, this is a breaking change (please indicate a breaking change in CHANGELOG.md and increment major revision).
[x] No, this is not a breaking change.
@chinkysight Thanks for the PR, would you mind rebasing this so that the tests can pass and I'll be happy to review.
|
2025-04-01T04:55:05.223696
| 2020-01-12T21:12:22
|
548627683
|
{
"authors": [
"Fizzadar",
"mkinney"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13951",
"repo": "Fizzadar/pyinfra",
"url": "https://github.com/Fizzadar/pyinfra/issues/229"
}
|
gharchive/issue
|
bug in apt.repo()?
I'm running into an issue using apt.repo().
Am I missing something?
$ cat mike.py
from pyinfra.modules import apt
SUDO = True
apt.repo(
{'Install VirtualBox repo'},
'deb https://download.virtualbox.org/virtualbox/debian bionic contrib',
)
$ rm pyinfra-debug.log
rm: pyinfra-debug.log: No such file or directory
$ pyinfra -vv --debug @docker/ubuntu:bionic mike.py
[pyinfra_cli.main] Checking potential directory:
[pyinfra_cli.main] Setting directory to:
--> Loading config...
--> Loading inventory...
[pyinfra_cli.inventory] Creating fake inventory...
[pyinfra.api.inventory] Expanding inventory connector: docker
The @docker connector is in Alpha!
[pyinfra_cli.inventory] Looking for group data in: group_data/all.py
[pyinfra_cli.inventory] Looking for group data in: group_data/bsd.py
[pyinfra.api.inventory] Expanding inventory connector: docker
--> Connecting to hosts...
localhost: >>> docker run -d ubuntu:bionic sleep 10000
localhost: a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05
[pyinfra.api.state] Activating host: ubuntu:bionic
--> Preparing operations...
Loading: mike.py
[pyinfra.api.operation] Adding operation, {'Install VirtualBox repo'}, called @ mike.py:7, opLines=(0, 7), opHash=97d8032458c50f8c46e458a4ca15c67a69be92b3
[pyinfra.api.facts] Getting fact: apt_sources (ensure_hosts: (ubuntu:bionic,))
[pyinfra.api.util] Building command (shell_executable: sh): docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "cat /etc/apt/sources.list /etc/apt/sources.list.d/*.list || true"
[pyinfra.api.connectors.local] --> Running command on localhost: sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "cat /etc/apt/sources.list /etc/apt/sources.list.d/*.list || true"'
[ubuntu:bionic] >>> sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "cat /etc/apt/sources.list /etc/apt/sources.list.d/*.list || true"'
[ubuntu:bionic] # See http://help.ubuntu.com/community/UpgradeNotes for how to upgrade to
[ubuntu:bionic] # newer versions of the distribution.
[ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic main restricted
[ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic main restricted
[ubuntu:bionic]
[ubuntu:bionic] ## Major bug fix updates produced after the final release of the
[ubuntu:bionic] ## distribution.
[ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic-updates main restricted
[ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic-updates main restricted
[ubuntu:bionic]
[ubuntu:bionic] ## N.B. software from this repository is ENTIRELY UNSUPPORTED by the Ubuntu
[ubuntu:bionic] ## team. Also, please note that software in universe WILL NOT receive any
[ubuntu:bionic] ## review or updates from the Ubuntu security team.
[ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic universe
[ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic universe
[ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic-updates universe
[ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic-updates universe
[ubuntu:bionic]
[ubuntu:bionic] ## N.B. software from this repository is ENTIRELY UNSUPPORTED by the Ubuntu
[ubuntu:bionic] ## team, and may not be under a free licence. Please satisfy yourself as to
[ubuntu:bionic] ## your rights to use the software. Also, please note that software in
[ubuntu:bionic] ## multiverse WILL NOT receive any review or updates from the Ubuntu
[ubuntu:bionic] ## security team.
[ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic multiverse
[ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic multiverse
[ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic-updates multiverse
[ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic-updates multiverse
[ubuntu:bionic]
[ubuntu:bionic] ## N.B. software from this repository may not have been tested as
[ubuntu:bionic] ## extensively as that contained in the main release, although it includes
[ubuntu:bionic] ## newer versions of some applications which may provide useful features.
[ubuntu:bionic] ## Also, please note that software in backports WILL NOT receive any review
[ubuntu:bionic] ## or updates from the Ubuntu security team.
[ubuntu:bionic] deb http://archive.ubuntu.com/ubuntu/ bionic-backports main restricted universe multiverse
[ubuntu:bionic] # deb-src http://archive.ubuntu.com/ubuntu/ bionic-backports main restricted universe multiverse
[ubuntu:bionic]
[ubuntu:bionic] ## Uncomment the following two lines to add software from Canonical's
[ubuntu:bionic] ## 'partner' repository.
[ubuntu:bionic] ## This software is not part of Ubuntu, but is offered by Canonical and the
[ubuntu:bionic] ## respective vendors as a service to Ubuntu users.
[ubuntu:bionic] # deb http://archive.canonical.com/ubuntu bionic partner
[ubuntu:bionic] # deb-src http://archive.canonical.com/ubuntu bionic partner
[ubuntu:bionic]
[ubuntu:bionic] deb http://security.ubuntu.com/ubuntu/ bionic-security main restricted
[ubuntu:bionic] # deb-src http://security.ubuntu.com/ubuntu/ bionic-security main restricted
[ubuntu:bionic] deb http://security.ubuntu.com/ubuntu/ bionic-security universe
[ubuntu:bionic] # deb-src http://security.ubuntu.com/ubuntu/ bionic-security universe
[ubuntu:bionic] deb http://security.ubuntu.com/ubuntu/ bionic-security multiverse
[ubuntu:bionic] # deb-src http://security.ubuntu.com/ubuntu/ bionic-security multiverse
[ubuntu:bionic] cat: '/etc/apt/sources.list.d/*.list': No such file or directory
[pyinfra.api.connectors.local] --> Waiting for exit status...
[pyinfra.api.connectors.local] --> Command exit status: 0
Loaded fact apt_sources
[pyinfra.api.facts] Getting fact: find_in_file (ensure_hosts: (ubuntu:bionic,))
[pyinfra.api.util] Building command (shell_executable: sh): docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" /etc/apt/sources.list || (find /etc/apt/sources.list -type f > /dev/null && echo "__pyinfra_exists_/etc/apt/sources.list")"
[pyinfra.api.connectors.local] --> Running command on localhost: sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" /etc/apt/sources.list || (find /etc/apt/sources.list -type f > /dev/null && echo "__pyinfra_exists_/etc/apt/sources.list")"'
[ubuntu:bionic] >>> sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" /etc/apt/sources.list || (find /etc/apt/sources.list -type f > /dev/null && echo "__pyinfra_exists_/etc/apt/sources.list")"'
| fact: find_in_file('/etc/apt/sources.list', '^.*deb https://download.virtualbox.org/virtualbox/debian bi / fact: find_in_file('/etc/apt/sources.list', '^.*deb https://download.virtualbox.org/virtualbox/debian bi - fact: find_in_file('/etc/apt/sources.list', '^.*deb https://download.virtualbox.org/virtualbox/debian bi \ fact: find_in_file('/etc/apt/sources.list', '^.*deb https://download.virtualbox.org/virtualbox/debian bi[pyinfra.api.connectors.local] --> Waiting for exit status...
[pyinfra.api.connectors.local] --> Command exit status: 1
Loaded fact find_in_file: ('/etc/apt/sources.list', '^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$')
[ubuntu:bionic] Ready: mike.py
--> Proposed changes:
Groups: @docker
[ubuntu:bionic] Operations: 1 Commands: 1
--> Beginning operation run...
--> Starting operation: Install VirtualBox repo
[pyinfra.api.operations] Starting operation Install VirtualBox repo on ubuntu:bionic
[pyinfra.api.util] Building command (shell_executable: sh): docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "# If the file now exists
if [ -f "/etc/apt/sources.list" ]; then
# Grep for the line, sed if matches
(grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" "/etc/apt/sources.list" && sed -i="" "s/^.*deb https:\/\/download.virtualbox.org\/virtualbox\/debian bionic contrib.*$//" /etc/apt/sources.list) || # Else echo
echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list
# No file, just echo
else
echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list
fi"
[pyinfra.api.connectors.local] --> Running command on localhost: sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "# If the file now exists
if [ -f "/etc/apt/sources.list" ]; then
# Grep for the line, sed if matches
(grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" "/etc/apt/sources.list" && sed -i="" "s/^.*deb https:\/\/download.virtualbox.org\/virtualbox\/debian bionic contrib.*$//" /etc/apt/sources.list) || # Else echo
echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list
# No file, just echo
else
echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list
fi"'
[ubuntu:bionic] >>> sh -c 'docker exec a426aa3919c74b6d09a5d839b4b89f5f1863114c67824842655044000cff3e05 sh -c "# If the file now exists
if [ -f "/etc/apt/sources.list" ]; then
# Grep for the line, sed if matches
(grep "^.*deb https://download.virtualbox.org/virtualbox/debian bionic contrib.*$" "/etc/apt/sources.list" && sed -i="" "s/^.*deb https:\/\/download.virtualbox.org\/virtualbox\/debian bionic contrib.*$//" /etc/apt/sources.list) || # Else echo
echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list
# No file, just echo
else
echo "deb https://download.virtualbox.org/virtualbox/debian bionic contrib" >> /etc/apt/sources.list
fi"'
[ubuntu:bionic] https://download.virtualbox.org/virtualbox/debian: 4: https://download.virtualbox.org/virtualbox/debian: Syntax error: end of file unexpected (expecting ")")
[pyinfra.api.connectors.local] --> Waiting for exit status...
[pyinfra.api.connectors.local] --> Command exit status: 2
[ubuntu:bionic] Error
[pyinfra.api.state] Failing hosts: ubuntu:bionic
localhost: >>> docker commit a426aa3919c7
localhost: sha256:db4497e361e507180b287c769f7777969102d1ffd5d009a6772ccfb9a4c218a6
localhost: >>> docker rm -f a426aa3919c7
localhost: a426aa3919c7
[ubuntu:bionic] docker build complete, image ID: db4497e361e5
--> pyinfra error: No hosts remaining!
$ ls pyinfra-debug.log
ls: pyinfra-debug.log: No such file or directory
$
Turns out this is an issue with Docker specifically - not quoting commands correctly; this should be resolved by https://github.com/Fizzadar/pyinfra/commit/0cc6433f6569d92596d5570423cba571208b7e0c.
|
2025-04-01T04:55:05.227372
| 2017-09-03T12:23:27
|
254871365
|
{
"authors": [
"Fizzadar",
"tobald"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13952",
"repo": "Fizzadar/pyinfra",
"url": "https://github.com/Fizzadar/pyinfra/pull/116"
}
|
gharchive/pull-request
|
Hackish: show "[sudo] password for user:" message when using @local
A naïve fix for letting sudo inform the user it waits for his password. Only tested with @local.
The following ticket details a more advanced take on the issue: https://github.com/Fizzadar/pyinfra/issues/40
@tobald sorry for never getting back to you on this one! I'm still not sure how or whether pyinfra should handle input. I'd like to do it as it makes the user experience far better; but the current plan in #40 is pretty involved and would make pyinfra lot more complicated :/
@tobald sorry for never getting back to you on this one! I'm still not
sure how or whether pyinfra should handle input. I'd like to do it as
it makes the user experience far better; but the current plan in #40
is pretty involved and would make pyinfra lot more complicated :/
No worries, I do not have a use for this feature anymore. Proper input
management is the way to go as stated in #40, and seems complicated.
Meanwhile we could detect/display something to the user, that was the
purpose of this commit.
|
2025-04-01T04:55:05.236585
| 2024-10-27T09:09:14
|
2616438623
|
{
"authors": [
"VERT1DAN"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13953",
"repo": "FlafyDev/auto-nong-indexes",
"url": "https://github.com/FlafyDev/auto-nong-indexes/issues/5478"
}
|
gharchive/issue
|
add F-777 - Sonic Blaster (DeCody remix) by DeCody
Song Name
F-777 - Sonic Blaster (DeCody remix)
Artist Name
DeCody
Source
Youtube
Youtube Link
9QV_T53XoEA
Direct File Link
No response
Song ID
574484
Start Offset [ms]
0
Extra
Created from Auto Nong.
Created for level: 26681070
Downloaded: Yes
accept
вс, 27 окт. 2024 г., 12:09 auto-nong[bot] @.***>:
Thanks for contributing!
Someone will soon review your request.
Here is the JSON result for the reviewer:
{
"name": "F-777 - Sonic Blaster (DeCody remix)",
"artist": "DeCody",
"source": "youtube",
"startOffset": 0,
"yt-id": "9QV_T53XoEA",
"songs": [
574484
]
}
To accept the request, a reviewer needs to write "accept".
—
Reply to this email directly, view it on GitHub
https://github.com/FlafyDev/auto-nong-indexes/issues/5478#issuecomment-2439918167,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/BMOTCG7SL5N6IC3ORUOC5GLZ5SUVVAVCNFSM6AAAAABQVUAT62VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDIMZZHEYTQMJWG4
.
You are receiving this because you authored the thread.Message ID:
@.***>
|
2025-04-01T04:55:05.245203
| 2023-05-31T16:28:42
|
1734589387
|
{
"authors": [
"khvn26",
"matthewelwell"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13954",
"repo": "Flagsmith/flagsmith",
"url": "https://github.com/Flagsmith/flagsmith/issues/2256"
}
|
gharchive/issue
|
Swagger is broken
Describe the bug
SwaggerGenerationError
your query_serializer contains fields that conflict with the filter_backend or paginator_class on the view - GET /api/v1/audit/
It's likely the type mismatch between filter params and the DRF serializer:
Filter:
{
description: '',
in: 'query',
name: 'is_system_event',
required: False,
> type: 'string' <
},
Serializer field:
{
in: 'query',
name: 'is_system_event',
required: False,
> type: 'boolean' <
},
The error is currently only reproducible for user id 14119.
Potentially related issue: https://github.com/axnsan12/drf-yasg/issues/514
To Reproduce
Steps to reproduce the behavior:
Login as user 14119.
Go to https://api.flagsmith.com/api/v1/docs/?format=openapi
Observe the HTTP 500 response.
Expected behavior
The server should render the OpenAPI schema.
A unit test should catch Swagger generation errors.
How are you running Flagsmith?
[ ] Self Hosted with Docker
[ ] Self Hosted with Kubernetes
[x] SaaS at flagsmith.com
[ ] Some other way
I have looked into this a bit further - I don't believe it is related to the error provided in the description since that was fixed here: https://github.com/Flagsmith/flagsmith/pull/2326.
It seems to only happen when users are logged in as a user that also has access to the django admin in the same browser session that they access the API docs.
The issue is still reproducible in prod. Tried logging out and in again to no avail.
Sorry, yes - it's still reproducible for sure, but I'm not convinced it's related to the exception in the description. It's definitely related to being logged in as a staff user though.
|
2025-04-01T04:55:05.256737
| 2019-08-07T15:26:13
|
477996991
|
{
"authors": [
"bootstraponline",
"doodla"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13955",
"repo": "Flank/flank",
"url": "https://github.com/Flank/flank/issues/588"
}
|
gharchive/issue
|
Realtime streaming of test results via firebase function
Prototype how to get firebase function streaming data to Flank.
https://github.com/firebase/functions-samples/tree/master/quickstarts/test-complete
This might not be possible given the function will have to push ( post ) data to some place, and most CI systems don't allow incoming traffic.
We're building a comprehensive build and test analytics platform that integrates with Flank. 😄 This is a research ticket that will help prove out those efforts.
Data will likely be pushed to a Firebase database.
For test analytics, we'll parse the final JUnit XML artifact as attached to the CI job. In this approach, the firebase function doesn't add value.
|
2025-04-01T04:55:05.293895
| 2021-06-19T15:26:58
|
925416032
|
{
"authors": [
"MarceloWatson",
"flax-dev"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13956",
"repo": "Flax-Network/flax-blockchain",
"url": "https://github.com/Flax-Network/flax-blockchain/issues/2"
}
|
gharchive/issue
|
API
Helo ! Is there any API or something similar so I can try to develop a site with information about Flax?
Hey, our website is https://flaxnetwork.org/. This is a fork of https://github.com/Chia-Network/chia-blockchain. You can see there for more info, but there is currently no user-facing API.
|
2025-04-01T04:55:05.307400
| 2022-07-18T13:53:59
|
1307998159
|
{
"authors": [
"dcconner",
"drosen20"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13957",
"repo": "FlexBE/flexbe_app",
"url": "https://github.com/FlexBE/flexbe_app/issues/80"
}
|
gharchive/issue
|
ROS2 OCS does not launch behavior
I am experimenting with flexBE in ROS2 Galactic and I ran into an issue with launching a behavior from the OCS GUI to run on a remote machine.
What I did:
followed the Basic Tutorials to create a hello world behavior that waits a few seconds and then prints to the screen
launched the behavior from the GUI after running ros2 launch flexbe_app flexbe_full.launch.py and observed the expected behavior
Copied the tutorials_behaviors repo (created in the tutorials) to the computer on my robot and ran colcon build to set up the workspace
On remote, launched the OCS with ros2 launch flexbe_app flexbe_ocs.launch.py. On robot, launched the behavior engine with ros2 launch flexbe_onboard behavior_onboard.launch.py.
Results:
The OCS was able to connect to the behavior engine. After seeing --- Behavior Engine ready! --- on the robot the OCS indicated --- Behavior Mirror ready! ---. But after clicking Start Execution I get an error
Failed to retrieve behavior from library: [-]
787353483
Dropped behavior start request because preparation failed.
Also if I switch the launch order (engine on robot first, then OCS on remote), when I click Start Execution the OCS hangs on the Waiting For Behavior screen with the message Start command has been sent. Now waiting for confirmation from onboard behavior...
This is an annoyance and something we plan to look at this summer. Likely it will involve a script to copy behaviors from the install folder to the source. Suggestions are appreciated.
|
2025-04-01T04:55:05.317759
| 2018-01-04T13:05:07
|
285979756
|
{
"authors": [
"LululuSir",
"raphaelschaad",
"reactxp"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13959",
"repo": "Flipboard/FLAnimatedImage",
"url": "https://github.com/Flipboard/FLAnimatedImage/pull/198"
}
|
gharchive/pull-request
|
If a GIF image has no Netscape 2.0 loop extension, it is meant to play once and then stop.
If a GIF image has no Netscape 2.0 loop extension, it is meant to play once and then stop.
Is there anyone else?
Is there anyone else?
yes
@LululuSir Is this the same as #102 ?
@LululuSir Is this the same as #102 ?
@LululuSir Is this the same as #102 ?
YES
This example GIF from this PR comment shows no Netscape 2.0 loop extension (when inspected with > gifsicle --extension-info) and thus shows each frame exactly once and then stops in Chrome/Safari and in FLAnimatedImage. This seems to be the correct behavior. Note that the definition of FLAnimatedImage's loopCount follows ImageIO's definition.
If you find a case where FLAnimatedImage behaves differently than Chrome/Safari, please provide the respective GIF file as an example. Then we can have a look at whether something would need to change in FLAnimatedImage.
|
2025-04-01T04:55:05.345958
| 2023-09-14T21:19:22
|
1897321542
|
{
"authors": [
"RyanDavies19",
"sanguinariojoe"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13960",
"repo": "FloatingArrayDesign/MoorDyn",
"url": "https://github.com/FloatingArrayDesign/MoorDyn/issues/121"
}
|
gharchive/issue
|
Local docs compiling
@sanguinariojoe do you have any idea what could be causing this error in compiling the documentation locally:
Patching output file 56/58
Patching output file 57/58
Patching output file 58/58
type lookup cache used 10518/65536 hits=122962 misses=10670
symbol lookup cache used 6206/65536 hits=83658 misses=6206
finished...
/Users/rdavies/Library/Python/3.9/lib/python/site-packages/urllib3/__init__.py:34: NotOpenSSLWarning: urllib3 v2.0 only supports OpenSSL 1.1.1+, currently the 'ssl' module is compiled with 'LibreSSL 2.8.3'. See: https://github.com/urllib3/urllib3/issues/3020
warnings.warn(
/Users/rdavies/Library/Python/3.9/lib/python/site-packages/breathe/project.py:116: RemovedInSphinx80Warning: Sphinx 8 will drop support for representing paths as strings. Use "pathlib.Path" or "os.fspath" instead.
self._default_build_dir = os.path.dirname(app.doctreedir.rstrip(os.sep))
making output directory... done
Exception occurred:
File "/Applications/Xcode.app/Contents/Developer/Library/Frameworks/Python3.framework/Versions/3.9/lib/python3.9/re.py", line 210, in sub
return _compile(pattern, flags).sub(repl, string, count)
TypeError: expected string or bytes-like object
The full traceback has been saved in /var/folders/46/4_pcs77x31j8w360rc7pfv6chrj_sl/T/sphinx-err-athjs7ze.log, if you want to report the issue to the developers.
Please also report this if it was a user error, so that a better error message can be provided next time.
A bug report can be filed in the tracker at <https://github.com/sphinx-doc/sphinx/issues>. Thanks!
make[2]: *** [docs/sphinx/index.html] Error 2
make[1]: *** [docs/CMakeFiles/Sphinx.dir/all] Error 2
make: *** [all] Error 2
The traceback that was saved shows the following:
# Platform: darwin; (macOS-13.5.2-x86_64-i386-64bit)
# Sphinx version: 7.2.6
# Python version: 3.9.6 (CPython)
# Docutils version: 0.18.1
# Jinja2 version: 3.1.2
# Pygments version: 2.15.1
# Last messages:
# Loaded extensions:
# Traceback:
Traceback (most recent call last):
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/cmd/build.py", line 293, in build_main
app = Sphinx(args.sourcedir, args.confdir, args.outputdir,
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/application.py", line 272, in __init__
self._init_builder()
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/application.py", line 342, in _init_builder
self.builder.init()
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/builders/html/__init__.py", line 219, in init
self.init_templates()
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/builders/html/__init__.py", line 270, in init_templates
self.theme = theme_factory.create(themename)
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py", line 227, in create
self.load_extra_theme(name)
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py", line 181, in load_extra_theme
self.load_external_theme(name)
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py", line 194, in load_external_theme
theme_entry_points = entry_points(group='sphinx.html_themes')
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/__init__.py", line 933, in entry_points
return EntryPoints(eps).select(**params)
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/__init__.py", line 930, in <genexpr>
eps = itertools.chain.from_iterable(
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/_itertools.py", line 16, in unique_everseen
k = key(element)
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/_py39compat.py", line 18, in normalized_name
return dist._normalized_name
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/__init__.py", line 855, in _normalized_name
or super()._normalized_name
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/__init__.py", line 473, in _normalized_name
return Prepared.normalize(self.name)
File "/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/__init__.py", line 773, in normalize
return re.sub(r"[-_.]+", "-", name).lower().replace('-', '_')
File "/Applications/Xcode.app/Contents/Developer/Library/Frameworks/Python3.framework/Versions/3.9/lib/python3.9/re.py", line 210, in sub
return _compile(pattern, flags).sub(repl, string, count)
TypeError: expected string or bytes-like object
Additionally early on in the process it says /bin/sh: sphinx-build: command not found
To get it to compile I am required to modify line 75 of build/docs/CMakeFiles/Sphinx.dir/build.make to remove the quotes around the path to python that are autogenerated.
This is confusing me because last week it was working fine and I was able to compile the docs locally without error. I havent changed any of the code other than .rst files, and made no modifications to my system as far as I can tell. I've tried uninstalling all the required packages and reinstalling them, as well as deleting the build folder and remaking it with cmake.
Pretty eclectic error, but it seems to me that you do not have the theme
installed:
sphinx_rtd_theme
As far as I remember it shall be installed with pip
On Thu, 14 Sept 2023, 23:19 RyanDavies19, @.***> wrote:
@sanguinariojoe https://github.com/sanguinariojoe do you have any idea
what could be causing this error in compiling the documentation locally:
`Patching output file 56/58
Patching output file 57/58
Patching output file 58/58
type lookup cache used 10518/65536 hits=122962 misses=10670
symbol lookup cache used 6206/65536 hits=83658 misses=6206
finished...
/Users/rdavies/Library/Python/3.9/lib/python/site-packages/urllib3/init.py:34:
NotOpenSSLWarning: urllib3 v2.0 only supports OpenSSL 1.1.1+, currently the
'ssl' module is compiled with 'LibreSSL 2.8.3'. See: urllib3/urllib3#3020
https://github.com/urllib3/urllib3/issues/3020
warnings.warn(
/Users/rdavies/Library/Python/3.9/lib/python/site-packages/breathe/project.py:116:
RemovedInSphinx80Warning: Sphinx 8 will drop support for representing paths
as strings. Use "pathlib.Path" or "os.fspath" instead.
self._default_build_dir = os.path.dirname(app.doctreedir.rstrip(os.sep))
making output directory... done
Exception occurred:
File
"/Applications/Xcode.app/Contents/Developer/Library/Frameworks/Python3.framework/Versions/3.9/lib/python3.9/re.py",
line 210, in sub
return _compile(pattern, flags).sub(repl, string, count)
TypeError: expected string or bytes-like object
The full traceback has been saved in
/var/folders/46/4_pcs77x31j8w360rc7pfv6chrj_sl/T/sphinx-err-athjs7ze.log,
if you want to report the issue to the developers.
Please also report this if it was a user error, so that a better error
message can be provided next time.
A bug report can be filed in the tracker at
https://github.com/sphinx-doc/sphinx/issues. Thanks!
make[2]: *** [docs/sphinx/index.html] Error 2
make[1]: *** [docs/CMakeFiles/Sphinx.dir/all] Error 2
make: *** [all] Error 2`
The traceback that was saved shows the following:
`# Platform: darwin; (macOS-13.5.2-x86_64-i386-64bit)
Sphinx version: 7.2.6 Python version: 3.9.6 (CPython) Docutils version:
0.18.1 Jinja2 version: 3.1.2 Pygments version: 2.15.1 Last messages: Loaded
extensions: Traceback:
Traceback (most recent call last):
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/cmd/build.py",
line 293, in build_main
app = Sphinx(args.sourcedir, args.confdir, args.outputdir,
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/application.py",
line 272, in init
self._init_builder()
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/application.py",
line 342, in _init_builder
self.builder.init()
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/builders/html/
init.py", line 219, in init
self.init_templates()
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/builders/html/
init.py", line 270, in init_templates
self.theme = theme_factory.create(themename)
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py",
line 227, in create
self.load_extra_theme(name)
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py",
line 181, in load_extra_theme
self.load_external_theme(name)
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/sphinx/theming.py",
line 194, in load_external_theme
theme_entry_points = entry_points(group='sphinx.html_themes')
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/
init.py", line 933, in entry_points
return EntryPoints(eps).select(**params)
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/
init.py", line 930, in
eps = itertools.chain.from_iterable(
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/_itertools.py",
line 16, in unique_everseen
k = key(element)
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/_py39compat.py",
line 18, in normalized_name
return dist._normalized_name
File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/
init.py", line 855, in _normalized_name
or super().
normalized_name File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/init.py",
line 473, in normalized_name return Prepared.normalize(self.name
http://self.name) File
"/Users/rdavies/Library/Python/3.9/lib/python/site-packages/importlib_metadata/init.py",
line 773, in normalize return re.sub(r"[-.]+", "-",
name).lower().replace('-', '')
File
"/Applications/Xcode.app/Contents/Developer/Library/Frameworks/Python3.framework/Versions/3.9/lib/python3.9/re.py",
line 210, in sub
return _compile(pattern, flags).sub(repl, string, count)
TypeError: expected string or bytes-like object`
—
Reply to this email directly, view it on GitHub
https://github.com/FloatingArrayDesign/MoorDyn/issues/121, or
unsubscribe
https://github.com/notifications/unsubscribe-auth/AAMXKKES3RWPYV5DIBTM6BDX2NYGLANCNFSM6AAAAAA4Y4B574
.
You are receiving this because you were mentioned.Message ID:
@.***>
@sanguinariojoe thanks, I was curious if you had ideas. Turns out when I installed Xcode with admin rights for something else, it created a second python copy. I had to go in and delete all that but things are working well now.
|
2025-04-01T04:55:05.394825
| 2024-02-01T11:53:40
|
2112327493
|
{
"authors": [
"elenaviter",
"hardillb"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13961",
"repo": "FlowFuse/nr-launcher",
"url": "https://github.com/FlowFuse/nr-launcher/issues/206"
}
|
gharchive/issue
|
Enhancing Node-RED Logging: Introducing Host Output Logging Capability
Description
At present, the logging mechanism for Node-RED projects directs all logs to the 'driver', which can be accessed through the FlowFuse Project logs interface. In environments where Kubernetes clusters are deployed and are integrated with monitoring solutions (such as log forwarding to OTEL/ FluentBit for pods, and then to traces/metrics/etc. to backends like Prometheus/Jaeger/, etc.), there is a valuable opportunity to streamline monitoring practices.
Enabling Node-RED projects to output their logs directly to the pod's console will facilitate a more unified approach to monitoring by leveraging existing log aggregation configurations.
https://github.com/FlowFuse/flowfuse/issues/3324
Thank you @hardillb - closing this one
|
2025-04-01T04:55:05.405227
| 2023-08-25T15:39:32
|
1867268042
|
{
"authors": [
"amzamani",
"chungyau97",
"isikhi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13962",
"repo": "FlowiseAI/Flowise",
"url": "https://github.com/FlowiseAI/Flowise/issues/832"
}
|
gharchive/issue
|
[BUG]yarn start not working
Describe the bug
Cloned the repo and followed the steps using yarn
To Reproduce
cloned and changed directory then=>
yarn install && yarn build && yarn start
install and build work fine with no error
when start is executed, error is encountered
Error
`
Tasks: 3 successful, 3 total
Cached: 0 cached, 3 total
Time: 2m25.321s
✨ Done in 146.13s.
yarn run v1.22.19
$ run-script-os
$ cd packages/server/bin && ./run start
(node:1212) [ERR_DLOPEN_FAILED] Error Plugin: flowise: dlopen(/Users/amzamani/Desktop/incoming/ai/Flowise/node_modules/canvas/build/Release/canvas.node, 1): Library not loaded: @loader_path/libpixman-<IP_ADDRESS>.dylib
Referenced from: /Users/amzamani/Desktop/incoming/ai/Flowise/node_modules/canvas/build/Release/canvas.node
Reason: no suitable image found. Did find:
/Users/amzamani/Desktop/incoming/ai/Flowise/node_modules/canvas/build/Release/libpixman-<IP_ADDRESS>.dylib: cannot load 'libpixman-<IP_ADDRESS>.dylib' (load command 0x80000034 is unknown)
/Users/amzamani/Desktop/incoming/ai/Flowise/node_modules/canvas/build/Release/libpixman-<IP_ADDRESS>.dylib: cannot load 'libpixman-<IP_ADDRESS>.dylib' (load command 0x80000034 is unknown)
module<EMAIL_ADDRESS>task: toCached
plugin: flowise
root: /Users/amzamani/Desktop/incoming/ai/Flowise/packages/server
See more details with DEBUG=*
(Use node --trace-warnings ... to show where the warning was created)
› Error: command start not found
error Command failed with exit code 2.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
error Command failed with exit code 2.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
Abus-MacBook-Air:Flowise amzamani$
`
MacOS Mojave- Version 10.14.4
Node- Version v18.17.1
Hi @amzamani.
I believe your issue is related https://github.com/FlowiseAI/Flowise/issues/556
but do try sudo yarn from install to start
I guess it is related with node-canvas. I did not faced this problem but i may suggest to this path to solve it:
Take a look: https://github.com/Automattic/node-canvas#compiling
Install them:
brew install pkg-config cairo pango libpng jpeg giflib librsvg pixman
then;
remove node_modules (remove build artifacts)
yarn install
yarn build
yarn dev
if it wont succeed i can also suggest to install this
npm install canvas --build-from-source
then reproduce steps starting with remove node_modules
|
2025-04-01T04:55:05.410262
| 2023-10-15T21:09:55
|
1944058658
|
{
"authors": [
"HenryHengZJ",
"matthias"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13963",
"repo": "FlowiseAI/Flowise",
"url": "https://github.com/FlowiseAI/Flowise/pull/1066"
}
|
gharchive/pull-request
|
Feature/PluginSystem
In this PR I propose a Plugin & hook system for Flowise
Plugins are packages, that currently live in /plugins - but they should also be installable via npm
It borrows ideas from the WordPress plugin and hook system. One of the most powerful forces why WordPress became such a popular OpenSource framework (despite a lot of not so likable things about WordPress code)
Within a plugin package we use package.json/main to point to the plugins main class which extends the FlowisePlugin class
It automatically loads components and credentials from the plugins directory by specifying nodesPath and credentialsPath
=> Check the components in the UI within a new "Custom" section
Under the hood it's using an event based hook system (based on EventEmitter) that adds the ability to collect and manipulate (filter) data, via emitting and listening to events (we call them hooks)
p.s.
Some changes where introduced by yarn lint-fix
p.p.s.
Sorry for the first broken commit/pr - pls ignore / delete
hey @matthias thanks a lot for the initiative! Am wondering whats the difference between plugin vs users creating their own node in packages/components/nodes?
hey @HenryHengZJ - thanks for looking into this.
I think it's the much cleaner separation than working directly in the same folder / code structure as the core project.
I thought about it since I'm working with Flowise. Coming from WordPress I was wondering why it's not more common to have a plugin/hook architecture in Node/JS projects. Then I investigated deeper into langchain recently - and discovered that they provide a mechanism to have custom nodes in a separate folder - which I liked very much. I elaborated this idea, because I could imagine that from a certain point you don't want to have 100th of nodes in the core / UI - but rather be able to install additional "packages" of functionality (not just nodes)
Currently it needs a very deep dive into the code until you know where to add what - and there is the constant need of updating and merging code bases. Adding own features by forking is possible, but I would rather not do it in the long run as it constantly requires to keep up with all the codebase (as you never know what could change) - and solve merging conflicts.
Actually adding / loading nodes from a plugin was just a demo use case of the hooks. They (hooks) are the important part, as they allow Flowise to add clear entry points how to add things. This could also be additional navigation points or more down the road.
Imagine other use cases, as adding / managing / observing sources. Adding alternative chat UIs. Adding api endpoints (routes).
Could be interesting for custom(er) projects based on Flowise. But could also be a mechanism to develop new features, before they become part of the core.
|
2025-04-01T04:55:05.546400
| 2023-05-13T00:07:16
|
1708355584
|
{
"authors": [
"corneliusyaovi",
"pastramahodu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13964",
"repo": "Flutterwave/Angular",
"url": "https://github.com/Flutterwave/Angular/pull/21"
}
|
gharchive/pull-request
|
Bug bounty test - PLEASE DO NOT APPROVE. ijoech
Bug bounty test - PLEASE DO NOT APPROVE. ijoech
Spam pull request
|
2025-04-01T04:55:05.560494
| 2023-04-13T18:11:50
|
1666905740
|
{
"authors": [
"gouttegd"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13965",
"repo": "FlyBase/drosophila-anatomy-developmental-ontology",
"url": "https://github.com/FlyBase/drosophila-anatomy-developmental-ontology/issues/1621"
}
|
gharchive/issue
|
Obsolete neuroepidermoblast
neuroepidermoblast is a subclass of neuroblast.
The term has no definition and is not used anywhere within FBbt. In FlyBase, it is used in one record (P{E(spl)m8-HLH-2.61}). The corresponding paper never uses the term neuroepidermoblast. It seems the term is intended to refer to cells of the proneural clusters of the neurectoderm – cells that have not yet committed to either the neuroblast or the epidermoblast fate.
The term does not seem used in the literature at all. PubMed returns zero result; Google returns a handful of results, which are all ultimately linked either to the FBbt term directly or to the equivalent term in CL (which is almost certainly a direct “copy” of the FBbt term).
I think the term should be obsoleted. In the one record where it has been used, it can probably be replaced by proneural cluster.
The term could also get a proper definition as “cell of a proneural cluster that will become either a neuroblast or an epidermoblast” or similar, rather than being obsoleted.
|
2025-04-01T04:55:05.637996
| 2017-07-12T18:41:56
|
242476859
|
{
"authors": [
"tas50"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13968",
"repo": "Foodcritic/foodcritic",
"url": "https://github.com/Foodcritic/foodcritic/pull/653"
}
|
gharchive/pull-request
|
FC001 should not alert on node.run_state[:foo]
Signed-off-by: Tim Smith<EMAIL_ADDRESS>
Resolves https://github.com/Foodcritic/foodcritic/issues/652
|
2025-04-01T04:55:05.810401
| 2024-12-24T07:59:21
|
2757371335
|
{
"authors": [
"JeyesHan",
"JinYue2015"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13969",
"repo": "FoundationVision/Infinity",
"url": "https://github.com/FoundationVision/Infinity/issues/10"
}
|
gharchive/issue
|
What does this line work for? Thanks.
https://github.com/FoundationVision/Infinity/blob/0ac8556fa2cc23c70c342785e53145c5d6b09d46/infinity/models/infinity.py#L433
This line is to quick jump to the implementation of [SelfAttnBlock.forward, CrossAttnBlock.forward]. It's for reference only and no other special usuage.
|
2025-04-01T04:55:05.817067
| 2024-08-07T03:20:39
|
2452193793
|
{
"authors": [
"Lucky-Light-Sun",
"iFighting"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13970",
"repo": "FoundationVision/VAR",
"url": "https://github.com/FoundationVision/VAR/issues/80"
}
|
gharchive/issue
|
Question about the training dataset - tokenzier
Hi, I recently read your paper “Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction” published in CVPR 2024. I am very interested in your method and have great respect for your contribution to the academic community.
We are currently doing some experiments and would like to ask you a question about the tokenizer part. Is the pre-training dataset you are using OpenImage V6 or another version? Current we are using OpenImage v6 to train tokenizer. So do you think it's fair to compare with VAR tokenizer?
Thank you again for your time and effort, and I look forward to your reply.
Hi, I recently read your paper “Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction” published in CVPR 2024. I am very interested in your method and have great respect for your contribution to the academic community.
We are currently doing some experiments and would like to ask you a question about the tokenizer part. Is the pre-training dataset you are using OpenImage V6 or another version? Current we are using OpenImage v6 to train tokenizer. So do you think it's fair to compare with VAR tokenizer?
Thank you again for your time and effort, and I look forward to your reply.
Hi, I recently read your paper “Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction” published in CVPR 2024. I am very interested in your method and have great respect for your contribution to the academic community.
We are currently doing some experiments and would like to ask you a question about the tokenizer part. Is the pre-training dataset you are using OpenImage V6 or another version? Current we are using OpenImage v6 to train tokenizer. So do you think it's fair to compare with VAR tokenizer?
Thank you again for your time and effort, and I look forward to your reply.
@Lucky-Light-Sun
Sorry for late reply.
The pre-training dataset we are using is OpenImage V6.
BTW, VAR is accepted as NeurIPS oral, not CVPR
|
2025-04-01T04:55:05.861271
| 2022-07-05T16:49:48
|
1294536571
|
{
"authors": [
"Pedrome203",
"andyowli"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13971",
"repo": "FranckFreiburger/vue-pdf",
"url": "https://github.com/FranckFreiburger/vue-pdf/issues/375"
}
|
gharchive/issue
|
Call the browser printing function, and a blank page will appear in PDF.
When invoking the browser print function, garbled code will appear at first, and then it will be modified according to pull requests #130, and the text will be displayed normally, but there will be an extra blank page between pages. Does anyone encounter the same problem? How to solve it?
Hi, were you able to solve it?
|
2025-04-01T04:55:05.866425
| 2019-01-28T02:42:30
|
403630238
|
{
"authors": [
"FrangSierra",
"KryptKode",
"abdul-hasib",
"gigiojr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13972",
"repo": "FrangSierra/RxFirebase",
"url": "https://github.com/FrangSierra/RxFirebase/issues/76"
}
|
gharchive/issue
|
subscribe when query is empty
I'm try to get an user profile with multiples nodes of a realtime database with this code:
public static void getUserProfile(String storeId, String userId, final Callback callback){
DatabaseReference userRef = RealtimeDatabase.getUserRef(userId);
DatabaseReference addressRef = RealtimeDatabase.getUserAddressRef(userId);
DatabaseReference contactRef = RealtimeDatabase.getUserContactRef(userId);
DatabaseReference storeUserRef = RealtimeDatabase.getStoreUserRef(storeId, userId);
DatabaseReference storeNameRef = RealtimeDatabase.getStoreRef(storeId).child(Store.PARAM_NAME);
RxFirebaseDatabase.observeSingleValueEvent(userRef, User.class)
.zipWith(RxFirebaseDatabase.observeSingleValueEvent(addressRef, Address.class), zipUserAddress)
.zipWith(RxFirebaseDatabase.observeSingleValueEvent(contactRef, UserContact.class), zipUserContact)
.zipWith(RxFirebaseDatabase.observeSingleValueEvent(storeUserRef, StoreUser.class), zipUserStore)
.zipWith(RxFirebaseDatabase.observeSingleValueEvent(storeNameRef, String.class), zipUserStoreName)
.subscribe(new Consumer<User>() {
@Override
public void accept(User user) {
callback.onSuccess(user);
}
}, new Consumer<Throwable>() {
@Override
public void accept(Throwable throwable) {
throwable.printStackTrace();
callback.onError(throwable);
}
});
}
This code works well to a complete profile, but when profile does not have address node or contact node, for example, the function in subscribe is never called and this is a problem for me now. Can you help me?
I am into the same issue, is there any update on this error?
I have created 3 flowable objects and then I called .zip to get the output of these 3 flowables. The issue is that it works well if all the 3 flowable returns value, if any of the query fails to find documents based on query .zip gets stuck and it never calls onNext or onComplete
@abdul-hasib Use Flowable.merge instead of Flowable.zip. The zip operator waits for all the observables to emit before it triggers onNext
@KryptKode is right. Merge instead of Zip should fix it.
Sorry for the delay on the response.
|
2025-04-01T04:55:05.868964
| 2021-11-24T23:53:06
|
1063006629
|
{
"authors": [
"Frankkkkk",
"demoze"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13973",
"repo": "Frankkkkk/python-pylontech",
"url": "https://github.com/Frankkkkk/python-pylontech/issues/8"
}
|
gharchive/issue
|
wonderfull,
wonderfull,
Exactly what I'am looking for to translate JK-BMS rs485 protocole (or other BMS) to pylontech rs485 protocole , and then have my voltronic/mppsolar inverter talk to my DIY battery.
Thank you.
Thanks @demoze ! 😀
|
2025-04-01T04:55:05.889498
| 2024-12-17T00:14:48
|
2743659541
|
{
"authors": [
"sitapriyamoorthi",
"tefirman"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13974",
"repo": "FredHutch/wdl-unit-tests",
"url": "https://github.com/FredHutch/wdl-unit-tests/pull/24"
}
|
gharchive/pull-request
|
Adding contribution guidelines and code of conduct
Adding Unit test workflow to contribution guide
Moved contribution guidelines from README to .github/CONTRIBUTING.md.
|
2025-04-01T04:55:05.898108
| 2024-03-28T22:14:00
|
2214256068
|
{
"authors": [
"Fredx87",
"anulman"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13975",
"repo": "Fredx87/openapi-io-ts",
"url": "https://github.com/Fredx87/openapi-io-ts/pull/27"
}
|
gharchive/pull-request
|
Fix request body file imports
Replace custom imports with the generic getImports() when writing request body files. I noticed that anything that referenced the DateFromISOString type was breaking.
Thank you @anulman for the recent PRs!
Unfortunately, I am not planning to maintain this library anymore, since I stopped using io-ts and fp-ts. I would like to rewrite the library for effectand@effect/schema`, but it is not going to happen soon.
That's all right! We've been maintaining our own fork at nmbrco/openapi-io-ts; was mostly trying to contrib back to your work in appreciation :)
Tbqh we were on zod before, but needed more out of generation than their ecosystem supported. Your work in comparison was much easier to hack on, even with the io-ts learning curve.
Thanks for publishing :)
|
2025-04-01T04:55:05.968328
| 2016-11-13T16:32:58
|
188979378
|
{
"authors": [
"Manish-Giri",
"bstonesifer"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13976",
"repo": "FreeCodeCamp/FreeCodeCamp",
"url": "https://github.com/FreeCodeCamp/FreeCodeCamp/issues/11643"
}
|
gharchive/issue
|
help please
Challenge Name
Issue Description
Browser Information
cant get the my test to pass.
i have compared my code to others and ask for help in chat..
Browser Name, Version: firefox newest...
At the very least, you haven't applied the smaller-image class to the img element, which is why your image isn't resizing.
Please visit the Help Room if you need assistance on this.
been to the help room.. its not helpful..
@bstonesifer Feel free to PM me on gitter if you're still stuck.
|
2025-04-01T04:55:05.970656
| 2015-08-18T22:37:29
|
101764302
|
{
"authors": [
"AryanJ-NYC",
"jmichb"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13977",
"repo": "FreeCodeCamp/FreeCodeCamp",
"url": "https://github.com/FreeCodeCamp/FreeCodeCamp/issues/2128"
}
|
gharchive/issue
|
Error seems to be in error
Challenge http://www.freecodecamp.com/challenges/waypoint-make-instances-of-objects-with-a-constructor-function has an issue. Please describe how to reproduce it, and include links to screenshots if possible.
To reproduce it, run the correct code. Then you'll see that the error reported seems to be, well, in error:
assert(typeof(myCar.engines) === 'number', 'The property engine of myCar should be a number');The property engine of myCar should be a number
See the screenshot.
engines, not engine. You used engine in your code.
|
2025-04-01T04:55:05.974472
| 2015-11-22T07:48:02
|
118249907
|
{
"authors": [
"3laa-Khalil-Alarabi",
"raisedadead"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13978",
"repo": "FreeCodeCamp/FreeCodeCamp",
"url": "https://github.com/FreeCodeCamp/FreeCodeCamp/issues/4631"
}
|
gharchive/issue
|
Arabic translation of Basic Algorithm Scripting
[ ] Meet Bonfire
[ ] Reverse a String
[ ] Factorialize a Number
[ ] Check for Palindromes
[ ] Find the Longest Word in a String
[ ] Title Case a Sentence
[ ] Return Largest Numbers in Arrays
[ ] Confirm the Ending
[ ] Repeat a string repeat a string
[ ] Truncate a string
[ ] Chunky Monkey
[ ] Slasher Flick
[ ] Mutations
[ ] Falsy Bouncer
[ ] Seek and Destroy
[ ] Where do I belong
Closing in favor of https://github.com/FreeCodeCamp/TranslationExpansion/issues/24
|
2025-04-01T04:55:05.978810
| 2016-01-11T03:03:11
|
125865182
|
{
"authors": [
"JBuji",
"KeaDC",
"sludge256"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13979",
"repo": "FreeCodeCamp/FreeCodeCamp",
"url": "https://github.com/FreeCodeCamp/FreeCodeCamp/issues/6046"
}
|
gharchive/issue
|
Waypoint: Add Placeholder Text to a Text Field ( bug in placeholder exercise)
has an issue.
User Agent is: Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.106 Safari/537.36.
Please describe how to reproduce this issue, and include links to screenshots if possible.
My code:
<link href="http://fonts.googleapis.com/css?family=Lobster" rel="stylesheet" type="text/css">
<style>
.red-text {
color: red;
}
h2 {
font-family: Lobster, Monospace;
}
p {
font-size: 16px;
font-family: Monospace;
}
.thick-green-border {
border-color: green;
border-width: 10px;
border-style: solid;
border-radius: 50%;
}
.smaller-image {
width: 100px;
}
</style>
<h2 class="red-text">CatPhotoApp</h2>
<p>Click here for <a href="#">cat photos</a>.</p>
<a href="#"><img class="smaller-image thick-green-border" alt="A cute orange cat lying on its back" src="https://bit.ly/fcc-relaxing-cat"></a>
<p>Things cats love:</p>
<ul>
<li>cat nip</li>
<li>laser pointers</li>
<li>lasagna</li>
</ul>
<p>Top 3 things cats hate:</p>
<ol>
<li>flea treatment</li>
<li>thunder</li>
<li>other cats</li>
</ol>
<input type="text">
<input type="text" placeholder="cat photo URL">
(the above line is reading as incorrect )
Name of exercise: Waypoint: Add Placeholder Text to a Text Field
You're supposed to add the placeholder text to the existing input not create a new input.
I passd "Add Placeholder Text to a Text Field"
challange puting TWO lines:
But it's strange.
I passed "Add Placeholder Text to a Text Field"
challenge putting TWO lines:
<input type="text" placeholder="cat photo URL">
<input type="text" placeholder="cat photo URL">
But it's strange. /FCC bug/
Sorry I'm new on FCC & I've lost this lines in my earlier Comment
|
2025-04-01T04:55:05.980969
| 2015-06-05T00:15:10
|
85341484
|
{
"authors": [
"MightyBoosh92",
"QuincyLarson"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13980",
"repo": "FreeCodeCamp/freecodecamp",
"url": "https://github.com/FreeCodeCamp/freecodecamp/issues/841"
}
|
gharchive/issue
|
Completed Challenge without creating "CSS blue" class
Challenge http://www.freecodecamp.com/challenges/waypoint-override-styles-with-important has an issue.
I was able to complete this challenge without doing the "Create the CSS class "blue-text"" step.
Thank you for your detailed explanation and screen shots. One of our campers has submitted a fix for this.
|
2025-04-01T04:55:05.982289
| 2016-12-22T11:51:07
|
197158096
|
{
"authors": [
"QuincyLarson",
"raisedadead"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13981",
"repo": "FreeCodeCamp/open2017",
"url": "https://github.com/FreeCodeCamp/open2017/pull/14"
}
|
gharchive/pull-request
|
feat(guests) : Add links to social profiles
Adds a social link from the twitter handles and falls back to Github handle.
P.S Demo available at https://raisedadead.com/open2017/
@raisedadead awesome! Well done!
|
2025-04-01T04:55:06.048345
| 2020-04-10T22:13:59
|
598117701
|
{
"authors": [
"jacobopantoja"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13982",
"repo": "FreeRDP/FreeRDP",
"url": "https://github.com/FreeRDP/FreeRDP/issues/6058"
}
|
gharchive/issue
|
USB redirection not working in non-Windows platforms
#It seems that commit ebdaea6f3b5e410b9a3380368d6023c52bed67cc introduces a change targeted to Windows builds, without effectively checking it, therefore breaking the USB redirection functionality in non-Windows builds.
When loading the urbdrc channel in Linux build, you receive the following error:
[00:01:01:900] [2961:2962] [WARN][com.freerdp.channels.urbdrc.client] - LIBUSB_OPTION_USE_USBDK Operation not supported or unimplemented on this platform [-12]
Simply adding the adequate ifdef fixes the problem
Pull request issued #6057
Fixed in 0f8437d1d86b2798edd46033974eb9678b6eecf0. Closing
|
2025-04-01T04:55:06.150384
| 2024-09-10T15:18:28
|
2516763491
|
{
"authors": [
"Frezyx",
"Luwirispok"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13983",
"repo": "Frezyx/talker",
"url": "https://github.com/Frezyx/talker/issues/260"
}
|
gharchive/issue
|
TalkerDataCard does not expand when changing title
Describe the bug
If you change the title in TalkerSettings in TalkerFlutter.init, TalkerDataCard stops responding to keystrokes, and responds to settings.
To Reproduce
Steps to reproduce the behavior:
Connect talker_flutter, talker_dio_logger packages
Create TalkerFlutter.init
In TalkerSettings, change the titles to custom headers (at least network requests)
Add TalkerDioLogger to dio.interceptors
Add TalkerScreen call to TalkerSettings
Make the request in the working application
Click on the network request logs in TalkerScreen
Error. TalkerDataCard with custom title has not changed
Expected behavior
Ability to expand TalkerDataCard and see detailed request information
Screenshots
https://github.com/user-attachments/assets/f20941ca-79f9-4c53-97e9-034672f550f1
Desktop (please complete the following information):
Smartphone (please complete the following information):
Device: Xiaomi MI 9T.
OS: Android 11.
Device: Blackview Active 8 Pro.
OS: Android 13.
Additional context
Spotted in log requests, responses, and network errors because it shows a lot of data.
Probably an issue with any logs with detailed data
Package version
talker_dio_logger: ^4.4.1
talker_bloc_logger: ^4.4.1
talker_flutter: ^4.4.1
[✓] Flutter (Channel stable, 3.24.1, on macOS 14.5 23F79 darwin-arm64, locale ru-RU)
• Flutter version 3.24.1 on channel stable at /Users/specialny/tools/flutter
• Upstream repository https://github.com/flutter/flutter.git
• Framework revision 5874a72aa4 (3 недели назад), 2024-08-20 16:46:00 -0500
• Engine revision c9b9d5780d
• Dart version 3.5.1
• DevTools version 2.37.2
[✓] Android toolchain - develop for Android devices (Android SDK version 34.0.0)
• Android SDK at /Users/specialny/Library/Android/sdk
• Platform android-34, build-tools 34.0.0
• Java binary at: /Applications/Android Studio.app/Contents/jbr/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 17.0.11+0-17.0.11b1207.24-11852314)
• All Android licenses accepted.
[✓] Xcode - develop for iOS and macOS (Xcode 15.4)
• Xcode at /Applications/Xcode.app/Contents/Developer
• Build 15F31d
• CocoaPods version 1.14.3
[✓] Chrome - develop for the web
• Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome
[✓] Android Studio (version 2024.1)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/9212-flutter
• Dart plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/6351-dart
• Java version OpenJDK Runtime Environment (build 17.0.11+0-17.0.11b1207.24-11852314)
[✓] VS Code (version 1.92.0)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.94.0
[✓] Connected device (4 available)
• Mi 9T (mobile) • e7b48232 • android-arm64 • Android 11 (API 30)
• macOS (desktop) • macos • darwin-arm64 • macOS 14.5 23F79 darwin-arm64
• Mac Designed for iPad (desktop) • mac-designed-for-ipad • darwin • macOS 14.5 23F79 darwin-arm64
• Chrome (web) • chrome • web-javascript • Google Chrome 128.0.6613.120
[✓] Network resources
• All expected network resources are available.
• No issues found!
Hello @Luwirispok !
Can you explain how request logs looks in your VSCode console ?
|
2025-04-01T04:55:06.160313
| 2024-03-01T14:17:50
|
2163511688
|
{
"authors": [
"jos0405",
"pavanobbeli"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13985",
"repo": "FriendlyDotCH/mautic-multi-domain",
"url": "https://github.com/FriendlyDotCH/mautic-multi-domain/issues/25"
}
|
gharchive/issue
|
Unable to install plugin
Team, I have manually installed this plugin https://github.com/FriendlyDotCH/mautic-multi-domain?tab=readme-ov-file and I am receiving an error when attempting to access Mautic. Could someone help me to resolve this issue?
Are you working with Mautic 4 or 5? This is not M5 compatible yet
|
2025-04-01T04:55:06.168369
| 2013-04-22T10:16:37
|
13473188
|
{
"authors": [
"bamarni",
"lyrixx"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13986",
"repo": "FriendsOfPHP/Sismo",
"url": "https://github.com/FriendsOfPHP/Sismo/issues/100"
}
|
gharchive/issue
|
Add author email to Commit object
What would you think about adding author's email to commit objects?
My use case is to add at runtime commit's author as a notifier recipient :
<?php
use Sismo\Commit;
use Sismo\Notifier\MailNotifier;
class Notifier extends MailNotifier
{
public function notify(Commit $commit)
{
/*
$authorEmail = $commit->getAuthorEmail(); // ???
$recipients = (array) $this->recipients;
if (!in_array($authorEmail, $recipients)) {
$this->recipients[] = $authorEmail;
}
*/
$previousCommit = $commit->getProject()->getLatestCommit();
if (!$commit->isSuccessful()
|| !$previousCommit
|| $previousCommit->getStatusCode() != $commit->getStatusCode()
) {
return parent::notify($commit);
}
return false;
}
}
ping @bamarni
hum it's hanging since a while :) in the meanwhile I switched to a cloud based solution so I can't implement this at the moment unfortunately..
|
2025-04-01T04:55:06.175559
| 2014-07-04T18:09:21
|
37181630
|
{
"authors": [
"perk11",
"stevro"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13987",
"repo": "FriendsOfSymfony/FOSCommentBundle",
"url": "https://github.com/FriendsOfSymfony/FOSCommentBundle/pull/441"
}
|
gharchive/pull-request
|
Using Symfony\Component\Validator\Context\LegacyExecutionContext
Quick fix so voting works in Symfony 2.5.
Has anyone provided a fix for this issue? Thanks.
|
2025-04-01T04:55:06.242744
| 2019-06-24T19:43:45
|
460062695
|
{
"authors": [
"fubar-coder",
"lfalck"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13988",
"repo": "FubarDevelopment/FtpServer",
"url": "https://github.com/FubarDevelopment/FtpServer/issues/72"
}
|
gharchive/issue
|
Error with validateScopes = true using version 3.0rc4
Hi!
When i build the service provider with validation of scopes like this:
services.BuildServiceProvider(validateScopes: true)
I get this error:
Cannot resolve scoped service 'FubarDev.FtpServer.Commands.IFtpCommandHandlerProvider' from root provider.
ValidateScopes is also true when using CreateDefaultBuilder in ASP.NET Core which is where i encountered the issue.
Here is a minimal example of the issue, if i remove validateScopes i can connect but with it the application crashes.
Great work with 3.0, looking foward to the release!
/Ludvig
Fixed in RC5 (should be available soon)
Great, thanks!
|
2025-04-01T04:55:06.264936
| 2015-06-02T06:57:51
|
83877883
|
{
"authors": [
"gokulkrishh",
"logeshpaul",
"rgksugan"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13989",
"repo": "FuelFrontend/generator-smacss",
"url": "https://github.com/FuelFrontend/generator-smacss/issues/46"
}
|
gharchive/issue
|
Not listed in yeoman.io
This generator is not listed in the yeoman.io site
Updated the package.json file based on the reply - https://twitter.com/Vaxilart/status/592434542503469056
@logeshpaul @rgksugan Our generator is know listed in yeoman.io site. Check the screenshot.
@gokulkrishh cool! There was an issue with the 'repo:url' in 'package.json' file. I fixed it and pushed it along with other fix we I did yesterday. Great that it fixed and listed!
Thanks for checking :+1:
Posting the line number here for reference https://github.com/FuelFrontend/generator-smacss/blob/9cb1ea13733101594b05abfd9b8d4ba488a7f96d/package.json#L35
Cool!!
|
2025-04-01T04:55:06.268133
| 2022-04-14T08:27:43
|
1204212908
|
{
"authors": [
"adlerjohn",
"binggh",
"mitchmindtree"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13990",
"repo": "FuelLabs/sway-vscode-plugin",
"url": "https://github.com/FuelLabs/sway-vscode-plugin/issues/42"
}
|
gharchive/issue
|
sway vscode plugin crashes on start up
The sway vscode plugin immediately crashes on startup.
This is probably due to the changes to the LSP feature becoming a plugin instead of a native command: https://github.com/FuelLabs/sway/pull/1178
Error: no such subcommand: `lsp`
[Info - 4:21:53 PM] Connection to server got closed. Server will restart.
Error: no such subcommand: `lsp`
[Info - 4:21:53 PM] Connection to server got closed. Server will restart.
Error: no such subcommand: `lsp`
[Info - 4:21:53 PM] Connection to server got closed. Server will restart.
Error: no such subcommand: `lsp`
[Info - 4:21:53 PM] Connection to server got closed. Server will restart.
Error: no such subcommand: `lsp`
[Error - 4:21:53 PM] Connection to server got closed. Server will not be restarted.
@binggh you might just need to install the forc-lsp plugin - could you try cargo install forc-lsp, then re-launch VS Code and see if that works?
I haven't looked into how VS Code plugins work yet myself (I do my editing with Vim), but perhaps there's some way we can make sure forc-lsp is installed during the plugin installation process.
I think this should mostly be resolved once we have binary distributions. In the meantime, if that is indeed the fix, then The Book can simply be update to include instruction to install the lsp and other important plugins.
@mitchmindtree oops, my bad - that fixed it 😅 will close this issue and update the docs at https://fuellabs.github.io/sway/latest/introduction/sway-toolchain.html
|
2025-04-01T04:55:06.273105
| 2021-01-20T15:06:05
|
790062444
|
{
"authors": [
"lknegendorf"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13991",
"repo": "Fuenfgeld/ATeamDatenmanagementUndArchivierung",
"url": "https://github.com/Fuenfgeld/ATeamDatenmanagementUndArchivierung/issues/22"
}
|
gharchive/issue
|
Identify dependencies in the data needed for our research
Depending on the results of #21!
For example code->description and reasoncode->reasondescription
These "translations" should not be in the main table but in the satellite-tables of our star scheme
CODE -> DESCRIPTION
CODE -> DESCRIPTION
|
2025-04-01T04:55:06.277422
| 2017-05-17T21:32:31
|
229493635
|
{
"authors": [
"coveralls",
"kangkyu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13992",
"repo": "Fullscreen/yt-support",
"url": "https://github.com/Fullscreen/yt-support/pull/6"
}
|
gharchive/pull-request
|
Add as_curl
I borrowed as_curl method from 'yt' gem. and think it could be useful for debugging projects which use 'yt-core' gem. Thank you
Coverage decreased (-4.3%) to 95.726% when pulling caaf5c2f40f6936623044eb48692b20cfbdedfb9 on kangkyu:as_curl into 38c6892b8a9a853b50fc47ef748f7a64dfcbe882 on Fullscreen:master.
Coverage remained the same at 100.0% when pulling 6133e0193fdf6d0327bf7b7d542e85c896b443ca on kangkyu:as_curl into 38c6892b8a9a853b50fc47ef748f7a64dfcbe882 on Fullscreen:master.
closes in favor of #8
|
2025-04-01T04:55:06.313197
| 2024-11-08T07:33:16
|
2643146600
|
{
"authors": [
"jamesbraza"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13994",
"repo": "Future-House/aviary",
"url": "https://github.com/Future-House/aviary/pull/117"
}
|
gharchive/pull-request
|
Fixing CI by downpinning uv<0.5
From this CI run, we get:
× No solution found when resolving dependencies for split
│ (python_full_version == '3.11.*' and platform_python_implementation ==
│ 'PyPy'):
╰─▶ Because only the following versions of paper-qa are available:
paper-qa<=5.0.0
paper-qa==5.0.1
...
paper-qa==5.3.2
and paper-qa==5.3.2 depends on fhaviary, we can conclude that
paper-qa>5.3.1 depends on fhaviary.
And because paper-qa>=5.0.0,<=5.3.1 depends on fhaviary, we can conclude
that paper-qa>=5.0.0 depends on fhaviary.
And because fhaviary[paperqa] depends on paper-qa>=5 and your workspace
requires fhaviary[paperqa], we can conclude that your workspace's
requirements are unsatisfiable.
hint: The package `paper-qa` depends on the package `fhaviary` but the
name is shadowed by one of your workspace members. Consider changing the
name of the workspace member.
Something about uv==0.5.0 broke our "circular" pinning with paper-qa (where paper-qa depends on fhaviary, and we have a convenience paperqa extra here). This PR just temporarily down-pins uv for this
Looks like test passed this time, but lint still fails. Need to dig deeper
|
2025-04-01T04:55:06.316430
| 2021-09-02T09:33:10
|
986449759
|
{
"authors": [
"Fyrd",
"j-hoffmann"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13995",
"repo": "Fyrd/caniuse",
"url": "https://github.com/Fyrd/caniuse/issues/6009"
}
|
gharchive/issue
|
Missing browser usage data for iOS in Switzerland
Usage data for Safari / Chrome on iOS devices seems to be missing. My screenshot below shows an example for browser versions 14.5 - 14.7, but all iOS versions are affected and are showing 0% for Switzerland. These numbers should be higher than global numbers since market share of iOS is generally higher in Switzerland.
Numbers for other browsers, including Safari on macOS do not seem to be affected.
I cannot tell exactly when this started happening, but it must be at least a few days.
I believe this was recently fixed. Note that you may need to delete the current data then re-import.
Thank you. I thought I had done a reimport, but trying it now helped. Problem fixed.
|
2025-04-01T04:55:06.326466
| 2022-10-10T20:19:09
|
1403668214
|
{
"authors": [
"dave-gantenbein",
"kannon92",
"suprjinx"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13996",
"repo": "G-Research/armada",
"url": "https://github.com/G-Research/armada/issues/1608"
}
|
gharchive/issue
|
ArmadaOperator: Add airflow task id to Armada annotations
When a user submits an Airflow task, we should add the airflow task id to the Armada annotations.
An example can be below:
jobSetId: test
priority: 100
namespace: batch-test
annotations:
armadaproject.io/taskRunId: 0a016adf-8325-1d6a-8183-268de7a625ea
@iJanki-gr Is there any other items in the airflow context that you would want added as annotations?
I believe everything is in place to add the annotations, we just need to read them from airflow context and create them in the job request.
Hey team! Please add your planning poker estimate with Zenhub @ClifHouck @dejanzele @headphonejames @kannon92 @richscott @Sharpz7 @suprjinx
@kannon92 it seems like airflow has a "task_id" and a "run_id" as separate identifiers -- which should be used for "taskRunId" in the description?
Please hold on this. I need to meet with Danielle to discuss airflow user request but it will have to happen after Kubecon.
So @iJanki-gr and I meant to sync up before I left for Kubecon. He has some ideas of what potential airflow users would want from this work but I think we need to see some DAGs and see what is possible.
At this moment, we use dagrunId as our job-set name. There was a request to dynamically use task name from armada job ids but I don't know if that is possible. Airflow seems to want static task names but it seems possible to do dynamic generation of dags but it seems to be an advanced feature so I think more information is necessary. Feel free to take lead on this if you want with Daniele.
|
2025-04-01T04:55:06.336670
| 2023-01-01T05:07:11
|
1515319947
|
{
"authors": [
"G0maa"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13997",
"repo": "G0maa/my-school-server",
"url": "https://github.com/G0maa/my-school-server/issues/32"
}
|
gharchive/issue
|
Use of middlewares for zod
As suggested by @Amr2812
Reference: https://github.com/G0maa/my-school-server/blob/a8971d8341a0ff4440dc9c9abbc886de34d3aaae/src/controllers/student.ts#L23-L24
and all other controllers.
Goal: Single middleware use, single schema to fulfill the above need.
Revisiting.
I've been trying to adapt this to the current code, but I haven't been able to.
So far what I have got is
const validateSchemas =
(schemas: AnyZodObject[]) =>
(req: Request, _res: Response, next: NextFunction) => {
// Error handling in errror-handler middleware
// await/promise is needed only if you use **async** `refinements` or `transformers`
for (const schema of schemas) {
schema.parse({
body: req.body,
query: req.query,
params: req.params,
});
}
return next();
};
And
export const ZStudentPost = z.object({
body: ZStudent,
});
I cant use transform since it will invalidate the state of req.body:
export const ZUserPost = z
.object({
body: ZUser,
})
.transform((obj) => {
return { ...obj, user: obj.body };
});
Which means, either:
Find a way to parse req.body twice populating it with the needed schema i.e. req.body.user & req.body.student
Redesign schemas around routes themselves rather than entities.
I used ChatGPT for 3-4 tries to refine the suggested code:
function validateAndExtract(schemas, properties) {
return (req, res, next) => {
try {
let parsed = {};
schemas.forEach(schema => {
parsed = Object.assign(parsed, schema.parse(req.body));
});
properties.forEach(property => {
req.body[property] = parsed[property];
});
next();
} catch (err) {
return res.status(400).json({ message: err.message });
}
}
}
i.e. to use
validateAndExtract([teacherSchema, userSchema], ['teacher', 'user'])
Revisited, tried & applied:
Here it is used to verify entire request, but manually giving types to req, this hugely depends on client sending different objects, user, student & studentDetails in req
https://github.com/G0maa/my-school-server/blob/576c72f15ca8e0440695c4eb724b1f794d07f6b9/src/controllers/teacher.ts#L55-L67
Here it is too much to use a middleware for a small task:
https://github.com/G0maa/my-school-server/blob/576c72f15ca8e0440695c4eb724b1f794d07f6b9/src/controllers/teacher.ts#L92-L101
And here you can't use it, req.query can't be sent as objects.
https://github.com/G0maa/my-school-server/blob/576c72f15ca8e0440695c4eb724b1f794d07f6b9/src/controllers/teacher.ts#L28-L47
3rd method
https://github.com/G0maa/my-school-server/blob/359143fb8524864d11340294aaa43a5e75f07a9f/src/controllers/fee.ts#L38-L56
For Reference, old:
https://github.com/G0maa/my-school-server/blob/def70826f7e5d43e8f683b0816e4e686d51e6192/src/validator/studyClass.validator.ts
Solved with #45
|
2025-04-01T04:55:06.361381
| 2019-12-16T14:29:46
|
538447637
|
{
"authors": [
"nicolsonaitken"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:13999",
"repo": "GCTC-NTGC/TalentCloud",
"url": "https://github.com/GCTC-NTGC/TalentCloud/issues/2363"
}
|
gharchive/issue
|
Bug - Unable to save changes to Profile b/c Password is incorrect.
Description
While logged into the Talent Cloud site, making changes to my Twitter handle, LinkedIn URL and tagline I am unable to save as there is an error message stating that: Current password isn't correct. As I am already logged in and made no changes to my password, I don't know why or how my password could be incorrect.
Technical Info
Device: Laptop
Operating System: MacOs
Browser: Chrome, Version 79.0.3945.79 (Official Build) (64-bit)
Steps to Reproduce
Steps to reproduce the behaviour. Example:
Go to Talent Cloud
Click on About Me
Scroll down to Account Information
Fill in additional details and Save Changes
Interaction is included within the video attached.
Video
Profile.mp4.zip
[ ] Regression test written.
Password is now within settings and the issue looks to be resolved.
|
2025-04-01T04:55:06.366602
| 2021-09-26T23:41:34
|
1007538455
|
{
"authors": [
"Suryanshomar7240",
"mohit355"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14000",
"repo": "GDSC-IIIT-Kalyani/Winter-of-Code-2.0-frontend",
"url": "https://github.com/GDSC-IIIT-Kalyani/Winter-of-Code-2.0-frontend/pull/22"
}
|
gharchive/pull-request
|
project Section
The Project section made is fully responsive but still, it can be improved in terms of animations filter option also add to filter those projects according to the topics
@Suryanshomar7240 Thanks for your contribution.
|
2025-04-01T04:55:06.374542
| 2021-07-12T08:54:38
|
941826515
|
{
"authors": [
"4ian",
"Bouh"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14001",
"repo": "GDevelopApp/GDevelop-extensions",
"url": "https://github.com/GDevelopApp/GDevelop-extensions/issues/181"
}
|
gharchive/issue
|
Remove The Kongregate Extension from The Extension Library
Description
From https://github.com/4ian/GDevelop/issues/2797
Kongregate does not accept games any more, which makes the extension useless. Probably add a gamejolt or newgrounds API extension to GDevelop 5 instead.
[x] Verification on Kongregate website
[x] Check how deleted extension act in GDevelop if we try to update it
Check how deleted extension act in GDevelop if we try to update it
Nothing bad should happen as existing games having it will continue to work, and it will be removed from the list of extensions, so it won't be listed and you can't "update" it.
Close by https://github.com/GDevelopApp/GDevelop-extensions/commit/a7f8ed33c5050d5ae9776bf8e48cc9a2c912485e
|
2025-04-01T04:55:06.377063
| 2021-10-08T19:28:35
|
1021402754
|
{
"authors": [
"Saad-Mahamood",
"tonysun9"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14002",
"repo": "GEM-benchmark/NL-Augmenter",
"url": "https://github.com/GEM-benchmark/NL-Augmenter/issues/316"
}
|
gharchive/issue
|
gender_neutral_rewrite Unresolved references to spaCy and Unresolved List reference
When running the gender_neutral_rewrite there are several unresolved references to the spacy_nlp variable. In particular on line:
Line 27: self.nlp = spacy_nlp if spacy_nlp else spacy.load("en_core_web_sm")
Please use from initialize import spacy_nlp to get a handle on the global spacy instance.
There is also an unresolved reference on Line 495: def generate(self, sentence: str) -> List[str]. List[str] is not resolvable. Should this be lower case? e.g. list[str]
Thanks for pointing this out. Made a PR here: https://github.com/GEM-benchmark/NL-Augmenter/pull/324
|
2025-04-01T04:55:06.379378
| 2018-01-23T16:48:16
|
290910256
|
{
"authors": [
"dbauszus-glx"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14003",
"repo": "GEOLYTIX/xyz",
"url": "https://github.com/GEOLYTIX/xyz/issues/19"
}
|
gharchive/issue
|
Generate database array from environmental settings
Currently the database array is hardcoded at the beginning of modules which use it.
const databases = {
xyz: pgp(process.env.POSTGRES),
ghs: pgp(process.env.POSTGRES_GHS)
};
This array should be generated from the launch settings. Split name at underscore with the second part being the name of the database connection which is used in this instance.
This is now possible. Database connection strings begin with DBS_ the keys are split at the underscore and added to a DBS array. The second part of the key is the name of the database connection which can be referenced in queries to the framework.
let pgp = require('pg-promise')({ promiseLib: require('bluebird'), noWarnings: true }); const DBS = {}; Object.keys(process.env).map(function (key) { if (key.split('_')[0] === 'DBS') DBS[key.split('_')[1]] = pgp(process.env[key]) });
|
2025-04-01T04:55:06.381478
| 2023-04-24T18:19:18
|
1681798348
|
{
"authors": [
"gmao-rreichle",
"mathomp4"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14004",
"repo": "GEOS-ESM/GEOS_Util",
"url": "https://github.com/GEOS-ESM/GEOS_Util/issues/15"
}
|
gharchive/issue
|
remap_upper.py should pass in stretch factors to interp_restarts
After talking with @bena-nasa, the right way to pass in stretch grid factors in remap_upper.py is to do so through command line arguments to interp_restarts.x.
If you do it this way, the restarts get global metadata for the stretch factor, lon, and lat.
Oh dear. I tried this and for some reason the winds (and other things) are very different when you a namelist vs options.
@bena-nasa is looking at this.
Addressed in #19
|
2025-04-01T04:55:06.389968
| 2024-12-20T15:33:51
|
2752969089
|
{
"authors": [
"Dooruk"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14005",
"repo": "GEOS-ESM/swell",
"url": "https://github.com/GEOS-ESM/swell/issues/490"
}
|
gharchive/issue
|
(SE) Utilizing more Cylc features
One of the main reasons we use Cylc is its design which includes many useful features for DA cycling. We currently exercise some of them but we should take advantage of more. I will create this as an epic and as users think of more features they can add this can expand.
Necessary ones:
[ ] Retry certain tasks in case they fail: Most of the time the failure of Run.. tasks are caused by filesystem issues rather than configuration problems. Cylc should try running a certain task at least one more time before giving up. This is especially important for suites involving ensembles as even if 31 members succeed and 1 fails the workflow will stall.
[ ] Hold before certain tasks: @rtodling mentioned this one. The ability for a suite to "hold" before a certain task, say RunJediVarExecutable so that different configurations could be tested swiftly without the need to swell create a whole new suite. JCSDA's https://github.com/JCSDA-internal/skylab uses this feature in a different workflow engine called EWOK.
[ ] Restart a failed suite from a certain cycle & task: This one is tricky, I know it is possible with Cylc but not sure how would this play out with Swell. After a suite cycled certain amount of days, we would want to continue from a that point rather than restarting altogether.
Optional but useful ones:
[ ] Workflow sending email if suite stops/fails. I was able to do this with a sandbox Cylc setup.
Case in point for retrying, the task had SLURM issues (pink square) but now running after two failed attempts and with zero changes. I happened to be monitoring:
|
2025-04-01T04:55:06.417554
| 2017-05-01T19:41:31
|
225508284
|
{
"authors": [
"dalcinl",
"v-dobrev"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14006",
"repo": "GLVis/glvis",
"url": "https://github.com/GLVis/glvis/pull/21"
}
|
gharchive/pull-request
|
Use a larger socketserver backlog [server-backlog]
Increasing the backlog should help in servicing requests from clients
running in parallel with large number of processes. While the value of
128 is still relatively small (and not the definitive cure for parallel
clients flooding the listening socket), it is usually the default maximum
in Linux and Darwin kernels:
$ uname
Linux
$ sysctl net.core.somaxconn
net.core.somaxconn = 128
$ uname
Darwin
$ sysctl kern.ipc.somaxconn
kern.ipc.somaxconn: 128
@stefanozampini, GLVis does not have a config file for options. The backlog can be made a command line option, if you want. I'm fine either way.
|
2025-04-01T04:55:06.669335
| 2020-08-18T16:04:43
|
681159501
|
{
"authors": [
"pjsharpe07"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14007",
"repo": "GSA/sdg-indicators-usa",
"url": "https://github.com/GSA/sdg-indicators-usa/pull/1011"
}
|
gharchive/pull-request
|
Release 20200818
Part of release 20200820
This is part of the 20200820 release
Going to close and reopen.
Just going to close it all together
|
2025-04-01T04:55:06.692507
| 2021-01-15T20:09:17
|
787163609
|
{
"authors": [
"Shaun-Regenbaum"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14008",
"repo": "GT-Jewish-DH/daf-render-lib",
"url": "https://github.com/GT-Jewish-DH/daf-render-lib/issues/10"
}
|
gharchive/issue
|
Create Example for Calling from Sefaria API
Create an example where you call from Sefaria's API.
We need to find a better way to simplify the example, we don't want to confuse people with all the processing junk. For now, I am going to push this off as it is more a documentation problem than simply creating an example.
We need to find a better way to simplify the example, we don't want to confuse people with all the processing junk. For now, I am going to push this off as it is more a documentation problem than simply creating an example.
|
2025-04-01T04:55:06.700548
| 2024-03-20T11:44:10
|
2197329375
|
{
"authors": [
"kyoi-the-fox"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14009",
"repo": "GTNewHorizons/BloodArsenal",
"url": "https://github.com/GTNewHorizons/BloodArsenal/pull/24"
}
|
gharchive/pull-request
|
add null checks for all sendPlayerInformation cases,to avoid crashes …
…if ritual owner is not online and
Altar is missing
Pedestals are missing
Item to enchant is missing
Enchantment books are missing
Not enough LP in the network to enchant items
Enchantment type can not be applied on item
flew under the radar because they only happen under specific conditions and were not part of the issue because of it
that would prevent it from working in it's entirety if the player owning the activation crystal is not on the server,this would still allow it to be used by other team members,but prevents the crash
|
2025-04-01T04:55:06.737852
| 2024-03-22T13:43:29
|
2202528804
|
{
"authors": [
"Mi-Peng",
"mtfishman"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14010",
"repo": "GTorlai/PastaQ.jl",
"url": "https://github.com/GTorlai/PastaQ.jl/issues/307"
}
|
gharchive/issue
|
How to calculate the entropy/entanglement metrics efficiently
Thanks for this awesome work. I'm wondering if PastaQ provides API to calculate the entropy and other entanglement metrics efficiently, or they are calculated manually, e.g., using eigenvalues to calculate von Neumann entropy, which is considered time-consuming when q-bit is large. Thanks.
I would take a look at the ITensor discussion forum (https://itensor.discourse.group/) and ask a question there if your question isn't answered in one of the posts already.
I'm closing since I think this is more of a user support question, and that functionality is outside of the scope of this package (we may add more convenient functions for how to do that in ITensor, that functionality probably would not go here).
|
2025-04-01T04:55:06.763803
| 2015-08-17T13:19:17
|
101419215
|
{
"authors": [
"aaronlevin",
"ocharles"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14012",
"repo": "Gabriel439/post-rfc",
"url": "https://github.com/Gabriel439/post-rfc/issues/5"
}
|
gharchive/issue
|
ad - automatic differentiation
This has always stood out as an interesting and uniquely-haskell library. I'm not sure where it should go.
I'd be happy to submit a PR with a suggestion of a good category. Data science?
https://hackage.haskell.org/package/ad
ad is generally useful in the machine learning field (though I also used it recently to do some root-finding I needed for graphics animations. It probably belongs in either machine learning or just numerics.
|
2025-04-01T04:55:06.778114
| 2020-06-22T01:13:18
|
642680759
|
{
"authors": [
"CanRau",
"arp242",
"peixian"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14013",
"repo": "GaiAma/Coding4GaiAma",
"url": "https://github.com/GaiAma/Coding4GaiAma/pull/26"
}
|
gharchive/pull-request
|
Add support for self-hosted goatcounter
Adds a selfHostUrl attribute to the pluginOptions, where a user can specify a selfhosted goatcounter URL.
Sorry for the delay and thanks for the PR 🙏
Personally I think it would be best if the plugin would always accept the full URL, rather than just a "code". In the documentation for the integration there's an "integrations" section now with:
The reason I use a full URL is so that it'll work in any use case: goatcounter.com, custom domains, self-host.
I found this because another customer trying to use this plugin emailed me with issues and the URL they ended up using is:
https://https//xxx.goatcounter.com/count.goatcounter.com/
which is obvious not correct heh.
Assuming you agree this would be a good idea, I don't know what the best way forward is here with regards to compatibility, but perhaps renaming selfHostUrl to just url or endpoint and keeping the code as a "hidden" undocumented option might be a good idea? I assume that at this point only @peixian is using this, so it can be changed pretty safely at this point?
Hey @arp242 thanks for your tip & insides 🙏 sorry for the late reply I didn't receive a notification, probably because it's closed, maybe in such cases it's needed to directly @ mention the person? 🤔 not sure about that 🤷♂️
@peixian Sorry for the delay, again, I now uploaded 0.4.0 to NPM 🎉 including your PR and will soonish make the needed changes suggested by Martin and upload those as 0.5.0 I guess.
Yea endpoint sounds appropriate as it's matching GoatCounter's terminology 👌
|
2025-04-01T04:55:06.813143
| 2015-11-27T11:12:43
|
119178724
|
{
"authors": [
"Galexrt",
"deevant07"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14014",
"repo": "Galexrt/docker-zulip",
"url": "https://github.com/Galexrt/docker-zulip/issues/51"
}
|
gharchive/issue
|
Zulip exited with code 1(stuck while waiting for DB connections)
Hi, I am not sure if the problem is with config but i am newbie to zulip and docker.
The docker-compose.yml is configured as below.
database:
image: "galexrt/zulip-postgresql-tsearchextras:latest"
environment:
DB_NAME: zulip
DB_USER: zulip
DB_PASS: zulip
volumes:
- "/opt/docker/zulip/postgresql/data:/var/lib/postgresql:rw"
memcached:
image: "sameersbn/memcached:latest"
restart: always
rabbitmq:
image: "rabbitmq:3.5.5"
hostname: zulip-rabbit
restart: always
environment:
RABBITMQ_DEFAULT_USER: "zulip"
RABBITMQ_DEFAULT_PASS: "zulip"
redis:
image: "sameersbn/redis:latest"
volumes:
- "/opt/docker/zulip/redis:/var/lib/redis:rw"
zulip:
image: "galexrt/zulip:1.3.7-2"
ports:
- "8081:8080"
- "8443:443"
links:
- database
- memcached
- rabbitmq
- redis
environment:
ZULIP_SETTINGS_EXTERNAL_HOST: "domain.com"
ZULIP_SETTINGS_ZULIP_ADMINISTRATOR<EMAIL_ADDRESS>ZULIP_SETTINGS_ADMIN_DOMAIN: "domain.com"
ZULIP_SETTINGS_NOREPLY_EMAIL_ADDRESS<EMAIL_ADDRESS>ZULIP_SETTINGS_DEFAULT_FROM_EMAIL: "Zulip<EMAIL_ADDRESS>ZULIP_SETTINGS_EMAIL_HOST: "<IP_ADDRESS>"
ZULIP_SETTINGS_EMAIL_HOST_USER: "testuser"
ZULIP_SETTINGS_EMAIL_HOST_PASSWORD: "xxxx"
ZULIP_USER_EMAIL<EMAIL_ADDRESS>ZULIP_USER_PASS: "xxxx"
ZULIP_USER_DOMAIN: "example.com"
volumes:
- "/opt/docker/zulip/zulip:/data:rw"
The part of the log file which highlights the problem is as follows.
zulip_1 | Setting database configuration ...
zulip_1 | Setting key "DATABASES", type "array".
zulip_1 | Database configuration succeeded.
zulip_1 | Setting caches configuration ...
zulip_1 | Setting key "CACHES", type "array".
zulip_1 | Caches configuration succeeded.
zulip_1 | Activating authentication backends ...
zulip_1 | Setting key "AUTHENTICATION_BACKENDS", type "array".
zulip_1 | Adding authentication backend "EmailAuthBackend".
zulip_1 | Authentication backend activation succeeded.
zulip_1 | Setting redis configuration ...
zulip_1 | Setting key "RATE_LIMITING", type "bool".
zulip_1 | Setting key "REDIS_HOST", type "string".
zulip_1 | Setting key "REDIS_HOST_PORT", type "int".
zulip_1 | Redis configuration succeeded.
zulip_1 | Setting rabbitmq configuration ...
zulip_1 | Setting key "RABBITMQ_HOST", type "string".
zulip_1 | Setting key "RABBITMQ_USERNAME", type "string".
zulip_1 | Rabbitmq configuration succeeded.
zulip_1 | Executing Zulip configuration ...
zulip_1 | Setting key "EMAIL_HOST_USER", type "string".
zulip_1 | Setting key "NOREPLY_EMAIL_ADDRESS", type "string".
zulip_1 | Setting key "EMAIL_HOST", type "string".
zulip_1 | Setting key "EXTERNAL_HOST", type "string".
zulip_1 | Setting key "EMAIL_HOST_PASSWORD", type "string".
zulip_1 | Setting key "ADMIN_DOMAIN", type "string".
zulip_1 | Setting key "DEFAULT_FROM_EMAIL", type "string".
zulip_1 | Setting key "ZULIP_ADMINISTRATOR", type "string".
zulip_1 | Zulip configuration succeeded.
zulip_1 | Auto backup enabled.
zulip_1 | === End Initial Configuration Phase ===
zulip_1 | === Begin Bootstrap Phase ===
zulip_1 | Waiting for database server to allow connections ...
zulip_1 | ...........................................................openstack_zulip_1 exited with code 1
Detailed log file is attached.
Kindly help in resolving the problem.
I am using the command sudo docker-compose up
docker-zulip.txt
Try again with the latest version https://github.com/Galexrt/docker-zulip/releases/tag/1.3.9
Hi,
I am not able to pull the image 1.3.9 version from quay.io/galexrt/zulip:v1.3.9 with following error:
Error response from daemon: unable to ping registry endpoint https://quay.io/v0/
v2 ping attempt failed with error: Get https://quay.io/v2/: x509: certificate signed by unknown authority
v1 ping attempt failed with error: Get https://quay.io/v1/_ping: x509: certificate signed by unknown authority
Can the image be created on docker.io as well so that its easy to download and proceed?
https://hub.docker.com/r/galexrt/zulip/tags/
The above link is showing all tags except 1.3.9
Thanks,
Deevan
@deevant07 I created the tags on the docker hub now. Should be built in about 30 minutes.
Thank you @Galexrt
Hi,
I am still facing the issue with version 1.3.9 with below error:
zulip_1 | Zulip configuration succeeded.
zulip_1 | Auto backup enabled.
zulip_1 | === End Initial Configuration Phase ===
zulip_1 | === Begin Bootstrap Phase ===
zulip_1 | Waiting for database server to allow connections ...
rabbitmq_1 |
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:18 ===
rabbitmq_1 | Starting RabbitMQ 3.5.5 on Erlang 18.1
rabbitmq_1 | Copyright (C) 2007-2015 Pivotal Software, Inc.
rabbitmq_1 | Licensed under the MPL. See http://www.rabbitmq.com/
rabbitmq_1 |
rabbitmq_1 | RabbitMQ 3.5.5. Copyright (C) 2007-2015 Pivotal Software, Inc.
rabbitmq_1 | ## ## Licensed under the MPL. See http://www.rabbitmq.com/
rabbitmq_1 | ## ##
rabbitmq_1 | ########## Logs: tty
rabbitmq_1 | ###### ## tty
rabbitmq_1 | ##########
rabbitmq_1 | Starting broker...
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:18 ===
rabbitmq_1 | node : rabbit@zulip-rabbit
rabbitmq_1 | home dir : /var/lib/rabbitmq
rabbitmq_1 | config file(s) : /etc/rabbitmq/rabbitmq.config
rabbitmq_1 | cookie hash : qRoOkIXORPV3bIHgNaTeww==
rabbitmq_1 | log : tty
rabbitmq_1 | sasl log : tty
rabbitmq_1 | database dir : /var/lib/rabbitmq/mnesia/rabbit@zulip-rabbit
rabbitmq_1 |
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:18 ===
rabbitmq_1 | Memory limit set to 1581MB of 3953MB total.
rabbitmq_1 |
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:18 ===
rabbitmq_1 | Disk free limit set to 50MB
rabbitmq_1 |
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 ===
rabbitmq_1 | Limiting to approx 524188 file handles (471767 sockets)
rabbitmq_1 |
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 ===
rabbitmq_1 | FHC read buffering: ON
rabbitmq_1 | FHC write buffering: ON
rabbitmq_1 |
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 ===
rabbitmq_1 | Priority queues enabled, real BQ is rabbit_variable_queue
rabbitmq_1 |
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 ===
rabbitmq_1 | msg_store_transient: using rabbit_msg_store_ets_index to provide index
rabbitmq_1 |
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 ===
rabbitmq_1 | msg_store_persistent: using rabbit_msg_store_ets_index to provide index
rabbitmq_1 |
rabbitmq_1 | =WARNING REPORT==== 1-Dec-2015::11:59:19 ===
rabbitmq_1 | msg_store_persistent: rebuilding indices from scratch
rabbitmq_1 |
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 ===
rabbitmq_1 | started TCP Listener on [::]:5672
rabbitmq_1 | completed with 0 plugins.
rabbitmq_1 |
rabbitmq_1 | =INFO REPORT==== 1-Dec-2015::11:59:19 ===
rabbitmq_1 | Server startup complete; 0 plugins started.
zulip_1 | ...........................................................openstack_zulip_1 exited with code 1
@deevant07 Can you provide your current docker-compose.yml? Remove all security data from it (passwords, etc.)
Hi,
Please find below the docker-compose.yml.
database:
image: "galexrt/zulip-postgresql-tsearchextras:latest"
environment:
DB_NAME: zulip
DB_USER: zulip
DB_PASS: zulip
volumes:
- "/opt/docker/zulip/postgresql/data:/var/lib/postgresql:rw"
memcached:
image: "sameersbn/memcached:latest"
restart: always
rabbitmq:
image: "rabbitmq:3.5.5"
hostname: zulip-rabbit
restart: always
environment:
RABBITMQ_DEFAULT_USER: "zulip"
RABBITMQ_DEFAULT_PASS: "zulip"
redis:
image: "sameersbn/redis:latest"
volumes:
- "/opt/docker/zulip/redis:/var/lib/redis:rw"
zulip:
image: "galexrt/zulip:v1.3.9"
ports:
- "8081:8080"
- "8443:443"
links:
- database
- memcached
- rabbitmq
- redis
environment:
ZULIP_SETTINGS_EXTERNAL_HOST: "domain.com"
ZULIP_SETTINGS_ZULIP_ADMINISTRATOR<EMAIL_ADDRESS> ZULIP_SETTINGS_ADMIN_DOMAIN: "domain.com"
ZULIP_SETTINGS_NOREPLY_EMAIL_ADDRESS<EMAIL_ADDRESS> ZULIP_SETTINGS_DEFAULT_FROM_EMAIL: "Zulip "
ZULIP_SETTINGS_EMAIL_HOST: "xxxxxx"
ZULIP_SETTINGS_EMAIL_HOST_USER: "xxxxxx"
ZULIP_SETTINGS_EMAIL_HOST_PASSWORD: "xxxx"
ZULIP_USER_EMAIL: "xxxxxxx"
ZULIP_USER_PASS: "xxxxxx"
ZULIP_USER_DOMAIN: "example.com"
volumes:
- "/opt/docker/zulip/zulip:/data:rw"
Thanks,
Deevan
@deevant07 You are missing some environment variables for the zulip container.
On the first look, you are missing the DB_* env variables. Please add them or download the neweset version of the docker-compose.yml https://github.com/Galexrt/docker-zulip/blob/master/docker-compose.yml and try again please.
Thanks @Galexrt
Installation is done successfully but facing one issue on signup.
Your e-mail does not match any existing open organization. Use a different e-mail address, or contact<EMAIL_ADDRESS>with questions.
I have gone through some links with the solution to edit the manage.py file but as i am running from docker i am not able to find the file on host machine directory.
@deevant07 You need to configure the email settings in the docker-compose.yml.
See https://github.com/Galexrt/docker-zulip/wiki/Configuration and https://github.com/Galexrt/docker-zulip/wiki/Zulip-Email-SMTP-Configuration.
Closing as of the "bug" is fixed.
|
2025-04-01T04:55:07.503940
| 2021-07-30T12:23:15
|
956698188
|
{
"authors": [
"EmeraldBlock",
"Gaming32"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14015",
"repo": "Gaming32/ArrayV-v4.0",
"url": "https://github.com/Gaming32/ArrayV-v4.0/pull/70"
}
|
gharchive/pull-request
|
Advanced Shuffle Editor
This editor allows you to chain shuffles as they run. This PR still allows you to use the old system if you like.
So is it ready then @EmeraldBlock?
I haven't yet tested the new changes, but if you think it's all fine, yeah
|
2025-04-01T04:55:07.505485
| 2023-03-13T16:26:51
|
1621835825
|
{
"authors": [
"Gaming32",
"Madis0"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14016",
"repo": "Gaming32/world-host",
"url": "https://github.com/Gaming32/world-host/issues/4"
}
|
gharchive/issue
|
Do other parties need the mod?
Those that want to access the hosted server, that is. Please clarify that in the description of the mod.
Currently there is a proxy IP that exists, but no way to view it. I might add that later.
This functionality is available in World Host 0.2.2.
|
2025-04-01T04:55:07.516858
| 2015-10-09T18:07:50
|
110709913
|
{
"authors": [
"angieskazka"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14017",
"repo": "Gapminder/vizabi",
"url": "https://github.com/Gapminder/vizabi/issues/644"
}
|
gharchive/issue
|
"Show" control, similar to "Find"
Need the same thing to be able to show just 2 country mountains (edit show field of state)
done in e1cda784054775313fa1e9c27861f4d5ec910587
|
2025-04-01T04:55:07.518836
| 2018-10-29T18:02:58
|
375146244
|
{
"authors": [
"Areloch",
"marauder2k7",
"wcbx"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14018",
"repo": "GarageGames/Torque3D",
"url": "https://github.com/GarageGames/Torque3D/pull/2281"
}
|
gharchive/pull-request
|
OpenALEffects
Ground work, this commit enables reverb zones and opens the slots for
other effects to be added later. Tested using the full template. Build
your openal dll and your full.exe use one of the environment presets and
it will work
i should add this also gives the option for exponential falloff
Seems to all work, did note a typo with "PresetFroest" presumably that's supposed to be PresetForest?
Also, if you could correct the indentations on the SFXEnvironment definitions, and the typo, i think it's good to go :)
yeah it was supposed to be preset forest lol sorry rushing to test it out was a bit excited.
bump
|
2025-04-01T04:55:07.522764
| 2023-12-07T23:21:20
|
2031675187
|
{
"authors": [
"WillEngler"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:14019",
"repo": "Garden-AI/garden",
"url": "https://github.com/Garden-AI/garden/issues/362"
}
|
gharchive/issue
|
Display and "save" the base image a notebook uses in a notebook cell
We have a little first draft system for pairing a notebook with a base image that looks like this ...
User says garden-ai notebook start foo.ipynb --base-image=3.10-torch
We save a mapping from current/directory/foo.ipynb to 3.10-torch in the user's local data
Then the user can say garden-ai notebook start foo.ipynb and we open up it up again in the right image.
This is nice as a quick first pass but ofc it's pretty fragile. If the user mvs the notebook to a different directory they'll be prompted again to specify the base image.
I suggest that when making a new notebook from a template we create a first cell that presents (in a nice user-readable format) what base image this notebook uses. If the user decides to change it later, that is well and good! They can just edit the cell.
This has a few benefits
We don't need a local DB mapping anymore
The user can move the notebook around or even send it to another machine and the notebook keeps its own metadata. (Especially useful if we're debugging someone else's notebook)
Assumptions:
Acceptance Criteria
Given..., when..., then...
Now that we let users specify a --requirements file, that's something we can also think about in this ticket. The base image we provide and the requirements a user has specified feel pretty linked in the user's mind. So we'll want to think about those holistically.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.