id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1495399017 | Org naming change
remove old org names
replace with new org name
closing as name change is too difficult to mediate and restarting a new branch would be easier
| gharchive/pull-request | 2022-12-14T00:51:14 | 2025-04-01T04:32:55.357066 | {
"authors": [
"jonreding2010"
],
"repo": "OpenMAQS/openmaqs-dotnet",
"url": "https://github.com/OpenMAQS/openmaqs-dotnet/pull/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
591888313 | iOS won’t show notifications if the app is in foreground
iOS doesn't show notification for apps in foreground
@Shreyasldl, I've fixed this bug in my PR, will keep this open until merge.
| gharchive/issue | 2020-04-01T12:45:44 | 2025-04-01T04:32:55.418358 | {
"authors": [
"Shreyasldl",
"s0l0ist"
],
"repo": "OpenMined/covid-alert",
"url": "https://github.com/OpenMined/covid-alert/issues/25",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
343655683 | Travis Support for Unit Tests
In Issue #11 , we created a basic Unit Testing Suite... in this issue, we want to support automated unit testing using Travis (with integration to Github) so that all submitted pull requests get tested automatically.
This should also include a travis build badge on the Readme.
This has been done already.
| gharchive/issue | 2018-07-23T14:24:59 | 2025-04-01T04:32:55.419397 | {
"authors": [
"cereallarceny",
"iamtrask"
],
"repo": "OpenMined/syft.js",
"url": "https://github.com/OpenMined/syft.js/issues/13",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
55432327 | OpenModsLib Server Crash on Join
I have been updating my modpack to 1.7.10
After hours of adding/removing mods, i found that just OpenModsLib alone causes a crash.
It happens upon trying to join the world.
http://pastebin.com/jj1pHsVD
Probably version mismatch. But I'm guessing, since you gave absolutely no information about used versions.
Also, you are using Bukkit/Spigot. It's also worth mentioning in future reports. Actually, crash occurs between vanilla and bukkit patched code, so I suggest investigating this.
While we have very limited support for non-vanilla server, I already know it's start fine. So it's very probably problem is on your side.
i have the same issue and i can give you what it if you need me to
this is the server console if you would like me to send you a photo or copy of the crash report file i can
https://clients.mcprohosting.com/knowledgebase.php?action=displayarticle&id=182
02.04 13:35:25 [Multicraft] Server stopped 02.04 13:35:23 [Multicraft] Not restarting crashed server. 02.04 13:35:23 [Multicraft] Error writing to server: [Errno 32] Broken pipe
02.04 13:35:23 [Multicraft] Error writing to server: [Errno 32] Broken pipe
02.04 13:35:23 [Multicraft] Stopping server!
02.04 13:35:23 [Multicraft] Server shut down
02.04 13:35:22 [Server] INFO The state engine was in incorrect state SERVER_STOPPING and forced into state SERVER_STOPPED. Errors may have been discarded.
02.04 13:35:22 [Server] INFO Holder lookups applied
02.04 13:35:22 [Server] INFO Applying holder lookups
02.04 13:35:22 [Server] INFO Unloading dimension 1 02.04
13:35:22 [Server] INFO Unloading dimension -1 02.04
13:35:22 [Server] INFO Unloading dimension 0 02.04
13:35:22 [Server] INFO Saving chunks for level 'DIM-1'/Nether 02.04
13:35:22 [Server] INFO Saving chunks for level 'DIM1'/The End 02.04
13:35:22 [Server] INFO Saving chunks for level 'Toast'/Overworld 02.04
13:35:22 [Server] INFO Saving worlds 02.04
13:35:22 [Server] INFO Saving players 02.04
13:35:22 [Server] INFO Disabling WorldEdit v6.0;3342-78f975b9 02.04
13:35:22 [Server] INFO Disabling Essentials v2.13.1 02.04
13:35:22 [Server] INFO Disabling EssentialsChat v2.13.1 02.04
13:35:22 [Server] INFO Disabling EssentialsProtect v2.13.1 02.04
13:35:22 [Server] INFO Disabling EssentialsAntiBuild v2.13.1 02.04
13:35:22 [Server] INFO Disabling EssentialsSpawn v2.13.1 02.04
13:35:22 [Server] INFO Stopping server 02.04
13:35:22 [Server] ERROR This crash report has been saved to: /./crash-reports/crash-2015-04-02_13.35.22-server.txt 02.04
13:35:22 [Server] INFO ... 12 more 02.04
13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:174) ~[launchwrapper-1.9.jar:?] 02.04
13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.runTransformers(LaunchClassLoader.java:276) ~[launchwrapper-1.9.jar:?] 02.04
13:35:22 [Server] INFO at openmods.OpenModsClassTransformer.transform(OpenModsClassTransformer.java:230) ~[OpenModsClassTransformer.class:?] 02.04
13:35:22 [Server] INFO at openmods.OpenModsClassTransformer.tryApplyContextWrapper(OpenModsClassTransformer.java:247) ~[OpenModsClassTransformer.class:?] 02.04
13:35:22 [Server] INFO at openmods.asm.VisitorHelper.apply(VisitorHelper.java:29) ~[OpenModsLib-1.7.10-0.6.jar:?] 02.04 13:35:22 [Server] INFO at openmods.OpenModsClassTransformer$8.createVisitor(OpenModsClassTransformer.java:250) ~[OpenModsClassTransformer$8.class:?] 02.04
13:35:22 [Server] INFO at openmods.context.ContextClassTransformer.(ContextClassTransformer.java:65) ~[ContextClassTransformer.class:?] 02.04
13:35:22 [Server] INFO at org.objectweb.asm.ClassVisitor.(Unknown Source) ~[asm-all-4.1.jar:4.1] 02.04 13:35:22 [Server] INFO Caused by: java.lang.IllegalArgumentException 02.04 13:35:22 [Server] INFO at java.lang.Thread.run(Unknown Source) [?:1.7.0_72] 02.04
13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:658) [MinecraftServer.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:798) ~[MinecraftServer.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:423) ~[lt.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:980) ~[MinecraftServer.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.network.NetworkSystem.func_151269_c(NetworkSystem.java:173) ~[nc.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.network.NetworkManager.func_74428_b(NetworkManager.java:247) ~[ej.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.network.NetHandlerLoginServer.func_147233_a(NetHandlerLoginServer.java:70) ~[nn.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.server.network.NetHandlerLoginServer.func_147326_c(NetHandlerLoginServer.java:130) ~[nn.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.server.management.ServerConfigurationManager.attemptLogin(ServerConfigurationManager.java:521) ~[oi.class:?] 02.04
13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04
13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:188) ~[launchwrapper-1.9.jar:?] 02.04
13:35:22 [Server] INFO java.lang.ClassNotFoundException: net.minecraft.server.management.ItemInWorldManager 02.04
13:35:22 [Server] ERROR Cause of unexpected exception was 02.04
13:35:22 [Server] INFO ... 10 more 02.04
13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:174) ~[launchwrapper-1.9.jar:?] 02.04
13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.runTransformers(LaunchClassLoader.java:276) ~[launchwrapper-1.9.jar:?] 02.04
13:35:22 [Server] INFO at openmods.OpenModsClassTransformer.transform(OpenModsClassTransformer.java:230) ~[OpenModsClassTransformer.class:?] 02.04
13:35:22 [Server] INFO at openmods.OpenModsClassTransformer.tryApplyContextWrapper(OpenModsClassTransformer.java:247) ~[OpenModsClassTransformer.class:?] 02.04
13:35:22 [Server] INFO at openmods.asm.VisitorHelper.apply(VisitorHelper.java:29) ~[OpenModsLib-1.7.10-0.6.jar:?] 02.04
13:35:22 [Server] INFO at openmods.OpenModsClassTransformer$8.createVisitor(OpenModsClassTransformer.java:250) ~[OpenModsClassTransformer$8.class:?] 02.04
13:35:22 [Server] INFO at openmods.context.ContextClassTransformer.(ContextClassTransformer.java:65) ~[ContextClassTransformer.class:?] 02.04
13:35:22 [Server] INFO at org.objectweb.asm.ClassVisitor.(Unknown Source) ~[asm-all-4.1.jar:4.1] 02.04
13:35:22 [Server] INFO Caused by: java.lang.IllegalArgumentException 02.04 13:35:22 [Server] INFO ... 10 more 02.04 13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:188) ~[launchwrapper-1.9.jar:?] 02.04 13:35:22 [Server] INFO Caused by: java.lang.ClassNotFoundException: net.minecraft.server.management.ItemInWorldManager 02.04 13:35:22 [Server] INFO at java.lang.Thread.run(Unknown Source) [?:1.7.0_72] 02.04
13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:658) [MinecraftServer.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:798) ~[MinecraftServer.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:423) ~[lt.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:980) ~[MinecraftServer.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.network.NetworkSystem.func_151269_c(NetworkSystem.java:173) ~[nc.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.network.NetworkManager.func_74428_b(NetworkManager.java:247) ~[ej.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.network.NetHandlerLoginServer.func_147233_a(NetHandlerLoginServer.java:70) ~[nn.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.server.network.NetHandlerLoginServer.func_147326_c(NetHandlerLoginServer.java:130) ~[nn.class:?] 02.04
13:35:22 [Server] INFO at net.minecraft.server.management.ServerConfigurationManager.attemptLogin(ServerConfigurationManager.java:521) ~[oi.class:?] 02.04
13:35:22 [Server] INFO java.lang.NoClassDefFoundError: net/minecraft/server/management/ItemInWorldManager 02.04
13:35:22 [Server] ERROR Encountered an unexpected exception 02.04
13:35:22 [Server] INFO openmods.OpenModsClassTransformer.tryApplyContextWrapper(OpenModsClassTransformer.java:246): Adding context wrappers to mx(net.minecraft.server.management.ItemInWorldManager) 02.04 13:35:22 [Server] INFO UUID of player Synthos_Colton is 7c2bf2e6-aecf-44d9-886e-072614862ab9
Oh, now I see what happening (but I had to spend some time to clean those scrambled logs). It's caused by ancient libraries in Cauldron. I guess I can work around that.
| gharchive/issue | 2015-01-25T22:04:26 | 2025-04-01T04:32:55.541061 | {
"authors": [
"boq",
"cjcoley2",
"enterprise12nx"
],
"repo": "OpenMods/OpenModsLib",
"url": "https://github.com/OpenMods/OpenModsLib/issues/32",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
863267303 | Investigate ways to improve the compilation environment on macOS
Currently, AbstractEquinoxTask copies tools.jar from the current JVM into the Notes bundle in order to have a working compilation environment. This has never been great, but it's particularly problematic when compiling with Java versions >8, which don't have a tools.jar as such.
It may be better to do similar to what the Open Liberty Runtime does and download an AdoptOpenJDK build of Java 8 as needed and use that, copying in the Notes.jar and websvc.jar from Notes.app or referencing them in the launch classpath. That'd be cumbersome, but it should be reliable at least until the AdoptOpenJDK URLs change (which they may with the move to Adoptium).
I investigated using the JDT by way of EclipseCompiler, but it looks like it is limited to working with filesystem-bound dependencies. There was a bug from way back about similar things, but that may have just been about compiling in-memory files (which works) and not about dependencies.
| gharchive/issue | 2021-04-20T21:42:09 | 2025-04-01T04:32:55.554021 | {
"authors": [
"jesse-gallagher"
],
"repo": "OpenNTF/org.openntf.nsfodp",
"url": "https://github.com/OpenNTF/org.openntf.nsfodp/issues/266",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
627664528 | ONE package definitions
Hi!
I couldn't find where the RPM/DEB package definitions (e.g. specfile for RPM) are maintained. Either I missed them or they are not tracked here: in the later case, would it be possible to publish them?
I'm currently playing with alpine-based ONE hosts and might end up doing some packaging work which could be upstreamed here.
Thanks!
Wow, it seems that repository was deleted (or moved to private repository).
Previously it was located on https://github.com/opennebula/packages/, now you can still find some forks on github, eg. my one with the latest changes which I managed to find: https://github.com/kvaps/packages
How come? Using the package feature was very useful for easy testing / deployment.
We are also baffled as to the disappearance of everything under https://github.com/OpenNebula/packages/releases and https://github.com/OpenNebula/packages without clear substitute anywhere else. Is anyone able to comment on why part of the project is now hidden?
Perhaps one of the owners / top contributers can chime in: @tinova @vholer @rsmontero @xorel @christian7007 @al3xhh
Hi guys! We are doing some changes in the structure of the repos, etc. in preparation for the release of OpenNebula 5.12 "Firecracker". We'll be publishing a proper announcement and a summary of all these changes in a couple of days max. Thanks for your patience!
@albert0marti excited to hear about the upcoming 5.12 release and congratulations to the team. However please explain the situation to people in the community who are working to test and setup OpenNebula 5.10.x and 5.11.x and need the packaging resources.
Right now it looks like resources have been hidden from the community or deleted by accident. Understand you may want to lock a branch for new releases but keeping existing code in other branches available would be really helpful.
If there has been a change of policy and you are withdrawing the packaging code then we would like some warning to prepare for that before the code is removed instead of afterwards. We were preparing for an upgrade when this happened. Thanks for your help.
Hi there!
Thank you all for your patience! Here you have the details about this decision: https://forum.opennebula.io/t/towards-a-stronger-opennebula-community/8506
Hope this helps to clarify the new scenario.
Just to make it clear: access to the packages repo is NOT one of the benefits that comes with the OpenNebula Subscription (https://opennebula.io/enterprise/). Although the repo is now private again, and only accessible in principle to the OpenNebula Team, active members of the community can request access.
We are new users of Opennebula and unfortunately this aggressive open-core model will make us reconsider our choice. Sorry to hear that this was the path you chose.
Hi @howels,
The OpenNebula Community Edition is a full-featured version of OpenNebula, and OpenNebula CE and EE are synchronized at the end of each maintenance cycle, so we are not speaking here about an open core model. Please refer to our Release Policy for more details.
The source code (https://github.com/OpenNebula/one/) and the binary and source packages that we produce for the Community Edition (http://downloads.opennebula.io/repo/) are publicly available, given that the CE is under the Apache License 2.0, but if you—for whatever reason—need access to our internal repo for package definitions, just send me a PM.
| gharchive/issue | 2020-05-30T06:32:19 | 2025-04-01T04:32:55.564055 | {
"authors": [
"Fnux",
"albert0marti",
"howels",
"kvaps",
"sambhavsaggi"
],
"repo": "OpenNebula/one",
"url": "https://github.com/OpenNebula/one/issues/4844",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1896672122 | Error: Missing required parameter: auth
Followed your running the project locally step by step. I am stuck on the below error
#200 already went through this. Double checked my github client id and secret key and they are fine
error Error: Missing required parameter: auth
[0] at new Replicate (/home/gaurav/openpipe/node_modules/.pnpm/replicate@0.12.3/node_modules/replicate/index.js:37:13)
[0] at eval (webpack-internal:///(api)/./src/modelProviders/replicate-llama2/getCompletion.ts:13:19)
[0] at process.processTicksAndRejections (node:internal/process/task_queues:95:5) {
[0] digest: undefined
[0] }
[0] - warn ../node_modules/.pnpm/@babel+standalone@7.22.9/node_modules/@babel/standalone/babel.js
[0] Critical dependency: the request of a dependency is an expression
[0]
Login seems to be successfull
What am I missing? because of the above auth error I am not able to use the app (seed and project name edit fails)
Update on this: It seems that I was missing Replicate Token API and after that it ran successfully
Yeah @DVGY07, if you could submit a PR to make the replicate token optional that would be awesome.
Sweet, it’s merged. Thanks for the PR!
| gharchive/issue | 2023-09-14T14:15:01 | 2025-04-01T04:32:55.575892 | {
"authors": [
"DVGY07",
"arcticfly"
],
"repo": "OpenPipe/OpenPipe",
"url": "https://github.com/OpenPipe/OpenPipe/issues/230",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
330549740 | Mass edit fix for blanks, numbers and booleans
This fix is for #1631 and #180
I'm going to make a cleaner PR - so closing this one
| gharchive/pull-request | 2018-06-08T07:31:15 | 2025-04-01T04:32:55.673617 | {
"authors": [
"ostephens"
],
"repo": "OpenRefine/OpenRefine",
"url": "https://github.com/OpenRefine/OpenRefine/pull/1641",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
745863793 | Temporary workaround set-env deprecation
This is an attempt to temporarily re-enable the set-env command, to make sure our CI stays green, before we figure out how to use the new system.
For #3341
Merging to master since GitHub workflows only run there (per our configuration).
| gharchive/pull-request | 2020-11-18T17:29:43 | 2025-04-01T04:32:55.674902 | {
"authors": [
"wetneb"
],
"repo": "OpenRefine/OpenRefine",
"url": "https://github.com/OpenRefine/OpenRefine/pull/3342",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1001978186 | Build for Java 11 and 16 as well
Attempt to check compatibility with newer versions of Java.
Of course we have a really tiny test suite on this project, so there are still tons of incompatibilities that can still exist.
Given that no one else is active here I will just merge this, release it and see if it magically makes OpenRefine work with Java 16.
I'll test the snapshot or beta this morning. Just woke up.
@thadguidry there is not much to test unfortunately, because updating OpenRefine to this new version of butterfly (1.2.0) does not solve the problem. So it could well be independent.
@wetneb OK understood
| gharchive/pull-request | 2021-09-21T07:17:52 | 2025-04-01T04:32:55.677273 | {
"authors": [
"thadguidry",
"wetneb"
],
"repo": "OpenRefine/simile-butterfly",
"url": "https://github.com/OpenRefine/simile-butterfly/pull/32",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
882079698 | Documentation for plugins when using docker compose (specifically: saml)
Is your feature request related to a problem? Please describe.
I have set up a working installation of OpenSlides which I would like to connect to our SAML2 SSO. However I have many questions regarding the configuration, especially how to handle Plugins. Do I have to run setup.sh every time I change settings? Do I have to edit any settings.py? If so, it seems I have to do a git stash, git pull, git stash pop before running the update.sh?
I enabled the SAML Plugin inside the docker/.env file. But I don't find any useful information where to set it up. The settings.pytells me that the .json is created on startup, but I cannot find it anywhere.
Describe the solution you'd like
I short and simple howto and where to look in order to setup the saml plugin with docker compose. Maybe a short overview and step by step list needed to update the installation.
Describe alternatives you've considered
Not using saml. Which would be pretty painful since I need to setup a lot of user manually in that case.
Additional context
The initial setup is very well described and documented, maybe an example nginx.conf reverse proxy config could be added:
location / {
proxy_buffering off;
proxy_pass http://localhost:8000;
}
location /ws {
proxy_pass http://localhost:8000/ws;
proxy_http_version 1.1;
proxy_set_header Upgrade $http_upgrade;
proxy_set_header Connection "Upgrade";
}
}
proxy_buffering off; seems to be very important! (Or maybe only for my case since I use 2 proxies :D )
Hi,
I have set up the files, the json has the following content:
{
"strict": true,
"debug": true,
"sp": {
"entityId": "<slides-url>",
"assertionConsumerService": {
"url": "<slides-url>/apps/saml/?acs",
"binding": "urn:oasis:names:tc:SAML:2.0:bindings:HTTP-POST"
},
"singleLogoutService": {
"url": "<slides-url>/apps/saml/?sls",
"binding": "urn:oasis:names:tc:SAML:2.0:bindings:HTTP-Redirect"
},
"NameIDFormat": "urn:oasis:names:tc:SAML:1.1:nameid-format:unspecified",
"x509cert": "",
"privateKey": ""
},
"idp": {
"entityId": "<simplesamlphp-url>",
"singleSignOnService": {
"url": "<simplesamlphp-url>/saml2/idp/SSOService.php",
"binding": "urn:oasis:names:tc:SAML:2.0:bindings:HTTP-Redirect"
},
"singleLogoutService": {
"url": "<simplesamlphp-url>/saml2/idp/SingleLogoutService.php",
"binding": "urn:oasis:names:tc:SAML:2.0:bindings:HTTP-Redirect"
},
"x509cert": "-----BEGIN CERTIFICATE-----<certificate>-----END CERTIFICATE-----"
},
"security": {
"nameIdEncrypted": false,
"authnRequestsSigned": true,
"logoutRequestSigned": false,
"logoutResponseSigned": false,
"signMetadata": true,
"wantMessagesSigned": false,
"wantAssertionsSigned": false,
"wantNameId" : true,
"wantNameIdEncrypted": false,
"wantAssertionsEncrypted": false,
"signatureAlgorithm": "http://www.w3.org/2000/09/xmldsig#rsa-sha1",
"digestAlgorithm": "http://www.w3.org/2000/09/xmldsig#sha1"
},
"contactPerson": {
"technical": {
"givenName": "Die Humanisten IT Team",
"emailAddress": "<removed>"
},
"support": {
"givenName": "Die Humanisten Support Team",
"emailAddress": "<removed>"
}
},
"organization": {
"en-US": {
"name": "OpenSlides",
"displayname": "OpenSlides",
"url": "http://openslides.org"
}
},
"generalSettings": {
"loginButtonText": "Login via Humanisten SSO",
"changePasswordUrl": "<removed>"
},
"attributeMapping": {
"UserID": ["username", true],
"FirstName": ["first_name", false],
"LastName": ["last_name", false]
},
"requestSettings": {},
"groups": {}
}
with manual setting the simplesamlphp configuration (is there a way to get the metadata xml for openslides?):
$metadata['<slides-url>'] = [
'entityid' => '<slides-url>',
'contacts' => [],
'metadata-set' => 'saml20-sp-remote',
'AssertionConsumerService' => [
[
'Binding' => 'urn:oasis:names:tc:SAML:2.0:bindings:HTTP-POST',
'Location' => '<slides-url>/apps/saml/?acs',
'index' => 1,
'isDefault' => true,
],
],
'SingleLogoutService' => [
[
'Binding' => 'urn:oasis:names:tc:SAML:2.0:bindings:HTTP-Redirect',
'Location' => '<slides-url>/apps/saml/?sls'
],
],
'NameIDFormat' => 'urn:oasis:names:tc:SAML:1.1:nameid-format:emailAddress',
'keys' => [
[
'encryption' => true,
'signing' => true,
'type' => 'X509Certificate',
'X509Certificate' =>
'<certificate>',
],
],
];
I do not get any SSO login button. Do I have to rebuild with make / build.sh when changing those settings?
You can get the metadata with <your-domain>/apps/saml/metadata/.
If you also have the sp.crt and sp.key, you have to regenerate your docker-compose.yml with the m4 command from the readme. Then, just a restart (docker-compose down and docker-compose up) should be sufficient. But a ./build.sh server is not harmful.
Ahh, I found the problem I had from the beginning. I used TRUE instead of True. Had a look into the .m4 file and noticed the problem. Now the saml settings appear in the .yml configuration!
Just one more issue: due to the proxy, the saml plugins refuses to accept the response:
The response was received at http://slides.diehumanisten.de:8000/apps/saml/ instead of https://slides.diehumanisten.de/apps/saml/?acs
Is there a way to set an allowlist? Or is the /apps/saml/?acs --> /apps/saml/ the Problem?
First, do you have https in "url": "<slides-url>/apps/saml/?acs" (assertionConsumerService) set? The missing ?acs is strange. Are you sure, that your identity provider redirecs correctly? IMO it is not the proxy, since in our instances with the same proxy it redirects correctly.
First, do you have https in "url": "<slides-url>/apps/saml/?acs" (assertionConsumerService) set? The missing ?acs is strange.
Yes, everything is https. And all the redirects lead to https inside my browser. That works.
The redirect internally is also pure https. I also changed all my proxies to use https (snakeoil, proxy_ignore_ssl internally). Additionally my hosts are set in a way that the domain is never "lost" during any hop. The only https --> http redirect is going on inside the docker guest.
Are you sure, that your identity provider redirecs correctly? IMO it is not the proxy, since in our instances with the same proxy it redirects correctly.
Actually, I am not. We are using simplesamlphp v1.19.0 with a bunch of other SPs connected without any issue. However none of them have a ? inside assertionConsumerService or any other config.
I did try again. The ?acs is not missing in the URL of my browser. It is definitely there.
I did try setting the assertionConsumerService to https://slides.diehumanisten.de/apps/saml/?acs=acs, because maybe the simplesamlphp did try to "fix" empty variables. The result is the following:
The response was received at http://slides.diehumanisten.de:8000/apps/saml/ instead of https://slides.diehumanisten.de/apps/saml/?acs=acs
(On another note, something breaks the caddyfile every time I start and stop the service, I have to run build.sh after each restart or it will complain about run: adapting config using caddyfile: /etc/caddy/Caddyfile:20 - Error during parsing: Unexpected EOF)
@FinnStutzenstein do you have any idea whats happening here? The ?acs is not the Problem. As you can see in the previous post the second displayed URL in the error message successfully changes. So I guess the http/https is the only issue. Since the only remaining redirection happens inside the container, the problem has to be inside the container.
No, sadly I do not have a clue. The caddy issue is fixed with #6083
Hello,
sorry for warming up this old thread, but we need openslides more than ever.
@FinnStutzenstein Could it be that I am chasing the wrong Error message. There is nothing indicating that the message
server_1 | The response was received at http://slides.diehumanisten.de:11080:8000/apps/saml/ instead of https://slides.diehumanisten.de/apps/saml/?acs
is indeed an error. I re-made out setup and found another Error:
server_1 | [2021-11-28 16:45:13 +0100] [377] [ERROR] django.request Internal Server Error: /apps/saml/
The full log:
server_1 | [2021-11-28 16:45:06 +0100] [377] [INFO] openslides.utils.timing [ywVU] request: 1: 0.01347 2: 0.00000 sum: 0.01347
pgbouncer_1 | 2021-11-28 15:45:07.003 7 LOG C-0x55b192781ad0: openslides/openslides@172.24.0.9:40302 login attempt: db=openslides user=openslides tls=no
server_1 | [2021-11-28 16:45:07 +0100] [377] [INFO] openslides.utils.timing [ywVU] request: 1: 0.00113 2: 0.00000 sum: 0.00113
server_1 | [2021-11-28 16:45:07 +0100] [383] [INFO] openslides.utils.timing [ywVU] request: 1: 0.17043 2: 0.00001 sum: 0.17044
server_1 | [2021-11-28 16:45:09 +0100] [381] [INFO] openslides.utils.timing [ywVU] request: 1: 0.01461 2: 0.00002 sum: 0.01463
client_1 | 172.24.0.11 - - [28/Nov/2021:15:45:10 +0000] "GET /ngsw-worker.js HTTP/1.1" 200 41983 "-" "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:94.0) Gecko/20100101 Firefox/94.0"
server_1 | The response was received at http://slides.diehumanisten.de:11080:8000/apps/saml/ instead of https://slides.diehumanisten.de/apps/saml/?acs
server_1 | [2021-11-28 16:45:13 +0100] [377] [INFO] openslides.utils.timing [ywVU] request: 1: 0.05524 2: 0.00000 sum: 0.05524
server_1 | [2021-11-28 16:45:13 +0100] [377] [ERROR] django.request Internal Server Error: /apps/saml/
client_1 | 172.24.0.11 - - [28/Nov/2021:15:45:14 +0000] "GET /ngsw-worker.js HTTP/1.1" 304 0 "-" "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:94.0) Gecko/20100101 Firefox/94.0"
pgnode1_1 | 2021-11-28 15:45:15,796 INFO reaped unknown pid 351
pgnode1_1 | 2021-11-28 15:45:15,796 INFO reaped unknown pid 351
client_1 | 172.24.0.11 - - [28/Nov/2021:15:45:18 +0000] "GET /ngsw.json?ngsw-cache-bust=0.9268512327499515 HTTP/1.1" 200 7119 "https://slides.diehumanisten.de/ngsw-worker.js" "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:94.0) Gecko/20100101 Firefox/94.0"
@jsangmeister @tsiegleauq Ping
I currently have the same issue ...
Update: solved it by letting the server run on port 443:
server:
<< : *default-osserver
depends_on:
- server-setup
environment:
<< : *default-osserver-env
secrets:
- django
- saml_cert
- saml_key
- saml_config
command:
- "gunicorn"
- "-w"
- "8"
- "--preload"
- "-t"
- "240"
- "-b"
- "0.0.0.0:443"
- "openslides.wsgi"
| gharchive/issue | 2021-05-09T09:44:10 | 2025-04-01T04:32:56.173491 | {
"authors": [
"ClundXIII",
"FinnStutzenstein",
"dorianim"
],
"repo": "OpenSlides/OpenSlides",
"url": "https://github.com/OpenSlides/OpenSlides/issues/6057",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
199209964 | Add new personal settings MOTIONS_ALLOW_AMENDMENTS_OF_AMENDMENTS.
Set settings value to 0 to disable new amendment button in motion detail
template. Default: 1.
Added changelog entry. Rebased.
Is it possible to use a boolean value here instead of 0 and 1?
No, it tested it. A python boolean "True/False" in settings.py is invalid (unknown value) in JavaScript (which requires "true/false"). So, 0 and 1 is just the easiest way.
Ah I see.
Please test a fix in line 187 of core/views.py: The variable client_settings contains a dict but we should transform it to JSON (i. e. a string) and render this string into the template. Then True should be exposed to true.
we should transform it to JSON (i. e. a string) and render this string into the template
How can I do this? Can you give me an example that I can try it?
Head
import json
Line 187
""".format(realm=realm, angular_modules=angular_modules, settings=json.dumps(client_settings), js_files=js_files) +
@normanjaeckel I added a new commit. Please review.
| gharchive/pull-request | 2017-01-06T14:44:30 | 2025-04-01T04:32:56.178346 | {
"authors": [
"emanuelschuetze",
"normanjaeckel"
],
"repo": "OpenSlides/OpenSlides",
"url": "https://github.com/OpenSlides/OpenSlides/pull/2821",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1124066498 | Fixed proto file.
@peb-adr Please check this again. The syntax of the current main branch proto file is definitly wrong. There is a > missing. So I tried the following fix.
See #157
@normanjaeckel
| gharchive/pull-request | 2022-02-04T10:53:33 | 2025-04-01T04:32:56.179892 | {
"authors": [
"normanjaeckel",
"peb-adr"
],
"repo": "OpenSlides/openslides-manage-service",
"url": "https://github.com/OpenSlides/openslides-manage-service/pull/158",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2622385706 | Not running on Arch Linux
There are multiple reports about the OpenTaiko Hub not running on Arch Linux systems, there might be additional configuration to do on the tauri configuration files in order to resolve this issue
Same issue here.
% ./OpenTaiko.Hub_0.1.4_amd64.AppImage
Could not create default EGL display: EGL_BAD_PARAMETER. Aborting...
^C
I'm having the same issue here, looking around it seems to be related to libwebkit2gtk
| gharchive/issue | 2024-10-29T21:42:29 | 2025-04-01T04:32:56.215804 | {
"authors": [
"0auBSQ",
"exentio",
"zDEFz"
],
"repo": "OpenTaiko/OpenTaiko-Hub",
"url": "https://github.com/OpenTaiko/OpenTaiko-Hub/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
436903502 | Rename comment visibility options, create a new ones, and update help text
-=Renames=-
Private => Applicant
Reviewers and Staff => Reviewers
Internal => Team
Public => All
-=To create=-
Applicant and Reviewers
Private
-=Help text changes=-
Applicant: Visible to applicant and team (default)
Reviewers: Visible to reviewers and team.
Applicants and Reviewers: Visible to applicants (including partners), reviewers, and team.
Team: Visible only to team.
Private: Visible only to author.
All: Visible to any user who has access to the submission. If the submission is public, that means he comment is publicly visible.
Relates to #1087
@danblah See how it looks in the screenshot below. Are all these options really needed? I'm afraid the users will be a bit overwhelmed by all the options. All users will not see all the options but staff will.
My suggestion is that we at least skip the new "Private" option that makes the comment visible only to the user. This demands a rewrite of the comment access function that now is based on groups.
If someone want to have private notes, why should they add it do the system?
All good to remove the private option.
On the interface, what if have it be a pop up box from a button titled "Visibility". On click, you have checkboxes for the following:
Applicant (default)
Team (only visible to users in the role Staff)
Partners
Reviewers
The help text below updates as options are selected to say:
Visible to: OTF team, Applicants, Partners, Reviewers (etc).
Then there's a save/submit or whatever button and it goes back to the comment edit screen.
@danblah So then we are down do:
Rename Private => Applicant
Rename Internal => Team
Rename Reviewers and Staff => Reviewers
Replace Public => Partners
Four options is what we have today so then the UI can stay the same.
Those are the right options and I like that they related to roles.
I am proposing a different interface though I don't know if its possible. Right now its a radio button with only one option that requires a lot of help text because I don't think its intuitive. I'm proposing that you have a set of checkboxes for the commenterd to select any of the above that apply to them, clearly indicating that those are who can see the comment. I think that makes would allow us to dismiss the help text all together. Is what I'm saying make sense here?
The new interface with checkboxes is not implemented but the names are updated and a partner option is added.
Will add a new issue for the improved interface.
Related: https://github.com/OpenTechFund/opentech.fund/issues/1445
blah asked for small word change from Applicant to Applicant(s) on July 11.
All features are visible and functional during test. Thanks!
| gharchive/issue | 2019-04-24T20:56:58 | 2025-04-01T04:32:56.225878 | {
"authors": [
"danblah",
"fourthletter",
"frjo"
],
"repo": "OpenTechFund/opentech.fund",
"url": "https://github.com/OpenTechFund/opentech.fund/issues/1174",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
163456480 | 20160701 111800 serial ports list
This branch does the following:
Removes auto-connect feature, and instead simply lists all available serial ports, and reports the list back to the frontend
Receives a serial port name from the front end, and attempts a connection to the port
Replaces all logging to reference a Logger object
Logging files are capped, and backup log files are auto-generated
The calibration values (created by the Pipette module) are now correctly being saved with the 'frozen' path, so they save to the correct spot when run as an executable
Looks good to me...
| gharchive/pull-request | 2016-07-01T19:05:08 | 2025-04-01T04:32:56.295339 | {
"authors": [
"SimplyAhmazing",
"andySigler"
],
"repo": "OpenTrons/otone_backend",
"url": "https://github.com/OpenTrons/otone_backend/pull/6",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2391181958 | Add description to data_source_connector.go file
Compared output of data source "cloudconnexa_connector" and Swagger - noticed that data source was missing "description" field
Should there be a test that would've caught this?
Theoretically yes) I'm not yet very familiar with testing, how it is done etc. So ... :wink:
| gharchive/pull-request | 2024-07-04T16:24:49 | 2025-04-01T04:32:56.298192 | {
"authors": [
"patoarvizu",
"sahaqaa"
],
"repo": "OpenVPN/terraform-provider-cloudconnexa",
"url": "https://github.com/OpenVPN/terraform-provider-cloudconnexa/pull/14",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
423747761 | New unit test organization and build strategy
We have outgrown our current unit test organization and build strategy. We currently have eleven tests that get compiled into eleven separate test executables. Ten of these are for testing the epanet2 library and one is for testing the epanet-output library.
The build is taking a long time because the Boost unit_test header file is very large and contains a lot of templates. We were compiling it 11 times. I have made some progress reorganizing things.
I propose the following. We have one test module for the Toolkit. Within the test module we will have separate test suites for each of the functional categories found in the Toolkit API. The test suites will be in separate files and will be compiled together into a single test executable at build time. We will link with the boost testing library to eliminate some of the time spent compiling the test headers.
These changes should help make things more manageable from both an organizational and build perspective.
I have three questions:
Will we be able to run the individual test files in stand-alone fashion on our local machine, suitably modified to run without Boost (as I have currently been doing), to allow us to debug CI failures when we submit PRs?
Has the successful build status of the dev branch been restored or will we have to wait until the new unit test strategy is implemented before we can submit any new PRs?
Are there alternatives to using the Boost unit test library that may be worth exploring or are we too far along in the game for that?
Good idea. If the combined test fails, will we know which part of it failed?
@LRossman
You are welcome to use your ingenuity to figure that out.
This is something I hope to check in later today or tomorrow.
I suppose it's never to late, provided you are willing to rewrite all the tests :smile:
@eladsal Yes boost test tells you where the failures are.
@michaeltryby you didn't really answer my first question. Let me rephrase it. Will the unit tests still exist as separate files in the \tests folder as they do now, which allows easy access to individual tests (and makes it easy to add new ones), or will they all be merged into a single file? I hope it's not the latter.
| gharchive/issue | 2019-03-21T14:08:17 | 2025-04-01T04:32:56.320802 | {
"authors": [
"LRossman",
"eladsal",
"michaeltryby"
],
"repo": "OpenWaterAnalytics/EPANET",
"url": "https://github.com/OpenWaterAnalytics/EPANET/issues/427",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1847610305 | Bug: Failed to Discover Devices after adding/updating the devices
Describe the Bug
The devices are not able to be found after adding/updating the children's devices.
I updated the existing device "switchbot bot" with different "device refresh rate".
I added a new device "switchbot bot" for the another new-buy bot.
After restarting the hb, the hb cannot find any devices nor hk showing any switchbot devices.
I have check the issue list, it seems that the below issue are similar
https://github.com/OpenWonderLabs/homebridge-switchbot/issues/775
https://github.com/OpenWonderLabs/homebridge-switchbot/issues/768
https://github.com/OpenWonderLabs/homebridge-switchbot/issues/757
Please help solve the bug, thank you =)
To Reproduce
Not sure
Expected Behavior
The homebridge-switchbot plugin should be able to get all the devices.
Relevant Log Output
[02/08/2023, 12:17:11] [SwitchBot] Bot: Bot Light Unknown statusCode: 500, Submit Bugs Here: ' + 'https://tinyurl.com/SwitchBotBug
[02/08/2023, 12:17:11] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Cannot read properties of undefined (reading 'power')"
[03/08/2023, 03:19:53] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Connect Timeout Error"
[03/08/2023, 03:20:03] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Connect Timeout Error"
[03/08/2023, 03:20:13] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Connect Timeout Error"
[03/08/2023, 03:20:23] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Connect Timeout Error"
[03/08/2023, 03:20:26] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:20:33] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:20:43] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:20:53] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:21:03] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:21:13] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:21:23] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:21:33] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:21:43] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:21:53] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:22:03] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com"
[03/08/2023, 03:22:13] [SwitchBot] Child bridge process ended
[03/08/2023, 03:22:13] [SwitchBot] Process Ended. Code: 1, Signal: null
[03/08/2023, 03:22:30] [SwitchBot] Initializing SwitchBot platform...
[03/08/2023, 03:22:30] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF
[03/08/2023, 03:22:31] [SwitchBot] Launched child bridge with PID 4131
[03/08/2023, 03:22:35] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.6.2 child bridge successfully
[03/08/2023, 03:22:35] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30}
[04/08/2023, 01:15:57] [SwitchBot] Initializing SwitchBot platform...
[04/08/2023, 01:15:57] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF
[04/08/2023, 01:15:59] [SwitchBot] Launched child bridge with PID 6361
[04/08/2023, 01:16:02] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[04/08/2023, 01:16:02] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30}
[04/08/2023, 01:16:07] [SwitchBot] Total SwitchBot Devices Found: 2
[04/08/2023, 01:16:07] [SwitchBot] Restoring existing accessory from cache: Air Conditioner DeviceID: 02-202307122218-42893239
[04/08/2023, 01:16:07] [SwitchBot] Air Conditioner: Air Conditioner Config: {"connectionType":"OpenAPI"}
[04/08/2023, 01:16:07] [SwitchBot] Total IR Devices Found: 1
[04/08/2023, 01:16:07] [SwitchBot] Restoring existing accessory from cache: Bot Light DeviceID: C13832304461
[04/08/2023, 01:16:07] [SwitchBot] Bot: Bot Light Config: {"mode":"press","deviceType":"switch","connectionType":"OpenAPI","logging":"standard","refreshRate":10,"offline":false}
[04/08/2023, 01:16:07] [SwitchBot] Bot: Bot Light Displaying as Switch
[04/08/2023, 01:16:14] [SwitchBot] Bot: Bot Light Set On: true
[04/08/2023, 01:16:29] [SwitchBot] Bot: Bot Light Set On: true
[04/08/2023, 01:29:04] [SwitchBot] Bot: Bot Light Set On: true
[04/08/2023, 10:41:50] [SwitchBot] Bot: Bot Light Set On: true
[04/08/2023, 22:18:28] [SwitchBot] Bot: Bot Light Set On: true
[04/08/2023, 22:50:57] [SwitchBot] Bot: Bot Light Set On: true
[06/08/2023, 00:07:39] [SwitchBot] Bot: Bot Light Set On: true
[06/08/2023, 00:17:50] [SwitchBot] Bot: Bot Light Set On: true
[06/08/2023, 10:42:14] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Cannot read properties of undefined (reading 'power')"
[06/08/2023, 19:25:51] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Cannot read properties of undefined (reading 'power')"
[07/08/2023, 01:29:29] [SwitchBot] Bot: Bot Light Set On: true
[08/08/2023, 22:25:06] [SwitchBot] Bot: Bot Light Set On: true
[09/08/2023, 21:00:04] [SwitchBot] Bot: Bot Light Unknown statusCode: 500, Submit Bugs Here: ' + 'https://tinyurl.com/SwitchBotBug
[09/08/2023, 21:00:04] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Cannot read properties of undefined (reading 'power')"
[11/08/2023, 01:32:33] [SwitchBot] Bot: Bot Light Set On: true
[11/08/2023, 09:55:28] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 09:55:28] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF
[11/08/2023, 09:55:30] [SwitchBot] Launched child bridge with PID 19374
[11/08/2023, 09:55:33] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 09:55:33] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30}
[11/08/2023, 09:55:35] [SwitchBot] Total SwitchBot Devices Found: 3
[11/08/2023, 09:55:35] [SwitchBot] Restoring existing accessory from cache: Air Conditioner DeviceID: 02-202307122218-42893239
[11/08/2023, 09:55:35] [SwitchBot] Air Conditioner: Air Conditioner Config: {"connectionType":"OpenAPI"}
[11/08/2023, 09:55:35] [SwitchBot] Total IR Devices Found: 1
[11/08/2023, 09:55:35] [SwitchBot] Restoring existing accessory from cache: Bot Light DeviceID: C13832304461
[11/08/2023, 09:55:35] [SwitchBot] Adding new accessory: Bot Door Bot DeviceID: CFF4619C509D
[11/08/2023, 09:55:35] [SwitchBot] Bot: Bot Light Config: {"mode":"press","deviceType":"switch","connectionType":"OpenAPI","logging":"standard","refreshRate":10,"offline":false}
[11/08/2023, 09:55:35] [SwitchBot] Bot: Bot Light Displaying as Switch
[11/08/2023, 09:55:35] [SwitchBot] Bot: Bot Door Config: {"mode":"switch","deviceType":"switch","connectionType":"OpenAPI","refreshRate":10}
[11/08/2023, 09:55:35] [SwitchBot] Bot: Bot Door Displaying as Switch
[11/08/2023, 09:56:38] [SwitchBot] Bot: Bot Door Set On: true
[11/08/2023, 09:56:54] [SwitchBot] Bot: Bot Door Set On: false
[11/08/2023, 09:56:58] [SwitchBot] Bot: Bot Door Set On: true
[11/08/2023, 09:57:09] [SwitchBot] Bot: Bot Door Set On: false
[11/08/2023, 09:57:22] [SwitchBot] Bot: Bot Door Set On: true
[11/08/2023, 09:57:38] [SwitchBot] Bot: Bot Door Set On: false
[11/08/2023, 09:59:18] [SwitchBot] Bot: Bot Door Set On: true
[11/08/2023, 09:59:24] [SwitchBot] Bot: Bot Door Set On: false
[11/08/2023, 10:02:06] [SwitchBot] Bot: Bot Door Set On: true
[11/08/2023, 10:02:22] [SwitchBot] Bot: Bot Door Set On: false
[11/08/2023, 14:57:50] [SwitchBot] Bot: Bot Door Set On: true
[11/08/2023, 14:59:55] [SwitchBot] Bot: Bot Door Set On: false
[11/08/2023, 23:03:34] [SwitchBot] Bot: Bot Door Set On: true
[11/08/2023, 23:03:47] [SwitchBot] Bot: Bot Door Set On: false
[11/08/2023, 23:05:09] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:05:09] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF
[11/08/2023, 23:05:11] [SwitchBot] Launched child bridge with PID 20503
[11/08/2023, 23:05:14] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:05:15] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30}
[11/08/2023, 23:10:29] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:10:29] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF
[11/08/2023, 23:10:31] [SwitchBot] Launched child bridge with PID 20599
[11/08/2023, 23:10:35] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:10:35] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30}
[11/08/2023, 23:13:53] [SwitchBot] Restarting child bridge...
[11/08/2023, 23:13:58] [SwitchBot] Child bridge process ended
[11/08/2023, 23:13:58] [SwitchBot] Process Ended. Code: 143, Signal: null
[11/08/2023, 23:14:05] [SwitchBot] Restarting Process...
[11/08/2023, 23:14:06] [SwitchBot] Launched child bridge with PID 20667
[11/08/2023, 23:14:09] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:14:09] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30}
[11/08/2023, 23:15:53] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:15:53] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF
[11/08/2023, 23:15:54] [SwitchBot] Launched child bridge with PID 20841
[11/08/2023, 23:15:58] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:15:58] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30}
[11/08/2023, 23:16:18] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:16:18] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF
[11/08/2023, 23:16:19] [SwitchBot] Launched child bridge with PID 20908
[11/08/2023, 23:16:23] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:16:23] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30}
[11/08/2023, 23:17:30] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:17:30] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF
[11/08/2023, 23:17:31] [SwitchBot] Launched child bridge with PID 21004
[11/08/2023, 23:17:35] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:17:35] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30}
[11/08/2023, 23:19:34] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:19:34] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF
[11/08/2023, 23:19:35] [SwitchBot] Launched child bridge with PID 21477
[11/08/2023, 23:19:39] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:19:39] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30}
[11/08/2023, 23:27:21] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:28:50] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:28:50] [SwitchBot] Initializing child bridge 0E:D2:C4:22:57:08
[11/08/2023, 23:28:52] [SwitchBot] Launched child bridge with PID 21790
[11/08/2023, 23:28:55] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:31:56] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:31:56] [SwitchBot] Initializing child bridge 0E:D2:C4:22:57:08
[11/08/2023, 23:31:58] [SwitchBot] Launched child bridge with PID 21871
[11/08/2023, 23:32:01] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:39:55] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:39:55] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41
[11/08/2023, 23:39:57] [SwitchBot] Launched child bridge with PID 22534
[11/08/2023, 23:40:00] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:41:50] [SwitchBot] Restarting child bridge...
[11/08/2023, 23:41:55] [SwitchBot] Child bridge process ended
[11/08/2023, 23:41:55] [SwitchBot] Process Ended. Code: 143, Signal: null
[11/08/2023, 23:42:02] [SwitchBot] Restarting Process...
[11/08/2023, 23:42:03] [SwitchBot] Launched child bridge with PID 22585
[11/08/2023, 23:42:06] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[11/08/2023, 23:44:11] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:44:11] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41
[11/08/2023, 23:44:13] [SwitchBot] Launched child bridge with PID 23104
[11/08/2023, 23:44:16] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[11/08/2023, 23:28:17] [SwitchBot] Initializing SwitchBot platform...
[11/08/2023, 23:28:17] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41
[11/08/2023, 23:28:18] [SwitchBot] Launched child bridge with PID 818
[11/08/2023, 23:47:10] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[12/08/2023, 00:06:47] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 00:06:47] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41
[12/08/2023, 00:06:49] [SwitchBot] Launched child bridge with PID 918
[12/08/2023, 00:06:53] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[12/08/2023, 00:18:03] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 00:18:03] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41
[12/08/2023, 00:18:05] [SwitchBot] Launched child bridge with PID 1147
[12/08/2023, 00:18:08] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[12/08/2023, 00:20:49] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 00:20:49] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41
[12/08/2023, 00:20:51] [SwitchBot] Launched child bridge with PID 1295
[12/08/2023, 00:20:54] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[12/08/2023, 00:23:28] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 00:23:28] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41
[12/08/2023, 00:23:30] [SwitchBot] Launched child bridge with PID 1470
[12/08/2023, 00:23:33] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[12/08/2023, 00:24:23] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 00:24:23] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41
[12/08/2023, 00:24:25] [SwitchBot] Launched child bridge with PID 1564
[12/08/2023, 00:24:28] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[12/08/2023, 00:25:55] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 00:25:55] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41
[12/08/2023, 00:25:57] [SwitchBot] Launched child bridge with PID 1702
[12/08/2023, 00:26:00] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[12/08/2023, 00:32:09] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 00:32:09] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41
[12/08/2023, 00:32:12] [SwitchBot] Launched child bridge with PID 1817
[12/08/2023, 00:32:15] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[12/08/2023, 00:50:22] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 00:50:22] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":15}
[12/08/2023, 00:53:37] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 00:53:37] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":15}
[12/08/2023, 00:55:25] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 00:55:25] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4
[12/08/2023, 00:55:27] [SwitchBot] Launched child bridge with PID 3691
[12/08/2023, 00:55:30] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[12/08/2023, 00:55:30] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":15}
[12/08/2023, 01:00:21] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 01:00:21] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4
[12/08/2023, 01:00:23] [SwitchBot] Launched child bridge with PID 3825
[12/08/2023, 01:00:27] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[12/08/2023, 01:00:27] [SwitchBot] [DEBUG] Using Config Logging: debug
[12/08/2023, 01:00:27] [SwitchBot] [DEBUG] Finished initializing platform:
[12/08/2023, 01:00:27] [SwitchBot] Platform Config: {"logging":"debug","refreshRate":15}
[12/08/2023, 01:00:27] [SwitchBot] [DEBUG] Config OK
[12/08/2023, 01:00:27] [SwitchBot] [DEBUG] Using Default Push Rate.
[12/08/2023, 01:00:28] [SwitchBot] [DEBUG] Executed didFinishLaunching callback
[12/08/2023, 01:00:31] [SwitchBot] [DEBUG] Devices: {}
[12/08/2023, 01:00:31] [SwitchBot] [DEBUG] Request successful, statusCode: 200
[12/08/2023, 01:00:31] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:00:31 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"f8c76a1c-190e-48cb-8bf7-d31a6fa2c66a","x-amz-apigw-id":"JgVfbHvMIAMEJBQ=","x-amzn-trace-id":"Root=1-64d6692f-61b99ffe79141ea3564ae157;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"}
[12/08/2023, 01:00:31] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":15,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}]
[12/08/2023, 01:00:31] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read property 'map' of undefined", Submit Bugs Here: https://tinyurl.com/SwitchBotBug
[12/08/2023, 01:00:31] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read property 'map' of undefined
[12/08/2023, 01:05:43] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 01:05:43] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4
[12/08/2023, 01:05:44] [SwitchBot] Launched child bridge with PID 4018
[12/08/2023, 01:05:48] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[12/08/2023, 01:05:48] [SwitchBot] [DEBUG] Finished initializing platform:
[12/08/2023, 01:05:48] [SwitchBot] [DEBUG] Using Config Logging: debug
[12/08/2023, 01:05:48] [SwitchBot] Platform Config: {"logging":"debug","refreshRate":6}
[12/08/2023, 01:05:48] [SwitchBot] [DEBUG] Using Default Push Rate.
[12/08/2023, 01:05:48] [SwitchBot] [DEBUG] Config OK
[12/08/2023, 01:05:49] [SwitchBot] [DEBUG] Executed didFinishLaunching callback
[12/08/2023, 01:05:50] [SwitchBot] [DEBUG] Devices: {}
[12/08/2023, 01:05:50] [SwitchBot] [DEBUG] Request successful, statusCode: 200
[12/08/2023, 01:05:50] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:05:50 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"56d969ef-a617-455a-bfc2-af2d4628a82c","x-amz-apigw-id":"JgWRSG56IAMEB7A=","x-amzn-trace-id":"Root=1-64d66a6e-28d5985205ff8f7b0b948d3a;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"}
[12/08/2023, 01:05:50] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":6,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}]
[12/08/2023, 01:05:50] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read property 'map' of undefined", Submit Bugs Here: https://tinyurl.com/SwitchBotBug
[12/08/2023, 01:05:50] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read property 'map' of undefined
[12/08/2023, 01:07:30] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 01:07:30] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4
[12/08/2023, 01:07:32] [SwitchBot] Launched child bridge with PID 4153
[12/08/2023, 01:07:35] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[12/08/2023, 01:07:35] [SwitchBot] [DEBUG] Using Config Logging: debug
[12/08/2023, 01:07:35] [SwitchBot] [DEBUG] Finished initializing platform:
[12/08/2023, 01:07:35] [SwitchBot] Platform Config: {"logging":"debug","refreshRate":6}
[12/08/2023, 01:07:35] [SwitchBot] [DEBUG] Using Default Push Rate.
[12/08/2023, 01:07:35] [SwitchBot] [DEBUG] Config OK
[12/08/2023, 01:07:36] [SwitchBot] [DEBUG] Executed didFinishLaunching callback
[12/08/2023, 01:07:39] [SwitchBot] [DEBUG] Devices: {}
[12/08/2023, 01:07:39] [SwitchBot] [DEBUG] Request successful, statusCode: 200
[12/08/2023, 01:07:39] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:07:39 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"5a6586c7-8f35-4fbf-b95d-ca62fe871743","x-amz-apigw-id":"JgWiXGH-IAMEJ8A=","x-amzn-trace-id":"Root=1-64d66adb-1cf5dc35044e47695c5d835c;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"}
[12/08/2023, 01:07:39] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":6,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}]
[12/08/2023, 01:07:39] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read property 'map' of undefined", Submit Bugs Here: https://tinyurl.com/SwitchBotBug
[12/08/2023, 01:07:39] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read property 'map' of undefined
[12/08/2023, 01:10:32] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 01:10:32] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4
[12/08/2023, 01:10:34] [SwitchBot] Launched child bridge with PID 5406
[12/08/2023, 01:10:37] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully
[12/08/2023, 01:10:37] [SwitchBot] [DEBUG] Using Config Logging: debug
[12/08/2023, 01:10:37] [SwitchBot] [DEBUG] Finished initializing platform:
[12/08/2023, 01:10:37] [SwitchBot] Platform Config: {"logging":"debug","refreshRate":6}
[12/08/2023, 01:10:37] [SwitchBot] [DEBUG] Using Default Push Rate.
[12/08/2023, 01:10:37] [SwitchBot] [DEBUG] Config OK
[12/08/2023, 01:10:38] [SwitchBot] [DEBUG] Executed didFinishLaunching callback
[12/08/2023, 01:10:40] [SwitchBot] [DEBUG] Devices: {}
[12/08/2023, 01:10:40] [SwitchBot] [DEBUG] Request successful, statusCode: 200
[12/08/2023, 01:10:40] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:10:40 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"ac52f94e-ee4c-446e-8c9b-78aa99a4ce66","x-amz-apigw-id":"JgW-iHJKIAMETRw=","x-amzn-trace-id":"Root=1-64d66b8f-396ff15e39a11ed262b961e4;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"}
[12/08/2023, 01:10:40] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":6,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}]
[12/08/2023, 01:10:40] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read properties of undefined (reading 'map')", Submit Bugs Here: https://tinyurl.com/SwitchBotBug
[12/08/2023, 01:10:40] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read properties of undefined (reading 'map')
[12/08/2023, 01:13:54] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 01:13:54] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4
[12/08/2023, 01:13:56] [SwitchBot] Launched child bridge with PID 6654
[12/08/2023, 01:13:59] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[12/08/2023, 01:13:59] [SwitchBot] [DEBUG] Using Config Logging: debug
[12/08/2023, 01:13:59] [SwitchBot] [DEBUG] Finished initializing platform:
[12/08/2023, 01:13:59] [SwitchBot] [DEBUG] Platform Config: {"logging":"debug","refreshRate":6}
[12/08/2023, 01:13:59] [SwitchBot] [DEBUG] Using Default Push Rate.
[12/08/2023, 01:13:59] [SwitchBot] [DEBUG] Config OK
[12/08/2023, 01:14:00] [SwitchBot] [DEBUG] Executed didFinishLaunching callback
[12/08/2023, 01:14:02] [SwitchBot] [DEBUG] Devices: {}
[12/08/2023, 01:14:02] [SwitchBot] [DEBUG] Request successful, statusCode: 200
[12/08/2023, 01:14:02] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:14:02 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"fc342878-cd12-4741-8da2-b83625fb834f","x-amz-apigw-id":"JgXeIF-7oAMFWyQ=","x-amzn-trace-id":"Root=1-64d66c5a-6efaa0f031f9a9b81e746a35;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"}
[12/08/2023, 01:14:02] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":6,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}]
[12/08/2023, 01:14:02] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read properties of undefined (reading 'map')", Submit Bugs Here: https://tinyurl.com/SwitchBotBug
[12/08/2023, 01:14:02] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read properties of undefined (reading 'map')
[12/08/2023, 01:15:12] [SwitchBot] Initializing SwitchBot platform...
[12/08/2023, 01:15:12] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4
[12/08/2023, 01:15:14] [SwitchBot] Launched child bridge with PID 6808
[12/08/2023, 01:15:17] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully
[12/08/2023, 01:15:17] [SwitchBot] [DEBUG] Using Config Logging: debug
[12/08/2023, 01:15:17] [SwitchBot] [DEBUG] Finished initializing platform:
[12/08/2023, 01:15:17] [SwitchBot] [DEBUG] Platform Config: {"logging":"debug","refreshRate":6}
[12/08/2023, 01:15:17] [SwitchBot] [DEBUG] Using Default Push Rate.
[12/08/2023, 01:15:17] [SwitchBot] [DEBUG] Config OK
[12/08/2023, 01:15:18] [SwitchBot] [DEBUG] Executed didFinishLaunching callback
[12/08/2023, 01:15:21] [SwitchBot] [DEBUG] Devices: {}
[12/08/2023, 01:15:21] [SwitchBot] [DEBUG] Request successful, statusCode: 200
[12/08/2023, 01:15:21] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:15:21 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"df281bbc-a423-418b-a3a4-f7ba0e8f93d1","x-amz-apigw-id":"JgXqeE00IAMErWw=","x-amzn-trace-id":"Root=1-64d66ca9-162845ec48b225120501fd7b;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"}
[12/08/2023, 01:15:21] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"BLE/OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":6,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}]
[12/08/2023, 01:15:21] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read properties of undefined (reading 'map')", Submit Bugs Here: https://tinyurl.com/SwitchBotBug
[12/08/2023, 01:15:21] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read properties of undefined (reading 'map')
Config for homebridge-switchbot
{
"bridge": {
"name": "Bill Homebridge",
"username": "0E:89:49:64:91:86",
"port": 51430,
"pin": "630-27-655"
},
"accessories": [],
"platforms": [
{
"name": "Config",
"port": 8581,
"auth": "form",
"theme": "auto",
"tempUnits": "c",
"lang": "auto",
"platform": "config"
},
{
"name": "SwitchBot",
"credentials": {
"token": "<hidden>",
"secret": "<hidden>",
"notice": "Keep your Token & Secret a secret!"
},
"options": {
"devices": [
{
"deviceId": "CF:F4:61:9C:50:9D",
"configDeviceName": "Bot Door",
"configDeviceType": "Bot",
"connectionType": "BLE/OpenAPI",
"bot": {
"mode": "switch",
"deviceType": "switch"
},
"firmware": "V6.3",
"refreshRate": 6,
"logging": "debug"
},
{
"deviceId": "C1:38:32:30:44:61",
"configDeviceName": "Bot Light",
"configDeviceType": "Bot",
"connectionType": "OpenAPI",
"bot": {
"mode": "press",
"deviceType": "switch"
},
"firmware": "V6.5",
"refreshRate": 15,
"external": false,
"logging": "debug"
}
],
"refreshRate": 6,
"logging": "debug"
},
"_bridge": {
"username": "0E:55:D6:94:91:A4",
"port": 57234
},
"platform": "SwitchBot"
}
],
"disabledPlugins": []
}
Screenshots
Device and Model
SwitchBot Bot
Node.js Version
v20.5.0
NPM Version
v9.8.0
Homebridge Version
v1.6.1
Homebridge Switchbot Plugin Version
v2.8.0-beta.12
Homebridge Config UI X Plugin Version
v4.50.4
Operating System
Raspbian
I'm having the same error.
+1. Was working fine then stopped working for no reason. Tried everything, removing and adding the plugin but now refuses to discover any device via API.
@mrlt8 have you experienced this at all?
The batteries in my bots are all dead so I haven't encountered this, but the API is pretty flakey and the culprit seems to be:
[02/08/2023, 12:17:11] [SwitchBot] Bot: Bot Light Unknown statusCode: 500, Submit Bugs Here: ' + 'https://tinyurl.com/SwitchBotBug
You'll probably want to assert the status code is 200 or throw an error around here to avoid the undefined message:
https://github.com/OpenWonderLabs/homebridge-switchbot/blob/09ff2520b0e2c7409a371e5c5c264aee17823a7e/src/device/bot.ts#L563-L566
or
at the end of the statusCode()?
something like:
if (statusCode !== 200) {
throw new Error(`Unexpected status code: ${statusCode}`);
}
As for discoverDevices, that one seems to return a 200, so maybe check the length of deviceLists instead of devicesAPI.body?
https://github.com/OpenWonderLabs/homebridge-switchbot/blob/472aef8bf69434f757c7cc573dc29d2b81d7bdf6/src/platform.ts#L295-L298
Pushed a beta for this:
As for discoverDevices, that one seems to return a 200, so maybe check the length of deviceLists instead of devicesAPI.body?
https://github.com/OpenWonderLabs/homebridge-switchbot/blob/472aef8bf69434f757c7cc573dc29d2b81d7bdf6/src/platform.ts#L295-L298
Can you try the latest beta:
Click here to find out how to install Beta
You'll probably want to assert the status code is 200 or throw an error around here to avoid the undefined message:
https://github.com/OpenWonderLabs/homebridge-switchbot/blob/09ff2520b0e2c7409a371e5c5c264aee17823a7e/src/device/bot.ts#L563-L566
or at the end of the statusCode()?
something like:
if (statusCode !== 200) {
throw new Error(`Unexpected status code: ${statusCode}`);
}
As for discoverDevices, that one seems to return a 200, so maybe check the length of deviceLists instead of devicesAPI.body?
https://github.com/OpenWonderLabs/homebridge-switchbot/blob/472aef8bf69434f757c7cc573dc29d2b81d7bdf6/src/platform.ts#L295-L298
Probably a good idea to also throw an error if the status code is not 200.
Modified both of these with beta.16
I would probably move the this.statusCode(statusCode); out of the if check and errorLog the statusCode:
this.statusCode(statusCode);
if (statusCode === 200) {
...
} else {
this.errorLog(`Failed <openAPIRefreshStatus/openAPIRefreshStatus>, statusCode: ${statusCode}`);
}
May also want to update all the other requests to catch any non-200 requests.
Since you already have the request wrapped in a try-catch, could potentially be easier to just update the statusCode method to throw an error instead, and the try-catch block should catch that:
async statusCode(statusCode: number): Promise<void> {
switch (statusCode) {
...
default:
this.infoLog(
`${this.device.deviceType}: ${this.accessory.displayName} Unknown statusCode: ` +
`${statusCode}, Submit Bugs Here: ' + 'https://tinyurl.com/SwitchBotBug`,
);
throw new Error(`Unknown Status Code: ${statusCode}`);
}
}
Hey @donavanbecker thanks for the latest beta! It is now working again!
Hi @donavanbecker thanks for the support.
I have updated to the latest version, v2.8.0-beta.19, but I still encounter the issue.
From my 3 days' observation, I found that my hb starts raising exceptions after 18:00 (UTC + 8) and it will recover after I wake up tomorrow after 8:00 (UTC + 8).
What could be the cause of making the plugin down every night?
Most likely an issue on the switchbot API. Might want to report it there: https://github.com/OpenWonderLabs/SwitchBotAPI, but support seems non-existent.
Or a network issue
I would probably move the this.statusCode(statusCode); out of the if check and errorLog the statusCode:
this.statusCode(statusCode);
if (statusCode === 200) {
...
} else {
this.errorLog(`Failed <openAPIRefreshStatus/openAPIRefreshStatus>, statusCode: ${statusCode}`);
}
May also want to update all the other requests to catch any non-200 requests.
Since you already have the request wrapped in a try-catch, could potentially be easier to just update the statusCode method to throw an error instead, and the try-catch block should catch that:
async statusCode(statusCode: number): Promise<void> {
switch (statusCode) {
...
default:
this.infoLog(
`${this.device.deviceType}: ${this.accessory.displayName} Unknown statusCode: ` +
`${statusCode}, Submit Bugs Here: ' + 'https://tinyurl.com/SwitchBotBug`,
);
throw new Error(`Unknown Status Code: ${statusCode}`);
}
}
@mrlt8 if you have time, wanna throw this in a PR?
@pinhao1994 is getting Unknown statusCode: 500 which is an "Internal Server Error" on the API: https://github.com/OpenWonderLabs/SwitchBotAPI#standard-http-error-codes
@pinhao1994 is getting Unknown statusCode: 500 which is an "Internal Server Error" on the API: https://github.com/OpenWonderLabs/SwitchBotAPI#standard-http-error-codes
@mrlt8 Could you give me more details on this?
I believe I am a stable network connection since my other smart devices (that are not Switchbot) still work perfectly. At the time when I am writing this message, I faced the same issue again. It is quite interesting that the SwitchBotAPI always die during the night.
I wonder if this could be caused by the request limit? Since I have a short pull time period, I could reach 1000 times API calls per day which caused the error.
Sounds like your theory is plausible! Mine stop working again and I have set it to 5 second API refresh rate, which would incur at least 17,280 calls per day.
I didn't think so at first as the API should return a 429 error.
However, I was doing some debugging and noticed that I kept getting a 200 response in Homebridge when taking my devices offline even though I would get a 171 when making a request directly to the API:
{"statusCode":171,"body":{},"message":"No hub record,is offline"}
It turns out the plug-in was only checking HTTP response and NOT the statusCode in the json response:
https://github.com/OpenWonderLabs/homebridge-switchbot/blob/472aef8bf69434f757c7cc573dc29d2b81d7bdf6/src/device/bot.ts#L538-L543
So we could potentially be ignoring some unexpected statusCode in the json response as long as the HTTP response was 200.
v2.8.0 is released
@nathennod @donavanbecker
I confirm the bug is caused by the 10k request limitation per day.
First, I increased the poll rate from 6 sec to 15 sec for 2 bots and the advanced settings. I did not face the issue after 10 pm (UTC+8). However, I meet the issue again at 7:30 am tomorrow morning, but the service recovers at 8:00 am (aka. 00:00 UTC). Therefore, I think although I have 3 poll rates to be exactly the same as 15 sec, I believe they could be polled separately and could still cause the issue. Lastly, I increase the rate to 60 sec and wish for the best.
@pinhao1994 Are you setting a refreshRate for each device? You can set different rates for each device.
I have 1 bot and I’ve set it to poll 30 seconds and so far so good!
@donavanbecker
I understand that we could set individual refreshRate for each device, but I suspect the devices are refreshed separately. That is to say, if I have one device, it could be refreshed as frequently as 9 sec which is equal to 9600 (< 10k) API calls per day.
My situation is that I have 2 switchbot bots, 1 IR for Air Conditioner, and 1 external IR for TV. Based on the settings, I could set individual refreshRate for each bot and one refreshRate in the "advanced settings" (I assume that the advanced settings refreshRate is for the IR devices".)
Now let's assume all devices will be refreshed "separately" in the background based on their refresh rates. Then, although I have max 10K API quota per day, each device could actually only be updated 10k / 4 = 2500 per day. That is to say, in the worst case, each device could only have refreshRate 86400 / 2500 ~= 35 sec.
Therefore, to optimize this part, we should first figure out the logic of refresh in the background.
If all the devices are refreshed together at once, then the cronjob refresh rate should be max(device1_rate, device2_rate, ..., deviceN_rate, 9) where 9 is the minimum rate to avoid reaching 10k limitation.
If all the devices are refreshed separately, then the cronjob refresh rate should be (86400 sec / (10k / num_of_devices)) = (86400 sec * num_of_devices) / 10k sec to avoid the worst case.
@pinhao1994 IR devices don't refresh data.
| gharchive/issue | 2023-08-12T00:54:33 | 2025-04-01T04:32:56.360705 | {
"authors": [
"donavanbecker",
"ievolk",
"mrlt8",
"nathennod",
"pinhao1994"
],
"repo": "OpenWonderLabs/homebridge-switchbot",
"url": "https://github.com/OpenWonderLabs/homebridge-switchbot/issues/789",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
} |
1052783477 | Bug: Switchbot plugin started crashing Homebridge today
Describe The Bug
Restarting Homebridge crashes when Switchbot plugin is configured
To Reproduce
Just restart. No changes to any of my SwitchBot devices or configuration have been made
Expected behavior
Shouldn't crash!
Relevant log output
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] {"statusCode":100,"body":{"deviceList":[{"deviceId":"98F4AB2F20EE","deviceName":"
Humidifier EE","deviceType":"Humidifier","enableCloudService":true,"hubDeviceId":"000000000000"},{"deviceId":"D8E9AA74909F","deviceName":"Li
ving Room Curtain Right","deviceType":"Curtain","hubDeviceId":"000000000000","curtainDevicesIds":["DFC80D28CE14","D8E9AA74909F"],"calibrate"
:true,"group":true,"master":false,"openDirection":"right"},{"deviceId":"DF7B6DA19FE7","deviceName":"Hub Mini E7","deviceType":"Hub Mini","hu
bDeviceId":"000000000000"},{"deviceId":"DFC80D28CE14","deviceName":"Living Room Curtain","deviceType":"Curtain","enableCloudService":true,"h
ubDeviceId":"DF7B6DA19FE7","curtainDevicesIds":["DFC80D28CE14","D8E9AA74909F"],"calibrate":true,"group":true,"master":true,"openDirection":"
left"},{"deviceId":"F037B7A5EED7","deviceName":"Bot D7","deviceType":"Bot","enableCloudService":true,"hubDeviceId":"DF7B6DA19FE7"},{"deviceI
d":"F1BAF4B4E58E","deviceName":"Living Room","deviceType":"Meter","enableCloudService":true,"hubDeviceId":"DF7B6DA19FE7"},{"deviceId":"F3745
EADBBFE","deviceName":"Bedroom","deviceType":"Meter","enableCloudService":true,"hubDeviceId":"DF7B6DA19FE7"}],"infraredRemoteList":[]},"mess
age":"success"}
[13/11/2021, 16:12:05] [SwitchBot] Total SwitchBot Devices Found: 7
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] SwitchBot Device Config Not Set: undefined
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Humidifier: 98F4AB2F20EE
[13/11/2021, 16:12:05] [SwitchBot] Restoring existing accessory from cache: Humidifier EE DeviceID: 98F4AB2F20EE
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Curtain: D8E9AA74909F
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] deviceId: D8E9AA74909F, curtainDevicesIds: DFC80D28CE14,D8E9AA74909F, master: fal
se, group: true, disable_group: undefined
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] [Curtain Config] disable_group: undefined
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Hub Mini: DF7B6DA19FE7
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Curtain: DFC80D28CE14
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] deviceId: DFC80D28CE14, curtainDevicesIds: DFC80D28CE14,D8E9AA74909F, master: tru
e, group: true, disable_group: undefined
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] [Curtain Config] disable_group: undefined
[13/11/2021, 16:12:05] [SwitchBot] Restoring existing accessory from cache: Living Room Curtain DeviceID: DFC80D28CE14
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Bot: F037B7A5EED7
[13/11/2021, 16:12:05] [SwitchBot] Restoring existing accessory from cache: Bot D7 DeviceID: F037B7A5EED7
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] undefined
[13/11/2021, 16:12:05] [SwitchBot] You must set your Bot to Press or Switch Mode
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Meter: F1BAF4B4E58E
[13/11/2021, 16:12:05] [SwitchBot] Restoring existing accessory from cache: Living Room DeviceID: F1BAF4B4E58E
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Meter: F3745EADBBFE
[13/11/2021, 16:12:05] [SwitchBot] Restoring existing accessory from cache: Bedroom DeviceID: F3745EADBBFE
[13/11/2021, 16:12:05] [SwitchBot] Total IR Devices Found: 0
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] IR Device Config Not Set: undefined
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Humidifier Humidifier EE CurrentTemperature: 0
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Humidifier uuid: 98F4AB2F20EE-Humidifier, (179d92bf-d829-46ee-b45f-954e1a4e6bf1)
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Curtain Living Room Curtain updateHomeKitCharacteristics - {"CurrentPosition":0,"
PositionState":2,"TargetPosition":0}
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Curtain Living Room Curtain CurrentAmbientLightLevel: undefined
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Curtain uuid: DFC80D28CE14-Curtain, (290ba6f5-8474-4a54-8d14-ce396a2e4251)
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Bot Bot D7 On: false
[13/11/2021, 16:12:05] [SwitchBot] Bot Bot D7 failed to refresh status, Error Message: "Cannot read properties of undefi
ned (reading 'updateCharacteristic')"
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Bot Bot D7, Error: {}
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Bot Bot D7 BatteryLevel: undefined
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Bot Bot D7 StatusLowBattery: undefined
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Bot uuid: F037B7A5EED7-Bot, (cffc0071-664c-4055-b95d-fc427c34c3a8)
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Meter uuid: F1BAF4B4E58E-Meter, (c79c02a1-a3a0-4f6f-ac00-41124aa94a2a)
[13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Meter uuid: F3745EADBBFE-Meter, (97065854-0ef6-426d-9a6f-5604eb4076d8)
[13/11/2021, 16:12:05] TypeError: Cannot read properties of undefined (reading 'updateCharacteristic')
at Bot.apiError (/usr/local/lib/node_modules/@switchbot/homebridge-switchbot/src/devices/bots.ts:356:18)
at Bot.openAPIRefreshStatus (/usr/local/lib/node_modules/@switchbot/homebridge-switchbot/src/devices/bots.ts:233:12)
at Bot.refreshStatus (/usr/local/lib/node_modules/@switchbot/homebridge-switchbot/src/devices/bots.ts:168:18)
at new Bot (/usr/local/lib/node_modules/@switchbot/homebridge-switchbot/src/devices/bots.ts:52:10)
at SwitchBotPlatform.createBot (/usr/local/lib/node_modules/@switchbot/homebridge-switchbot/src/platform.ts:440:9)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
[13/11/2021, 16:12:05] Got SIGTERM, shutting down Homebridge...
Config for homebridge-switchbot
"name": "SwitchBot",
"credentials": {
"openToken": "secret",
"notice": "Keep your token a secret!"
},
"options": {
"refreshRate": 600,
"debug": "debug"
},
"platform": "SwitchBot"
},
Screenshots
None
Device & Model
N/A
Node.js Version
v16.13.0
NPM Version
v8.1.0
Homebridge Version
v1.3.6
Homebridge Switchbot Plugin Version
v1.0.0
Homebridge Config UI X Plugin Version
v4.41.2
Operating System
Raspbian
Same error here.
I just update for the new version of the Switch bot plugin and crash my homebridge instance.
duplicate of #77
I'm still having the same issue even when updated to v1.0.1
@deflao Pleas post more logs because more then likely it is a different issue.
| gharchive/issue | 2021-11-13T21:54:35 | 2025-04-01T04:32:56.370856 | {
"authors": [
"deflao",
"donavanbecker",
"nfsilva999",
"nyjklein"
],
"repo": "OpenWonderLabs/homebridge-switchbot",
"url": "https://github.com/OpenWonderLabs/homebridge-switchbot/issues/79",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
} |
503398881 | Running a live relay requires 10 ETH???
username@computer:~/servers$ npx oz-gsn run-relayer
Could not connect to node at http://localhost:8545 (Error: Found no accounts with sufficient balance (10000000000000000000 wei)).
Is there a way to bring that down a bit? 10000000000000000000 wei is 10 ETH and that's like $1800 USD pushing it out of the reach of a lot of people that would like to join up. That's the same cash outlay as a good quality mining rig. Why do we gotta be so rich to help out?
Thanx!
The helpers look for an account with 10 eth because you did not specify --from, so the library has to pick one of your unlocked accounts, and it does so by looking for enough balance that it knows transactions won't fail. It is true however that this number could be lower, but the selecting of a default account feature was intended for testing environments: you'd probably want to make explicit which account transactions will come from.
That said, running a relayer does require a deposit on the RelayHub contract of 1 ETH: this is to protect users of the GSN from malicious relayers. You can read more about attack vectors and how they are mitigated (partly using this deposit) on the relevant EIP section: https://eips.ethereum.org/EIPS/eip-1613#attacks-and-mitigations
Thank you, this is helpful information. Wouldn't it be better to have the program either require a "from" field or present the user with a list of accounts it found?
Also according to this... https://gist.github.com/spalladino/f516b46bdefb5e025c5d28daf7ad1491 it is possible to use infura as the provider and keep a local wallet seperately. But I can't figure out that option with oz-gsn.
Ah, yes - if your intent is to run a relayer on a real testnet, then I'd suggest following the guide on that gist. While more involved, you'll achieve a more robust result than simply running a process from JavaScript. run-relayer was originally intended to easily run a relayer locally on a testnet for testing purposes, not production, which is why e.g. it has a default from value.
| gharchive/issue | 2019-10-07T11:43:11 | 2025-04-01T04:32:56.381193 | {
"authors": [
"edukaded",
"nventuro"
],
"repo": "OpenZeppelin/openzeppelin-gsn-helpers",
"url": "https://github.com/OpenZeppelin/openzeppelin-gsn-helpers/issues/23",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
438386349 | Add backwards compatilibty with ERC20 for ERC777
A major feature of ERC777 is that it can be used as if it were an ERC20 token: we should implement this interface to enable this use case.
Due to ERC777's state variables being private, the most straightforward way to achieve this would be by having this backward compatibility baked-in into the contract. We could then provide a second contract that disables the ERC20 functions (by calling revert on them) for users to opt-out of this compatibility.
I don't think we should provide the non-ERC20-compatible contract for now. Let's wait and see if there is demand for it.
| gharchive/issue | 2019-04-29T15:52:49 | 2025-04-01T04:32:56.382764 | {
"authors": [
"frangio",
"nventuro"
],
"repo": "OpenZeppelin/openzeppelin-solidity",
"url": "https://github.com/OpenZeppelin/openzeppelin-solidity/issues/1731",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2594649552 | [Feature]: ERC1155Burnable extension
What is the feature you would like to see?
Add ERC1155Burnable extension (solidity).
Contribution Guidelines
[X] I agree to follow this project's Contribution Guidelines
Blocked by #261
Bringing back @programskillforverification code.
| gharchive/issue | 2024-10-17T12:51:31 | 2025-04-01T04:32:56.384668 | {
"authors": [
"bidzyyys"
],
"repo": "OpenZeppelin/rust-contracts-stylus",
"url": "https://github.com/OpenZeppelin/rust-contracts-stylus/issues/362",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
251025379 | pob white crash screen on booting up.
so when i updated pob to the latest version everything was fine until i rebooted my pc every time i open up pob it will just be a big white screen and say that its not responding i tried opening pob as an admin but same problem...
every time i reinstall pob it work until i reboot my pc and then its the same problem please help.
i play poe alot and need pob to make my builds and try out stuff
Am I to assume from your comment in #450 that you're having exactly the same issue (i.e it only occurs when opening on a non-primary monitor)? If so, then I'll close this.
| gharchive/issue | 2017-08-17T18:04:02 | 2025-04-01T04:32:56.386057 | {
"authors": [
"Novadesu",
"Openarl"
],
"repo": "Openarl/PathOfBuilding",
"url": "https://github.com/Openarl/PathOfBuilding/issues/570",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
692174592 | PD: Air gap on dispense
Background
As a user I may want to air gap after a dispense to prevent drips/contamination between wells (in a multi dispense) and during pipette movement across the deck.
Acceptance Criteria
[ ] Add a checkbox for air gap after Blow out (https://www.figma.com/file/0dTEbGNsKYgNwKEW6LLfzJ/Airgap-Delay?node-id=50%3A850)
[ ] Add tooltip on hover (https://www.figma.com/file/0dTEbGNsKYgNwKEW6LLfzJ/Airgap-Delay?node-id=104%3A0)
[ ] Min validation Air gap default is the pipette's min. If there is not enough room left in the tip the default will be the largest value possible between the pipette's min and 0. If a user goes beneath the pipette's recommended minimum the user will get a dismissible warning.
[ ] Max validation
The remaining volume in a tip after the first (or only) dispense is as follows:
(Tip capacity - occupied volume of tip before the dispense) + first dispense volume + air gap volume.
If the user is doing a multi aspirate or 1:1 Transfer then the remaining volume in the entire tip capacity. The max is therefore the tip's capacity.
If a user is doing a multi dispense then the space for an airgap after the first dispense is any left over room in the tip after the aspirate, plus what was just dispensed (first dispense volume + air gap volume). This number gets bigger as the tip does more dispenses, but to keep things simple we will set the max at the smallest air gap.
[ ] Order of operations for single transfer (https://www.figma.com/file/0dTEbGNsKYgNwKEW6LLfzJ/Airgap-Delay?node-id=453%3A6016)
[ ] Order of operations for multi dispense (https://www.figma.com/file/0dTEbGNsKYgNwKEW6LLfzJ/Airgap-Delay?node-id=453%3A6016)
[ ] Order of operations for multi dispense (https://www.figma.com/file/0dTEbGNsKYgNwKEW6LLfzJ/Airgap-Delay?node-id=453%3A6016)
In terms of breaking this down into more technical, bite sized pieces, I see the tickets as:
Add checkbox + tooltip for air gap on dispense (UI)
Add air gap dispense min validation for air gap on dispense (all three paths)
Add air gap dispense max validation for 1:1 transfer and multi aspirate
Add air gap dispense max validation for multi dispense
Add atomic commands for air gap dispense for single transfer
Add atomic commands for air gap dispense for multi aspirate
Add atomic commands for air gap dispense for multi dispense
Closing this as it is broken down into:
#6347
#6500
#6508
#6510
#6512
Note, still need a ticket for max air gap volume validation for multi dispense. Need to clarify whether it is okay if we clamp the maximum value down to whatever the max air gap aspirate value is. @howisthisnamenottakenyet
| gharchive/issue | 2020-09-03T18:11:23 | 2025-04-01T04:32:56.394419 | {
"authors": [
"howisthisnamenottakenyet",
"shlokamin"
],
"repo": "Opentrons/opentrons",
"url": "https://github.com/Opentrons/opentrons/issues/6482",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1164165728 | Show a link to view Robot release notes
give users a way to see the release notes for the current robot version. In the Robot Settings > Advanced > Robot Server version section, display a link to see release notes. On click, a modal will display the release notes for the current robot version.
Nice-to-have for 5.1
This has been solved by #10010
| gharchive/issue | 2022-03-09T16:29:37 | 2025-04-01T04:32:56.395777 | {
"authors": [
"emilywools",
"koji",
"mikecardarelli"
],
"repo": "Opentrons/opentrons",
"url": "https://github.com/Opentrons/opentrons/issues/9666",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1826797759 | docs(api,robot-server): Document errors list as having no more than 1 element
Overview
In the public (HTTP-facing) models for runs and analyses, you can access the errors of individual commands. In addition, there's an array of top-level errors. That top-level array is used to indicate fatal errors, including those that happen in between commands.
This documents, officially, that the top-level array will only ever have zero elements or one element. It will never contain multiple errors. The fact that it's an array, instead of a nullable object, is just a historical quirk.
This goes towards RSS-146 by clarifying how we should expose different kinds of run failures.
Rationale
There is no use for errors to have multiple elements today.
This array is left over from a time before individual commands held their own errors. Today, commands have a self-contained error field, but in þe olden tymes, they had an errorId reference that pointed into this array. This was problematic for a few reasons, such as inherent duplication.
PR #9715 fixed this. Since then, errors has only had up to 1 element.
Having multiple elements is confusing, and muddies what the field is supposed to be doing, conceptually. One internal comment describes it as "a list of fatal errors." But there can only be one fatal error, by definition.
For the case where multiple things contributed to a run's failure, and we want to convey them all, we now have @sfoster1's EnumeratedError architecture. We should commit to that tree-based API instead of keeping around a competing array-based API.
According to @shlokamin, the app just does, and has always just done, something like errors[0].
If we really need to have multiple elements in this list later on, we can just revert this documentation change. We can do that revert without breaking any clients.
Test Plan
None needed.
Changelog
Update the public HTTP documentation in run and analysis models, and internal comments.
Review requests
Do we agree with this?
Risk assessment
No risk.
Got an in-person approval from @sanni-t.
| gharchive/pull-request | 2023-07-28T17:22:52 | 2025-04-01T04:32:56.401628 | {
"authors": [
"SyntaxColoring"
],
"repo": "Opentrons/opentrons",
"url": "https://github.com/Opentrons/opentrons/pull/13191",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2198511262 | feat(app): populate ChooseRobotSlideout with runtime parameters
closes AUTH-98
closes AUTH-99
closes AUTH-101
Overview
Populate second page of ChooseRobotSlideout with RTPs.
Add link for restoring values to default.
Save state for in an object that will be used in createRunFromProtocolSource
Still todo:
error handling for numeric input fields
sending updated RTP object along with srcFiles during protocol run creation
Test Plan
select any protocol (this PR uses mock data until analysis provides RTP)
select Start setup button to open ChooseRobot slideout
select any robot and 'Continue to parameters' footer button
observe toggle, numerical input fields, and dropdown menus set to their default values (NOTE: dropdown fields will initialize with the first choice whose value matches its RTP's default value, regardless of its position in the array of choices)
observe tooltip on disabled 'Restore default values' link when all values are default
observe that all values are restored to default after any RTPs are changed and 'Restore default values' link is clicked
Changelog
Review requests
authorship stakeholders
Risk assessment
Could you update the tests?
@koji @jerader Thank you guys so much for the detailed reviews! I have made the changes you suggested/commented where appropriate. I will push one more commit for adding test coverage for slideouts passing and rendering RTPs and then can merge once approved.
| gharchive/pull-request | 2024-03-20T20:35:23 | 2025-04-01T04:32:56.407307 | {
"authors": [
"koji",
"ncdiehl11"
],
"repo": "Opentrons/opentrons",
"url": "https://github.com/Opentrons/opentrons/pull/14706",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
380444859 | refactor(api): more functions use hc
Introduce a ton of changes to inject hardware controller instances everywhere, depend on singletons less and be able to run on both v1 and v2 apis.
This does not extend to deck calibration, protocol sessions (including labwarecalibration), or testing tools.
In addition, there’s work done on the tests so that tests can more easily run on both versions of the api while relying on simple fixtures; async_client and main_router and the things that depend on them, for instance, now will run on both apis unless marked otherwise.
Many small additions have been made to both the new and old APIs to get them to work a bit more like each other to make the logic necessary in dependent systems like the servers to be a little more similar between versions.
** Review Requests **
Look it over. I know there's a lot of it but most of it really can't be separated out. Surface questions you have. I appreciate that this might take a while to review.
This PR should actually now come up on a robot with the feature flag set. Try it out by pushing it, connecting to the robot, and using the run app to switch the use protocol api v2 feature flag. When you reset the robot, it should be using protocol API 2.
For all of the changes here, the only external facing parts of the system that use this are the endpoints in /control. Try connecting to the robot and flipping lights, homing, refreshing connected modules, that sort of thing.
Closes #2245
Performed a few tests on the robot, here are my notes.
With Flag ON
[ ] Modules still work
[ ] Change pipettes works
[ ] Deck calibration works
[ ] Homing works
[ ] Lights work
With Flag OFF
[ ] share=True
appears to be broken (don't think it's necessarily as a result of this PR)
[ ] Modules still work
[ ] Change pipettes works
[ ] Deck calibration works
I got an alarm: Hard limit error on the Z (robot was not super off on height upon starting) -- can someone test on their robot to see if they have issues?
[ ] Homing works
[ ] Lights work
| gharchive/pull-request | 2018-11-13T22:05:05 | 2025-04-01T04:32:56.412802 | {
"authors": [
"Laura-Danielle",
"sfoster1"
],
"repo": "Opentrons/opentrons",
"url": "https://github.com/Opentrons/opentrons/pull/2675",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1045508802 | Op 7901
We prefer small, well tested pull requests.
Please refer to Contributing to Spinnaker.
When filling out a pull request, please consider the following:
Follow the commit message conventions found here.
Provide a descriptive summary for your changes.
If it fixes a bug or resolves a feature request, be sure to link to that issue.
Add inline code comments to changes that might not be obvious.
Squash your commits as you keep adding changes.
Add a comment to @spinnaker/reviewers for review if your issue has been outstanding for more than 3 days.
Note that we are unlikely to accept pull requests that add features without prior discussion. The best way to propose a feature is to open an issue first and discuss your ideas there before implementing them.
https://devopsmx.atlassian.net/browse/OP-7901
This PR triggered this build in this execution, please merge to master only if this succeeds: https://spin.prod.opsmx.com/#/applications/gitprtrigger/executions/details/01FKQF0BWWFYKBXH7PQ9KXR16G
| gharchive/pull-request | 2021-11-05T07:11:29 | 2025-04-01T04:32:56.425916 | {
"authors": [
"Pranav-b-7",
"opsmxuser"
],
"repo": "OpsMx/gate",
"url": "https://github.com/OpsMx/gate/pull/75",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1404126821 | Add backendRoles to userrolebinding
Fixes https://github.com/Opster/opensearch-k8s-operator/issues/294
@idanl21 @prudhvigodithi Are you OK to merge this one?
Hey @swoehrl-mw i'm good with this. Thanks for the PR @AleksZimin
| gharchive/pull-request | 2022-10-11T07:18:27 | 2025-04-01T04:32:56.427349 | {
"authors": [
"AleksZimin",
"prudhvigodithi",
"swoehrl-mw"
],
"repo": "Opster/opensearch-k8s-operator",
"url": "https://github.com/Opster/opensearch-k8s-operator/pull/313",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1890691226 | add readme for speculative decoding
add readme for speculative decoding under ./scripts/speculative_decoding/README.md
add specific cli examples
| gharchive/pull-request | 2023-09-11T14:50:10 | 2025-04-01T04:32:56.432682 | {
"authors": [
"wheresmyhair"
],
"repo": "OptimalScale/LMFlow",
"url": "https://github.com/OptimalScale/LMFlow/pull/641",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
549009081 | Prepare v4.4.1
[x] Run linters;
[x] Run compilers;
[x] Run tests;
[x] Check documentation site: examples and contents;
[x] Test cross-browser compatibility locally and with BrowserStack:
Firefox ESR
IE11
Latest Edge, Chrome, Firefox, Safari
iOS Safari
Chrome & Firefox on Android
[x] Including RTL mode;
[x] Ask for reviews and accessibility testing;
[x] npm run release-version to bump version number
[x] npm run release to compile dist, update SRI hashes in doc and package the release
[x] Prepare changelog:
install Conventionnal Changelog globally
conventional-changelog -p angular -i CHANGELOG.md -s
maintain a ship list … #287
[x] Wait for brand approval or suggestions about current items in navigation
[x] implement changes (in components, Boostwtach, examples…)
[x] re-run npm run release
[x] update Ship List + changelog
Then:
[ ] commit and push with a chore(release) commit message
[ ] merge (on v4-dev for now)
[ ] tag your version
[ ] create a GitHub release:
attach zip file
paste CHANGELOG in the release's description
[ ] publish documentation on gh-pages
copy _gh_pages/ top level files (index.html, redirects.json, sitemap.xml…) and _gh_pages/docs/4.4 to the gh-pages branch
ensure Algolia indexes new release content (probably requires a PR)
[ ] npm pack then npm publish
if you are already logged in NPM (with a personnal account, for example), you'd better use a repository scoped .npmrc file
[ ] publish on Nuget
[ ] check release on NPM, Nuget, Packagist…
Close #287 on merge.
New comments to check:
[x] Reduce the height of the Supra bar by 2px to 40px.
[ ] Align the first navigation label ‘Personal’ so it is left aligned to the Orange logo.
[ ] Increase the spacing between the labels in the Supra bar to 15px.
[ ] The space between the account icon and the language selection (in the supra bar) needs to be 30px.
[ ] The icons in the global header need to move up 2px.
| gharchive/pull-request | 2020-01-13T15:35:37 | 2025-04-01T04:32:56.453756 | {
"authors": [
"ffoodd"
],
"repo": "Orange-OpenSource/Orange-Boosted-Bootstrap",
"url": "https://github.com/Orange-OpenSource/Orange-Boosted-Bootstrap/pull/285",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1507943882 | Proposal: further tags for output cache
https://github.com/OrchardCMS/Orchard/blob/028e2e413b28c46bfcc6e07417a6c28b23bfb523/src/Orchard.Web/Modules/Orchard.OutputCache/Filters/OutputCacheFilter.cs#L261
OutputCache is currently using the Ids of the displayed ContentItems as tags for the OutputCache.
In a few of our applications, we've found this somewhat limiting, so we've had to find workarounds, in cases where a content that isn't, strictly speaking, being displayed nonetheless is relevant to what should be shown, and a change in its state should cause the output cache to be evicted. A common example would be some setting or configuration item.
We propose adding an interface allowing injection of additional tags, so it's easier for features to control the eviction of elements from output cache.
The new interface would look something like:
public interface ITagsProvider : IDependency {
IEnumerable<string> GetTags();
}
In the OutputCacheFilter we would then inject a IEnumerable<ITagsProvider > _tagsProviders, that we would use (syntax is approximate):
Tags = new[] { _invariantCacheKey }
.Union(cacheItemTags.Distinct())
.Union(._tagsProviders.SelectMany(tp => tp.GetTags()).Distinct())
.ToArray(),
Each feature where we would implement one of these would then be able to cause its own cache evictions through its own tags.
What about the other way around, a service in the output cache that allows to add custom tags to the current request. This way modules won't have to store what has changed until the output cache calls them.
If I understand what you are saying, you mean something that works like the IDisplayedContentItemHandler, but rather than being invoked only on the BuildDisplay (because it's a ContentHandler), it would be possible to invoke it at any point in the code.
public interface ITagAdderService : IDependency {
void AddTags(params string[] tags);
IEnumerable<string> GetTags();
}
Anywhere in code:
[...]
_tagAdderService("foo", "bar");
[...]
In the OutputCacheFilter we would then inject a ITagAdderService _tagAdderService, that we would use like:
Tags = new[] { _invariantCacheKey }
.Union(cacheItemTags.Distinct())
.Union(._tagAdderService.Select(tp => tp.GetTags()).Distinct())
.ToArray(),
Is that correct?
Exactly.
Look around orchard core, we might already be doing that.
| gharchive/issue | 2022-12-22T14:02:07 | 2025-04-01T04:32:56.462460 | {
"authors": [
"MatteoPiovanelli-Laser",
"sebastienros"
],
"repo": "OrchardCMS/Orchard",
"url": "https://github.com/OrchardCMS/Orchard/issues/8641",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1422221507 | 1.4. UX Report | Evaluative & Formative Research
[x] Usability testing of wireframes: Grant issuers and recipients (link report and survey)
[ ] Collect feedback on improvements (link report)
[ ] iterate (updated wireframes, figma link)
The UI for usability testing is currently in development. See issue #3
| gharchive/issue | 2022-10-25T10:14:11 | 2025-04-01T04:32:56.489663 | {
"authors": [
"XyloDrone"
],
"repo": "OrdumLTD/UX-UI",
"url": "https://github.com/OrdumLTD/UX-UI/issues/4",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
356092581 | 📙 Remove all but ipfs urls from listing pictures before indexing.
Instead of removing all pictures data, let's keep IPFS urls.
Some merge conflicts due to recent schema changes.
Also, there has been some recent updates to the DataStore so I don't expect any listing to have pictures as Data URLs anymore. See this code.
So I'm not sure we need to do any filtering anymore ? Except if it is to guard against listing data uploaded directly by partners/malicious actors and that bypassed origin-js code.
But maybe for that we should add some additional guards in marketplace.getListing to detect those potentially malformed/malicious listings ?
I'll close this PR, no need for it anymore.
We do need to make sure we always keep it a throwing error in origin.js when loading a listing using http/https images, since anyone could track people by using them.
| gharchive/pull-request | 2018-08-31T19:04:29 | 2025-04-01T04:32:56.516510 | {
"authors": [
"DanielVF",
"franckc"
],
"repo": "OriginProtocol/origin-js",
"url": "https://github.com/OriginProtocol/origin-js/pull/394",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
330867949 | Missing template / detailed description for ERP customers how to export the GS1 XML file in an automated way or process
Expected Behavior
Provide detailed concept and implementation examples for the large ERP software vendors how the XML export should be implemented, ideally without customization of the ERP. Customization is moving the customer away from the standard and is therefore seen as bad practice.
Actual Behavior
A customer using Navision or SAP ERP will not be willing to implement the XML export from the scratch or the reinvent the XML export wheel per customer.
No detailed description how the XML export should be implemented for a large ERP vendors like Navision, Infor or SAP.
https://github.com/OriginTrail/ot-node/wiki/ERP-Customization
Specifications
https://community.dynamics.com/ax/b/axvanyakashperuk/archive/2014/09/16/tutorial-generating-shipping-labels-using-the-gs1-sscc-18-barcode-format
https://help.sap.com/saphelp_me60/helpdata/EN/f7/86c1536ca9b54ce10000000a174cb4/frameset.htm
More detailed instructions can be found here
https://github.com/OriginTrail/ot-node/wiki/Data-Structure-Guidelines
You can also find templates (examples) here:
https://github.com/OriginTrail/ot-node/tree/develop/importers
Thanks for covering my question,
However, I am not sure if we are on the same page regarding this issue.
Image a meeting w. a ERP customer and the CIO asks his staff to implement the importer in an ERP systems.
I am pretty sure that a average Joe ERP basis supported will be lost using the current installation.
They will ask what OT does recommend, XSD or JSON? As well they will ask which data-structures in their ERP contains the required fields.
They need a step by step installation guide how to implement the exporter for their ERP systems.
Otherwise a consultancy will charge a small project for implementing the exporter on a ERP.
| gharchive/issue | 2018-06-09T08:54:13 | 2025-04-01T04:32:56.523109 | {
"authors": [
"ConcernedOriginTrailInvestor",
"schrink"
],
"repo": "OriginTrail/ot-node",
"url": "https://github.com/OriginTrail/ot-node/issues/182",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1213244706 | Scroll Physics
Can you please add physics property for more customization.
@ElsayedDev
Hmm..
There is not enough explanation Can you give me a more detailed example?
| gharchive/issue | 2022-04-23T08:36:43 | 2025-04-01T04:32:56.524261 | {
"authors": [
"ElsayedDev",
"Origogi"
],
"repo": "Origogi/Vertical_Card_Pager",
"url": "https://github.com/Origogi/Vertical_Card_Pager/issues/5",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1205420374 | Workflow test
Merging workflow test branch
Fixing dependa bot yml
| gharchive/pull-request | 2022-04-15T09:15:12 | 2025-04-01T04:32:56.524973 | {
"authors": [
"OrionH"
],
"repo": "OrionH/Tag-Bot",
"url": "https://github.com/OrionH/Tag-Bot/pull/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2083866601 | 🛑 RoboCheck is down
In 094f548, RoboCheck (https://robocheck.ru) was down:
HTTP code: 0
Response time: 0 ms
Resolved: RoboCheck is back up in aabd42f after 6 days.
| gharchive/issue | 2024-01-16T12:35:43 | 2025-04-01T04:32:56.540412 | {
"authors": [
"OsintUK"
],
"repo": "OsintUK/Up-or-Down",
"url": "https://github.com/OsintUK/Up-or-Down/issues/1518",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1562358784 | 🛑 HakePok is down
In 2ee6778, HakePok (http://hakepok.ru) was down:
HTTP code: 500
Response time: 821 ms
Resolved: HakePok is back up in ff2b2c0.
| gharchive/issue | 2023-01-30T12:31:20 | 2025-04-01T04:32:56.542499 | {
"authors": [
"OsintUK"
],
"repo": "OsintUK/Up-or-Down",
"url": "https://github.com/OsintUK/Up-or-Down/issues/477",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
259787664 | New winrm upload functionality
Two threads here:
New winrm_upload function:
This allows file transfer through WinRM sreaming. It's not speedy for big binary files, but works well for
text/config files. There are unit and integration tests covering it.
The origin of this PR is that we need to provide sysprep unattend.xml files to be able to change the SID
in windows machines in a fully automated way. It's a bit frustrating, but sysprep is the only supported
tool that performs that task, and the only way to fully automate it is to provide config in xml form that
gets read after the reboot (Otherwise you get undesired side effects like Administrator password reset
that basically breaks the downstream of your automation pipeline). The reason behind changing SID's
is that they have to be unique if you want to have them in the same domain, and cloning windows
boxes from a Vsphere template sadly gives them the same SID. It all looks a bit too complicated, but
this opens a big door for further Windows automation.
Fix and improve integration tests:
I have fixed some race conditions that appeared recently in the integration tests. I've also made
them faster by optimizing the number of times we create and delete the vms and I've made Jenkins to
run them in parallel (Python 2 and 3). Builds are now ~5 times faster: 14 minutes against 1 hour and 6
minutes.
I ran a few tests of different files. When I started to upload the jre-8u144-windows-x64.exe, it failed with an error similar to above. It failed in different places, sometimes after 4K, after 20K or in its furthest run after 208K.
(As an aside it copied at slightly over 2KB/s)
Alternatives are:
Install SSH, which is a bit clunky in Windows, as you need to do it through 3rd parties
Setup an SFTP server with Windows IIS
Mount a file share
All of them require adding extra stuff to our windows templates or doing it at runtime, which seems to be overkill to send a config file, which is the use case here. So I thought re-using winRM could do the trick, as discussed here https://github.com/diyan/pywinrm/issues/18. I knew it was slow, but I was not expecting it to fail with binaries.
I'll take a look to see if I can improve it.
By the way, speed is limited by the maximum size that WinRM allows for remote powershell scripts, which is something of the same magnitude of the steps default value of the winrm_upload function. So I don't think we can do much better there unless I am missing something.
I cannot find a reliable way of guessing whether a file is a text file or not. Some guys recommend using mimetypes, some others looking at null bytes. I'll give a test to that java exe to see if I can debug the error.
@edsharp I have found the problem and I think I have fixed it. Apparently there's a race condition where you can have the previous write process still open when you try to add more bytes to the file and that makes it fail, so you need to provide a small delay for that case. I tested it with your java binary until I got tired of waiting after 10-15 minutes with no errors at 2% and 1MB (It would have taken hours at such speed), and then I tried other binaries and they all went good, with the right sha256.
I have added after your last review:
File being read in chunks
Fix for the race condition and the test to cover that case
Better debug information for the RemoteCommandError exceptions so they also show stdout/stderr and not just the command that failed and the status code
Some other fixes/improvements to the Jenkinsfile that I saw on the fly
Hmm first attempt failed:
vm.winrm_upload(local_path='/Users/edward/Downloads/jre-8u144-windows-x64.exe',remote_path='c:\\jre-8u144-windows-x64.exe')
vcdriver_vm_winrm_username: Administrator
vcdriver_vm_winrm_password:
Copying "/Users/edward/Downloads/jre-8u144-windows-x64.exe" to "c:\jre-8u144-windows-x64.exe" ... [ ] 0 %Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "vcdriver/config.py", line 114, in wrapper
return function(*args, **kwargs)
File "vcdriver/vm.py", line 398, in winrm_upload
raise WinRmError(script, status_code)
vcdriver.exceptions.WinRmError: Remote execution of "add-content -value $([System.Convert]::FromBase64String("aOgaEAAAi0MoiYQkkAAAAEiNVCRoSI1MJDjo4YsAAJBIg7wkgAAAABByCkiLTCRo6Av7AABIibQkgAAAAEiJfCR4QIh8JGhIg7wkoAEAABByDUiLjCSIAQAA6OH6AABMjYQkYAEAAEiNlCTAAAAASI2MJJgAAADorGsAAJAzyf8Vq1wCAEiJtCQIAQAASIm8JAABAABAiLwk8AAAAEiJtCQ4AQAASIm8JDABAABAiLwkIAEAAEyLwLpkAAAASI2MJPAAAADo8qwAAJBIjZQkmAAAAEiNjCTwAAAA6JyxAACQSIO8JDgBAAAQcg1Ii4wkIAEAAOhH+gAASIm0JDgBAABIibwkMAEAAECIvCQgAQAASIO8JAgBAAAQcg1Ii4wk8AAAAOgX+gAASI2UJMAAAABIjUwkaOjRhgAAkEiLVCRwSItMJGhIK9FIuGdmZmZmZmZmSPfqSMH6BEiLwkjB6D9IA9C7AQAAAEg703URSI2UJJgAAADorcoAAITAdQNAit9IjUwkaOgEDQAAhNt0QEiNlCSYAAAASI1MJGjo+vf//5BIjVQkaEiNjCSYAAAA6Ef6//+JRCQwSIO8JIAAAAAQcklIi0wkaOhy+QAA6z26AwAAADkVvV4DAH8wSI0FfJwCAEiJRCQoSI0FoJwCAEiJRCQgRI1Kd0yNBeCaAgBIjQ2RXgMA6Izy//+QSIO8JLAAAAAQcg1Ii4wkmAAAAOgb+QAASIm0JLAAAABIibwkqAAAAECIvCSYAAAASI1MJDjo/YYAAJBIjUwkOOg+CgAAkEiDvCTYAAAAEHINSIuMJMAAAADo1fgAAEiJtCTYAAAASIm8JNAAAABAiLwkwAAAAEiDvCR4AQAAEHIOSIuMJGABAADopfgAAJCLRCQwSIuMJLgBAABIM8zowO4AAEyNnCTAAQAASYtbEEmLcxhJi+Nfw8zMzEiD7DhIjQ2KmQIA/xUvWgIASINkJCAATI1EJCBIjRV1mQIAuQIAAAD/FfpZAgCFwHQeSItMJCBIjRVqmQIA/xUcWgIASIXAdAe5AAgAAP/QSIPEOOkH/P//zMzMQFNIg+wgSIvZSIsJSIXJdAXo+vYAAEiDIwBIg8QgW8NIi0H4SGNQBEiNBUWcAgBIiUQK+MPMzMxIjQWVmwIASIkB6QniAADMSIlcJAhXSIPsIEiLeWhIjQWXmwIASIvZSIkBSIX/dBBIi8/oxOIAAEiLz+ik9wAASI1LCEiLXCQwSIPEIF/pgegAAMxIg8EI6ZfoAADMzMxIg8EI6ZPoAADMzMwzwMPMg8j/w0BTSIPsIEiLAUiL2f9QMIPK/zvCdBdIi0NYARBIi0tASIsRSI1CAUiJAQ+2EovCSA==")) -encoding byte -path c:\jre-8u144-windows-x64.exe" failed with exit code 1
| gharchive/pull-request | 2017-09-22T11:38:33 | 2025-04-01T04:32:56.556411 | {
"authors": [
"Lantero",
"edsharp"
],
"repo": "Osirium/vcdriver",
"url": "https://github.com/Osirium/vcdriver/pull/3",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
368288919 | Cannot connect to tm2 source
When using a tm2 source I get the following error (shortened stacktrace):
File "XXX/postile/postile.py", line 57, in setup_db
ConnectionRefusedError: [Errno 111] Connection refused
sys:1: RuntimeWarning: coroutine 'Loop.create_server' was never awaited
I use the following command: postile --cors --tm2 ../../tilelive-swift/openmaptiles.tm2source/data.yml
When I specify the db connection settings myself using postile --pgpassword ... it works.
@boldtrn sorry for the delay, the PostgreSQL connection parameters needs to be passed on command line, they are not read from the data.yml.
The tm2source is a format used mainly for the old Mapbox Studio Classic software and is only used partially here (like in OpenMapTiles project).
Thanks for the update :+1:
| gharchive/issue | 2018-10-09T16:13:10 | 2025-04-01T04:32:56.558867 | {
"authors": [
"boldtrn",
"ldgeo"
],
"repo": "Oslandia/postile",
"url": "https://github.com/Oslandia/postile/issues/6",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
420635771 | minor change to 'create_ensemble' utility to support mf_datasets
Please check if the PR fulfills these requirements
[ ] This PR addresses an already opened issue (for bug fixes / features)
[ ] Tests for the changes have been added (for bug fixes / features)
[ ] Docs have been added / updated (for bug fixes / features)
What kind of change does this PR introduce? (Bug fix, feature, docs update, etc.)
support mf_dataset when creating ensembles
Does this PR introduce a breaking change? (Has there been an API change?)
no
Pull Request Test Coverage Report for Build 577
8 of 12 (66.67%) changed or added relevant lines in 1 file are covered.
No unchanged relevant lines lost coverage.
Overall coverage decreased (-0.3%) to 86.364%
Changes Missing Coverage
Covered Lines
Changed/Added Lines
%
xclim/utils.py
8
12
66.67%
Totals
Change from base Build 559:
-0.3%
Covered Lines:
874
Relevant Lines:
1012
💛 - Coveralls
| gharchive/pull-request | 2019-03-13T17:47:27 | 2025-04-01T04:32:56.587791 | {
"authors": [
"coveralls",
"tlogan2000"
],
"repo": "Ouranosinc/xclim",
"url": "https://github.com/Ouranosinc/xclim/pull/175",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
103894116 | Fetch unconfirmed utxos as well
Can't see a reason why not. :-/
Thanks! Keep in mind it might take a few hours for the changes to update on the site due to cloudflares cache.
| gharchive/pull-request | 2015-08-30T01:52:23 | 2025-04-01T04:32:56.590375 | {
"authors": [
"OutCast3k",
"dabura667"
],
"repo": "OutCast3k/coinbin",
"url": "https://github.com/OutCast3k/coinbin/pull/25",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2128108382 | InternetStatus and InternetConnection class not found
Why after adding this library into our project i can't use InternetStatus and InternetConnection?
source:
https://github.com/OutdatedGuy/internet_connection_checker_plus/blob/main/example/lib/pages/listen_to_stream.dart
I think you have not installed the package properly, can you provide the snippet code for me to debug? Cuz I can make it work on every platform on my side.
@OutdatedGuy
You're mistaken. I installed the library and got an error. Then I realized that the documentation related to this library is outdated, and the names of classes and methods have changed, but they haven't been updated in the documentation.
Can you provide your code where you are trying this package? Also the version of the package you are using and the output of flutter doctor -v.
@OutdatedGuy
After installing the library i tested only the codes of implemented source code inside sample
[✓] Flutter (Channel stable, 3.16.5, on macOS 14.2.1 23C71 darwin-arm64, locale en-US)
• Flutter version 3.16.5 on channel stable at /Users/mahdipishguy/Desktop/home/develop/flutter
• Upstream repository https://github.com/flutter/flutter.git
• Framework revision 78666c8dc5 (8 weeks ago), 2023-12-19 16:14:14 -0800
• Engine revision 3f3e560236
• Dart version 3.2.3
• DevTools version 2.28.4
Can you provide detailed steps from start till the error cuz I am unable to find any issues.
Try to use InternetConnectionCheckerPlus in place of InternetConnection and InternetConnectionStatus in place of InternetStatus
@pishguy looks like you have installed a outdated version of this package. Try adding version ^2.0.0 or greater.
Closing as not an issue.
Feel free to comment if you want to re-open this issue.
| gharchive/issue | 2024-02-10T05:25:38 | 2025-04-01T04:32:56.598098 | {
"authors": [
"AbdNz",
"OutdatedGuy",
"pishguy"
],
"repo": "OutdatedGuy/internet_connection_checker_plus",
"url": "https://github.com/OutdatedGuy/internet_connection_checker_plus/issues/28",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1045833867 | Move MiniMap event handling code
Reference: #650, PR #1118
Move event handling code for mini map clicks from MapViewState to MiniMap.
Looks like there's a vcpkg error:
Computing installation plan...
Error: yasm-tool[core] is only supported on 'native & !uwp'
I'm guessing there's a bug in a recent vcpkg update.
| gharchive/pull-request | 2021-11-05T13:10:37 | 2025-04-01T04:32:56.599987 | {
"authors": [
"DanRStevens"
],
"repo": "OutpostUniverse/OPHD",
"url": "https://github.com/OutpostUniverse/OPHD/pull/1129",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2166947167 | Feature/sales
pull request de factories seeders tables and basic models and controllers
ok
| gharchive/pull-request | 2024-03-04T14:12:00 | 2025-04-01T04:32:56.614948 | {
"authors": [
"HelenaDR84",
"maicaocaa"
],
"repo": "P10MarketplaceDelisana/Delisana_Back",
"url": "https://github.com/P10MarketplaceDelisana/Delisana_Back/pull/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
251431285 | Compare region viewer: doesn't load for private genomes
During the Germany Workshop, several participants noted that the compare region viewer wasn't working for private genomes.
-Maulik
i have tried with my private genome and got error like below
this is fixed on alpha. will be deployed with other fixes soon.
working well on production now.
| gharchive/issue | 2017-08-19T15:19:41 | 2025-04-01T04:32:56.652992 | {
"authors": [
"hyoo",
"mshukla1"
],
"repo": "PATRIC3/patric3_website",
"url": "https://github.com/PATRIC3/patric3_website/issues/1664",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
657503804 | Retire PDX custom loaders
Description:
As a developer, i want to use the universal loader to ensure data loading via templates only.
XDog should be ran on all of the PDX Net data sets. Then each of these exports need to be validated and ran in the Updog to ensure that no data is being loss from the XDog. Additionally, the PDX Net data sets are very sparse and it needs to be ensured that the Updog with run the data sets without throwing runtime exceptions.
Yomi made the very excellent PDX-Transformer that will export both PDMR and JAX data sets which are on feeds. The PDMR exporter needs to be cleaned up and validated before using it. The Jackson Lab is in development of their own JAX to PDX Finder pipeline so it is unclear if it is necessary to spend any time cleaning the PDX Finder in-house tool.
Relevant ticket for PDMR: #296
Jira ticket PDXI-864
Acceptance criteria:
Data sets exported and validated for Updog:
[ ] PDXNet-HCI-BCM
[ ] PDXNet-WUSTL
[ ] PDXNet-Wistar-MDAnderson-Penn
[ ] PDXNet-MDAnderson
[ ] PDMR
[ ] JAX
[ ] Custom Loader code is retired on Dev
this will require some clean-up
execute after release
split into separate tickets
Export of HCI-BCM is blocked by #343
@zperova Done
Great work, @Afollet Thank you!
| gharchive/issue | 2020-07-15T16:55:10 | 2025-04-01T04:32:56.731908 | {
"authors": [
"Afollet",
"zperova"
],
"repo": "PDXFinder/pdxfinder",
"url": "https://github.com/PDXFinder/pdxfinder/issues/184",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
110850151 | SQL Error
Description
[Insert description of issue here]
Detailed Information
Is available here
Similar to #2193
Follow my instructions that I've posted on #2193 and reply here on your results.
Unfortunately PermissionsEx is currently unmaintained and no further updates are planned. Because of this, all pull requests are being closed. If you are looking for an actively maintained permissions plugin, take a look at LuckPerms. It is compatible with major server platforms and can import data from PEX.
| gharchive/issue | 2015-10-11T11:17:36 | 2025-04-01T04:32:56.741325 | {
"authors": [
"CzEMC",
"tteeee",
"zml2008"
],
"repo": "PEXPlugins/PermissionsEx",
"url": "https://github.com/PEXPlugins/PermissionsEx/issues/2185",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
190531944 | MYSQL Bug
Here:
http://minilebenmc.de/Bugs/pex/1.png
http://minilebenmc.de/Bugs/pex/2.png
Try deleting all the tables related to PEX (permissions, permissions_entity and permissions_inheritance). Let the plugin remake the tables
Unfortunately PermissionsEx is currently unmaintained and no further updates are planned. Because of this, all pull requests are being closed. If you are looking for an actively maintained permissions plugin, take a look at LuckPerms. It is compatible with major server platforms and can import data from PEX.
| gharchive/issue | 2016-11-20T01:11:28 | 2025-04-01T04:32:56.743529 | {
"authors": [
"LartyHD",
"RubeVi",
"zml2008"
],
"repo": "PEXPlugins/PermissionsEx",
"url": "https://github.com/PEXPlugins/PermissionsEx/issues/2586",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
766320169 | Behaviour of @x.setter and @x.deleter
In using a CleverDict instance to hold passwords (in conjunction with keyring) I noticed the setter and delete decorators didn't appear to be working:
@property
def password(self):
return keyring.get_password(account, self.username)
@password.setter
def password(self, value):
print("Setting!")
keyring.set_password(account, self.username, value)
@password.deleter
def password(self):
keyring.delete_password(account, username)
>>> value = "mysecret"
>>> keyring.set_password(CleverSession.choices[self.url], self.username, value)
>>> self.password
'mysecret'
# set
>>> self.password = "newpassword"
>>> self.password
'mysecret'
>>> self['password']
'newpassword'
# delete
>>> del self.password
>>> self.password
'mysecret'
>>> self['password']
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "C:\Users\Peter\AppData\Roaming\Python\Python39\site-packages\cleverdict\cleverdict.py", line 168, in __getitem__
name = self.get_key(name)
File "C:\Users\Peter\AppData\Roaming\Python\Python39\site-packages\cleverdict\cleverdict.py", line 310, in get_key
raise KeyError(name)
KeyError: 'password'
So I'm just wondering if this is something that has to be worked around, a limitation of CleverDict, or if there's something even more Clever we can do in CleverDict itself so that @x.setter and @x.deleter work as expected right out of the box?
It is tricky to try and combine properties with setattr and getattr and getattribute.
This stackoverflow article describes it quite well:
https://stackoverflow.com/questions/15750522/class-properties-and-setattr
Labelled as "wontfix" after Version 1.8.0 simply because the main authors don't currently have the time for non essential enhancements. If there's a huge uptake in people using CleverDict and asking for this feature that may change, or if you're reading this why not Fork, develop a solution, and contribute a Pull Request yourself?
Labelled as "wontfix" after Version 1.8.0 simply because the main authors don't currently have the time for non essential enhancements. If there's a huge uptake in people using CleverDict and asking for this feature that may change, or if you're reading this why not Fork, develop a solution, and contribute a Pull Request yourself?
import keyring
account = "PyPI"
class User:
def init(self):
self.username = "Pfython"
@property
def password(self):
return keyring.get_password(account, self.username)
@password.setter
def password(self, value):
keyring.set_password(account, self.username, value)
@password.deleter
def password(self):
keyring.delete_password(account, username)
| gharchive/issue | 2020-12-14T11:26:08 | 2025-04-01T04:32:56.754408 | {
"authors": [
"PFython",
"salabim"
],
"repo": "PFython/cleverdict",
"url": "https://github.com/PFython/cleverdict/issues/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
492263358 | Ravines generation doesn't turn off
OTG Plugin or Mod
[x] OTG Mod for Forge
[ ] OTG Plugin for Spigot
OTG Version
[ ] 1.12.2_v8_r1
[ ] 1.12.2_v7_DEV_r26
[x] 1.12.2-v6
[ ] 1.12.2-v5
[ ] 1.11.2-v8
[ ] 1.11.2-v9
[ ] 1.10.2-v21
[ ] Other - specify here
Forge or Spigot Versions
Forge 14.23.5.2814
What other plugins are you running?
Only decorative/technical mods:Optifine,BetterFoliage,JourneyMap,TellMe,TreeCapitator,NEI+JEI
Description
There were a lot of ravines everywhere, so I started changing values in WorldConfig.ini,however,it doesn't help.I had to set option RavineRarity: 0.0 in WorldConfig.ini, but it also doesn't work,ravines are still found in every biome,even very small,like beach,sometimes up to 3 in the same place,crossshaped.I want ravines to happen in the world,but not so often,and not in every biome.I didn't change cave settings,they are still default.
Is there any way to change ravines frequency?
If you need to look at WorldConfig.ini,I attached it(in txt format),but it is still work-in-progress and a big mess:)
WorldConfig.txt
Ravine rarity shouldn't have a decimal in it, try setting it to 0 instead.
Thank you very much! It helped ^_^ It was really stupid mistake,I hadn't even thought about decimal!
| gharchive/issue | 2019-09-11T14:10:19 | 2025-04-01T04:32:56.759741 | {
"authors": [
"Coll1234567",
"Darsenia"
],
"repo": "PG85/OpenTerrainGenerator",
"url": "https://github.com/PG85/OpenTerrainGenerator/issues/320",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
648471395 | The link to the Blitz XML is a dead link
On the page https://pgm.dev/docs/modules/objectives/blitz/ the link to the Blitz XML is a dead link, returns a GitHub 404 as that no longer exists. (https://maps.oc.tc/Blitz/blitz-global.xml)
Sorry for the wait. This has been fixed in https://github.com/PGMDev/Website/commit/ea04bbc0d0c837f534d07ee8121688137a3bf1d2.
| gharchive/issue | 2020-06-30T20:29:31 | 2025-04-01T04:32:56.761572 | {
"authors": [
"IcedCode",
"Indicardo"
],
"repo": "PGMDev/Website",
"url": "https://github.com/PGMDev/Website/issues/5",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2533603489 | PSR12 4.4 Methods and Functions close braces same line
Is your feature request related to a problem?
It is related to PHP8 and PSR12 standard
Describe the solution you'd like
As described in PSR12 item 4.4
If a function or method contains no statements or comments (such as an empty no-op implementation or when using constructor property promotion), then the body SHOULD be abbreviated as {} and placed on the same line as the previous symbol, separated by a space. For example:
class Point
{
public function __construct(private int $x, private int $y) {}
// ...
}
class Point
{
public function __construct(
public readonly int $x,
public readonly int $y,
) {}
}
Additional context (optional)
Exists some discussion and "suggestions" in this topic - https://github.com/squizlabs/PHP_CodeSniffer/issues/3291
[x] I have read the Contribution Guidelines and this is not a support question.
Don't know where you're getting your information from, but this is not part of PSR12.
This is the literal text of PSR12 section 4.4:
4.4 Methods and Functions
Visibility MUST be declared on all methods.
Method names MUST NOT be prefixed with a single underscore to indicate protected or private visibility. That is, an underscore prefix explicitly has no meaning.
Method and function names MUST NOT be declared with space after the method name. The opening brace MUST go on its own line, and the closing brace MUST go on the next line following the body. There MUST NOT be a space after the opening parenthesis, and there MUST NOT be a space before the closing parenthesis.
A method declaration looks like the following. Note the placement of parentheses, commas, spaces, and braces:
... code sample...
https://www.php-fig.org/psr/psr-12/#44-methods-and-functions
I took it from - https://www.php-fig.org/per/coding-style/ that can be accessed in the page https://www.php-fig.org/psr/psr-12/ and click on the link "PER Coding Style" on the right top menu (Additional info)
@junior-paytech PSR-PER is not the same as PSR12 and for those things for which PSR12 has its own rules, PSR-12 sniffs cannot be updated to follow PSR-PER as in that case, the PSR 12 sniff would no longer enforce PSR12.
I understand your point, but I believe that the PER is more "actual" and aligned with the current php development.
Anyway, thank you your support.
| gharchive/issue | 2024-09-18T12:26:30 | 2025-04-01T04:32:56.791457 | {
"authors": [
"jrfnl",
"junior-paytech"
],
"repo": "PHPCSStandards/PHP_CodeSniffer",
"url": "https://github.com/PHPCSStandards/PHP_CodeSniffer/issues/618",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2036195110 | Explaination on The YoloX Body-Head-Hand output dimension?
Issue Type
Documentation Feature Request
OS
Other
OS architecture
armv7
Programming Language
Other
Framework
TensorFlowLite
Model name and Weights/Checkpoints URL
YoloX Body-Head-Hand
https://github.com/PINTO0309/PINTO_model_zoo/tree/main/426_YOLOX-Body-Head-Hand
Description
First of all thanks and congratulation for your excellent work, on YOLO.
I have a question regarding the output size, though.
@PINTO0309 You say you limit the output to 20 boxes : How come we obtain a float32[60,7] output then?
shouldn't we get something like 20x [class, score, x1, y1, x2, y2] = 20 x 5 floats?
Relevant Log Output
No response
URL or source code for simple inference testing code
No response
As detailed in the README.
max output boxes per class
20 (Body) + 20 (Head) + 20 (Hand) = 60 boxes
| gharchive/issue | 2023-12-11T17:41:51 | 2025-04-01T04:32:56.912400 | {
"authors": [
"PINTO0309",
"kho-bluefrogrobotics"
],
"repo": "PINTO0309/PINTO_model_zoo",
"url": "https://github.com/PINTO0309/PINTO_model_zoo/issues/381",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
561056495 | [IIS] WinAcme stuck on "[DBUG] Scanning IIS sites"
Bug Description
Win Acme stuck on "[DBUG] Scanning IIS sites", won't renew it through task neither through the client.
Expected behavior
Win Acme should automatically renew certificates, as it did in the past.
Log
[VERB] Verbose mode logging enabled
[VERB] Looking for settings.json in C:\win-acme
[DBUG] Config folder: C:\ProgramData\win-acme\acme-v02.api.letsencrypt.org
[DBUG] Log path: C:\ProgramData\win-acme\acme-v02.api.letsencrypt.org\Log
[DBUG] Cache path: C:\ProgramData\win-acme\acme-v02.api.letsencrypt.org\Certificates
[INFO] A simple Windows ACMEv2 client (WACS)
[INFO] Software version 2.1.0.539 (RELEASE, PLUGGABLE)
[INFO] IIS version 10.0
[INFO] Running with administrator credentials
[INFO] Scheduled task looks healthy
[INFO] Please report issues at https://github.com/PKISharp/win-acme
[VERB] Test for international support: 語言 язык لغة
[VERB] Verbose mode logging enabled
[VERB] Arguments: --verbose
[DBUG] Renewal period: 55 days
N: Create new certificate (simple for IIS)
M: Create new certificate (full options)
L: List scheduled renewals
R: Renew scheduled
S: Renew specific
A: Renew all
O: More options...
Q: Quit
Please choose from the menu: s
1: [IISSite] suporte - renewed 1 time, due now
2: [IISSite] autenticador - renewed 2 times, due now
3: [IISSite] suporte_homologacao - renewed 3 times, due now
: Back
Which renewal would you like to run?: 1
[VERB] Sending e-mails True
[DBUG] Scanning IIS sites
Platform:
OS: Windows Server 2019, English
Version: 2.1.0.539 (RELEASE, PLUGGABLE)
Additional context
It's not generating errors in the Event Viewer or the win acme logs either. I'm not experienced in server administration so I'm kinda lost on where to look to try and debug this issue.
Setting the --test flag also leaves me stuck with the same issue.
I already tried leaving it on for hours and it simply won't do a thing.
Do you have a lot of sites and/or a lot of bindings? I've got no idea why it would hang at that point without any feedback, unless IIS is unhealthy for some reason, but I assume all your sites are up and running?
As a first step I would advise you to try the latest release. If that doesn't work then I may need to ask for a copy of your applicationHost.config to debug the issue.
I upgraded to the latest version and got a little further on the process. It did scan my websites but the requests were not responding to the letsencrypt servers.
I ended up fixing the issue by rebooting my machine, which is hosted at AWS.
Thanks for the quick response!
| gharchive/issue | 2020-02-06T14:41:32 | 2025-04-01T04:32:56.933287 | {
"authors": [
"Incognitowski",
"WouterTinus"
],
"repo": "PKISharp/win-acme",
"url": "https://github.com/PKISharp/win-acme/issues/1401",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
408013074 | Update YMap.js
Добавил дополнительную опцию для возможности более подробной отладки ошибок.
Спасибо!
| gharchive/pull-request | 2019-02-08T05:37:01 | 2025-04-01T04:32:56.949112 | {
"authors": [
"PNKBizz",
"VadimMalykhin"
],
"repo": "PNKBizz/vue-yandex-map",
"url": "https://github.com/PNKBizz/vue-yandex-map/pull/114",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1193820829 | Add in AML specific signatures
AML does not have standard deconvolution, instead there is a 17 gene stemness signature and some other subtypes
https://pubmed.ncbi.nlm.nih.gov/27926740/
and
https://www.sciencedirect.com/science/article/pii/S0092867419300947
Creating an issue to add these signatures to a single file and 'signature_matrices' code so we can deconvolve them./
This could be closed, but i'm still getting an xcell issue.
| gharchive/issue | 2022-04-05T23:00:11 | 2025-04-01T04:32:56.950951 | {
"authors": [
"sgosline"
],
"repo": "PNNL-CompBio/decomprolute",
"url": "https://github.com/PNNL-CompBio/decomprolute/issues/186",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1748088919 | Potential deprecation of concolor
I'm posting this as this project is reported as a dependent of concolor and wanted to both give you a heads up to prepare and an opportunity to discuss in case we should re-evaluate this.
We are proposing deprecating concolor in favor of either
anstream
application-specific solutions built on colorchoice, colorchoice-clap, and your choice of environment detection crates including anstyle-query and supports-color.
We welcome feedback on this topic at rust-cli/concolor#47
Thank you! I commented on the linked issue.
| gharchive/issue | 2023-06-08T15:00:42 | 2025-04-01T04:32:57.107368 | {
"authors": [
"epage",
"max-sixty"
],
"repo": "PRQL/prql",
"url": "https://github.com/PRQL/prql/issues/2765",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1610173228 | feat: upper and lower function
Fixes #1793
This is great! Even the func interp now uses better naming (in my opinion)!
Agree! Thanks @Jelenkee !
| gharchive/pull-request | 2023-03-05T12:03:56 | 2025-04-01T04:32:57.108586 | {
"authors": [
"Jelenkee",
"aljazerzen",
"max-sixty"
],
"repo": "PRQL/prql",
"url": "https://github.com/PRQL/prql/pull/2019",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
608621121 | Different endpoints to verify token/logout
@ChetanKhanna Please merge similar end points for guest/verified user.
This is done in 8e05d594f0dc497951262255aef17f4bccfb9eba
| gharchive/issue | 2020-04-28T21:15:32 | 2025-04-01T04:32:57.109399 | {
"authors": [
"ChetanKhanna",
"ritiktaneja"
],
"repo": "PS-Division-BITS/Chat",
"url": "https://github.com/PS-Division-BITS/Chat/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2673206909 | k-space preconditioning for pdhg
Accelerating Non-Cartesian MRI Reconstruction Convergence using k-space Preconditioning by Frank Ong, Martin Uecker, and Michael Lustig arxiv suggests to use pdhg with k-space preconditoning.
For a diagonal precondition as proposed, this would in practice resulting the dual stepsize sigma being a tensor in our implementation.
See also how to do their experiments https://github.com/mikgroup/kspace_precond/blob/master/Sense Reconstruction.ipynb -- they set sigpy's pdhg stepsize to the precondition
We would also need their methodof obtaining the single- or multi-channel preconditioner.
Accelerating Non-Cartesian MRI Reconstruction Convergence using k-space Preconditioning by Frank Ong, Martin Uecker, and Michael Lustig arxiv suggests to use pdhg with k-space preconditoning.
For a diagonal precondition as proposed, this would in practice resulting the dual stepsize sigma being a tensor in our implementation. See also how to do their experiments https://github.com/mikgroup/kspace_precond/blob/master/Sense Reconstruction.ipynb -- they set sigpy's pdhg stepsize to the precondition
We would also need their methodof obtaining the single- or multi-channel preconditioner.
I agree that would be of high interest... for example, for our TV-unrolling, it would be nice to be able to only use a few iterations of PDHG and already have converged to a decent solution. This would make the approach applicable to non-Cartesian MRI as well, which we on purpose did not consider yet because of those reasons.
| gharchive/issue | 2024-11-19T18:55:34 | 2025-04-01T04:32:57.129597 | {
"authors": [
"fzimmermann89",
"koflera"
],
"repo": "PTB-MR/mrpro",
"url": "https://github.com/PTB-MR/mrpro/issues/545",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1228727633 | Cambiar_rol/hay cambios en login(actualiza)
pequeños cambios en el login(puede ser) y opcion de cambiar rol del trabajador en pagina de trabajadores donde solo el admin tendra acceso.
No entiendo muy bien este cambio, ya tenemos una manera de modificar los datos del trabajador cuando accedamos a el. Los cambios que debías hacer eran en la pagina del trabajador, comprobando el rol del usuario que quiere cambiar la info.
No se a que tarjeta del trello hace referencia esto pero no me suena
| gharchive/pull-request | 2022-05-07T21:31:36 | 2025-04-01T04:32:57.130892 | {
"authors": [
"Ar1e1",
"alfman99"
],
"repo": "PTIN2022/GeSyS-Front",
"url": "https://github.com/PTIN2022/GeSyS-Front/pull/44",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
212351903 | How to remove lsm303d_accel driver
I have a problem with my pixhawk.
My pixhawk lsm303d_accel sensor is dead.
So I try to remove lsm303d accel driver .
But I fail to remove.
Can you help me?
How to remove lsm303d accel driver?
We will not help support broken hardware setups.
@winszenYoung It will automatically not start if its broken. However, if the sensor is not really dead but half-working, you should consider retiring this board. Its not safe to fly.
| gharchive/issue | 2017-03-07T07:29:22 | 2025-04-01T04:32:57.154618 | {
"authors": [
"LorenzMeier",
"mhkabir",
"winszenYoung"
],
"repo": "PX4/Firmware",
"url": "https://github.com/PX4/Firmware/issues/6734",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
682641205 | Fix Mavros Mission tests for VTOL
Describe problem solved by this pull request
Recently Mavros Mission Tests have been failing for VTOLs in Firmware.
Describe your solution
Update the container from ROS melodic to ROS noetic to run the tests as the same environment as the rest of the tests
Failure:
[ERROR] [1597924035.046962627]: [registerPublisher] Failed to contact master at [7f9311d946d5:44367]. Retrying...
Traceback (most recent call last):
File "/opt/ros/noetic/lib/gazebo_ros/spawn_model", line 20, in <module>
import rospy
File "/opt/ros/noetic/lib/python3/dist-packages/rospy/__init__.py", line 49, in <module>
from .client import spin, myargv, init_node, \
File "/opt/ros/noetic/lib/python3/dist-packages/rospy/client.py", line 52, in <module>
import roslib
File "<frozen importlib._bootstrap>", line 991, in _find_and_load
File "<frozen importlib._bootstrap>", line 971, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 906, in _find_spec
KeyboardInterrupt
@TSC21 Updating to noetic might not be a good idea, since we don't have catkin on focal working yet
any progress on this end @Jaeyoung-Lim ?
| gharchive/pull-request | 2020-08-20T11:31:53 | 2025-04-01T04:32:57.157016 | {
"authors": [
"Jaeyoung-Lim",
"mrpollo"
],
"repo": "PX4/Firmware",
"url": "https://github.com/PX4/Firmware/pull/15587",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
120186583 | Make posix sitl work on my mac
@LorenzMeier please double check my PX4_ISFINITE changes and va_arg type change. I'm not sure if the va_arg change breaks it for others.
@ SITL-CI test this please
This is all looking good.
| gharchive/pull-request | 2015-12-03T14:44:32 | 2025-04-01T04:32:57.158268 | {
"authors": [
"AndreasAntener",
"LorenzMeier"
],
"repo": "PX4/Firmware",
"url": "https://github.com/PX4/Firmware/pull/3313",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
203459286 | Remove the notion of legacy from the api and save some FLASH space too
@LorenzMeier
This remove the notion of legacy from the api. The board level code will perform the translation to legacy format on the STM32. New targets will not need to do this as there is no case where the serial number were used by mfg for tracking.
It maintains the legacy incorrect selection of significant digits of the UUID in the PX4 code base.
This is done to avoid the ripple effects changing the IDs used on STM32 base platforms. [See] (https://github.com/PX4/Firmware/commit/f126668dbd41f10667a8759be7937dd822549828#diff-8da0cce1d272772f52e0d92c6f4343a5R83)
Jenkins test this please.
From your comment I gather you only kept the display of legacy. If that conclusion would be correct: However, the byte order is also inverted in storage IIRC and via MAVLink. Please compare the UUID as sent via MAVLInk in the MAVLink inspector of QGroundControl. It should stay the same for STM32 as it will blow up existing configs.
If you kept the representation for STM32 then this looks like its ready to go.
Jenkins test this please.
1.5.0 Release RC1
nsh> ver all
HW arch: PX4FMU_V2
FW git-hash: 3d7906647e5c6f382bf6c574927bac88d752874f
FW version: v1.4.4rc1 (1.4.4 192), 17040576
OS version: 6.27 (4294967295)
Build datetime: Aug 26 2016 09:41:20
Toolchain: 4.8.4 20140725 (release) [ARM/embedded-4_8-branch revision 213147]
MCU: STM32F42x, rev. 3
UID: 400038:33355110:32383336
nsh> fmu id
ERROR [fmu] Board serial:
00400038 33355110 32383336
Current master
nsh> ver all
HW arch: PX4FMU_V2
FW git-hash: 264589b2cc8bc69e95fc63580e2ad94be109c016
FW version: 1.5.2 0 (17105408)
OS: NuttX
OS version: Release 1.8.0 (17301759)
Build datetime: Jan 28 2017 05:19:48
Build uri: localhost
Toolchain: GNU GCC, 5.4.1 20160609 (release) [ARM/embedded-5-branch revision 237715]
MCU: STM32F42x, rev. 3
UID: 400038:33355110:32383336
nsh> fmu id
Board serial:
400038 33355110 32383336
nsh>
This Branch
nsh> ver all
HW arch: PX4FMU_V2
FW git-hash: b1b8e9b87a96b0b032237e31bb91f4e182a86d69
FW version: 1.5.2 0 (17105408)
OS: NuttX
OS version: Release 1.8.0 (17301759)
Build datetime: Jan 28 2017 05:49:58
Build uri: localhost
Toolchain: GNU GCC, 5.4.1 20160609 (release) [ARM/embedded-5-branch revision 237715]
MCU: STM32F42x, rev. 3
UID: 400038:33355110:32383336
nsh> fmu id
Board serial:
00400038 33355110 32383336
nsh> WARN [commander] Not ready to fly: Sensors not set up correctly
@LorenzMeier - Please DNM yet - I have more commits to add here....
This branch post discussion
nsh> ver all
HW arch: PX4FMU_V2
FW git-hash: 256851bc27a9445e5fb24e5a0ec3516cc34d37d6
FW version: 1.5.2 0 (17105408)
OS: NuttX
OS version: Release 1.8.0 (17301759)
Build datetime: Jan 28 2017 08:20:45
Build uri: localhost
Toolchain: GNU GCC, 5.4.1 20160609 (release) [ARM/embedded-5-branch revision 237715]
MFGUID: 323833363335511000400038
MCU: STM32F42x, rev. 3
UID: 400038:33355110:32383336
nsh> fmu id
FMU: unrecognized command id, try:
mode_gpio, mode_pwm, mode_pwm4, test, sensor_reset [milliseconds], i2c <bus> <hz>, bind
@LorenzMeier - ready!
CI catched some code issues:
/Users/travis/build/PX4/Firmware/src/systemcmds/ver/ver.c:60:19: fatal error: variable 'mfg_uid_str' is not needed and will not be emitted [-Wunneeded-internal-declaration]
static const char mfg_uid_str[] = "mfguid";
^
1 error generated.
@LorenzMeier
Now that we added the mfguid and display in in ver I cleaned up the uavcan usage and removed all the exposed reordering from the API and include files. The stm32 board code only needs to do the maintain the legacy format.
I have tested this on px4fmu-v3 and a new 2.4.6 board w/ rev 3 silicon as well as the esc35-v1 hw bootloader.
Node id allocation is AOK with these changes.
nsh> ver all
HW arch: PX4FMU_V2
FW git-hash: 3fe5c4f2372298bf1c44fa655e344d632a31afcd
FW version: 1.6.0 0 (17170432)
OS: NuttX
OS version: Release 1.8.0 (17301759)
Build datetime: Jan 30 2017 13:23:35
Build uri: localhost
Toolchain: GNU GCC, 5.4.1 20160609 (release) [ARM/embedded-5-branch revision 237715]
MFGUID: 323833363335511000400038
MCU: STM32F42x, rev. 3
UID: 400038:33355110:32383336
nsh> ver uid
UID: 400038:33355110:32383336
nsh> ver mfguid
MFGUID: 323833363335511000400038
@LorenzMeier Please give this a second look.
@LorenzMeier Architectural question. Do we want to pad out a 96 bit mfg to 128 bit with leading 0's.
The we can use 32 character string in the logging and add the to QGC as an mfguid[32]?
Rebased and force-pushed.
@davids5 Padding is fine.
I still will merge this once it passes.
| gharchive/pull-request | 2017-01-26T19:04:31 | 2025-04-01T04:32:57.168270 | {
"authors": [
"LorenzMeier",
"davids5"
],
"repo": "PX4/Firmware",
"url": "https://github.com/PX4/Firmware/pull/6441",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
506470253 | Enhancement: Increased robustness against 0 signal quality and cleanup range data handling
This PR attempts to cleanup the checks on the range finder data. Additionally it adds a hysteresis to declaring range finder data valid after signal quality was zero.
Specifically:
whenever signal quality of the range data was 0 we require non-zero signal quality for 3 seconds before we accept the data to be used. This is to improve robustness against a known sensor failure mode where the signal quality toggles between 0 and non-zero values (see plot below)
moved range finder data checks to a central location and created new file range_finder_checks.cpp in order to avoid loading control.cpp
replaced _rng_hgt_faulty with _range_hgt_valid in order to avoid negation all over the place
use better function name
moved updateRangeDataContinuity() out of the terrain estimator and into range_finder_checks.cpp
I will provide SITL test data shortly.
Plot of range finder failure mode:
Notice how signal quality erratically jumps between 0 and 95% in the middle of the flight. The height reported by the sensor at that point was completely wrong.
Signed-off-by: RomanBapst bapstroman@gmail.com
The 3 seconds hold-off sounds like it is engineered for a specific sensor - can this be made a parameter and be set for that sensor?
Plot demonstrating 3 second hysteresis after signal quality was 0:
Log: https://mail.google.com/mail/u/0/#inbox/FMfcgxwDrlVmwVjrGGZSjKnTPrbflWXr
Plot demonstrating a height reset to range finder (simulated a step in range finder data of 5m):
Log: https://logs.px4.io/plot_app?log=d3182c9d-01e2-4d15-940f-00b7d2190af2
@MaEtUgR I addressed your comments, they were good suggestions. Please have another look.
@bresch Sorry I meant you, not @MaEtUgR
@priseborough Could you please have a look as well?
Thanks @RomanBapst
| gharchive/pull-request | 2019-10-14T06:30:24 | 2025-04-01T04:32:57.195960 | {
"authors": [
"LorenzMeier",
"RomanBapst",
"priseborough"
],
"repo": "PX4/ecl",
"url": "https://github.com/PX4/ecl/pull/655",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1302120674 | Can "integrated charts" be used with streamlit-aggrid?
I can see from the showcase example that e.g. groups and pivot-tables work. Before buying a license, I would like to know if the enterprise "inegrated charts" feature will work when AGGrid is used via Streamlit.
same request here with Tree Data feature of aggrid enterprise
Same here. Any plans to support live streaming updates, and integrated charting. This will be amazing.
I hope there is a way for you to give us a referral link where you get something whenever one of us buys an enterprise license from ag-grid.
Same here. Any plans to support live streaming updates, and integrated charting. This will be amazing.
I hope there is a way for you to give us a referral link where you get something whenever one of us buys an enterprise license from ag-grid.
I've solved this problem, later this week I'll post my code here.
Thanks @littlepanda0716 and look forward to your solution.
`import streamlit as st
from st_aggrid import AgGrid, GridOptionsBuilder, JsCode
import pandas as pd
if name == "main":
df = pd.DataFrame({"Name": ["A", "B", "C", "D"], "Age": ["80.0%", "0.0%", "20.0%", "None"]})
st.write(df)
gb = GridOptionsBuilder.from_dataframe(df)
gridOptions = gb.build()
gridOptions["columnDefs"][1] = {"field": "Age",
"cellRenderer": "agSparklineCellRenderer",
'valueGetter': 'JSON.parse("["+data.Age.replace("%","]"))',
"cellRendererParams": {
"sparklineOptions": {
"type": 'bar',
"fill": '#4fa2d9',
"valueAxisDomain": [0.0, 100.0],
"label": {
"enabled": True,
"color": 'black',
"fontSize": 10,
"fontWeight": 'bold',
"formatter": JsCode("""function(params){
return `${params.value}%`;}"""),
"placement": "insideBase",
},
"tooltip": {
"enabled": False,
},
}
}
}
r = AgGrid(
df,
gridOptions=gridOptions,
height=500,
allow_unsafe_jscode=True,
enable_enterprise_modules=True,
filter=True,
# update_mode=GridUpdateMode.SELECTION_CHANGED,
theme="material",
)`
import streamlit as st
from st_aggrid import AgGrid, GridOptionsBuilder, JsCode
import pandas as pd
if __name__ == "__main__":
df = pd.DataFrame({"Name": ["A", "B", "C", "D"], "Age": ["80.0%", "0.0%", "20.0%", "None"]})
st.write(df)
gb = GridOptionsBuilder.from_dataframe(df)
gridOptions = gb.build()
gridOptions["columnDefs"][1] = {"field": "Age",
"cellRenderer": "agSparklineCellRenderer",
'valueGetter': 'JSON.parse("["+data.Age.replace("%","]"))',
"cellRendererParams": {
"sparklineOptions": {
"type": 'bar',
"fill": '#4fa2d9',
"valueAxisDomain": [0.0, 100.0],
"label": {
"enabled": True,
"color": 'black',
"fontSize": 10,
"fontWeight": 'bold',
"formatter": JsCode("""function(params){
return `${params.value}%`;}"""),
"placement": "insideBase",
},
"tooltip": {
"enabled": False,
},
}
}
}
r = AgGrid(
df,
gridOptions=gridOptions,
height=500,
allow_unsafe_jscode=True,
enable_enterprise_modules=True,
filter=True,
# update_mode=GridUpdateMode.SELECTION_CHANGED,
theme="material",
)
Hope this may help you @mzeidhassan
Thanks a million @littlepanda0716 ! This is very nice of you. I appreciate it.
| gharchive/issue | 2022-07-12T14:08:55 | 2025-04-01T04:32:57.204607 | {
"authors": [
"fluence-world",
"littlepanda0716",
"mzeidhassan"
],
"repo": "PablocFonseca/streamlit-aggrid",
"url": "https://github.com/PablocFonseca/streamlit-aggrid/issues/110",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
308887058 | unzip uses too much memories
Dear developer,
We found that unzip module 'rr_ctg_track.py' try to read all .las files into memory and we had around 20T .las files. It's hard to find a machine with so large memories. Do you have any suggestion to avoid loading all data into memories?
3ku!
This issue is created by me and is the same problem. You can get more information from it (https://github.com/marbl/canu/issues/838).
The memory is probably not consumed by rr_ctg_track.py directly. That program spawns LA4Falcon for each .las file. You will have a number of instances of LA4Falcon running equal to your --n-core argument. (And each will be under a different sub-process, so the memory used by rr_ctg_track will be cloned. That's probably not an problem, but you can look at the forked python procs on your machine.)
Each LA4Falcon loads the entire DAZZLER DB, which is probably your problem. (Look at the file 0-rawreads/.raw_reads.bps) There are 2 solutions:
Hack our code to load the DB from dev/shm. (Non-trivial, but one user has done this.)
Use --n-core=0. (Same as 1, but simpler, since it avoids the whole "multiprocessing" module.)
You can experiment with various values of --n-core.
Also, your unzip might be out-of-date. You could try the Falcon-unzip binary tarball, as the GitHub code is not up-to-date.
https://github.com/PacificBiosciences/FALCON_unzip/wiki/Binaries
The explaination is really clear and I believe the solution you provided must be useful.
I'll try as you said. 3ku.
| gharchive/issue | 2018-03-27T09:19:41 | 2025-04-01T04:32:57.210133 | {
"authors": [
"nottwy",
"pb-cdunn"
],
"repo": "PacificBiosciences/FALCON_unzip",
"url": "https://github.com/PacificBiosciences/FALCON_unzip/issues/111",
"license": "BSD-3-Clause-Clear",
"license_type": "permissive",
"license_source": "github-api"
} |
101098721 | more pbsmrtpipe integration
The most important part (which I forgot in the commit messages) is using ZMW ranges for chunked datasets.
Hold on, I may need that index file generation after all...
Can we avoid outputting that index? By restricting the number of interdependencies and separating functionally distinct components software is much easier to write, read, debug and maintain.
Yeah, we really need a central, general solution for this problem.
And I think I figured out the general solution, using a decorator defined by pbcommnad
should I go ahead and merge then?
yes, the ZMW range is critical
| gharchive/pull-request | 2015-08-14T20:35:45 | 2025-04-01T04:32:57.212279 | {
"authors": [
"evolvedmicrobe",
"natechols"
],
"repo": "PacificBiosciences/pbccs",
"url": "https://github.com/PacificBiosciences/pbccs/pull/53",
"license": "BSD-3-Clause-Clear",
"license_type": "permissive",
"license_source": "github-api"
} |
307222428 | getting the loss of a audio file with a certain transcription
Any chance to get a guidance on hoe to receive the ctc_loss of a (audio,transcrption) pair via the python wrapper?
Thanks in advance!
Please refer to this method
https://github.com/PaddlePaddle/DeepSpeech/blob/dad1c2727e375beb05dd747b06b04d0438e4a78d/model_utils/model.py#L158-L174
It tried it already...do you have any documentation or example how to use
this function..
any combination I tried failed, like: infer_data=[(speech,txt)] or
[speech,txt] or (speech,txt) with different errors or dumps..
On Thu, Mar 22, 2018 at 8:53 AM, Yibing Liu notifications@github.com
wrote:
Please refer to this method
https://github.com/PaddlePaddle/DeepSpeech/blob/
dad1c2727e375beb05dd747b06b04d0438e4a78d/model_utils/model.py#L158-L174
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/PaddlePaddle/DeepSpeech/issues/186#issuecomment-375197523,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AE_DSVEVfIXKrCSyea5aEHzgDXT3m_Ouks5tg0pbgaJpZM4SzidE
.
@alanbekker Have you figured it out ? I incurred the same problem, looking forward your reply, thanks.
I have figured it out , just set the DataGenerator's keep_transcription_text to False and make the same preprocess as in infer_batch_probs():
data_generator = DataGenerator(
vocab_filepath=args.vocab_path,
mean_std_filepath=args.mean_std_path,
augmentation_config='{}',
specgram_type=args.specgram_type,
num_threads=8,
keep_transcription_text=False)
| gharchive/issue | 2018-03-21T12:21:17 | 2025-04-01T04:32:57.236629 | {
"authors": [
"alanbekker",
"hui001",
"kuke"
],
"repo": "PaddlePaddle/DeepSpeech",
"url": "https://github.com/PaddlePaddle/DeepSpeech/issues/186",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1182271042 | 模型预测或模型推理的结果,没有保存的TXT文件。
PaddleClas/tools/infer.py
PaddleCals/deploy/python/predict_cls.py
预测或推理路径下的所有图像,结果没有保存的txt文件。
目前尚未提供该功能,如有需要,可以修改源码,或是在执行推理命令时将结果重定向到指定文件中,再对结果进行处理:
python python/predict_cls.py -c configs/inference_cls.yaml > output.txt
Example
import ast
import re
with open("output.txt") as f:
ftxt = f.read()
pattern = r"{'class_ids':.*?}"
pre_results = re.findall(pattern, ftxt)
print(len(pre_results))
pre_data = []
for pre_result in pre_results:
pre_dict = ast.literal_eval(pre_result) # ["class_ids", "scores", "file_name", "label_names"]
pre_data.append([pre_dict["file_name"].split("/")[-1],\
pre_dict["class_ids"][0], pre_dict["scores"][0]])
columns = ["img_path", "label", "scores"]
pre_df = pd.DataFrame(data=pre_data,
index=range(len(pre_data)),
columns=columns)
pre_df.drop(columns=['scores'],inplace=True) # 删除 scores 列
pre_df.sort_values(["img_path"], inplace=True) # 排序
pre_df.to_csv('submission.csv',index=None,header=None)
| gharchive/issue | 2022-03-27T02:36:03 | 2025-04-01T04:32:57.300225 | {
"authors": [
"CPones",
"TingquanGao",
"Wulx2050"
],
"repo": "PaddlePaddle/PaddleClas",
"url": "https://github.com/PaddlePaddle/PaddleClas/issues/1788",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2173033299 | [Question]: UIE信息抽取做 关系抽取任务,出现幻想,文本中没有出现的信息,会出现幻觉
请提出你的问题
调用的是paddlepaddle uie-base 模型进行推理,配置关系如下,及其结果如下,有好的解决方案吗?
这个是模型效果的问题,需要通过UIE微调方式来提升效果。
微调 通过增加负样例 依然还是有这样的问题,不觉得是负样例的问题,猜测可能和uie本身预测的结构有关系
您好,UIE相关的技术已经停止更新了,欢迎使用LLM来解决您的业务问题
| gharchive/issue | 2024-03-07T06:10:32 | 2025-04-01T04:32:57.321165 | {
"authors": [
"827648313",
"w5688414",
"wawltor"
],
"repo": "PaddlePaddle/PaddleNLP",
"url": "https://github.com/PaddlePaddle/PaddleNLP/issues/8068",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1758099051 | [Performance optimization] add to_static feature for all generative model
PR types
Performance optimization
PR changes
APIs
Description
完善 generation_utils 中的生成相关逻辑。
[x] 消除静态图子图
[x] 优化生成性能
[x] 添加动转静相关单测
可以在下一个 pr 中将 动转静的相关逻辑添加到基类当中去,这样其他模型也可以快速添加对应生成配置。
| gharchive/pull-request | 2023-06-15T06:01:59 | 2025-04-01T04:32:57.323657 | {
"authors": [
"wj-Mcat"
],
"repo": "PaddlePaddle/PaddleNLP",
"url": "https://github.com/PaddlePaddle/PaddleNLP/pull/6186",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2262465511 | add Pose and YaRN
PR types
New features
PR changes
Models
Description
Added the Pose algorithm, including its data processing and training scripts.
Added YaRN, one of the RoPE extension methods.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.whf seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
cd PaddleNLP
pre-commit install
commit之前执行,自动修改格式问题
| gharchive/pull-request | 2024-04-25T01:37:06 | 2025-04-01T04:32:57.327730 | {
"authors": [
"CLAassistant",
"lugimzzz",
"whf313"
],
"repo": "PaddlePaddle/PaddleNLP",
"url": "https://github.com/PaddlePaddle/PaddleNLP/pull/8327",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1039609131 | The third-party dynamic library (mklml.dll) that Paddle depends on is not configured
环境
请提供您使用的PaddlePaddle和PaddleX的版本号
paddlepaddle-gpu 2.1.3.post112
paddleslim 2.1.1
paddlex 2.0.0
请提供您使用的操作系统信息,如Linux/Windows/MacOS
WIN10
请问您使用的Python版本是?
python 3.7
请问您使用的CUDA/cuDNN的版本号是?
cuda11.2 cudnn8
2021-10-29 19:06:26 [INFO] Model[FasterRCNN] loaded.
Exception in thread CAM_BENDI_00:
Traceback (most recent call last):
File "C:\Users\hello\threading.py", line 932, in _bootstrap_inner
File "C:\Users\hello\threading.py", line 870, in run
File "C:\Users\hello\hello_class.py", line 4999, in Work_threadai_0
File "C:\Users\hello\paddlex\deploy.py", line 64, in __init__
self.predictor = self.create_predictor(
File "C:\Users\hello\paddlex\deploy.py", line 126, in create_predictor
predictor = create_predictor(config)
RuntimeError: (PreconditionNotMet) The third-party dynamic library (mklml.dll) that Paddle depends on is not configured correctly. (error code is 182)
Suggestions:
1. Check if the third-party dynamic library (e.g. CUDA, CUDNN) is installed correctly and its version is matched with paddlepaddle you installed.
2. Configure third-party dynamic library environment variables as follows:
- Linux: set LD_LIBRARY_PATH by `export LD_LIBRARY_PATH=...`
- Windows: set PATH by `set PATH=XXX; (at C:\home\workspace\Paddle_release3\paddle\fluid\platform\dynload\dynamic_loader.cc:265)
用户安装了线描相机的软件和驱动之后 paddle就报错
python环境是这样的问题
之前打包好独立的EXE环境 也是一样的错误
conda新建环境不行
重新安装cuda+cudnn 也不行
@monkeycc 你好,请问你这个问题解决了么? 我是使用multiprocessing模块时,遇到了这个报错。。用threading模块就不会。。感觉很迷糊。。---我是用paddlepaddle库做文本检测的,没有用到paddleX...我只是想请教下,你是怎么解决这个问题的。。
常规情况下,我使用GPU是没有报错的。。
@lmw0320
直接把线描相机的软件和驱动 删除就好了
CPU加速的模块 有冲突
multiprocessing模块时,遇到了这个报错。。用线程模块就不会
multiprocessing的时候,初始化模型放在主程序,预测放在multiprocessing进程
没太明白您的意思:
要删除线描相机的软件和驱动?? 哪个才是??我貌似没安装过啊?(win10企业版)
我的线程和进程操作,都是你说的那样,初始化模型放在主程序,然后把这个变量传入具体的进程函数中。进程函数内进行具体的预测操作。
paddlepaddle验证通过,
Running verify PaddlePaddle program ...
W1208 13:21:04.446607 14284 gpu_resources.cc:61] Please NOTE: device: 0, GPU Compute Capability: 8.6, Driver API Version: 12.0, Runtime API Version: 11.7
W1208 13:21:04.450608 14284 gpu_resources.cc:91] device: 0, cuDNN Version: 8.4.
PaddlePaddle works well on 1 GPU.
PaddlePaddle works well on 1 GPUs.
但运行Paddleocr的时候报错,
RuntimeError: (PreconditionNotMet) The third-party dynamic library (mklml.dll) that Paddle depends on is not configured correctly.(error code is 126)
没有多进程,没有多线程。
paddlepaddle-gpu==2.4.0.post117
paddleocr 2.6.1.1
您好,请问您解决了吗?
应该是某个软件
增加了系统环境
可以删除一部分系统环境设置
排除一下
https://learn.microsoft.com/ja-jp/cognitive-toolkit/setup-mkl-on-windows
downloading from here worked for me
| gharchive/issue | 2021-10-29T14:09:27 | 2025-04-01T04:32:57.382090 | {
"authors": [
"Heryk13",
"MiShiDeHaiLuo",
"changmaoguai",
"lmw0320",
"monkeycc"
],
"repo": "PaddlePaddle/PaddleX",
"url": "https://github.com/PaddlePaddle/PaddleX/issues/1232",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
973314566 | support 3 types model
support 3 types model
LGTM
| gharchive/pull-request | 2021-08-18T06:15:30 | 2025-04-01T04:32:57.383304 | {
"authors": [
"HexToString"
],
"repo": "PaddlePaddle/Serving",
"url": "https://github.com/PaddlePaddle/Serving/pull/1354",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1201332751 | paddle.optimizer.lr.LRScheduler
文档中英文链接
https://www.paddlepaddle.org.cn/documentation/docs/en/api/paddle/optimizer/lr/LRScheduler_en.html
https://www.paddlepaddle.org.cn/documentation/docs/zh/api/paddle/optimizer/lr/LRScheduler_cn.html
红线问题
https://www.paddlepaddle.org.cn/documentation/docs/en/api/paddle/optimizer/lr/LRScheduler_en.html#:~:text=the schedulers state.-,get_lr(),-For those subclass 英文部分没有代码示例,中文有
结构问题
No response
内容问题
中文部分
https://www.paddlepaddle.org.cn/documentation/docs/zh/api/paddle/optimizer/lr/LRScheduler_cn.html#:~:text=请参考 cn_api_paddle_optimizer_lr_PolynomialDecay 未给出链接
https://www.paddlepaddle.org.cn/documentation/docs/zh/api/paddle/optimizer/lr/LRScheduler_cn.html#:~:text=请参考-,基类LRScheduler,-的任意子 格式问题,代码和汉字混在一起了
https://www.paddlepaddle.org.cn/documentation/docs/zh/api/paddle/optimizer/lr/LRScheduler_cn.html#:~:text=学习率策略,-,导出基类的方法为,-form paddle.optimizer 应该为引用而非导出
英文
https://www.paddlepaddle.org.cn/documentation/docs/en/api/paddle/optimizer/lr/LRScheduler_en.html#:~:text=Can be set to restart training 缺乏主语
https://www.paddlepaddle.org.cn/documentation/docs/en/api/paddle/optimizer/lr/LRScheduler_en.html#:~:text=“last_epoch%2C last_lr” 引用代码格式问题
https://www.paddlepaddle.org.cn/documentation/docs/en/api/paddle/optimizer/lr/LRScheduler_en.html#:~:text=set_dict(state_dict)-,Loads the schedulers state.,-get_lr() 应该是loading吧??
格式问题
No response
主观评价
总体OK,英文缺失部分代码示例
请:
1、修复 PolynomialDecay: 多项式衰减,即得到的学习率为初始学习率和给定最终学习之间由多项式计算权重定比分点的插值。请参考 cn_api_paddle_optimizer_lr_PolynomialDecay 中链接问题
2、修改 基类LRScheduler 的格式问题
3、导出基类的方法为 -> 引用基类的方法为
| gharchive/issue | 2022-04-12T07:48:28 | 2025-04-01T04:32:57.390943 | {
"authors": [
"BeyondYourself",
"TCChenlong"
],
"repo": "PaddlePaddle/docs",
"url": "https://github.com/PaddlePaddle/docs/issues/4620",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1382897089 | Add Solace bindings and extension fields
Add Solace bindings and extension fields
The purpose of this change is to allow the parser to handle solace bindings (new) and extension fields of the form x-my-extension-field in the spec files
Changes
Added capability to handle extension fields in objects as properties lists
See com.asyncapi.v2.model.info.Info.java for an example
any unknown field will attempt to be added as a string property
e.g. x-ep-my-property
There is no validation on the field names: yy-my-property will also be accepted
added "solace" bindings
com.asyncapi.v2.binding.solace for examples
Project config changes
Includes update to Java11
It can be downgraded back to Java 8, but would need to be re-tested
Updated all pom versions to 1.0.0-EAP-1 -> 1.0.1-EAP-1
Updated java.version 8 -> 11
Updated jackson.version 2.13.2 -> 2.13.3
kotlin.version 1.6.20 -> 1.7.10
dokka.version 0.10.1 -> 1.7.0
maven-source-plugin 2.2.1 -> 3.2.1
maven-javadoc-plugin 2.9.1 -> 3.3.2
ToDo:
Dependency stanza in README.md files should be updated to SemVer 1.0.1-EAP-1
Class SchemaGeneratorMojo not found
Hi @dennis-brinley,
Thanks for your awesome PR. Will merge it in this month
| gharchive/pull-request | 2022-09-22T19:11:20 | 2025-04-01T04:32:57.421033 | {
"authors": [
"Pakisan",
"dennis-brinley"
],
"repo": "Pakisan/jasyncapi",
"url": "https://github.com/Pakisan/jasyncapi/pull/6",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
271406957 | 点击浏览器刷新按钮后页面空白什么都没有
store.dispatch('GenerateRoutes', store.getters.roles).then(() => {
console.log('路由表', store.getters.addRouters)
router.addRoutes(store.getters.addRouters) // 动态添加可访问的路由
// next({...to}) // hack方法,确保addRutes已完成
next() // 直接使用该方法刷新页面后不会保留在历史页,而页面是空白的
})
如果直接使用next(),在页面中刷新后页面什么都没有,只有一个#app跟节点.
如果使用next({...to})的话页面就正常,刷新后还是在历史页面,
虽然文档写的hack方法,确保addRutes已完成,但是这儿还是不要理解,麻烦楼主详细说一下好么,非常感谢
主要是 addRoutes 没有callback function 所以 你不能确定你 next() 之后进去的页面已成功被 addRoutes 添加。 而next(xxx) 会重新再次进入 router.beforeEach 这个钩子 这时候 addRoutes 已success 所以能直接 next()
| gharchive/issue | 2017-11-06T09:23:32 | 2025-04-01T04:32:57.445768 | {
"authors": [
"PanJiaChen",
"wxxtqk"
],
"repo": "PanJiaChen/vue-element-admin",
"url": "https://github.com/PanJiaChen/vue-element-admin/issues/276",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
532468541 | 关于右侧窗体
Question(提问)
关于右侧窗体,我想请教下,在该窗体里面的页面中如果有下拉框,弹出消息等,都会被该窗体遮挡屏蔽,是否有办法解决?
谢谢!
@PanJiaChen
解决,设置下z-index
| gharchive/issue | 2019-12-04T05:59:21 | 2025-04-01T04:32:57.447064 | {
"authors": [
"moonose"
],
"repo": "PanJiaChen/vue-element-admin",
"url": "https://github.com/PanJiaChen/vue-element-admin/issues/2806",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
694780537 | 当点击 tagsView 右键菜单 '关闭其他' 或者 '关闭所有' 选项时,功能虽然正常,但是会触发设置了 affix: true 的固定标签刷新。
Bug report(问题描述)
当点击 tagsView 右键 '关闭其他' 或者 '关闭所有' 选项时,功能虽然正常,但是会触发设置了 affix: true 的固定标签刷新。
Steps to reproduce(问题复现步骤)
设置 src/router/modules/table.js 中的 drag-table 路由为固定标签:
{
path: 'drag-table',
component: () => import('@/views/table/drag-table'),
name: 'DragTable',
meta: { title: 'Drag Table', affix: true }
}
npm run dev 运行项目
浏览器打开项目,随便打开几个未固定标签,然后点击未固定标签的右键菜单中的 '关闭其他' 或者 '关闭所有' 选项,
这时未固定标签会关闭,但是再次点击 drag-table 标签(当选择的是 ‘关闭所有’ 选项时不需要点击 drag-table 标签)就会刷新
Screenshot or Gif(截图或动态图)
Other relevant information(格外信息)
Your OS: win7 x64
Node.js version: v12.16.3
vue-element-admin version: 4.4.0
已自行修改代码解决
| gharchive/issue | 2020-09-07T07:05:44 | 2025-04-01T04:32:57.451094 | {
"authors": [
"leffss"
],
"repo": "PanJiaChen/vue-element-admin",
"url": "https://github.com/PanJiaChen/vue-element-admin/issues/3416",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
363649976 | Nombre de Tablas
Cambiar nombre de las tablas a minúsculas
Hecho
| gharchive/issue | 2018-09-25T15:54:24 | 2025-04-01T04:32:57.453361 | {
"authors": [
"JoseEmmanuelGC",
"Mgmez"
],
"repo": "Pandiurex/Foraneos-UDG-Api",
"url": "https://github.com/Pandiurex/Foraneos-UDG-Api/issues/36",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
172560001 | Images resizing not correct
Questions
Expected Behavior
When I put images in the timeline, I'd like to see them resized to fit the circles.
Actual Behavior
If the images are not sized well enough, you see ellipsoidal images too much inside the circles or too much outside.
Steps to Reproduce the Issue
Put new images in the /img/timeline/ folder
Change _config.yml
jekyll serve
Display the timeline
Specifications
Every version of the browser, with { Personal } up to date
I've got two ways to solve this issue:
The easy one: modifying the CSS so that the image fits whatever the original dimensions are. You can set the images of the timeline with these properties:
width: 100%;
height: 100%;
Creating a "generate timeline images" script which resizes the images correctly, why not creating new images, "image.png" giving "timeline_image.png".
I'll work on this issue, so I'm open to any comment.
The images should be square in order to be rendered correctly, but I like that you're trying to tackle that problem. Did you come up with a good enough solution?
| gharchive/issue | 2016-08-22T21:30:39 | 2025-04-01T04:32:57.457717 | {
"authors": [
"PanosSakkos",
"djcaesar9114"
],
"repo": "PanosSakkos/personal-jekyll-theme",
"url": "https://github.com/PanosSakkos/personal-jekyll-theme/issues/193",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
239465233 | Rework the help in order to add more detailled description
Since the description of every commands is displayed when using f!help category name, descriptions must be short. However there is already few commands with arguments which might need a more detailled description, so @ParadoxOrigins , add a detailled description and a short one
that is done
| gharchive/issue | 2017-06-29T12:40:23 | 2025-04-01T04:32:57.551112 | {
"authors": [
"ParadoxOrigins"
],
"repo": "ParadoxOrigins/FelixBot",
"url": "https://github.com/ParadoxOrigins/FelixBot/issues/23",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
217202950 | Revisit data
I'd like to be able to revisit my results.
Links #5 #6 #10 #14
closed via #81
| gharchive/issue | 2017-03-27T10:32:54 | 2025-04-01T04:32:57.562705 | {
"authors": [
"ClairePUK",
"ivanmauricio"
],
"repo": "ParkisonsUK-fuse/what3things",
"url": "https://github.com/ParkisonsUK-fuse/what3things/issues/17",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
158181931 | Can we use serverURL with parse/node?
I am trying to migrate my application using a self hosted parse-server setup. I have some test scripts that I run in NodeJS using the parse/node module. However when I specify a server URL, I get an error trying to connect to the parse server.
does the parse/node module support serverURL?
I am trying this:
Parse.initialize('API KEY WOULD GO HERE');
var serverURL = 'https://some.server.url/parse';
Parse.serverURL = serverURL;
and I get this error:
ParseError {
code: 100,
message: 'XMLHttpRequest failed: "Unable to connect to the Parse API"' }
It's supported. That's how all of the integration tests run, so I can guarantee that hasn't broken.
I'd make sure that your server running your node app can actually see your parse server, and that there are no networking complications, but I doubt there's an SDK bug here.
@andrewimm Thanks, that's what I figured also. I have a web application running, which can see the parse server and the browser/client can access the parse server at the same URI, so based on that I do believe it is reachable. I will dig in some more to see what else might be going on.
Well, good luck digging into this. I'd suggest trying to build a simple node script and seeing if it can make requests to your server using the https module. There might be some crazy permissions mismatch going on.
here is what I discovered... if I set the serverURL to http instead of https, I can connect the the remote parse-server. Why? In the browser, I point to https and it works fine.
The web service is deployed on AWS, sitting behind a load balancer. The interesting thing is that in the parse-server instance on the server, I use http://localhost/parse as the serverURL. The load balancer handles https termination I guess, and passes the traffic over http internally? Atleast that is what I think is going on... but again, when I point to https from the browser, all good.
Any ideas here?
You mentioned making requests to the server using the https module... what did you mean by that? is there a specific module I need to use to enable a TLS session using parse/node?
Nah, I just meant writing a simple node program that makes an http requests to your server, and sees if it can get any response. I was referring to node's builtin https module, there's no configuration necessary on the parse/node side.
In most AWS setups with ELB, the ELB is the termination point for SSL (for speed reasons). Anything behind the load-balancer should be using http in its path, to connect to port 80, versus https, which will point to port 443. If it were going back out to the internet before reconnecting (for instance, to get the benefit of the load balancer), you'd still need the https protocol in there.
I want to keep Github issues focused on SDK bugs and improvements, so I'll be closing out this task, since I'm convinced it's tied to your infrastructure and topology.
| gharchive/issue | 2016-06-02T16:18:20 | 2025-04-01T04:32:57.603172 | {
"authors": [
"andrewimm",
"ndnguru"
],
"repo": "ParsePlatform/Parse-SDK-JS",
"url": "https://github.com/ParsePlatform/Parse-SDK-JS/issues/290",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
55541153 | Python 3 Support
Possibly not going into 0.8.2-01, but we should be sure to not forget about this.
Work in progress https://github.com/Parsely/pykafka/pull/144
+1
A lot has happened since work on this started, and the current most promising pull request is here thanks to @sontek. Figuring out a testing workflow around this will be my first priority once #177 is merged.
@sontek's branch is looking pretty good now - as far as I'm concerned it's ready for merging. We'll need to add a bit of documentation about the preferred way to run tests under three interpreters.
#231
#246 merged! :tada: :tada: :tada: :tada: :partly_sunny: :tada:
:+1:
| gharchive/issue | 2015-01-26T21:37:33 | 2025-04-01T04:32:57.610988 | {
"authors": [
"emmett9001",
"kbourgoin",
"mcandre",
"rduplain"
],
"repo": "Parsely/pykafka",
"url": "https://github.com/Parsely/pykafka/issues/120",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
862869909 | PS0001 is redundant
This is now covered by CS4014:
Note that CS4014 also does not report violations when .ConfigureAwait(false) is present:
And it also does not report violations when the task is assigned:
... so it's functionally identical to PS0001.
Bizarrely, CS4014 doesn't show on some calls:
So I guess we still need PS0001 to be sure. 🤷
| gharchive/issue | 2021-04-20T14:19:31 | 2025-04-01T04:32:57.632451 | {
"authors": [
"adamralph"
],
"repo": "Particular/Particular.Analyzers",
"url": "https://github.com/Particular/Particular.Analyzers/issues/90",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1108619743 | Revert "Improvement/updates ruby to 2.6.8"
Reverts Path-Check/gaen-mobile#946
Reverting until major Ruby upgrade
@si1k Fixed!
| gharchive/pull-request | 2022-01-19T22:23:08 | 2025-04-01T04:32:57.714175 | {
"authors": [
"mxMarkowitz",
"si1k"
],
"repo": "Path-Check/gaen-mobile",
"url": "https://github.com/Path-Check/gaen-mobile/pull/961",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
785432855 | Build a pipeline to download data from PubMed
Description
Build a pipeline to download data from PubMed.
// Fill in detail - What is the goal of this issue? How will this information be used?
Specification
Details
Fields to include
PMID
Title
Abstract
Persisting information
filesystem
Format
name: PMID; body: newline-delim title, text
Tasks
[x] Create separate remote on GitHub for this capability (https://github.com/PathwayCommons/pubmed-dl)
[ ] Create script to extract article information and text
[ ] Persist extracted information
[ ] // Fill in detail
Notes
Existing Tools/packages
Wrapper for EUTILS i.e. https://github.com/mfcovington/pubmed-lookup
Considerations
How long does it take to download?
How long does it take to index?
How much memory is used in indexing?
Superceded by https://github.com/PathwayCommons/pubmed-dl/issues/1
Superceded by https://github.com/PathwayCommons/pubmed-dl/issues/1
| gharchive/issue | 2021-01-13T20:38:18 | 2025-04-01T04:32:57.741761 | {
"authors": [
"jvwong"
],
"repo": "PathwayCommons/semantic-search",
"url": "https://github.com/PathwayCommons/semantic-search/issues/43",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2303207365 | BUG: Search not working
Is there an existing issue for this?
[X] I have searched the existing issues
What happened?
@PatilHarshh
Search function not working
Record
[X] I agree to follow this project's Code of Conduct
[X] I'm a GSSOC'24 contributor
[X] I want to work on this issue
check once again !!
| gharchive/issue | 2024-05-17T16:54:27 | 2025-04-01T04:32:57.744481 | {
"authors": [
"PatilHarshh",
"dhruv8433"
],
"repo": "PatilHarshh/Kaam-Do",
"url": "https://github.com/PatilHarshh/Kaam-Do/issues/68",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1200682447 | Running Test for Get Entrance Fee Doesn't Work
I tried run my test for getEntranceFee function and got an error. I can't figure out whats going on here.
My test_lottery.py file :
from brownie import Lottery, accounts, config, network
from web3 import Web3
def test_get_entrance_fee():
account = accounts[0]
lottery = Lottery.deploy(
config["networks"][network.show_active()]["eth_usd_price_feed"],
{"from": account},
)
assert lottery.getEntranceFee() > Web3.toWei(0.018, "ether")
assert lottery.getEntranceFee() < Web3.toWei(0.022, "ether")
my config file:
dependencies:
- smartcontractkit/chainlink-brownie-contracts@1.1.1
compiler:
solc:
remappings:
- "@chainlink=smartcontractkit/chainlink-brownie-contracts@1.1.1"
networks:
mainnet-fork:
eth_usd_price_feed: "0x5f4eC3Df9cbd43714FE2740f5E3616155c5b8419"
output error:
__________________________________________ test_get_entrance_fee ___________________________________________
def test_get_entrance_fee():
account = accounts[0]
lottery = Lottery.deploy(
config["networks"][network.show_active()]["eth_usd_price_feed"],
{"from": account},
)
> assert lottery.getEntranceFee() > Web3.toWei(0.018, "ether")
tests\test_lottery.py:12:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
C:\Users\ssd\.local\pipx\venvs\eth-brownie\lib\site-packages\brownie\network\multicall.py:115: in _proxy_call result = ContractCall.__call__(*args, **kwargs) # type: ignore
C:\Users\ssd\.local\pipx\venvs\eth-brownie\lib\site-packages\brownie\network\contract.py:1751: in __call__
return self.call(*args, block_identifier=block_identifier, override=override)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <ContractCall 'getEntranceFee()'>, block_identifier = None, override = None, args = ()
tx = {'allow_revert': None, 'data': '0x09bc33a7', 'from': '0x66aB6D9362d4F35596279692F0251Db635165871', 'gas': None, ...}
def call(
self, *args: Tuple, block_identifier: Union[int, str, bytes] = None, override: Dict = None
) -> Any:
"""
Call the contract method without broadcasting a transaction.
Arguments
---------
*args
Contract method inputs. You can optionally provide a
dictionary of transaction properties as the last arg.
block_identifier : int | str | bytes, optional
A block number or hash that the call is executed at. If not given, the
latest block used. Raises `ValueError` if this value is too far in the
past and you are not using an archival node.
override : dict, optional
A mapping from addresses to balance, nonce, code, state, stateDiff
overrides for the context of the call.
Returns
-------
Contract method return value(s).
"""
args, tx = _get_tx(self._owner, args)
if tx["from"]:
tx["from"] = str(tx["from"])
del tx["required_confs"]
tx.update({"to": self._address, "data": self.encode_input(*args)})
try:
data = web3.eth.call({k: v for k, v in tx.items() if v}, block_identifier, override)
except ValueError as e:
> raise VirtualMachineError(e) from None
E brownie.exceptions.VirtualMachineError: revert
C:\Users\ssd\.local\pipx\venvs\eth-brownie\lib\site-packages\brownie\network\contract.py:1542: VirtualMachineError
========================================= short test summary info ==========================================
FAILED tests/test_lottery.py::test_get_entrance_fee - brownie.exceptions.VirtualMachineError: revert
============================================ 1 failed in 8.88s =============================================
Terminating local RPC client...
Could you ask on the main repo? Thanks!
https://github.com/smartcontractkit/full-blockchain-solidity-course-py
| gharchive/issue | 2022-04-11T23:45:02 | 2025-04-01T04:32:57.752628 | {
"authors": [
"PatrickAlphaC",
"bananlabs"
],
"repo": "PatrickAlphaC/smartcontract-lottery",
"url": "https://github.com/PatrickAlphaC/smartcontract-lottery/issues/58",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
229908194 | Does it work for Xcode 8.3.2?
I tried my project and some public project i.e Alamofire. But it does not generate anythings.
Console:
./generate-objc-dependencies-to-json.rb -d -s "" > origin.js ; open index.html
find: /Users/username/Library/Caches/appCode*/DerivedData: No such file or directory
There were 91 directories found
Last modifications were in
/Users/username/Library/Developer/Xcode/DerivedData/Alamofire-dhtndbmjkaofkwemgqqalpcgngke/Build/Intermediates/Alamofire.build/Debug/Alamofire macOS.build/Objects-normal/x86_64
directory at
2017-05-19 18:38:31 +1000
Should work fine. You can try change something in the source and try to rebuild project.
Should look like this
| gharchive/issue | 2017-05-19T08:41:26 | 2025-04-01T04:32:57.756113 | {
"authors": [
"PaulTaykalo",
"huy-le"
],
"repo": "PaulTaykalo/objc-dependency-visualizer",
"url": "https://github.com/PaulTaykalo/objc-dependency-visualizer/issues/43",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2141026672 | Обновить документацию.
Обновить информацию по автору yoomoney-api.
https://github.com/AlekseyKorshuk/yoomoney-api
Its unofficial python library...
| gharchive/issue | 2024-02-18T15:01:51 | 2025-04-01T04:32:57.762764 | {
"authors": [
"PavelShaura",
"franciscoMY"
],
"repo": "PavelShaura/Autoseller_bot_aiogram_3",
"url": "https://github.com/PavelShaura/Autoseller_bot_aiogram_3/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1057326296 | feat: adds owner commands
Adds Owner only command like:
evals -> Run evals
load/unload/reload plugins
Looks good to me.
Need a review from @CaffeineDuck .
Please review and approve the changes so this can be merged.
Pull Request has been stagnant for quite a while now.
This is hindering the development of the bot.
Merging this PR.
| gharchive/pull-request | 2021-11-18T12:57:08 | 2025-04-01T04:32:57.769948 | {
"authors": [
"thenishantsapkota"
],
"repo": "Peacebot-Development/Peacebot-v2",
"url": "https://github.com/Peacebot-Development/Peacebot-v2/pull/24",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
318766057 | Use Orchid to maintain documentation and landing page site
Orchid is a tool I recently released which is perfect for managing the documentation of Java projects, and Pebble is the ideal candidate. Orchid itself uses Pebble as its primary template engine, so setting up Pebble with Orchid docs not only will improve the documentation as a whole, but also serve as a great example of what can be done with Pebble.
This issue will track my progress getting Pebble set up with Orchid, and also serve as a discussion around the site.
Progress
[x] Landing page(s)
[x] Full wiki, migrated from the current Github Wiki
[x] Auto-generated listings of/documentation for:
[x] Tags
[x] Filters
[x] Functions
[x] Tests
[x] Operators
[x] Complete Javadoc API documentation
[x] Changelog
[ ] Link to older versions of docs
[x] Automatic build/deploy from Travis CI to Github Pages
[x] Host each version's docs in a subfolder to maintain docs history
Discussion Points
Orchid runs as a Gradle plugin, so using Orchid will require Pebble to either migrate from Maven to Gradle as its sole build system, or use both in the same project (Maven for normal build/test, Gradle for Orchid build). Alternatively, I would like to build a Maven plugin eventually but don't know how; if anyone has experience/would like to help out building a Maven Orchid plugin, we could continue using only Maven here
What should the site look like? Keep the theme used for the current docs, use a default Orchid theme, or find/create a new theme just for Pebble
https://github.com/JavaEden/Orchid
Alright, I've got the initial groundwork laid out for this on the docs branch of my fork, and have a preview of the site deployed to https://pebbletemplates.netlify.com/ . Not much of the actual content has been migrated yet, but I've got it generating empty pages for everything registered in the CoreExtension as a start, currently, along with Javadocs.
@cjbrooks12 I'd prefer to stick with Maven for pebble project. I know it really well vs gradle.
I agree that an actual documentation website would be easier to use than the current wiki where the menu is not user friendly.
Any news on what you started @cjbrooks12 ?
Sorry for the delay with this, there were a number of bugfixes and features Orchid needed before it could really do well as a great project documentation tool. But I have just gotten a few of my other small projects set up fully with Orchid docs, along with Strikt, and feel like it's stable enough and now has the features needed to get Pebble set up with what it needs. I'll get these remaining tasks finished out soon (gotta get those 5 PRs!).
@cjbrooks12 Looking at the strict docs they seem to be build using freemarker, not pebble?
I've been working with the maintainer to replace JBake with Orchid, the PR is still open
@cjbrooks12 is there something I could do to help you migrate the docs to Orchid? (gotta get those 5 PRs too :))
Do you know how to make Maven plugins? My current plan is to use a kscript like this to package everything up so that it doesn't need Gradle to run, but it would be nicer to have a Maven plugin which can run Orchid as part of the normal Pebble deploy process.
Yes, I have written a couple plugins at work, I'll give it a try :)
Very cool!! Maybe add a footer? (at least there's none in the mobile layout)
The landing page could be a little bit improved, especially if it becomes the landing page of pebbletemplates.io, but it's not really related to Orchid nor to this PR :) This can be done later.
Yeah I just used the current Wiki's landing page as the Orchid site homepage as well. And this theme is set up so the "footer" is at the bottom of the sidebar (which becomes a drawer on mobile). But yeah, the copyright on it should be updated to PebbleTemplates, is there anything else that should go in it?
Also, now that I'm thinking about it, it would be nice to have a contributor list in the site, and maybe changelogs as well. I'm not sure how the current release process goes and if it automatically creates a Github Release, but Orchid's changelog plugin can create the release here at the same time the site is deployed, in addition to showing the full list of releases. Less manual work required for releasing new versions of Pebble 😉
The release process doesn't generate a Github release. I run mvn release:prepare release:perform which build everything and push it to sonatype. After that, I manually create the Github release and update documentation with the new version.
It would be awesome to integrate the github release based on a Changelog to the release process and i'm open to hear about it
@cjbrooks12 Can I close this issue ? I see that there's still an uncompleted task.
Perhaps now http://pebbletemplates.io could show the new site instead of redirecting to the github project?
Yeah on next release
| gharchive/issue | 2018-04-30T01:00:46 | 2025-04-01T04:32:57.784175 | {
"authors": [
"DirkLachowski",
"bjansen",
"cjbrooks12",
"ebussieres",
"hectorlf"
],
"repo": "PebbleTemplates/pebble",
"url": "https://github.com/PebbleTemplates/pebble/issues/337",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.