id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1495399017
Org naming change remove old org names replace with new org name closing as name change is too difficult to mediate and restarting a new branch would be easier
gharchive/pull-request
2022-12-14T00:51:14
2025-04-01T04:32:55.357066
{ "authors": [ "jonreding2010" ], "repo": "OpenMAQS/openmaqs-dotnet", "url": "https://github.com/OpenMAQS/openmaqs-dotnet/pull/31", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
591888313
iOS won’t show notifications if the app is in foreground iOS doesn't show notification for apps in foreground @Shreyasldl, I've fixed this bug in my PR, will keep this open until merge.
gharchive/issue
2020-04-01T12:45:44
2025-04-01T04:32:55.418358
{ "authors": [ "Shreyasldl", "s0l0ist" ], "repo": "OpenMined/covid-alert", "url": "https://github.com/OpenMined/covid-alert/issues/25", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
343655683
Travis Support for Unit Tests In Issue #11 , we created a basic Unit Testing Suite... in this issue, we want to support automated unit testing using Travis (with integration to Github) so that all submitted pull requests get tested automatically. This should also include a travis build badge on the Readme. This has been done already.
gharchive/issue
2018-07-23T14:24:59
2025-04-01T04:32:55.419397
{ "authors": [ "cereallarceny", "iamtrask" ], "repo": "OpenMined/syft.js", "url": "https://github.com/OpenMined/syft.js/issues/13", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
55432327
OpenModsLib Server Crash on Join I have been updating my modpack to 1.7.10 After hours of adding/removing mods, i found that just OpenModsLib alone causes a crash. It happens upon trying to join the world. http://pastebin.com/jj1pHsVD Probably version mismatch. But I'm guessing, since you gave absolutely no information about used versions. Also, you are using Bukkit/Spigot. It's also worth mentioning in future reports. Actually, crash occurs between vanilla and bukkit patched code, so I suggest investigating this. While we have very limited support for non-vanilla server, I already know it's start fine. So it's very probably problem is on your side. i have the same issue and i can give you what it if you need me to this is the server console if you would like me to send you a photo or copy of the crash report file i can https://clients.mcprohosting.com/knowledgebase.php?action=displayarticle&id=182 02.04 13:35:25 [Multicraft] Server stopped 02.04 13:35:23 [Multicraft] Not restarting crashed server. 02.04 13:35:23 [Multicraft] Error writing to server: [Errno 32] Broken pipe 02.04 13:35:23 [Multicraft] Error writing to server: [Errno 32] Broken pipe 02.04 13:35:23 [Multicraft] Stopping server! 02.04 13:35:23 [Multicraft] Server shut down 02.04 13:35:22 [Server] INFO The state engine was in incorrect state SERVER_STOPPING and forced into state SERVER_STOPPED. Errors may have been discarded. 02.04 13:35:22 [Server] INFO Holder lookups applied 02.04 13:35:22 [Server] INFO Applying holder lookups 02.04 13:35:22 [Server] INFO Unloading dimension 1 02.04 13:35:22 [Server] INFO Unloading dimension -1 02.04 13:35:22 [Server] INFO Unloading dimension 0 02.04 13:35:22 [Server] INFO Saving chunks for level 'DIM-1'/Nether 02.04 13:35:22 [Server] INFO Saving chunks for level 'DIM1'/The End 02.04 13:35:22 [Server] INFO Saving chunks for level 'Toast'/Overworld 02.04 13:35:22 [Server] INFO Saving worlds 02.04 13:35:22 [Server] INFO Saving players 02.04 13:35:22 [Server] INFO Disabling WorldEdit v6.0;3342-78f975b9 02.04 13:35:22 [Server] INFO Disabling Essentials v2.13.1 02.04 13:35:22 [Server] INFO Disabling EssentialsChat v2.13.1 02.04 13:35:22 [Server] INFO Disabling EssentialsProtect v2.13.1 02.04 13:35:22 [Server] INFO Disabling EssentialsAntiBuild v2.13.1 02.04 13:35:22 [Server] INFO Disabling EssentialsSpawn v2.13.1 02.04 13:35:22 [Server] INFO Stopping server 02.04 13:35:22 [Server] ERROR This crash report has been saved to: /./crash-reports/crash-2015-04-02_13.35.22-server.txt 02.04 13:35:22 [Server] INFO ... 12 more 02.04 13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:174) ~[launchwrapper-1.9.jar:?] 02.04 13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.runTransformers(LaunchClassLoader.java:276) ~[launchwrapper-1.9.jar:?] 02.04 13:35:22 [Server] INFO at openmods.OpenModsClassTransformer.transform(OpenModsClassTransformer.java:230) ~[OpenModsClassTransformer.class:?] 02.04 13:35:22 [Server] INFO at openmods.OpenModsClassTransformer.tryApplyContextWrapper(OpenModsClassTransformer.java:247) ~[OpenModsClassTransformer.class:?] 02.04 13:35:22 [Server] INFO at openmods.asm.VisitorHelper.apply(VisitorHelper.java:29) ~[OpenModsLib-1.7.10-0.6.jar:?] 02.04 13:35:22 [Server] INFO at openmods.OpenModsClassTransformer$8.createVisitor(OpenModsClassTransformer.java:250) ~[OpenModsClassTransformer$8.class:?] 02.04 13:35:22 [Server] INFO at openmods.context.ContextClassTransformer.(ContextClassTransformer.java:65) ~[ContextClassTransformer.class:?] 02.04 13:35:22 [Server] INFO at org.objectweb.asm.ClassVisitor.(Unknown Source) ~[asm-all-4.1.jar:4.1] 02.04 13:35:22 [Server] INFO Caused by: java.lang.IllegalArgumentException 02.04 13:35:22 [Server] INFO at java.lang.Thread.run(Unknown Source) [?:1.7.0_72] 02.04 13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:658) [MinecraftServer.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:798) ~[MinecraftServer.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:423) ~[lt.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:980) ~[MinecraftServer.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.network.NetworkSystem.func_151269_c(NetworkSystem.java:173) ~[nc.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.network.NetworkManager.func_74428_b(NetworkManager.java:247) ~[ej.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.network.NetHandlerLoginServer.func_147233_a(NetHandlerLoginServer.java:70) ~[nn.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.network.NetHandlerLoginServer.func_147326_c(NetHandlerLoginServer.java:130) ~[nn.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.management.ServerConfigurationManager.attemptLogin(ServerConfigurationManager.java:521) ~[oi.class:?] 02.04 13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:188) ~[launchwrapper-1.9.jar:?] 02.04 13:35:22 [Server] INFO java.lang.ClassNotFoundException: net.minecraft.server.management.ItemInWorldManager 02.04 13:35:22 [Server] ERROR Cause of unexpected exception was 02.04 13:35:22 [Server] INFO ... 10 more 02.04 13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:174) ~[launchwrapper-1.9.jar:?] 02.04 13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.runTransformers(LaunchClassLoader.java:276) ~[launchwrapper-1.9.jar:?] 02.04 13:35:22 [Server] INFO at openmods.OpenModsClassTransformer.transform(OpenModsClassTransformer.java:230) ~[OpenModsClassTransformer.class:?] 02.04 13:35:22 [Server] INFO at openmods.OpenModsClassTransformer.tryApplyContextWrapper(OpenModsClassTransformer.java:247) ~[OpenModsClassTransformer.class:?] 02.04 13:35:22 [Server] INFO at openmods.asm.VisitorHelper.apply(VisitorHelper.java:29) ~[OpenModsLib-1.7.10-0.6.jar:?] 02.04 13:35:22 [Server] INFO at openmods.OpenModsClassTransformer$8.createVisitor(OpenModsClassTransformer.java:250) ~[OpenModsClassTransformer$8.class:?] 02.04 13:35:22 [Server] INFO at openmods.context.ContextClassTransformer.(ContextClassTransformer.java:65) ~[ContextClassTransformer.class:?] 02.04 13:35:22 [Server] INFO at org.objectweb.asm.ClassVisitor.(Unknown Source) ~[asm-all-4.1.jar:4.1] 02.04 13:35:22 [Server] INFO Caused by: java.lang.IllegalArgumentException 02.04 13:35:22 [Server] INFO ... 10 more 02.04 13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source) ~[?:1.7.0_72] 02.04 13:35:22 [Server] INFO at net.minecraft.launchwrapper.LaunchClassLoader.findClass(LaunchClassLoader.java:188) ~[launchwrapper-1.9.jar:?] 02.04 13:35:22 [Server] INFO Caused by: java.lang.ClassNotFoundException: net.minecraft.server.management.ItemInWorldManager 02.04 13:35:22 [Server] INFO at java.lang.Thread.run(Unknown Source) [?:1.7.0_72] 02.04 13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:658) [MinecraftServer.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:798) ~[MinecraftServer.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:423) ~[lt.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:980) ~[MinecraftServer.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.network.NetworkSystem.func_151269_c(NetworkSystem.java:173) ~[nc.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.network.NetworkManager.func_74428_b(NetworkManager.java:247) ~[ej.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.network.NetHandlerLoginServer.func_147233_a(NetHandlerLoginServer.java:70) ~[nn.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.network.NetHandlerLoginServer.func_147326_c(NetHandlerLoginServer.java:130) ~[nn.class:?] 02.04 13:35:22 [Server] INFO at net.minecraft.server.management.ServerConfigurationManager.attemptLogin(ServerConfigurationManager.java:521) ~[oi.class:?] 02.04 13:35:22 [Server] INFO java.lang.NoClassDefFoundError: net/minecraft/server/management/ItemInWorldManager 02.04 13:35:22 [Server] ERROR Encountered an unexpected exception 02.04 13:35:22 [Server] INFO openmods.OpenModsClassTransformer.tryApplyContextWrapper(OpenModsClassTransformer.java:246): Adding context wrappers to mx(net.minecraft.server.management.ItemInWorldManager) 02.04 13:35:22 [Server] INFO UUID of player Synthos_Colton is 7c2bf2e6-aecf-44d9-886e-072614862ab9 Oh, now I see what happening (but I had to spend some time to clean those scrambled logs). It's caused by ancient libraries in Cauldron. I guess I can work around that.
gharchive/issue
2015-01-25T22:04:26
2025-04-01T04:32:55.541061
{ "authors": [ "boq", "cjcoley2", "enterprise12nx" ], "repo": "OpenMods/OpenModsLib", "url": "https://github.com/OpenMods/OpenModsLib/issues/32", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
863267303
Investigate ways to improve the compilation environment on macOS Currently, AbstractEquinoxTask copies tools.jar from the current JVM into the Notes bundle in order to have a working compilation environment. This has never been great, but it's particularly problematic when compiling with Java versions >8, which don't have a tools.jar as such. It may be better to do similar to what the Open Liberty Runtime does and download an AdoptOpenJDK build of Java 8 as needed and use that, copying in the Notes.jar and websvc.jar from Notes.app or referencing them in the launch classpath. That'd be cumbersome, but it should be reliable at least until the AdoptOpenJDK URLs change (which they may with the move to Adoptium). I investigated using the JDT by way of EclipseCompiler, but it looks like it is limited to working with filesystem-bound dependencies. There was a bug from way back about similar things, but that may have just been about compiling in-memory files (which works) and not about dependencies.
gharchive/issue
2021-04-20T21:42:09
2025-04-01T04:32:55.554021
{ "authors": [ "jesse-gallagher" ], "repo": "OpenNTF/org.openntf.nsfodp", "url": "https://github.com/OpenNTF/org.openntf.nsfodp/issues/266", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
627664528
ONE package definitions Hi! I couldn't find where the RPM/DEB package definitions (e.g. specfile for RPM) are maintained. Either I missed them or they are not tracked here: in the later case, would it be possible to publish them? I'm currently playing with alpine-based ONE hosts and might end up doing some packaging work which could be upstreamed here. Thanks! Wow, it seems that repository was deleted (or moved to private repository). Previously it was located on https://github.com/opennebula/packages/, now you can still find some forks on github, eg. my one with the latest changes which I managed to find: https://github.com/kvaps/packages How come? Using the package feature was very useful for easy testing / deployment. We are also baffled as to the disappearance of everything under https://github.com/OpenNebula/packages/releases and https://github.com/OpenNebula/packages without clear substitute anywhere else. Is anyone able to comment on why part of the project is now hidden? Perhaps one of the owners / top contributers can chime in: @tinova @vholer @rsmontero @xorel @christian7007 @al3xhh Hi guys! We are doing some changes in the structure of the repos, etc. in preparation for the release of OpenNebula 5.12 "Firecracker". We'll be publishing a proper announcement and a summary of all these changes in a couple of days max. Thanks for your patience! @albert0marti excited to hear about the upcoming 5.12 release and congratulations to the team. However please explain the situation to people in the community who are working to test and setup OpenNebula 5.10.x and 5.11.x and need the packaging resources. Right now it looks like resources have been hidden from the community or deleted by accident. Understand you may want to lock a branch for new releases but keeping existing code in other branches available would be really helpful. If there has been a change of policy and you are withdrawing the packaging code then we would like some warning to prepare for that before the code is removed instead of afterwards. We were preparing for an upgrade when this happened. Thanks for your help. Hi there! Thank you all for your patience! Here you have the details about this decision: https://forum.opennebula.io/t/towards-a-stronger-opennebula-community/8506 Hope this helps to clarify the new scenario. Just to make it clear: access to the packages repo is NOT one of the benefits that comes with the OpenNebula Subscription (https://opennebula.io/enterprise/). Although the repo is now private again, and only accessible in principle to the OpenNebula Team, active members of the community can request access. We are new users of Opennebula and unfortunately this aggressive open-core model will make us reconsider our choice. Sorry to hear that this was the path you chose. Hi @howels, The OpenNebula Community Edition is a full-featured version of OpenNebula, and OpenNebula CE and EE are synchronized at the end of each maintenance cycle, so we are not speaking here about an open core model. Please refer to our Release Policy for more details. The source code (https://github.com/OpenNebula/one/) and the binary and source packages that we produce for the Community Edition (http://downloads.opennebula.io/repo/) are publicly available, given that the CE is under the Apache License 2.0, but if you—for whatever reason—need access to our internal repo for package definitions, just send me a PM.
gharchive/issue
2020-05-30T06:32:19
2025-04-01T04:32:55.564055
{ "authors": [ "Fnux", "albert0marti", "howels", "kvaps", "sambhavsaggi" ], "repo": "OpenNebula/one", "url": "https://github.com/OpenNebula/one/issues/4844", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1896672122
Error: Missing required parameter: auth Followed your running the project locally step by step. I am stuck on the below error #200 already went through this. Double checked my github client id and secret key and they are fine error Error: Missing required parameter: auth [0] at new Replicate (/home/gaurav/openpipe/node_modules/.pnpm/replicate@0.12.3/node_modules/replicate/index.js:37:13) [0] at eval (webpack-internal:///(api)/./src/modelProviders/replicate-llama2/getCompletion.ts:13:19) [0] at process.processTicksAndRejections (node:internal/process/task_queues:95:5) { [0] digest: undefined [0] } [0] - warn ../node_modules/.pnpm/@babel+standalone@7.22.9/node_modules/@babel/standalone/babel.js [0] Critical dependency: the request of a dependency is an expression [0] Login seems to be successfull What am I missing? because of the above auth error I am not able to use the app (seed and project name edit fails) Update on this: It seems that I was missing Replicate Token API and after that it ran successfully Yeah @DVGY07, if you could submit a PR to make the replicate token optional that would be awesome. Sweet, it’s merged. Thanks for the PR!
gharchive/issue
2023-09-14T14:15:01
2025-04-01T04:32:55.575892
{ "authors": [ "DVGY07", "arcticfly" ], "repo": "OpenPipe/OpenPipe", "url": "https://github.com/OpenPipe/OpenPipe/issues/230", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
330549740
Mass edit fix for blanks, numbers and booleans This fix is for #1631 and #180 I'm going to make a cleaner PR - so closing this one
gharchive/pull-request
2018-06-08T07:31:15
2025-04-01T04:32:55.673617
{ "authors": [ "ostephens" ], "repo": "OpenRefine/OpenRefine", "url": "https://github.com/OpenRefine/OpenRefine/pull/1641", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
745863793
Temporary workaround set-env deprecation This is an attempt to temporarily re-enable the set-env command, to make sure our CI stays green, before we figure out how to use the new system. For #3341 Merging to master since GitHub workflows only run there (per our configuration).
gharchive/pull-request
2020-11-18T17:29:43
2025-04-01T04:32:55.674902
{ "authors": [ "wetneb" ], "repo": "OpenRefine/OpenRefine", "url": "https://github.com/OpenRefine/OpenRefine/pull/3342", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1001978186
Build for Java 11 and 16 as well Attempt to check compatibility with newer versions of Java. Of course we have a really tiny test suite on this project, so there are still tons of incompatibilities that can still exist. Given that no one else is active here I will just merge this, release it and see if it magically makes OpenRefine work with Java 16. I'll test the snapshot or beta this morning. Just woke up. @thadguidry there is not much to test unfortunately, because updating OpenRefine to this new version of butterfly (1.2.0) does not solve the problem. So it could well be independent. @wetneb OK understood
gharchive/pull-request
2021-09-21T07:17:52
2025-04-01T04:32:55.677273
{ "authors": [ "thadguidry", "wetneb" ], "repo": "OpenRefine/simile-butterfly", "url": "https://github.com/OpenRefine/simile-butterfly/pull/32", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
882079698
Documentation for plugins when using docker compose (specifically: saml) Is your feature request related to a problem? Please describe. I have set up a working installation of OpenSlides which I would like to connect to our SAML2 SSO. However I have many questions regarding the configuration, especially how to handle Plugins. Do I have to run setup.sh every time I change settings? Do I have to edit any settings.py? If so, it seems I have to do a git stash, git pull, git stash pop before running the update.sh? I enabled the SAML Plugin inside the docker/.env file. But I don't find any useful information where to set it up. The settings.pytells me that the .json is created on startup, but I cannot find it anywhere. Describe the solution you'd like I short and simple howto and where to look in order to setup the saml plugin with docker compose. Maybe a short overview and step by step list needed to update the installation. Describe alternatives you've considered Not using saml. Which would be pretty painful since I need to setup a lot of user manually in that case. Additional context The initial setup is very well described and documented, maybe an example nginx.conf reverse proxy config could be added: location / { proxy_buffering off; proxy_pass http://localhost:8000; } location /ws { proxy_pass http://localhost:8000/ws; proxy_http_version 1.1; proxy_set_header Upgrade $http_upgrade; proxy_set_header Connection "Upgrade"; } } proxy_buffering off; seems to be very important! (Or maybe only for my case since I use 2 proxies :D ) Hi, I have set up the files, the json has the following content: { "strict": true, "debug": true, "sp": { "entityId": "<slides-url>", "assertionConsumerService": { "url": "<slides-url>/apps/saml/?acs", "binding": "urn:oasis:names:tc:SAML:2.0:bindings:HTTP-POST" }, "singleLogoutService": { "url": "<slides-url>/apps/saml/?sls", "binding": "urn:oasis:names:tc:SAML:2.0:bindings:HTTP-Redirect" }, "NameIDFormat": "urn:oasis:names:tc:SAML:1.1:nameid-format:unspecified", "x509cert": "", "privateKey": "" }, "idp": { "entityId": "<simplesamlphp-url>", "singleSignOnService": { "url": "<simplesamlphp-url>/saml2/idp/SSOService.php", "binding": "urn:oasis:names:tc:SAML:2.0:bindings:HTTP-Redirect" }, "singleLogoutService": { "url": "<simplesamlphp-url>/saml2/idp/SingleLogoutService.php", "binding": "urn:oasis:names:tc:SAML:2.0:bindings:HTTP-Redirect" }, "x509cert": "-----BEGIN CERTIFICATE-----<certificate>-----END CERTIFICATE-----" }, "security": { "nameIdEncrypted": false, "authnRequestsSigned": true, "logoutRequestSigned": false, "logoutResponseSigned": false, "signMetadata": true, "wantMessagesSigned": false, "wantAssertionsSigned": false, "wantNameId" : true, "wantNameIdEncrypted": false, "wantAssertionsEncrypted": false, "signatureAlgorithm": "http://www.w3.org/2000/09/xmldsig#rsa-sha1", "digestAlgorithm": "http://www.w3.org/2000/09/xmldsig#sha1" }, "contactPerson": { "technical": { "givenName": "Die Humanisten IT Team", "emailAddress": "<removed>" }, "support": { "givenName": "Die Humanisten Support Team", "emailAddress": "<removed>" } }, "organization": { "en-US": { "name": "OpenSlides", "displayname": "OpenSlides", "url": "http://openslides.org" } }, "generalSettings": { "loginButtonText": "Login via Humanisten SSO", "changePasswordUrl": "<removed>" }, "attributeMapping": { "UserID": ["username", true], "FirstName": ["first_name", false], "LastName": ["last_name", false] }, "requestSettings": {}, "groups": {} } with manual setting the simplesamlphp configuration (is there a way to get the metadata xml for openslides?): $metadata['<slides-url>'] = [ 'entityid' => '<slides-url>', 'contacts' => [], 'metadata-set' => 'saml20-sp-remote', 'AssertionConsumerService' => [ [ 'Binding' => 'urn:oasis:names:tc:SAML:2.0:bindings:HTTP-POST', 'Location' => '<slides-url>/apps/saml/?acs', 'index' => 1, 'isDefault' => true, ], ], 'SingleLogoutService' => [ [ 'Binding' => 'urn:oasis:names:tc:SAML:2.0:bindings:HTTP-Redirect', 'Location' => '<slides-url>/apps/saml/?sls' ], ], 'NameIDFormat' => 'urn:oasis:names:tc:SAML:1.1:nameid-format:emailAddress', 'keys' => [ [ 'encryption' => true, 'signing' => true, 'type' => 'X509Certificate', 'X509Certificate' => '<certificate>', ], ], ]; I do not get any SSO login button. Do I have to rebuild with make / build.sh when changing those settings? You can get the metadata with <your-domain>/apps/saml/metadata/. If you also have the sp.crt and sp.key, you have to regenerate your docker-compose.yml with the m4 command from the readme. Then, just a restart (docker-compose down and docker-compose up) should be sufficient. But a ./build.sh server is not harmful. Ahh, I found the problem I had from the beginning. I used TRUE instead of True. Had a look into the .m4 file and noticed the problem. Now the saml settings appear in the .yml configuration! Just one more issue: due to the proxy, the saml plugins refuses to accept the response: The response was received at http://slides.diehumanisten.de:8000/apps/saml/ instead of https://slides.diehumanisten.de/apps/saml/?acs Is there a way to set an allowlist? Or is the /apps/saml/?acs --> /apps/saml/ the Problem? First, do you have https in "url": "<slides-url>/apps/saml/?acs" (assertionConsumerService) set? The missing ?acs is strange. Are you sure, that your identity provider redirecs correctly? IMO it is not the proxy, since in our instances with the same proxy it redirects correctly. First, do you have https in "url": "<slides-url>/apps/saml/?acs" (assertionConsumerService) set? The missing ?acs is strange. Yes, everything is https. And all the redirects lead to https inside my browser. That works. The redirect internally is also pure https. I also changed all my proxies to use https (snakeoil, proxy_ignore_ssl internally). Additionally my hosts are set in a way that the domain is never "lost" during any hop. The only https --> http redirect is going on inside the docker guest. Are you sure, that your identity provider redirecs correctly? IMO it is not the proxy, since in our instances with the same proxy it redirects correctly. Actually, I am not. We are using simplesamlphp v1.19.0 with a bunch of other SPs connected without any issue. However none of them have a ? inside assertionConsumerService or any other config. I did try again. The ?acs is not missing in the URL of my browser. It is definitely there. I did try setting the assertionConsumerService to https://slides.diehumanisten.de/apps/saml/?acs=acs, because maybe the simplesamlphp did try to "fix" empty variables. The result is the following: The response was received at http://slides.diehumanisten.de:8000/apps/saml/ instead of https://slides.diehumanisten.de/apps/saml/?acs=acs (On another note, something breaks the caddyfile every time I start and stop the service, I have to run build.sh after each restart or it will complain about run: adapting config using caddyfile: /etc/caddy/Caddyfile:20 - Error during parsing: Unexpected EOF) @FinnStutzenstein do you have any idea whats happening here? The ?acs is not the Problem. As you can see in the previous post the second displayed URL in the error message successfully changes. So I guess the http/https is the only issue. Since the only remaining redirection happens inside the container, the problem has to be inside the container. No, sadly I do not have a clue. The caddy issue is fixed with #6083 Hello, sorry for warming up this old thread, but we need openslides more than ever. @FinnStutzenstein Could it be that I am chasing the wrong Error message. There is nothing indicating that the message server_1 | The response was received at http://slides.diehumanisten.de:11080:8000/apps/saml/ instead of https://slides.diehumanisten.de/apps/saml/?acs is indeed an error. I re-made out setup and found another Error: server_1 | [2021-11-28 16:45:13 +0100] [377] [ERROR] django.request Internal Server Error: /apps/saml/ The full log: server_1 | [2021-11-28 16:45:06 +0100] [377] [INFO] openslides.utils.timing [ywVU] request: 1: 0.01347 2: 0.00000 sum: 0.01347 pgbouncer_1 | 2021-11-28 15:45:07.003 7 LOG C-0x55b192781ad0: openslides/openslides@172.24.0.9:40302 login attempt: db=openslides user=openslides tls=no server_1 | [2021-11-28 16:45:07 +0100] [377] [INFO] openslides.utils.timing [ywVU] request: 1: 0.00113 2: 0.00000 sum: 0.00113 server_1 | [2021-11-28 16:45:07 +0100] [383] [INFO] openslides.utils.timing [ywVU] request: 1: 0.17043 2: 0.00001 sum: 0.17044 server_1 | [2021-11-28 16:45:09 +0100] [381] [INFO] openslides.utils.timing [ywVU] request: 1: 0.01461 2: 0.00002 sum: 0.01463 client_1 | 172.24.0.11 - - [28/Nov/2021:15:45:10 +0000] "GET /ngsw-worker.js HTTP/1.1" 200 41983 "-" "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:94.0) Gecko/20100101 Firefox/94.0" server_1 | The response was received at http://slides.diehumanisten.de:11080:8000/apps/saml/ instead of https://slides.diehumanisten.de/apps/saml/?acs server_1 | [2021-11-28 16:45:13 +0100] [377] [INFO] openslides.utils.timing [ywVU] request: 1: 0.05524 2: 0.00000 sum: 0.05524 server_1 | [2021-11-28 16:45:13 +0100] [377] [ERROR] django.request Internal Server Error: /apps/saml/ client_1 | 172.24.0.11 - - [28/Nov/2021:15:45:14 +0000] "GET /ngsw-worker.js HTTP/1.1" 304 0 "-" "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:94.0) Gecko/20100101 Firefox/94.0" pgnode1_1 | 2021-11-28 15:45:15,796 INFO reaped unknown pid 351 pgnode1_1 | 2021-11-28 15:45:15,796 INFO reaped unknown pid 351 client_1 | 172.24.0.11 - - [28/Nov/2021:15:45:18 +0000] "GET /ngsw.json?ngsw-cache-bust=0.9268512327499515 HTTP/1.1" 200 7119 "https://slides.diehumanisten.de/ngsw-worker.js" "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:94.0) Gecko/20100101 Firefox/94.0" @jsangmeister @tsiegleauq Ping I currently have the same issue ... Update: solved it by letting the server run on port 443: server: << : *default-osserver depends_on: - server-setup environment: << : *default-osserver-env secrets: - django - saml_cert - saml_key - saml_config command: - "gunicorn" - "-w" - "8" - "--preload" - "-t" - "240" - "-b" - "0.0.0.0:443" - "openslides.wsgi"
gharchive/issue
2021-05-09T09:44:10
2025-04-01T04:32:56.173491
{ "authors": [ "ClundXIII", "FinnStutzenstein", "dorianim" ], "repo": "OpenSlides/OpenSlides", "url": "https://github.com/OpenSlides/OpenSlides/issues/6057", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
199209964
Add new personal settings MOTIONS_ALLOW_AMENDMENTS_OF_AMENDMENTS. Set settings value to 0 to disable new amendment button in motion detail template. Default: 1. Added changelog entry. Rebased. Is it possible to use a boolean value here instead of 0 and 1? No, it tested it. A python boolean "True/False" in settings.py is invalid (unknown value) in JavaScript (which requires "true/false"). So, 0 and 1 is just the easiest way. Ah I see. Please test a fix in line 187 of core/views.py: The variable client_settings contains a dict but we should transform it to JSON (i. e. a string) and render this string into the template. Then True should be exposed to true. we should transform it to JSON (i. e. a string) and render this string into the template How can I do this? Can you give me an example that I can try it? Head import json Line 187 """.format(realm=realm, angular_modules=angular_modules, settings=json.dumps(client_settings), js_files=js_files) + @normanjaeckel I added a new commit. Please review.
gharchive/pull-request
2017-01-06T14:44:30
2025-04-01T04:32:56.178346
{ "authors": [ "emanuelschuetze", "normanjaeckel" ], "repo": "OpenSlides/OpenSlides", "url": "https://github.com/OpenSlides/OpenSlides/pull/2821", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1124066498
Fixed proto file. @peb-adr Please check this again. The syntax of the current main branch proto file is definitly wrong. There is a > missing. So I tried the following fix. See #157 @normanjaeckel
gharchive/pull-request
2022-02-04T10:53:33
2025-04-01T04:32:56.179892
{ "authors": [ "normanjaeckel", "peb-adr" ], "repo": "OpenSlides/openslides-manage-service", "url": "https://github.com/OpenSlides/openslides-manage-service/pull/158", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2622385706
Not running on Arch Linux There are multiple reports about the OpenTaiko Hub not running on Arch Linux systems, there might be additional configuration to do on the tauri configuration files in order to resolve this issue Same issue here. % ./OpenTaiko.Hub_0.1.4_amd64.AppImage Could not create default EGL display: EGL_BAD_PARAMETER. Aborting... ^C I'm having the same issue here, looking around it seems to be related to libwebkit2gtk
gharchive/issue
2024-10-29T21:42:29
2025-04-01T04:32:56.215804
{ "authors": [ "0auBSQ", "exentio", "zDEFz" ], "repo": "OpenTaiko/OpenTaiko-Hub", "url": "https://github.com/OpenTaiko/OpenTaiko-Hub/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
436903502
Rename comment visibility options, create a new ones, and update help text -=Renames=- Private => Applicant Reviewers and Staff => Reviewers Internal => Team Public => All -=To create=- Applicant and Reviewers Private -=Help text changes=- Applicant: Visible to applicant and team (default) Reviewers: Visible to reviewers and team. Applicants and Reviewers: Visible to applicants (including partners), reviewers, and team. Team: Visible only to team. Private: Visible only to author. All: Visible to any user who has access to the submission. If the submission is public, that means he comment is publicly visible. Relates to #1087 @danblah See how it looks in the screenshot below. Are all these options really needed? I'm afraid the users will be a bit overwhelmed by all the options. All users will not see all the options but staff will. My suggestion is that we at least skip the new "Private" option that makes the comment visible only to the user. This demands a rewrite of the comment access function that now is based on groups. If someone want to have private notes, why should they add it do the system? All good to remove the private option. On the interface, what if have it be a pop up box from a button titled "Visibility". On click, you have checkboxes for the following: Applicant (default) Team (only visible to users in the role Staff) Partners Reviewers The help text below updates as options are selected to say: Visible to: OTF team, Applicants, Partners, Reviewers (etc). Then there's a save/submit or whatever button and it goes back to the comment edit screen. @danblah So then we are down do: Rename Private => Applicant Rename Internal => Team Rename Reviewers and Staff => Reviewers Replace Public => Partners Four options is what we have today so then the UI can stay the same. Those are the right options and I like that they related to roles. I am proposing a different interface though I don't know if its possible. Right now its a radio button with only one option that requires a lot of help text because I don't think its intuitive. I'm proposing that you have a set of checkboxes for the commenterd to select any of the above that apply to them, clearly indicating that those are who can see the comment. I think that makes would allow us to dismiss the help text all together. Is what I'm saying make sense here? The new interface with checkboxes is not implemented but the names are updated and a partner option is added. Will add a new issue for the improved interface. Related: https://github.com/OpenTechFund/opentech.fund/issues/1445 blah asked for small word change from Applicant to Applicant(s) on July 11. All features are visible and functional during test. Thanks!
gharchive/issue
2019-04-24T20:56:58
2025-04-01T04:32:56.225878
{ "authors": [ "danblah", "fourthletter", "frjo" ], "repo": "OpenTechFund/opentech.fund", "url": "https://github.com/OpenTechFund/opentech.fund/issues/1174", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
163456480
20160701 111800 serial ports list This branch does the following: Removes auto-connect feature, and instead simply lists all available serial ports, and reports the list back to the frontend Receives a serial port name from the front end, and attempts a connection to the port Replaces all logging to reference a Logger object Logging files are capped, and backup log files are auto-generated The calibration values (created by the Pipette module) are now correctly being saved with the 'frozen' path, so they save to the correct spot when run as an executable Looks good to me...
gharchive/pull-request
2016-07-01T19:05:08
2025-04-01T04:32:56.295339
{ "authors": [ "SimplyAhmazing", "andySigler" ], "repo": "OpenTrons/otone_backend", "url": "https://github.com/OpenTrons/otone_backend/pull/6", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2391181958
Add description to data_source_connector.go file Compared output of data source "cloudconnexa_connector" and Swagger - noticed that data source was missing "description" field Should there be a test that would've caught this? Theoretically yes) I'm not yet very familiar with testing, how it is done etc. So ... :wink:
gharchive/pull-request
2024-07-04T16:24:49
2025-04-01T04:32:56.298192
{ "authors": [ "patoarvizu", "sahaqaa" ], "repo": "OpenVPN/terraform-provider-cloudconnexa", "url": "https://github.com/OpenVPN/terraform-provider-cloudconnexa/pull/14", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
423747761
New unit test organization and build strategy We have outgrown our current unit test organization and build strategy. We currently have eleven tests that get compiled into eleven separate test executables. Ten of these are for testing the epanet2 library and one is for testing the epanet-output library. The build is taking a long time because the Boost unit_test header file is very large and contains a lot of templates. We were compiling it 11 times. I have made some progress reorganizing things. I propose the following. We have one test module for the Toolkit. Within the test module we will have separate test suites for each of the functional categories found in the Toolkit API. The test suites will be in separate files and will be compiled together into a single test executable at build time. We will link with the boost testing library to eliminate some of the time spent compiling the test headers. These changes should help make things more manageable from both an organizational and build perspective. I have three questions: Will we be able to run the individual test files in stand-alone fashion on our local machine, suitably modified to run without Boost (as I have currently been doing), to allow us to debug CI failures when we submit PRs? Has the successful build status of the dev branch been restored or will we have to wait until the new unit test strategy is implemented before we can submit any new PRs? Are there alternatives to using the Boost unit test library that may be worth exploring or are we too far along in the game for that? Good idea. If the combined test fails, will we know which part of it failed? @LRossman You are welcome to use your ingenuity to figure that out. This is something I hope to check in later today or tomorrow. I suppose it's never to late, provided you are willing to rewrite all the tests :smile: @eladsal Yes boost test tells you where the failures are. @michaeltryby you didn't really answer my first question. Let me rephrase it. Will the unit tests still exist as separate files in the \tests folder as they do now, which allows easy access to individual tests (and makes it easy to add new ones), or will they all be merged into a single file? I hope it's not the latter.
gharchive/issue
2019-03-21T14:08:17
2025-04-01T04:32:56.320802
{ "authors": [ "LRossman", "eladsal", "michaeltryby" ], "repo": "OpenWaterAnalytics/EPANET", "url": "https://github.com/OpenWaterAnalytics/EPANET/issues/427", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1847610305
Bug: Failed to Discover Devices after adding/updating the devices Describe the Bug The devices are not able to be found after adding/updating the children's devices. I updated the existing device "switchbot bot" with different "device refresh rate". I added a new device "switchbot bot" for the another new-buy bot. After restarting the hb, the hb cannot find any devices nor hk showing any switchbot devices. I have check the issue list, it seems that the below issue are similar https://github.com/OpenWonderLabs/homebridge-switchbot/issues/775 https://github.com/OpenWonderLabs/homebridge-switchbot/issues/768 https://github.com/OpenWonderLabs/homebridge-switchbot/issues/757 Please help solve the bug, thank you =) To Reproduce Not sure Expected Behavior The homebridge-switchbot plugin should be able to get all the devices. Relevant Log Output [02/08/2023, 12:17:11] [SwitchBot] Bot: Bot Light Unknown statusCode: 500, Submit Bugs Here: ' + 'https://tinyurl.com/SwitchBotBug [02/08/2023, 12:17:11] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Cannot read properties of undefined (reading 'power')" [03/08/2023, 03:19:53] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Connect Timeout Error" [03/08/2023, 03:20:03] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Connect Timeout Error" [03/08/2023, 03:20:13] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Connect Timeout Error" [03/08/2023, 03:20:23] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Connect Timeout Error" [03/08/2023, 03:20:26] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:20:33] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:20:43] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:20:53] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:21:03] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:21:13] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:21:23] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:21:33] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:21:43] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:21:53] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:22:03] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "getaddrinfo ENOTFOUND api.switch-bot.com" [03/08/2023, 03:22:13] [SwitchBot] Child bridge process ended [03/08/2023, 03:22:13] [SwitchBot] Process Ended. Code: 1, Signal: null [03/08/2023, 03:22:30] [SwitchBot] Initializing SwitchBot platform... [03/08/2023, 03:22:30] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF [03/08/2023, 03:22:31] [SwitchBot] Launched child bridge with PID 4131 [03/08/2023, 03:22:35] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.6.2 child bridge successfully [03/08/2023, 03:22:35] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30} [04/08/2023, 01:15:57] [SwitchBot] Initializing SwitchBot platform... [04/08/2023, 01:15:57] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF [04/08/2023, 01:15:59] [SwitchBot] Launched child bridge with PID 6361 [04/08/2023, 01:16:02] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [04/08/2023, 01:16:02] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30} [04/08/2023, 01:16:07] [SwitchBot] Total SwitchBot Devices Found: 2 [04/08/2023, 01:16:07] [SwitchBot] Restoring existing accessory from cache: Air Conditioner DeviceID: 02-202307122218-42893239 [04/08/2023, 01:16:07] [SwitchBot] Air Conditioner: Air Conditioner Config: {"connectionType":"OpenAPI"} [04/08/2023, 01:16:07] [SwitchBot] Total IR Devices Found: 1 [04/08/2023, 01:16:07] [SwitchBot] Restoring existing accessory from cache: Bot Light DeviceID: C13832304461 [04/08/2023, 01:16:07] [SwitchBot] Bot: Bot Light Config: {"mode":"press","deviceType":"switch","connectionType":"OpenAPI","logging":"standard","refreshRate":10,"offline":false} [04/08/2023, 01:16:07] [SwitchBot] Bot: Bot Light Displaying as Switch [04/08/2023, 01:16:14] [SwitchBot] Bot: Bot Light Set On: true [04/08/2023, 01:16:29] [SwitchBot] Bot: Bot Light Set On: true [04/08/2023, 01:29:04] [SwitchBot] Bot: Bot Light Set On: true [04/08/2023, 10:41:50] [SwitchBot] Bot: Bot Light Set On: true [04/08/2023, 22:18:28] [SwitchBot] Bot: Bot Light Set On: true [04/08/2023, 22:50:57] [SwitchBot] Bot: Bot Light Set On: true [06/08/2023, 00:07:39] [SwitchBot] Bot: Bot Light Set On: true [06/08/2023, 00:17:50] [SwitchBot] Bot: Bot Light Set On: true [06/08/2023, 10:42:14] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Cannot read properties of undefined (reading 'power')" [06/08/2023, 19:25:51] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Cannot read properties of undefined (reading 'power')" [07/08/2023, 01:29:29] [SwitchBot] Bot: Bot Light Set On: true [08/08/2023, 22:25:06] [SwitchBot] Bot: Bot Light Set On: true [09/08/2023, 21:00:04] [SwitchBot] Bot: Bot Light Unknown statusCode: 500, Submit Bugs Here: ' + 'https://tinyurl.com/SwitchBotBug [09/08/2023, 21:00:04] [SwitchBot] Bot: Bot Light failed openAPIRefreshStatus with OpenAPI Connection, Error Message: "Cannot read properties of undefined (reading 'power')" [11/08/2023, 01:32:33] [SwitchBot] Bot: Bot Light Set On: true [11/08/2023, 09:55:28] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 09:55:28] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF [11/08/2023, 09:55:30] [SwitchBot] Launched child bridge with PID 19374 [11/08/2023, 09:55:33] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 09:55:33] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30} [11/08/2023, 09:55:35] [SwitchBot] Total SwitchBot Devices Found: 3 [11/08/2023, 09:55:35] [SwitchBot] Restoring existing accessory from cache: Air Conditioner DeviceID: 02-202307122218-42893239 [11/08/2023, 09:55:35] [SwitchBot] Air Conditioner: Air Conditioner Config: {"connectionType":"OpenAPI"} [11/08/2023, 09:55:35] [SwitchBot] Total IR Devices Found: 1 [11/08/2023, 09:55:35] [SwitchBot] Restoring existing accessory from cache: Bot Light DeviceID: C13832304461 [11/08/2023, 09:55:35] [SwitchBot] Adding new accessory: Bot Door Bot DeviceID: CFF4619C509D [11/08/2023, 09:55:35] [SwitchBot] Bot: Bot Light Config: {"mode":"press","deviceType":"switch","connectionType":"OpenAPI","logging":"standard","refreshRate":10,"offline":false} [11/08/2023, 09:55:35] [SwitchBot] Bot: Bot Light Displaying as Switch [11/08/2023, 09:55:35] [SwitchBot] Bot: Bot Door Config: {"mode":"switch","deviceType":"switch","connectionType":"OpenAPI","refreshRate":10} [11/08/2023, 09:55:35] [SwitchBot] Bot: Bot Door Displaying as Switch [11/08/2023, 09:56:38] [SwitchBot] Bot: Bot Door Set On: true [11/08/2023, 09:56:54] [SwitchBot] Bot: Bot Door Set On: false [11/08/2023, 09:56:58] [SwitchBot] Bot: Bot Door Set On: true [11/08/2023, 09:57:09] [SwitchBot] Bot: Bot Door Set On: false [11/08/2023, 09:57:22] [SwitchBot] Bot: Bot Door Set On: true [11/08/2023, 09:57:38] [SwitchBot] Bot: Bot Door Set On: false [11/08/2023, 09:59:18] [SwitchBot] Bot: Bot Door Set On: true [11/08/2023, 09:59:24] [SwitchBot] Bot: Bot Door Set On: false [11/08/2023, 10:02:06] [SwitchBot] Bot: Bot Door Set On: true [11/08/2023, 10:02:22] [SwitchBot] Bot: Bot Door Set On: false [11/08/2023, 14:57:50] [SwitchBot] Bot: Bot Door Set On: true [11/08/2023, 14:59:55] [SwitchBot] Bot: Bot Door Set On: false [11/08/2023, 23:03:34] [SwitchBot] Bot: Bot Door Set On: true [11/08/2023, 23:03:47] [SwitchBot] Bot: Bot Door Set On: false [11/08/2023, 23:05:09] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:05:09] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF [11/08/2023, 23:05:11] [SwitchBot] Launched child bridge with PID 20503 [11/08/2023, 23:05:14] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:05:15] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30} [11/08/2023, 23:10:29] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:10:29] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF [11/08/2023, 23:10:31] [SwitchBot] Launched child bridge with PID 20599 [11/08/2023, 23:10:35] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:10:35] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30} [11/08/2023, 23:13:53] [SwitchBot] Restarting child bridge... [11/08/2023, 23:13:58] [SwitchBot] Child bridge process ended [11/08/2023, 23:13:58] [SwitchBot] Process Ended. Code: 143, Signal: null [11/08/2023, 23:14:05] [SwitchBot] Restarting Process... [11/08/2023, 23:14:06] [SwitchBot] Launched child bridge with PID 20667 [11/08/2023, 23:14:09] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:14:09] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30} [11/08/2023, 23:15:53] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:15:53] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF [11/08/2023, 23:15:54] [SwitchBot] Launched child bridge with PID 20841 [11/08/2023, 23:15:58] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:15:58] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30} [11/08/2023, 23:16:18] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:16:18] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF [11/08/2023, 23:16:19] [SwitchBot] Launched child bridge with PID 20908 [11/08/2023, 23:16:23] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:16:23] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30} [11/08/2023, 23:17:30] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:17:30] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF [11/08/2023, 23:17:31] [SwitchBot] Launched child bridge with PID 21004 [11/08/2023, 23:17:35] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:17:35] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30} [11/08/2023, 23:19:34] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:19:34] [SwitchBot] Initializing child bridge 0E:84:DC:C8:E2:AF [11/08/2023, 23:19:35] [SwitchBot] Launched child bridge with PID 21477 [11/08/2023, 23:19:39] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:19:39] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":30} [11/08/2023, 23:27:21] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:28:50] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:28:50] [SwitchBot] Initializing child bridge 0E:D2:C4:22:57:08 [11/08/2023, 23:28:52] [SwitchBot] Launched child bridge with PID 21790 [11/08/2023, 23:28:55] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:31:56] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:31:56] [SwitchBot] Initializing child bridge 0E:D2:C4:22:57:08 [11/08/2023, 23:31:58] [SwitchBot] Launched child bridge with PID 21871 [11/08/2023, 23:32:01] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:39:55] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:39:55] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41 [11/08/2023, 23:39:57] [SwitchBot] Launched child bridge with PID 22534 [11/08/2023, 23:40:00] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:41:50] [SwitchBot] Restarting child bridge... [11/08/2023, 23:41:55] [SwitchBot] Child bridge process ended [11/08/2023, 23:41:55] [SwitchBot] Process Ended. Code: 143, Signal: null [11/08/2023, 23:42:02] [SwitchBot] Restarting Process... [11/08/2023, 23:42:03] [SwitchBot] Launched child bridge with PID 22585 [11/08/2023, 23:42:06] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [11/08/2023, 23:44:11] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:44:11] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41 [11/08/2023, 23:44:13] [SwitchBot] Launched child bridge with PID 23104 [11/08/2023, 23:44:16] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [11/08/2023, 23:28:17] [SwitchBot] Initializing SwitchBot platform... [11/08/2023, 23:28:17] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41 [11/08/2023, 23:28:18] [SwitchBot] Launched child bridge with PID 818 [11/08/2023, 23:47:10] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [12/08/2023, 00:06:47] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 00:06:47] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41 [12/08/2023, 00:06:49] [SwitchBot] Launched child bridge with PID 918 [12/08/2023, 00:06:53] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [12/08/2023, 00:18:03] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 00:18:03] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41 [12/08/2023, 00:18:05] [SwitchBot] Launched child bridge with PID 1147 [12/08/2023, 00:18:08] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [12/08/2023, 00:20:49] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 00:20:49] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41 [12/08/2023, 00:20:51] [SwitchBot] Launched child bridge with PID 1295 [12/08/2023, 00:20:54] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [12/08/2023, 00:23:28] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 00:23:28] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41 [12/08/2023, 00:23:30] [SwitchBot] Launched child bridge with PID 1470 [12/08/2023, 00:23:33] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [12/08/2023, 00:24:23] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 00:24:23] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41 [12/08/2023, 00:24:25] [SwitchBot] Launched child bridge with PID 1564 [12/08/2023, 00:24:28] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [12/08/2023, 00:25:55] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 00:25:55] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41 [12/08/2023, 00:25:57] [SwitchBot] Launched child bridge with PID 1702 [12/08/2023, 00:26:00] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [12/08/2023, 00:32:09] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 00:32:09] [SwitchBot] Initializing child bridge 0E:71:DC:1D:3A:41 [12/08/2023, 00:32:12] [SwitchBot] Launched child bridge with PID 1817 [12/08/2023, 00:32:15] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [12/08/2023, 00:50:22] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 00:50:22] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":15} [12/08/2023, 00:53:37] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 00:53:37] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":15} [12/08/2023, 00:55:25] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 00:55:25] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4 [12/08/2023, 00:55:27] [SwitchBot] Launched child bridge with PID 3691 [12/08/2023, 00:55:30] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [12/08/2023, 00:55:30] [SwitchBot] Platform Config: {"logging":"standard","refreshRate":15} [12/08/2023, 01:00:21] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 01:00:21] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4 [12/08/2023, 01:00:23] [SwitchBot] Launched child bridge with PID 3825 [12/08/2023, 01:00:27] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [12/08/2023, 01:00:27] [SwitchBot] [DEBUG] Using Config Logging: debug [12/08/2023, 01:00:27] [SwitchBot] [DEBUG] Finished initializing platform: [12/08/2023, 01:00:27] [SwitchBot] Platform Config: {"logging":"debug","refreshRate":15} [12/08/2023, 01:00:27] [SwitchBot] [DEBUG] Config OK [12/08/2023, 01:00:27] [SwitchBot] [DEBUG] Using Default Push Rate. [12/08/2023, 01:00:28] [SwitchBot] [DEBUG] Executed didFinishLaunching callback [12/08/2023, 01:00:31] [SwitchBot] [DEBUG] Devices: {} [12/08/2023, 01:00:31] [SwitchBot] [DEBUG] Request successful, statusCode: 200 [12/08/2023, 01:00:31] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:00:31 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"f8c76a1c-190e-48cb-8bf7-d31a6fa2c66a","x-amz-apigw-id":"JgVfbHvMIAMEJBQ=","x-amzn-trace-id":"Root=1-64d6692f-61b99ffe79141ea3564ae157;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"} [12/08/2023, 01:00:31] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":15,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}] [12/08/2023, 01:00:31] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read property 'map' of undefined", Submit Bugs Here: https://tinyurl.com/SwitchBotBug [12/08/2023, 01:00:31] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read property 'map' of undefined [12/08/2023, 01:05:43] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 01:05:43] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4 [12/08/2023, 01:05:44] [SwitchBot] Launched child bridge with PID 4018 [12/08/2023, 01:05:48] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [12/08/2023, 01:05:48] [SwitchBot] [DEBUG] Finished initializing platform: [12/08/2023, 01:05:48] [SwitchBot] [DEBUG] Using Config Logging: debug [12/08/2023, 01:05:48] [SwitchBot] Platform Config: {"logging":"debug","refreshRate":6} [12/08/2023, 01:05:48] [SwitchBot] [DEBUG] Using Default Push Rate. [12/08/2023, 01:05:48] [SwitchBot] [DEBUG] Config OK [12/08/2023, 01:05:49] [SwitchBot] [DEBUG] Executed didFinishLaunching callback [12/08/2023, 01:05:50] [SwitchBot] [DEBUG] Devices: {} [12/08/2023, 01:05:50] [SwitchBot] [DEBUG] Request successful, statusCode: 200 [12/08/2023, 01:05:50] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:05:50 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"56d969ef-a617-455a-bfc2-af2d4628a82c","x-amz-apigw-id":"JgWRSG56IAMEB7A=","x-amzn-trace-id":"Root=1-64d66a6e-28d5985205ff8f7b0b948d3a;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"} [12/08/2023, 01:05:50] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":6,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}] [12/08/2023, 01:05:50] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read property 'map' of undefined", Submit Bugs Here: https://tinyurl.com/SwitchBotBug [12/08/2023, 01:05:50] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read property 'map' of undefined [12/08/2023, 01:07:30] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 01:07:30] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4 [12/08/2023, 01:07:32] [SwitchBot] Launched child bridge with PID 4153 [12/08/2023, 01:07:35] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [12/08/2023, 01:07:35] [SwitchBot] [DEBUG] Using Config Logging: debug [12/08/2023, 01:07:35] [SwitchBot] [DEBUG] Finished initializing platform: [12/08/2023, 01:07:35] [SwitchBot] Platform Config: {"logging":"debug","refreshRate":6} [12/08/2023, 01:07:35] [SwitchBot] [DEBUG] Using Default Push Rate. [12/08/2023, 01:07:35] [SwitchBot] [DEBUG] Config OK [12/08/2023, 01:07:36] [SwitchBot] [DEBUG] Executed didFinishLaunching callback [12/08/2023, 01:07:39] [SwitchBot] [DEBUG] Devices: {} [12/08/2023, 01:07:39] [SwitchBot] [DEBUG] Request successful, statusCode: 200 [12/08/2023, 01:07:39] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:07:39 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"5a6586c7-8f35-4fbf-b95d-ca62fe871743","x-amz-apigw-id":"JgWiXGH-IAMEJ8A=","x-amzn-trace-id":"Root=1-64d66adb-1cf5dc35044e47695c5d835c;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"} [12/08/2023, 01:07:39] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":6,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}] [12/08/2023, 01:07:39] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read property 'map' of undefined", Submit Bugs Here: https://tinyurl.com/SwitchBotBug [12/08/2023, 01:07:39] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read property 'map' of undefined [12/08/2023, 01:10:32] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 01:10:32] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4 [12/08/2023, 01:10:34] [SwitchBot] Launched child bridge with PID 5406 [12/08/2023, 01:10:37] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.7.1 child bridge successfully [12/08/2023, 01:10:37] [SwitchBot] [DEBUG] Using Config Logging: debug [12/08/2023, 01:10:37] [SwitchBot] [DEBUG] Finished initializing platform: [12/08/2023, 01:10:37] [SwitchBot] Platform Config: {"logging":"debug","refreshRate":6} [12/08/2023, 01:10:37] [SwitchBot] [DEBUG] Using Default Push Rate. [12/08/2023, 01:10:37] [SwitchBot] [DEBUG] Config OK [12/08/2023, 01:10:38] [SwitchBot] [DEBUG] Executed didFinishLaunching callback [12/08/2023, 01:10:40] [SwitchBot] [DEBUG] Devices: {} [12/08/2023, 01:10:40] [SwitchBot] [DEBUG] Request successful, statusCode: 200 [12/08/2023, 01:10:40] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:10:40 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"ac52f94e-ee4c-446e-8c9b-78aa99a4ce66","x-amz-apigw-id":"JgW-iHJKIAMETRw=","x-amzn-trace-id":"Root=1-64d66b8f-396ff15e39a11ed262b961e4;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"} [12/08/2023, 01:10:40] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":6,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}] [12/08/2023, 01:10:40] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read properties of undefined (reading 'map')", Submit Bugs Here: https://tinyurl.com/SwitchBotBug [12/08/2023, 01:10:40] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read properties of undefined (reading 'map') [12/08/2023, 01:13:54] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 01:13:54] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4 [12/08/2023, 01:13:56] [SwitchBot] Launched child bridge with PID 6654 [12/08/2023, 01:13:59] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [12/08/2023, 01:13:59] [SwitchBot] [DEBUG] Using Config Logging: debug [12/08/2023, 01:13:59] [SwitchBot] [DEBUG] Finished initializing platform: [12/08/2023, 01:13:59] [SwitchBot] [DEBUG] Platform Config: {"logging":"debug","refreshRate":6} [12/08/2023, 01:13:59] [SwitchBot] [DEBUG] Using Default Push Rate. [12/08/2023, 01:13:59] [SwitchBot] [DEBUG] Config OK [12/08/2023, 01:14:00] [SwitchBot] [DEBUG] Executed didFinishLaunching callback [12/08/2023, 01:14:02] [SwitchBot] [DEBUG] Devices: {} [12/08/2023, 01:14:02] [SwitchBot] [DEBUG] Request successful, statusCode: 200 [12/08/2023, 01:14:02] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:14:02 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"fc342878-cd12-4741-8da2-b83625fb834f","x-amz-apigw-id":"JgXeIF-7oAMFWyQ=","x-amzn-trace-id":"Root=1-64d66c5a-6efaa0f031f9a9b81e746a35;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"} [12/08/2023, 01:14:02] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":6,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}] [12/08/2023, 01:14:02] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read properties of undefined (reading 'map')", Submit Bugs Here: https://tinyurl.com/SwitchBotBug [12/08/2023, 01:14:02] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read properties of undefined (reading 'map') [12/08/2023, 01:15:12] [SwitchBot] Initializing SwitchBot platform... [12/08/2023, 01:15:12] [SwitchBot] Initializing child bridge 0E:55:D6:94:91:A4 [12/08/2023, 01:15:14] [SwitchBot] Launched child bridge with PID 6808 [12/08/2023, 01:15:17] [SwitchBot] Loaded @switchbot/homebridge-switchbot v2.8.0-beta.12 child bridge successfully [12/08/2023, 01:15:17] [SwitchBot] [DEBUG] Using Config Logging: debug [12/08/2023, 01:15:17] [SwitchBot] [DEBUG] Finished initializing platform: [12/08/2023, 01:15:17] [SwitchBot] [DEBUG] Platform Config: {"logging":"debug","refreshRate":6} [12/08/2023, 01:15:17] [SwitchBot] [DEBUG] Using Default Push Rate. [12/08/2023, 01:15:17] [SwitchBot] [DEBUG] Config OK [12/08/2023, 01:15:18] [SwitchBot] [DEBUG] Executed didFinishLaunching callback [12/08/2023, 01:15:21] [SwitchBot] [DEBUG] Devices: {} [12/08/2023, 01:15:21] [SwitchBot] [DEBUG] Request successful, statusCode: 200 [12/08/2023, 01:15:21] [SwitchBot] [DEBUG] Headers: {"date":"Fri, 11 Aug 2023 17:15:21 GMT","content-type":"application/json","content-length":"73","connection":"keep-alive","x-amzn-requestid":"df281bbc-a423-418b-a3a4-f7ba0e8f93d1","x-amz-apigw-id":"JgXqeE00IAMErWw=","x-amzn-trace-id":"Root=1-64d66ca9-162845ec48b225120501fd7b;Sampled=0;lineage=c8c2b0f2:0|bf95bacf:0"} [12/08/2023, 01:15:21] [SwitchBot] [DEBUG] SwitchBot Device Config Set: [{"deviceId":"CF:F4:61:9C:50:9D","configDeviceName":"Bot Door","configDeviceType":"Bot","connectionType":"BLE/OpenAPI","bot":{"mode":"switch","deviceType":"switch"},"firmware":"V6.3","refreshRate":6,"logging":"debug"},{"deviceId":"C1:38:32:30:44:61","configDeviceName":"Bot Light","configDeviceType":"Bot","connectionType":"OpenAPI","bot":{"mode":"press","deviceType":"switch"},"firmware":"V6.5","refreshRate":15,"external":false,"logging":"debug"}] [12/08/2023, 01:15:21] [SwitchBot] [DEBUG] Failed to Discover Devices, Error Message: "Cannot read properties of undefined (reading 'map')", Submit Bugs Here: https://tinyurl.com/SwitchBotBug [12/08/2023, 01:15:21] [SwitchBot] [DEBUG] Failed to Discover Devices, Error: TypeError: Cannot read properties of undefined (reading 'map') Config for homebridge-switchbot { "bridge": { "name": "Bill Homebridge", "username": "0E:89:49:64:91:86", "port": 51430, "pin": "630-27-655" }, "accessories": [], "platforms": [ { "name": "Config", "port": 8581, "auth": "form", "theme": "auto", "tempUnits": "c", "lang": "auto", "platform": "config" }, { "name": "SwitchBot", "credentials": { "token": "<hidden>", "secret": "<hidden>", "notice": "Keep your Token & Secret a secret!" }, "options": { "devices": [ { "deviceId": "CF:F4:61:9C:50:9D", "configDeviceName": "Bot Door", "configDeviceType": "Bot", "connectionType": "BLE/OpenAPI", "bot": { "mode": "switch", "deviceType": "switch" }, "firmware": "V6.3", "refreshRate": 6, "logging": "debug" }, { "deviceId": "C1:38:32:30:44:61", "configDeviceName": "Bot Light", "configDeviceType": "Bot", "connectionType": "OpenAPI", "bot": { "mode": "press", "deviceType": "switch" }, "firmware": "V6.5", "refreshRate": 15, "external": false, "logging": "debug" } ], "refreshRate": 6, "logging": "debug" }, "_bridge": { "username": "0E:55:D6:94:91:A4", "port": 57234 }, "platform": "SwitchBot" } ], "disabledPlugins": [] } Screenshots Device and Model SwitchBot Bot Node.js Version v20.5.0 NPM Version v9.8.0 Homebridge Version v1.6.1 Homebridge Switchbot Plugin Version v2.8.0-beta.12 Homebridge Config UI X Plugin Version v4.50.4 Operating System Raspbian I'm having the same error. +1. Was working fine then stopped working for no reason. Tried everything, removing and adding the plugin but now refuses to discover any device via API. @mrlt8 have you experienced this at all? The batteries in my bots are all dead so I haven't encountered this, but the API is pretty flakey and the culprit seems to be: [02/08/2023, 12:17:11] [SwitchBot] Bot: Bot Light Unknown statusCode: 500, Submit Bugs Here: ' + 'https://tinyurl.com/SwitchBotBug You'll probably want to assert the status code is 200 or throw an error around here to avoid the undefined message: https://github.com/OpenWonderLabs/homebridge-switchbot/blob/09ff2520b0e2c7409a371e5c5c264aee17823a7e/src/device/bot.ts#L563-L566 or at the end of the statusCode()? something like: if (statusCode !== 200) { throw new Error(`Unexpected status code: ${statusCode}`); } As for discoverDevices, that one seems to return a 200, so maybe check the length of deviceLists instead of devicesAPI.body? https://github.com/OpenWonderLabs/homebridge-switchbot/blob/472aef8bf69434f757c7cc573dc29d2b81d7bdf6/src/platform.ts#L295-L298 Pushed a beta for this: As for discoverDevices, that one seems to return a 200, so maybe check the length of deviceLists instead of devicesAPI.body? https://github.com/OpenWonderLabs/homebridge-switchbot/blob/472aef8bf69434f757c7cc573dc29d2b81d7bdf6/src/platform.ts#L295-L298 Can you try the latest beta: Click here to find out how to install Beta You'll probably want to assert the status code is 200 or throw an error around here to avoid the undefined message: https://github.com/OpenWonderLabs/homebridge-switchbot/blob/09ff2520b0e2c7409a371e5c5c264aee17823a7e/src/device/bot.ts#L563-L566 or at the end of the statusCode()? something like: if (statusCode !== 200) { throw new Error(`Unexpected status code: ${statusCode}`); } As for discoverDevices, that one seems to return a 200, so maybe check the length of deviceLists instead of devicesAPI.body? https://github.com/OpenWonderLabs/homebridge-switchbot/blob/472aef8bf69434f757c7cc573dc29d2b81d7bdf6/src/platform.ts#L295-L298 Probably a good idea to also throw an error if the status code is not 200. Modified both of these with beta.16 I would probably move the this.statusCode(statusCode); out of the if check and errorLog the statusCode: this.statusCode(statusCode); if (statusCode === 200) { ... } else { this.errorLog(`Failed <openAPIRefreshStatus/openAPIRefreshStatus>, statusCode: ${statusCode}`); } May also want to update all the other requests to catch any non-200 requests. Since you already have the request wrapped in a try-catch, could potentially be easier to just update the statusCode method to throw an error instead, and the try-catch block should catch that: async statusCode(statusCode: number): Promise<void> { switch (statusCode) { ... default: this.infoLog( `${this.device.deviceType}: ${this.accessory.displayName} Unknown statusCode: ` + `${statusCode}, Submit Bugs Here: ' + 'https://tinyurl.com/SwitchBotBug`, ); throw new Error(`Unknown Status Code: ${statusCode}`); } } Hey @donavanbecker thanks for the latest beta! It is now working again! Hi @donavanbecker thanks for the support. I have updated to the latest version, v2.8.0-beta.19, but I still encounter the issue. From my 3 days' observation, I found that my hb starts raising exceptions after 18:00 (UTC + 8) and it will recover after I wake up tomorrow after 8:00 (UTC + 8). What could be the cause of making the plugin down every night? Most likely an issue on the switchbot API. Might want to report it there: https://github.com/OpenWonderLabs/SwitchBotAPI, but support seems non-existent. Or a network issue I would probably move the this.statusCode(statusCode); out of the if check and errorLog the statusCode: this.statusCode(statusCode); if (statusCode === 200) { ... } else { this.errorLog(`Failed <openAPIRefreshStatus/openAPIRefreshStatus>, statusCode: ${statusCode}`); } May also want to update all the other requests to catch any non-200 requests. Since you already have the request wrapped in a try-catch, could potentially be easier to just update the statusCode method to throw an error instead, and the try-catch block should catch that: async statusCode(statusCode: number): Promise<void> { switch (statusCode) { ... default: this.infoLog( `${this.device.deviceType}: ${this.accessory.displayName} Unknown statusCode: ` + `${statusCode}, Submit Bugs Here: ' + 'https://tinyurl.com/SwitchBotBug`, ); throw new Error(`Unknown Status Code: ${statusCode}`); } } @mrlt8 if you have time, wanna throw this in a PR? @pinhao1994 is getting Unknown statusCode: 500 which is an "Internal Server Error" on the API: https://github.com/OpenWonderLabs/SwitchBotAPI#standard-http-error-codes @pinhao1994 is getting Unknown statusCode: 500 which is an "Internal Server Error" on the API: https://github.com/OpenWonderLabs/SwitchBotAPI#standard-http-error-codes @mrlt8 Could you give me more details on this? I believe I am a stable network connection since my other smart devices (that are not Switchbot) still work perfectly. At the time when I am writing this message, I faced the same issue again. It is quite interesting that the SwitchBotAPI always die during the night. I wonder if this could be caused by the request limit? Since I have a short pull time period, I could reach 1000 times API calls per day which caused the error. Sounds like your theory is plausible! Mine stop working again and I have set it to 5 second API refresh rate, which would incur at least 17,280 calls per day. I didn't think so at first as the API should return a 429 error. However, I was doing some debugging and noticed that I kept getting a 200 response in Homebridge when taking my devices offline even though I would get a 171 when making a request directly to the API: {"statusCode":171,"body":{},"message":"No hub record,is offline"} It turns out the plug-in was only checking HTTP response and NOT the statusCode in the json response: https://github.com/OpenWonderLabs/homebridge-switchbot/blob/472aef8bf69434f757c7cc573dc29d2b81d7bdf6/src/device/bot.ts#L538-L543 So we could potentially be ignoring some unexpected statusCode in the json response as long as the HTTP response was 200. v2.8.0 is released @nathennod @donavanbecker I confirm the bug is caused by the 10k request limitation per day. First, I increased the poll rate from 6 sec to 15 sec for 2 bots and the advanced settings. I did not face the issue after 10 pm (UTC+8). However, I meet the issue again at 7:30 am tomorrow morning, but the service recovers at 8:00 am (aka. 00:00 UTC). Therefore, I think although I have 3 poll rates to be exactly the same as 15 sec, I believe they could be polled separately and could still cause the issue. Lastly, I increase the rate to 60 sec and wish for the best. @pinhao1994 Are you setting a refreshRate for each device? You can set different rates for each device. I have 1 bot and I’ve set it to poll 30 seconds and so far so good! @donavanbecker I understand that we could set individual refreshRate for each device, but I suspect the devices are refreshed separately. That is to say, if I have one device, it could be refreshed as frequently as 9 sec which is equal to 9600 (< 10k) API calls per day. My situation is that I have 2 switchbot bots, 1 IR for Air Conditioner, and 1 external IR for TV. Based on the settings, I could set individual refreshRate for each bot and one refreshRate in the "advanced settings" (I assume that the advanced settings refreshRate is for the IR devices".) Now let's assume all devices will be refreshed "separately" in the background based on their refresh rates. Then, although I have max 10K API quota per day, each device could actually only be updated 10k / 4 = 2500 per day. That is to say, in the worst case, each device could only have refreshRate 86400 / 2500 ~= 35 sec. Therefore, to optimize this part, we should first figure out the logic of refresh in the background. If all the devices are refreshed together at once, then the cronjob refresh rate should be max(device1_rate, device2_rate, ..., deviceN_rate, 9) where 9 is the minimum rate to avoid reaching 10k limitation. If all the devices are refreshed separately, then the cronjob refresh rate should be (86400 sec / (10k / num_of_devices)) = (86400 sec * num_of_devices) / 10k sec to avoid the worst case. @pinhao1994 IR devices don't refresh data.
gharchive/issue
2023-08-12T00:54:33
2025-04-01T04:32:56.360705
{ "authors": [ "donavanbecker", "ievolk", "mrlt8", "nathennod", "pinhao1994" ], "repo": "OpenWonderLabs/homebridge-switchbot", "url": "https://github.com/OpenWonderLabs/homebridge-switchbot/issues/789", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
1052783477
Bug: Switchbot plugin started crashing Homebridge today Describe The Bug Restarting Homebridge crashes when Switchbot plugin is configured To Reproduce Just restart. No changes to any of my SwitchBot devices or configuration have been made Expected behavior Shouldn't crash! Relevant log output [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] {"statusCode":100,"body":{"deviceList":[{"deviceId":"98F4AB2F20EE","deviceName":" Humidifier EE","deviceType":"Humidifier","enableCloudService":true,"hubDeviceId":"000000000000"},{"deviceId":"D8E9AA74909F","deviceName":"Li ving Room Curtain Right","deviceType":"Curtain","hubDeviceId":"000000000000","curtainDevicesIds":["DFC80D28CE14","D8E9AA74909F"],"calibrate" :true,"group":true,"master":false,"openDirection":"right"},{"deviceId":"DF7B6DA19FE7","deviceName":"Hub Mini E7","deviceType":"Hub Mini","hu bDeviceId":"000000000000"},{"deviceId":"DFC80D28CE14","deviceName":"Living Room Curtain","deviceType":"Curtain","enableCloudService":true,"h ubDeviceId":"DF7B6DA19FE7","curtainDevicesIds":["DFC80D28CE14","D8E9AA74909F"],"calibrate":true,"group":true,"master":true,"openDirection":" left"},{"deviceId":"F037B7A5EED7","deviceName":"Bot D7","deviceType":"Bot","enableCloudService":true,"hubDeviceId":"DF7B6DA19FE7"},{"deviceI d":"F1BAF4B4E58E","deviceName":"Living Room","deviceType":"Meter","enableCloudService":true,"hubDeviceId":"DF7B6DA19FE7"},{"deviceId":"F3745 EADBBFE","deviceName":"Bedroom","deviceType":"Meter","enableCloudService":true,"hubDeviceId":"DF7B6DA19FE7"}],"infraredRemoteList":[]},"mess age":"success"} [13/11/2021, 16:12:05] [SwitchBot] Total SwitchBot Devices Found: 7 [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] SwitchBot Device Config Not Set: undefined [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Humidifier: 98F4AB2F20EE [13/11/2021, 16:12:05] [SwitchBot] Restoring existing accessory from cache: Humidifier EE DeviceID: 98F4AB2F20EE [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Curtain: D8E9AA74909F [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] deviceId: D8E9AA74909F, curtainDevicesIds: DFC80D28CE14,D8E9AA74909F, master: fal se, group: true, disable_group: undefined [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] [Curtain Config] disable_group: undefined [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Hub Mini: DF7B6DA19FE7 [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Curtain: DFC80D28CE14 [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] deviceId: DFC80D28CE14, curtainDevicesIds: DFC80D28CE14,D8E9AA74909F, master: tru e, group: true, disable_group: undefined [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] [Curtain Config] disable_group: undefined [13/11/2021, 16:12:05] [SwitchBot] Restoring existing accessory from cache: Living Room Curtain DeviceID: DFC80D28CE14 [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Bot: F037B7A5EED7 [13/11/2021, 16:12:05] [SwitchBot] Restoring existing accessory from cache: Bot D7 DeviceID: F037B7A5EED7 [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] undefined [13/11/2021, 16:12:05] [SwitchBot] You must set your Bot to Press or Switch Mode [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Meter: F1BAF4B4E58E [13/11/2021, 16:12:05] [SwitchBot] Restoring existing accessory from cache: Living Room DeviceID: F1BAF4B4E58E [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Discovered Meter: F3745EADBBFE [13/11/2021, 16:12:05] [SwitchBot] Restoring existing accessory from cache: Bedroom DeviceID: F3745EADBBFE [13/11/2021, 16:12:05] [SwitchBot] Total IR Devices Found: 0 [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] IR Device Config Not Set: undefined [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Humidifier Humidifier EE CurrentTemperature: 0 [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Humidifier uuid: 98F4AB2F20EE-Humidifier, (179d92bf-d829-46ee-b45f-954e1a4e6bf1) [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Curtain Living Room Curtain updateHomeKitCharacteristics - {"CurrentPosition":0," PositionState":2,"TargetPosition":0} [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Curtain Living Room Curtain CurrentAmbientLightLevel: undefined [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Curtain uuid: DFC80D28CE14-Curtain, (290ba6f5-8474-4a54-8d14-ce396a2e4251) [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Bot Bot D7 On: false [13/11/2021, 16:12:05] [SwitchBot] Bot Bot D7 failed to refresh status, Error Message: "Cannot read properties of undefi ned (reading 'updateCharacteristic')" [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Bot Bot D7, Error: {} [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Bot Bot D7 BatteryLevel: undefined [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Bot Bot D7 StatusLowBattery: undefined [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Bot uuid: F037B7A5EED7-Bot, (cffc0071-664c-4055-b95d-fc427c34c3a8) [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Meter uuid: F1BAF4B4E58E-Meter, (c79c02a1-a3a0-4f6f-ac00-41124aa94a2a) [13/11/2021, 16:12:05] [SwitchBot] [DEBUG] Meter uuid: F3745EADBBFE-Meter, (97065854-0ef6-426d-9a6f-5604eb4076d8) [13/11/2021, 16:12:05] TypeError: Cannot read properties of undefined (reading 'updateCharacteristic') at Bot.apiError (/usr/local/lib/node_modules/@switchbot/homebridge-switchbot/src/devices/bots.ts:356:18) at Bot.openAPIRefreshStatus (/usr/local/lib/node_modules/@switchbot/homebridge-switchbot/src/devices/bots.ts:233:12) at Bot.refreshStatus (/usr/local/lib/node_modules/@switchbot/homebridge-switchbot/src/devices/bots.ts:168:18) at new Bot (/usr/local/lib/node_modules/@switchbot/homebridge-switchbot/src/devices/bots.ts:52:10) at SwitchBotPlatform.createBot (/usr/local/lib/node_modules/@switchbot/homebridge-switchbot/src/platform.ts:440:9) at processTicksAndRejections (node:internal/process/task_queues:96:5) [13/11/2021, 16:12:05] Got SIGTERM, shutting down Homebridge... Config for homebridge-switchbot "name": "SwitchBot", "credentials": { "openToken": "secret", "notice": "Keep your token a secret!" }, "options": { "refreshRate": 600, "debug": "debug" }, "platform": "SwitchBot" }, Screenshots None Device & Model N/A Node.js Version v16.13.0 NPM Version v8.1.0 Homebridge Version v1.3.6 Homebridge Switchbot Plugin Version v1.0.0 Homebridge Config UI X Plugin Version v4.41.2 Operating System Raspbian Same error here. I just update for the new version of the Switch bot plugin and crash my homebridge instance. duplicate of #77 I'm still having the same issue even when updated to v1.0.1 @deflao Pleas post more logs because more then likely it is a different issue.
gharchive/issue
2021-11-13T21:54:35
2025-04-01T04:32:56.370856
{ "authors": [ "deflao", "donavanbecker", "nfsilva999", "nyjklein" ], "repo": "OpenWonderLabs/homebridge-switchbot", "url": "https://github.com/OpenWonderLabs/homebridge-switchbot/issues/79", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
503398881
Running a live relay requires 10 ETH??? username@computer:~/servers$ npx oz-gsn run-relayer Could not connect to node at http://localhost:8545 (Error: Found no accounts with sufficient balance (10000000000000000000 wei)). Is there a way to bring that down a bit? 10000000000000000000 wei is 10 ETH and that's like $1800 USD pushing it out of the reach of a lot of people that would like to join up. That's the same cash outlay as a good quality mining rig. Why do we gotta be so rich to help out? Thanx! The helpers look for an account with 10 eth because you did not specify --from, so the library has to pick one of your unlocked accounts, and it does so by looking for enough balance that it knows transactions won't fail. It is true however that this number could be lower, but the selecting of a default account feature was intended for testing environments: you'd probably want to make explicit which account transactions will come from. That said, running a relayer does require a deposit on the RelayHub contract of 1 ETH: this is to protect users of the GSN from malicious relayers. You can read more about attack vectors and how they are mitigated (partly using this deposit) on the relevant EIP section: https://eips.ethereum.org/EIPS/eip-1613#attacks-and-mitigations Thank you, this is helpful information. Wouldn't it be better to have the program either require a "from" field or present the user with a list of accounts it found? Also according to this... https://gist.github.com/spalladino/f516b46bdefb5e025c5d28daf7ad1491 it is possible to use infura as the provider and keep a local wallet seperately. But I can't figure out that option with oz-gsn. Ah, yes - if your intent is to run a relayer on a real testnet, then I'd suggest following the guide on that gist. While more involved, you'll achieve a more robust result than simply running a process from JavaScript. run-relayer was originally intended to easily run a relayer locally on a testnet for testing purposes, not production, which is why e.g. it has a default from value.
gharchive/issue
2019-10-07T11:43:11
2025-04-01T04:32:56.381193
{ "authors": [ "edukaded", "nventuro" ], "repo": "OpenZeppelin/openzeppelin-gsn-helpers", "url": "https://github.com/OpenZeppelin/openzeppelin-gsn-helpers/issues/23", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
438386349
Add backwards compatilibty with ERC20 for ERC777 A major feature of ERC777 is that it can be used as if it were an ERC20 token: we should implement this interface to enable this use case. Due to ERC777's state variables being private, the most straightforward way to achieve this would be by having this backward compatibility baked-in into the contract. We could then provide a second contract that disables the ERC20 functions (by calling revert on them) for users to opt-out of this compatibility. I don't think we should provide the non-ERC20-compatible contract for now. Let's wait and see if there is demand for it.
gharchive/issue
2019-04-29T15:52:49
2025-04-01T04:32:56.382764
{ "authors": [ "frangio", "nventuro" ], "repo": "OpenZeppelin/openzeppelin-solidity", "url": "https://github.com/OpenZeppelin/openzeppelin-solidity/issues/1731", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2594649552
[Feature]: ERC1155Burnable extension What is the feature you would like to see? Add ERC1155Burnable extension (solidity). Contribution Guidelines [X] I agree to follow this project's Contribution Guidelines Blocked by #261 Bringing back @programskillforverification code.
gharchive/issue
2024-10-17T12:51:31
2025-04-01T04:32:56.384668
{ "authors": [ "bidzyyys" ], "repo": "OpenZeppelin/rust-contracts-stylus", "url": "https://github.com/OpenZeppelin/rust-contracts-stylus/issues/362", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
251025379
pob white crash screen on booting up. so when i updated pob to the latest version everything was fine until i rebooted my pc every time i open up pob it will just be a big white screen and say that its not responding i tried opening pob as an admin but same problem... every time i reinstall pob it work until i reboot my pc and then its the same problem please help. i play poe alot and need pob to make my builds and try out stuff Am I to assume from your comment in #450 that you're having exactly the same issue (i.e it only occurs when opening on a non-primary monitor)? If so, then I'll close this.
gharchive/issue
2017-08-17T18:04:02
2025-04-01T04:32:56.386057
{ "authors": [ "Novadesu", "Openarl" ], "repo": "Openarl/PathOfBuilding", "url": "https://github.com/Openarl/PathOfBuilding/issues/570", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
692174592
PD: Air gap on dispense Background As a user I may want to air gap after a dispense to prevent drips/contamination between wells (in a multi dispense) and during pipette movement across the deck. Acceptance Criteria [ ] Add a checkbox for air gap after Blow out (https://www.figma.com/file/0dTEbGNsKYgNwKEW6LLfzJ/Airgap-Delay?node-id=50%3A850) [ ] Add tooltip on hover (https://www.figma.com/file/0dTEbGNsKYgNwKEW6LLfzJ/Airgap-Delay?node-id=104%3A0) [ ] Min validation Air gap default is the pipette's min. If there is not enough room left in the tip the default will be the largest value possible between the pipette's min and 0. If a user goes beneath the pipette's recommended minimum the user will get a dismissible warning. [ ] Max validation The remaining volume in a tip after the first (or only) dispense is as follows: (Tip capacity - occupied volume of tip before the dispense) + first dispense volume + air gap volume. If the user is doing a multi aspirate or 1:1 Transfer then the remaining volume in the entire tip capacity. The max is therefore the tip's capacity. If a user is doing a multi dispense then the space for an airgap after the first dispense is any left over room in the tip after the aspirate, plus what was just dispensed (first dispense volume + air gap volume). This number gets bigger as the tip does more dispenses, but to keep things simple we will set the max at the smallest air gap. [ ] Order of operations for single transfer (https://www.figma.com/file/0dTEbGNsKYgNwKEW6LLfzJ/Airgap-Delay?node-id=453%3A6016) [ ] Order of operations for multi dispense (https://www.figma.com/file/0dTEbGNsKYgNwKEW6LLfzJ/Airgap-Delay?node-id=453%3A6016) [ ] Order of operations for multi dispense (https://www.figma.com/file/0dTEbGNsKYgNwKEW6LLfzJ/Airgap-Delay?node-id=453%3A6016) In terms of breaking this down into more technical, bite sized pieces, I see the tickets as: Add checkbox + tooltip for air gap on dispense (UI) Add air gap dispense min validation for air gap on dispense (all three paths) Add air gap dispense max validation for 1:1 transfer and multi aspirate Add air gap dispense max validation for multi dispense Add atomic commands for air gap dispense for single transfer Add atomic commands for air gap dispense for multi aspirate Add atomic commands for air gap dispense for multi dispense Closing this as it is broken down into: #6347 #6500 #6508 #6510 #6512 Note, still need a ticket for max air gap volume validation for multi dispense. Need to clarify whether it is okay if we clamp the maximum value down to whatever the max air gap aspirate value is. @howisthisnamenottakenyet
gharchive/issue
2020-09-03T18:11:23
2025-04-01T04:32:56.394419
{ "authors": [ "howisthisnamenottakenyet", "shlokamin" ], "repo": "Opentrons/opentrons", "url": "https://github.com/Opentrons/opentrons/issues/6482", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1164165728
Show a link to view Robot release notes give users a way to see the release notes for the current robot version. In the Robot Settings > Advanced > Robot Server version section, display a link to see release notes. On click, a modal will display the release notes for the current robot version. Nice-to-have for 5.1 This has been solved by #10010
gharchive/issue
2022-03-09T16:29:37
2025-04-01T04:32:56.395777
{ "authors": [ "emilywools", "koji", "mikecardarelli" ], "repo": "Opentrons/opentrons", "url": "https://github.com/Opentrons/opentrons/issues/9666", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1826797759
docs(api,robot-server): Document errors list as having no more than 1 element Overview In the public (HTTP-facing) models for runs and analyses, you can access the errors of individual commands. In addition, there's an array of top-level errors. That top-level array is used to indicate fatal errors, including those that happen in between commands. This documents, officially, that the top-level array will only ever have zero elements or one element. It will never contain multiple errors. The fact that it's an array, instead of a nullable object, is just a historical quirk. This goes towards RSS-146 by clarifying how we should expose different kinds of run failures. Rationale There is no use for errors to have multiple elements today. This array is left over from a time before individual commands held their own errors. Today, commands have a self-contained error field, but in þe olden tymes, they had an errorId reference that pointed into this array. This was problematic for a few reasons, such as inherent duplication. PR #9715 fixed this. Since then, errors has only had up to 1 element. Having multiple elements is confusing, and muddies what the field is supposed to be doing, conceptually. One internal comment describes it as "a list of fatal errors." But there can only be one fatal error, by definition. For the case where multiple things contributed to a run's failure, and we want to convey them all, we now have @sfoster1's EnumeratedError architecture. We should commit to that tree-based API instead of keeping around a competing array-based API. According to @shlokamin, the app just does, and has always just done, something like errors[0]. If we really need to have multiple elements in this list later on, we can just revert this documentation change. We can do that revert without breaking any clients. Test Plan None needed. Changelog Update the public HTTP documentation in run and analysis models, and internal comments. Review requests Do we agree with this? Risk assessment No risk. Got an in-person approval from @sanni-t.
gharchive/pull-request
2023-07-28T17:22:52
2025-04-01T04:32:56.401628
{ "authors": [ "SyntaxColoring" ], "repo": "Opentrons/opentrons", "url": "https://github.com/Opentrons/opentrons/pull/13191", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2198511262
feat(app): populate ChooseRobotSlideout with runtime parameters closes AUTH-98 closes AUTH-99 closes AUTH-101 Overview Populate second page of ChooseRobotSlideout with RTPs. Add link for restoring values to default. Save state for in an object that will be used in createRunFromProtocolSource Still todo: error handling for numeric input fields sending updated RTP object along with srcFiles during protocol run creation Test Plan select any protocol (this PR uses mock data until analysis provides RTP) select Start setup button to open ChooseRobot slideout select any robot and 'Continue to parameters' footer button observe toggle, numerical input fields, and dropdown menus set to their default values (NOTE: dropdown fields will initialize with the first choice whose value matches its RTP's default value, regardless of its position in the array of choices) observe tooltip on disabled 'Restore default values' link when all values are default observe that all values are restored to default after any RTPs are changed and 'Restore default values' link is clicked Changelog Review requests authorship stakeholders Risk assessment Could you update the tests? @koji @jerader Thank you guys so much for the detailed reviews! I have made the changes you suggested/commented where appropriate. I will push one more commit for adding test coverage for slideouts passing and rendering RTPs and then can merge once approved.
gharchive/pull-request
2024-03-20T20:35:23
2025-04-01T04:32:56.407307
{ "authors": [ "koji", "ncdiehl11" ], "repo": "Opentrons/opentrons", "url": "https://github.com/Opentrons/opentrons/pull/14706", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
380444859
refactor(api): more functions use hc Introduce a ton of changes to inject hardware controller instances everywhere, depend on singletons less and be able to run on both v1 and v2 apis. This does not extend to deck calibration, protocol sessions (including labwarecalibration), or testing tools. In addition, there’s work done on the tests so that tests can more easily run on both versions of the api while relying on simple fixtures; async_client and main_router and the things that depend on them, for instance, now will run on both apis unless marked otherwise. Many small additions have been made to both the new and old APIs to get them to work a bit more like each other to make the logic necessary in dependent systems like the servers to be a little more similar between versions. ** Review Requests ** Look it over. I know there's a lot of it but most of it really can't be separated out. Surface questions you have. I appreciate that this might take a while to review. This PR should actually now come up on a robot with the feature flag set. Try it out by pushing it, connecting to the robot, and using the run app to switch the use protocol api v2 feature flag. When you reset the robot, it should be using protocol API 2. For all of the changes here, the only external facing parts of the system that use this are the endpoints in /control. Try connecting to the robot and flipping lights, homing, refreshing connected modules, that sort of thing. Closes #2245 Performed a few tests on the robot, here are my notes. With Flag ON [ ] Modules still work [ ] Change pipettes works [ ] Deck calibration works [ ] Homing works [ ] Lights work With Flag OFF [ ] share=True appears to be broken (don't think it's necessarily as a result of this PR) [ ] Modules still work [ ] Change pipettes works [ ] Deck calibration works I got an alarm: Hard limit error on the Z (robot was not super off on height upon starting) -- can someone test on their robot to see if they have issues? [ ] Homing works [ ] Lights work
gharchive/pull-request
2018-11-13T22:05:05
2025-04-01T04:32:56.412802
{ "authors": [ "Laura-Danielle", "sfoster1" ], "repo": "Opentrons/opentrons", "url": "https://github.com/Opentrons/opentrons/pull/2675", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1045508802
Op 7901 We prefer small, well tested pull requests. Please refer to Contributing to Spinnaker. When filling out a pull request, please consider the following: Follow the commit message conventions found here. Provide a descriptive summary for your changes. If it fixes a bug or resolves a feature request, be sure to link to that issue. Add inline code comments to changes that might not be obvious. Squash your commits as you keep adding changes. Add a comment to @spinnaker/reviewers for review if your issue has been outstanding for more than 3 days. Note that we are unlikely to accept pull requests that add features without prior discussion. The best way to propose a feature is to open an issue first and discuss your ideas there before implementing them. https://devopsmx.atlassian.net/browse/OP-7901 This PR triggered this build in this execution, please merge to master only if this succeeds: https://spin.prod.opsmx.com/#/applications/gitprtrigger/executions/details/01FKQF0BWWFYKBXH7PQ9KXR16G
gharchive/pull-request
2021-11-05T07:11:29
2025-04-01T04:32:56.425916
{ "authors": [ "Pranav-b-7", "opsmxuser" ], "repo": "OpsMx/gate", "url": "https://github.com/OpsMx/gate/pull/75", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1404126821
Add backendRoles to userrolebinding Fixes https://github.com/Opster/opensearch-k8s-operator/issues/294 @idanl21 @prudhvigodithi Are you OK to merge this one? Hey @swoehrl-mw i'm good with this. Thanks for the PR @AleksZimin
gharchive/pull-request
2022-10-11T07:18:27
2025-04-01T04:32:56.427349
{ "authors": [ "AleksZimin", "prudhvigodithi", "swoehrl-mw" ], "repo": "Opster/opensearch-k8s-operator", "url": "https://github.com/Opster/opensearch-k8s-operator/pull/313", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1890691226
add readme for speculative decoding add readme for speculative decoding under ./scripts/speculative_decoding/README.md add specific cli examples
gharchive/pull-request
2023-09-11T14:50:10
2025-04-01T04:32:56.432682
{ "authors": [ "wheresmyhair" ], "repo": "OptimalScale/LMFlow", "url": "https://github.com/OptimalScale/LMFlow/pull/641", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
549009081
Prepare v4.4.1 [x] Run linters; [x] Run compilers; [x] Run tests; [x] Check documentation site: examples and contents; [x] Test cross-browser compatibility locally and with BrowserStack: Firefox ESR IE11 Latest Edge, Chrome, Firefox, Safari iOS Safari Chrome & Firefox on Android [x] Including RTL mode; [x] Ask for reviews and accessibility testing; [x] npm run release-version to bump version number [x] npm run release to compile dist, update SRI hashes in doc and package the release [x] Prepare changelog: install Conventionnal Changelog globally conventional-changelog -p angular -i CHANGELOG.md -s maintain a ship list … #287 [x] Wait for brand approval or suggestions about current items in navigation [x] implement changes (in components, Boostwtach, examples…) [x] re-run npm run release [x] update Ship List + changelog Then: [ ] commit and push with a chore(release) commit message [ ] merge (on v4-dev for now) [ ] tag your version [ ] create a GitHub release: attach zip file paste CHANGELOG in the release's description [ ] publish documentation on gh-pages copy _gh_pages/ top level files (index.html, redirects.json, sitemap.xml…) and _gh_pages/docs/4.4 to the gh-pages branch ensure Algolia indexes new release content (probably requires a PR) [ ] npm pack then npm publish if you are already logged in NPM (with a personnal account, for example), you'd better use a repository scoped .npmrc file [ ] publish on Nuget [ ] check release on NPM, Nuget, Packagist… Close #287 on merge. New comments to check: [x] Reduce the height of the Supra bar by 2px to 40px. [ ] Align the first navigation label ‘Personal’ so it is left aligned to the Orange logo. [ ] Increase the spacing between the labels in the Supra bar to 15px. [ ] The space between the account icon and the language selection (in the supra bar) needs to be 30px. [ ] The icons in the global header need to move up 2px.
gharchive/pull-request
2020-01-13T15:35:37
2025-04-01T04:32:56.453756
{ "authors": [ "ffoodd" ], "repo": "Orange-OpenSource/Orange-Boosted-Bootstrap", "url": "https://github.com/Orange-OpenSource/Orange-Boosted-Bootstrap/pull/285", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1507943882
Proposal: further tags for output cache https://github.com/OrchardCMS/Orchard/blob/028e2e413b28c46bfcc6e07417a6c28b23bfb523/src/Orchard.Web/Modules/Orchard.OutputCache/Filters/OutputCacheFilter.cs#L261 OutputCache is currently using the Ids of the displayed ContentItems as tags for the OutputCache. In a few of our applications, we've found this somewhat limiting, so we've had to find workarounds, in cases where a content that isn't, strictly speaking, being displayed nonetheless is relevant to what should be shown, and a change in its state should cause the output cache to be evicted. A common example would be some setting or configuration item. We propose adding an interface allowing injection of additional tags, so it's easier for features to control the eviction of elements from output cache. The new interface would look something like: public interface ITagsProvider : IDependency { IEnumerable<string> GetTags(); } In the OutputCacheFilter we would then inject a IEnumerable<ITagsProvider > _tagsProviders, that we would use (syntax is approximate): Tags = new[] { _invariantCacheKey } .Union(cacheItemTags.Distinct()) .Union(._tagsProviders.SelectMany(tp => tp.GetTags()).Distinct()) .ToArray(), Each feature where we would implement one of these would then be able to cause its own cache evictions through its own tags. What about the other way around, a service in the output cache that allows to add custom tags to the current request. This way modules won't have to store what has changed until the output cache calls them. If I understand what you are saying, you mean something that works like the IDisplayedContentItemHandler, but rather than being invoked only on the BuildDisplay (because it's a ContentHandler), it would be possible to invoke it at any point in the code. public interface ITagAdderService : IDependency { void AddTags(params string[] tags); IEnumerable<string> GetTags(); } Anywhere in code: [...] _tagAdderService("foo", "bar"); [...] In the OutputCacheFilter we would then inject a ITagAdderService _tagAdderService, that we would use like: Tags = new[] { _invariantCacheKey } .Union(cacheItemTags.Distinct()) .Union(._tagAdderService.Select(tp => tp.GetTags()).Distinct()) .ToArray(), Is that correct? Exactly. Look around orchard core, we might already be doing that.
gharchive/issue
2022-12-22T14:02:07
2025-04-01T04:32:56.462460
{ "authors": [ "MatteoPiovanelli-Laser", "sebastienros" ], "repo": "OrchardCMS/Orchard", "url": "https://github.com/OrchardCMS/Orchard/issues/8641", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1422221507
1.4. UX Report | Evaluative & Formative Research [x] Usability testing of wireframes: Grant issuers and recipients (link report and survey) [ ] Collect feedback on improvements (link report) [ ] iterate (updated wireframes, figma link) The UI for usability testing is currently in development. See issue #3
gharchive/issue
2022-10-25T10:14:11
2025-04-01T04:32:56.489663
{ "authors": [ "XyloDrone" ], "repo": "OrdumLTD/UX-UI", "url": "https://github.com/OrdumLTD/UX-UI/issues/4", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
356092581
📙 Remove all but ipfs urls from listing pictures before indexing. Instead of removing all pictures data, let's keep IPFS urls. Some merge conflicts due to recent schema changes. Also, there has been some recent updates to the DataStore so I don't expect any listing to have pictures as Data URLs anymore. See this code. So I'm not sure we need to do any filtering anymore ? Except if it is to guard against listing data uploaded directly by partners/malicious actors and that bypassed origin-js code. But maybe for that we should add some additional guards in marketplace.getListing to detect those potentially malformed/malicious listings ? I'll close this PR, no need for it anymore. We do need to make sure we always keep it a throwing error in origin.js when loading a listing using http/https images, since anyone could track people by using them.
gharchive/pull-request
2018-08-31T19:04:29
2025-04-01T04:32:56.516510
{ "authors": [ "DanielVF", "franckc" ], "repo": "OriginProtocol/origin-js", "url": "https://github.com/OriginProtocol/origin-js/pull/394", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
330867949
Missing template / detailed description for ERP customers how to export the GS1 XML file in an automated way or process Expected Behavior Provide detailed concept and implementation examples for the large ERP software vendors how the XML export should be implemented, ideally without customization of the ERP. Customization is moving the customer away from the standard and is therefore seen as bad practice. Actual Behavior A customer using Navision or SAP ERP will not be willing to implement the XML export from the scratch or the reinvent the XML export wheel per customer. No detailed description how the XML export should be implemented for a large ERP vendors like Navision, Infor or SAP. https://github.com/OriginTrail/ot-node/wiki/ERP-Customization Specifications https://community.dynamics.com/ax/b/axvanyakashperuk/archive/2014/09/16/tutorial-generating-shipping-labels-using-the-gs1-sscc-18-barcode-format https://help.sap.com/saphelp_me60/helpdata/EN/f7/86c1536ca9b54ce10000000a174cb4/frameset.htm More detailed instructions can be found here https://github.com/OriginTrail/ot-node/wiki/Data-Structure-Guidelines You can also find templates (examples) here: https://github.com/OriginTrail/ot-node/tree/develop/importers Thanks for covering my question, However, I am not sure if we are on the same page regarding this issue. Image a meeting w. a ERP customer and the CIO asks his staff to implement the importer in an ERP systems. I am pretty sure that a average Joe ERP basis supported will be lost using the current installation. They will ask what OT does recommend, XSD or JSON? As well they will ask which data-structures in their ERP contains the required fields. They need a step by step installation guide how to implement the exporter for their ERP systems. Otherwise a consultancy will charge a small project for implementing the exporter on a ERP.
gharchive/issue
2018-06-09T08:54:13
2025-04-01T04:32:56.523109
{ "authors": [ "ConcernedOriginTrailInvestor", "schrink" ], "repo": "OriginTrail/ot-node", "url": "https://github.com/OriginTrail/ot-node/issues/182", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1213244706
Scroll Physics Can you please add physics property for more customization. @ElsayedDev Hmm.. There is not enough explanation Can you give me a more detailed example?
gharchive/issue
2022-04-23T08:36:43
2025-04-01T04:32:56.524261
{ "authors": [ "ElsayedDev", "Origogi" ], "repo": "Origogi/Vertical_Card_Pager", "url": "https://github.com/Origogi/Vertical_Card_Pager/issues/5", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
1205420374
Workflow test Merging workflow test branch Fixing dependa bot yml
gharchive/pull-request
2022-04-15T09:15:12
2025-04-01T04:32:56.524973
{ "authors": [ "OrionH" ], "repo": "OrionH/Tag-Bot", "url": "https://github.com/OrionH/Tag-Bot/pull/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2083866601
🛑 RoboCheck is down In 094f548, RoboCheck (https://robocheck.ru) was down: HTTP code: 0 Response time: 0 ms Resolved: RoboCheck is back up in aabd42f after 6 days.
gharchive/issue
2024-01-16T12:35:43
2025-04-01T04:32:56.540412
{ "authors": [ "OsintUK" ], "repo": "OsintUK/Up-or-Down", "url": "https://github.com/OsintUK/Up-or-Down/issues/1518", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1562358784
🛑 HakePok is down In 2ee6778, HakePok (http://hakepok.ru) was down: HTTP code: 500 Response time: 821 ms Resolved: HakePok is back up in ff2b2c0.
gharchive/issue
2023-01-30T12:31:20
2025-04-01T04:32:56.542499
{ "authors": [ "OsintUK" ], "repo": "OsintUK/Up-or-Down", "url": "https://github.com/OsintUK/Up-or-Down/issues/477", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
259787664
New winrm upload functionality Two threads here: New winrm_upload function: This allows file transfer through WinRM sreaming. It's not speedy for big binary files, but works well for text/config files. There are unit and integration tests covering it. The origin of this PR is that we need to provide sysprep unattend.xml files to be able to change the SID in windows machines in a fully automated way. It's a bit frustrating, but sysprep is the only supported tool that performs that task, and the only way to fully automate it is to provide config in xml form that gets read after the reboot (Otherwise you get undesired side effects like Administrator password reset that basically breaks the downstream of your automation pipeline). The reason behind changing SID's is that they have to be unique if you want to have them in the same domain, and cloning windows boxes from a Vsphere template sadly gives them the same SID. It all looks a bit too complicated, but this opens a big door for further Windows automation. Fix and improve integration tests: I have fixed some race conditions that appeared recently in the integration tests. I've also made them faster by optimizing the number of times we create and delete the vms and I've made Jenkins to run them in parallel (Python 2 and 3). Builds are now ~5 times faster: 14 minutes against 1 hour and 6 minutes. I ran a few tests of different files. When I started to upload the jre-8u144-windows-x64.exe, it failed with an error similar to above. It failed in different places, sometimes after 4K, after 20K or in its furthest run after 208K. (As an aside it copied at slightly over 2KB/s) Alternatives are: Install SSH, which is a bit clunky in Windows, as you need to do it through 3rd parties Setup an SFTP server with Windows IIS Mount a file share All of them require adding extra stuff to our windows templates or doing it at runtime, which seems to be overkill to send a config file, which is the use case here. So I thought re-using winRM could do the trick, as discussed here https://github.com/diyan/pywinrm/issues/18. I knew it was slow, but I was not expecting it to fail with binaries. I'll take a look to see if I can improve it. By the way, speed is limited by the maximum size that WinRM allows for remote powershell scripts, which is something of the same magnitude of the steps default value of the winrm_upload function. So I don't think we can do much better there unless I am missing something. I cannot find a reliable way of guessing whether a file is a text file or not. Some guys recommend using mimetypes, some others looking at null bytes. I'll give a test to that java exe to see if I can debug the error. @edsharp I have found the problem and I think I have fixed it. Apparently there's a race condition where you can have the previous write process still open when you try to add more bytes to the file and that makes it fail, so you need to provide a small delay for that case. I tested it with your java binary until I got tired of waiting after 10-15 minutes with no errors at 2% and 1MB (It would have taken hours at such speed), and then I tried other binaries and they all went good, with the right sha256. I have added after your last review: File being read in chunks Fix for the race condition and the test to cover that case Better debug information for the RemoteCommandError exceptions so they also show stdout/stderr and not just the command that failed and the status code Some other fixes/improvements to the Jenkinsfile that I saw on the fly Hmm first attempt failed: vm.winrm_upload(local_path='/Users/edward/Downloads/jre-8u144-windows-x64.exe',remote_path='c:\\jre-8u144-windows-x64.exe') vcdriver_vm_winrm_username: Administrator vcdriver_vm_winrm_password: Copying "/Users/edward/Downloads/jre-8u144-windows-x64.exe" to "c:\jre-8u144-windows-x64.exe" ... [ ] 0 %Traceback (most recent call last): File "<stdin>", line 1, in <module> File "vcdriver/config.py", line 114, in wrapper return function(*args, **kwargs) File "vcdriver/vm.py", line 398, in winrm_upload raise WinRmError(script, status_code) vcdriver.exceptions.WinRmError: Remote execution of "add-content -value $([System.Convert]::FromBase64String("aOgaEAAAi0MoiYQkkAAAAEiNVCRoSI1MJDjo4YsAAJBIg7wkgAAAABByCkiLTCRo6Av7AABIibQkgAAAAEiJfCR4QIh8JGhIg7wkoAEAABByDUiLjCSIAQAA6OH6AABMjYQkYAEAAEiNlCTAAAAASI2MJJgAAADorGsAAJAzyf8Vq1wCAEiJtCQIAQAASIm8JAABAABAiLwk8AAAAEiJtCQ4AQAASIm8JDABAABAiLwkIAEAAEyLwLpkAAAASI2MJPAAAADo8qwAAJBIjZQkmAAAAEiNjCTwAAAA6JyxAACQSIO8JDgBAAAQcg1Ii4wkIAEAAOhH+gAASIm0JDgBAABIibwkMAEAAECIvCQgAQAASIO8JAgBAAAQcg1Ii4wk8AAAAOgX+gAASI2UJMAAAABIjUwkaOjRhgAAkEiLVCRwSItMJGhIK9FIuGdmZmZmZmZmSPfqSMH6BEiLwkjB6D9IA9C7AQAAAEg703URSI2UJJgAAADorcoAAITAdQNAit9IjUwkaOgEDQAAhNt0QEiNlCSYAAAASI1MJGjo+vf//5BIjVQkaEiNjCSYAAAA6Ef6//+JRCQwSIO8JIAAAAAQcklIi0wkaOhy+QAA6z26AwAAADkVvV4DAH8wSI0FfJwCAEiJRCQoSI0FoJwCAEiJRCQgRI1Kd0yNBeCaAgBIjQ2RXgMA6Izy//+QSIO8JLAAAAAQcg1Ii4wkmAAAAOgb+QAASIm0JLAAAABIibwkqAAAAECIvCSYAAAASI1MJDjo/YYAAJBIjUwkOOg+CgAAkEiDvCTYAAAAEHINSIuMJMAAAADo1fgAAEiJtCTYAAAASIm8JNAAAABAiLwkwAAAAEiDvCR4AQAAEHIOSIuMJGABAADopfgAAJCLRCQwSIuMJLgBAABIM8zowO4AAEyNnCTAAQAASYtbEEmLcxhJi+Nfw8zMzEiD7DhIjQ2KmQIA/xUvWgIASINkJCAATI1EJCBIjRV1mQIAuQIAAAD/FfpZAgCFwHQeSItMJCBIjRVqmQIA/xUcWgIASIXAdAe5AAgAAP/QSIPEOOkH/P//zMzMQFNIg+wgSIvZSIsJSIXJdAXo+vYAAEiDIwBIg8QgW8NIi0H4SGNQBEiNBUWcAgBIiUQK+MPMzMxIjQWVmwIASIkB6QniAADMSIlcJAhXSIPsIEiLeWhIjQWXmwIASIvZSIkBSIX/dBBIi8/oxOIAAEiLz+ik9wAASI1LCEiLXCQwSIPEIF/pgegAAMxIg8EI6ZfoAADMzMxIg8EI6ZPoAADMzMwzwMPMg8j/w0BTSIPsIEiLAUiL2f9QMIPK/zvCdBdIi0NYARBIi0tASIsRSI1CAUiJAQ+2EovCSA==")) -encoding byte -path c:\jre-8u144-windows-x64.exe" failed with exit code 1
gharchive/pull-request
2017-09-22T11:38:33
2025-04-01T04:32:56.556411
{ "authors": [ "Lantero", "edsharp" ], "repo": "Osirium/vcdriver", "url": "https://github.com/Osirium/vcdriver/pull/3", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
368288919
Cannot connect to tm2 source When using a tm2 source I get the following error (shortened stacktrace): File "XXX/postile/postile.py", line 57, in setup_db ConnectionRefusedError: [Errno 111] Connection refused sys:1: RuntimeWarning: coroutine 'Loop.create_server' was never awaited I use the following command: postile --cors --tm2 ../../tilelive-swift/openmaptiles.tm2source/data.yml When I specify the db connection settings myself using postile --pgpassword ... it works. @boldtrn sorry for the delay, the PostgreSQL connection parameters needs to be passed on command line, they are not read from the data.yml. The tm2source is a format used mainly for the old Mapbox Studio Classic software and is only used partially here (like in OpenMapTiles project). Thanks for the update :+1:
gharchive/issue
2018-10-09T16:13:10
2025-04-01T04:32:56.558867
{ "authors": [ "boldtrn", "ldgeo" ], "repo": "Oslandia/postile", "url": "https://github.com/Oslandia/postile/issues/6", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
420635771
minor change to 'create_ensemble' utility to support mf_datasets Please check if the PR fulfills these requirements [ ] This PR addresses an already opened issue (for bug fixes / features) [ ] Tests for the changes have been added (for bug fixes / features) [ ] Docs have been added / updated (for bug fixes / features) What kind of change does this PR introduce? (Bug fix, feature, docs update, etc.) support mf_dataset when creating ensembles Does this PR introduce a breaking change? (Has there been an API change?) no Pull Request Test Coverage Report for Build 577 8 of 12 (66.67%) changed or added relevant lines in 1 file are covered. No unchanged relevant lines lost coverage. Overall coverage decreased (-0.3%) to 86.364% Changes Missing Coverage Covered Lines Changed/Added Lines % xclim/utils.py 8 12 66.67% Totals Change from base Build 559: -0.3% Covered Lines: 874 Relevant Lines: 1012 💛 - Coveralls
gharchive/pull-request
2019-03-13T17:47:27
2025-04-01T04:32:56.587791
{ "authors": [ "coveralls", "tlogan2000" ], "repo": "Ouranosinc/xclim", "url": "https://github.com/Ouranosinc/xclim/pull/175", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
103894116
Fetch unconfirmed utxos as well Can't see a reason why not. :-/ Thanks! Keep in mind it might take a few hours for the changes to update on the site due to cloudflares cache.
gharchive/pull-request
2015-08-30T01:52:23
2025-04-01T04:32:56.590375
{ "authors": [ "OutCast3k", "dabura667" ], "repo": "OutCast3k/coinbin", "url": "https://github.com/OutCast3k/coinbin/pull/25", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2128108382
InternetStatus and InternetConnection class not found Why after adding this library into our project i can't use InternetStatus and InternetConnection? source: https://github.com/OutdatedGuy/internet_connection_checker_plus/blob/main/example/lib/pages/listen_to_stream.dart I think you have not installed the package properly, can you provide the snippet code for me to debug? Cuz I can make it work on every platform on my side. @OutdatedGuy You're mistaken. I installed the library and got an error. Then I realized that the documentation related to this library is outdated, and the names of classes and methods have changed, but they haven't been updated in the documentation. Can you provide your code where you are trying this package? Also the version of the package you are using and the output of flutter doctor -v. @OutdatedGuy After installing the library i tested only the codes of implemented source code inside sample [✓] Flutter (Channel stable, 3.16.5, on macOS 14.2.1 23C71 darwin-arm64, locale en-US) • Flutter version 3.16.5 on channel stable at /Users/mahdipishguy/Desktop/home/develop/flutter • Upstream repository https://github.com/flutter/flutter.git • Framework revision 78666c8dc5 (8 weeks ago), 2023-12-19 16:14:14 -0800 • Engine revision 3f3e560236 • Dart version 3.2.3 • DevTools version 2.28.4 Can you provide detailed steps from start till the error cuz I am unable to find any issues. Try to use InternetConnectionCheckerPlus in place of InternetConnection and InternetConnectionStatus in place of InternetStatus @pishguy looks like you have installed a outdated version of this package. Try adding version ^2.0.0 or greater. Closing as not an issue. Feel free to comment if you want to re-open this issue.
gharchive/issue
2024-02-10T05:25:38
2025-04-01T04:32:56.598098
{ "authors": [ "AbdNz", "OutdatedGuy", "pishguy" ], "repo": "OutdatedGuy/internet_connection_checker_plus", "url": "https://github.com/OutdatedGuy/internet_connection_checker_plus/issues/28", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1045833867
Move MiniMap event handling code Reference: #650, PR #1118 Move event handling code for mini map clicks from MapViewState to MiniMap. Looks like there's a vcpkg error: Computing installation plan... Error: yasm-tool[core] is only supported on 'native & !uwp' I'm guessing there's a bug in a recent vcpkg update.
gharchive/pull-request
2021-11-05T13:10:37
2025-04-01T04:32:56.599987
{ "authors": [ "DanRStevens" ], "repo": "OutpostUniverse/OPHD", "url": "https://github.com/OutpostUniverse/OPHD/pull/1129", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2166947167
Feature/sales pull request de factories seeders tables and basic models and controllers ok
gharchive/pull-request
2024-03-04T14:12:00
2025-04-01T04:32:56.614948
{ "authors": [ "HelenaDR84", "maicaocaa" ], "repo": "P10MarketplaceDelisana/Delisana_Back", "url": "https://github.com/P10MarketplaceDelisana/Delisana_Back/pull/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
251431285
Compare region viewer: doesn't load for private genomes During the Germany Workshop, several participants noted that the compare region viewer wasn't working for private genomes. -Maulik i have tried with my private genome and got error like below this is fixed on alpha. will be deployed with other fixes soon. working well on production now.
gharchive/issue
2017-08-19T15:19:41
2025-04-01T04:32:56.652992
{ "authors": [ "hyoo", "mshukla1" ], "repo": "PATRIC3/patric3_website", "url": "https://github.com/PATRIC3/patric3_website/issues/1664", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
657503804
Retire PDX custom loaders Description: As a developer, i want to use the universal loader to ensure data loading via templates only. XDog should be ran on all of the PDX Net data sets. Then each of these exports need to be validated and ran in the Updog to ensure that no data is being loss from the XDog. Additionally, the PDX Net data sets are very sparse and it needs to be ensured that the Updog with run the data sets without throwing runtime exceptions. Yomi made the very excellent PDX-Transformer that will export both PDMR and JAX data sets which are on feeds. The PDMR exporter needs to be cleaned up and validated before using it. The Jackson Lab is in development of their own JAX to PDX Finder pipeline so it is unclear if it is necessary to spend any time cleaning the PDX Finder in-house tool. Relevant ticket for PDMR: #296 Jira ticket PDXI-864 Acceptance criteria: Data sets exported and validated for Updog: [ ] PDXNet-HCI-BCM [ ] PDXNet-WUSTL [ ] PDXNet-Wistar-MDAnderson-Penn [ ] PDXNet-MDAnderson [ ] PDMR [ ] JAX [ ] Custom Loader code is retired on Dev this will require some clean-up execute after release split into separate tickets Export of HCI-BCM is blocked by #343 @zperova Done Great work, @Afollet Thank you!
gharchive/issue
2020-07-15T16:55:10
2025-04-01T04:32:56.731908
{ "authors": [ "Afollet", "zperova" ], "repo": "PDXFinder/pdxfinder", "url": "https://github.com/PDXFinder/pdxfinder/issues/184", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
110850151
SQL Error Description [Insert description of issue here] Detailed Information Is available here Similar to #2193 Follow my instructions that I've posted on #2193 and reply here on your results. Unfortunately PermissionsEx is currently unmaintained and no further updates are planned. Because of this, all pull requests are being closed. If you are looking for an actively maintained permissions plugin, take a look at LuckPerms. It is compatible with major server platforms and can import data from PEX.
gharchive/issue
2015-10-11T11:17:36
2025-04-01T04:32:56.741325
{ "authors": [ "CzEMC", "tteeee", "zml2008" ], "repo": "PEXPlugins/PermissionsEx", "url": "https://github.com/PEXPlugins/PermissionsEx/issues/2185", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
190531944
MYSQL Bug Here: http://minilebenmc.de/Bugs/pex/1.png http://minilebenmc.de/Bugs/pex/2.png Try deleting all the tables related to PEX (permissions, permissions_entity and permissions_inheritance). Let the plugin remake the tables Unfortunately PermissionsEx is currently unmaintained and no further updates are planned. Because of this, all pull requests are being closed. If you are looking for an actively maintained permissions plugin, take a look at LuckPerms. It is compatible with major server platforms and can import data from PEX.
gharchive/issue
2016-11-20T01:11:28
2025-04-01T04:32:56.743529
{ "authors": [ "LartyHD", "RubeVi", "zml2008" ], "repo": "PEXPlugins/PermissionsEx", "url": "https://github.com/PEXPlugins/PermissionsEx/issues/2586", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
766320169
Behaviour of @x.setter and @x.deleter In using a CleverDict instance to hold passwords (in conjunction with keyring) I noticed the setter and delete decorators didn't appear to be working: @property def password(self): return keyring.get_password(account, self.username) @password.setter def password(self, value): print("Setting!") keyring.set_password(account, self.username, value) @password.deleter def password(self): keyring.delete_password(account, username) >>> value = "mysecret" >>> keyring.set_password(CleverSession.choices[self.url], self.username, value) >>> self.password 'mysecret' # set >>> self.password = "newpassword" >>> self.password 'mysecret' >>> self['password'] 'newpassword' # delete >>> del self.password >>> self.password 'mysecret' >>> self['password'] Traceback (most recent call last): File "<stdin>", line 1, in <module> File "C:\Users\Peter\AppData\Roaming\Python\Python39\site-packages\cleverdict\cleverdict.py", line 168, in __getitem__ name = self.get_key(name) File "C:\Users\Peter\AppData\Roaming\Python\Python39\site-packages\cleverdict\cleverdict.py", line 310, in get_key raise KeyError(name) KeyError: 'password' So I'm just wondering if this is something that has to be worked around, a limitation of CleverDict, or if there's something even more Clever we can do in CleverDict itself so that @x.setter and @x.deleter work as expected right out of the box? It is tricky to try and combine properties with setattr and getattr and getattribute. This stackoverflow article describes it quite well: https://stackoverflow.com/questions/15750522/class-properties-and-setattr Labelled as "wontfix" after Version 1.8.0 simply because the main authors don't currently have the time for non essential enhancements. If there's a huge uptake in people using CleverDict and asking for this feature that may change, or if you're reading this why not Fork, develop a solution, and contribute a Pull Request yourself? Labelled as "wontfix" after Version 1.8.0 simply because the main authors don't currently have the time for non essential enhancements. If there's a huge uptake in people using CleverDict and asking for this feature that may change, or if you're reading this why not Fork, develop a solution, and contribute a Pull Request yourself? import keyring account = "PyPI" class User: def init(self): self.username = "Pfython" @property def password(self): return keyring.get_password(account, self.username) @password.setter def password(self, value): keyring.set_password(account, self.username, value) @password.deleter def password(self): keyring.delete_password(account, username)
gharchive/issue
2020-12-14T11:26:08
2025-04-01T04:32:56.754408
{ "authors": [ "PFython", "salabim" ], "repo": "PFython/cleverdict", "url": "https://github.com/PFython/cleverdict/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
492263358
Ravines generation doesn't turn off OTG Plugin or Mod [x] OTG Mod for Forge [ ] OTG Plugin for Spigot OTG Version [ ] 1.12.2_v8_r1 [ ] 1.12.2_v7_DEV_r26 [x] 1.12.2-v6 [ ] 1.12.2-v5 [ ] 1.11.2-v8 [ ] 1.11.2-v9 [ ] 1.10.2-v21 [ ] Other - specify here Forge or Spigot Versions Forge 14.23.5.2814 What other plugins are you running? Only decorative/technical mods:Optifine,BetterFoliage,JourneyMap,TellMe,TreeCapitator,NEI+JEI Description There were a lot of ravines everywhere, so I started changing values in WorldConfig.ini,however,it doesn't help.I had to set option RavineRarity: 0.0 in WorldConfig.ini, but it also doesn't work,ravines are still found in every biome,even very small,like beach,sometimes up to 3 in the same place,crossshaped.I want ravines to happen in the world,but not so often,and not in every biome.I didn't change cave settings,they are still default. Is there any way to change ravines frequency? If you need to look at WorldConfig.ini,I attached it(in txt format),but it is still work-in-progress and a big mess:) WorldConfig.txt Ravine rarity shouldn't have a decimal in it, try setting it to 0 instead. Thank you very much! It helped ^_^ It was really stupid mistake,I hadn't even thought about decimal!
gharchive/issue
2019-09-11T14:10:19
2025-04-01T04:32:56.759741
{ "authors": [ "Coll1234567", "Darsenia" ], "repo": "PG85/OpenTerrainGenerator", "url": "https://github.com/PG85/OpenTerrainGenerator/issues/320", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
648471395
The link to the Blitz XML is a dead link On the page https://pgm.dev/docs/modules/objectives/blitz/ the link to the Blitz XML is a dead link, returns a GitHub 404 as that no longer exists. (https://maps.oc.tc/Blitz/blitz-global.xml) Sorry for the wait. This has been fixed in https://github.com/PGMDev/Website/commit/ea04bbc0d0c837f534d07ee8121688137a3bf1d2.
gharchive/issue
2020-06-30T20:29:31
2025-04-01T04:32:56.761572
{ "authors": [ "IcedCode", "Indicardo" ], "repo": "PGMDev/Website", "url": "https://github.com/PGMDev/Website/issues/5", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2533603489
PSR12 4.4 Methods and Functions close braces same line Is your feature request related to a problem? It is related to PHP8 and PSR12 standard Describe the solution you'd like As described in PSR12 item 4.4 If a function or method contains no statements or comments (such as an empty no-op implementation or when using constructor property promotion), then the body SHOULD be abbreviated as {} and placed on the same line as the previous symbol, separated by a space. For example: class Point { public function __construct(private int $x, private int $y) {} // ... } class Point { public function __construct( public readonly int $x, public readonly int $y, ) {} } Additional context (optional) Exists some discussion and "suggestions" in this topic - https://github.com/squizlabs/PHP_CodeSniffer/issues/3291 [x] I have read the Contribution Guidelines and this is not a support question. Don't know where you're getting your information from, but this is not part of PSR12. This is the literal text of PSR12 section 4.4: 4.4 Methods and Functions Visibility MUST be declared on all methods. Method names MUST NOT be prefixed with a single underscore to indicate protected or private visibility. That is, an underscore prefix explicitly has no meaning. Method and function names MUST NOT be declared with space after the method name. The opening brace MUST go on its own line, and the closing brace MUST go on the next line following the body. There MUST NOT be a space after the opening parenthesis, and there MUST NOT be a space before the closing parenthesis. A method declaration looks like the following. Note the placement of parentheses, commas, spaces, and braces: ... code sample... https://www.php-fig.org/psr/psr-12/#44-methods-and-functions I took it from - https://www.php-fig.org/per/coding-style/ that can be accessed in the page https://www.php-fig.org/psr/psr-12/ and click on the link "PER Coding Style" on the right top menu (Additional info) @junior-paytech PSR-PER is not the same as PSR12 and for those things for which PSR12 has its own rules, PSR-12 sniffs cannot be updated to follow PSR-PER as in that case, the PSR 12 sniff would no longer enforce PSR12. I understand your point, but I believe that the PER is more "actual" and aligned with the current php development. Anyway, thank you your support.
gharchive/issue
2024-09-18T12:26:30
2025-04-01T04:32:56.791457
{ "authors": [ "jrfnl", "junior-paytech" ], "repo": "PHPCSStandards/PHP_CodeSniffer", "url": "https://github.com/PHPCSStandards/PHP_CodeSniffer/issues/618", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2036195110
Explaination on The YoloX Body-Head-Hand output dimension? Issue Type Documentation Feature Request OS Other OS architecture armv7 Programming Language Other Framework TensorFlowLite Model name and Weights/Checkpoints URL YoloX Body-Head-Hand https://github.com/PINTO0309/PINTO_model_zoo/tree/main/426_YOLOX-Body-Head-Hand Description First of all thanks and congratulation for your excellent work, on YOLO. I have a question regarding the output size, though. @PINTO0309 You say you limit the output to 20 boxes : How come we obtain a float32[60,7] output then? shouldn't we get something like 20x [class, score, x1, y1, x2, y2] = 20 x 5 floats? Relevant Log Output No response URL or source code for simple inference testing code No response As detailed in the README. max output boxes per class 20 (Body) + 20 (Head) + 20 (Hand) = 60 boxes
gharchive/issue
2023-12-11T17:41:51
2025-04-01T04:32:56.912400
{ "authors": [ "PINTO0309", "kho-bluefrogrobotics" ], "repo": "PINTO0309/PINTO_model_zoo", "url": "https://github.com/PINTO0309/PINTO_model_zoo/issues/381", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
561056495
[IIS] WinAcme stuck on "[DBUG] Scanning IIS sites" Bug Description Win Acme stuck on "[DBUG] Scanning IIS sites", won't renew it through task neither through the client. Expected behavior Win Acme should automatically renew certificates, as it did in the past. Log [VERB] Verbose mode logging enabled [VERB] Looking for settings.json in C:\win-acme [DBUG] Config folder: C:\ProgramData\win-acme\acme-v02.api.letsencrypt.org [DBUG] Log path: C:\ProgramData\win-acme\acme-v02.api.letsencrypt.org\Log [DBUG] Cache path: C:\ProgramData\win-acme\acme-v02.api.letsencrypt.org\Certificates [INFO] A simple Windows ACMEv2 client (WACS) [INFO] Software version 2.1.0.539 (RELEASE, PLUGGABLE) [INFO] IIS version 10.0 [INFO] Running with administrator credentials [INFO] Scheduled task looks healthy [INFO] Please report issues at https://github.com/PKISharp/win-acme [VERB] Test for international support: 語言 язык لغة [VERB] Verbose mode logging enabled [VERB] Arguments: --verbose [DBUG] Renewal period: 55 days N: Create new certificate (simple for IIS) M: Create new certificate (full options) L: List scheduled renewals R: Renew scheduled S: Renew specific A: Renew all O: More options... Q: Quit Please choose from the menu: s 1: [IISSite] suporte - renewed 1 time, due now 2: [IISSite] autenticador - renewed 2 times, due now 3: [IISSite] suporte_homologacao - renewed 3 times, due now : Back Which renewal would you like to run?: 1 [VERB] Sending e-mails True [DBUG] Scanning IIS sites Platform: OS: Windows Server 2019, English Version: 2.1.0.539 (RELEASE, PLUGGABLE) Additional context It's not generating errors in the Event Viewer or the win acme logs either. I'm not experienced in server administration so I'm kinda lost on where to look to try and debug this issue. Setting the --test flag also leaves me stuck with the same issue. I already tried leaving it on for hours and it simply won't do a thing. Do you have a lot of sites and/or a lot of bindings? I've got no idea why it would hang at that point without any feedback, unless IIS is unhealthy for some reason, but I assume all your sites are up and running? As a first step I would advise you to try the latest release. If that doesn't work then I may need to ask for a copy of your applicationHost.config to debug the issue. I upgraded to the latest version and got a little further on the process. It did scan my websites but the requests were not responding to the letsencrypt servers. I ended up fixing the issue by rebooting my machine, which is hosted at AWS. Thanks for the quick response!
gharchive/issue
2020-02-06T14:41:32
2025-04-01T04:32:56.933287
{ "authors": [ "Incognitowski", "WouterTinus" ], "repo": "PKISharp/win-acme", "url": "https://github.com/PKISharp/win-acme/issues/1401", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
408013074
Update YMap.js Добавил дополнительную опцию для возможности более подробной отладки ошибок. Спасибо!
gharchive/pull-request
2019-02-08T05:37:01
2025-04-01T04:32:56.949112
{ "authors": [ "PNKBizz", "VadimMalykhin" ], "repo": "PNKBizz/vue-yandex-map", "url": "https://github.com/PNKBizz/vue-yandex-map/pull/114", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1193820829
Add in AML specific signatures AML does not have standard deconvolution, instead there is a 17 gene stemness signature and some other subtypes https://pubmed.ncbi.nlm.nih.gov/27926740/ and https://www.sciencedirect.com/science/article/pii/S0092867419300947 Creating an issue to add these signatures to a single file and 'signature_matrices' code so we can deconvolve them./ This could be closed, but i'm still getting an xcell issue.
gharchive/issue
2022-04-05T23:00:11
2025-04-01T04:32:56.950951
{ "authors": [ "sgosline" ], "repo": "PNNL-CompBio/decomprolute", "url": "https://github.com/PNNL-CompBio/decomprolute/issues/186", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1748088919
Potential deprecation of concolor I'm posting this as this project is reported as a dependent of concolor and wanted to both give you a heads up to prepare and an opportunity to discuss in case we should re-evaluate this. We are proposing deprecating concolor in favor of either anstream application-specific solutions built on colorchoice, colorchoice-clap, and your choice of environment detection crates including anstyle-query and supports-color. We welcome feedback on this topic at rust-cli/concolor#47 Thank you! I commented on the linked issue.
gharchive/issue
2023-06-08T15:00:42
2025-04-01T04:32:57.107368
{ "authors": [ "epage", "max-sixty" ], "repo": "PRQL/prql", "url": "https://github.com/PRQL/prql/issues/2765", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1610173228
feat: upper and lower function Fixes #1793 This is great! Even the func interp now uses better naming (in my opinion)! Agree! Thanks @Jelenkee !
gharchive/pull-request
2023-03-05T12:03:56
2025-04-01T04:32:57.108586
{ "authors": [ "Jelenkee", "aljazerzen", "max-sixty" ], "repo": "PRQL/prql", "url": "https://github.com/PRQL/prql/pull/2019", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
608621121
Different endpoints to verify token/logout @ChetanKhanna Please merge similar end points for guest/verified user. This is done in 8e05d594f0dc497951262255aef17f4bccfb9eba
gharchive/issue
2020-04-28T21:15:32
2025-04-01T04:32:57.109399
{ "authors": [ "ChetanKhanna", "ritiktaneja" ], "repo": "PS-Division-BITS/Chat", "url": "https://github.com/PS-Division-BITS/Chat/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2673206909
k-space preconditioning for pdhg Accelerating Non-Cartesian MRI Reconstruction Convergence using k-space Preconditioning by Frank Ong, Martin Uecker, and Michael Lustig arxiv suggests to use pdhg with k-space preconditoning. For a diagonal precondition as proposed, this would in practice resulting the dual stepsize sigma being a tensor in our implementation. See also how to do their experiments https://github.com/mikgroup/kspace_precond/blob/master/Sense Reconstruction.ipynb -- they set sigpy's pdhg stepsize to the precondition We would also need their methodof obtaining the single- or multi-channel preconditioner. Accelerating Non-Cartesian MRI Reconstruction Convergence using k-space Preconditioning by Frank Ong, Martin Uecker, and Michael Lustig arxiv suggests to use pdhg with k-space preconditoning. For a diagonal precondition as proposed, this would in practice resulting the dual stepsize sigma being a tensor in our implementation. See also how to do their experiments https://github.com/mikgroup/kspace_precond/blob/master/Sense Reconstruction.ipynb -- they set sigpy's pdhg stepsize to the precondition We would also need their methodof obtaining the single- or multi-channel preconditioner. I agree that would be of high interest... for example, for our TV-unrolling, it would be nice to be able to only use a few iterations of PDHG and already have converged to a decent solution. This would make the approach applicable to non-Cartesian MRI as well, which we on purpose did not consider yet because of those reasons.
gharchive/issue
2024-11-19T18:55:34
2025-04-01T04:32:57.129597
{ "authors": [ "fzimmermann89", "koflera" ], "repo": "PTB-MR/mrpro", "url": "https://github.com/PTB-MR/mrpro/issues/545", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1228727633
Cambiar_rol/hay cambios en login(actualiza) pequeños cambios en el login(puede ser) y opcion de cambiar rol del trabajador en pagina de trabajadores donde solo el admin tendra acceso. No entiendo muy bien este cambio, ya tenemos una manera de modificar los datos del trabajador cuando accedamos a el. Los cambios que debías hacer eran en la pagina del trabajador, comprobando el rol del usuario que quiere cambiar la info. No se a que tarjeta del trello hace referencia esto pero no me suena
gharchive/pull-request
2022-05-07T21:31:36
2025-04-01T04:32:57.130892
{ "authors": [ "Ar1e1", "alfman99" ], "repo": "PTIN2022/GeSyS-Front", "url": "https://github.com/PTIN2022/GeSyS-Front/pull/44", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
212351903
How to remove lsm303d_accel driver I have a problem with my pixhawk. My pixhawk lsm303d_accel sensor is dead. So I try to remove lsm303d accel driver . But I fail to remove. Can you help me? How to remove lsm303d accel driver? We will not help support broken hardware setups. @winszenYoung It will automatically not start if its broken. However, if the sensor is not really dead but half-working, you should consider retiring this board. Its not safe to fly.
gharchive/issue
2017-03-07T07:29:22
2025-04-01T04:32:57.154618
{ "authors": [ "LorenzMeier", "mhkabir", "winszenYoung" ], "repo": "PX4/Firmware", "url": "https://github.com/PX4/Firmware/issues/6734", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
682641205
Fix Mavros Mission tests for VTOL Describe problem solved by this pull request Recently Mavros Mission Tests have been failing for VTOLs in Firmware. Describe your solution Update the container from ROS melodic to ROS noetic to run the tests as the same environment as the rest of the tests Failure: [ERROR] [1597924035.046962627]: [registerPublisher] Failed to contact master at [7f9311d946d5:44367]. Retrying... Traceback (most recent call last): File "/opt/ros/noetic/lib/gazebo_ros/spawn_model", line 20, in <module> import rospy File "/opt/ros/noetic/lib/python3/dist-packages/rospy/__init__.py", line 49, in <module> from .client import spin, myargv, init_node, \ File "/opt/ros/noetic/lib/python3/dist-packages/rospy/client.py", line 52, in <module> import roslib File "<frozen importlib._bootstrap>", line 991, in _find_and_load File "<frozen importlib._bootstrap>", line 971, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 906, in _find_spec KeyboardInterrupt @TSC21 Updating to noetic might not be a good idea, since we don't have catkin on focal working yet any progress on this end @Jaeyoung-Lim ?
gharchive/pull-request
2020-08-20T11:31:53
2025-04-01T04:32:57.157016
{ "authors": [ "Jaeyoung-Lim", "mrpollo" ], "repo": "PX4/Firmware", "url": "https://github.com/PX4/Firmware/pull/15587", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
120186583
Make posix sitl work on my mac @LorenzMeier please double check my PX4_ISFINITE changes and va_arg type change. I'm not sure if the va_arg change breaks it for others. @ SITL-CI test this please This is all looking good.
gharchive/pull-request
2015-12-03T14:44:32
2025-04-01T04:32:57.158268
{ "authors": [ "AndreasAntener", "LorenzMeier" ], "repo": "PX4/Firmware", "url": "https://github.com/PX4/Firmware/pull/3313", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
203459286
Remove the notion of legacy from the api and save some FLASH space too @LorenzMeier This remove the notion of legacy from the api. The board level code will perform the translation to legacy format on the STM32. New targets will not need to do this as there is no case where the serial number were used by mfg for tracking. It maintains the legacy incorrect selection of significant digits of the UUID in the PX4 code base. This is done to avoid the ripple effects changing the IDs used on STM32 base platforms. [See] (https://github.com/PX4/Firmware/commit/f126668dbd41f10667a8759be7937dd822549828#diff-8da0cce1d272772f52e0d92c6f4343a5R83) Jenkins test this please. From your comment I gather you only kept the display of legacy. If that conclusion would be correct: However, the byte order is also inverted in storage IIRC and via MAVLink. Please compare the UUID as sent via MAVLInk in the MAVLink inspector of QGroundControl. It should stay the same for STM32 as it will blow up existing configs. If you kept the representation for STM32 then this looks like its ready to go. Jenkins test this please. 1.5.0 Release RC1 nsh> ver all HW arch: PX4FMU_V2 FW git-hash: 3d7906647e5c6f382bf6c574927bac88d752874f FW version: v1.4.4rc1 (1.4.4 192), 17040576 OS version: 6.27 (4294967295) Build datetime: Aug 26 2016 09:41:20 Toolchain: 4.8.4 20140725 (release) [ARM/embedded-4_8-branch revision 213147] MCU: STM32F42x, rev. 3 UID: 400038:33355110:32383336 nsh> fmu id ERROR [fmu] Board serial: 00400038 33355110 32383336 Current master nsh> ver all HW arch: PX4FMU_V2 FW git-hash: 264589b2cc8bc69e95fc63580e2ad94be109c016 FW version: 1.5.2 0 (17105408) OS: NuttX OS version: Release 1.8.0 (17301759) Build datetime: Jan 28 2017 05:19:48 Build uri: localhost Toolchain: GNU GCC, 5.4.1 20160609 (release) [ARM/embedded-5-branch revision 237715] MCU: STM32F42x, rev. 3 UID: 400038:33355110:32383336 nsh> fmu id Board serial: 400038 33355110 32383336 nsh> This Branch nsh> ver all HW arch: PX4FMU_V2 FW git-hash: b1b8e9b87a96b0b032237e31bb91f4e182a86d69 FW version: 1.5.2 0 (17105408) OS: NuttX OS version: Release 1.8.0 (17301759) Build datetime: Jan 28 2017 05:49:58 Build uri: localhost Toolchain: GNU GCC, 5.4.1 20160609 (release) [ARM/embedded-5-branch revision 237715] MCU: STM32F42x, rev. 3 UID: 400038:33355110:32383336 nsh> fmu id Board serial: 00400038 33355110 32383336 nsh> WARN [commander] Not ready to fly: Sensors not set up correctly @LorenzMeier - Please DNM yet - I have more commits to add here.... This branch post discussion nsh> ver all HW arch: PX4FMU_V2 FW git-hash: 256851bc27a9445e5fb24e5a0ec3516cc34d37d6 FW version: 1.5.2 0 (17105408) OS: NuttX OS version: Release 1.8.0 (17301759) Build datetime: Jan 28 2017 08:20:45 Build uri: localhost Toolchain: GNU GCC, 5.4.1 20160609 (release) [ARM/embedded-5-branch revision 237715] MFGUID: 323833363335511000400038 MCU: STM32F42x, rev. 3 UID: 400038:33355110:32383336 nsh> fmu id FMU: unrecognized command id, try: mode_gpio, mode_pwm, mode_pwm4, test, sensor_reset [milliseconds], i2c <bus> <hz>, bind @LorenzMeier - ready! CI catched some code issues: /Users/travis/build/PX4/Firmware/src/systemcmds/ver/ver.c:60:19: fatal error: variable 'mfg_uid_str' is not needed and will not be emitted [-Wunneeded-internal-declaration] static const char mfg_uid_str[] = "mfguid"; ^ 1 error generated. @LorenzMeier Now that we added the mfguid and display in in ver I cleaned up the uavcan usage and removed all the exposed reordering from the API and include files. The stm32 board code only needs to do the maintain the legacy format. I have tested this on px4fmu-v3 and a new 2.4.6 board w/ rev 3 silicon as well as the esc35-v1 hw bootloader. Node id allocation is AOK with these changes. nsh> ver all HW arch: PX4FMU_V2 FW git-hash: 3fe5c4f2372298bf1c44fa655e344d632a31afcd FW version: 1.6.0 0 (17170432) OS: NuttX OS version: Release 1.8.0 (17301759) Build datetime: Jan 30 2017 13:23:35 Build uri: localhost Toolchain: GNU GCC, 5.4.1 20160609 (release) [ARM/embedded-5-branch revision 237715] MFGUID: 323833363335511000400038 MCU: STM32F42x, rev. 3 UID: 400038:33355110:32383336 nsh> ver uid UID: 400038:33355110:32383336 nsh> ver mfguid MFGUID: 323833363335511000400038 @LorenzMeier Please give this a second look. @LorenzMeier Architectural question. Do we want to pad out a 96 bit mfg to 128 bit with leading 0's. The we can use 32 character string in the logging and add the to QGC as an mfguid[32]? Rebased and force-pushed. @davids5 Padding is fine. I still will merge this once it passes.
gharchive/pull-request
2017-01-26T19:04:31
2025-04-01T04:32:57.168270
{ "authors": [ "LorenzMeier", "davids5" ], "repo": "PX4/Firmware", "url": "https://github.com/PX4/Firmware/pull/6441", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
506470253
Enhancement: Increased robustness against 0 signal quality and cleanup range data handling This PR attempts to cleanup the checks on the range finder data. Additionally it adds a hysteresis to declaring range finder data valid after signal quality was zero. Specifically: whenever signal quality of the range data was 0 we require non-zero signal quality for 3 seconds before we accept the data to be used. This is to improve robustness against a known sensor failure mode where the signal quality toggles between 0 and non-zero values (see plot below) moved range finder data checks to a central location and created new file range_finder_checks.cpp in order to avoid loading control.cpp replaced _rng_hgt_faulty with _range_hgt_valid in order to avoid negation all over the place use better function name moved updateRangeDataContinuity() out of the terrain estimator and into range_finder_checks.cpp I will provide SITL test data shortly. Plot of range finder failure mode: Notice how signal quality erratically jumps between 0 and 95% in the middle of the flight. The height reported by the sensor at that point was completely wrong. Signed-off-by: RomanBapst bapstroman@gmail.com The 3 seconds hold-off sounds like it is engineered for a specific sensor - can this be made a parameter and be set for that sensor? Plot demonstrating 3 second hysteresis after signal quality was 0: Log: https://mail.google.com/mail/u/0/#inbox/FMfcgxwDrlVmwVjrGGZSjKnTPrbflWXr Plot demonstrating a height reset to range finder (simulated a step in range finder data of 5m): Log: https://logs.px4.io/plot_app?log=d3182c9d-01e2-4d15-940f-00b7d2190af2 @MaEtUgR I addressed your comments, they were good suggestions. Please have another look. @bresch Sorry I meant you, not @MaEtUgR @priseborough Could you please have a look as well? Thanks @RomanBapst
gharchive/pull-request
2019-10-14T06:30:24
2025-04-01T04:32:57.195960
{ "authors": [ "LorenzMeier", "RomanBapst", "priseborough" ], "repo": "PX4/ecl", "url": "https://github.com/PX4/ecl/pull/655", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1302120674
Can "integrated charts" be used with streamlit-aggrid? I can see from the showcase example that e.g. groups and pivot-tables work. Before buying a license, I would like to know if the enterprise "inegrated charts" feature will work when AGGrid is used via Streamlit. same request here with Tree Data feature of aggrid enterprise Same here. Any plans to support live streaming updates, and integrated charting. This will be amazing. I hope there is a way for you to give us a referral link where you get something whenever one of us buys an enterprise license from ag-grid. Same here. Any plans to support live streaming updates, and integrated charting. This will be amazing. I hope there is a way for you to give us a referral link where you get something whenever one of us buys an enterprise license from ag-grid. I've solved this problem, later this week I'll post my code here. Thanks @littlepanda0716 and look forward to your solution. `import streamlit as st from st_aggrid import AgGrid, GridOptionsBuilder, JsCode import pandas as pd if name == "main": df = pd.DataFrame({"Name": ["A", "B", "C", "D"], "Age": ["80.0%", "0.0%", "20.0%", "None"]}) st.write(df) gb = GridOptionsBuilder.from_dataframe(df) gridOptions = gb.build() gridOptions["columnDefs"][1] = {"field": "Age", "cellRenderer": "agSparklineCellRenderer", 'valueGetter': 'JSON.parse("["+data.Age.replace("%","]"))', "cellRendererParams": { "sparklineOptions": { "type": 'bar', "fill": '#4fa2d9', "valueAxisDomain": [0.0, 100.0], "label": { "enabled": True, "color": 'black', "fontSize": 10, "fontWeight": 'bold', "formatter": JsCode("""function(params){ return `${params.value}%`;}"""), "placement": "insideBase", }, "tooltip": { "enabled": False, }, } } } r = AgGrid( df, gridOptions=gridOptions, height=500, allow_unsafe_jscode=True, enable_enterprise_modules=True, filter=True, # update_mode=GridUpdateMode.SELECTION_CHANGED, theme="material", )` import streamlit as st from st_aggrid import AgGrid, GridOptionsBuilder, JsCode import pandas as pd if __name__ == "__main__": df = pd.DataFrame({"Name": ["A", "B", "C", "D"], "Age": ["80.0%", "0.0%", "20.0%", "None"]}) st.write(df) gb = GridOptionsBuilder.from_dataframe(df) gridOptions = gb.build() gridOptions["columnDefs"][1] = {"field": "Age", "cellRenderer": "agSparklineCellRenderer", 'valueGetter': 'JSON.parse("["+data.Age.replace("%","]"))', "cellRendererParams": { "sparklineOptions": { "type": 'bar', "fill": '#4fa2d9', "valueAxisDomain": [0.0, 100.0], "label": { "enabled": True, "color": 'black', "fontSize": 10, "fontWeight": 'bold', "formatter": JsCode("""function(params){ return `${params.value}%`;}"""), "placement": "insideBase", }, "tooltip": { "enabled": False, }, } } } r = AgGrid( df, gridOptions=gridOptions, height=500, allow_unsafe_jscode=True, enable_enterprise_modules=True, filter=True, # update_mode=GridUpdateMode.SELECTION_CHANGED, theme="material", ) Hope this may help you @mzeidhassan Thanks a million @littlepanda0716 ! This is very nice of you. I appreciate it.
gharchive/issue
2022-07-12T14:08:55
2025-04-01T04:32:57.204607
{ "authors": [ "fluence-world", "littlepanda0716", "mzeidhassan" ], "repo": "PablocFonseca/streamlit-aggrid", "url": "https://github.com/PablocFonseca/streamlit-aggrid/issues/110", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
308887058
unzip uses too much memories Dear developer, We found that unzip module 'rr_ctg_track.py' try to read all .las files into memory and we had around 20T .las files. It's hard to find a machine with so large memories. Do you have any suggestion to avoid loading all data into memories? 3ku! This issue is created by me and is the same problem. You can get more information from it (https://github.com/marbl/canu/issues/838). The memory is probably not consumed by rr_ctg_track.py directly. That program spawns LA4Falcon for each .las file. You will have a number of instances of LA4Falcon running equal to your --n-core argument. (And each will be under a different sub-process, so the memory used by rr_ctg_track will be cloned. That's probably not an problem, but you can look at the forked python procs on your machine.) Each LA4Falcon loads the entire DAZZLER DB, which is probably your problem. (Look at the file 0-rawreads/.raw_reads.bps) There are 2 solutions: Hack our code to load the DB from dev/shm. (Non-trivial, but one user has done this.) Use --n-core=0. (Same as 1, but simpler, since it avoids the whole "multiprocessing" module.) You can experiment with various values of --n-core. Also, your unzip might be out-of-date. You could try the Falcon-unzip binary tarball, as the GitHub code is not up-to-date. https://github.com/PacificBiosciences/FALCON_unzip/wiki/Binaries The explaination is really clear and I believe the solution you provided must be useful. I'll try as you said. 3ku.
gharchive/issue
2018-03-27T09:19:41
2025-04-01T04:32:57.210133
{ "authors": [ "nottwy", "pb-cdunn" ], "repo": "PacificBiosciences/FALCON_unzip", "url": "https://github.com/PacificBiosciences/FALCON_unzip/issues/111", "license": "BSD-3-Clause-Clear", "license_type": "permissive", "license_source": "github-api" }
101098721
more pbsmrtpipe integration The most important part (which I forgot in the commit messages) is using ZMW ranges for chunked datasets. Hold on, I may need that index file generation after all... Can we avoid outputting that index? By restricting the number of interdependencies and separating functionally distinct components software is much easier to write, read, debug and maintain. Yeah, we really need a central, general solution for this problem. And I think I figured out the general solution, using a decorator defined by pbcommnad should I go ahead and merge then? yes, the ZMW range is critical
gharchive/pull-request
2015-08-14T20:35:45
2025-04-01T04:32:57.212279
{ "authors": [ "evolvedmicrobe", "natechols" ], "repo": "PacificBiosciences/pbccs", "url": "https://github.com/PacificBiosciences/pbccs/pull/53", "license": "BSD-3-Clause-Clear", "license_type": "permissive", "license_source": "github-api" }
307222428
getting the loss of a audio file with a certain transcription Any chance to get a guidance on hoe to receive the ctc_loss of a (audio,transcrption) pair via the python wrapper? Thanks in advance! Please refer to this method https://github.com/PaddlePaddle/DeepSpeech/blob/dad1c2727e375beb05dd747b06b04d0438e4a78d/model_utils/model.py#L158-L174 It tried it already...do you have any documentation or example how to use this function.. any combination I tried failed, like: infer_data=[(speech,txt)] or [speech,txt] or (speech,txt) with different errors or dumps.. On Thu, Mar 22, 2018 at 8:53 AM, Yibing Liu notifications@github.com wrote: Please refer to this method https://github.com/PaddlePaddle/DeepSpeech/blob/ dad1c2727e375beb05dd747b06b04d0438e4a78d/model_utils/model.py#L158-L174 — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/PaddlePaddle/DeepSpeech/issues/186#issuecomment-375197523, or mute the thread https://github.com/notifications/unsubscribe-auth/AE_DSVEVfIXKrCSyea5aEHzgDXT3m_Ouks5tg0pbgaJpZM4SzidE . @alanbekker Have you figured it out ? I incurred the same problem, looking forward your reply, thanks. I have figured it out , just set the DataGenerator's keep_transcription_text to False and make the same preprocess as in infer_batch_probs(): data_generator = DataGenerator( vocab_filepath=args.vocab_path, mean_std_filepath=args.mean_std_path, augmentation_config='{}', specgram_type=args.specgram_type, num_threads=8, keep_transcription_text=False)
gharchive/issue
2018-03-21T12:21:17
2025-04-01T04:32:57.236629
{ "authors": [ "alanbekker", "hui001", "kuke" ], "repo": "PaddlePaddle/DeepSpeech", "url": "https://github.com/PaddlePaddle/DeepSpeech/issues/186", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1182271042
模型预测或模型推理的结果,没有保存的TXT文件。 PaddleClas/tools/infer.py PaddleCals/deploy/python/predict_cls.py 预测或推理路径下的所有图像,结果没有保存的txt文件。 目前尚未提供该功能,如有需要,可以修改源码,或是在执行推理命令时将结果重定向到指定文件中,再对结果进行处理: python python/predict_cls.py -c configs/inference_cls.yaml > output.txt Example import ast import re with open("output.txt") as f: ftxt = f.read() pattern = r"{'class_ids':.*?}" pre_results = re.findall(pattern, ftxt) print(len(pre_results)) pre_data = [] for pre_result in pre_results: pre_dict = ast.literal_eval(pre_result) # ["class_ids", "scores", "file_name", "label_names"] pre_data.append([pre_dict["file_name"].split("/")[-1],\ pre_dict["class_ids"][0], pre_dict["scores"][0]]) columns = ["img_path", "label", "scores"] pre_df = pd.DataFrame(data=pre_data, index=range(len(pre_data)), columns=columns) pre_df.drop(columns=['scores'],inplace=True) # 删除 scores 列 pre_df.sort_values(["img_path"], inplace=True) # 排序 pre_df.to_csv('submission.csv',index=None,header=None)
gharchive/issue
2022-03-27T02:36:03
2025-04-01T04:32:57.300225
{ "authors": [ "CPones", "TingquanGao", "Wulx2050" ], "repo": "PaddlePaddle/PaddleClas", "url": "https://github.com/PaddlePaddle/PaddleClas/issues/1788", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2173033299
[Question]: UIE信息抽取做 关系抽取任务,出现幻想,文本中没有出现的信息,会出现幻觉 请提出你的问题 调用的是paddlepaddle uie-base 模型进行推理,配置关系如下,及其结果如下,有好的解决方案吗? 这个是模型效果的问题,需要通过UIE微调方式来提升效果。 微调 通过增加负样例 依然还是有这样的问题,不觉得是负样例的问题,猜测可能和uie本身预测的结构有关系 您好,UIE相关的技术已经停止更新了,欢迎使用LLM来解决您的业务问题
gharchive/issue
2024-03-07T06:10:32
2025-04-01T04:32:57.321165
{ "authors": [ "827648313", "w5688414", "wawltor" ], "repo": "PaddlePaddle/PaddleNLP", "url": "https://github.com/PaddlePaddle/PaddleNLP/issues/8068", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1758099051
[Performance optimization] add to_static feature for all generative model PR types Performance optimization PR changes APIs Description 完善 generation_utils 中的生成相关逻辑。 [x] 消除静态图子图 [x] 优化生成性能 [x] 添加动转静相关单测 可以在下一个 pr 中将 动转静的相关逻辑添加到基类当中去,这样其他模型也可以快速添加对应生成配置。
gharchive/pull-request
2023-06-15T06:01:59
2025-04-01T04:32:57.323657
{ "authors": [ "wj-Mcat" ], "repo": "PaddlePaddle/PaddleNLP", "url": "https://github.com/PaddlePaddle/PaddleNLP/pull/6186", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2262465511
add Pose and YaRN PR types New features PR changes Models Description Added the Pose algorithm, including its data processing and training scripts. Added YaRN, one of the RoPE extension methods. Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.whf seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it. cd PaddleNLP pre-commit install commit之前执行,自动修改格式问题
gharchive/pull-request
2024-04-25T01:37:06
2025-04-01T04:32:57.327730
{ "authors": [ "CLAassistant", "lugimzzz", "whf313" ], "repo": "PaddlePaddle/PaddleNLP", "url": "https://github.com/PaddlePaddle/PaddleNLP/pull/8327", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1039609131
The third-party dynamic library (mklml.dll) that Paddle depends on is not configured 环境 请提供您使用的PaddlePaddle和PaddleX的版本号 paddlepaddle-gpu 2.1.3.post112 paddleslim 2.1.1 paddlex 2.0.0 请提供您使用的操作系统信息,如Linux/Windows/MacOS WIN10 请问您使用的Python版本是? python 3.7 请问您使用的CUDA/cuDNN的版本号是? cuda11.2 cudnn8 2021-10-29 19:06:26 [INFO] Model[FasterRCNN] loaded. Exception in thread CAM_BENDI_00: Traceback (most recent call last): File "C:\Users\hello\threading.py", line 932, in _bootstrap_inner File "C:\Users\hello\threading.py", line 870, in run File "C:\Users\hello\hello_class.py", line 4999, in Work_threadai_0 File "C:\Users\hello\paddlex\deploy.py", line 64, in __init__ self.predictor = self.create_predictor( File "C:\Users\hello\paddlex\deploy.py", line 126, in create_predictor predictor = create_predictor(config) RuntimeError: (PreconditionNotMet) The third-party dynamic library (mklml.dll) that Paddle depends on is not configured correctly. (error code is 182) Suggestions: 1. Check if the third-party dynamic library (e.g. CUDA, CUDNN) is installed correctly and its version is matched with paddlepaddle you installed. 2. Configure third-party dynamic library environment variables as follows: - Linux: set LD_LIBRARY_PATH by `export LD_LIBRARY_PATH=...` - Windows: set PATH by `set PATH=XXX; (at C:\home\workspace\Paddle_release3\paddle\fluid\platform\dynload\dynamic_loader.cc:265) 用户安装了线描相机的软件和驱动之后 paddle就报错 python环境是这样的问题 之前打包好独立的EXE环境 也是一样的错误 conda新建环境不行 重新安装cuda+cudnn 也不行 @monkeycc 你好,请问你这个问题解决了么? 我是使用multiprocessing模块时,遇到了这个报错。。用threading模块就不会。。感觉很迷糊。。---我是用paddlepaddle库做文本检测的,没有用到paddleX...我只是想请教下,你是怎么解决这个问题的。。 常规情况下,我使用GPU是没有报错的。。 @lmw0320 直接把线描相机的软件和驱动 删除就好了 CPU加速的模块 有冲突 multiprocessing模块时,遇到了这个报错。。用线程模块就不会 multiprocessing的时候,初始化模型放在主程序,预测放在multiprocessing进程 没太明白您的意思: 要删除线描相机的软件和驱动?? 哪个才是??我貌似没安装过啊?(win10企业版) 我的线程和进程操作,都是你说的那样,初始化模型放在主程序,然后把这个变量传入具体的进程函数中。进程函数内进行具体的预测操作。 paddlepaddle验证通过, Running verify PaddlePaddle program ... W1208 13:21:04.446607 14284 gpu_resources.cc:61] Please NOTE: device: 0, GPU Compute Capability: 8.6, Driver API Version: 12.0, Runtime API Version: 11.7 W1208 13:21:04.450608 14284 gpu_resources.cc:91] device: 0, cuDNN Version: 8.4. PaddlePaddle works well on 1 GPU. PaddlePaddle works well on 1 GPUs. 但运行Paddleocr的时候报错, RuntimeError: (PreconditionNotMet) The third-party dynamic library (mklml.dll) that Paddle depends on is not configured correctly.(error code is 126) 没有多进程,没有多线程。 paddlepaddle-gpu==2.4.0.post117 paddleocr 2.6.1.1 您好,请问您解决了吗? 应该是某个软件 增加了系统环境 可以删除一部分系统环境设置 排除一下 https://learn.microsoft.com/ja-jp/cognitive-toolkit/setup-mkl-on-windows downloading from here worked for me
gharchive/issue
2021-10-29T14:09:27
2025-04-01T04:32:57.382090
{ "authors": [ "Heryk13", "MiShiDeHaiLuo", "changmaoguai", "lmw0320", "monkeycc" ], "repo": "PaddlePaddle/PaddleX", "url": "https://github.com/PaddlePaddle/PaddleX/issues/1232", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
973314566
support 3 types model support 3 types model LGTM
gharchive/pull-request
2021-08-18T06:15:30
2025-04-01T04:32:57.383304
{ "authors": [ "HexToString" ], "repo": "PaddlePaddle/Serving", "url": "https://github.com/PaddlePaddle/Serving/pull/1354", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1201332751
paddle.optimizer.lr.LRScheduler 文档中英文链接 https://www.paddlepaddle.org.cn/documentation/docs/en/api/paddle/optimizer/lr/LRScheduler_en.html https://www.paddlepaddle.org.cn/documentation/docs/zh/api/paddle/optimizer/lr/LRScheduler_cn.html 红线问题 https://www.paddlepaddle.org.cn/documentation/docs/en/api/paddle/optimizer/lr/LRScheduler_en.html#:~:text=the schedulers state.-,get_lr(),-For those subclass 英文部分没有代码示例,中文有 结构问题 No response 内容问题 中文部分 https://www.paddlepaddle.org.cn/documentation/docs/zh/api/paddle/optimizer/lr/LRScheduler_cn.html#:~:text=请参考 cn_api_paddle_optimizer_lr_PolynomialDecay 未给出链接 https://www.paddlepaddle.org.cn/documentation/docs/zh/api/paddle/optimizer/lr/LRScheduler_cn.html#:~:text=请参考-,基类LRScheduler,-的任意子 格式问题,代码和汉字混在一起了 https://www.paddlepaddle.org.cn/documentation/docs/zh/api/paddle/optimizer/lr/LRScheduler_cn.html#:~:text=学习率策略,-,导出基类的方法为,-form paddle.optimizer 应该为引用而非导出 英文 https://www.paddlepaddle.org.cn/documentation/docs/en/api/paddle/optimizer/lr/LRScheduler_en.html#:~:text=Can be set to restart training 缺乏主语 https://www.paddlepaddle.org.cn/documentation/docs/en/api/paddle/optimizer/lr/LRScheduler_en.html#:~:text=“last_epoch%2C last_lr” 引用代码格式问题 https://www.paddlepaddle.org.cn/documentation/docs/en/api/paddle/optimizer/lr/LRScheduler_en.html#:~:text=set_dict(state_dict)-,Loads the schedulers state.,-get_lr() 应该是loading吧?? 格式问题 No response 主观评价 总体OK,英文缺失部分代码示例 请: 1、修复 PolynomialDecay: 多项式衰减,即得到的学习率为初始学习率和给定最终学习之间由多项式计算权重定比分点的插值。请参考 cn_api_paddle_optimizer_lr_PolynomialDecay 中链接问题 2、修改 基类LRScheduler 的格式问题 3、导出基类的方法为 -> 引用基类的方法为
gharchive/issue
2022-04-12T07:48:28
2025-04-01T04:32:57.390943
{ "authors": [ "BeyondYourself", "TCChenlong" ], "repo": "PaddlePaddle/docs", "url": "https://github.com/PaddlePaddle/docs/issues/4620", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1382897089
Add Solace bindings and extension fields Add Solace bindings and extension fields The purpose of this change is to allow the parser to handle solace bindings (new) and extension fields of the form x-my-extension-field in the spec files Changes Added capability to handle extension fields in objects as properties lists See com.asyncapi.v2.model.info.Info.java for an example any unknown field will attempt to be added as a string property e.g. x-ep-my-property There is no validation on the field names: yy-my-property will also be accepted added "solace" bindings com.asyncapi.v2.binding.solace for examples Project config changes Includes update to Java11 It can be downgraded back to Java 8, but would need to be re-tested Updated all pom versions to 1.0.0-EAP-1 -> 1.0.1-EAP-1 Updated java.version 8 -> 11 Updated jackson.version 2.13.2 -> 2.13.3 kotlin.version 1.6.20 -> 1.7.10 dokka.version 0.10.1 -> 1.7.0 maven-source-plugin 2.2.1 -> 3.2.1 maven-javadoc-plugin 2.9.1 -> 3.3.2 ToDo: Dependency stanza in README.md files should be updated to SemVer 1.0.1-EAP-1 Class SchemaGeneratorMojo not found Hi @dennis-brinley, Thanks for your awesome PR. Will merge it in this month
gharchive/pull-request
2022-09-22T19:11:20
2025-04-01T04:32:57.421033
{ "authors": [ "Pakisan", "dennis-brinley" ], "repo": "Pakisan/jasyncapi", "url": "https://github.com/Pakisan/jasyncapi/pull/6", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
271406957
点击浏览器刷新按钮后页面空白什么都没有 store.dispatch('GenerateRoutes', store.getters.roles).then(() => { console.log('路由表', store.getters.addRouters) router.addRoutes(store.getters.addRouters) // 动态添加可访问的路由 // next({...to}) // hack方法,确保addRutes已完成 next() // 直接使用该方法刷新页面后不会保留在历史页,而页面是空白的 }) 如果直接使用next(),在页面中刷新后页面什么都没有,只有一个#app跟节点. 如果使用next({...to})的话页面就正常,刷新后还是在历史页面, 虽然文档写的hack方法,确保addRutes已完成,但是这儿还是不要理解,麻烦楼主详细说一下好么,非常感谢 主要是 addRoutes 没有callback function 所以 你不能确定你 next() 之后进去的页面已成功被 addRoutes 添加。 而next(xxx) 会重新再次进入 router.beforeEach 这个钩子 这时候 addRoutes 已success 所以能直接 next()
gharchive/issue
2017-11-06T09:23:32
2025-04-01T04:32:57.445768
{ "authors": [ "PanJiaChen", "wxxtqk" ], "repo": "PanJiaChen/vue-element-admin", "url": "https://github.com/PanJiaChen/vue-element-admin/issues/276", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
532468541
关于右侧窗体 Question(提问) 关于右侧窗体,我想请教下,在该窗体里面的页面中如果有下拉框,弹出消息等,都会被该窗体遮挡屏蔽,是否有办法解决? 谢谢! @PanJiaChen 解决,设置下z-index
gharchive/issue
2019-12-04T05:59:21
2025-04-01T04:32:57.447064
{ "authors": [ "moonose" ], "repo": "PanJiaChen/vue-element-admin", "url": "https://github.com/PanJiaChen/vue-element-admin/issues/2806", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
694780537
当点击 tagsView 右键菜单 '关闭其他' 或者 '关闭所有' 选项时,功能虽然正常,但是会触发设置了 affix: true 的固定标签刷新。 Bug report(问题描述) 当点击 tagsView 右键 '关闭其他' 或者 '关闭所有' 选项时,功能虽然正常,但是会触发设置了 affix: true 的固定标签刷新。 Steps to reproduce(问题复现步骤) 设置 src/router/modules/table.js 中的 drag-table 路由为固定标签: { path: 'drag-table', component: () => import('@/views/table/drag-table'), name: 'DragTable', meta: { title: 'Drag Table', affix: true } } npm run dev 运行项目 浏览器打开项目,随便打开几个未固定标签,然后点击未固定标签的右键菜单中的 '关闭其他' 或者 '关闭所有' 选项, 这时未固定标签会关闭,但是再次点击 drag-table 标签(当选择的是 ‘关闭所有’ 选项时不需要点击 drag-table 标签)就会刷新 Screenshot or Gif(截图或动态图) Other relevant information(格外信息) Your OS: win7 x64 Node.js version: v12.16.3 vue-element-admin version: 4.4.0 已自行修改代码解决
gharchive/issue
2020-09-07T07:05:44
2025-04-01T04:32:57.451094
{ "authors": [ "leffss" ], "repo": "PanJiaChen/vue-element-admin", "url": "https://github.com/PanJiaChen/vue-element-admin/issues/3416", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
363649976
Nombre de Tablas Cambiar nombre de las tablas a minúsculas Hecho
gharchive/issue
2018-09-25T15:54:24
2025-04-01T04:32:57.453361
{ "authors": [ "JoseEmmanuelGC", "Mgmez" ], "repo": "Pandiurex/Foraneos-UDG-Api", "url": "https://github.com/Pandiurex/Foraneos-UDG-Api/issues/36", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
172560001
Images resizing not correct Questions Expected Behavior When I put images in the timeline, I'd like to see them resized to fit the circles. Actual Behavior If the images are not sized well enough, you see ellipsoidal images too much inside the circles or too much outside. Steps to Reproduce the Issue Put new images in the /img/timeline/ folder Change _config.yml jekyll serve Display the timeline Specifications Every version of the browser, with { Personal } up to date I've got two ways to solve this issue: The easy one: modifying the CSS so that the image fits whatever the original dimensions are. You can set the images of the timeline with these properties: width: 100%; height: 100%; Creating a "generate timeline images" script which resizes the images correctly, why not creating new images, "image.png" giving "timeline_image.png". I'll work on this issue, so I'm open to any comment. The images should be square in order to be rendered correctly, but I like that you're trying to tackle that problem. Did you come up with a good enough solution?
gharchive/issue
2016-08-22T21:30:39
2025-04-01T04:32:57.457717
{ "authors": [ "PanosSakkos", "djcaesar9114" ], "repo": "PanosSakkos/personal-jekyll-theme", "url": "https://github.com/PanosSakkos/personal-jekyll-theme/issues/193", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
239465233
Rework the help in order to add more detailled description Since the description of every commands is displayed when using f!help category name, descriptions must be short. However there is already few commands with arguments which might need a more detailled description, so @ParadoxOrigins , add a detailled description and a short one that is done
gharchive/issue
2017-06-29T12:40:23
2025-04-01T04:32:57.551112
{ "authors": [ "ParadoxOrigins" ], "repo": "ParadoxOrigins/FelixBot", "url": "https://github.com/ParadoxOrigins/FelixBot/issues/23", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
217202950
Revisit data I'd like to be able to revisit my results. Links #5 #6 #10 #14 closed via #81
gharchive/issue
2017-03-27T10:32:54
2025-04-01T04:32:57.562705
{ "authors": [ "ClairePUK", "ivanmauricio" ], "repo": "ParkisonsUK-fuse/what3things", "url": "https://github.com/ParkisonsUK-fuse/what3things/issues/17", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
158181931
Can we use serverURL with parse/node? I am trying to migrate my application using a self hosted parse-server setup. I have some test scripts that I run in NodeJS using the parse/node module. However when I specify a server URL, I get an error trying to connect to the parse server. does the parse/node module support serverURL? I am trying this: Parse.initialize('API KEY WOULD GO HERE'); var serverURL = 'https://some.server.url/parse'; Parse.serverURL = serverURL; and I get this error: ParseError { code: 100, message: 'XMLHttpRequest failed: "Unable to connect to the Parse API"' } It's supported. That's how all of the integration tests run, so I can guarantee that hasn't broken. I'd make sure that your server running your node app can actually see your parse server, and that there are no networking complications, but I doubt there's an SDK bug here. @andrewimm Thanks, that's what I figured also. I have a web application running, which can see the parse server and the browser/client can access the parse server at the same URI, so based on that I do believe it is reachable. I will dig in some more to see what else might be going on. Well, good luck digging into this. I'd suggest trying to build a simple node script and seeing if it can make requests to your server using the https module. There might be some crazy permissions mismatch going on. here is what I discovered... if I set the serverURL to http instead of https, I can connect the the remote parse-server. Why? In the browser, I point to https and it works fine. The web service is deployed on AWS, sitting behind a load balancer. The interesting thing is that in the parse-server instance on the server, I use http://localhost/parse as the serverURL. The load balancer handles https termination I guess, and passes the traffic over http internally? Atleast that is what I think is going on... but again, when I point to https from the browser, all good. Any ideas here? You mentioned making requests to the server using the https module... what did you mean by that? is there a specific module I need to use to enable a TLS session using parse/node? Nah, I just meant writing a simple node program that makes an http requests to your server, and sees if it can get any response. I was referring to node's builtin https module, there's no configuration necessary on the parse/node side. In most AWS setups with ELB, the ELB is the termination point for SSL (for speed reasons). Anything behind the load-balancer should be using http in its path, to connect to port 80, versus https, which will point to port 443. If it were going back out to the internet before reconnecting (for instance, to get the benefit of the load balancer), you'd still need the https protocol in there. I want to keep Github issues focused on SDK bugs and improvements, so I'll be closing out this task, since I'm convinced it's tied to your infrastructure and topology.
gharchive/issue
2016-06-02T16:18:20
2025-04-01T04:32:57.603172
{ "authors": [ "andrewimm", "ndnguru" ], "repo": "ParsePlatform/Parse-SDK-JS", "url": "https://github.com/ParsePlatform/Parse-SDK-JS/issues/290", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
55541153
Python 3 Support Possibly not going into 0.8.2-01, but we should be sure to not forget about this. Work in progress https://github.com/Parsely/pykafka/pull/144 +1 A lot has happened since work on this started, and the current most promising pull request is here thanks to @sontek. Figuring out a testing workflow around this will be my first priority once #177 is merged. @sontek's branch is looking pretty good now - as far as I'm concerned it's ready for merging. We'll need to add a bit of documentation about the preferred way to run tests under three interpreters. #231 #246 merged! :tada: :tada: :tada: :tada: :partly_sunny: :tada: :+1:
gharchive/issue
2015-01-26T21:37:33
2025-04-01T04:32:57.610988
{ "authors": [ "emmett9001", "kbourgoin", "mcandre", "rduplain" ], "repo": "Parsely/pykafka", "url": "https://github.com/Parsely/pykafka/issues/120", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
862869909
PS0001 is redundant This is now covered by CS4014: Note that CS4014 also does not report violations when .ConfigureAwait(false) is present: And it also does not report violations when the task is assigned: ... so it's functionally identical to PS0001. Bizarrely, CS4014 doesn't show on some calls: So I guess we still need PS0001 to be sure. 🤷
gharchive/issue
2021-04-20T14:19:31
2025-04-01T04:32:57.632451
{ "authors": [ "adamralph" ], "repo": "Particular/Particular.Analyzers", "url": "https://github.com/Particular/Particular.Analyzers/issues/90", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1108619743
Revert "Improvement/updates ruby to 2.6.8" Reverts Path-Check/gaen-mobile#946 Reverting until major Ruby upgrade @si1k Fixed!
gharchive/pull-request
2022-01-19T22:23:08
2025-04-01T04:32:57.714175
{ "authors": [ "mxMarkowitz", "si1k" ], "repo": "Path-Check/gaen-mobile", "url": "https://github.com/Path-Check/gaen-mobile/pull/961", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
785432855
Build a pipeline to download data from PubMed Description Build a pipeline to download data from PubMed. // Fill in detail - What is the goal of this issue? How will this information be used? Specification Details Fields to include PMID Title Abstract Persisting information filesystem Format name: PMID; body: newline-delim title, text Tasks [x] Create separate remote on GitHub for this capability (https://github.com/PathwayCommons/pubmed-dl) [ ] Create script to extract article information and text [ ] Persist extracted information [ ] // Fill in detail Notes Existing Tools/packages Wrapper for EUTILS i.e. https://github.com/mfcovington/pubmed-lookup Considerations How long does it take to download? How long does it take to index? How much memory is used in indexing? Superceded by https://github.com/PathwayCommons/pubmed-dl/issues/1 Superceded by https://github.com/PathwayCommons/pubmed-dl/issues/1
gharchive/issue
2021-01-13T20:38:18
2025-04-01T04:32:57.741761
{ "authors": [ "jvwong" ], "repo": "PathwayCommons/semantic-search", "url": "https://github.com/PathwayCommons/semantic-search/issues/43", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2303207365
BUG: Search not working Is there an existing issue for this? [X] I have searched the existing issues What happened? @PatilHarshh Search function not working Record [X] I agree to follow this project's Code of Conduct [X] I'm a GSSOC'24 contributor [X] I want to work on this issue check once again !!
gharchive/issue
2024-05-17T16:54:27
2025-04-01T04:32:57.744481
{ "authors": [ "PatilHarshh", "dhruv8433" ], "repo": "PatilHarshh/Kaam-Do", "url": "https://github.com/PatilHarshh/Kaam-Do/issues/68", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1200682447
Running Test for Get Entrance Fee Doesn't Work I tried run my test for getEntranceFee function and got an error. I can't figure out whats going on here. My test_lottery.py file : from brownie import Lottery, accounts, config, network from web3 import Web3 def test_get_entrance_fee(): account = accounts[0] lottery = Lottery.deploy( config["networks"][network.show_active()]["eth_usd_price_feed"], {"from": account}, ) assert lottery.getEntranceFee() > Web3.toWei(0.018, "ether") assert lottery.getEntranceFee() < Web3.toWei(0.022, "ether") my config file: dependencies: - smartcontractkit/chainlink-brownie-contracts@1.1.1 compiler: solc: remappings: - "@chainlink=smartcontractkit/chainlink-brownie-contracts@1.1.1" networks: mainnet-fork: eth_usd_price_feed: "0x5f4eC3Df9cbd43714FE2740f5E3616155c5b8419" output error: __________________________________________ test_get_entrance_fee ___________________________________________ def test_get_entrance_fee(): account = accounts[0] lottery = Lottery.deploy( config["networks"][network.show_active()]["eth_usd_price_feed"], {"from": account}, ) > assert lottery.getEntranceFee() > Web3.toWei(0.018, "ether") tests\test_lottery.py:12: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ C:\Users\ssd\.local\pipx\venvs\eth-brownie\lib\site-packages\brownie\network\multicall.py:115: in _proxy_call result = ContractCall.__call__(*args, **kwargs) # type: ignore C:\Users\ssd\.local\pipx\venvs\eth-brownie\lib\site-packages\brownie\network\contract.py:1751: in __call__ return self.call(*args, block_identifier=block_identifier, override=override) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <ContractCall 'getEntranceFee()'>, block_identifier = None, override = None, args = () tx = {'allow_revert': None, 'data': '0x09bc33a7', 'from': '0x66aB6D9362d4F35596279692F0251Db635165871', 'gas': None, ...} def call( self, *args: Tuple, block_identifier: Union[int, str, bytes] = None, override: Dict = None ) -> Any: """ Call the contract method without broadcasting a transaction. Arguments --------- *args Contract method inputs. You can optionally provide a dictionary of transaction properties as the last arg. block_identifier : int | str | bytes, optional A block number or hash that the call is executed at. If not given, the latest block used. Raises `ValueError` if this value is too far in the past and you are not using an archival node. override : dict, optional A mapping from addresses to balance, nonce, code, state, stateDiff overrides for the context of the call. Returns ------- Contract method return value(s). """ args, tx = _get_tx(self._owner, args) if tx["from"]: tx["from"] = str(tx["from"]) del tx["required_confs"] tx.update({"to": self._address, "data": self.encode_input(*args)}) try: data = web3.eth.call({k: v for k, v in tx.items() if v}, block_identifier, override) except ValueError as e: > raise VirtualMachineError(e) from None E brownie.exceptions.VirtualMachineError: revert C:\Users\ssd\.local\pipx\venvs\eth-brownie\lib\site-packages\brownie\network\contract.py:1542: VirtualMachineError ========================================= short test summary info ========================================== FAILED tests/test_lottery.py::test_get_entrance_fee - brownie.exceptions.VirtualMachineError: revert ============================================ 1 failed in 8.88s ============================================= Terminating local RPC client... Could you ask on the main repo? Thanks! https://github.com/smartcontractkit/full-blockchain-solidity-course-py
gharchive/issue
2022-04-11T23:45:02
2025-04-01T04:32:57.752628
{ "authors": [ "PatrickAlphaC", "bananlabs" ], "repo": "PatrickAlphaC/smartcontract-lottery", "url": "https://github.com/PatrickAlphaC/smartcontract-lottery/issues/58", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
229908194
Does it work for Xcode 8.3.2? I tried my project and some public project i.e Alamofire. But it does not generate anythings. Console: ./generate-objc-dependencies-to-json.rb -d -s "" > origin.js ; open index.html find: /Users/username/Library/Caches/appCode*/DerivedData: No such file or directory There were 91 directories found Last modifications were in /Users/username/Library/Developer/Xcode/DerivedData/Alamofire-dhtndbmjkaofkwemgqqalpcgngke/Build/Intermediates/Alamofire.build/Debug/Alamofire macOS.build/Objects-normal/x86_64 directory at 2017-05-19 18:38:31 +1000 Should work fine. You can try change something in the source and try to rebuild project. Should look like this
gharchive/issue
2017-05-19T08:41:26
2025-04-01T04:32:57.756113
{ "authors": [ "PaulTaykalo", "huy-le" ], "repo": "PaulTaykalo/objc-dependency-visualizer", "url": "https://github.com/PaulTaykalo/objc-dependency-visualizer/issues/43", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2141026672
Обновить документацию. Обновить информацию по автору yoomoney-api. https://github.com/AlekseyKorshuk/yoomoney-api Its unofficial python library...
gharchive/issue
2024-02-18T15:01:51
2025-04-01T04:32:57.762764
{ "authors": [ "PavelShaura", "franciscoMY" ], "repo": "PavelShaura/Autoseller_bot_aiogram_3", "url": "https://github.com/PavelShaura/Autoseller_bot_aiogram_3/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1057326296
feat: adds owner commands Adds Owner only command like: evals -> Run evals load/unload/reload plugins Looks good to me. Need a review from @CaffeineDuck . Please review and approve the changes so this can be merged. Pull Request has been stagnant for quite a while now. This is hindering the development of the bot. Merging this PR.
gharchive/pull-request
2021-11-18T12:57:08
2025-04-01T04:32:57.769948
{ "authors": [ "thenishantsapkota" ], "repo": "Peacebot-Development/Peacebot-v2", "url": "https://github.com/Peacebot-Development/Peacebot-v2/pull/24", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
318766057
Use Orchid to maintain documentation and landing page site Orchid is a tool I recently released which is perfect for managing the documentation of Java projects, and Pebble is the ideal candidate. Orchid itself uses Pebble as its primary template engine, so setting up Pebble with Orchid docs not only will improve the documentation as a whole, but also serve as a great example of what can be done with Pebble. This issue will track my progress getting Pebble set up with Orchid, and also serve as a discussion around the site. Progress [x] Landing page(s) [x] Full wiki, migrated from the current Github Wiki [x] Auto-generated listings of/documentation for: [x] Tags [x] Filters [x] Functions [x] Tests [x] Operators [x] Complete Javadoc API documentation [x] Changelog [ ] Link to older versions of docs [x] Automatic build/deploy from Travis CI to Github Pages [x] Host each version's docs in a subfolder to maintain docs history Discussion Points Orchid runs as a Gradle plugin, so using Orchid will require Pebble to either migrate from Maven to Gradle as its sole build system, or use both in the same project (Maven for normal build/test, Gradle for Orchid build). Alternatively, I would like to build a Maven plugin eventually but don't know how; if anyone has experience/would like to help out building a Maven Orchid plugin, we could continue using only Maven here What should the site look like? Keep the theme used for the current docs, use a default Orchid theme, or find/create a new theme just for Pebble https://github.com/JavaEden/Orchid Alright, I've got the initial groundwork laid out for this on the docs branch of my fork, and have a preview of the site deployed to https://pebbletemplates.netlify.com/ . Not much of the actual content has been migrated yet, but I've got it generating empty pages for everything registered in the CoreExtension as a start, currently, along with Javadocs. @cjbrooks12 I'd prefer to stick with Maven for pebble project. I know it really well vs gradle. I agree that an actual documentation website would be easier to use than the current wiki where the menu is not user friendly. Any news on what you started @cjbrooks12 ? Sorry for the delay with this, there were a number of bugfixes and features Orchid needed before it could really do well as a great project documentation tool. But I have just gotten a few of my other small projects set up fully with Orchid docs, along with Strikt, and feel like it's stable enough and now has the features needed to get Pebble set up with what it needs. I'll get these remaining tasks finished out soon (gotta get those 5 PRs!). @cjbrooks12 Looking at the strict docs they seem to be build using freemarker, not pebble? I've been working with the maintainer to replace JBake with Orchid, the PR is still open @cjbrooks12 is there something I could do to help you migrate the docs to Orchid? (gotta get those 5 PRs too :)) Do you know how to make Maven plugins? My current plan is to use a kscript like this to package everything up so that it doesn't need Gradle to run, but it would be nicer to have a Maven plugin which can run Orchid as part of the normal Pebble deploy process. Yes, I have written a couple plugins at work, I'll give it a try :) Very cool!! Maybe add a footer? (at least there's none in the mobile layout) The landing page could be a little bit improved, especially if it becomes the landing page of pebbletemplates.io, but it's not really related to Orchid nor to this PR :) This can be done later. Yeah I just used the current Wiki's landing page as the Orchid site homepage as well. And this theme is set up so the "footer" is at the bottom of the sidebar (which becomes a drawer on mobile). But yeah, the copyright on it should be updated to PebbleTemplates, is there anything else that should go in it? Also, now that I'm thinking about it, it would be nice to have a contributor list in the site, and maybe changelogs as well. I'm not sure how the current release process goes and if it automatically creates a Github Release, but Orchid's changelog plugin can create the release here at the same time the site is deployed, in addition to showing the full list of releases. Less manual work required for releasing new versions of Pebble 😉 The release process doesn't generate a Github release. I run mvn release:prepare release:perform which build everything and push it to sonatype. After that, I manually create the Github release and update documentation with the new version. It would be awesome to integrate the github release based on a Changelog to the release process and i'm open to hear about it @cjbrooks12 Can I close this issue ? I see that there's still an uncompleted task. Perhaps now http://pebbletemplates.io could show the new site instead of redirecting to the github project? Yeah on next release
gharchive/issue
2018-04-30T01:00:46
2025-04-01T04:32:57.784175
{ "authors": [ "DirkLachowski", "bjansen", "cjbrooks12", "ebussieres", "hectorlf" ], "repo": "PebbleTemplates/pebble", "url": "https://github.com/PebbleTemplates/pebble/issues/337", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }