id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
45719067
Adds rudimentary support for Arch Linux I am working in a VirtualBox, so I haven't executed the test suite against Arch Linux. Possibly some combination of attribute options will require further adjustment to work on this platform. After making these changes everything I did test (by hand) is working quite happily. Hello! There was a pretty heavy refactor in 0.40.0.... I suspect chef-client should detect systemd and try to use it on Arch. I'm going to close this issue for now, since this isn't going to merge cleanly.... if the new provider detection stuff in 12.4 doesn't work, you may need to add a line to the provider in libraries and send a new PR. Thanks! -s
gharchive/pull-request
2014-10-14T07:13:44
2025-04-01T04:33:37.494459
{ "authors": [ "logankoester", "someara" ], "repo": "bflad/chef-docker", "url": "https://github.com/bflad/chef-docker/pull/213", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
3221013
ImportError: cannot import name HTTPSConnection Hello, After installing this plugin from Package Control on a Ubuntu 11.10 system, I get this in the console: Reloading plugin /home/nchambrier/.config/sublime-text-2/Packages/sublime-github/gist.py Traceback (most recent call last): File "./sublime_plugin.py", line 62, in reload_plugin File "./gist.py", line 10, in <module> import requests File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/__init__.py", line 26, in <module> from . import utils File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/utils.py", line 19, in <module> from .compat import parse_http_list as _parse_list_header File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/compat.py", line 86, in <module> from .packages.oreos.monkeys import SimpleCookie File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/__init__.py", line 3, in <module> from . import urllib3 File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/__init__.py", line 16, in <module> from .connectionpool import ( File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/connectionpool.py", line 22, in <module> from httplib import HTTPConnection, HTTPSConnection, HTTPException ImportError: cannot import name HTTPSConnection The thing I don't understand is that I have only one version of Python installed, and when I run python REPL I can see the import working: $ python Python 2.7.2+ (default, Oct 4 2011, 20:03:08) [GCC 4.6.1] on linux2 Type "help", "copyright", "credits" or "license" for more information. >>> from httplib import HTTPConnection, HTTPSConnection, HTTPException >>> HTTPSConnection <class httplib.HTTPSConnection at 0x8cdee9c> Thanks for the report. I think the problem is the version of python that Sublime Text uses doesn't have SSL support. I worked around this before by using curl; guess I'll have to do that again. Yep I checked out closed issues after, but it's highly probable that #2 is related to this issue ;) On 14 February 2012 17:01, Brad Greenlee < reply@reply.github.com wrote: Thanks for the report. I think the problem is the version of python that Sublime Text uses doesn't have SSL support. I worked around this before by using curl; guess I'll have to do that again. Reply to this email directly or view it on GitHub: https://github.com/bgreenlee/sublime-github/issues/6#issuecomment-3963694 -- Nicolas Chambrier, aka naholyr Blog : http://naholyr.fr Formateur Clever Institut : http://clever-institut.com/formateur/nicolas-chambrier Ok, I've pushed a fix--re-adding the curl fallback (so you need to have curl installed). Upgrade to version 2.0.2 and let me know if it works for you. I've only tested it on OS X; if you have issues, I'll set up a Ubuntu VM to test with. Upgraded to 2.0.3 and still the same :( Here are the startup errors: Reloading plugin /home/nchambrier/.config/sublime-text-2/Packages/sublime-github/gist.py Traceback (most recent call last): File "./sublime_plugin.py", line 62, in reload_plugin File "./gist.py", line 7, in <module> import sublime_requests as requests File "./sublime_requests.py", line 5, in <module> import requests File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/__init__.py", line 26, in <module> from . import utils File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/utils.py", line 19, in <module> from .compat import parse_http_list as _parse_list_header File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/compat.py", line 86, in <module> from .packages.oreos.monkeys import SimpleCookie File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/__init__.py", line 3, in <module> from . import urllib3 File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/__init__.py", line 16, in <module> from .connectionpool import ( File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/connectionpool.py", line 25, in <module> from httplib import HTTPConnection, HTTPSConnection, HTTPException ImportError: cannot import name HTTPSConnection Reloading plugin /home/nchambrier/.config/sublime-text-2/Packages/sublime-github/sublime_requests.py Traceback (most recent call last): File "./sublime_plugin.py", line 62, in reload_plugin File "./sublime_requests.py", line 5, in <module> import requests File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/__init__.py", line 26, in <module> from . import utils File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/utils.py", line 19, in <module> from .compat import parse_http_list as _parse_list_header File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/compat.py", line 86, in <module> from .packages.oreos.monkeys import SimpleCookie File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/__init__.py", line 3, in <module> from . import urllib3 File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/__init__.py", line 16, in <module> from .connectionpool import ( File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/connectionpool.py", line 25, in <module> from httplib import HTTPConnection, HTTPSConnection, HTTPException ImportError: cannot import name HTTPSConnection I'll check out if I can fix it myself and git you a patch, so you don't have to bother with a VM. I think it will still be hard after that. I naively removed "HTTPSConnection, " from "urllib3.connectionpool", and I then get other errors: Reloading plugin /home/nchambrier/.config/sublime-text-2/Packages/sublime-github/gist.py Traceback (most recent call last): File "./sublime_plugin.py", line 62, in reload_plugin File "./gist.py", line 7, in <module> import sublime_requests as requests File "./sublime_requests.py", line 5, in <module> import requests File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/__init__.py", line 26, in <module> from . import utils File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/utils.py", line 19, in <module> from .compat import parse_http_list as _parse_list_header File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/compat.py", line 86, in <module> from .packages.oreos.monkeys import SimpleCookie File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/__init__.py", line 3, in <module> from . import urllib3 File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/__init__.py", line 16, in <module> from .connectionpool import ( File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/connectionpool.py", line 42, in <module> from .request import RequestMethods File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/request.py", line 12, in <module> from .filepost import encode_multipart_formdata File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/filepost.py", line 15, in <module> from io import BytesIO File ".\io.py", line 63, in <module> ImportError: No module named _fileio Reloading plugin /home/nchambrier/.config/sublime-text-2/Packages/sublime-github/sublime_requests.py Traceback (most recent call last): File "./sublime_plugin.py", line 62, in reload_plugin File "./sublime_requests.py", line 5, in <module> import requests File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/__init__.py", line 26, in <module> from . import utils File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/utils.py", line 19, in <module> from .compat import parse_http_list as _parse_list_header File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/compat.py", line 86, in <module> from .packages.oreos.monkeys import SimpleCookie File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/__init__.py", line 3, in <module> from . import urllib3 File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/__init__.py", line 16, in <module> from .connectionpool import ( File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/connectionpool.py", line 42, in <module> from .request import RequestMethods File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/request.py", line 12, in <module> from .filepost import encode_multipart_formdata File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/filepost.py", line 15, in <module> from io import BytesIO File ".\io.py", line 63, in <module> ImportError: No module named _fileio I'm not sure yet it those errors are due to the fact I removed HTTPSConnection from connectionpool, or if it's yet another missing component from my system :-\ You definitely don't want to remove HTTPSConnection entirely--that will break OS X. But it does sound like there might be other things missing. It shouldn't be hard to work around, since Linux doesn't need any of that anyway. I'm downloading Ubuntu 11.10 now, so I'll have a fix soon. Ok, give it a shot. Remember you need curl installed, so apt-get install curl if you haven't already. I had a similar problem where the issue was openssl was updated but python was not recompiled, basically whenever you update openssl you have to recompile python. https://github.com/jarun/googler/issues/312
gharchive/issue
2012-02-14T15:43:12
2025-04-01T04:33:37.513483
{ "authors": [ "bgreenlee", "ishandutta2007", "naholyr" ], "repo": "bgreenlee/sublime-github", "url": "https://github.com/bgreenlee/sublime-github/issues/6", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
241249376
libcrnn not found When I run /src/demo.lua with th , it gives me this error ```module 'libcrnn' not found:No LuaRocks module found for libcrnn no field package.preload['libcrnn'] no file '/home/akash/.luarocks/share/lua/5.1/libcrnn.lua' no file '/home/akash/.luarocks/share/lua/5.1/libcrnn/init.lua' no file '/home/akash/torch/install/share/lua/5.1/libcrnn.lua' no file '/home/akash/torch/install/share/lua/5.1/libcrnn/init.lua' no file './libcrnn.lua' no file '/home/akash/torch/install/share/luajit-2.1.0-beta1/libcrnn.lua' no file '/usr/local/share/lua/5.1/libcrnn.lua' no file '/usr/local/share/lua/5.1/libcrnn/init.lua' no file '/home/akash/.luarocks/lib/lua/5.1/libcrnn.so' no file '/home/akash/torch/install/lib/lua/5.1/libcrnn.so' no file '/home/akash/torch/install/lib/libcrnn.so' no file './libcrnn.so' no file '/usr/local/lib/lua/5.1/libcrnn.so' no file '/usr/local/lib/lua/5.1/loadall.so' Its looking for the .so file that is in your src directory after you've successfully built the project. Same issue. There is .so file even then showing the same error @JiteshPshah ,hello,I am facing the same problem ,have you run successfully ,if so,can you give me some advice, thanks a lot ^_^ Hi Did you guys fix this?
gharchive/issue
2017-07-07T12:29:04
2025-04-01T04:33:37.518580
{ "authors": [ "DecentMakeover", "JiteshPshah", "dvlshah", "lxx1884896", "stalagmite7" ], "repo": "bgshih/crnn", "url": "https://github.com/bgshih/crnn/issues/69", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2047188007
🛑 Hacker News is down In 7e19267, Hacker News (https://news.ycombinator.com) was down: HTTP code: 502 Response time: 362 ms Resolved: Hacker News is back up in cc04d68 after 11 minutes.
gharchive/issue
2023-12-18T18:13:00
2025-04-01T04:33:37.521235
{ "authors": [ "bguivarch" ], "repo": "bguivarch/testuptime2", "url": "https://github.com/bguivarch/testuptime2/issues/100", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
712948644
added A-star path finding algorithm in the projects This is my project where we can visualize the A-star path finding algorithm Wow, amazing code! Wow, amazing code! Thanks :) @bhaveshlohana
gharchive/pull-request
2020-10-01T15:25:23
2025-04-01T04:33:37.563528
{ "authors": [ "bhaveshlohana", "shikharmaxx" ], "repo": "bhaveshlohana/HacktoberFest2020-Contributions", "url": "https://github.com/bhaveshlohana/HacktoberFest2020-Contributions/pull/330", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
351469430
Problem: Gnome-Terminal displays segment color after newline if using two newline - segments Hi, please see following screenshot: As you can see, the color of the status segment is displayed after the newline. I'm using ohmyzsh and my .zshrc config is as follows: POWERLEVEL9K_LEFT_PROMPT_ELEMENTS=(context dir status newline vcs newline) POWERLEVEL9K_RIGHT_PROMPT_ELEMENTS=(time) POWERLEVEL9K_DISABLE_RPROMPT=true POWERLEVEL9K_CUSTOM_PROMPTICON="echo \>" # HIDE current user in prompt DEFAULT_USER=`whoami` # load solarized dir colors eval `dircolors ~/.dircolors/dircolors.ansi-dark` It doesn't matter what segments I use. If using two newline statements, this error occurs. Does anybody know a solution to this or is it just not recommended to use two newline statements? Thanks ;) I have the exact same problem. I'm on Ubuntu 18.04, using oh-my-zsh. All other configurations seem to work as expected, but this little annoyance keeps showing up. Hm, interesting. And weird that it's actually showing the segment BG color of the segment on the top line, not of the segment on the second line. @bytekeks @ChrisSeattle - Are you guys using a release or the master branch, by chance? Can you give the next branch a shot and let us know if the behavior is the same? @bytekeks - Whoa, really? The latest master fixed it for you? @ChrisSeattle - Can you test & confirm as well? @bhilburn indeed. The issue is gone :) It seems, that using no seperator icons lead to this problem in the old version. POWERLEVEL9K_LEFT_SEGMENT_SEPARATOR="" POWERLEVEL9K_RIGHT_SEGMENT_SEPARATOR="" In the current master, leaving SEPERATOR empty will just be ignored and the default is used, right? Great! I'm so happy to hear the issue is resolved in the next branch. That means a fix will be included in the next release. @bytekeks - If you don't define the separator at all, then the default is used. If you define it to be empty, then it will be empty -- which was probably the cause, per your suspicion =) I'm closing this issue since it's resolved in next. Please re-open if you are still experiencing this!
gharchive/issue
2018-08-17T06:13:18
2025-04-01T04:33:37.569149
{ "authors": [ "ChrisSeattle", "bhilburn", "bytekeks" ], "repo": "bhilburn/powerlevel9k", "url": "https://github.com/bhilburn/powerlevel9k/issues/957", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
128277971
Reworked the AWS-Beanstalk segment This PR replaces #162 Perfect! Thank you, @dritter!!
gharchive/pull-request
2016-01-23T00:38:04
2025-04-01T04:33:37.570298
{ "authors": [ "bhilburn", "dritter" ], "repo": "bhilburn/powerlevel9k", "url": "https://github.com/bhilburn/powerlevel9k/pull/203", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
156396200
Refactor location installation path code Now it is possible to set the installation path by setting POWERLEVEL9K_INSTALLATION_PATH. If set, this path will always be used. If it was not set manually, we try to determine the path automatically. This fixes #264 @protist could you double-check? Especially with your ZSH 5.0.7. That would be great. Thanks. Patching works perfectly on my zsh 5.0.7. Thank you for the quick fix. Lots of changes here, not all of which I can easily test. The code looks good, though! Merging into next for further dev testing. Thanks so much, @dritter! Looks like it fixed @protist's issue, as well, which is awesome. Nice work.
gharchive/pull-request
2016-05-24T00:37:05
2025-04-01T04:33:37.572438
{ "authors": [ "bhilburn", "dritter", "protist" ], "repo": "bhilburn/powerlevel9k", "url": "https://github.com/bhilburn/powerlevel9k/pull/266", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
238651155
Fixed prompt hangs on kubectl version command This resolves issue #553 when prompt hangs on kubectl version Ran into this issue when using this feature on next based on the great work jtaylor007.jt@gmail.com did. @derekbassett - Thanks so much for taking the time to fix this issue! I appreciate the thorough explanation both here and in #553. This is a great PR. Welcome to the club of P9k contributors! =)
gharchive/pull-request
2017-06-26T19:53:52
2025-04-01T04:33:37.574350
{ "authors": [ "bhilburn", "derekbassett" ], "repo": "bhilburn/powerlevel9k", "url": "https://github.com/bhilburn/powerlevel9k/pull/554", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1054361959
dwi description seems wrong: T2w instead of T2*w here it says that "Diffusion-weighted imaging contrast (specialized T2* weighting)." but given that DWI is based on a spin echo (vs. gradient echo) EPI, I would call it a "specialized T2 weighting" instead. Pinging @oesteban, who might be able to weigh in on this. If it is something that needs to be changed, we can do that in the suffixes schema file: https://github.com/bids-standard/bids-specification/blob/30c61b7eb473cccd2414d3947d38487995720f67/src/schema/objects/suffixes.yaml#L426-L429
gharchive/issue
2021-11-16T02:22:55
2025-04-01T04:33:37.621460
{ "authors": [ "jcohenadad", "tsalo" ], "repo": "bids-standard/bids-specification", "url": "https://github.com/bids-standard/bids-specification/issues/930", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
195777933
Add OS package install option Add an option to install Wildfly from an OS package Vs from downloaded tar Coverage remained the same at 73.83% when pulling 213711b174bbece2b049eb7e3ae5043506ce56f3 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master. Coverage remained the same at 73.83% when pulling 213711b174bbece2b049eb7e3ae5043506ce56f3 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master. Hey @zipkid, this is pretty much the code in PR #137 Do you know @jstremick ? I suppose the same suggestions I gave there applies here: remove install_method parameter and use package if it's provided and URL otherwise remove code to set ownership of wildfly directory. I believe it's package responsability. What do you think? Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master. Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master. Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master. Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master. Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master. Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master. Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master. @jairojunior any thoughts on this? The fourth comment is mine. @coveralls is kind of annoying. :smile: Damn... I missed yours in the forest.... I'll look into your comment :-) Your comment makes complete sense. I hope to be able to modify my code according to that soonish... Awesome. Feel free to disagree though, it's just a suggestion based exclusively on code review. :smile: Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master. Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master. Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master. Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master. Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master. Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master. Awesome. Thanks.
gharchive/pull-request
2016-12-15T11:09:08
2025-04-01T04:33:37.655174
{ "authors": [ "coveralls", "jairojunior", "zipkid" ], "repo": "biemond/biemond-wildfly", "url": "https://github.com/biemond/biemond-wildfly/pull/165", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
108217754
better validation when saving paired fastqs logError for LENIENT throw for `STRICT Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins/job/ADAM-prb/926/ Test PASSed. Other than the small nit, LGTM! nit should be fixed Thanks @ryan-williams! I'll leave this open for comments, but if we don't hear from anyone else tomorrow, I'll merge. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins/job/ADAM-prb/931/ Test PASSed. Thanks @ryan-williams!
gharchive/pull-request
2015-09-24T21:28:21
2025-04-01T04:33:37.720232
{ "authors": [ "AmplabJenkins", "fnothaft", "ryan-williams" ], "repo": "bigdatagenomics/adam", "url": "https://github.com/bigdatagenomics/adam/pull/830", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
96604773
Rename to bigdatagenomics/formats Small nit: I feel like bigdatagenomics/bdg-formats is redundant. Would anyone be opposed to renaming to bigdatagenomics/formats? I will leave this open for a week and make the changes if no one is opposed. We decided against this; it would lead to a JAR with a name that was too generic.
gharchive/issue
2015-07-22T16:08:54
2025-04-01T04:33:37.721391
{ "authors": [ "fnothaft" ], "repo": "bigdatagenomics/bdg-formats", "url": "https://github.com/bigdatagenomics/bdg-formats/issues/60", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
53577987
Deal properly with merging and partitioning data. For example, the 1000 Genomes VCF data is organized into individual files by chromosome. But this obviously causes the files to differ substantially in size. My ideal would be to merge into a single file, or to split the genome into equal sized bins based on locus and sort the data into it. This is being tackled in https://github.com/bigdatagenomics/eggo/issues/30 - closing.
gharchive/issue
2015-01-06T23:34:22
2025-04-01T04:33:37.722695
{ "authors": [ "laserson", "tomwhite" ], "repo": "bigdatagenomics/eggo", "url": "https://github.com/bigdatagenomics/eggo/issues/2", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2026219490
Updated AbsoluteLink I was receiving the following error on silverstripe 5.1, php 8.1: Fatal error: Declaration of Bigfork\SilverStripeOAuth\Client\Control\Controller::AbsoluteLink() must be compatible with SilverStripe\Control\RequestHandler::AbsoluteLink($action = '') in C:\xampp-8.1\htdocs\silverstripe-install\vendor\bigfork\silverstripe-oauth\src\Control\Controller.php on line 79 ERROR [UNKNOWN TYPE, ERRNO 64]: Declaration of Bigfork\SilverStripeOAuth\Client\Control\Controller::AbsoluteLink() must be compatible with SilverStripe\Control\RequestHandler::AbsoluteLink($action = '') IN GET / Line 79 in C:\xampp-8.1\htdocs\silverstripe-install\vendor\bigfork\silverstripe-oauth\src\Control\Controller.php Source ====== 70: $backUrl = Director::absoluteBaseURL(); 71: } 72: 73: return $backUrl; 74: } 75: 76: /** 77: * @return string 78: */ * 79: public function AbsoluteLink() 80: { 81: return static::join_links(Director::absoluteBaseURL(), $this->Link()); 82: } 83: 84: /** 85: * This takes parameters like the provider, scopes and callback url, builds an authentication Trace ===== SilverStripe\Logging\DetailedErrorFormatter->output(64, Declaration of Bigfork\SilverStripeOAuth\Client\Control\Controller::AbsoluteLink() must be compatible with SilverStripe\Control\RequestHandler::AbsoluteLink($action = ''), C:\xampp-8.1\htdocs\silverstripe-install\vendor\bigfork\silverstripe-oauth\src\Control\Controller.php, 79, ) DetailedErrorFormatter.php:55 SilverStripe\Logging\DetailedErrorFormatter->format(Monolog\LogRecord) AbstractProcessingHandler.php:42 Monolog\Handler\AbstractProcessingHandler->handle(Monolog\LogRecord) Logger.php:389 Monolog\Logger->addRecord(Monolog\Level, Fatal Error (E_COMPILE_ERROR): Declaration of Bigfork\SilverStripeOAuth\Client\Control\Controller::AbsoluteLink() must be compatible with SilverStripe\Control\RequestHandler::AbsoluteLink($action = ''), Array) Logger.php:579 Monolog\Logger->log(Monolog\Level, Fatal Error (E_COMPILE_ERROR): Declaration of Bigfork\SilverStripeOAuth\Client\Control\Controller::AbsoluteLink() must be compatible with SilverStripe\Control\RequestHandler::AbsoluteLink($action = ''), Array) ErrorHandler.php:247 Monolog\ErrorHandler->handleFatalError() I just updated the AbsoluteLink function with $action Thanks, tagged in 2.2.1
gharchive/pull-request
2023-12-05T13:37:47
2025-04-01T04:33:37.725098
{ "authors": [ "aletail", "kinglozzer" ], "repo": "bigfork/silverstripe-oauth", "url": "https://github.com/bigfork/silverstripe-oauth/pull/19", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2069989821
别买了,我买了没用的,大家别花冤枉钱,一直是下载中0% 你下载是什么,具体的下载地址提供一下 不用当心没有用,如果是软件问题,肯定能解决,如果有问题及时反馈,100%能免单 有问题,可以支付宝上留言 有问题,请支付宝上留言 好的我去支付宝留言
gharchive/issue
2024-01-08T08:52:05
2025-04-01T04:33:37.726985
{ "authors": [ "bigintpro", "yyxxiiaaoo" ], "repo": "bigintpro/csdn_downloader", "url": "https://github.com/bigintpro/csdn_downloader/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
328432446
fixes for linux Fixes two problems on linux: "ambiguous use of FE_UNDERFLOW" fixed with explicit variable type Lack of support for CFNumberGetType. Reimplemented using casts I'm okay with both changes regardless of CI - I'll get that fixed shortly. Thank you!
gharchive/pull-request
2018-06-01T08:47:25
2025-04-01T04:33:37.728289
{ "authors": [ "mlilback", "zwaldowski" ], "repo": "bignerdranch/Freddy", "url": "https://github.com/bignerdranch/Freddy/pull/279", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
106652271
Improved expand/collapse performance When expanding/collapsing a group of list items, notifyDataSetChanged is called multiple times. It would perform better if notifyDataSetChanged was just called once after all of the items were expanded/collapsed. Resolved in 2.1.0.
gharchive/issue
2015-09-15T21:30:15
2025-04-01T04:33:37.729352
{ "authors": [ "dgreenhalgh" ], "repo": "bignerdranch/expandable-recycler-view", "url": "https://github.com/bignerdranch/expandable-recycler-view/issues/72", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
87306993
re work on report report enhancement currently only covers steploggable. we will need to add code to cover the objectfinder and objectaction report needs to be considered as well
gharchive/issue
2015-06-11T11:14:42
2025-04-01T04:33:37.734896
{ "authors": [ "peidong-hu" ], "repo": "bigtester/automation-test-engine", "url": "https://github.com/bigtester/automation-test-engine/issues/164", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1025561587
Login feature test Adds Wallaby for Feature/Integration testing Adds an example test for Login [x] Works locally [ ] Works in CI @drteeth have you tried https://hexdocs.pm/phoenix_live_view/Phoenix.LiveViewTest.html at all? I think that it can do everything Wallaby can for liveviews, and is faster / doesn't need the chrome dependency. The only thing that it wouldn't cover is the test you wrote here which is a "deadview" but that's because when I wrote the login page LiveView didn't have hooks yet - it can be rewritten as a liveview. Feels on trying some LiveView tests to see if you like them? So browser tests more for like full integration testing, and liveview tests for "browser unit tests?"
gharchive/pull-request
2021-10-13T18:22:07
2025-04-01T04:33:37.746736
{ "authors": [ "drteeth", "mveytsman" ], "repo": "bikebrigade/dispatch", "url": "https://github.com/bikebrigade/dispatch/pull/5", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
177145865
Oracle DB support Dear Bill, how are you, i hope you are fine. i am php developer and not in java development at all so please be patient to my questions. i did download the oracle driver ojdbc6.jar and put it in all WEB-INF\lib for all webapps and also did configure Oracle driver in the database table qc_data_source_driver with the following class name oracle.jdbc.driver.OracleDriver and in the table qc_data_source_conf i configure the JDBC_URL jdbc:oracle:thin:username/password@database:1521:xe i tried to change the configuration file applicationContext-dataSource.properties in each webapps like the following: dataSource.driverClass=oracle.jdbc.driver.OracleDriver dataSource.jdbcUrl=jdbc:oracle:thin:@//darabase:1521/XE dataSource.user=username dataSource.password=password dataSource.maxPoolSize=48 dataSource.minPoolSize=1 dataSource.initialPoolSize=1 dataSource.maxIdleTime=120 after restart the server nothing appears in the page except the top bar with the application dropdown and only the refresh option is there so could you please guide me how to make it work migrating to the oracle database? thanks for your time. Config qc_data_source_conf and put ojdbc6.jar is only support OLAP for QCHARTS-WEB, if you want all bambooBSC system support Oracle db, you need jump current mysql bbcore data to Oracle, thanks. I want to know some questions: Do you want to use the OLAP features like(01 - Basic query, 02 - Analytics) Or want bambooBSC use MySQL database change to Oracle-DB please answer to me. 1 or 2. thanks. Dear chen, It is 2 i want to move totally to use oracle db instead of the current MYSQL database. OK, i will to try how to jump Oracle-DB, i will response issues if jump MySQL to Oracle-DB is success complete. That would be grate and tell me if there is anything I can do from my side. I tested all day The main problem now is: Some of the bambooBSC table field name is a reserved word in Oracle-DB, like SYS_ID So if want to transfer to Oracle, then the table mapper Object and table field name must be change. I think currently is no way support Oracle-DB in 0.6.6 release. Support Oracle-DB plan i want move do in next version 0.7.0 . i did manage to migrate the Database form MySQL to Oracle - using Oracle SQL Developer - with errors in the fields that marked not null but has null value, otherwise everything work perfectly. if the migration did not work the try to copy the tables one by one and that at lest create the table empty. and here is a drop form Oracle SQL Developer for the table TB_SYS: CREATE_ TABLE "BBCORE"."TB_SYS" ( "OID" CHAR(36 CHAR) NOT NULL ENABLE, "SYS_ID" VARCHAR2(10 CHAR) NOT NULL ENABLE, "NAME" VARCHAR2(100 CHAR) NOT NULL ENABLE, "HOST" VARCHAR2(200 CHAR) NOT NULL ENABLE, "CONTEXT_PATH" VARCHAR2(100 CHAR) NOT NULL ENABLE, "IS_LOCAL" VARCHAR2(1 CHAR) NOT NULL ENABLE, "ICON" VARCHAR2(20 CHAR) NOT NULL ENABLE, "CUSERID" VARCHAR2(24 CHAR) NOT NULL ENABLE, "CDATE" DATE NOT NULL ENABLE, "UUSERID" VARCHAR2(24 CHAR), "UDATE" DATE ) SEGMENT CREATION IMMEDIATE PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT) TABLESPACE "SYSTEM" ; i want to correct the JDBC_URL to jdbc:oracle:thin:username/password@Server:1521:xe i has manage to migrate the Database form MySQL to Oracle completed. but in bambooBSC runtime run SQL error, because SYS_ID cannot in SQL SELECT command with Oracle-DB expression version. This SQL command cannot run with my Oracle-DB expression version. It need for bambooBSC. select tbsys0_.OID as OID1_2_, tbsys0_.CDATE as CDATE2_2_, tbsys0_.CONTEXT_PATH as CONTEXT_PATH3_2_, tbsys0_.CUSERID as CUSERID4_2_, tbsys0_.HOST as HOST5_2_, tbsys0_.ICON as ICON6_2_, tbsys0_.IS_LOCAL as IS_LOCAL7_2_, tbsys0_.NAME as NAME8_2_, tbsys0_.SYS_ID as SYS_ID9_2_, tbsys0_.UDATE as UDATE10_2_, tbsys0_.UUSERID as UUSERID11_2_ from tb_sys tbsys0_ where 1=1 and tbsys0_.SYS_ID='CORE'
gharchive/issue
2016-09-15T10:52:28
2025-04-01T04:33:37.798963
{ "authors": [ "billchen198318", "maltourky" ], "repo": "billchen198318/bamboobsc", "url": "https://github.com/billchen198318/bamboobsc/issues/34", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
226692016
Fix block comments. Fixes #9 by assigning the leadingComments and comments from the property to type rather than type.value. Thanks!
gharchive/pull-request
2017-05-05T21:06:37
2025-04-01T04:33:37.803238
{ "authors": [ "billyvg", "wdhorton" ], "repo": "billyvg/codemod-proptypes-to-flow", "url": "https://github.com/billyvg/codemod-proptypes-to-flow/pull/10", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
167913020
Sorting Issue There seems to be some issue with the sorting algorithm: This seems like a duplicate of https://github.com/billyvg/pokemon-journal/issues/6 and kind of fixed in https://github.com/billyvg/pokemon-journal/pull/17 Yep, this should be fixed in #17
gharchive/issue
2016-07-27T17:40:27
2025-04-01T04:33:37.805128
{ "authors": [ "Nicte", "billyvg", "onilton" ], "repo": "billyvg/pokemon-journal", "url": "https://github.com/billyvg/pokemon-journal/issues/7", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
345666872
Honesty Store Logo - Remove it NAME OF PAGE/COMPONENT All Screens PART OF DESIGN NOT MATCHING Remove the honesty store logo Logo is still present in UX spec - need confirmation of where to replace with back buttons to continue @isaac-lapworth https://drive.google.com/file/d/16ee202YhVJ1f1-3c9zMwaO-ZTJLBvQxI/view?usp=sharing please find the spec without the HS logo linked here. The Honesty Store logo has been removed from all pages
gharchive/issue
2018-07-30T08:51:26
2025-04-01T04:33:37.824893
{ "authors": [ "KyleADay", "isaac-lapworth", "siobhanobrien" ], "repo": "bin-nuance-bat/project", "url": "https://github.com/bin-nuance-bat/project/issues/222", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
404396596
stringify_keys returns whole config when used with a Hash value I've a config which has one value as a Hash as follows: PlannerConfig.driving => {"average_speed"=>50, "driver_off_duty_time"=>10, "max_driver_on_duty_time"=>14, "day_one_cap_value"=>300, "max_continuous_drive_time"=>10, "max_waiting_time"=>48, "mandatory_rest_after_drive"=>0.5, "sleeping_time"=>10} PlannerConfig is using SettingsLogic to the values from a yaml file. The values are: distance: redis_cache: true ts: max_iterations: 1000 max_unchanged_iterations: 200 thread_count: 1 timeout: 5 shuffle: 5 cost_tolerance: 1.0 mode_selection: false cost_calculation: true ignore_capacity: false constraints: max_number_of_stops: 3 out_of_route_items_attributes: [ {distance_value: 0, percent: 40} ] driving: average_speed: 50 driver_off_duty_time: 10 max_driver_on_duty_time: 14 day_one_cap_value: 300 max_continuous_drive_time: 11 max_waiting_time: 48 mandatory_rest_after_drive: 0.5 max_continuous_drive_time: 10 sleeping_time: 10 task: key: 'planner:running' When passing this to an AR object as a parameter for initialization, it throws an error with missing attributes: [1] pry(main)> Optimization::DrivingHoursConfiguration.new(PlannerConfig.driving) ActiveRecord::UnknownAttributeError: unknown attribute 'distance' for Optimization::DrivingHoursConfiguration. from /home/ashish/.rvm/gems/ruby-2.3.8/ruby/2.3.0/gems/activerecord-4.2.11/lib/active_record/attribute_assignment.rb:59:in `rescue in _assign_attribute' This happens in Rails 4.2.x, doesn't happen in Rails 4.1.x. On investigation I found out that when using stringify_keys the output is not as expected: [2] pry(main)> PlannerConfig.driving => {"average_speed"=>50, "driver_off_duty_time"=>10, "max_driver_on_duty_time"=>14, "day_one_cap_value"=>300, "max_continuous_drive_time"=>10, "max_waiting_time"=>48, "mandatory_rest_after_drive"=>0.5, "sleeping_time"=>10} [3] pry(main)> PlannerConfig.driving.stringify_keys => {"distance"=>{"redis_cache"=>true}, "ts"=>{"max_iterations"=>2500, "max_unchanged_iterations"=>1000}, "constraints"=>{"max_number_of_stops"=>3, "out_of_route_items_attributes"=>[{"distance_value"=>0, "percent"=>40}]}, "driving"=> {"average_speed"=>50, "driver_off_duty_time"=>10, "max_driver_on_duty_time"=>14, "day_one_cap_value"=>300, "max_continuous_drive_time"=>10, "max_waiting_time"=>48, "mandatory_rest_after_drive"=>0.5, "sleeping_time"=>10}, "task"=>{"key"=>"planner:running"}, "average_speed"=>50, "driver_off_duty_time"=>10, "max_driver_on_duty_time"=>14, "day_one_cap_value"=>300, "max_continuous_drive_time"=>10, "max_waiting_time"=>48, "mandatory_rest_after_drive"=>0.5, "sleeping_time"=>10} This is happening due to change in activesupport where we're using self.class.new instead of populating a hash, reference commit here. So it does PlannerConfig.new which adds all the keys and hence getting this error. For now I am doing Optimization::DrivingHoursConfiguration.new(PlannerConfig.driving.to_h). Moved to https://github.com/settingslogic/settingslogic/issues/19.
gharchive/issue
2019-01-29T17:12:31
2025-04-01T04:33:37.907683
{ "authors": [ "ashish91" ], "repo": "binarylogic/settingslogic", "url": "https://github.com/binarylogic/settingslogic/issues/85", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
443768257
automatic run nodejs and throw some error pyspider version: 0.3.10 Operating system: windows 10 Start up command: pyspider Expected behavior normal run Actual behavior throw Error: Cannot find module 'express' * N exec 2 nodejs process and exit over over again How to reproduce install nodejs into windows path then run pyspider solved remove nodejs from windows path You'll need to install node components: express, puppeteer, and body-parser before starting @snail-lb no ... remove node.exe from path running no error I was getting the same error. The documentation is not updated. In pyspider/fetcher/puppeteer_fetcher.js you can find this: const express = require("express"); const puppeteer = require('puppeteer'); const bodyParser = require('body-parser'); I have installed the 3 node packages within the pyspider folder: $ npm install express puppeteer body-parser and the error is gone.
gharchive/issue
2019-05-14T08:06:13
2025-04-01T04:33:37.920567
{ "authors": [ "huaji0353", "hubitor", "snail-lb" ], "repo": "binux/pyspider", "url": "https://github.com/binux/pyspider/issues/898", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1175407475
Attachments: save to cache folder instead of Downloads Currently, attachments are directly downloaded to the Download folder, which can be annoying for ephemeral images and such. It'd be nice to have a way to configure and/or move/copy files to the Download folder on demand. Done in https://github.com/binwiederhier/ntfy-android/commit/8339bc9c2a385d5ee685deaba8bb0eef4be80c7c, will be in 1.11.0
gharchive/issue
2022-03-21T13:41:53
2025-04-01T04:33:37.921941
{ "authors": [ "binwiederhier" ], "repo": "binwiederhier/ntfy", "url": "https://github.com/binwiederhier/ntfy/issues/181", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
411795026
Реализовать сериализацию PDB в Text Спецификация здесь: http://www.wwpdb.org/documentation/file-format-content/format33/v3.3.html Added in https://github.com/biocad/cobot-io/commit/e53b634e79ebf770624eca4d421b5cc7f80436b3
gharchive/issue
2019-02-19T08:08:59
2025-04-01T04:33:37.930665
{ "authors": [ "AlexKaneRUS", "zmactep" ], "repo": "biocad/cobot-io", "url": "https://github.com/biocad/cobot-io/issues/7", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1050797964
Add debug_model function This adds a helper function to debug a model, which returns test inputs, outputs, expected outputs as well as relevant intermediates. In order to implement it I have changed the pre-and-postprocessing calls in PredictionPipeline s.t. they also return the stats parameter; otherwise it's not possible to use the postprocessing function. returning the stats is good and it's just internal anyway. 👍
gharchive/pull-request
2021-11-11T09:56:51
2025-04-01T04:33:38.039845
{ "authors": [ "FynnBe", "constantinpape" ], "repo": "bioimage-io/core-bioimage-io-python", "url": "https://github.com/bioimage-io/core-bioimage-io-python/pull/141", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
806599876
API can persist data-processing progress Background Link to issue https://biomage.atlassian.net/browse/BIOMAGE-500 https://biomage.atlassian.net/browse/BIOMAGE-534 Link to staging deployment URL https://ui-gk4wi78my831kkcxbf13nz8gkh.scp-staging.biomage.net/ Links to any Pull Requests related to this N/A Anything else the reviewers should know about the changes here Joined the two tickets together are they both involved some refactoring of the API code. Changes Code changes Refactored the API schemas to be in separate files for easier readability of diffs in merge requests. Added an additional schema for the meta part of the Redux store. Persisting processing steps can then be done by the existing Redux action for saving the actual steps' settings. Definition of DONE Your changes will be ready for merging after each of the steps below have been completed: Testing [x] Unit tests written [x] Tested locally with Inframock [ ] Deployed to staging To set up easy local testing with inframock, follow the instructions here: https://github.com/biomage-ltd/inframock To deploy to the staging environment, follow the instructions here: https://github.com/biomage-ltd/biomage-utils Documentation updates Is all relevant documentation updated to reflect the proposed changes in this PR? [ ] Relevant Github READMEs updated [ ] Relevant wiki pages created/updated Approvers [ ] Approved by a member of the core engineering team [ ] (UX changes) Approved by vickymorrison (this is her username, tag her if you need approval) Just before merging: [ ] After the PR is approved, the unstage script in here: https://github.com/biomage-ltd/biomage-utils is executed. This script cleans up your deployment to staging Optional [ ] Photo of a cute animal attached to this PR Staging available at https://ui-gk4wi78my831kkcxbf13nz8gkh.scp-staging.biomage.net/. Looks all good to me. Some changes requested to the API schema are better allocated to task on persisting DataIntegration in DynamoDB @ivababukova approved the changes but can't approve via GitHub now.
gharchive/pull-request
2021-02-11T17:35:09
2025-04-01T04:33:38.068716
{ "authors": [ "aerlaut", "marcellp" ], "repo": "biomage-ltd/api", "url": "https://github.com/biomage-ltd/api/pull/45", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1736983816
Access to biomarker/data, api, and sparql Please provide Daniall, Robel, Jonathon, Pat, and Karina access to hive-lab/biochemistry.edu/biomarker/data hive-lab/biochemistry.edu/biomarker/api hive-lab/biochemistry.edu/biomarker/sparql I don't think this is something I can do. I am not familiar with these. Would this be a task for Robel, or someone else? Daniel Lyman, PhD Senior Research Associate Cancer Biomarkers Project Lead, HIVE-Lab MS in Bioinformatics and Biochemistry Program Coordinator Department of Biochemistry and Molecular Medicine School of Medicine & Health Sciences The George Washington University Ross Hall 2300 Eye Street N.W. Washington, DC 20037 Phone lab: 202-994-3639 MS Program Office hours: T/W 9:00am-5:00pm On Thu, Jun 8, 2023 at 12:09 PM Hadley King @.***> wrote: I don't think this is something I can do. I am not familiar with these. — Reply to this email directly, view it on GitHub https://github.com/biomarker-ontology/Biomarker-Database-Curation/issues/8#issuecomment-1582943068, or unsubscribe https://github.com/notifications/unsubscribe-auth/A4WWYURATDB4N7ND334CTQDXKH2LBANCNFSM6AAAAAAYXMVESI . You are receiving this because you were assigned.Message ID: @.*** .com> Probably. I am not even sure what the paths you supplied are referencing. Where is that? What resource is it? I'll reassigne to Daniall and Robel Daniall Please work with Robel on these issues. Sean has been able to complete these and is working on sparql
gharchive/issue
2023-06-01T19:54:09
2025-04-01T04:33:38.076073
{ "authors": [ "DaniallMasood", "HadleyKing", "danlymangw" ], "repo": "biomarker-ontology/Biomarker-Database-Curation", "url": "https://github.com/biomarker-ontology/Biomarker-Database-Curation/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2737076560
🐛 Biome extension & CLI not consistent VS Code version 1.93.1 Extension version 2.3.0 Biome version 1.9.4 Operating system [ ] Windows [X] macOS [ ] Linux Description Biome CLI (via package.json) and Biome extensions have contradictory behaviors when it come to formatting. Steps to reproduce Here is my biome.json: { "$schema": "https://biomejs.dev/schemas/1.9.4/schema.json", "vcs": { "enabled": true, "clientKind": "git", "useIgnoreFile": true }, "formatter": { "enabled": true, "useEditorconfig": true, "formatWithErrors": true, "indentStyle": "tab", "lineEnding": "lf", "indentWidth": 2, "lineWidth": 100, "attributePosition": "auto", "bracketSpacing": true, "ignore": ["**/dist", "**/coverage"] }, "organizeImports": { "enabled": false }, "linter": { "enabled": false }, "javascript": { "formatter": { "enabled": true, "jsxQuoteStyle": "double", "quoteProperties": "asNeeded", "trailingCommas": "all", "semicolons": "always", "arrowParentheses": "asNeeded", "bracketSameLine": false, "quoteStyle": "single" } }, "json": { "formatter": { "enabled": true, "trailingCommas": "none" } }, "files": { "include": [ "src/**/*.ts", "src/**/*.tsx", "src/**/*.js", "src/**/*.jsx", "src/**/*.json", "test/**/*.ts", "test/**/*.tsx", "test/**/*.js", "test/**/*.jsx", "test/**/*.json" ], "ignore": ["**/node_modules", "**/dist", "**/coverage"] } } Install biome (package & extension) and add the following scripts: "format": "biome check ." run format document via vscode, then run pnpm format --fix Contradictory & self-cancelling behaviors. I believe it started with a recent update. https://github.com/user-attachments/assets/96edd425-564f-4ffa-a063-c130f61dfec2 Expected behavior Expected the same behavior whether vscode, or biome package. Does this issue occur when using the CLI directly? Yes Link to a minimal reproduction No response Logs No response I have also encountered out of sync issue with the extension 2.3.0 with @biomejs/biome 1.9.4 in both windsurf and cursor editors. I resolved this by adding a [project].code-workspace config file at the root of my monorepo something like this : { "folders": [ { "name": "references", "path": "references" }, { "name": "workers", "path": "workers" }, { "name": "app", "path": "app" } ], "settings": { "files.exclude": { "./backup": true } } } Then open the codebase with workspace from the config file, and the extension starts to work in sync with each biome.json. tip: I setup the local .vscode with settings.json in each workspace to enable customized biome behavior
gharchive/issue
2024-12-12T23:38:39
2025-04-01T04:33:38.083573
{ "authors": [ "FaresKi", "lumpinif" ], "repo": "biomejs/biome-vscode", "url": "https://github.com/biomejs/biome-vscode/issues/444", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
247288949
Segmentation Fault Hi, I have tried running Hera on few samples, but it always crashes with a segmentation fault. No other log output is available, so I cannot be more precise about the error. It seems to happen at random times (I have rerun the same sample multiple times). Any ideas? Is there any possibility for you to share your data with us so we could reproduce and fix the issue? Also, could you tell us a little bit more information about the system that you run hera on? Unfortunately, I only have clinical in-house data (which I cannot share). I will try to dig up some RNA data from the public domain and see if I can reproduce the error. Stay tuned! I run Hera as Slurm jobs on a CentOS HPC cluster (http://genome.au.dk/system-info/). Thank you for your bug report & help to trace the source of the error! Hi micknudsen, Jemalloc on Centos could be the cause of the instability. We currently turn it off for stability purpose. Hence, the program should consume a little more memory (1-2GB). Please pull or clone again to check if it work. Best, Bioturing Algorithm Team. Micknudsen, Hopefully, the new fix solves the segmentation fault problem? I am running a test right now. I will let you know as soon as it finishes or (hopefully not) crashes. Problem solved. Thanks!
gharchive/issue
2017-08-02T06:45:56
2025-04-01T04:33:38.101149
{ "authors": [ "GinnyAquarius", "bioturing", "kspham", "micknudsen" ], "repo": "bioturing/hera", "url": "https://github.com/bioturing/hera/issues/6", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2146110471
🛑 Bangla Plex API is down In c120fc0, Bangla Plex API (https://banglaplexapi.bymirrorx.eu.org/popular/) was down: HTTP code: 0 Response time: 0 ms Resolved: Bangla Plex API is back up in d0c936e after 23 minutes.
gharchive/issue
2024-02-21T08:30:37
2025-04-01T04:33:38.106619
{ "authors": [ "bipuldey19" ], "repo": "bipuldey19/pinger", "url": "https://github.com/bipuldey19/pinger/issues/540", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
200762976
How to debug error responses? Greetings, and thank you for writing this plugin! I've included this plugin in my ionic v1 app, inject it into my service as $cordovaFacebookLogin, and try to authenticate to Facebook like this: $cordovaFacebookLogin.login({ permissions: [] }) When I testing on iOS 10.2 I get back JSON that looks like this: {"userID":"1234567890","cancelled":true,"success":false,"accessToken":"token-here","error":true} But I didn't cancel! When I debug the access token here it says it's valid: https://developers.facebook.com/tools/debug/accesstoken My question is: how do I best determine what the error is here? In case it helps, when I run that function the Facebook app opens, and I see a page that is pretty-much blank except for the word "cancel" in white at the top. Then after a few seconds the Facebook app closes and returns to my app with the above JSON. Thanks again! Yeah, I'm getting the same issue here, and having trouble debugging... I actually got a successful response the first time I ran the facebook authentication, but since then I can only get the onFailure method to fire...
gharchive/issue
2017-01-14T00:14:07
2025-04-01T04:33:38.150379
{ "authors": [ "Wade-McDaniel", "samasmith89" ], "repo": "bisrael/cordova-plugin-facebook", "url": "https://github.com/bisrael/cordova-plugin-facebook/issues/23", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1199697306
Remove submodule source packages once published Regarding ROS2: Remove the following git submodules once the packages are published via apt: [ ] ros-rolling-vision-msgs [ ] ros-rolling-soccer-vision-2d-msgs Update bit-bots/bitbots_tools/pull/126 once resolved.
gharchive/issue
2022-04-11T09:41:05
2025-04-01T04:33:38.151946
{ "authors": [ "jaagut" ], "repo": "bit-bots/bitbots_meta", "url": "https://github.com/bit-bots/bitbots_meta/issues/106", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1777789419
github-merge: omit ACKs in quoted comments or with markdown indentation Avoid counting an ACK if it is in a quoted comment or with markdown indentation, when present in a review before the actual ACK. See https://github.com/bitcoin/bitcoin/pull/27334#issuecomment-1610203248 for an example. Concept ACK 9333d5f61bf85a38123a56c9bd63ef162eeec1b2 ACK 9333d5f61bf85a38123a56c9bd63ef162eeec1b2
gharchive/pull-request
2023-06-27T21:37:15
2025-04-01T04:33:38.185004
{ "authors": [ "achow101", "jonatack", "katesalazar" ], "repo": "bitcoin-core/bitcoin-maintainer-tools", "url": "https://github.com/bitcoin-core/bitcoin-maintainer-tools/pull/154", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2178450879
Newsletters: add 293 (2024-03-13) [ ] Lede, releases/RCs, topic entries @harding [ ] Bitcoin Core PR Review Club @LarryRuane I know the review club section is light on questions (only 3), I'm having trouble finding questions that aren't too much in the weeds and also I was out of town all weekend. I have to step out for a few hours but will try to add more questions after I get back home at around 9PM MDT. @glozow - if you have time, please feel free to add review club questions or make any changes to the ones I just pushed, no need to check with me, thanks. Made edits and added lede and releases/RCs. I didn't review added content or add topic links; I'll try to get to those in a few hours.
gharchive/pull-request
2024-03-11T07:44:19
2025-04-01T04:33:38.304663
{ "authors": [ "LarryRuane", "harding" ], "repo": "bitcoinops/bitcoinops.github.io", "url": "https://github.com/bitcoinops/bitcoinops.github.io/pull/1549", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
105570181
0.11B -stealth-mode is not mentioned in command line usage bitcoinxt-qt -? | grep stealth is blank. It does work though, connecting to a 0.11A node using -stealth-mode makes it report. receive version message: /Satoshi:0.11.0/: version 70002, blocks=222611, us=0.0.0.0:0, peerid=2959 Added documentation in https://github.com/bitcoinxt/bitcoinxt/pull/85
gharchive/issue
2015-09-09T10:37:46
2025-04-01T04:33:38.306233
{ "authors": [ "Beanow", "jlopp" ], "repo": "bitcoinxt/bitcoinxt", "url": "https://github.com/bitcoinxt/bitcoinxt/issues/62", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
930704675
Implementation of SWP-08 routing protocol Hey! I would like companion to implement the SWP-08 router protocol. Its a protocol supported by alot of different router-systems like ross ultrix, direct out matrix-routing and lots more. I want to be able to perform routes and read out routing info. Then be able to use the feedback from the router to make a tallysystem. Documentation apllies to both SWP-02 and 08 https://wwwapps.grassvalley.com/docs/Manuals/sam/Protocols and MIBs/Router Control Protocols SW-P-88 Issue 4b.pdf Now i have been testing a bit and here are som points: How can i get the names from the router onto my buttons? should they be stored in variables? I want to have a feedback that i can but on the sources that change color when the destination it is routed to is selected. It should be a variable with name that is the source of the selected destination. I used the "select destination" and when I choose one destination and then another, the previuosly chosen destination inherit the now selected destinations destination(complicated) FYI: Although the version linked at https://we.tl/t-5nurvA8mcQ is a different size, it has the same version number as the last one of 2.2.0 (2.2.0-061e92b1-3423) Re: level multiselect. The observed behaviour of the buttons in the emulator do not seem to reflect multiselect. In normal button mode, each button will flash the active (yellow top bar) indication when pressed and the feed back will light up the last button pushed. In Latch/Toggle button mode, each button will light the active (yellow top bar) indication when pressed; however, it will never turn off. The feed back will still light up the last button pushed. I would guess that the buttons would want to be normal rather than using the predefined Latch/Toggle logic? Is this behavior simply an anomaly of the emulator that is not present with an active router controller? Also, this may be a feature request; but, I see the ability to select more than one level per action & feedback is present; would it be possible to define overlapping level buttons? For example, having four buttons for three levels where the first three are toggle on/off and the fourth is an override to toggle all three selected on or off. Here would be the definitions for these buttons for testing. Button name: L1 Press Action: SW-P-08: Select Levels 1 Feedback: SW-P-08: Selected Levels 1 Button name: L2 Press Action: SW-P-08: Select Levels 2 Feedback: SW-P-08: Selected Levels 2 Button name: L3 Press Action: SW-P-08: Select Levels 3 Feedback: SW-P-08: Selected Levels 3 Button name: All Levels Press Action: SW-P-08: Select Levels 1,2,3 Feedback: SW-P-08: Selected Levels 1,2,3 Selected levels always need to be visible via feedback for those who need to be able to do split routes. As mentioned before, most people will only ever want simultaneous routing; however, for those who need to be able to break out Audio and Video levels, they need to see the current level status directly on the panel via feedback. Turning off all level selection is a common operational procedure rather than a true lockout. Many router control panels are in areas on desks that can be bumped and being able to deselect levels reduces issues; however, usually, the same button is also used to select all levels again rather than to select each level individually so as to minimize the number of intentional button pushes. So a clear option is only useful with a select all option to go with it. Re: lockouts, while SW-P-08 has a robust set of destination lock and router lock options, it is probably best not to support those at all; as those can be much more complex than routing itself, and, if someone needs that level of control they should probably be solving that problem direct with the specific router vendor. I think that using the existing lockout options within Companion is more than sufficient. I think I understand that the selection level buttons need to be additive, so that combinations can be created. This will take a bit of thought. One thing I have done already is the make the clear action more configurable so it can clear only levels if necessary. Thanks. I also notice that there is an additional method to get feedback beyond the way I mentioned above. Feedback: internal: Check variable value SW-P-08:Level - Selected levels = [] That method has available operators of [ =, !=, >, and < ]; however, does not have [AND] [OR] options so it is not clear how that would support multi level feedback. The Feedback: SW-P-08: Selected Levels method on the other hand seems to support multiple levels already so that may not need any further changes. I currently have 2 modes of operation. Firstly: level select and deselct which can take multiple levels source select destination select take or clear This makes one 'set' of actions. Then we also have: crosspoint a level/source/destination in a single action I'm wondering if this second mode takes the level selection in the first method or ignores it? My vote would be to keep the operational modes separate if possible. @evenover do you have a preference? If a single button push is defined, it should have the ability to have control over all aspects of that action regardless of what state the control surface was previously in. For the other mode of operation, the default mode should be that all levels are selected at startup or after a Clear command. The ability to split levels is a more advanced method and should only be enabled if explicitly requested. For example, if I walk up to a control surface and do not know what the previous condition was, this is the typical expected sequence. Hit Clear. {Reset to default condition.} Select a Destination. {This allows the ability to see what is currently routed there.} Select a new Source. Hit the Take button. {Modifying the levels is optional and can be changed at any point before the Take.} This should result in all of levels of the selected source being routed to that destination. If the levels are changed for a route those can then persist until cleared or manually changed. BTW: Some control systems implement a timeout function to Clear the routing presets after a period in case someone has a complex route setup and then walks away, which could cause unexpected results for the next user. It seems to be a good idea @chibiconsulting. A timeout-function would also be great. The way i am planing to use this module is to have kind of a catergory select. So i'm going to build separate pages for Video, audio and then video+audio(video+8 audio probably). This way i am not going to make use of turning levels on and off but just add level-selection to the button with the destination. I think this protocol actually only supports routing 1 level at the time, thats probably why its hard to get feedback of multiple levels selected. I have managed to obtain the latest full documentation of swp08 with extended routing/labels from Grass Valley, but i have to check that my employer do not have a NDA with Grass before i can supply it to you. Hopefully its not a big deal and i will give it to you as soon as i know. @evenover It's great that you have the documentation but remember that companion and this module is open source so GV need to be aware that it's trivial to reverse engineer their spec from the code in this module. Hi guys, in order to get this module out to more potential users it has been included in the next beta of companion as the basic routing functionality is all working. This doesn't mean I won't make the more complex changes requested above. Thats great! Any updates on the features: Change color on background of source of destination when destination is selected variables for label names which can be added to buttons variable with the source name of selected destination 1 and 2 are pretty much the same features as the videohub-module. I took a look in the manual for ultrix and it only supports 8-char labels, but 16char umds of the sources. In the extended documentation its possible to ask for 4, 8 or 12. Hopefully i can disclose it with you asap. I've redone the level select/deselect actions so they work as toggles, next step is to make feedbacks work for those and then the list above. The level commands have been rewritten to work as follows: Select Level: add the level(s) in the action to the current selection DeSelect Level: remove the level(s) in the action from the current selection Feedback: Change button colour if all the level(s) in the feedback are in the current selection You would normally put all of the above on one phyiscal button and enable latch/toggle mode. The clear action re-enables all levels up to the number set in the module configuration. Due to all the changes you will need to delete any previous level actions and feedbacks from your configuration before trying this! https://we.tl/t-RqZEdtxs3c Having done more testing the above method can sometimes get in a mess when the latch state gets out of step with the level state. To avoid this there's a new action which will simply toggle (invert) the level(s) in the action each time it is pressed. This is much preferable to using the button latch mode. https://we.tl/t-LmLxyBaQ26 I like the SW-P-08: Toggle Levels action, nice. Can still see some people using the SW-P-08: Select Levels and SW-P-08: De-Select Levels actions, hope those stay as well. @evenover How can i get the names from the router onto my buttons? should they be stored in variables? I would think that if available by feedback it would be enough, or do you have something else in mind? i.e. Feedback: SW-P-08: Name of target. or Name of Source and Name of Destination for example. 2./3. I want to have a feedback that i can put on the sources that change color when the destination it is routed to is selected. Something like Feedback: SW-P-08: Source routed to Destination? Would need to query the selected destination for the current Source. What happens if the routed Source changes after the button is updated? Could use a lot of cycles to keep checking. I used the "select destination" and when I choose one destination and then another, the previously chosen destination inherit the now selected destinations destination(complicated) I think that Evertz implements this as a virtual destination which is a destination that can then be used as a source somewhere else. I do not recall SW-P-08 ever having that capability. Might make sense as a separate, more generic request to provide that function for all router protocols. If a range of virtual sources/destinations are defined as being linked and held separate from the router, requests to / from those can be translated on the fly and dropped back to the native router protocol. Not a trivial problem by any means. How can i get the names from the router onto my buttons? should they be stored in variables? I would think that if available by feedback it would be enough, or do you have something else in mind? i.e. Feedback: SW-P-08: Name of target. or Name of Source and Name of Destination for example. Feedbacks are really intended for modifying the button colour or text, I think what is being asked here is that the button label can be set from a variable, please confirm. I want to have a feedback that i can put on the sources that change color when the destination it is routed to is selected. Something like Feedback: SW-P-08: Source routed to Destination? Would need to query the selected destination for the current Source. What happens if the routed Source changes after the button is updated? Could use a lot of cycles to keep checking. Currently the router sends a message every time a route changes (look in the debug log to see these) so I need to write some code to keep track of all these messages. What makes it complicated compared to say the videohub is keeping track of all the potential breakaway routes. I used the "select destination" and when I choose one destination and then another, the previously chosen destination inherit the now selected destinations destination(complicated) I think that Evertz implements this as a virtual destination which is a destination that can then be used as a source somewhere else. I do not recall SW-P-08 ever having that capability. Might make sense as a separate, more generic request to provide that function for all router protocols. If a range of virtual sources/destinations are defined as being linked and held separate from the router, requests to / from those can be translated on the fly and dropped back to the native router protocol. Not a trivial problem by any means. This is a quite advanced feature and in my opinion companion is generally aimed at casual or supplementary use cases. For this kind of complexity you should probably be using the Ross panels or similar. Now this module has a repository we can log all these requests separately and this thread is getting too long. I have transferred the feature requests I am aware of to new issues. Please check and update as required. Please log new bugs and features separately. I can't find the repository, could you tell me where to find it? All the wetransfer links are obviously expired. Thanks Now this module has a repository we can log all these requests separately and this thread is getting too long. I have transferred the feature requests I am aware of to new issues. Please check and update as required. Please log new bugs and features separately. @scp85 The latest build is in the beta companion app - https://bitfocus.io/companion/download/builds/ is there any IP based emulator available for SWP-08 routing protocol? is there any IP based emulator available for SWP-08 routing protocol? Not that I'm aware of. What were you thinking using it for? I’m trying to figure out ways to control an Swp-08 audio/video router (grassvalley (used to be Snell Andvanced Media) Sirius S830)) with a PC/ audio desk.. On 26 Jan 2022, at 14:57, evenover @.***> wrote:  is there any IP based emulator available for SWP-08 routing protocol? Not that I'm aware of. What were you thinking using it for? — Reply to this email directly, view it on GitHub, or unsubscribe. Triage notifications on the go with GitHub Mobile for iOS or Android. You are receiving this because you were mentioned. This module should be working fine. I've done some testing on my Ross Ultrix and it looks good for now. If you have over 1000 sources and destinations there are still some developments that need to happen. Ouch! 8000 I/o for me. I’ll still give it a go tho On 26 Jan 2022, at 14:59, Paolo Scillieri @.***> wrote:  I’m trying to figure out ways to control an Swp-08 audio/video router (grassvalley (used to be Snell Andvanced Media) Sirius S830)) with a PC/ audio desk.. On 26 Jan 2022, at 14:57, evenover @.***> wrote:  is there any IP based emulator available for SWP-08 routing protocol? Not that I'm aware of. What were you thinking using it for? — Reply to this email directly, view it on GitHub, or unsubscribe. Triage notifications on the go with GitHub Mobile for iOS or Android. You are receiving this because you were mentioned. If we get the extended comands implemented it will be able to control 65536 sources and destinations. I'm afraid that because I don't require the extended commands in my use cases the development time would need to be paid for somehow. I understand, any idea on what kind of payment? I might be able to fix some funding. I think £200 would cover it. Awesome, that should be no problem! Would it then be possible to see if you could get the "change background color of selected source" working? Maybe... let's do the extended commands first. Please email me to discuss payment mail-at-pjdaniel-dot-org-dot-uk Any chance you would know if this is able to work with a nVision NV9000 Router controller from Grass Valley? Any chance you would know if this is able to work with a nVision NV9000 Router controller from Grass Valley? Hi, I think it could be working, NV9000 is on swp02. This is router protocol swp08 is remote protocol and we are using swp08 for evertz switcher control thru CEREBRUM automatization via swp08 Thanks, hmmmm now I just have to figure out how to set that up lol Hello team, I'm curious to know whether it's possible to disconnect a crosspoint using the Pro-Bel SW-P-08 protocol. Can a crosspoint connection be directly removed, or is the standard procedure to always redirect the input to a specific output? Moreover, is there a way to work-around this behavior? For example, could we route the input to a designated output that can be deactivated on the physical device? Thank you very much in advance ! No. This is not specific to SW-P-8 control, all routers have this constraint. For Video or Audio routers, the standard practice is to route a specific input to any outputs that no longer require the previous input. Examples would be a blank screen and silent audio or bars and tone. Most systems already have a source for those anyway. This is due to the physics of the crosspoint switching mechanism and associated amplification within the device. To put it another way, an unterminated source being output by a router would cause undesirable operation. ;) Note that data routers are an exception, where the router itself should define a disconnect source and you would need to check the specifics with the data sheet for those devices; however, from a protocol perspective, it still places a crosspoint select message using whatever the disconnect source is defined as. Thank you @chibiconsulting for your prompt and informative response; I was seeking for anything like this... All uncertainties are resolved! Thanks, hmmmm now I just have to figure out how to set that up lol I'm wandering if you managed to get it to work ? Because I'm using a NV9000 too but it's not working for me for now Thanks Thanks, hmmmm now I just have to figure out how to set that up lol I'm wandering if you managed to get it to work ? Because I'm using a NV9000 too but it's not working for me for now Thanks Hi we're using SW-P-08 generic and works for us, just fill IP and port what we are using and it works. Truth is that we are using semi-software Cerebrum for manage control of matrix. Means Elgato->swp08->Cerebrum->swp08->Evertz @daniep01 @evenover is this module in good shape now? I'll arrange funding if it can be improved in some way. Hi William, I can probably speak to that reasonably having done a bit of work on the module recently. It works reasonably but the implementation is not very complete. A few obvious areas for improvement: Request and process Tally Dump Better multilevel crosspoint tallies Protect / Extended Protect Set / Tally / Request Tally Dump One of the potential pitfalls with SW P 08 is that there is a large number of commands specified, but few if any devices support all of them. I faced a similar issue with the SW P 02, and only implemented the scope I could test. For instance this is the portions of SW P 08 that a specific Direct Out device supports. Perhaps a quarter of all commands listed, which is probably not complete to begin with. Hello, I’m experiencing issues with the SWP-08 and the Ultrix FR-12 Frame. While I can send control commands, I’m not receiving any feedback or information from the equipment. Has anyone worked with Ross Ultrix systems and encountered a similar issue? Any troubleshooting suggestions would be greatly appreciated. Thanks! Yep, we use it with a smaller Ultrix. I can have a look at the Ultrix side config next week and see if there is anything notable we had to do. Can you please check the module logs for return data, and let me know what you find. Hi Guys, Im a solutions architect with the Hyerconverged team at Ross. The FR12 uses the Ultricore BCS as a Control system. Make sure you are sending your commands to the Ultricore BCS and not to the frame directly. Also it is possible that the probel service is not enabled. We implimented new security features that disables most services by default now (version 5.5 and up). On the BCS under product info, services. I've tested this today with Companion 3.4.3 and Ultricore BCS / Ultrix version 5.6 and 6.2 (with an FR12, FR5, FR2) and it works as expected. Hi Guys, Im a solutions architect with the Hyerconverged team at Ross. The FR12 uses the Ultricore BCS as a Control system. Make sure you are sending your commands to the Ultricore BCS and not to the frame directly. Also it is possible that the probel service is not enabled. We implimented new security features that disables most services by default now (version 5.5 and up). On the BCS under product info, services. I've tested this today with Companion 3.4.3 and Ultricore BCS / Ultrix version 5.6 and 6.2 (with an FR12, FR5, FR2) and it works as expected. Hello, indeed my issue was there; I was not pointing to the virtual IP address of the cluster. The grid now appears perfectly in the module. Thank you!
gharchive/issue
2021-05-15T12:07:26
2025-04-01T04:33:38.368916
{ "authors": [ "Corent1CSN-LAB", "Gbou7", "InertiaImpact", "chibiconsulting", "daniep01", "engelCZE", "evenover", "karthickvkl", "mplakalovi1", "phillipivan", "pomerle", "scp85", "willosof" ], "repo": "bitfocus/companion-module-generic-swp08", "url": "https://github.com/bitfocus/companion-module-generic-swp08/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
895793231
add BitBasicList component (#269 ) This closes #269 Since the assignee of the related issue is changed, I'm closing this PR.
gharchive/pull-request
2021-05-19T19:15:13
2025-04-01T04:33:38.378002
{ "authors": [ "msynk", "zoha-shobbar" ], "repo": "bitfoundation/bitframework", "url": "https://github.com/bitfoundation/bitframework/pull/409", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2025484970
Enter amount : input field appears disabled Even after entering a value, the input field still appears disabled, retaining the disabled appearance despite the entered value. Resolved,
gharchive/issue
2023-12-05T07:42:38
2025-04-01T04:33:38.379766
{ "authors": [ "Deveshshankar", "akshay8551" ], "repo": "bithyve/bitcoin-keeper", "url": "https://github.com/bithyve/bitcoin-keeper/issues/3503", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
940695184
Swan Buy integration Mostly done, need to be tested Verified this on dev app v1.8.0(317) Verified this Swan Buy integration in dev app v1.8.0(317) It is working properly in android device. In ios it is asking to open in "Hexa Stage"? and when we click on "Open" it gives a msg as ""Hexa stage" Beta has expired" and after clicking on "OK" when we again click on "Authorize Hexa wallet" it throws an error as {"message : "invalid_request"}. Still the Swan is not working on both the device android ans ios as well. -Tested on dev app v1.8.0(319) Uploading 21-07-13-19-48-24.mp4… I was able to complete Swan journey in #317. The deeplink didn't open the modal in #319. @cakesoft-swati - are any deep links working? [ ] swan deeplink not opening from the app Now the deeplink is working for swan account. Tested in dev app v1.8.5(325)
gharchive/issue
2021-07-09T11:52:21
2025-04-01T04:33:38.383914
{ "authors": [ "AliMeer", "antuz123", "cakesoft-swati" ], "repo": "bithyve/hexa", "url": "https://github.com/bithyve/hexa/issues/3732", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
137392737
Update forms.py to fix issue #1 improve compatibility with Django version 1.7+ @Kyle-Falconer Thanks a lot!
gharchive/pull-request
2016-02-29T22:08:08
2025-04-01T04:33:38.385125
{ "authors": [ "Kyle-Falconer", "Tyrdall" ], "repo": "bitmazk/django-multilingual-survey", "url": "https://github.com/bitmazk/django-multilingual-survey/pull/2", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
70525435
can.view.href This pull request adds the can-href attribute for easy linking. Not ready to merge yet, finalization incoming. Nevermind, needs to be minor anyway.
gharchive/pull-request
2015-04-23T22:10:35
2025-04-01T04:33:38.515223
{ "authors": [ "daffl" ], "repo": "bitovi/canjs", "url": "https://github.com/bitovi/canjs/pull/1641", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
59858252
Move less plugin into stealjs/system-less I've moved this issue out to stealjs/system-less#1. We should discuss anything about this issue not related to actual core steal features over there. Gonna keep this ticket open for tracking purposes. As part of this, we should actually move the steal plugin out of this repository and out into that one and make sure it's easily installable/accessible. Perhaps we might still want to include the file in steal releases, as well? Yes, I think we should keep it in the releases but moving it to this new repo is a good idea. That way people who want a more current version don't have to wait for a steal release. Also, there's a couple of less tests that can probably be moved. There's a test that checks for steal's integration with less... that one should not be moved, but any that just tests the functionality of the plugin can, I think. This one should be pretty easy for someone looking to contribute. What needs to be done is: [ ] - Move ext/less.js and ext/less-*.js to the system-less repo. [ ] - Move less tests there as well (note: not this one which tests that less is loaded by default) [ ] - Setup CI for that repo [ ] - Update steal Gruntfile to move less.js, less-*.js to the ext folder where it was before. We do the same sort of thing for system-npm so you can use that as an example. This will make it easier to keep the less plugin up-to-date with Less releases.
gharchive/issue
2015-03-04T20:39:32
2025-04-01T04:33:38.518940
{ "authors": [ "matthewp", "zkat" ], "repo": "bitovi/steal", "url": "https://github.com/bitovi/steal/issues/346", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
82437951
Error when creating payout: This endpoint does not support the pos facade I tried the example found here which unfortunately result in an error using testnet. {"error":"This endpoint does not support the `pos` facade"} The request looks as follows POST https://test.bitpay.com:443/payouts HTTP/1.1 ... {"token":"__REPLACED__","amount":10,"currency":"USD","instructions":[{"label":"Test","address":"__REPLACED__","amount":10}],"effectiveDate":"2015-05-29T14:05:38.000Z","pricingMethod":"bitcoinbestbuy","guid":"__REPLACED__","nonce":__REPLACED__,"reference":"123","notificationURL":"__REPLACED__"} Note: REPLACED was placed by me to not disclose private information. @4ley the payouts endpoint requires the payroll facade. As I understood, I have to set it on the token class. Thats what I did. Here is my code: $instruction = new \Bitpay\PayoutInstruction(); $instruction ->setAmount(10) ->setAddress('__REPLACED__') ->setLabel('Test'); $time = gmdate("Y-m-d\TH:i:s\.", time()+3000)."000Z"; $token = new \Bitpay\Token(); $token ->setFacade('payroll') ->setToken('__REPLACED__'); $payout = new \Bitpay\Payout(); $payout ->setEffectiveDate($time) ->setCurrency(new \Bitpay\Currency('USD')) ->setPricingMethod('bitcoinbestbuy') ->setReference('__REPLACED__') ->setNotificationUrl('__REPLACED__') ->setToken($token) ->addInstruction($instruction); Ok, gotcha. It sounds like your account was paired with a POS token originally. That's the least-privileged access token (and the default one created). However, if you make a call to getTokens() it should respond with all of the access tokens associated with your account. If you do have a payroll token, use that one, otherwise the default POS token will be returned and used. Here's the example code for that call: https://github.com/bitpay/php-bitpay-client/blob/master/examples/GetTokens.php Hm, very strange the $client->getTokens() returns NULL @4ley , have you created any tokens? You might want to follow our tutorials. In 001.php, keys are generated and stored. In 002.php, a (POS) token is created. You'll want to modify the create token snippet to the following to create a payroll token: $token = $client->createToken( array( 'facade' => 'payroll', 'label' => 'You can insert a label here', 'id' => (string) $sin, ) ); That'll create a payroll token, but you'll still need to pair it with your account. At the bottom of the same file, after $persistThisValue = $token->getToken(); add $pairingCode = $token->getPairingCode(); And then print out $pairingCode however you like. Log in to your BitPay account and then go to Access Tokens and paste the pairing code and click Find. If everything worked, your account will now be associated with that payroll facade token. Your code gives me the following error: PHP Fatal error: Uncaught exception 'Bitpay\Client\ArgumentException' with message 'pairing code is not legal' in /home/www/includes/bitpay/vendor/bitpay/php-client/src/Bitpay/Client/Client.php:474 What version of the PHP-Bitpay-Client are you using? The VERSION file says 2.0.0 To use non-POS facade tokens, you'll need to be using the unreleased master branch. To do this, in your composer json, the require for our client should like like the following: ... "require": { ... "bitpay/php-client": "dev-master" ... }, ... where the ... are other possible code snippets Okay I made the update, now it gives me PHP Fatal error: Uncaught exception 'Bitpay\Client\BitpayException' with message '500: Facade not allowed' in /home/www/master/includes/bitpay/vendor/bitpay/php-client/src/Bitpay/Client/Client.php:502 I just checked, and you're right, the payroll facade isn't listed. It requires manual approval to get a token for that facade, so you'll need to contact support@bitpay.com with a request for approval. Thanks. I've sent a request.
gharchive/issue
2015-05-29T13:54:43
2025-04-01T04:33:38.532067
{ "authors": [ "4ley", "ionux", "sambohler" ], "repo": "bitpay/php-bitpay-client", "url": "https://github.com/bitpay/php-bitpay-client/issues/145", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
232783729
reconciliation error while trying to fetch reconciliation report getting error like "This endpoint does not support the merchant facade". reconciliation is for merchant facade itself. Fetching a settlement report (https://bitpay.com/api#resource-Settlements) is a 2 step approach: Please use the merchant facade to fetch the settlement report summary for a date range. The response will contain a resource token. Please use the resource token to fetch the settlement report details.
gharchive/issue
2017-06-01T07:03:18
2025-04-01T04:33:38.534284
{ "authors": [ "pieterpoorthuis", "pravithapk" ], "repo": "bitpay/php-bitpay-client", "url": "https://github.com/bitpay/php-bitpay-client/issues/230", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1293725743
[feature] add "MysqlDatabase" "MysqlUser" Add delete policy add "MysqlDatabase" "MysqlUser" Add delete policy Sometimes deleting these resources is very dangerous, so adding a certain label can reduce data loss I publish new branch to megre.review plz @cndoit18
gharchive/pull-request
2022-07-05T03:01:30
2025-04-01T04:33:38.535535
{ "authors": [ "jiuker" ], "repo": "bitpoke/mysql-operator", "url": "https://github.com/bitpoke/mysql-operator/pull/828", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
139445823
Quote tag and branch names Branches with characters like the single-quote (') cause builds to hang indefinitely. This should solve that. Are you sure that there's a problem with this step? We know about an escaping issue, but that's in https://github.com/bitrise-tools/bitrise-machine or in one of the sub-tools it uses to communicate with the build VMs. We'll fix that this week. Thanks - I just know that something hangs somewhere. You're right, it's actually an escaping issue in one of the core tools used for pushing the parameter into the build VM. We'll fix it ASAP! Yay!
gharchive/pull-request
2016-03-09T01:29:12
2025-04-01T04:33:38.542054
{ "authors": [ "MisterRager", "viktorbenei" ], "repo": "bitrise-io/steps-git-clone", "url": "https://github.com/bitrise-io/steps-git-clone/pull/11", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1573342170
Revert to 6.2.2 Checklist [x] I've read and followed the Contribution Guidelines [x] step.yml and README.md is updated with the changes (if needed) Version Requires a PATCH version update Context Users were experiencing an issue where upon checking out the merge ref for a PR the merge ref was referencing an old commit which is incorrect. Changes Copied the code from the version 6.2.2 release and kept the changes for GIT_CLONE_COMMIT_COMMITTER_NAME and GIT_CLONE_COMMIT_COMMITTER_EMAIL. Investigation details Decisions This is not a minor version change. When migrating from 6.x to 7.x the behavior was changed so in some cases you could the explicit depth, and manual_merge was removed. With bringing back the behavior from 6.x, people who have changed their settings for 7.x now have to change them once again. And a lot of people do not check the changelog of minor version changes. Especially one with as little explanation as a cryptic "Revert to 6.2.2 by @vshah23 in #203" This is not a minor version change. When migrating from 6.x to 7.x the behavior was changed so in some cases you could remove the explicit clone depth, and manual_merge was removed. With bringing back the behavior from 6.x, people who have changed their settings for 7.x now have to change them once again. And a lot of people do not check the changelog of minor version changes. Especially one with as little explanation as a cryptic "Revert to 6.2.2 by @vshah23 in #203" Agreed, we had clone depth set to -1 as per the previous instructions, and this has now just started failing with fatal: depth -1 is not a positive number
gharchive/pull-request
2023-02-06T21:49:13
2025-04-01T04:33:38.548727
{ "authors": [ "ilockett", "vincentisambart", "vshah23" ], "repo": "bitrise-steplib/steps-git-clone", "url": "https://github.com/bitrise-steplib/steps-git-clone/pull/203", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
916818317
Update pureconfig to 0.16.0 Updates com.github.pureconfig:pureconfig from 0.12.3 to 0.16.0. GitHub Release Notes - Changelog - Version Diff I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "com.github.pureconfig", artifactId = "pureconfig" } ] labels: library-update, semver-minor Superseded by #300.
gharchive/pull-request
2021-06-10T02:14:48
2025-04-01T04:33:38.552363
{ "authors": [ "scala-steward" ], "repo": "bitrockteam/kafka-dvs-streams", "url": "https://github.com/bitrockteam/kafka-dvs-streams/pull/275", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1059170977
compile secp256k1 error. ./build_secp256k1_ios.sh Cloning into 'secp256k1_ios'... remote: Enumerating objects: 2615, done. remote: Total 2615 (delta 0), reused 0 (delta 0), pack-reused 2615 Receiving objects: 100% (2615/2615), 1.10 MiB | 260.00 KiB/s, done. Resolving deltas: 100% (1865/1865), done. make: *** No rule to make target distclean'. Stop. building thin libraries... [arm64] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. [armv7s] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. [x86_64] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. [i386] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. [armv7] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. building fat binaries... ./build_secp256k1_ios.sh: line 86: cd: /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/btspp_output/thin-libs/arm64/lib: No such file or directory error: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/lipo: no input files specified usage: lipo <input_file> [ ...] command is one of: -archs -create [-arch_blank <arch_type>] -detailed_info -extract <arch_type> [-extract <arch_type> ...] -extract_family <arch_type> [-extract_family <arch_type> ...] -info -remove <arch_type> [-remove <arch_type> ...] -replace <arch_type> <file_name> [-replace <arch_type> <file_name> ...] -thin <arch_type> -verify_arch <arch_type> ... options are one or more of: -arch <arch_type> <input_file> -hideARM64 -output <output_file> -segalign <arch_type> cp: /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/btspp_output/thin-libs/arm64/include: No such file or directory done. target dir: secp256k1_ios/btspp_output/fat-libs how to fix it? thank you. build for android is same error: ./build_secp256k1_android.sh Cloning into 'secp256k1_android'... remote: Enumerating objects: 2615, done. remote: Total 2615 (delta 0), reused 0 (delta 0), pack-reused 2615 Receiving objects: 100% (2615/2615), 1.10 MiB | 180.00 KiB/s, done. Resolving deltas: 100% (1865/1865), done. make: *** No rule to make target distclean'. Stop. building all arch libraries... [arm64] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. [armv7] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. [x86] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. [x86_64] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target `install'. Stop. done. target dir: secp256k1_android/btspp_output Please copy it to the MAC directory to run ------------------ 原始邮件 ------------------ 发件人: "bitshares/bitshares-mobile-app" @.>; 发送时间: 2021年11月20日(星期六) 晚上11:43 @.>; @.***>; 主题: Re: [bitshares/bitshares-mobile-app] compile secp256k1 error. (Issue #9) build for android is same error: ./build_secp256k1_android.sh Cloning into 'secp256k1_android'... remote: Enumerating objects: 2615, done. remote: Total 2615 (delta 0), reused 0 (delta 0), pack-reused 2615 Receiving objects: 100% (2615/2615), 1.10 MiB | 180.00 KiB/s, done. Resolving deltas: 100% (1865/1865), done. make: *** No rule to make target distclean'. Stop. building all arch libraries... [arm64] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. [armv7] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. [x86] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop. [x86_64] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target `install'. Stop. done. target dir: secp256k1_android/btspp_output — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub, or unsubscribe. Triage notifications on the go with GitHub Mobile for iOS or Android. Thank you for your reply. where is MAC directory? MAC is another repo? I use MacOS, only git clone this repo, and compile it with the guide document. The repository is cloned to run under the macOS path, not in the /Volume xxx directory. ------------------ Original ------------------ From: bitcard @.> Date: Tue,Nov 23,2021 9:39 AM To: bitshares/bitshares-mobile-app @.> Cc: YANG @.>, Comment @.> Subject: Re: [bitshares/bitshares-mobile-app] compile secp256k1 error. (Issue#9) Thank you for your reply. where is MAC directory? MAC is another repo? I use MacOS, only git clone this repo, and compile it with the guide document. — You are receiving this because you commented. Reply to this email directly, view it on GitHub, or unsubscribe. Triage notifications on the go with GitHub Mobile for iOS or Android.
gharchive/issue
2021-11-20T15:41:32
2025-04-01T04:33:38.581644
{ "authors": [ "bitcard", "hanomirin" ], "repo": "bitshares/bitshares-mobile-app", "url": "https://github.com/bitshares/bitshares-mobile-app/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
217932233
include time to the deadline email notification As an Applicant I want to see the time of the deadline in the submission deadline email, so I know the exact time my submission is due https://trello.com/c/PTMmcKOR/555-qae17imp-as-an-applicant-i-want-to-see-the-time-of-the-deadline-in-the-submission-deadline-email-so-i-know-the-exact-time-my-sub :peter:
gharchive/pull-request
2017-03-29T16:24:08
2025-04-01T04:33:38.638393
{ "authors": [ "dreamfall", "rusllonrails" ], "repo": "bitzesty/qae", "url": "https://github.com/bitzesty/qae/pull/1190", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
939475858
get_analog_in only shows volts? We would like our unit to read 4...20 mA current but looking at the source code it seems it only displays volts? even then comparing it to the multi-meter, the AIN1 is receiving 22 volts, and get_analog_in is showing 10.647, how do we read the current coming in at AIN1 and AIN2? The Monarco HAT can support 4-20mA signals, however this is not supported in this repository at this moment. You can easily convert 2-10V to a 4-20mA signal by using a 500 Ohm resistor, please read this Also, you should not connect a voltage higher than 10V to the analog inputs as this could overload and break the hardware. I ran two 1000Ohm resistors in parallel, to produce a 500Ohm resistor, it worked fine, but the EM fields in the area are messing with the 10vdc signal and introducing noise, this is fine for now, but we would really appreciate 4-20mA signals in the near future
gharchive/issue
2021-07-08T05:10:21
2025-04-01T04:33:38.647786
{ "authors": [ "bjarnekvae", "vendablefall" ], "repo": "bjarnekvae/pymonarco-hat", "url": "https://github.com/bjarnekvae/pymonarco-hat/issues/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
701366644
[Feature request] Alingments: AMR <-> AMR and AMR <-> Sentence Hi, first, I want to say: You deserve a medal for creating this library. It is the first time I installed an AMR parser without getting a little headache :-). Also it's a nice idea to wrap the noRECAT variant of GSII and ablate all external java preprocessing. I think the noRECAT version may also be more robust. I have two suggestions of which I think they would be cool to have in an amrlib: AMR2sent alignment: As far as I know, there exist aligners (for instance as pre-processing of JAMR parser), that align AMR nodes to tokens. Since often lemmas of the sentence are projected into the AMR graph, a simple string match, maybe with some additional rules, could make up a first solid method. Maybe there are other methods that are more suitable and also easy-to-use. AMR2AMR variable alignment: This could be useful, e.g., for computing AMR metrics or enabling sentence retrieval via AMR parsed corpora or sentence similarity computation via AMR. It is an NP hard problem but can be implemented via hill climbing maximizing triple matching. I have been working on this lately, here is a repo containing AMR metrics (Smatch and S2match) that are based on this alignment. Both alignments should be quite easy to implement in the lib, since it's all native python. (It could be worthwhile, though, to make the alignment faster, e .g., using cython, since it can be very slow for graphs with many variables) Alas, these are just suggestions which may or may not be useful to have (in some near or distant future). Again, thanks for your awesome amrlib! I partially ported the JAMR aligner a while back and I can add this to the lib. I'd like to find the source for the aligner used for the LDC data but so far haven't been able to. I'll dig into this a bit next weeek as I'm OOO for a while The integrated aligner would be very helpful! Question to anyone interested in the alignment functionality... I have a very simple rule based, "word aligner" from JAMR that I ported to python. This is not the more complicated "span aligner" (ie.. phrases) that is the default alignment method in JAMR. That code is much more complicated (and in scala) so it would be a fair amount of work to port it into amrlib. Question: Is a simple "word aligner" enough or would this only be useful if we have full span alignments? There are better span aligners out there (model based) since the early (rule based) JAMR one. I found the code for an ISI aligner (which may have been used to annotate LDC2020T02). That code is basically a bunch of scripts to train/run a model using the C++ MGIZA++ library. It would be a bit of a project, but not completely unreasonable, to make this usable either as part of amrlib, or more likely, as a stand-alone lib. Question: How important (ie.. how much effort is it worth) to include a span aligner and if you think its important, can you give me the use cases for it? The reason why I want an aligner in amrlib is that "JAMR" is hard to install and somehow hard to integrate with python scripts. I have seen several papers using JAMR because it is considered as a light aligner compared to model-based ones. So I think as long as the aligner performance is reasonable and easy to install, people like me would love to use it. I agree with @lujiaying . I also think that, from reading the AMR guidelines, there is no "theory of alignment". Therefore, maybe the token-node alignment may even be a bit more clear-cut than node-span. E.g. # The frog jumps. (j / jump-01 :arg0 (f /frog) With token alignment it's clear that (f,frog) and (j, jumps) are correct alignments, however both (f, the frog) and (f, frog) can be considered correct in span alignment. On the other hand, maybe span has advantages when concepts are really abstract In sum, I think that a lightweight node-token alignment that performs reasonable would be a very good start, since it anchors the AMR in the sentence, which may already be quite helpful for some tasks. I have updated the master project with Rule Based Word aligner, similar in function to JAMR's word aligner. If you have comments or bugs, please post a new issue. (update to pip install to follow).
gharchive/issue
2020-09-14T19:28:54
2025-04-01T04:33:38.655776
{ "authors": [ "bjascob", "flipz357", "lujiaying" ], "repo": "bjascob/amrlib", "url": "https://github.com/bjascob/amrlib/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1582852933
Simple way to disable AuthProvider in development environment? Is there a simple way to render the AuthProvider non-functional (but present) if process.env.NODE_ENV === "development"? Additionally, since useAuth() is a hook, it can't be conditionally called based on whether the environment is development, so I can't just omit AuthProvider in dev and not call useAuth() Hm, what's the use-case for this? Simplify frontend development? Yeah, to skip the login requirement during development. I've achieved this now with a combination of: If environment is dev, set autoSignIn to false If environment is dev, return app directly instead of a login prompt in App.tsx
gharchive/issue
2023-02-13T18:28:44
2025-04-01T04:33:38.660232
{ "authors": [ "cobraz", "denniskrq" ], "repo": "bjerkio/oidc-react", "url": "https://github.com/bjerkio/oidc-react/issues/941", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
183187895
Got examples to work a bit more obviously Added npm scripts: npm run ex:svg npm run ex:group Great, thanks. Published as 2.4.2
gharchive/pull-request
2016-10-15T04:22:00
2025-04-01T04:33:38.662001
{ "authors": [ "bjnortier", "skawaguchi" ], "repo": "bjnortier/dxf", "url": "https://github.com/bjnortier/dxf/pull/14", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1633616029
Update to Cranelift 0.94 Fixes https://github.com/bjorn3/rustc_codegen_cranelift/issues/1249 through https://github.com/bytecodealliance/wasmtime/pull/5619 Fixes https://github.com/bjorn3/rustc_codegen_cranelift/issues/1349 through https://github.com/gimli-rs/object/pull/475. cc @PROMETHIA-27 and @frederikhors
gharchive/pull-request
2023-03-21T10:29:38
2025-04-01T04:33:38.663984
{ "authors": [ "bjorn3" ], "repo": "bjorn3/rustc_codegen_cranelift", "url": "https://github.com/bjorn3/rustc_codegen_cranelift/pull/1363", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
236169276
Offscreen rendering with MSAA How can I draw in a color texture of a framebuffer with MSAA enabled? bgfx::reset(width, height, BGFX_RESET_MSAA_X16); // .... bgfx::TextureHandle m_gbufferTex[2]; m_gbufferTex[0] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::BGRA8, BGFX_TEXTURE_RT | BGFX_TEXTURE_U_CLAMP | BGFX_TEXTURE_V_CLAMP); m_gbufferTex[1] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::D24, BGFX_TEXTURE_RT_WRITE_ONLY); bgfx::FrameBufferHandle m_gbuffer = bgfx::createFrameBuffer(BX_COUNTOF(m_gbufferTex), m_gbufferTex, true); // .... bgfx::readTexture(bgfx::getTexture(m_gbuffer, 0), mem); Rendering into a window works perfectly well with MSAA, but with this code my render result is not antialiased. Here are the debug infos of the cubes example: http://imgur.com/a/csdRo See example-09-hdr it uses MSAA... Toggle it with F8: https://github.com/bkaradzic/bgfx/blob/master/examples/09-hdr/hdr.cpp#L285 ((msaa + 1) << BGFX_TEXTURE_RT_MSAA_SHIFT) If I add this flag to my framebuffer textures, nothing gets rendered at all. I modified my code to match the example, but no progress. I'm also rendering only simple geometry with very basic shaders. Do you have any idea what the problem could be? Maybe reading directly from the framebuffers color texture? In the example the texture is used in other shader and at the and directly drawn to the backbuffer. So I made a minimalistic example, rendering a cube: int main(int argc, char** argv) { uint32_t width = 1024; uint32_t height = 1024; SDL_Window* window = SDL_CreateWindow("bgfx_ros", SDL_WINDOWPOS_UNDEFINED, SDL_WINDOWPOS_UNDEFINED, width, height, SDL_WINDOW_RESIZABLE); SDL_SysWMinfo wmi; SDL_VERSION(&wmi.version); if (!SDL_GetWindowWMInfo(window, &wmi) ) { std::cerr << "couldn't get wm info" << std::endl; return 1; } bgfx::PlatformData pd; pd.ndt = wmi.info.x11.display; pd.nwh = reinterpret_cast<void*>((uintptr_t)wmi.info.x11.window); pd.context = NULL; pd.backBuffer = NULL; pd.backBufferDS = NULL; bgfx::setPlatformData(pd); if (!bgfx::init(bgfx::RendererType::Count, BGFX_PCI_ID_NONE)) { std::cerr << "bgfx init failed" << std::endl; return -1; } bgfx::reset(width, height, BGFX_RESET_MSAA_X16); bgfx::setViewClear(0, BGFX_CLEAR_COLOR | BGFX_CLEAR_DEPTH, 0x00000000, 1.0f, 0); std::vector<uint8_t> vresult; bgfx::ShaderHandle vhandle; createShaderFromFile("vs_object_instanced.bin", vresult, vhandle); std::vector<uint8_t> fresult; bgfx::ShaderHandle fhandle; createShaderFromFile("fs_object.bin", fresult, fhandle); bgfx::ProgramHandle program; const bool destroy_shader = true; program = bgfx::createProgram(vhandle, fhandle, destroy_shader); if (!isValid(program)) { std::cerr << "creating shader program failed" << std::endl; return 1; } // Create vertex stream declaration. PosVertex::init(); std::vector<PosVertex> vertices; std::vector<uint16_t> indices; loadOBJ("/home/david/Desktop/cube.obj", vertices, indices); bgfx::VertexBufferHandle vbh = bgfx::createVertexBuffer(bgfx::makeRef(static_cast<void*>(vertices.data()), vertices.size() * sizeof(PosVertex)), PosVertex::ms_decl); bgfx::IndexBufferHandle ibh = bgfx::createIndexBuffer(bgfx::makeRef(static_cast<void*>(indices.data()), indices.size() * sizeof(uint16_t))); **uint32_t msaa = (BGFX_RESET_MSAA_X16&BGFX_RESET_MSAA_MASK)>>BGFX_RESET_MSAA_SHIFT; bgfx::TextureHandle m_gbufferTex[2]; // does not work - MSAA m_gbufferTex[0] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::BGRA8, ((msaa + 1) << BGFX_TEXTURE_RT_MSAA_SHIFT) | BGFX_TEXTURE_U_CLAMP | BGFX_TEXTURE_V_CLAMP); m_gbufferTex[1] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::D24, ((msaa + 1) << BGFX_TEXTURE_RT_MSAA_SHIFT) | BGFX_TEXTURE_RT_WRITE_ONLY ); // WORKS - no MSAA //m_gbufferTex[0] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::BGRA8, BGFX_TEXTURE_U_CLAMP | BGFX_TEXTURE_V_CLAMP); //m_gbufferTex[1] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::D24, BGFX_TEXTURE_RT_WRITE_ONLY );** bgfx::FrameBufferHandle m_gbuffer = bgfx::createFrameBuffer(BX_COUNTOF(m_gbufferTex), m_gbufferTex, true); float eye[3] = { 2.0f, 2.0f, -5.0f }; float at[3] = { 0.0f, 0.0f, 0.0f }; float view[16]; bx::mtxLookAtRh(view, eye, at); float proj[16];; bx::mtxProjRh(proj, 60.f, 1.f, 0.1f, 100.0f, true); bgfx::setViewFrameBuffer(0, m_gbuffer); bgfx::setViewTransform(0, view, proj); bgfx::setViewRect(0, 0, 0, uint16_t(width), uint16_t(height)); if (!bgfx::isValid(m_gbuffer) ) { throw std::runtime_error("framebuffer not valid!"); } // This dummy draw call is here to make sure that view 0 is cleared // if no other draw calls are submitted to view 0. bgfx::touch(0); // Set vertex and index buffer. bgfx::setVertexBuffer(vbh); bgfx::setIndexBuffer(ibh); // Set render states and submit bgfx::setState(BGFX_STATE_RGB_WRITE|BGFX_STATE_ALPHA_WRITE); bgfx::submit(0, program); void* mem = malloc(width*height*4); bgfx::readTexture(m_gbufferTex[0], mem); bgfx::frame(); bgfx::destroyIndexBuffer(ibh); bgfx::destroyVertexBuffer(vbh); bgfx::destroyProgram(program); bgfx::destroyFrameBuffer(m_gbuffer); bgfx::destroyTexture(m_gbufferTex[0]); bgfx::destroyTexture(m_gbufferTex[1]); bgfx::shutdown(); // get texture cv::Mat renderResult(height, width, CV_8UC4, mem); // flip it (correction, since the opengl stores the data differently) cv::flip(renderResult, renderResult, 0); cv::imshow("testim", renderResult); cv::waitKey(0); return 0; } This is the rendering result without MSAA: http://imgur.com/a/2D0H8 What am I doing wrong? :( bgfx::setState(BGFX_STATE_RGB_WRITE|BGFX_STATE_ALPHA_WRITE); add |BGFX_STATE_MSAA Just tested it and no success. But thanks for looking at it! Textures are missing BGFX_TEXTURE_RT This is wrong: uint32_t msaa = (BGFX_RESET_MSAA_X16&BGFX_RESET_MSAA_MASK)>>BGFX_RESET_MSAA_SHIFT; It should be 0 for 2x, 1 for 4x etc. or just use BGFX_TEXTURE_RT_MSAA_X16 directly instead of ((msaa + 1) << BGFX_TEXTURE_RT_MSAA_SHIFT). Thanks for the hint, but adding BGFX_TEXTURE_RT_MSAA_X16 to the color texture or the depth texture (or both) yields to a black screen. So, the same result as the mess with MSAA_MASK+MSAA_SHIFT (which I took from the hdr example)... I want to do rendering on a server application (with msaa) and read the result in CPU memory for further processing. Maybe I can get past this problem by just reading the render result from the SDL window. This seems not very elegant to me, but MSAA works when I do on-screen rendering. . Try copying what the hdr example does with bgfx::readTexture. You create a separate texture with BGFX_TEXTURE_READ_BACK flag and blit the multisampled texture into that. Also, you should call bgfx::frame until it returns a frame number >= the return value from bgfx::readTexture. See the documentation on bgfx::readTexture. Thank you so much! The trick is to blit the color texture of the framebuffer onto another texture, as you suggested! I assumed I could read the fbo color texture directly, because it worked without MSAA. That was a long struggle. Thanks for your help!! You should use SAMPLER2DMS if you want to read MSAA directly.
gharchive/issue
2017-06-15T12:05:38
2025-04-01T04:33:38.678955
{ "authors": [ "bkaradzic", "jpcy", "tuxa" ], "repo": "bkaradzic/bgfx", "url": "https://github.com/bkaradzic/bgfx/issues/1159", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
278717166
CMake as an official build option A very large number, if not most C++ projects are based on the CMake build system. It would be a huge benefit to these projects if bgfx, in addition to GENie, also officially supported the CMake build system. @JoshuaBrookover has a fork of bgfx , (https://github.com/JoshuaBrookover/bgfx.cmake), that adds CMake. Would it be any way possible to incorporate Joshua's CMake system into the official bgfx repository? The addition of a supported CMake build would greatly increase the target audience of bgfx. I know of a number of developers who are hesitant to try bgfx because their projects are based are CMake, and using a forked version instead of the official bgfx is less likely to remain stable in the long term. I'm in no way suggesting removing or changing GENie, just offering a official option of a CMake build in addition to the GENie build. I would be happy to help to maintain the CMake build. Why not just use compiled version? To my knowledge, there are no bgfx compiled binary packages for RHEL or OSX (2 of our primary dev platforms). We support Windows, Mac, RHEL and Ubuntu. We currently build all our 3rd party dependencies, because traditionally it's been simpler to include them as git submodules, and include them as part of the build as opposed to trying to script a variety of different package managers. Yes, I understand where your use case is coming from, and I think it's a completely valid use case for some libraries. Though in this case bgfx has a lot of 3rdparty deps and generate not only a library but also some binaries, so some folks here (including me) just went with building bgfx and including lib/bin files + headers in the main repository, like here for example https://github.com/beardsvibe/leengine/tree/master/3rdparty/bgfx This approach removes a lot of headache trying to fight build system on every platform every time I build my project, instead I just need to do it once when I'm building bgfx (which is once in a while :) ). I'm all for an officially maintained CMake build system. Mine is far from perfect. I think @bkaradzic doesn't want to deal with it though. It's more likely GENie will automatically spit out CMake scripts for you eventually. I'm not too familiar with how GENie works though. I worked at a company where we used CMake (hence the need for bgfx.cmake). Building bgfx from source each time was very beneficial to us as it allowed us to modify the source code and debug problems with minimum hassle. That's really where my project is most useful. I don't want maintain CMake configs at all. I looked into CMake 10 years ago before I start using Premake and I didn't like it. Fact that it's popular doesn't mean I want to use it. But I'm fine if there is someone who wants to do that in external repo. The addition of a supported CMake build would greatly increase the target audience of bgfx. I know of a number of developers who are hesitant to try bgfx because their projects are based are CMake, and using a forked version instead of the official bgfx is less likely to remain stable in the long term. I'm totally fine with that. If someone chooses to write their own renderer just based on project generator (GENie is not build system, rather it's project generator) used by bgfx, that means that bgfx doesn't really offer them anything they need. Btw, GENie has CMake project generator, but it's minimal required for CLion to work with it: https://github.com/bkaradzic/GENie/tree/master/src/actions/cmake You can contribute there if you want bgfx to support CMake, that's the only way to be in main repo... ;) At least you're not ashamed to admit that your opinion of CMake is based on an experience from 10 years ago. A LOT has changed since then. A LOT has changed since then. Not my opinion about it. https://github.com/widberg/bgfx.cmake is valid? @Yairama Ask there...
gharchive/issue
2017-12-02T20:59:40
2025-04-01T04:33:38.687905
{ "authors": [ "AndySomogyi", "JoshuaBrookover", "Yairama", "bkaradzic", "jimon", "minexew" ], "repo": "bkaradzic/bgfx", "url": "https://github.com/bkaradzic/bgfx/issues/1284", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
110376998
Variables not being interpolated when using Procfile Hello so I have a .env with: MONGO_PORT=17017 TEST=${MONGO_PORT} MONGOHQ_URL="mongodb://localhost:${MONGO_PORT}/database" and a script: #!/usr/bin/env ruby require File.expand_path('../../config/environment', __FILE__) puts Rails.env puts ENV['TEST'] puts ENV['MONGOHQ_URL'] If I run it with bundle exec ./bin/test, the output is correct development 17017 mongodb://localhost:17017/database but if I add to a Procfile test: bundle exec ./bin/test and run foreman start test then the variable is not interpolated: 14:32:51 test.1 | started with pid 17789 14:32:57 test.1 | development 14:32:57 test.1 | ${MONGO_PORT} 14:32:57 test.1 | mongodb://localhost:${MONGO_PORT}/database 14:32:57 test.1 | exited with code 0 14:32:57 system | sending SIGTERM to all processes This might not be a dotenv issue, but I'm kinda lost on how to debug this, any place I should be looking at? Also I should add that the .env is being loaded just not interpolating the variables. So after debugging inside dotenv I realised that the parsing and substitutions were happening. The issue here is that foreman will also load a .env file if it finds it, and it seems dotenv give priority to already defined env variables, which is ok. Here the fix is to move those variables into a .env.local or .env.<environment> file.
gharchive/issue
2015-10-08T05:34:22
2025-04-01T04:33:38.691865
{ "authors": [ "GriffinHeart" ], "repo": "bkeepers/dotenv", "url": "https://github.com/bkeepers/dotenv/issues/218", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
90470627
jobs processed even when queue is not running I have an app with Rails 4.2.2 (using mongoid) running qu from master branch, and I'm seeing this strange behavior. On my model_obserer I have this code def after_save(message) room = "room_#{message.room.uid}" SendMessageJob.create(room, message.id) end I started my server, but didn't ran the command bundle exec rake qu:work but the messages were still delivered to the rooms... This is unexpected, is there something wrong? Can you provide a small rails app/example that shows the issue? It is really hard to say based on the tiny bit of code you have provided. well, there is not much than that, actually. I installed qu and qu-mongoid, moved the method responsible to send the message to a Job and enqueued on after_save...that's it I still having this issue. Here is the whole process: message_controller def create message = Message.new(message_params) message.room = room message.user = current_user message.save MetricServices.message_sent(current_user.uid, room: message.room.name) render nothing: true end message_observer def after_save(message) room = "room_#{message.room.uid}" SendMessageJob.create message.id end send_message_job class SendMessageJob < Qu::Job def initialize(message_id) @message_id = message_id end def perform message = Message.find @message_id room = "room_#{message.room.uid}" WebsocketRails[room].trigger :new_message, message.present.to_html end end I'm starting the app with rails server and JUST THIS...the message is still processed. Also, I didn't saw anything on my mongo database. is this right? should I see something, like a queue collection or whatever? anything? A couple things. I'm not 100% positive that master is stable right now. We've been tweaking things for a while to make some big breaking API changes for the long term good. It should be stable, but if it is, the API will change soon, so I wouldn't pin to master. I'd pin to the latest release and upgrade later. Second, where are you configuring qu? It seems like the qu-immediate backend is being used or something. I'll try to change to the latest stable and actually, I haven't any config file for qu...only the bundle exec rake qu:work on the Procfile but it is commented, so it is not executing
gharchive/issue
2015-06-23T19:27:18
2025-04-01T04:33:38.697042
{ "authors": [ "jnunemaker", "luizkowalski" ], "repo": "bkeepers/qu", "url": "https://github.com/bkeepers/qu/issues/102", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
765609363
[Gally]: master <- dev Automatically created by Git-Ally :tada: This PR is included in version 1.10.202 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket: :tada: This PR is included in version 1.10.202 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2020-12-13T19:24:03
2025-04-01T04:33:38.735520
{ "authors": [ "MrsFlux" ], "repo": "blackflux/lambda-example", "url": "https://github.com/blackflux/lambda-example/pull/1627", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
322856571
Purchases API Hey guys, I noticed there's no implementation for the purchases API (https://dev.recurly.com/docs/create-purchase). Any plans on adding the support for it? I most likely won't get to it any time soon. @MOZGIII This was just merged into master.
gharchive/issue
2018-05-14T15:02:51
2025-04-01T04:33:38.746177
{ "authors": [ "MOZGIII", "cristiangraz" ], "repo": "blacklightcms/recurly", "url": "https://github.com/blacklightcms/recurly/issues/50", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2764185890
Miningcore Solo Mining pool based on rmt-miningcore - Pool wallet to miners wallet tranfer fails I have setup miningcore solo mining pool on docker container that is working fine. Though facing following issue: As per above image pending shares transferred to pending balance but these are not transferring to PAID balance with following error. [2024-12-31 03:45:02.9038] [I] [khaaaicpu] [Bitcoin Payout Handler] Paying 110 CPU to 1 addresses [2024-12-31 03:45:03.1267] [E] [khaaaicpu] [Bitcoin Payout Handler] sendmany returned error: Fee estimation failed. Fallbackfee is disabled. Wait a few blocks or enable -fallbackfee. code -6 Any advise how what to change in config.json file. thanks Hi there, We use GitHub issues as a place to track bugs and other development-related issues. Please see the link below to our dedicated support line: Help Center : Click Here Ticket ID: WR240 Note: Click on the live chat icon at the bottom corner of the page to start a conversation. @KHPak2023 Dear user I understand that you have been experiencing some issues and I want to ensure you get the support you need as quickly as possible by using our live chat option. Support : Support 1. Access the Live Chat: • Navigate to our website’s home page. • Look for the chat icon located at the bottom right corner of the screen and start a 💬. Hello, I have no idea which coin you are trying to mine here, it would have been a good idea to at least provide its name. The error which is returned by the node, seems to actually offer a good options, run your node with the option -fallbackfee Hello @KHPak2023 A support ticket has been opened for you. Kindly Talk to Support on the Live support page so that your issue can be looked into. CLICK HERE Hello, I have no idea which coin you are trying to mine here, it would have been a good idea to at least provide its name. The error which is returned by the node, seems to actually offer a good option, run your node with the option -fallbackfee Thank you very much for response. The coin is cpuchain Thanks it is done by adding fallback=0.00001 to coin.conf file. Thanks for prompt response and help
gharchive/issue
2024-12-31T09:55:58
2025-04-01T04:33:38.758029
{ "authors": [ "Justin-skies", "KHPak2023", "MatthewDEVo", "blackmennewstyle", "mark8700-bit" ], "repo": "blackmennewstyle/miningcore", "url": "https://github.com/blackmennewstyle/miningcore/issues/205", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1481587176
Quick edit YAML properties from dataview table (like in database folder) Is your feature request related to a problem? Please describe. A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] I'm always frustrated when I create a dataview table & need to visit each individual file & use metaedit to change YAML properties. Describe the solution you'd like A clear and concise description of what you want to happen. Ability to change YAML properties for files in dataview table in a similar fashion to that in the database folder plugin i.e. with an assigned & appropriate data type (dropdowns, date & time, etc). Describe alternatives you've considered A clear and concise description of any alternative solutions or features you've considered. Using database folder - issue is that it can't be embedded properly into other docs. Using metaedit plugin - issue is that it still requires navigating away from the file to change the frontmatter & also does not support ease of use data types with drop downs, dates & times etc. Hello, this feature request is already created multiple times, like in #749 or #116. Closing this as duplicate in favor of the existing ones.
gharchive/issue
2022-12-07T10:35:31
2025-04-01T04:33:38.762364
{ "authors": [ "danrfletcher", "s-blu" ], "repo": "blacksmithgu/obsidian-dataview", "url": "https://github.com/blacksmithgu/obsidian-dataview/issues/1644", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1110383347
Typo: Task instead of Calendar in a heading Please provide a link to the documentation page and section https://blacksmithgu.github.io/obsidian-dataview/query/queries/?query=title#task-queries_1 Describe the problem The says "Task Query", even though is should be "Calendar Query" Thanks! Fixed for 0.4.23.
gharchive/issue
2022-01-21T11:52:07
2025-04-01T04:33:38.764477
{ "authors": [ "blacksmithgu", "chrisgrieser" ], "repo": "blacksmithgu/obsidian-dataview", "url": "https://github.com/blacksmithgu/obsidian-dataview/issues/799", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1228716821
🛑 Discord Akary Bot is down In db19567, Discord Akary Bot (https://discord-bot-akary.blackveriun.repl.co) was down: HTTP code: 0 Response time: 0 ms Resolved: Discord Akary Bot is back up in 36f8c16.
gharchive/issue
2022-05-07T20:31:24
2025-04-01T04:33:38.771255
{ "authors": [ "blackveriun" ], "repo": "blackveriun/status", "url": "https://github.com/blackveriun/status/issues/135", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1931807108
🛑 WSMM Bitbucket is down In fe34e52, WSMM Bitbucket (https://scm.wsmm.de) was down: HTTP code: 503 Response time: 447 ms Resolved: WSMM Bitbucket is back up in cdfdf0c after 10 minutes.
gharchive/issue
2023-10-08T13:31:49
2025-04-01T04:33:38.773543
{ "authors": [ "blacky1707" ], "repo": "blacky1707/uptime", "url": "https://github.com/blacky1707/uptime/issues/127", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1334184486
Revert "Restore JSON decode for ws_api (#329)" Related card PR: https://github.com/dermotduffy/frigate-hass-card/pull/796 Thanks @uvjustin !
gharchive/pull-request
2022-08-10T07:14:44
2025-04-01T04:33:38.779543
{ "authors": [ "dermotduffy", "uvjustin" ], "repo": "blakeblackshear/frigate-hass-integration", "url": "https://github.com/blakeblackshear/frigate-hass-integration/pull/330", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
356346797
JS Modules export? Hi @blakeembrey! This module is really popular! It's also nice and small, but I was wondering if you'd be open to providing a JS Modules version in addition to CommonJS? I'd be happy to PR. Let me know if you have a preference for tooling/etc. @developit Happy to accept a PR 👍 Added with https://github.com/blakeembrey/array-flatten/releases/tag/v3.0.0. @blakeembrey This breaks existing usage code for the old CommonJS pattern, just as an FYI. Where var flatten = require('array-flatten'); used to return the function directly, it now returns an object with the function defined as the object's flatten property. Existing calls to the flatten function directly need to be changed to call the method on the returned object instead. Otherwise seeing "Warning: flatten is not a function" error. Not sure if there is a way to accommodate both usage patterns, or if packages using the old version need to either update or manually specify the older version of this package as a dependency, but I think a lot of people will have builds that break due to this. (grunt-assemble is affected) I know, that’s why this was released as V3. If it affects anything, they’re using dependencies incorrectly. I also don’t see this used in grunt assemble either. @blakeembrey Thanks for taking the time to reply to my comment. I tracked the issue to this package from grunt-assemble > resolve-dep > arrayify-compact > array-flatten Admittedly some of the versions of those packages that were in use on my affected project are 4+ years old and it's likely that there are issues with dependency management somewhere in that tree. I was able to fix the issue on my end by adding version 2.1.2 of your module as a dependency in my package.json, even though I'm not using it directly in my build. arrayify-compact lists version 1.0.1 as a dependency but appears to have been pulling in the latest version instead. Didn't mean to imply there was an issue on your end, just wanted to point that out in case anyone else came across the same problem. Thanks again!
gharchive/issue
2018-09-03T03:25:22
2025-04-01T04:33:38.829787
{ "authors": [ "Lohdro", "blakeembrey", "developit" ], "repo": "blakeembrey/array-flatten", "url": "https://github.com/blakeembrey/array-flatten/issues/19", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
243085814
clear ng2-datepicker after form submit I want to clear the date field after form submit. I tried this.expire_date = null; where my model is expire_date. But it doesn't work. How can I do it? Obviously I don't want to manually click the clear button in the ui. This will solve your problem. delete this.expire_date['formatted']; tried delete this.expire_date['formatted']; and it does not work. this works for me! <ng2-datepicker [options]="options" [inputEvents]="validFormdatepickerInputEvents" formControlName="valid_from" [(ngModel)]="package.valid_from" class="text_area"> this.package.valid_from = new DateModel(); if binded to [(ngModel)] you reset it just like any other variable in the form. You can easily clear selected date using ElementRef: <ng-datepicker [(ngModel)]="empty_date" [options]="options1" #date1></ng-datepicker> <button (click)="date1.displayValue=''">Clear Date</button> OR You can clear it from component also: @ViewChild('date1') date1: ElementRef; resetDate() { this.date1['displayValue']=''; }
gharchive/issue
2017-07-14T19:10:41
2025-04-01T04:33:38.862022
{ "authors": [ "ascuet", "jkuri", "kuldeeps1ngh", "nilasissen", "vaibhavpatil111", "zjx" ], "repo": "bleenco/ng2-datepicker", "url": "https://github.com/bleenco/ng2-datepicker/issues/244", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1795512305
Endpoints for LocalStateQuery via REST The LocalStateQuery Ouroboros protocol has many client endpoints. GetCurrentEra GetSystemStart GetChainBlockNo GetChainPoint GetEraHistory GetEpochNo GetCurrentProtocolParams GetGenesisConfig This depends on blinklabs-io/gouroboros#315 for underlying support for the query communication with the Cardano Node. Merging ChainBlockNo, EpochNo, and ChainPoint into a single Tip to more closely match other tools and reduce the necessary API calls. The first part of this is in #134 and the second part will be in another pull request. The second part is in #143 with a caveat of a bug in gOuroboros around the Genesis Config.
gharchive/issue
2023-07-09T18:30:20
2025-04-01T04:33:38.877181
{ "authors": [ "musik-c", "wolf31o2" ], "repo": "blinklabs-io/cardano-node-api", "url": "https://github.com/blinklabs-io/cardano-node-api/issues/25", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
365159246
Don't prompt for React Router if exists. When using blix add the user is prompted with do you want to add React Router even if they already have react-router in their project. Recommended solution - Do a quick check for react-router in the user's package.json prior to running the script. I'll also go after the empty App folder bug
gharchive/issue
2018-09-29T19:13:02
2025-04-01T04:33:38.888397
{ "authors": [ "DBULL7", "alocke12992" ], "repo": "blixjs/blix", "url": "https://github.com/blixjs/blix/issues/338", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2628396201
Fix validator functions https://sepolia.etherscan.io/address/0xaa186e7c5483ed18cEd7C23eA3c60b3BC929F4d6#code No sepolia! Straight base
gharchive/pull-request
2024-11-01T06:03:29
2025-04-01T04:33:38.891411
{ "authors": [ "Mitchell139", "bitcoinbrisbane" ], "repo": "block52/poker-vm", "url": "https://github.com/block52/poker-vm/pull/20", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
371246139
Hide legacy proposals #1986 Closes #23 .I added new method hideProposals() for checking and hiding legacy proposals, also I added checkbox and hideLegacyProposals state @sschiessl-bcp Please review it , when you have time. Please reference the corresponding issue in the PR https://github.com/bitshares/bitshares-ui/issues/1986 I think the recognition is off. Please adjust: [ ] Proposals with almost same name and same payout account. [ ] if both are not active, hide older one [ ] if one is active, hide other one. Example: 1.14.74 and 1.14.73. [ ] Add a manually defined list of ids that are to be hidden [ ] If the start date of the proposal is more than one month in the past, and less than 10 million votes if one is active, hide other one. Example: 1.14.74 and 1.14.73. I am not sure that understood correctly. Should we display some active proposals when hide proposals is enabled? if one is active, hide other one. Example: 1.14.74 and 1.14.73. Consider the header of the proposal tab Listed under "Proposed": Inactive Listed under "Active": Active I adjusted filter according to your`s comments. Please review it when you have time.
gharchive/pull-request
2018-10-17T20:07:01
2025-04-01T04:33:38.896724
{ "authors": [ "VerevkinAlexander", "sschiessl-bcp" ], "repo": "blockchainprojects/bitshares-ui", "url": "https://github.com/blockchainprojects/bitshares-ui/pull/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
200177893
Fix output for insufficient funds to actually output the needed amount The output prints the current balance instead of the fee Looks good. Thanks for catching this!
gharchive/pull-request
2017-01-11T18:59:05
2025-04-01T04:33:38.940459
{ "authors": [ "jcnelson", "vbrandl" ], "repo": "blockstack/blockstack-cli", "url": "https://github.com/blockstack/blockstack-cli/pull/186", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
776892210
Broken on recent versions of discourse This plugin is broken in recent versions (true as of 2.5.6) of discourse. The Stacks Foundation is looking for help with updating the plugin. (https://github.com/stacksgov/Stacks-Grants/issues/29) Related issues/info: https://github.com/blockstack/blockstack.org/issues/392 https://github.com/blockstack/blockstack.org/issues/1159 https://github.com/blockstack/omniauth-blockstack/blob/master/lib/omniauth/strategies/blockstack.rb#L39-L42 I spent a couple hours taking a look at this today. @wileyj writes: Pretty sure the issue is here: https://github.com/blockstack/omniauth-blockstack/blob/master/lib/omniauth/strategies/blockstack.rb#L39-L42 One issue is here. What's supposed to happen is when /auth/blockstack?manifest=true is fetched by the authenticator, the manifest json is returned. It appears discourse changed behavior such that it instead returns an HTML page with a button the user has to click on before the JSON is returned by the omniauth strategy. I imagine this has something to do with the discourse folks trying to address cross-site scripting, but it breaks our authentication flow. One short-term way to work around this to put the discourse installation behind a proxy and intercept requests to /auth/blockstack?manifest=true and return the manifest json. A longer term solution would be really digging into the internals of discourse to figure a better way to do this - there might not be away to do it in a supported fashion. A second issue was reported by @tvr132b here: https://github.com/blockstack/blockstack.org/issues/1159#issue-687503148 This issue is with the content security policy added by discourse. You need to disable to included content security policy on discourse or add the unsafe-inline keyword (unclear how to do this in on discourse in a supported way). hmm, the proxy idea might be the path of least resistance here. I think we can give it a shot, and see if it works hmm, the proxy idea might be the path of least resistance here. I think we can give it a shot, and see if it works Cool! Let me know how it works! Cool! Let me know how it works!
gharchive/issue
2020-12-31T09:32:20
2025-04-01T04:33:38.947835
{ "authors": [ "larrysalibra", "wileyj" ], "repo": "blockstack/discourse-blockstack", "url": "https://github.com/blockstack/discourse-blockstack/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
766373447
How To Generate Tx And Sign offline? Hello, how can I generate a transaction body and sign it offline? I want to complete a simple function of transferring STX. The content in the document is generated on the client side. I want to try the offline method. I see a similar method called makeUnsignedSTXTokenTransfer, but it is not very suitable . What I want is to first generate an unsigned transaction body, sign it by the sender, then serialize the signed transaction body, and broadcast the serialized result. Is there such a way? I'm not sure, looking forward to your reply, thanks @Wangmmx Please use the forum.stacks.org or discord for these open questions and close this issue. If the documentation at https://docs.blockstack.org/understand-stacks/transactions#serialization is not sufficient you can open an issue on github.com/blockstacks/docs You don't understand what I mean. I need to confirm to the developer instead of the forum. There are many types of transaction bodies in the document. I want to know if there is a way to assemble transaction bodies and sign them offline. It is a simple one-signature account transfer: https://docs.blockstack.org/understand-stacks/transactions#stacks-token-transfer, here is to put the senderKey in the transaction body to build, is there any first to build How about the unsigned transaction body, senderKey to sign separately? @Wangmmx makeUnsignedSTXTokenTransfer does exactly what you want, creating a transaction without the private key. What makes it not very suitable? @Wangmmx makeUnsignedSTXTokenTransfer does exactly what you want, creating a transaction without the private key. What makes it not very suitable? Because this method also uses generator to construct the transaction body, it returns a promise. I want to use a more normal method instead of promise. I have basically solved this problem and constructed a StackTransaction class to complete it. By the way, if you use the way you said, how to put the signature into the transaction body? In other words, what is the signature comes from? thank you for your reply Can you explain why it's not possible to use a promise for your use case? If the nonce and fee are supplied as inputs to the transaction builder, it does not need to be async. Once you have an unsigned transaction, you can sign it using the TransactionSigner class. It's my own problem. I try to avoid using asynchronous methods in my projects. If I quote directly, I need await or handle promise. I have successfully constructed the transaction body and signed it, thank you for your reply, I responded too late. I would like to ask how to broadcast transactions? I want to use the rpc interface /v2/transactions to broadcast. The rawTx parameter is the serialized transaction body, which looks like 8080000000..., but it keeps reporting errors: {Error: "Failed to deserialize posted transaction: Failed to parse transaction authorization : unrecognized auth flags 101"}, I used the api interface broadcastRawTransaction to be successful, may I ask why? Your request body is probably not formatted correctly. Take a look at broadcastTransaction in @stacks/transactions Okay, I made it, thank you very much! I forgot to add the url to the request, thanks again! The successful broadcast transaction was not found on the browser. I found that the last broadcast was successful 10 days ago. Is the testnet unavailable now? There were some issues with Testnet in the last 24 hours. It should be resolved now.
gharchive/issue
2020-12-14T12:24:11
2025-04-01T04:33:38.956836
{ "authors": [ "Wangmmx", "friedger", "yknl" ], "repo": "blockstack/stacks.js", "url": "https://github.com/blockstack/stacks.js/issues/884", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
708897829
Provide option for developers to require browser extension for app authentication During a conversation with @radicleart, he expressed interest in requiring that all users of his apps authenticate with the Blockstack browser extension instead of the hosted authenticator. This would ensure that those users get habituated during both onboarding and sign in to using an installation-based authenticator instead of a dynamically loaded, centrally hosted one, resolving concerns that the latter's use could lead to greater phishing or hacking exposure as discussed on the forum. Providing an option for developers to require the browser extension for authentication in their apps would equip them to make their own choice in regards to the relevant tradeoffs. I'm closing this out since our designs for Connect 2.0 will make the extension option primary for all new users.
gharchive/issue
2020-09-25T12:29:53
2025-04-01T04:33:38.959348
{ "authors": [ "markmhx" ], "repo": "blockstack/ux", "url": "https://github.com/blockstack/ux/issues/607", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2595428797
🛑 Maxthon API is down In 88b8215, Maxthon API (https://api.maxthon.com) was down: HTTP code: 500 Response time: 218 ms Resolved: Maxthon API is back up in df2965b after 13 minutes.
gharchive/issue
2024-10-17T18:19:49
2025-04-01T04:33:38.961762
{ "authors": [ "bloodchen" ], "repo": "bloodchen/upptime", "url": "https://github.com/bloodchen/upptime/issues/1614", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1010053225
Add atomWithDangerouslyAllowMutability signatures Context Signatures in Recoil.resi interface file were missing for Recoil.atomWithDangerouslyAllowMutability Related issues: #30 #24 Related PRs: #29 What does this PR do? It adds proper signatures to the interface file (Recoil.res) Nice catch! Thanks for your PR Released in v2.1.1
gharchive/pull-request
2021-09-28T18:04:21
2025-04-01T04:33:38.963705
{ "authors": [ "bloodyowl", "searleser97" ], "repo": "bloodyowl/rescript-recoil", "url": "https://github.com/bloodyowl/rescript-recoil/pull/31", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
386269075
Drop diamond Remove disk monitoring in graphite since the disk metrics are supplied by diamond. Diamond not only provides disk metrics, but also [CPU|https://github.com/bloomberg/chef-bach/pull/1317/files#diff-9abfd340ffd8489bd5f812ee7e0b10feL186], memory, and network. Some of the alarms we currently have in zabbix is based on memory. I would suggest disabling disk metrics collector rather than drop diamond as a whole. dropped the memory queries in zabbix too. We don't have zabbix configured to alert on CPU or network so it's safe to drop as well.
gharchive/pull-request
2018-11-30T16:55:06
2025-04-01T04:33:38.974345
{ "authors": [ "aespinosa", "leochen4891" ], "repo": "bloomberg/chef-bach", "url": "https://github.com/bloomberg/chef-bach/pull/1317", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2601360392
When i try to join a game it says roblox failed to launch Acknowledgement of preliminary instructions [X] I have read the preliminary instructions, and I am certain that my problem has not already been addressed. [X] I have thoroughly looked through the available Wiki articles and could not find a solution to my problem. [X] I am using the latest version of Bloxstrap. [X] I did not answer truthfully to all the above checkboxes. Bloxstrap Version v2.8.0 What problem did you encounter? When i try to join a game it says roblox failed to launch and takes me to the help website and i dont know why Bloxstrap Log N/A Runs fine, but is encounted as bug. re: i have also been encountering this, yet roblox launches just fine dunno why it's happening I have the same issue, anyone have a solution for this? That's a bug which is gonna be fixed in the new bloxstrap update Addressed in Wiki
gharchive/issue
2024-10-21T05:20:06
2025-04-01T04:33:38.979898
{ "authors": [ "FallenMxru", "Spoonus", "Vad1mCSO2", "lukaku112", "pizzaboxer", "unfunny1" ], "repo": "bloxstraplabs/bloxstrap", "url": "https://github.com/bloxstraplabs/bloxstrap/issues/3381", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2668146372
[BUG] the bloxstrap is running but it isn't displaying Acknowledgement of preliminary instructions [X] I have read the preliminary instructions, and I am certain that my problem has not already been addressed. [X] I have thoroughly looked through the available Wiki articles and could not find a solution to my problem. [X] I am using the latest version of Bloxstrap. [ ] I did not answer truthfully to all the above checkboxes. Bloxstrap Version v2.8.1 What problem did you encounter? I open Bloxstrap, and it's displayed on Task Manager but it doesn't display on the screen Bloxstrap Log N/A Client related issues are no longer accepted.
gharchive/issue
2024-11-18T11:02:16
2025-04-01T04:33:38.982530
{ "authors": [ "TranBao0707", "bluepilledgreat" ], "repo": "bloxstraplabs/bloxstrap", "url": "https://github.com/bloxstraplabs/bloxstrap/issues/3780", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1037622954
Constant latency spikes Hello, I am seeing a pattern with my AWS speed tests showing jitter/latency spikes up to roughly 120ms. The jitter amount is from about 60ms, since that is my average. I have a new internet installation and was using AWS Speed test to gauge the reliability. I was wondering if this was normal. Screenshot attached. Same here Currently AWS latency is measured by performing a http get on a storage blob file in target region. The spikes might be something specific to AWS storage access as same issue does not occur on same code deployed for Azure latency test does. I will dig why this could happen, I am also thinking to change average to pencential so the result would make more sense. Same here: @blrchen I am also thinking to change average to pencential so the result would make more sense. I like this idea. You asked for help. I would like to assist, but not sure how I can assist you.
gharchive/issue
2021-10-27T16:28:24
2025-04-01T04:33:38.985898
{ "authors": [ "Yuvix25", "axljones", "blrchen", "cdp401", "pwprince" ], "repo": "blrchen/aws-speed-test", "url": "https://github.com/blrchen/aws-speed-test/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1174336021
Support for image file formats WebP and AVIF See the following closed issue: https://github.com/bludit/bludit/issues/1168 Supported in Bludit v3.14.1.
gharchive/issue
2022-03-19T19:17:13
2025-04-01T04:33:38.989736
{ "authors": [ "clickwork-git", "dignajar" ], "repo": "bludit/bludit", "url": "https://github.com/bludit/bludit/issues/1414", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
163836761
Test on OS X Update documentation accordingly. Work and instructions here: https://github.com/blue-jay/blueprint/issues/36
gharchive/issue
2016-07-05T12:15:54
2025-04-01T04:33:38.990778
{ "authors": [ "josephspurrier" ], "repo": "blue-jay/jay", "url": "https://github.com/blue-jay/jay/issues/18", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
180180587
Console Warnings in Bluekit Example When I launch the Bluekit Example (and also with my app) I get the following console warnings. Should I be concern? warning.js:36 Warning: Unknown prop 'styles' on <div> tag. Remove this prop from the element. For details, see https://fb.me/react-unknown-prop in div (created by PropsTable) in div (created by PropsTable) in div (created by PropsTable) in PropsTable (created by PropsSidebar) in div (created by PropsSidebar) in div (created by PropsSidebar) in PropsSidebar (created by Page) in div (created by Page) in div (created by StyleRoot) in StyleRoot (created by Page) in Page (created by Content) in div (created by Content) in Content (created by Page) in div (created by Page) in div (created by StyleRoot) in StyleRoot (created by Page) in Page (created by StateProvider) in StateProvider (created by App) in App printWarning @ warning.js:36 warning.js:36 Warning: Failed prop type: Invalid prop 'name' of type 'string' supplied to 'SourceCode', expected 'object'. in SourceCode (created by Preview) in Preview (created by Page) in div (created by Page) in div (created by StyleRoot) in StyleRoot (created by Page) in Page (created by Content) in div (created by Content) in Content (created by Page) in div (created by Page) in div (created by StyleRoot) in StyleRoot (created by Page) in Page (created by StateProvider) in StateProvider (created by App) in App printWarning @ warning.js:36 @andrerecinto These warnings are directly in BlueKit or BlueKit example components. So you shouldn't be concerned. We hope to get rid of them next week in a new release. Thanks for the feedback! Ok good to know. Thank you.
gharchive/issue
2016-09-29T22:36:34
2025-04-01T04:33:38.992916
{ "authors": [ "andrerecinto", "imtoo" ], "repo": "blueberryapps/react-bluekit", "url": "https://github.com/blueberryapps/react-bluekit/issues/69", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
138345769
Stop using bluebox gem mirror for now It's under a lot of pressure and needs some time off to sort through its life. :+1:
gharchive/pull-request
2016-03-04T01:15:28
2025-04-01T04:33:39.006667
{ "authors": [ "j2sol", "pwplusnick" ], "repo": "blueboxgroup/ursula", "url": "https://github.com/blueboxgroup/ursula/pull/1652", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2044152428
Work only first 11 seconds device-2023-12-15-205512.webm Fix PR: https://github.com/blueboy-tm/flutter_v2ray/pull/31
gharchive/issue
2023-12-15T18:01:40
2025-04-01T04:33:39.008238
{ "authors": [ "zikwall" ], "repo": "blueboy-tm/flutter_v2ray", "url": "https://github.com/blueboy-tm/flutter_v2ray/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
499563030
IndexError: List Index out of range Traceback (most recent call last): File "/usr/local/lib64/python3.6/site-packages/tornado/web.py", line 1699, in _execute result = await result File "/usr/local/lib/python3.6/site-packages/samlauthenticator/samlauthenticator.py", line 713, in get login_handler_self) File "/usr/local/lib/python3.6/site-packages/samlauthenticator/samlauthenticator.py", line 634, in _get_redirect_from_metadata_and_redirect handler_self.redirect(redirect_link_getter(saml_metadata_etree)[0], permanent=False) IndexError: list index out of range I am using the simple configuration where: c.JupyterHub.authenticator_class = 'samlauthenticator.SAMLAuthenticator' Where the SAML IdP's metadata is stored. c.SAMLAuthenticator.metadata_filepath = '/PATH/TO/MYFILE.xml' Hi @jameholme! This looks like an issue with how the authenticator interfaces with your Metadata xml. Would it be possible to post the metadata xml file? If not, could you check what value your metadata xml has at the following XPath? //md:SingleSignOnService[@Binding='urn:oasis:names:tc:SAML:2.0:bindings:HTTP-Redirect']/@Location How I'm walking through this issue: It looks like the error is being thrown here, and I'm not entirely clear what the XPath is since it's dynamically defined above that line. So reading up the code (since there isn't much down), we find that we are being called from here. It looks like that line sets the element to md:SingleSignOnService, and then we can more-or-less reconstruct the XPath from there. If you want to double-check my work, I would encourage you to turn on debug logging for the JupyterHub server - I'm not sure how to do this, but it seems like it should be straightforward? <md:SingleSignOnService Binding="urn:oasis:names:tc:SAML:2.0:bindings:HTTP-POST" Location="URL REMOVED FOR REASONS"/> Thanks for the help, this is my first rodeo @jameholme if this is your first SAML Rodeo, welcome! There's some really cool stuff about the tech - for example, the metadata xml is usually publicly available (here's a link to one of the metadata xmls from the services that I usually use, it should download a file). Just like it's ok to spread public keys around to whoever needs them, it's fine to spread the xml metadata around unless you have a company policy against it. Looking at the XML snippet you posted, it looks like we're using different SAML bindings. It looks like you're using an HTTP-POST binding, and I'm looking for an HTTP-Redirect binding. The authenticator might work if you alter your metadata xml to say HTTP-Redirect where it currently says HTTP-POST, but this is a fairly off-the-rails configuration that we're setting up. I think the best course of action would be to leave this issue open so that we know that we need to add HTTP-POST binding in the future, and let me know if you aren't able to get going with the advice in this post. Sound good? I have changed it to redirect, but it did not change anything and I still received those same errors. Below is the XML with the stuff I was told to remove before I could send it to you (again thanks for your help!): <?xml version="1.0"?> -<md:EntityDescriptor xmlns:md="urn:oasis:names:tc:SAML:2.0:metadata" entityID="COMPANYURL.COM" cacheDuration="DURATION" ID="ID"> -<ds:Signature xmlns:ds="http://www.w3.org/2000/09/xmldsig#"> -<ds:SignedInfo> <ds:CanonicalizationMethod Algorithm="http://www.w3.org/2001/10/xml-exc-c14n#"/> <ds:SignatureMethod Algorithm="http://www.w3.org/2001/04/xmldsig-more#rsa-sha256"/> -<ds:Reference URI="URI"> -<ds:Transforms> <ds:Transform Algorithm="http://www.w3.org/2000/09/xmldsig#enveloped-signature"/> <ds:Transform Algorithm="http://www.w3.org/2001/10/xml-exc-c14n#"/> </ds:Transforms> <ds:DigestMethod Algorithm="http://www.w3.org/2001/04/xmlenc#sha256"/> <ds:DigestValue> VALUE </ds:DigestValue> </ds:Reference> </ds:SignedInfo> <ds:SignatureValue> CERT </ds:SignatureValue> -<ds:KeyInfo> -<ds:X509Data> <ds:X509Certificate> CERT </ds:X509Certificate> </ds:X509Data> </ds:KeyInfo> </ds:Signature> -<md:IDPSSODescriptor WantAuthnRequestsSigned="true" protocolSupportEnumeration="urn:oasis:names:tc:SAML:2.0:protocol"> -<md:KeyDescriptor use="signing"> -<ds:KeyInfo xmlns:ds="http://www.w3.org/2000/09/xmldsig#"> -<ds:X509Data> <ds:X509Certificate>CERT</ds:X509Certificate> </ds:X509Data> </ds:KeyInfo> </md:KeyDescriptor> <md:NameIDFormat>urn:oasis:names:tc:SAML:1.1:nameid-format:unspecified</md:NameIDFormat> <md:SingleSignOnService Location="COMPANYURL.COM" Binding="urn:oasis:names:tc:SAML:2.0:bindings:HTTP-REDIRECT"/> <saml:Attribute xmlns:saml="urn:oasis:names:tc:SAML:2.0:assertion" NameFormat="urn:oasis:names:tc:SAML:2.0:attrname-format:basic" Name="employeeID"/> <saml:Attribute xmlns:saml="urn:oasis:names:tc:SAML:2.0:assertion" NameFormat="urn:oasis:names:tc:SAML:2.0:attrname-format:basic" Name="memberOf"/> <saml:Attribute xmlns:saml="urn:oasis:names:tc:SAML:2.0:assertion" NameFormat="urn:oasis:names:tc:SAML:2.0:attrname-format:basic" Name="mail"/> <saml:Attribute xmlns:saml="urn:oasis:names:tc:SAML:2.0:assertion" NameFormat="urn:oasis:names:tc:SAML:2.0:attrname-format:basic" Name="title"/> <saml:Attribute xmlns:saml="urn:oasis:names:tc:SAML:2.0:assertion" NameFormat="urn:oasis:names:tc:SAML:2.0:attrname-format:basic" Name="displayName"/> </md:IDPSSODescriptor> <md:ContactPerson contactType="administrative"/> </md:EntityDescriptor> Ok, two things - I'm not 100% sure how Python's XPath works, but it may be that the HTTP-Redirect value has to be case-matched exactly. Check this pdf around line 1026 to see the exact values for this field. Bad news - currently we don't support signing the SAML Request. We can put that into the backlog, but I'm not sure when I'm going to get to working on it. This is a great opportunity for a pull request :)
gharchive/issue
2019-09-27T17:31:26
2025-04-01T04:33:39.017108
{ "authors": [ "distortedsignal", "jameholme" ], "repo": "bluedatainc/jupyterhub-samlauthenticator", "url": "https://github.com/bluedatainc/jupyterhub-samlauthenticator/issues/38", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1748606
Error Insert file name to database Error Insert file name to database // this is the database configuration you need to setup before being able to use a database $db_host = "localhost"; // usually localhost $db_user = "user_name"; // the user you created when you create a database or a table, i don't remember $db_pwd = "password"; // that user's password $db_database= "database_name"; // the name of the database you created // this starts the database update script $mysqli = new mysqli($db_host, $db_user, $db_pwd, $db_database); // this is a mysqli prepared statement which does the update to the database with the filename of the uploaded image // add this at the end of the handle_file_upload method, before the line: } else { $file->error = $error; if($stmt = $mysqli->prepare("INSERT INTO uploaded_images (id, uploaded_image) VALUES (NULL, ?) ")) { $stmt->bind_param('s', $file_name); $stmt->execute(); $stmt->close(); }else{ // this removes the image, in case there's an error with the updating of the database, // so you don't stay with uploaded images that are not in the database $file_path = $this->options['upload_dir'].$file_name; $success = is_file($file_path) && $file_name[0] !== '.' && unlink($file_path); if ($success) { foreach($this->options['image_versions'] as $version => $options) { $file = $options['upload_dir'].$file_name; if (is_file($file)) { unlink($file); } } } } $mysqli->close(); Error Message: 0003.JPG 34.93 KB Error: SyntaxError: JSON.parse Delete How to solve ? Best regards I tried your code and it gave me: SyntaxError: JSON.parse: unexpected character Have you succeeded in getting the file name into the DB?
gharchive/issue
2011-09-27T03:20:51
2025-04-01T04:33:39.023777
{ "authors": [ "cadnet", "cnodom" ], "repo": "blueimp/jQuery-File-Upload", "url": "https://github.com/blueimp/jQuery-File-Upload/issues/650", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
190507191
Set font size for buttons to a reasonable size This only applies to users with javascript disabled as the styling is otherwise overwritten by bootstrap. 200px size was introduced in 76ca40b044b7660f17a77ff82f7fcf14b51878a8 which mentions #2635 but the issue appears to be no longer available so I don't know what the point of the incredibly large size was. For reference, this is how the demo looks with JS disabled: Thanks @tribut. There's a separate file where I fixed this issue.
gharchive/pull-request
2016-11-19T17:04:36
2025-04-01T04:33:39.025780
{ "authors": [ "blueimp", "tribut" ], "repo": "blueimp/jQuery-File-Upload", "url": "https://github.com/blueimp/jQuery-File-Upload/pull/3464", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1974851431
HLS.js usage: bugfix, stability improvements and resilience improvement I propose the following changes to the HLS.js: BUGFIX: pass "video" element to create() on restart triggering fatal-error STABILITY: await MEDIA_ATTACHED event before performing "loadSource" on HLS STABILITY: await MANIFEST_PARSED event before performing "play" on video element RESILIENCE: on "MEDIA_ERROR" event perform "recoverMediaError" on HLS Code was now adjusted. merged, thanks
gharchive/pull-request
2023-11-02T18:44:16
2025-04-01T04:33:39.047489
{ "authors": [ "aler9", "rse" ], "repo": "bluenviron/mediamtx", "url": "https://github.com/bluenviron/mediamtx/pull/2631", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2652629079
Guideline for use of "tag" versus "label" Steps to Reproduce There are references to both "label" as well as "tag" in the English translation. Are these different concepts? If so, what are their semantics? If not, can they be merged to one word? In Dutch, a "tag" is often translated as "label", as is the English "label". Sample of "Label" Sample of "Tag" Attachments No response What platform(s) does this occur on? iOS, Android, Web (Desktop), Web (Mobile) Device Info No response What version of the app are you using? main Additional Information No response They're distinct concepts. Labels are annotations on users and content. They can be used to hide, warn, and categorize the network. "Labels" are mainly a moderation tool, like "Rude" or "Intolerance", but a few can also be self-applied to warn about adult and graphic media (which then is hidden or behind a click-through by default). There are also positive labels for example by https://bsky.app/profile/yardcrow.com. Hashtags are "tags", but "tags" is a bit more general, since the protocol also allows you to add tags to posts outside of the text and there (formally) without a #. Bluesky doesn't yet have UI for the latter though, and they may well end up with a purely visual "#" since they more or less do behave just like hashtags. Thank you! Labels are for labeling, tags are hastags and for content (text). I will add it to the glossary.
gharchive/issue
2024-11-12T16:02:59
2025-04-01T04:33:39.058827
{ "authors": [ "Tamschi", "monty241" ], "repo": "bluesky-social/social-app", "url": "https://github.com/bluesky-social/social-app/issues/6262", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1926004184
flaky tests Sometimes CI tests fail. Flaky tests are bad and we should aim to not have any. Here is an example of such a test failing on one instance, but passing when rerun: https://github.com/bluesky/ophyd-async/actions/runs/10741400139/job/29791703029 FAILED tests/epics/adsimdetector/test_sim.py::test_detector_writes_to_file - bluesky.utils.FailedStatus: <AsyncStatus, task: <coroutine object StandardDetector.trigger at 0x7fecaa70fe60>, errored: TimeoutError()> I haven't seen flaky tests for several months so going to close this. If they come back, I've wondered about detecting them with infrequent use of pytest-repeat Flaky tests have re-appeared, I'm going to re-open this ticket. pytest plugin for generating CTRF report for the above https://pypi.org/project/pytest-json-ctrf/ I think all runs, why not? Identifying flaky tests vs. tests that were temporarily broken by unfinished PRs? @callumforrester I've enabled on this draft PR (scroll down to summary section): it hasn't marked any tests as flaky yet, and doesn't provide any easier method of seeing what's flaky/broken. @coretl for comment too https://github.com/bluesky/ophyd-async/actions/runs/10809455998?pr=567 Tango tests are apparently flaky because of Tango test infrastructure and there is nothing we can do to fix that. I suggest we report test failures on this ticket and make PRs that @pytest.mark.flaky(reruns=2) individual flaky tests until the CI is reliable again. The first PR that does this needs to add pytest-rerunfailures to the dev dependencies to get the marker.
gharchive/issue
2023-10-04T11:29:19
2025-04-01T04:33:39.063311
{ "authors": [ "DiamondJoseph", "callumforrester", "coretl", "rosesyrett" ], "repo": "bluesky/ophyd-async", "url": "https://github.com/bluesky/ophyd-async/issues/23", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
623633980
set serde derive to use serde untagged Think the way serde is derived is a non-intuitive way. As I think most people do not want the left and right tags for the values. Suggest adding untagged to the serde derive. However this would be a breaking change. I do not belive many people are using the serde fetaure as its now implemented. This will allow us to do something like: type ComplicatedType = HashMap<String, Either<Vec<String>, String>> Some json that would be parsed now would be { "some_key": "hello", "some_other_random_key": ["hello","world!"] } Either is just one two-case enum, and if you have a particular use case, you should of course make an enum that works for you. It is not obvious if tagged or untagged should be the default for a general two-case enum, but I would lean towards tagged. Think Either with serde is usefull when working with rapid prototyping. You get so much for free, like Display... That beeing sed, I agree that Either with conflicting types makes it hard, thow I do not see people using either with simular types for (de)serializing. FWIW, this was also discussed in #22. I opened this issue as I thought tagged was a strange default. Do also think that the serde-untagged feature is looking like a good solution. Ok, it will certainly not be a feature for the reasons clarified in #22. Closing as I think this is done the right way, reading further throught it once more.
gharchive/issue
2020-05-23T10:56:11
2025-04-01T04:33:39.068809
{ "authors": [ "bluss", "cuviper", "meltinglava" ], "repo": "bluss/either", "url": "https://github.com/bluss/either/issues/46", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
312657587
Use mongo state to update timestamp The timestamp method should show when underlying database has changed. Currently mongo driver only watches changes that it makes, but if two arachne servers point at the same mongo db, they won't recognize when the db has changed. Need to use something like https://docs.mongodb.com/manual/changeStreams/ to update timestamp This same problem also exists for elasticsearch Will https://stackoverflow.com/questions/27255596/elasticsearch-index-last-update-time work for elasticsearch ?
gharchive/issue
2018-04-09T19:32:06
2025-04-01T04:33:39.078486
{ "authors": [ "adamstruck", "kellrott" ], "repo": "bmeg/arachne", "url": "https://github.com/bmeg/arachne/issues/88", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }