id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
45719067 | Adds rudimentary support for Arch Linux
I am working in a VirtualBox, so I haven't executed the test suite against Arch Linux. Possibly some combination of attribute options will require further adjustment to work on this platform.
After making these changes everything I did test (by hand) is working quite happily.
Hello!
There was a pretty heavy refactor in 0.40.0.... I suspect chef-client should detect systemd and try to use it on Arch.
I'm going to close this issue for now, since this isn't going to merge cleanly.... if the new provider detection stuff in 12.4 doesn't work, you may need to add a line to the provider in libraries and send a new PR.
Thanks!
-s
| gharchive/pull-request | 2014-10-14T07:13:44 | 2025-04-01T04:33:37.494459 | {
"authors": [
"logankoester",
"someara"
],
"repo": "bflad/chef-docker",
"url": "https://github.com/bflad/chef-docker/pull/213",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
3221013 | ImportError: cannot import name HTTPSConnection
Hello,
After installing this plugin from Package Control on a Ubuntu 11.10 system, I get this in the console:
Reloading plugin /home/nchambrier/.config/sublime-text-2/Packages/sublime-github/gist.py
Traceback (most recent call last):
File "./sublime_plugin.py", line 62, in reload_plugin
File "./gist.py", line 10, in <module>
import requests
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/__init__.py", line 26, in <module>
from . import utils
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/utils.py", line 19, in <module>
from .compat import parse_http_list as _parse_list_header
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/compat.py", line 86, in <module>
from .packages.oreos.monkeys import SimpleCookie
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/__init__.py", line 3, in <module>
from . import urllib3
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/__init__.py", line 16, in <module>
from .connectionpool import (
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/connectionpool.py", line 22, in <module>
from httplib import HTTPConnection, HTTPSConnection, HTTPException
ImportError: cannot import name HTTPSConnection
The thing I don't understand is that I have only one version of Python installed, and when I run python REPL I can see the import working:
$ python
Python 2.7.2+ (default, Oct 4 2011, 20:03:08)
[GCC 4.6.1] on linux2
Type "help", "copyright", "credits" or "license" for more information.
>>> from httplib import HTTPConnection, HTTPSConnection, HTTPException
>>> HTTPSConnection
<class httplib.HTTPSConnection at 0x8cdee9c>
Thanks for the report. I think the problem is the version of python that Sublime Text uses doesn't have SSL support. I worked around this before by using curl; guess I'll have to do that again.
Yep I checked out closed issues after, but it's highly probable that #2 is
related to this issue ;)
On 14 February 2012 17:01, Brad Greenlee <
reply@reply.github.com
wrote:
Thanks for the report. I think the problem is the version of python that
Sublime Text uses doesn't have SSL support. I worked around this before by
using curl; guess I'll have to do that again.
Reply to this email directly or view it on GitHub:
https://github.com/bgreenlee/sublime-github/issues/6#issuecomment-3963694
--
Nicolas Chambrier, aka naholyr
Blog : http://naholyr.fr
Formateur Clever Institut :
http://clever-institut.com/formateur/nicolas-chambrier
Ok, I've pushed a fix--re-adding the curl fallback (so you need to have curl installed). Upgrade to version 2.0.2 and let me know if it works for you. I've only tested it on OS X; if you have issues, I'll set up a Ubuntu VM to test with.
Upgraded to 2.0.3 and still the same :(
Here are the startup errors:
Reloading plugin /home/nchambrier/.config/sublime-text-2/Packages/sublime-github/gist.py
Traceback (most recent call last):
File "./sublime_plugin.py", line 62, in reload_plugin
File "./gist.py", line 7, in <module>
import sublime_requests as requests
File "./sublime_requests.py", line 5, in <module>
import requests
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/__init__.py", line 26, in <module>
from . import utils
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/utils.py", line 19, in <module>
from .compat import parse_http_list as _parse_list_header
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/compat.py", line 86, in <module>
from .packages.oreos.monkeys import SimpleCookie
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/__init__.py", line 3, in <module>
from . import urllib3
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/__init__.py", line 16, in <module>
from .connectionpool import (
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/connectionpool.py", line 25, in <module>
from httplib import HTTPConnection, HTTPSConnection, HTTPException
ImportError: cannot import name HTTPSConnection
Reloading plugin /home/nchambrier/.config/sublime-text-2/Packages/sublime-github/sublime_requests.py
Traceback (most recent call last):
File "./sublime_plugin.py", line 62, in reload_plugin
File "./sublime_requests.py", line 5, in <module>
import requests
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/__init__.py", line 26, in <module>
from . import utils
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/utils.py", line 19, in <module>
from .compat import parse_http_list as _parse_list_header
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/compat.py", line 86, in <module>
from .packages.oreos.monkeys import SimpleCookie
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/__init__.py", line 3, in <module>
from . import urllib3
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/__init__.py", line 16, in <module>
from .connectionpool import (
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/connectionpool.py", line 25, in <module>
from httplib import HTTPConnection, HTTPSConnection, HTTPException
ImportError: cannot import name HTTPSConnection
I'll check out if I can fix it myself and git you a patch, so you don't have to bother with a VM.
I think it will still be hard after that. I naively removed "HTTPSConnection, " from "urllib3.connectionpool", and I then get other errors:
Reloading plugin /home/nchambrier/.config/sublime-text-2/Packages/sublime-github/gist.py
Traceback (most recent call last):
File "./sublime_plugin.py", line 62, in reload_plugin
File "./gist.py", line 7, in <module>
import sublime_requests as requests
File "./sublime_requests.py", line 5, in <module>
import requests
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/__init__.py", line 26, in <module>
from . import utils
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/utils.py", line 19, in <module>
from .compat import parse_http_list as _parse_list_header
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/compat.py", line 86, in <module>
from .packages.oreos.monkeys import SimpleCookie
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/__init__.py", line 3, in <module>
from . import urllib3
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/__init__.py", line 16, in <module>
from .connectionpool import (
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/connectionpool.py", line 42, in <module>
from .request import RequestMethods
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/request.py", line 12, in <module>
from .filepost import encode_multipart_formdata
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/filepost.py", line 15, in <module>
from io import BytesIO
File ".\io.py", line 63, in <module>
ImportError: No module named _fileio
Reloading plugin /home/nchambrier/.config/sublime-text-2/Packages/sublime-github/sublime_requests.py
Traceback (most recent call last):
File "./sublime_plugin.py", line 62, in reload_plugin
File "./sublime_requests.py", line 5, in <module>
import requests
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/__init__.py", line 26, in <module>
from . import utils
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/utils.py", line 19, in <module>
from .compat import parse_http_list as _parse_list_header
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/compat.py", line 86, in <module>
from .packages.oreos.monkeys import SimpleCookie
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/__init__.py", line 3, in <module>
from . import urllib3
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/__init__.py", line 16, in <module>
from .connectionpool import (
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/connectionpool.py", line 42, in <module>
from .request import RequestMethods
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/request.py", line 12, in <module>
from .filepost import encode_multipart_formdata
File "/home/nchambrier/.config/sublime-text-2/Packages/sublime-github/lib/requests/packages/urllib3/filepost.py", line 15, in <module>
from io import BytesIO
File ".\io.py", line 63, in <module>
ImportError: No module named _fileio
I'm not sure yet it those errors are due to the fact I removed HTTPSConnection from connectionpool, or if it's yet another missing component from my system :-\
You definitely don't want to remove HTTPSConnection entirely--that will break OS X. But it does sound like there might be other things missing. It shouldn't be hard to work around, since Linux doesn't need any of that anyway. I'm downloading Ubuntu 11.10 now, so I'll have a fix soon.
Ok, give it a shot. Remember you need curl installed, so apt-get install curl if you haven't already.
I had a similar problem where the issue was openssl was updated but python was not recompiled, basically whenever you update openssl you have to recompile python.
https://github.com/jarun/googler/issues/312
| gharchive/issue | 2012-02-14T15:43:12 | 2025-04-01T04:33:37.513483 | {
"authors": [
"bgreenlee",
"ishandutta2007",
"naholyr"
],
"repo": "bgreenlee/sublime-github",
"url": "https://github.com/bgreenlee/sublime-github/issues/6",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
241249376 | libcrnn not found
When I run /src/demo.lua with th , it gives me this error
```module 'libcrnn' not found:No LuaRocks module found for libcrnn
no field package.preload['libcrnn']
no file '/home/akash/.luarocks/share/lua/5.1/libcrnn.lua'
no file '/home/akash/.luarocks/share/lua/5.1/libcrnn/init.lua'
no file '/home/akash/torch/install/share/lua/5.1/libcrnn.lua'
no file '/home/akash/torch/install/share/lua/5.1/libcrnn/init.lua'
no file './libcrnn.lua'
no file '/home/akash/torch/install/share/luajit-2.1.0-beta1/libcrnn.lua'
no file '/usr/local/share/lua/5.1/libcrnn.lua'
no file '/usr/local/share/lua/5.1/libcrnn/init.lua'
no file '/home/akash/.luarocks/lib/lua/5.1/libcrnn.so'
no file '/home/akash/torch/install/lib/lua/5.1/libcrnn.so'
no file '/home/akash/torch/install/lib/libcrnn.so'
no file './libcrnn.so'
no file '/usr/local/lib/lua/5.1/libcrnn.so'
no file '/usr/local/lib/lua/5.1/loadall.so'
Its looking for the .so file that is in your src directory after you've successfully built the project.
Same issue. There is .so file even then showing the same error
@JiteshPshah ,hello,I am facing the same problem ,have you run successfully ,if so,can you give me some advice, thanks a lot ^_^
Hi
Did you guys fix this?
| gharchive/issue | 2017-07-07T12:29:04 | 2025-04-01T04:33:37.518580 | {
"authors": [
"DecentMakeover",
"JiteshPshah",
"dvlshah",
"lxx1884896",
"stalagmite7"
],
"repo": "bgshih/crnn",
"url": "https://github.com/bgshih/crnn/issues/69",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2047188007 | 🛑 Hacker News is down
In 7e19267, Hacker News (https://news.ycombinator.com) was down:
HTTP code: 502
Response time: 362 ms
Resolved: Hacker News is back up in cc04d68 after 11 minutes.
| gharchive/issue | 2023-12-18T18:13:00 | 2025-04-01T04:33:37.521235 | {
"authors": [
"bguivarch"
],
"repo": "bguivarch/testuptime2",
"url": "https://github.com/bguivarch/testuptime2/issues/100",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
712948644 | added A-star path finding algorithm in the projects
This is my project where we can visualize the A-star path finding algorithm
Wow, amazing code!
Wow, amazing code!
Thanks :) @bhaveshlohana
| gharchive/pull-request | 2020-10-01T15:25:23 | 2025-04-01T04:33:37.563528 | {
"authors": [
"bhaveshlohana",
"shikharmaxx"
],
"repo": "bhaveshlohana/HacktoberFest2020-Contributions",
"url": "https://github.com/bhaveshlohana/HacktoberFest2020-Contributions/pull/330",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
351469430 | Problem: Gnome-Terminal displays segment color after newline if using two newline - segments
Hi,
please see following screenshot:
As you can see, the color of the status segment is displayed after the newline.
I'm using ohmyzsh and my .zshrc config is as follows:
POWERLEVEL9K_LEFT_PROMPT_ELEMENTS=(context dir status newline vcs newline)
POWERLEVEL9K_RIGHT_PROMPT_ELEMENTS=(time)
POWERLEVEL9K_DISABLE_RPROMPT=true
POWERLEVEL9K_CUSTOM_PROMPTICON="echo \>"
# HIDE current user in prompt
DEFAULT_USER=`whoami`
# load solarized dir colors
eval `dircolors ~/.dircolors/dircolors.ansi-dark`
It doesn't matter what segments I use. If using two newline statements, this error occurs.
Does anybody know a solution to this or is it just not recommended to use two newline statements?
Thanks ;)
I have the exact same problem. I'm on Ubuntu 18.04, using oh-my-zsh. All other configurations seem to work as expected, but this little annoyance keeps showing up.
Hm, interesting. And weird that it's actually showing the segment BG color of the segment on the top line, not of the segment on the second line.
@bytekeks @ChrisSeattle - Are you guys using a release or the master branch, by chance? Can you give the next branch a shot and let us know if the behavior is the same?
@bytekeks - Whoa, really? The latest master fixed it for you?
@ChrisSeattle - Can you test & confirm as well?
@bhilburn indeed. The issue is gone :)
It seems, that using no seperator icons lead to this problem in the old version.
POWERLEVEL9K_LEFT_SEGMENT_SEPARATOR=""
POWERLEVEL9K_RIGHT_SEGMENT_SEPARATOR=""
In the current master, leaving SEPERATOR empty will just be ignored and the default is used, right?
Great! I'm so happy to hear the issue is resolved in the next branch. That means a fix will be included in the next release.
@bytekeks - If you don't define the separator at all, then the default is used. If you define it to be empty, then it will be empty -- which was probably the cause, per your suspicion =)
I'm closing this issue since it's resolved in next. Please re-open if you are still experiencing this!
| gharchive/issue | 2018-08-17T06:13:18 | 2025-04-01T04:33:37.569149 | {
"authors": [
"ChrisSeattle",
"bhilburn",
"bytekeks"
],
"repo": "bhilburn/powerlevel9k",
"url": "https://github.com/bhilburn/powerlevel9k/issues/957",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
128277971 | Reworked the AWS-Beanstalk segment
This PR replaces #162
Perfect! Thank you, @dritter!!
| gharchive/pull-request | 2016-01-23T00:38:04 | 2025-04-01T04:33:37.570298 | {
"authors": [
"bhilburn",
"dritter"
],
"repo": "bhilburn/powerlevel9k",
"url": "https://github.com/bhilburn/powerlevel9k/pull/203",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
156396200 | Refactor location installation path code
Now it is possible to set the installation path by setting POWERLEVEL9K_INSTALLATION_PATH. If set, this path will always be used. If it was not set manually, we try to determine the path automatically.
This fixes #264
@protist could you double-check? Especially with your ZSH 5.0.7. That would be great. Thanks.
Patching works perfectly on my zsh 5.0.7. Thank you for the quick fix.
Lots of changes here, not all of which I can easily test. The code looks good, though! Merging into next for further dev testing.
Thanks so much, @dritter! Looks like it fixed @protist's issue, as well, which is awesome. Nice work.
| gharchive/pull-request | 2016-05-24T00:37:05 | 2025-04-01T04:33:37.572438 | {
"authors": [
"bhilburn",
"dritter",
"protist"
],
"repo": "bhilburn/powerlevel9k",
"url": "https://github.com/bhilburn/powerlevel9k/pull/266",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
238651155 | Fixed prompt hangs on kubectl version command
This resolves issue #553 when prompt hangs on kubectl version
Ran into this issue when using this feature on next based on the great work jtaylor007.jt@gmail.com did.
@derekbassett - Thanks so much for taking the time to fix this issue! I appreciate the thorough explanation both here and in #553.
This is a great PR. Welcome to the club of P9k contributors! =)
| gharchive/pull-request | 2017-06-26T19:53:52 | 2025-04-01T04:33:37.574350 | {
"authors": [
"bhilburn",
"derekbassett"
],
"repo": "bhilburn/powerlevel9k",
"url": "https://github.com/bhilburn/powerlevel9k/pull/554",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1054361959 | dwi description seems wrong: T2w instead of T2*w
here it says that "Diffusion-weighted imaging contrast (specialized T2* weighting)."
but given that DWI is based on a spin echo (vs. gradient echo) EPI, I would call it a "specialized T2 weighting" instead.
Pinging @oesteban, who might be able to weigh in on this.
If it is something that needs to be changed, we can do that in the suffixes schema file:
https://github.com/bids-standard/bids-specification/blob/30c61b7eb473cccd2414d3947d38487995720f67/src/schema/objects/suffixes.yaml#L426-L429
| gharchive/issue | 2021-11-16T02:22:55 | 2025-04-01T04:33:37.621460 | {
"authors": [
"jcohenadad",
"tsalo"
],
"repo": "bids-standard/bids-specification",
"url": "https://github.com/bids-standard/bids-specification/issues/930",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
195777933 | Add OS package install option
Add an option to install Wildfly from an OS package Vs from downloaded tar
Coverage remained the same at 73.83% when pulling 213711b174bbece2b049eb7e3ae5043506ce56f3 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master.
Coverage remained the same at 73.83% when pulling 213711b174bbece2b049eb7e3ae5043506ce56f3 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master.
Hey @zipkid, this is pretty much the code in PR #137
Do you know @jstremick ?
I suppose the same suggestions I gave there applies here:
remove install_method parameter and use package if it's provided and URL otherwise
remove code to set ownership of wildfly directory. I believe it's package responsability.
What do you think?
Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master.
Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master.
Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master.
Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master.
Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master.
Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master.
Coverage remained the same at 73.83% when pulling f78628b1e0a902c451903d5e9784d88d0712f033 on vrtdev:feature/os_package_install into cd998a141c872f94b3a0b2ac66739f712025d0b6 on biemond:master.
@jairojunior any thoughts on this?
The fourth comment is mine. @coveralls is kind of annoying. :smile:
Damn... I missed yours in the forest....
I'll look into your comment :-)
Your comment makes complete sense.
I hope to be able to modify my code according to that soonish...
Awesome. Feel free to disagree though, it's just a suggestion based exclusively on code review. :smile:
Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master.
Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master.
Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master.
Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master.
Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master.
Coverage remained the same at 73.987% when pulling 2f81d319ef066ddd565540112f491d4a7c1f0888 on vrtdev:feature/os_package_install into e609be57a6eea9ef0773e4a71de7c6872bddf10d on biemond:master.
Awesome. Thanks.
| gharchive/pull-request | 2016-12-15T11:09:08 | 2025-04-01T04:33:37.655174 | {
"authors": [
"coveralls",
"jairojunior",
"zipkid"
],
"repo": "biemond/biemond-wildfly",
"url": "https://github.com/biemond/biemond-wildfly/pull/165",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
108217754 | better validation when saving paired fastqs
logError for LENIENT
throw for `STRICT
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins/job/ADAM-prb/926/
Test PASSed.
Other than the small nit, LGTM!
nit should be fixed
Thanks @ryan-williams! I'll leave this open for comments, but if we don't hear from anyone else tomorrow, I'll merge.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins/job/ADAM-prb/931/
Test PASSed.
Thanks @ryan-williams!
| gharchive/pull-request | 2015-09-24T21:28:21 | 2025-04-01T04:33:37.720232 | {
"authors": [
"AmplabJenkins",
"fnothaft",
"ryan-williams"
],
"repo": "bigdatagenomics/adam",
"url": "https://github.com/bigdatagenomics/adam/pull/830",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
96604773 | Rename to bigdatagenomics/formats
Small nit: I feel like bigdatagenomics/bdg-formats is redundant. Would anyone be opposed to renaming to bigdatagenomics/formats? I will leave this open for a week and make the changes if no one is opposed.
We decided against this; it would lead to a JAR with a name that was too generic.
| gharchive/issue | 2015-07-22T16:08:54 | 2025-04-01T04:33:37.721391 | {
"authors": [
"fnothaft"
],
"repo": "bigdatagenomics/bdg-formats",
"url": "https://github.com/bigdatagenomics/bdg-formats/issues/60",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
53577987 | Deal properly with merging and partitioning data.
For example, the 1000 Genomes VCF data is organized into individual files by chromosome. But this obviously causes the files to differ substantially in size. My ideal would be to merge into a single file, or to split the genome into equal sized bins based on locus and sort the data into it.
This is being tackled in https://github.com/bigdatagenomics/eggo/issues/30 - closing.
| gharchive/issue | 2015-01-06T23:34:22 | 2025-04-01T04:33:37.722695 | {
"authors": [
"laserson",
"tomwhite"
],
"repo": "bigdatagenomics/eggo",
"url": "https://github.com/bigdatagenomics/eggo/issues/2",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2026219490 | Updated AbsoluteLink
I was receiving the following error on silverstripe 5.1, php 8.1:
Fatal error: Declaration of Bigfork\SilverStripeOAuth\Client\Control\Controller::AbsoluteLink() must be compatible with SilverStripe\Control\RequestHandler::AbsoluteLink($action = '') in C:\xampp-8.1\htdocs\silverstripe-install\vendor\bigfork\silverstripe-oauth\src\Control\Controller.php on line 79
ERROR [UNKNOWN TYPE, ERRNO 64]: Declaration of Bigfork\SilverStripeOAuth\Client\Control\Controller::AbsoluteLink() must be compatible with SilverStripe\Control\RequestHandler::AbsoluteLink($action = '') IN GET / Line 79 in C:\xampp-8.1\htdocs\silverstripe-install\vendor\bigfork\silverstripe-oauth\src\Control\Controller.php Source ====== 70: $backUrl = Director::absoluteBaseURL(); 71: } 72: 73: return $backUrl; 74: } 75: 76: /** 77: * @return string 78: */ * 79: public function AbsoluteLink() 80: { 81: return static::join_links(Director::absoluteBaseURL(), $this->Link()); 82: } 83: 84: /** 85: * This takes parameters like the provider, scopes and callback url, builds an authentication Trace ===== SilverStripe\Logging\DetailedErrorFormatter->output(64, Declaration of Bigfork\SilverStripeOAuth\Client\Control\Controller::AbsoluteLink() must be compatible with SilverStripe\Control\RequestHandler::AbsoluteLink($action = ''), C:\xampp-8.1\htdocs\silverstripe-install\vendor\bigfork\silverstripe-oauth\src\Control\Controller.php, 79, ) DetailedErrorFormatter.php:55 SilverStripe\Logging\DetailedErrorFormatter->format(Monolog\LogRecord) AbstractProcessingHandler.php:42 Monolog\Handler\AbstractProcessingHandler->handle(Monolog\LogRecord) Logger.php:389 Monolog\Logger->addRecord(Monolog\Level, Fatal Error (E_COMPILE_ERROR): Declaration of Bigfork\SilverStripeOAuth\Client\Control\Controller::AbsoluteLink() must be compatible with SilverStripe\Control\RequestHandler::AbsoluteLink($action = ''), Array) Logger.php:579 Monolog\Logger->log(Monolog\Level, Fatal Error (E_COMPILE_ERROR): Declaration of Bigfork\SilverStripeOAuth\Client\Control\Controller::AbsoluteLink() must be compatible with SilverStripe\Control\RequestHandler::AbsoluteLink($action = ''), Array) ErrorHandler.php:247 Monolog\ErrorHandler->handleFatalError()
I just updated the AbsoluteLink function with $action
Thanks, tagged in 2.2.1
| gharchive/pull-request | 2023-12-05T13:37:47 | 2025-04-01T04:33:37.725098 | {
"authors": [
"aletail",
"kinglozzer"
],
"repo": "bigfork/silverstripe-oauth",
"url": "https://github.com/bigfork/silverstripe-oauth/pull/19",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2069989821 | 别买了,我买了没用的,大家别花冤枉钱,一直是下载中0%
你下载是什么,具体的下载地址提供一下
不用当心没有用,如果是软件问题,肯定能解决,如果有问题及时反馈,100%能免单
有问题,可以支付宝上留言
有问题,请支付宝上留言
好的我去支付宝留言
| gharchive/issue | 2024-01-08T08:52:05 | 2025-04-01T04:33:37.726985 | {
"authors": [
"bigintpro",
"yyxxiiaaoo"
],
"repo": "bigintpro/csdn_downloader",
"url": "https://github.com/bigintpro/csdn_downloader/issues/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
328432446 | fixes for linux
Fixes two problems on linux:
"ambiguous use of FE_UNDERFLOW" fixed with explicit variable type
Lack of support for CFNumberGetType. Reimplemented using casts
I'm okay with both changes regardless of CI - I'll get that fixed shortly. Thank you!
| gharchive/pull-request | 2018-06-01T08:47:25 | 2025-04-01T04:33:37.728289 | {
"authors": [
"mlilback",
"zwaldowski"
],
"repo": "bignerdranch/Freddy",
"url": "https://github.com/bignerdranch/Freddy/pull/279",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
106652271 | Improved expand/collapse performance
When expanding/collapsing a group of list items, notifyDataSetChanged is called multiple times. It would perform better if notifyDataSetChanged was just called once after all of the items were expanded/collapsed.
Resolved in 2.1.0.
| gharchive/issue | 2015-09-15T21:30:15 | 2025-04-01T04:33:37.729352 | {
"authors": [
"dgreenhalgh"
],
"repo": "bignerdranch/expandable-recycler-view",
"url": "https://github.com/bignerdranch/expandable-recycler-view/issues/72",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
87306993 | re work on report
report enhancement currently only covers steploggable. we will need to add code to cover the objectfinder and objectaction
report needs to be considered as well
| gharchive/issue | 2015-06-11T11:14:42 | 2025-04-01T04:33:37.734896 | {
"authors": [
"peidong-hu"
],
"repo": "bigtester/automation-test-engine",
"url": "https://github.com/bigtester/automation-test-engine/issues/164",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1025561587 | Login feature test
Adds Wallaby for Feature/Integration testing
Adds an example test for Login
[x] Works locally
[ ] Works in CI
@drteeth have you tried https://hexdocs.pm/phoenix_live_view/Phoenix.LiveViewTest.html at all?
I think that it can do everything Wallaby can for liveviews, and is faster / doesn't need the chrome dependency.
The only thing that it wouldn't cover is the test you wrote here which is a "deadview" but that's because when I wrote the login page LiveView didn't have hooks yet - it can be rewritten as a liveview.
Feels on trying some LiveView tests to see if you like them?
So browser tests more for like full integration testing, and liveview tests for "browser unit tests?"
| gharchive/pull-request | 2021-10-13T18:22:07 | 2025-04-01T04:33:37.746736 | {
"authors": [
"drteeth",
"mveytsman"
],
"repo": "bikebrigade/dispatch",
"url": "https://github.com/bikebrigade/dispatch/pull/5",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
177145865 | Oracle DB support
Dear Bill,
how are you, i hope you are fine.
i am php developer and not in java development at all so please be patient to my questions.
i did download the oracle driver ojdbc6.jar and put it in all WEB-INF\lib for all webapps and also did configure Oracle driver in the database table qc_data_source_driver with the following class name oracle.jdbc.driver.OracleDriver and in the table qc_data_source_conf i configure the JDBC_URL jdbc:oracle:thin:username/password@database:1521:xe
i tried to change the configuration file applicationContext-dataSource.properties in each webapps like the following:
dataSource.driverClass=oracle.jdbc.driver.OracleDriver
dataSource.jdbcUrl=jdbc:oracle:thin:@//darabase:1521/XE
dataSource.user=username
dataSource.password=password
dataSource.maxPoolSize=48
dataSource.minPoolSize=1
dataSource.initialPoolSize=1
dataSource.maxIdleTime=120
after restart the server nothing appears in the page except the top bar with the application dropdown and only the refresh option is there
so could you please guide me how to make it work migrating to the oracle database?
thanks for your time.
Config qc_data_source_conf and put ojdbc6.jar is only support OLAP for QCHARTS-WEB, if you want all bambooBSC system support Oracle db, you need jump current mysql bbcore data to Oracle, thanks.
I want to know some questions:
Do you want to use the OLAP features like(01 - Basic query, 02 - Analytics)
Or want bambooBSC use MySQL database change to Oracle-DB
please answer to me. 1 or 2.
thanks.
Dear chen,
It is 2 i want to move totally to use oracle db instead of the current MYSQL database.
OK, i will to try how to jump Oracle-DB, i will response issues if jump MySQL to Oracle-DB is success complete.
That would be grate and tell me if there is anything I can do from my side.
I tested all day
The main problem now is:
Some of the bambooBSC table field name is a reserved word in Oracle-DB, like SYS_ID
So if want to transfer to Oracle, then the table mapper Object and table field name must be change.
I think currently is no way support Oracle-DB in 0.6.6 release.
Support Oracle-DB plan i want move do in next version 0.7.0 .
i did manage to migrate the Database form MySQL to Oracle - using Oracle SQL Developer - with errors in the fields that marked not null but has null value, otherwise everything work perfectly.
if the migration did not work the try to copy the tables one by one and that at lest create the table empty.
and here is a drop form Oracle SQL Developer for the table TB_SYS:
CREATE_ TABLE "BBCORE"."TB_SYS" ( "OID" CHAR(36 CHAR) NOT NULL ENABLE, "SYS_ID" VARCHAR2(10 CHAR) NOT NULL ENABLE, "NAME" VARCHAR2(100 CHAR) NOT NULL ENABLE, "HOST" VARCHAR2(200 CHAR) NOT NULL ENABLE, "CONTEXT_PATH" VARCHAR2(100 CHAR) NOT NULL ENABLE, "IS_LOCAL" VARCHAR2(1 CHAR) NOT NULL ENABLE, "ICON" VARCHAR2(20 CHAR) NOT NULL ENABLE, "CUSERID" VARCHAR2(24 CHAR) NOT NULL ENABLE, "CDATE" DATE NOT NULL ENABLE, "UUSERID" VARCHAR2(24 CHAR), "UDATE" DATE ) SEGMENT CREATION IMMEDIATE PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT) TABLESPACE "SYSTEM" ;
i want to correct the JDBC_URL to jdbc:oracle:thin:username/password@Server:1521:xe
i has manage to migrate the Database form MySQL to Oracle completed.
but in bambooBSC runtime run SQL error, because SYS_ID cannot in SQL SELECT command with Oracle-DB expression version.
This SQL command cannot run with my Oracle-DB expression version.
It need for bambooBSC.
select
tbsys0_.OID as OID1_2_,
tbsys0_.CDATE as CDATE2_2_,
tbsys0_.CONTEXT_PATH as CONTEXT_PATH3_2_,
tbsys0_.CUSERID as CUSERID4_2_,
tbsys0_.HOST as HOST5_2_,
tbsys0_.ICON as ICON6_2_,
tbsys0_.IS_LOCAL as IS_LOCAL7_2_,
tbsys0_.NAME as NAME8_2_,
tbsys0_.SYS_ID as SYS_ID9_2_,
tbsys0_.UDATE as UDATE10_2_,
tbsys0_.UUSERID as UUSERID11_2_
from
tb_sys tbsys0_
where
1=1
and tbsys0_.SYS_ID='CORE'
| gharchive/issue | 2016-09-15T10:52:28 | 2025-04-01T04:33:37.798963 | {
"authors": [
"billchen198318",
"maltourky"
],
"repo": "billchen198318/bamboobsc",
"url": "https://github.com/billchen198318/bamboobsc/issues/34",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
226692016 | Fix block comments.
Fixes #9 by assigning the leadingComments and comments from the property to type rather than type.value.
Thanks!
| gharchive/pull-request | 2017-05-05T21:06:37 | 2025-04-01T04:33:37.803238 | {
"authors": [
"billyvg",
"wdhorton"
],
"repo": "billyvg/codemod-proptypes-to-flow",
"url": "https://github.com/billyvg/codemod-proptypes-to-flow/pull/10",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
167913020 | Sorting Issue
There seems to be some issue with the sorting algorithm:
This seems like a duplicate of https://github.com/billyvg/pokemon-journal/issues/6 and kind of fixed in https://github.com/billyvg/pokemon-journal/pull/17
Yep, this should be fixed in #17
| gharchive/issue | 2016-07-27T17:40:27 | 2025-04-01T04:33:37.805128 | {
"authors": [
"Nicte",
"billyvg",
"onilton"
],
"repo": "billyvg/pokemon-journal",
"url": "https://github.com/billyvg/pokemon-journal/issues/7",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
345666872 | Honesty Store Logo - Remove it
NAME OF PAGE/COMPONENT
All Screens
PART OF DESIGN NOT MATCHING
Remove the honesty store logo
Logo is still present in UX spec - need confirmation of where to replace with back buttons to continue
@isaac-lapworth https://drive.google.com/file/d/16ee202YhVJ1f1-3c9zMwaO-ZTJLBvQxI/view?usp=sharing please find the spec without the HS logo linked here.
The Honesty Store logo has been removed from all pages
| gharchive/issue | 2018-07-30T08:51:26 | 2025-04-01T04:33:37.824893 | {
"authors": [
"KyleADay",
"isaac-lapworth",
"siobhanobrien"
],
"repo": "bin-nuance-bat/project",
"url": "https://github.com/bin-nuance-bat/project/issues/222",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
404396596 | stringify_keys returns whole config when used with a Hash value
I've a config which has one value as a Hash as follows:
PlannerConfig.driving
=> {"average_speed"=>50,
"driver_off_duty_time"=>10,
"max_driver_on_duty_time"=>14,
"day_one_cap_value"=>300,
"max_continuous_drive_time"=>10,
"max_waiting_time"=>48,
"mandatory_rest_after_drive"=>0.5,
"sleeping_time"=>10}
PlannerConfig is using SettingsLogic to the values from a yaml file. The values are:
distance:
redis_cache: true
ts:
max_iterations: 1000
max_unchanged_iterations: 200
thread_count: 1
timeout: 5
shuffle: 5
cost_tolerance: 1.0
mode_selection: false
cost_calculation: true
ignore_capacity: false
constraints:
max_number_of_stops: 3
out_of_route_items_attributes: [
{distance_value: 0, percent: 40}
]
driving:
average_speed: 50
driver_off_duty_time: 10
max_driver_on_duty_time: 14
day_one_cap_value: 300
max_continuous_drive_time: 11
max_waiting_time: 48
mandatory_rest_after_drive: 0.5
max_continuous_drive_time: 10
sleeping_time: 10
task:
key: 'planner:running'
When passing this to an AR object as a parameter for initialization, it throws an error with missing attributes:
[1] pry(main)> Optimization::DrivingHoursConfiguration.new(PlannerConfig.driving)
ActiveRecord::UnknownAttributeError: unknown attribute 'distance' for Optimization::DrivingHoursConfiguration.
from /home/ashish/.rvm/gems/ruby-2.3.8/ruby/2.3.0/gems/activerecord-4.2.11/lib/active_record/attribute_assignment.rb:59:in `rescue in _assign_attribute'
This happens in Rails 4.2.x, doesn't happen in Rails 4.1.x. On investigation I found out that when using stringify_keys the output is not as expected:
[2] pry(main)> PlannerConfig.driving
=> {"average_speed"=>50,
"driver_off_duty_time"=>10,
"max_driver_on_duty_time"=>14,
"day_one_cap_value"=>300,
"max_continuous_drive_time"=>10,
"max_waiting_time"=>48,
"mandatory_rest_after_drive"=>0.5,
"sleeping_time"=>10}
[3] pry(main)> PlannerConfig.driving.stringify_keys
=> {"distance"=>{"redis_cache"=>true},
"ts"=>{"max_iterations"=>2500, "max_unchanged_iterations"=>1000},
"constraints"=>{"max_number_of_stops"=>3, "out_of_route_items_attributes"=>[{"distance_value"=>0, "percent"=>40}]},
"driving"=>
{"average_speed"=>50,
"driver_off_duty_time"=>10,
"max_driver_on_duty_time"=>14,
"day_one_cap_value"=>300,
"max_continuous_drive_time"=>10,
"max_waiting_time"=>48,
"mandatory_rest_after_drive"=>0.5,
"sleeping_time"=>10},
"task"=>{"key"=>"planner:running"},
"average_speed"=>50,
"driver_off_duty_time"=>10,
"max_driver_on_duty_time"=>14,
"day_one_cap_value"=>300,
"max_continuous_drive_time"=>10,
"max_waiting_time"=>48,
"mandatory_rest_after_drive"=>0.5,
"sleeping_time"=>10}
This is happening due to change in activesupport where we're using self.class.new instead of populating a hash, reference commit here. So it does PlannerConfig.new which adds all the keys and hence getting this error.
For now I am doing Optimization::DrivingHoursConfiguration.new(PlannerConfig.driving.to_h).
Moved to https://github.com/settingslogic/settingslogic/issues/19.
| gharchive/issue | 2019-01-29T17:12:31 | 2025-04-01T04:33:37.907683 | {
"authors": [
"ashish91"
],
"repo": "binarylogic/settingslogic",
"url": "https://github.com/binarylogic/settingslogic/issues/85",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
443768257 | automatic run nodejs and throw some error
pyspider version: 0.3.10
Operating system: windows 10
Start up command: pyspider
Expected behavior
normal run
Actual behavior
throw Error: Cannot find module 'express' * N
exec 2 nodejs process and exit over over again
How to reproduce
install nodejs into windows path
then run pyspider
solved
remove nodejs from windows path
You'll need to install node components: express, puppeteer, and body-parser before starting
@snail-lb
no ... remove node.exe from path
running no error
I was getting the same error. The documentation is not updated.
In pyspider/fetcher/puppeteer_fetcher.js you can find this:
const express = require("express");
const puppeteer = require('puppeteer');
const bodyParser = require('body-parser');
I have installed the 3 node packages within the pyspider folder:
$ npm install express puppeteer body-parser
and the error is gone.
| gharchive/issue | 2019-05-14T08:06:13 | 2025-04-01T04:33:37.920567 | {
"authors": [
"huaji0353",
"hubitor",
"snail-lb"
],
"repo": "binux/pyspider",
"url": "https://github.com/binux/pyspider/issues/898",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1175407475 | Attachments: save to cache folder instead of Downloads
Currently, attachments are directly downloaded to the Download folder, which can be annoying for ephemeral images and such. It'd be nice to have a way to configure and/or move/copy files to the Download folder on demand.
Done in https://github.com/binwiederhier/ntfy-android/commit/8339bc9c2a385d5ee685deaba8bb0eef4be80c7c, will be in 1.11.0
| gharchive/issue | 2022-03-21T13:41:53 | 2025-04-01T04:33:37.921941 | {
"authors": [
"binwiederhier"
],
"repo": "binwiederhier/ntfy",
"url": "https://github.com/binwiederhier/ntfy/issues/181",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
411795026 | Реализовать сериализацию PDB в Text
Спецификация здесь: http://www.wwpdb.org/documentation/file-format-content/format33/v3.3.html
Added in https://github.com/biocad/cobot-io/commit/e53b634e79ebf770624eca4d421b5cc7f80436b3
| gharchive/issue | 2019-02-19T08:08:59 | 2025-04-01T04:33:37.930665 | {
"authors": [
"AlexKaneRUS",
"zmactep"
],
"repo": "biocad/cobot-io",
"url": "https://github.com/biocad/cobot-io/issues/7",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1050797964 | Add debug_model function
This adds a helper function to debug a model, which returns test inputs, outputs, expected outputs as well as relevant intermediates.
In order to implement it I have changed the pre-and-postprocessing calls in PredictionPipeline s.t. they also return the stats parameter; otherwise it's not possible to use the postprocessing function.
returning the stats is good and it's just internal anyway. 👍
| gharchive/pull-request | 2021-11-11T09:56:51 | 2025-04-01T04:33:38.039845 | {
"authors": [
"FynnBe",
"constantinpape"
],
"repo": "bioimage-io/core-bioimage-io-python",
"url": "https://github.com/bioimage-io/core-bioimage-io-python/pull/141",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
806599876 | API can persist data-processing progress
Background
Link to issue
https://biomage.atlassian.net/browse/BIOMAGE-500
https://biomage.atlassian.net/browse/BIOMAGE-534
Link to staging deployment URL
https://ui-gk4wi78my831kkcxbf13nz8gkh.scp-staging.biomage.net/
Links to any Pull Requests related to this
N/A
Anything else the reviewers should know about the changes here
Joined the two tickets together are they both involved some refactoring of the API code.
Changes
Code changes
Refactored the API schemas to be in separate files for easier readability of diffs in merge requests.
Added an additional schema for the meta part of the Redux store.
Persisting processing steps can then be done by the existing Redux action for saving the actual steps' settings.
Definition of DONE
Your changes will be ready for merging after each of the steps below have been completed:
Testing
[x] Unit tests written
[x] Tested locally with Inframock
[ ] Deployed to staging
To set up easy local testing with inframock, follow the instructions here: https://github.com/biomage-ltd/inframock
To deploy to the staging environment, follow the instructions here: https://github.com/biomage-ltd/biomage-utils
Documentation updates
Is all relevant documentation updated to reflect the proposed changes in this PR?
[ ] Relevant Github READMEs updated
[ ] Relevant wiki pages created/updated
Approvers
[ ] Approved by a member of the core engineering team
[ ] (UX changes) Approved by vickymorrison (this is her username, tag her if you need approval)
Just before merging:
[ ] After the PR is approved, the unstage script in here: https://github.com/biomage-ltd/biomage-utils is executed. This script cleans up your deployment to staging
Optional
[ ] Photo of a cute animal attached to this PR
Staging available at https://ui-gk4wi78my831kkcxbf13nz8gkh.scp-staging.biomage.net/.
Looks all good to me. Some changes requested to the API schema are better allocated to task on persisting DataIntegration in DynamoDB
@ivababukova approved the changes but can't approve via GitHub now.
| gharchive/pull-request | 2021-02-11T17:35:09 | 2025-04-01T04:33:38.068716 | {
"authors": [
"aerlaut",
"marcellp"
],
"repo": "biomage-ltd/api",
"url": "https://github.com/biomage-ltd/api/pull/45",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1736983816 | Access to biomarker/data, api, and sparql
Please provide Daniall, Robel, Jonathon, Pat, and Karina access to
hive-lab/biochemistry.edu/biomarker/data
hive-lab/biochemistry.edu/biomarker/api
hive-lab/biochemistry.edu/biomarker/sparql
I don't think this is something I can do. I am not familiar with these.
Would this be a task for Robel, or someone else?
Daniel Lyman, PhD
Senior Research Associate
Cancer Biomarkers Project Lead, HIVE-Lab
MS in Bioinformatics and Biochemistry Program Coordinator
Department of Biochemistry and Molecular Medicine
School of Medicine & Health Sciences
The George Washington University
Ross Hall
2300 Eye Street N.W.
Washington, DC 20037
Phone lab: 202-994-3639
MS Program Office hours: T/W 9:00am-5:00pm
On Thu, Jun 8, 2023 at 12:09 PM Hadley King @.***>
wrote:
I don't think this is something I can do. I am not familiar with these.
—
Reply to this email directly, view it on GitHub
https://github.com/biomarker-ontology/Biomarker-Database-Curation/issues/8#issuecomment-1582943068,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/A4WWYURATDB4N7ND334CTQDXKH2LBANCNFSM6AAAAAAYXMVESI
.
You are receiving this because you were assigned.Message ID:
@.***
.com>
Probably. I am not even sure what the paths you supplied are referencing. Where is that? What resource is it?
I'll reassigne to Daniall and Robel
Daniall
Please work with Robel on these issues.
Sean has been able to complete these and is working on sparql
| gharchive/issue | 2023-06-01T19:54:09 | 2025-04-01T04:33:38.076073 | {
"authors": [
"DaniallMasood",
"HadleyKing",
"danlymangw"
],
"repo": "biomarker-ontology/Biomarker-Database-Curation",
"url": "https://github.com/biomarker-ontology/Biomarker-Database-Curation/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2737076560 | 🐛 Biome extension & CLI not consistent
VS Code version
1.93.1
Extension version
2.3.0
Biome version
1.9.4
Operating system
[ ] Windows
[X] macOS
[ ] Linux
Description
Biome CLI (via package.json) and Biome extensions have contradictory behaviors when it come to formatting.
Steps to reproduce
Here is my biome.json:
{
"$schema": "https://biomejs.dev/schemas/1.9.4/schema.json",
"vcs": {
"enabled": true,
"clientKind": "git",
"useIgnoreFile": true
},
"formatter": {
"enabled": true,
"useEditorconfig": true,
"formatWithErrors": true,
"indentStyle": "tab",
"lineEnding": "lf",
"indentWidth": 2,
"lineWidth": 100,
"attributePosition": "auto",
"bracketSpacing": true,
"ignore": ["**/dist", "**/coverage"]
},
"organizeImports": {
"enabled": false
},
"linter": {
"enabled": false
},
"javascript": {
"formatter": {
"enabled": true,
"jsxQuoteStyle": "double",
"quoteProperties": "asNeeded",
"trailingCommas": "all",
"semicolons": "always",
"arrowParentheses": "asNeeded",
"bracketSameLine": false,
"quoteStyle": "single"
}
},
"json": {
"formatter": {
"enabled": true,
"trailingCommas": "none"
}
},
"files": {
"include": [
"src/**/*.ts",
"src/**/*.tsx",
"src/**/*.js",
"src/**/*.jsx",
"src/**/*.json",
"test/**/*.ts",
"test/**/*.tsx",
"test/**/*.js",
"test/**/*.jsx",
"test/**/*.json"
],
"ignore": ["**/node_modules", "**/dist", "**/coverage"]
}
}
Install biome (package & extension) and add the following scripts:
"format": "biome check ."
run format document via vscode, then run pnpm format --fix
Contradictory & self-cancelling behaviors. I believe it started with a recent update.
https://github.com/user-attachments/assets/96edd425-564f-4ffa-a063-c130f61dfec2
Expected behavior
Expected the same behavior whether vscode, or biome package.
Does this issue occur when using the CLI directly?
Yes
Link to a minimal reproduction
No response
Logs
No response
I have also encountered out of sync issue with the extension 2.3.0 with @biomejs/biome 1.9.4 in both windsurf and cursor editors.
I resolved this by adding a [project].code-workspace config file at the root of my monorepo something like this :
{
"folders": [
{
"name": "references",
"path": "references"
},
{
"name": "workers",
"path": "workers"
},
{
"name": "app",
"path": "app"
}
],
"settings": {
"files.exclude": {
"./backup": true
}
}
}
Then open the codebase with workspace from the config file, and the extension starts to work in sync with each biome.json.
tip: I setup the local .vscode with settings.json in each workspace to enable customized biome behavior
| gharchive/issue | 2024-12-12T23:38:39 | 2025-04-01T04:33:38.083573 | {
"authors": [
"FaresKi",
"lumpinif"
],
"repo": "biomejs/biome-vscode",
"url": "https://github.com/biomejs/biome-vscode/issues/444",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
247288949 | Segmentation Fault
Hi,
I have tried running Hera on few samples, but it always crashes with a segmentation fault. No other log output is available, so I cannot be more precise about the error. It seems to happen at random times (I have rerun the same sample multiple times).
Any ideas?
Is there any possibility for you to share your data with us so we could reproduce and fix the issue?
Also, could you tell us a little bit more information about the system that you run hera on?
Unfortunately, I only have clinical in-house data (which I cannot share). I will try to dig up some RNA data from the public domain and see if I can reproduce the error. Stay tuned!
I run Hera as Slurm jobs on a CentOS HPC cluster (http://genome.au.dk/system-info/).
Thank you for your bug report & help to trace the source of the error!
Hi micknudsen,
Jemalloc on Centos could be the cause of the instability. We currently turn it off for stability purpose. Hence, the program should consume a little more memory (1-2GB). Please pull or clone again to check if it work.
Best,
Bioturing Algorithm Team.
Micknudsen, Hopefully, the new fix solves the segmentation fault problem?
I am running a test right now. I will let you know as soon as it finishes or (hopefully not) crashes.
Problem solved. Thanks!
| gharchive/issue | 2017-08-02T06:45:56 | 2025-04-01T04:33:38.101149 | {
"authors": [
"GinnyAquarius",
"bioturing",
"kspham",
"micknudsen"
],
"repo": "bioturing/hera",
"url": "https://github.com/bioturing/hera/issues/6",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2146110471 | 🛑 Bangla Plex API is down
In c120fc0, Bangla Plex API (https://banglaplexapi.bymirrorx.eu.org/popular/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Bangla Plex API is back up in d0c936e after 23 minutes.
| gharchive/issue | 2024-02-21T08:30:37 | 2025-04-01T04:33:38.106619 | {
"authors": [
"bipuldey19"
],
"repo": "bipuldey19/pinger",
"url": "https://github.com/bipuldey19/pinger/issues/540",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
200762976 | How to debug error responses?
Greetings, and thank you for writing this plugin!
I've included this plugin in my ionic v1 app, inject it into my service as $cordovaFacebookLogin, and try to authenticate to Facebook like this:
$cordovaFacebookLogin.login({ permissions: [] })
When I testing on iOS 10.2 I get back JSON that looks like this:
{"userID":"1234567890","cancelled":true,"success":false,"accessToken":"token-here","error":true}
But I didn't cancel! When I debug the access token here it says it's valid:
https://developers.facebook.com/tools/debug/accesstoken
My question is: how do I best determine what the error is here?
In case it helps, when I run that function the Facebook app opens, and I see a page that is pretty-much blank except for the word "cancel" in white at the top. Then after a few seconds the Facebook app closes and returns to my app with the above JSON.
Thanks again!
Yeah, I'm getting the same issue here, and having trouble debugging...
I actually got a successful response the first time I ran the facebook authentication, but since then I can only get the onFailure method to fire...
| gharchive/issue | 2017-01-14T00:14:07 | 2025-04-01T04:33:38.150379 | {
"authors": [
"Wade-McDaniel",
"samasmith89"
],
"repo": "bisrael/cordova-plugin-facebook",
"url": "https://github.com/bisrael/cordova-plugin-facebook/issues/23",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1199697306 | Remove submodule source packages once published
Regarding ROS2:
Remove the following git submodules once the packages are published via apt:
[ ] ros-rolling-vision-msgs
[ ] ros-rolling-soccer-vision-2d-msgs
Update bit-bots/bitbots_tools/pull/126 once resolved.
| gharchive/issue | 2022-04-11T09:41:05 | 2025-04-01T04:33:38.151946 | {
"authors": [
"jaagut"
],
"repo": "bit-bots/bitbots_meta",
"url": "https://github.com/bit-bots/bitbots_meta/issues/106",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1777789419 | github-merge: omit ACKs in quoted comments or with markdown indentation
Avoid counting an ACK if it is in a quoted comment or with markdown indentation, when present in a review before the actual ACK. See https://github.com/bitcoin/bitcoin/pull/27334#issuecomment-1610203248 for an example.
Concept ACK 9333d5f61bf85a38123a56c9bd63ef162eeec1b2
ACK 9333d5f61bf85a38123a56c9bd63ef162eeec1b2
| gharchive/pull-request | 2023-06-27T21:37:15 | 2025-04-01T04:33:38.185004 | {
"authors": [
"achow101",
"jonatack",
"katesalazar"
],
"repo": "bitcoin-core/bitcoin-maintainer-tools",
"url": "https://github.com/bitcoin-core/bitcoin-maintainer-tools/pull/154",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2178450879 | Newsletters: add 293 (2024-03-13)
[ ] Lede, releases/RCs, topic entries @harding
[ ] Bitcoin Core PR Review Club @LarryRuane
I know the review club section is light on questions (only 3), I'm having trouble finding questions that aren't too much in the weeds and also I was out of town all weekend. I have to step out for a few hours but will try to add more questions after I get back home at around 9PM MDT.
@glozow - if you have time, please feel free to add review club questions or make any changes to the ones I just pushed, no need to check with me, thanks.
Made edits and added lede and releases/RCs. I didn't review added content or add topic links; I'll try to get to those in a few hours.
| gharchive/pull-request | 2024-03-11T07:44:19 | 2025-04-01T04:33:38.304663 | {
"authors": [
"LarryRuane",
"harding"
],
"repo": "bitcoinops/bitcoinops.github.io",
"url": "https://github.com/bitcoinops/bitcoinops.github.io/pull/1549",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
105570181 | 0.11B -stealth-mode is not mentioned in command line usage
bitcoinxt-qt -? | grep stealth is blank.
It does work though, connecting to a 0.11A node using -stealth-mode makes it report.
receive version message: /Satoshi:0.11.0/: version 70002, blocks=222611, us=0.0.0.0:0, peerid=2959
Added documentation in https://github.com/bitcoinxt/bitcoinxt/pull/85
| gharchive/issue | 2015-09-09T10:37:46 | 2025-04-01T04:33:38.306233 | {
"authors": [
"Beanow",
"jlopp"
],
"repo": "bitcoinxt/bitcoinxt",
"url": "https://github.com/bitcoinxt/bitcoinxt/issues/62",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
930704675 | Implementation of SWP-08 routing protocol
Hey!
I would like companion to implement the SWP-08 router protocol. Its a protocol supported by alot of different router-systems like ross ultrix, direct out matrix-routing and lots more.
I want to be able to perform routes and read out routing info. Then be able to use the feedback from the router to make a tallysystem.
Documentation apllies to both SWP-02 and 08
https://wwwapps.grassvalley.com/docs/Manuals/sam/Protocols and MIBs/Router Control Protocols SW-P-88 Issue 4b.pdf
Now i have been testing a bit and here are som points:
How can i get the names from the router onto my buttons? should they be stored in variables?
I want to have a feedback that i can but on the sources that change color when the destination it is routed to is selected.
It should be a variable with name that is the source of the selected destination.
I used the "select destination" and when I choose one destination and then another, the previuosly chosen destination inherit the now selected destinations destination(complicated)
FYI: Although the version linked at https://we.tl/t-5nurvA8mcQ is a different size, it has the same version number as the last one of 2.2.0 (2.2.0-061e92b1-3423)
Re: level multiselect.
The observed behaviour of the buttons in the emulator do not seem to reflect multiselect. In normal button mode, each button will flash the active (yellow top bar) indication when pressed and the feed back will light up the last button pushed. In Latch/Toggle button mode, each button will light the active (yellow top bar) indication when pressed; however, it will never turn off. The feed back will still light up the last button pushed. I would guess that the buttons would want to be normal rather than using the predefined Latch/Toggle logic?
Is this behavior simply an anomaly of the emulator that is not present with an active router controller?
Also, this may be a feature request; but, I see the ability to select more than one level per action & feedback is present; would it be possible to define overlapping level buttons?
For example, having four buttons for three levels where the first three are toggle on/off and the fourth is an override to toggle all three selected on or off. Here would be the definitions for these buttons for testing.
Button name: L1
Press Action: SW-P-08: Select Levels 1
Feedback: SW-P-08: Selected Levels 1
Button name: L2
Press Action: SW-P-08: Select Levels 2
Feedback: SW-P-08: Selected Levels 2
Button name: L3
Press Action: SW-P-08: Select Levels 3
Feedback: SW-P-08: Selected Levels 3
Button name: All Levels
Press Action: SW-P-08: Select Levels 1,2,3
Feedback: SW-P-08: Selected Levels 1,2,3
Selected levels always need to be visible via feedback for those who need to be able to do split routes. As mentioned before, most people will only ever want simultaneous routing; however, for those who need to be able to break out Audio and Video levels, they need to see the current level status directly on the panel via feedback.
Turning off all level selection is a common operational procedure rather than a true lockout. Many router control panels are in areas on desks that can be bumped and being able to deselect levels reduces issues; however, usually, the same button is also used to select all levels again rather than to select each level individually so as to minimize the number of intentional button pushes. So a clear option is only useful with a select all option to go with it.
Re: lockouts, while SW-P-08 has a robust set of destination lock and router lock options, it is probably best not to support those at all; as those can be much more complex than routing itself, and, if someone needs that level of control they should probably be solving that problem direct with the specific router vendor. I think that using the existing lockout options within Companion is more than sufficient.
I think I understand that the selection level buttons need to be additive, so that combinations can be created. This will take a bit of thought.
One thing I have done already is the make the clear action more configurable so it can clear only levels if necessary.
Thanks.
I also notice that there is an additional method to get feedback beyond the way I mentioned above.
Feedback: internal: Check variable value SW-P-08:Level - Selected levels = []
That method has available operators of [ =, !=, >, and < ]; however, does not have [AND] [OR] options so it is not clear how that would support multi level feedback.
The Feedback: SW-P-08: Selected Levels method on the other hand seems to support multiple levels already so that may not need any further changes.
I currently have 2 modes of operation. Firstly:
level select and deselct which can take multiple levels
source select
destination select
take or clear
This makes one 'set' of actions. Then we also have:
crosspoint a level/source/destination in a single action
I'm wondering if this second mode takes the level selection in the first method or ignores it?
My vote would be to keep the operational modes separate if possible. @evenover do you have a preference? If a single button push is defined, it should have the ability to have control over all aspects of that action regardless of what state the control surface was previously in.
For the other mode of operation, the default mode should be that all levels are selected at startup or after a Clear command. The ability to split levels is a more advanced method and should only be enabled if explicitly requested. For example, if I walk up to a control surface and do not know what the previous condition was, this is the typical expected sequence.
Hit Clear. {Reset to default condition.}
Select a Destination. {This allows the ability to see what is currently routed there.}
Select a new Source.
Hit the Take button. {Modifying the levels is optional and can be changed at any point before the Take.}
This should result in all of levels of the selected source being routed to that destination. If the levels are changed for a route those can then persist until cleared or manually changed. BTW: Some control systems implement a timeout function to Clear the routing presets after a period in case someone has a complex route setup and then walks away, which could cause unexpected results for the next user.
It seems to be a good idea @chibiconsulting. A timeout-function would also be great.
The way i am planing to use this module is to have kind of a catergory select. So i'm going to build separate pages for Video, audio and then video+audio(video+8 audio probably). This way i am not going to make use of turning levels on and off but just add level-selection to the button with the destination.
I think this protocol actually only supports routing 1 level at the time, thats probably why its hard to get feedback of multiple levels selected.
I have managed to obtain the latest full documentation of swp08 with extended routing/labels from Grass Valley, but i have to check that my employer do not have a NDA with Grass before i can supply it to you. Hopefully its not a big deal and i will give it to you as soon as i know.
@evenover It's great that you have the documentation but remember that companion and this module is open source so GV need to be aware that it's trivial to reverse engineer their spec from the code in this module.
Hi guys, in order to get this module out to more potential users it has been included in the next beta of companion as the basic routing functionality is all working. This doesn't mean I won't make the more complex changes requested above.
Thats great!
Any updates on the features:
Change color on background of source of destination when destination is selected
variables for label names which can be added to buttons
variable with the source name of selected destination
1 and 2 are pretty much the same features as the videohub-module.
I took a look in the manual for ultrix and it only supports 8-char labels, but 16char umds of the sources. In the extended documentation its possible to ask for 4, 8 or 12. Hopefully i can disclose it with you asap.
I've redone the level select/deselect actions so they work as toggles, next step is to make feedbacks work for those and then the list above.
The level commands have been rewritten to work as follows:
Select Level: add the level(s) in the action to the current selection
DeSelect Level: remove the level(s) in the action from the current selection
Feedback: Change button colour if all the level(s) in the feedback are in the current selection
You would normally put all of the above on one phyiscal button and enable latch/toggle mode. The clear action re-enables all levels up to the number set in the module configuration.
Due to all the changes you will need to delete any previous level actions and feedbacks from your configuration before trying this!
https://we.tl/t-RqZEdtxs3c
Having done more testing the above method can sometimes get in a mess when the latch state gets out of step with the level state. To avoid this there's a new action which will simply toggle (invert) the level(s) in the action each time it is pressed. This is much preferable to using the button latch mode.
https://we.tl/t-LmLxyBaQ26
I like the SW-P-08: Toggle Levels action, nice. Can still see some people using the SW-P-08: Select Levels and SW-P-08: De-Select Levels actions, hope those stay as well.
@evenover
How can i get the names from the router onto my buttons? should they be stored in variables?
I would think that if available by feedback it would be enough, or do you have something else in mind?
i.e. Feedback: SW-P-08: Name of target. or Name of Source and Name of Destination for example.
2./3. I want to have a feedback that i can put on the sources that change color when the destination it is routed to is selected.
Something like Feedback: SW-P-08: Source routed to Destination? Would need to query the selected destination for the current Source. What happens if the routed Source changes after the button is updated? Could use a lot of cycles to keep checking.
I used the "select destination" and when I choose one destination and then another, the previously chosen destination inherit the now selected destinations destination(complicated)
I think that Evertz implements this as a virtual destination which is a destination that can then be used as a source somewhere else. I do not recall SW-P-08 ever having that capability. Might make sense as a separate, more generic request to provide that function for all router protocols. If a range of virtual sources/destinations are defined as being linked and held separate from the router, requests to / from those can be translated on the fly and dropped back to the native router protocol. Not a trivial problem by any means.
How can i get the names from the router onto my buttons? should they be stored in variables?
I would think that if available by feedback it would be enough, or do you have something else in mind?
i.e. Feedback: SW-P-08: Name of target. or Name of Source and Name of Destination for example.
Feedbacks are really intended for modifying the button colour or text, I think what is being asked here is that the button label can be set from a variable, please confirm.
I want to have a feedback that i can put on the sources that change color when the destination it is routed to is selected.
Something like Feedback: SW-P-08: Source routed to Destination? Would need to query the selected destination for the current Source. What happens if the routed Source changes after the button is updated? Could use a lot of cycles to keep checking.
Currently the router sends a message every time a route changes (look in the debug log to see these) so I need to write some code to keep track of all these messages. What makes it complicated compared to say the videohub is keeping track of all the potential breakaway routes.
I used the "select destination" and when I choose one destination and then another, the previously chosen destination inherit the now selected destinations destination(complicated)
I think that Evertz implements this as a virtual destination which is a destination that can then be used as a source somewhere else. I do not recall SW-P-08 ever having that capability. Might make sense as a separate, more generic request to provide that function for all router protocols. If a range of virtual sources/destinations are defined as being linked and held separate from the router, requests to / from those can be translated on the fly and dropped back to the native router protocol. Not a trivial problem by any means.
This is a quite advanced feature and in my opinion companion is generally aimed at casual or supplementary use cases. For this kind of complexity you should probably be using the Ross panels or similar.
Now this module has a repository we can log all these requests separately and this thread is getting too long. I have transferred the feature requests I am aware of to new issues. Please check and update as required. Please log new bugs and features separately.
I can't find the repository, could you tell me where to find it? All the wetransfer links are obviously expired.
Thanks
Now this module has a repository we can log all these requests separately and this thread is getting too long. I have transferred the feature requests I am aware of to new issues. Please check and update as required. Please log new bugs and features separately.
@scp85 The latest build is in the beta companion app - https://bitfocus.io/companion/download/builds/
is there any IP based emulator available for SWP-08 routing protocol?
is there any IP based emulator available for SWP-08 routing protocol?
Not that I'm aware of.
What were you thinking using it for?
I’m trying to figure out ways to control an Swp-08 audio/video router (grassvalley (used to be Snell Andvanced Media) Sirius S830)) with a PC/ audio desk..
On 26 Jan 2022, at 14:57, evenover @.***> wrote:
is there any IP based emulator available for SWP-08 routing protocol?
Not that I'm aware of.
What were you thinking using it for?
—
Reply to this email directly, view it on GitHub, or unsubscribe.
Triage notifications on the go with GitHub Mobile for iOS or Android.
You are receiving this because you were mentioned.
This module should be working fine. I've done some testing on my Ross Ultrix and it looks good for now. If you have over 1000 sources and destinations there are still some developments that need to happen.
Ouch! 8000 I/o for me.
I’ll still give it a go tho
On 26 Jan 2022, at 14:59, Paolo Scillieri @.***> wrote:
I’m trying to figure out ways to control an Swp-08 audio/video router (grassvalley (used to be Snell Andvanced Media) Sirius S830)) with a PC/ audio desk..
On 26 Jan 2022, at 14:57, evenover @.***> wrote:
is there any IP based emulator available for SWP-08 routing protocol?
Not that I'm aware of.
What were you thinking using it for?
—
Reply to this email directly, view it on GitHub, or unsubscribe.
Triage notifications on the go with GitHub Mobile for iOS or Android.
You are receiving this because you were mentioned.
If we get the extended comands implemented it will be able to control 65536 sources and destinations.
I'm afraid that because I don't require the extended commands in my use cases the development time would need to be paid for somehow.
I understand, any idea on what kind of payment? I might be able to fix some funding.
I think £200 would cover it.
Awesome, that should be no problem!
Would it then be possible to see if you could get the "change background color of selected source" working?
Maybe... let's do the extended commands first. Please email me to discuss payment mail-at-pjdaniel-dot-org-dot-uk
Any chance you would know if this is able to work with a nVision NV9000 Router controller from Grass Valley?
Any chance you would know if this is able to work with a nVision NV9000 Router controller from Grass Valley?
Hi, I think it could be working, NV9000 is on swp02. This is router protocol swp08 is remote protocol and we are using swp08 for evertz switcher control thru CEREBRUM automatization via swp08
Thanks, hmmmm now I just have to figure out how to set that up lol
Hello team, I'm curious to know whether it's possible to disconnect a crosspoint using the Pro-Bel SW-P-08 protocol. Can a crosspoint connection be directly removed, or is the standard procedure to always redirect the input to a specific output?
Moreover, is there a way to work-around this behavior? For example, could we route the input to a designated output that can be deactivated on the physical device?
Thank you very much in advance !
No. This is not specific to SW-P-8 control, all routers have this constraint.
For Video or Audio routers, the standard practice is to route a specific input to any outputs that no longer require the previous input. Examples would be a blank screen and silent audio or bars and tone. Most systems already have a source for those anyway.
This is due to the physics of the crosspoint switching mechanism and associated amplification within the device. To put it another way, an unterminated source being output by a router would cause undesirable operation. ;)
Note that data routers are an exception, where the router itself should define a disconnect source and you would need to check the specifics with the data sheet for those devices; however, from a protocol perspective, it still places a crosspoint select message using whatever the disconnect source is defined as.
Thank you @chibiconsulting for your prompt and informative response; I was seeking for anything like this...
All uncertainties are resolved!
Thanks, hmmmm now I just have to figure out how to set that up lol
I'm wandering if you managed to get it to work ? Because I'm using a NV9000 too but it's not working for me for now
Thanks
Thanks, hmmmm now I just have to figure out how to set that up lol
I'm wandering if you managed to get it to work ? Because I'm using a NV9000 too but it's not working for me for now
Thanks
Hi we're using SW-P-08 generic and works for us, just fill IP and port what we are using and it works. Truth is that we are using semi-software Cerebrum for manage control of matrix. Means Elgato->swp08->Cerebrum->swp08->Evertz
@daniep01 @evenover is this module in good shape now? I'll arrange funding if it can be improved in some way.
Hi William,
I can probably speak to that reasonably having done a bit of work on the module recently. It works reasonably but the implementation is not very complete. A few obvious areas for improvement:
Request and process Tally Dump
Better multilevel crosspoint tallies
Protect / Extended Protect Set / Tally / Request Tally Dump
One of the potential pitfalls with SW P 08 is that there is a large number of commands specified, but few if any devices support all of them. I faced a similar issue with the SW P 02, and only implemented the scope I could test.
For instance this is the portions of SW P 08 that a specific Direct Out device supports. Perhaps a quarter of all commands listed, which is probably not complete to begin with.
Hello,
I’m experiencing issues with the SWP-08 and the Ultrix FR-12 Frame. While I can send control commands, I’m not receiving any feedback or information from the equipment.
Has anyone worked with Ross Ultrix systems and encountered a similar issue? Any troubleshooting suggestions would be greatly appreciated.
Thanks!
Yep, we use it with a smaller Ultrix. I can have a look at the Ultrix side config next week and see if there is anything notable we had to do.
Can you please check the module logs for return data, and let me know what you find.
Hi Guys, Im a solutions architect with the Hyerconverged team at Ross. The FR12 uses the Ultricore BCS as a Control system. Make sure you are sending your commands to the Ultricore BCS and not to the frame directly. Also it is possible that the probel service is not enabled. We implimented new security features that disables most services by default now (version 5.5 and up). On the BCS under product info, services.
I've tested this today with Companion 3.4.3 and Ultricore BCS / Ultrix version 5.6 and 6.2 (with an FR12, FR5, FR2) and it works as expected.
Hi Guys, Im a solutions architect with the Hyerconverged team at Ross. The FR12 uses the Ultricore BCS as a Control system. Make sure you are sending your commands to the Ultricore BCS and not to the frame directly. Also it is possible that the probel service is not enabled. We implimented new security features that disables most services by default now (version 5.5 and up). On the BCS under product info, services.
I've tested this today with Companion 3.4.3 and Ultricore BCS / Ultrix version 5.6 and 6.2 (with an FR12, FR5, FR2) and it works as expected.
Hello, indeed my issue was there; I was not pointing to the virtual IP address of the cluster. The grid now appears perfectly in the module. Thank you!
| gharchive/issue | 2021-05-15T12:07:26 | 2025-04-01T04:33:38.368916 | {
"authors": [
"Corent1CSN-LAB",
"Gbou7",
"InertiaImpact",
"chibiconsulting",
"daniep01",
"engelCZE",
"evenover",
"karthickvkl",
"mplakalovi1",
"phillipivan",
"pomerle",
"scp85",
"willosof"
],
"repo": "bitfocus/companion-module-generic-swp08",
"url": "https://github.com/bitfocus/companion-module-generic-swp08/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
895793231 | add BitBasicList component (#269 )
This closes #269
Since the assignee of the related issue is changed, I'm closing this PR.
| gharchive/pull-request | 2021-05-19T19:15:13 | 2025-04-01T04:33:38.378002 | {
"authors": [
"msynk",
"zoha-shobbar"
],
"repo": "bitfoundation/bitframework",
"url": "https://github.com/bitfoundation/bitframework/pull/409",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2025484970 | Enter amount : input field appears disabled
Even after entering a value, the input field still appears disabled, retaining the disabled appearance despite the entered value.
Resolved,
| gharchive/issue | 2023-12-05T07:42:38 | 2025-04-01T04:33:38.379766 | {
"authors": [
"Deveshshankar",
"akshay8551"
],
"repo": "bithyve/bitcoin-keeper",
"url": "https://github.com/bithyve/bitcoin-keeper/issues/3503",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
940695184 | Swan Buy integration
Mostly done, need to be tested
Verified this on dev app v1.8.0(317)
Verified this Swan Buy integration in dev app v1.8.0(317)
It is working properly in android device.
In ios it is asking to open in "Hexa Stage"? and when we click on "Open" it gives a msg as ""Hexa stage" Beta has expired" and after clicking on "OK" when we again click on "Authorize Hexa wallet" it throws an error as {"message : "invalid_request"}.
Still the Swan is not working on both the device android ans ios as well.
-Tested on dev app v1.8.0(319)
Uploading 21-07-13-19-48-24.mp4…
I was able to complete Swan journey in #317. The deeplink didn't open the modal in #319.
@cakesoft-swati - are any deep links working?
[ ] swan deeplink not opening from the app
Now the deeplink is working for swan account.
Tested in dev app v1.8.5(325)
| gharchive/issue | 2021-07-09T11:52:21 | 2025-04-01T04:33:38.383914 | {
"authors": [
"AliMeer",
"antuz123",
"cakesoft-swati"
],
"repo": "bithyve/hexa",
"url": "https://github.com/bithyve/hexa/issues/3732",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
137392737 | Update forms.py to fix issue #1
improve compatibility with Django version 1.7+
@Kyle-Falconer Thanks a lot!
| gharchive/pull-request | 2016-02-29T22:08:08 | 2025-04-01T04:33:38.385125 | {
"authors": [
"Kyle-Falconer",
"Tyrdall"
],
"repo": "bitmazk/django-multilingual-survey",
"url": "https://github.com/bitmazk/django-multilingual-survey/pull/2",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
70525435 | can.view.href
This pull request adds the can-href attribute for easy linking.
Not ready to merge yet, finalization incoming.
Nevermind, needs to be minor anyway.
| gharchive/pull-request | 2015-04-23T22:10:35 | 2025-04-01T04:33:38.515223 | {
"authors": [
"daffl"
],
"repo": "bitovi/canjs",
"url": "https://github.com/bitovi/canjs/pull/1641",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
59858252 | Move less plugin into stealjs/system-less
I've moved this issue out to stealjs/system-less#1. We should discuss anything about this issue not related to actual core steal features over there. Gonna keep this ticket open for tracking purposes.
As part of this, we should actually move the steal plugin out of this repository and out into that one and make sure it's easily installable/accessible. Perhaps we might still want to include the file in steal releases, as well?
Yes, I think we should keep it in the releases but moving it to this new repo is a good idea. That way people who want a more current version don't have to wait for a steal release.
Also, there's a couple of less tests that can probably be moved. There's a test that checks for steal's integration with less... that one should not be moved, but any that just tests the functionality of the plugin can, I think.
This one should be pretty easy for someone looking to contribute. What needs to be done is:
[ ] - Move ext/less.js and ext/less-*.js to the system-less repo.
[ ] - Move less tests there as well (note: not this one which tests that less is loaded by default)
[ ] - Setup CI for that repo
[ ] - Update steal Gruntfile to move less.js, less-*.js to the ext folder where it was before. We do the same sort of thing for system-npm so you can use that as an example.
This will make it easier to keep the less plugin up-to-date with Less releases.
| gharchive/issue | 2015-03-04T20:39:32 | 2025-04-01T04:33:38.518940 | {
"authors": [
"matthewp",
"zkat"
],
"repo": "bitovi/steal",
"url": "https://github.com/bitovi/steal/issues/346",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
82437951 | Error when creating payout: This endpoint does not support the pos facade
I tried the example found here which unfortunately result in an error using testnet.
{"error":"This endpoint does not support the `pos` facade"}
The request looks as follows
POST https://test.bitpay.com:443/payouts HTTP/1.1
...
{"token":"__REPLACED__","amount":10,"currency":"USD","instructions":[{"label":"Test","address":"__REPLACED__","amount":10}],"effectiveDate":"2015-05-29T14:05:38.000Z","pricingMethod":"bitcoinbestbuy","guid":"__REPLACED__","nonce":__REPLACED__,"reference":"123","notificationURL":"__REPLACED__"}
Note: REPLACED was placed by me to not disclose private information.
@4ley the payouts endpoint requires the payroll facade.
As I understood, I have to set it on the token class. Thats what I did. Here is my code:
$instruction = new \Bitpay\PayoutInstruction();
$instruction
->setAmount(10)
->setAddress('__REPLACED__')
->setLabel('Test');
$time = gmdate("Y-m-d\TH:i:s\.", time()+3000)."000Z";
$token = new \Bitpay\Token();
$token
->setFacade('payroll')
->setToken('__REPLACED__');
$payout = new \Bitpay\Payout();
$payout
->setEffectiveDate($time)
->setCurrency(new \Bitpay\Currency('USD'))
->setPricingMethod('bitcoinbestbuy')
->setReference('__REPLACED__')
->setNotificationUrl('__REPLACED__')
->setToken($token)
->addInstruction($instruction);
Ok, gotcha. It sounds like your account was paired with a POS token originally. That's the least-privileged access token (and the default one created). However, if you make a call to getTokens() it should respond with all of the access tokens associated with your account. If you do have a payroll token, use that one, otherwise the default POS token will be returned and used. Here's the example code for that call: https://github.com/bitpay/php-bitpay-client/blob/master/examples/GetTokens.php
Hm, very strange the $client->getTokens() returns NULL
@4ley , have you created any tokens? You might want to follow our tutorials.
In 001.php, keys are generated and stored.
In 002.php, a (POS) token is created. You'll want to modify the create token snippet to the following to create a payroll token:
$token = $client->createToken(
array(
'facade' => 'payroll',
'label' => 'You can insert a label here',
'id' => (string) $sin,
)
);
That'll create a payroll token, but you'll still need to pair it with your account. At the bottom of the same file, after $persistThisValue = $token->getToken(); add $pairingCode = $token->getPairingCode();
And then print out $pairingCode however you like. Log in to your BitPay account and then go to Access Tokens and paste the pairing code and click Find. If everything worked, your account will now be associated with that payroll facade token.
Your code gives me the following error:
PHP Fatal error: Uncaught exception 'Bitpay\Client\ArgumentException' with message 'pairing code is not legal' in /home/www/includes/bitpay/vendor/bitpay/php-client/src/Bitpay/Client/Client.php:474
What version of the PHP-Bitpay-Client are you using?
The VERSION file says 2.0.0
To use non-POS facade tokens, you'll need to be using the unreleased master branch.
To do this, in your composer json, the require for our client should like like the following:
...
"require": {
...
"bitpay/php-client": "dev-master"
...
},
...
where the ... are other possible code snippets
Okay I made the update, now it gives me
PHP Fatal error: Uncaught exception 'Bitpay\Client\BitpayException' with message '500: Facade not allowed' in /home/www/master/includes/bitpay/vendor/bitpay/php-client/src/Bitpay/Client/Client.php:502
I just checked, and you're right, the payroll facade isn't listed. It requires manual approval to get a token for that facade, so you'll need to contact support@bitpay.com with a request for approval.
Thanks. I've sent a request.
| gharchive/issue | 2015-05-29T13:54:43 | 2025-04-01T04:33:38.532067 | {
"authors": [
"4ley",
"ionux",
"sambohler"
],
"repo": "bitpay/php-bitpay-client",
"url": "https://github.com/bitpay/php-bitpay-client/issues/145",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
232783729 | reconciliation error
while trying to fetch reconciliation report getting error like "This endpoint does not support the merchant facade".
reconciliation is for merchant facade itself.
Fetching a settlement report (https://bitpay.com/api#resource-Settlements) is a 2 step approach:
Please use the merchant facade to fetch the settlement report summary for a date range.
The response will contain a resource token. Please use the resource token to fetch the settlement report details.
| gharchive/issue | 2017-06-01T07:03:18 | 2025-04-01T04:33:38.534284 | {
"authors": [
"pieterpoorthuis",
"pravithapk"
],
"repo": "bitpay/php-bitpay-client",
"url": "https://github.com/bitpay/php-bitpay-client/issues/230",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1293725743 | [feature] add "MysqlDatabase" "MysqlUser" Add delete policy
add
"MysqlDatabase" "MysqlUser" Add delete policy
Sometimes deleting these resources is very dangerous, so adding a certain label can reduce data loss
I publish new branch to megre.review plz @cndoit18
| gharchive/pull-request | 2022-07-05T03:01:30 | 2025-04-01T04:33:38.535535 | {
"authors": [
"jiuker"
],
"repo": "bitpoke/mysql-operator",
"url": "https://github.com/bitpoke/mysql-operator/pull/828",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
139445823 | Quote tag and branch names
Branches with characters like the single-quote (') cause builds to hang indefinitely. This should solve that.
Are you sure that there's a problem with this step? We know about an escaping issue, but that's in https://github.com/bitrise-tools/bitrise-machine or in one of the sub-tools it uses to communicate with the build VMs.
We'll fix that this week.
Thanks - I just know that something hangs somewhere.
You're right, it's actually an escaping issue in one of the core tools used for pushing the parameter into the build VM. We'll fix it ASAP!
Yay!
| gharchive/pull-request | 2016-03-09T01:29:12 | 2025-04-01T04:33:38.542054 | {
"authors": [
"MisterRager",
"viktorbenei"
],
"repo": "bitrise-io/steps-git-clone",
"url": "https://github.com/bitrise-io/steps-git-clone/pull/11",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1573342170 | Revert to 6.2.2
Checklist
[x] I've read and followed the Contribution Guidelines
[x] step.yml and README.md is updated with the changes (if needed)
Version
Requires a PATCH version update
Context
Users were experiencing an issue where upon checking out the merge ref for a PR the merge ref was referencing an old commit which is incorrect.
Changes
Copied the code from the version 6.2.2 release and kept the changes for GIT_CLONE_COMMIT_COMMITTER_NAME and GIT_CLONE_COMMIT_COMMITTER_EMAIL.
Investigation details
Decisions
This is not a minor version change.
When migrating from 6.x to 7.x the behavior was changed so in some cases you could the explicit depth, and manual_merge was removed. With bringing back the behavior from 6.x, people who have changed their settings for 7.x now have to change them once again. And a lot of people do not check the changelog of minor version changes. Especially one with as little explanation as a cryptic "Revert to 6.2.2 by @vshah23 in #203"
This is not a minor version change. When migrating from 6.x to 7.x the behavior was changed so in some cases you could remove the explicit clone depth, and manual_merge was removed. With bringing back the behavior from 6.x, people who have changed their settings for 7.x now have to change them once again. And a lot of people do not check the changelog of minor version changes. Especially one with as little explanation as a cryptic "Revert to 6.2.2 by @vshah23 in #203"
Agreed, we had clone depth set to -1 as per the previous instructions, and this has now just started failing with fatal: depth -1 is not a positive number
| gharchive/pull-request | 2023-02-06T21:49:13 | 2025-04-01T04:33:38.548727 | {
"authors": [
"ilockett",
"vincentisambart",
"vshah23"
],
"repo": "bitrise-steplib/steps-git-clone",
"url": "https://github.com/bitrise-steplib/steps-git-clone/pull/203",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
916818317 | Update pureconfig to 0.16.0
Updates com.github.pureconfig:pureconfig from 0.12.3 to 0.16.0.
GitHub Release Notes - Changelog - Version Diff
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "com.github.pureconfig", artifactId = "pureconfig" } ]
labels: library-update, semver-minor
Superseded by #300.
| gharchive/pull-request | 2021-06-10T02:14:48 | 2025-04-01T04:33:38.552363 | {
"authors": [
"scala-steward"
],
"repo": "bitrockteam/kafka-dvs-streams",
"url": "https://github.com/bitrockteam/kafka-dvs-streams/pull/275",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1059170977 | compile secp256k1 error.
./build_secp256k1_ios.sh
Cloning into 'secp256k1_ios'...
remote: Enumerating objects: 2615, done.
remote: Total 2615 (delta 0), reused 0 (delta 0), pack-reused 2615
Receiving objects: 100% (2615/2615), 1.10 MiB | 260.00 KiB/s, done.
Resolving deltas: 100% (1865/1865), done.
make: *** No rule to make target distclean'. Stop. building thin libraries... [arm64] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: syntax error near unexpected token ('
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop.
[armv7s] building...
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])'
make: *** No rule to make target install'. Stop. [x86_64] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: syntax error near unexpected token ('
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop.
[i386] building...
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])'
make: *** No rule to make target install'. Stop. [armv7] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: syntax error near unexpected token ('
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop.
building fat binaries...
./build_secp256k1_ios.sh: line 86: cd: /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/btspp_output/thin-libs/arm64/lib: No such file or directory
error: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/lipo: no input files specified
usage: lipo <input_file> [ ...]
command is one of:
-archs
-create [-arch_blank <arch_type>]
-detailed_info
-extract <arch_type> [-extract <arch_type> ...]
-extract_family <arch_type> [-extract_family <arch_type> ...]
-info
-remove <arch_type> [-remove <arch_type> ...]
-replace <arch_type> <file_name> [-replace <arch_type> <file_name> ...]
-thin <arch_type>
-verify_arch <arch_type> ...
options are one or more of:
-arch <arch_type> <input_file>
-hideARM64
-output <output_file>
-segalign <arch_type>
cp: /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_ios/btspp_output/thin-libs/arm64/include: No such file or directory
done. target dir: secp256k1_ios/btspp_output/fat-libs
how to fix it? thank you.
build for android is same error:
./build_secp256k1_android.sh
Cloning into 'secp256k1_android'...
remote: Enumerating objects: 2615, done.
remote: Total 2615 (delta 0), reused 0 (delta 0), pack-reused 2615
Receiving objects: 100% (2615/2615), 1.10 MiB | 180.00 KiB/s, done.
Resolving deltas: 100% (1865/1865), done.
make: *** No rule to make target distclean'. Stop. building all arch libraries... [arm64] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token ('
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop.
[armv7] building...
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])'
make: *** No rule to make target install'. Stop. [x86] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token ('
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop.
[x86_64] building...
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])'
make: *** No rule to make target `install'. Stop.
done. target dir: secp256k1_android/btspp_output
Please copy it to the MAC directory to run
------------------ 原始邮件 ------------------
发件人: "bitshares/bitshares-mobile-app" @.>;
发送时间: 2021年11月20日(星期六) 晚上11:43
@.>;
@.***>;
主题: Re: [bitshares/bitshares-mobile-app] compile secp256k1 error. (Issue #9)
build for android is same error:
./build_secp256k1_android.sh
Cloning into 'secp256k1_android'...
remote: Enumerating objects: 2615, done.
remote: Total 2615 (delta 0), reused 0 (delta 0), pack-reused 2615
Receiving objects: 100% (2615/2615), 1.10 MiB | 180.00 KiB/s, done.
Resolving deltas: 100% (1865/1865), done.
make: *** No rule to make target distclean'. Stop. building all arch libraries... [arm64] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token ('
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop.
[armv7] building...
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])'
make: *** No rule to make target install'. Stop. [x86] building... /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token ('
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])' make: *** No rule to make target install'. Stop.
[x86_64] building...
/Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: syntax error near unexpected token (' /Volumes/T7Touch/github/crypto/bitshares-mobile-app/shell/secp256k1_android/configure: line 1: m4trace:configure.ac:2: -1- AC_INIT([libsecp256k1], [0.1])'
make: *** No rule to make target `install'. Stop.
done. target dir: secp256k1_android/btspp_output
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub, or unsubscribe.
Triage notifications on the go with GitHub Mobile for iOS or Android.
Thank you for your reply. where is MAC directory? MAC is another repo? I use MacOS, only git clone this repo, and compile it with the guide document.
The repository is cloned to run under the macOS path, not in the /Volume xxx directory.
------------------ Original ------------------
From: bitcard @.>
Date: Tue,Nov 23,2021 9:39 AM
To: bitshares/bitshares-mobile-app @.>
Cc: YANG @.>, Comment @.>
Subject: Re: [bitshares/bitshares-mobile-app] compile secp256k1 error. (Issue#9)
Thank you for your reply. where is MAC directory? MAC is another repo? I use MacOS, only git clone this repo, and compile it with the guide document.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub, or unsubscribe.
Triage notifications on the go with GitHub Mobile for iOS or Android.
| gharchive/issue | 2021-11-20T15:41:32 | 2025-04-01T04:33:38.581644 | {
"authors": [
"bitcard",
"hanomirin"
],
"repo": "bitshares/bitshares-mobile-app",
"url": "https://github.com/bitshares/bitshares-mobile-app/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
217932233 | include time to the deadline email notification
As an Applicant I want to see the time of the deadline in the submission deadline email, so I know the exact time my submission is due
https://trello.com/c/PTMmcKOR/555-qae17imp-as-an-applicant-i-want-to-see-the-time-of-the-deadline-in-the-submission-deadline-email-so-i-know-the-exact-time-my-sub
:peter:
| gharchive/pull-request | 2017-03-29T16:24:08 | 2025-04-01T04:33:38.638393 | {
"authors": [
"dreamfall",
"rusllonrails"
],
"repo": "bitzesty/qae",
"url": "https://github.com/bitzesty/qae/pull/1190",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
939475858 | get_analog_in only shows volts?
We would like our unit to read 4...20 mA current but looking at the source code it seems it only displays volts?
even then comparing it to the multi-meter, the AIN1 is receiving 22 volts, and get_analog_in is showing 10.647, how do we read the current coming in at AIN1 and AIN2?
The Monarco HAT can support 4-20mA signals, however this is not supported in this repository at this moment.
You can easily convert 2-10V to a 4-20mA signal by using a 500 Ohm resistor, please read this
Also, you should not connect a voltage higher than 10V to the analog inputs as this could overload and break the hardware.
I ran two 1000Ohm resistors in parallel, to produce a 500Ohm resistor, it worked fine, but the EM fields in the area are messing with the 10vdc signal and introducing noise, this is fine for now, but we would really appreciate 4-20mA signals in the near future
| gharchive/issue | 2021-07-08T05:10:21 | 2025-04-01T04:33:38.647786 | {
"authors": [
"bjarnekvae",
"vendablefall"
],
"repo": "bjarnekvae/pymonarco-hat",
"url": "https://github.com/bjarnekvae/pymonarco-hat/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
701366644 | [Feature request] Alingments: AMR <-> AMR and AMR <-> Sentence
Hi,
first, I want to say: You deserve a medal for creating this library. It is the first time I installed an AMR parser without getting a little headache :-). Also it's a nice idea to wrap the noRECAT variant of GSII and ablate all external java preprocessing. I think the noRECAT version may also be more robust.
I have two suggestions of which I think they would be cool to have in an amrlib:
AMR2sent alignment: As far as I know, there exist aligners (for instance as pre-processing of JAMR parser), that align AMR nodes to tokens. Since often lemmas of the sentence are projected into the AMR graph, a simple string match, maybe with some additional rules, could make up a first solid method. Maybe there are other methods that are more suitable and also easy-to-use.
AMR2AMR variable alignment: This could be useful, e.g., for computing AMR metrics or enabling sentence retrieval via AMR parsed corpora or sentence similarity computation via AMR. It is an NP hard problem but can be implemented via hill climbing maximizing triple matching. I have been working on this lately, here is a repo containing AMR metrics (Smatch and S2match) that are based on this alignment. Both alignments should be quite easy to implement in the lib, since it's all native python. (It could be worthwhile, though, to make the alignment faster, e .g., using cython, since it can be very slow for graphs with many variables)
Alas, these are just suggestions which may or may not be useful to have (in some near or distant future). Again, thanks for your awesome amrlib!
I partially ported the JAMR aligner a while back and I can add this to the lib. I'd like to find the source for the aligner used for the LDC data but so far haven't been able to.
I'll dig into this a bit next weeek as I'm OOO for a while
The integrated aligner would be very helpful!
Question to anyone interested in the alignment functionality...
I have a very simple rule based, "word aligner" from JAMR that I ported to python. This is not the more complicated "span aligner" (ie.. phrases) that is the default alignment method in JAMR. That code is much more complicated (and in scala) so it would be a fair amount of work to port it into amrlib.
Question: Is a simple "word aligner" enough or would this only be useful if we have full span alignments?
There are better span aligners out there (model based) since the early (rule based) JAMR one. I found the code for an ISI aligner (which may have been used to annotate LDC2020T02). That code is basically a bunch of scripts to train/run a model using the C++ MGIZA++ library. It would be a bit of a project, but not completely unreasonable, to make this usable either as part of amrlib, or more likely, as a stand-alone lib.
Question: How important (ie.. how much effort is it worth) to include a span aligner and if you think its important, can you give me the use cases for it?
The reason why I want an aligner in amrlib is that "JAMR" is hard to install and somehow hard to integrate with python scripts.
I have seen several papers using JAMR because it is considered as a light aligner compared to model-based ones. So I think as long as the aligner performance is reasonable and easy to install, people like me would love to use it.
I agree with @lujiaying .
I also think that, from reading the AMR guidelines, there is no "theory of alignment". Therefore, maybe the token-node alignment may even be a bit more clear-cut than node-span.
E.g.
# The frog jumps.
(j / jump-01
:arg0 (f /frog)
With token alignment it's clear that (f,frog) and (j, jumps) are correct alignments, however both (f, the frog) and (f, frog) can be considered correct in span alignment.
On the other hand, maybe span has advantages when concepts are really abstract
In sum, I think that a lightweight node-token alignment that performs reasonable would be a very good start, since it anchors the AMR in the sentence, which may already be quite helpful for some tasks.
I have updated the master project with Rule Based Word aligner, similar in function to JAMR's word aligner. If you have comments or bugs, please post a new issue. (update to pip install to follow).
| gharchive/issue | 2020-09-14T19:28:54 | 2025-04-01T04:33:38.655776 | {
"authors": [
"bjascob",
"flipz357",
"lujiaying"
],
"repo": "bjascob/amrlib",
"url": "https://github.com/bjascob/amrlib/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1582852933 | Simple way to disable AuthProvider in development environment?
Is there a simple way to render the AuthProvider non-functional (but present) if process.env.NODE_ENV === "development"? Additionally, since useAuth() is a hook, it can't be conditionally called based on whether the environment is development, so I can't just omit AuthProvider in dev and not call useAuth()
Hm, what's the use-case for this? Simplify frontend development?
Yeah, to skip the login requirement during development. I've achieved this now with a combination of:
If environment is dev, set autoSignIn to false
If environment is dev, return app directly instead of a login prompt in App.tsx
| gharchive/issue | 2023-02-13T18:28:44 | 2025-04-01T04:33:38.660232 | {
"authors": [
"cobraz",
"denniskrq"
],
"repo": "bjerkio/oidc-react",
"url": "https://github.com/bjerkio/oidc-react/issues/941",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
183187895 | Got examples to work a bit more obviously
Added npm scripts:
npm run ex:svg
npm run ex:group
Great, thanks. Published as 2.4.2
| gharchive/pull-request | 2016-10-15T04:22:00 | 2025-04-01T04:33:38.662001 | {
"authors": [
"bjnortier",
"skawaguchi"
],
"repo": "bjnortier/dxf",
"url": "https://github.com/bjnortier/dxf/pull/14",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1633616029 | Update to Cranelift 0.94
Fixes https://github.com/bjorn3/rustc_codegen_cranelift/issues/1249 through https://github.com/bytecodealliance/wasmtime/pull/5619
Fixes https://github.com/bjorn3/rustc_codegen_cranelift/issues/1349 through https://github.com/gimli-rs/object/pull/475.
cc @PROMETHIA-27 and @frederikhors
| gharchive/pull-request | 2023-03-21T10:29:38 | 2025-04-01T04:33:38.663984 | {
"authors": [
"bjorn3"
],
"repo": "bjorn3/rustc_codegen_cranelift",
"url": "https://github.com/bjorn3/rustc_codegen_cranelift/pull/1363",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
236169276 | Offscreen rendering with MSAA
How can I draw in a color texture of a framebuffer with MSAA enabled?
bgfx::reset(width, height, BGFX_RESET_MSAA_X16);
// ....
bgfx::TextureHandle m_gbufferTex[2];
m_gbufferTex[0] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::BGRA8, BGFX_TEXTURE_RT | BGFX_TEXTURE_U_CLAMP | BGFX_TEXTURE_V_CLAMP);
m_gbufferTex[1] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::D24, BGFX_TEXTURE_RT_WRITE_ONLY);
bgfx::FrameBufferHandle m_gbuffer = bgfx::createFrameBuffer(BX_COUNTOF(m_gbufferTex), m_gbufferTex, true);
// ....
bgfx::readTexture(bgfx::getTexture(m_gbuffer, 0), mem);
Rendering into a window works perfectly well with MSAA, but with this code my render result is not antialiased. Here are the debug infos of the cubes example: http://imgur.com/a/csdRo
See example-09-hdr it uses MSAA... Toggle it with F8: https://github.com/bkaradzic/bgfx/blob/master/examples/09-hdr/hdr.cpp#L285
((msaa + 1) << BGFX_TEXTURE_RT_MSAA_SHIFT)
If I add this flag to my framebuffer textures, nothing gets rendered at all. I modified my code to match the example, but no progress. I'm also rendering only simple geometry with very basic shaders.
Do you have any idea what the problem could be? Maybe reading directly from the framebuffers color texture? In the example the texture is used in other shader and at the and directly drawn to the backbuffer.
So I made a minimalistic example, rendering a cube:
int main(int argc, char** argv)
{
uint32_t width = 1024;
uint32_t height = 1024;
SDL_Window* window = SDL_CreateWindow("bgfx_ros", SDL_WINDOWPOS_UNDEFINED,
SDL_WINDOWPOS_UNDEFINED, width, height,
SDL_WINDOW_RESIZABLE);
SDL_SysWMinfo wmi;
SDL_VERSION(&wmi.version);
if (!SDL_GetWindowWMInfo(window, &wmi) )
{
std::cerr << "couldn't get wm info" << std::endl;
return 1;
}
bgfx::PlatformData pd;
pd.ndt = wmi.info.x11.display;
pd.nwh = reinterpret_cast<void*>((uintptr_t)wmi.info.x11.window);
pd.context = NULL;
pd.backBuffer = NULL;
pd.backBufferDS = NULL;
bgfx::setPlatformData(pd);
if (!bgfx::init(bgfx::RendererType::Count, BGFX_PCI_ID_NONE))
{
std::cerr << "bgfx init failed" << std::endl;
return -1;
}
bgfx::reset(width, height, BGFX_RESET_MSAA_X16);
bgfx::setViewClear(0, BGFX_CLEAR_COLOR | BGFX_CLEAR_DEPTH, 0x00000000, 1.0f, 0);
std::vector<uint8_t> vresult;
bgfx::ShaderHandle vhandle;
createShaderFromFile("vs_object_instanced.bin", vresult, vhandle);
std::vector<uint8_t> fresult;
bgfx::ShaderHandle fhandle;
createShaderFromFile("fs_object.bin", fresult, fhandle);
bgfx::ProgramHandle program;
const bool destroy_shader = true;
program = bgfx::createProgram(vhandle, fhandle, destroy_shader);
if (!isValid(program))
{
std::cerr << "creating shader program failed" << std::endl;
return 1;
}
// Create vertex stream declaration.
PosVertex::init();
std::vector<PosVertex> vertices;
std::vector<uint16_t> indices;
loadOBJ("/home/david/Desktop/cube.obj", vertices, indices);
bgfx::VertexBufferHandle vbh = bgfx::createVertexBuffer(bgfx::makeRef(static_cast<void*>(vertices.data()), vertices.size() * sizeof(PosVertex)), PosVertex::ms_decl);
bgfx::IndexBufferHandle ibh = bgfx::createIndexBuffer(bgfx::makeRef(static_cast<void*>(indices.data()), indices.size() * sizeof(uint16_t)));
**uint32_t msaa = (BGFX_RESET_MSAA_X16&BGFX_RESET_MSAA_MASK)>>BGFX_RESET_MSAA_SHIFT;
bgfx::TextureHandle m_gbufferTex[2];
// does not work - MSAA
m_gbufferTex[0] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::BGRA8, ((msaa + 1) << BGFX_TEXTURE_RT_MSAA_SHIFT) | BGFX_TEXTURE_U_CLAMP | BGFX_TEXTURE_V_CLAMP);
m_gbufferTex[1] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::D24, ((msaa + 1) << BGFX_TEXTURE_RT_MSAA_SHIFT) | BGFX_TEXTURE_RT_WRITE_ONLY );
// WORKS - no MSAA
//m_gbufferTex[0] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::BGRA8, BGFX_TEXTURE_U_CLAMP | BGFX_TEXTURE_V_CLAMP);
//m_gbufferTex[1] = bgfx::createTexture2D(uint16_t(width), uint16_t(height), false, 1, bgfx::TextureFormat::D24, BGFX_TEXTURE_RT_WRITE_ONLY );**
bgfx::FrameBufferHandle m_gbuffer = bgfx::createFrameBuffer(BX_COUNTOF(m_gbufferTex), m_gbufferTex, true);
float eye[3] = { 2.0f, 2.0f, -5.0f };
float at[3] = { 0.0f, 0.0f, 0.0f };
float view[16];
bx::mtxLookAtRh(view, eye, at);
float proj[16];;
bx::mtxProjRh(proj, 60.f, 1.f, 0.1f, 100.0f, true);
bgfx::setViewFrameBuffer(0, m_gbuffer);
bgfx::setViewTransform(0, view, proj);
bgfx::setViewRect(0, 0, 0, uint16_t(width), uint16_t(height));
if (!bgfx::isValid(m_gbuffer) ) {
throw std::runtime_error("framebuffer not valid!");
}
// This dummy draw call is here to make sure that view 0 is cleared
// if no other draw calls are submitted to view 0.
bgfx::touch(0);
// Set vertex and index buffer.
bgfx::setVertexBuffer(vbh);
bgfx::setIndexBuffer(ibh);
// Set render states and submit
bgfx::setState(BGFX_STATE_RGB_WRITE|BGFX_STATE_ALPHA_WRITE);
bgfx::submit(0, program);
void* mem = malloc(width*height*4);
bgfx::readTexture(m_gbufferTex[0], mem);
bgfx::frame();
bgfx::destroyIndexBuffer(ibh);
bgfx::destroyVertexBuffer(vbh);
bgfx::destroyProgram(program);
bgfx::destroyFrameBuffer(m_gbuffer);
bgfx::destroyTexture(m_gbufferTex[0]);
bgfx::destroyTexture(m_gbufferTex[1]);
bgfx::shutdown();
// get texture
cv::Mat renderResult(height, width, CV_8UC4, mem);
// flip it (correction, since the opengl stores the data differently)
cv::flip(renderResult, renderResult, 0);
cv::imshow("testim", renderResult);
cv::waitKey(0);
return 0;
}
This is the rendering result without MSAA: http://imgur.com/a/2D0H8
What am I doing wrong? :(
bgfx::setState(BGFX_STATE_RGB_WRITE|BGFX_STATE_ALPHA_WRITE); add |BGFX_STATE_MSAA
Just tested it and no success. But thanks for looking at it!
Textures are missing BGFX_TEXTURE_RT
This is wrong:
uint32_t msaa = (BGFX_RESET_MSAA_X16&BGFX_RESET_MSAA_MASK)>>BGFX_RESET_MSAA_SHIFT;
It should be 0 for 2x, 1 for 4x etc. or just use BGFX_TEXTURE_RT_MSAA_X16 directly instead of ((msaa + 1) << BGFX_TEXTURE_RT_MSAA_SHIFT).
Thanks for the hint, but adding BGFX_TEXTURE_RT_MSAA_X16 to the color texture or the depth texture (or both) yields to a black screen. So, the same result as the mess with MSAA_MASK+MSAA_SHIFT (which I took from the hdr example)...
I want to do rendering on a server application (with msaa) and read the result in CPU memory for further processing. Maybe I can get past this problem by just reading the render result from the SDL window. This seems not very elegant to me, but MSAA works when I do on-screen rendering.
.
Try copying what the hdr example does with bgfx::readTexture. You create a separate texture with BGFX_TEXTURE_READ_BACK flag and blit the multisampled texture into that.
Also, you should call bgfx::frame until it returns a frame number >= the return value from bgfx::readTexture. See the documentation on bgfx::readTexture.
Thank you so much! The trick is to blit the color texture of the framebuffer onto another texture, as you suggested! I assumed I could read the fbo color texture directly, because it worked without MSAA.
That was a long struggle. Thanks for your help!!
You should use SAMPLER2DMS if you want to read MSAA directly.
| gharchive/issue | 2017-06-15T12:05:38 | 2025-04-01T04:33:38.678955 | {
"authors": [
"bkaradzic",
"jpcy",
"tuxa"
],
"repo": "bkaradzic/bgfx",
"url": "https://github.com/bkaradzic/bgfx/issues/1159",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
278717166 | CMake as an official build option
A very large number, if not most C++ projects are based on the CMake build system. It would be a huge benefit to these projects if bgfx, in addition to GENie, also officially supported the CMake build system.
@JoshuaBrookover has a fork of bgfx , (https://github.com/JoshuaBrookover/bgfx.cmake), that adds CMake. Would it be any way possible to incorporate Joshua's CMake system into the official bgfx repository?
The addition of a supported CMake build would greatly increase the target audience of bgfx. I know of a number of developers who are hesitant to try bgfx because their projects are based are CMake, and using a forked version instead of the official bgfx is less likely to remain stable in the long term.
I'm in no way suggesting removing or changing GENie, just offering a official option of a CMake build in addition to the GENie build.
I would be happy to help to maintain the CMake build.
Why not just use compiled version?
To my knowledge, there are no bgfx compiled binary packages for RHEL or OSX (2 of our primary dev platforms). We support Windows, Mac, RHEL and Ubuntu. We currently build all our 3rd party dependencies, because traditionally it's been simpler to include them as git submodules, and include them as part of the build as opposed to trying to script a variety of different package managers.
Yes, I understand where your use case is coming from, and I think it's a completely valid use case for some libraries. Though in this case bgfx has a lot of 3rdparty deps and generate not only a library but also some binaries, so some folks here (including me) just went with building bgfx and including lib/bin files + headers in the main repository, like here for example https://github.com/beardsvibe/leengine/tree/master/3rdparty/bgfx
This approach removes a lot of headache trying to fight build system on every platform every time I build my project, instead I just need to do it once when I'm building bgfx (which is once in a while :) ).
I'm all for an officially maintained CMake build system. Mine is far from perfect. I think @bkaradzic doesn't want to deal with it though. It's more likely GENie will automatically spit out CMake scripts for you eventually. I'm not too familiar with how GENie works though.
I worked at a company where we used CMake (hence the need for bgfx.cmake). Building bgfx from source each time was very beneficial to us as it allowed us to modify the source code and debug problems with minimum hassle. That's really where my project is most useful.
I don't want maintain CMake configs at all. I looked into CMake 10 years ago before I start using Premake and I didn't like it. Fact that it's popular doesn't mean I want to use it. But I'm fine if there is someone who wants to do that in external repo.
The addition of a supported CMake build would greatly increase the target audience of bgfx.
I know of a number of developers who are hesitant to try bgfx because their projects are based
are CMake, and using a forked version instead of the official bgfx is less likely to remain stable
in the long term.
I'm totally fine with that. If someone chooses to write their own renderer just based on project generator (GENie is not build system, rather it's project generator) used by bgfx, that means that bgfx doesn't really offer them anything they need.
Btw, GENie has CMake project generator, but it's minimal required for CLion to work with it:
https://github.com/bkaradzic/GENie/tree/master/src/actions/cmake
You can contribute there if you want bgfx to support CMake, that's the only way to be in main repo... ;)
At least you're not ashamed to admit that your opinion of CMake is based on an experience from 10 years ago. A LOT has changed since then.
A LOT has changed since then.
Not my opinion about it.
https://github.com/widberg/bgfx.cmake is valid?
@Yairama Ask there...
| gharchive/issue | 2017-12-02T20:59:40 | 2025-04-01T04:33:38.687905 | {
"authors": [
"AndySomogyi",
"JoshuaBrookover",
"Yairama",
"bkaradzic",
"jimon",
"minexew"
],
"repo": "bkaradzic/bgfx",
"url": "https://github.com/bkaradzic/bgfx/issues/1284",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
110376998 | Variables not being interpolated when using Procfile
Hello so I have a .env with:
MONGO_PORT=17017
TEST=${MONGO_PORT}
MONGOHQ_URL="mongodb://localhost:${MONGO_PORT}/database"
and a script:
#!/usr/bin/env ruby
require File.expand_path('../../config/environment', __FILE__)
puts Rails.env
puts ENV['TEST']
puts ENV['MONGOHQ_URL']
If I run it with bundle exec ./bin/test, the output is correct
development
17017
mongodb://localhost:17017/database
but if I add to a Procfile test: bundle exec ./bin/test and run foreman start test then the variable is not interpolated:
14:32:51 test.1 | started with pid 17789
14:32:57 test.1 | development
14:32:57 test.1 | ${MONGO_PORT}
14:32:57 test.1 | mongodb://localhost:${MONGO_PORT}/database
14:32:57 test.1 | exited with code 0
14:32:57 system | sending SIGTERM to all processes
This might not be a dotenv issue, but I'm kinda lost on how to debug this, any place I should be looking at?
Also I should add that the .env is being loaded just not interpolating the variables.
So after debugging inside dotenv I realised that the parsing and substitutions were happening.
The issue here is that foreman will also load a .env file if it finds it, and it seems dotenv give priority to already defined env variables, which is ok.
Here the fix is to move those variables into a .env.local or .env.<environment> file.
| gharchive/issue | 2015-10-08T05:34:22 | 2025-04-01T04:33:38.691865 | {
"authors": [
"GriffinHeart"
],
"repo": "bkeepers/dotenv",
"url": "https://github.com/bkeepers/dotenv/issues/218",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
90470627 | jobs processed even when queue is not running
I have an app with Rails 4.2.2 (using mongoid) running qu from master branch, and I'm seeing this strange behavior. On my model_obserer I have this code
def after_save(message)
room = "room_#{message.room.uid}"
SendMessageJob.create(room, message.id)
end
I started my server, but didn't ran the command bundle exec rake qu:work but the messages were still delivered to the rooms...
This is unexpected, is there something wrong?
Can you provide a small rails app/example that shows the issue? It is really hard to say based on the tiny bit of code you have provided.
well, there is not much than that, actually. I installed qu and qu-mongoid, moved the method responsible to send the message to a Job and enqueued on after_save...that's it
I still having this issue. Here is the whole process:
message_controller
def create
message = Message.new(message_params)
message.room = room
message.user = current_user
message.save
MetricServices.message_sent(current_user.uid, room: message.room.name)
render nothing: true
end
message_observer
def after_save(message)
room = "room_#{message.room.uid}"
SendMessageJob.create message.id
end
send_message_job
class SendMessageJob < Qu::Job
def initialize(message_id)
@message_id = message_id
end
def perform
message = Message.find @message_id
room = "room_#{message.room.uid}"
WebsocketRails[room].trigger :new_message, message.present.to_html
end
end
I'm starting the app with rails server and JUST THIS...the message is still processed.
Also, I didn't saw anything on my mongo database. is this right? should I see something, like a queue collection or whatever?
anything?
A couple things. I'm not 100% positive that master is stable right now. We've been tweaking things for a while to make some big breaking API changes for the long term good. It should be stable, but if it is, the API will change soon, so I wouldn't pin to master. I'd pin to the latest release and upgrade later.
Second, where are you configuring qu? It seems like the qu-immediate backend is being used or something.
I'll try to change to the latest stable and actually, I haven't any config file for qu...only the bundle exec rake qu:work on the Procfile but it is commented, so it is not executing
| gharchive/issue | 2015-06-23T19:27:18 | 2025-04-01T04:33:38.697042 | {
"authors": [
"jnunemaker",
"luizkowalski"
],
"repo": "bkeepers/qu",
"url": "https://github.com/bkeepers/qu/issues/102",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
765609363 | [Gally]: master <- dev
Automatically created by Git-Ally
:tada: This PR is included in version 1.10.202 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
:tada: This PR is included in version 1.10.202 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
| gharchive/pull-request | 2020-12-13T19:24:03 | 2025-04-01T04:33:38.735520 | {
"authors": [
"MrsFlux"
],
"repo": "blackflux/lambda-example",
"url": "https://github.com/blackflux/lambda-example/pull/1627",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
322856571 | Purchases API
Hey guys, I noticed there's no implementation for the purchases API (https://dev.recurly.com/docs/create-purchase).
Any plans on adding the support for it?
I most likely won't get to it any time soon.
@MOZGIII This was just merged into master.
| gharchive/issue | 2018-05-14T15:02:51 | 2025-04-01T04:33:38.746177 | {
"authors": [
"MOZGIII",
"cristiangraz"
],
"repo": "blacklightcms/recurly",
"url": "https://github.com/blacklightcms/recurly/issues/50",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2764185890 | Miningcore Solo Mining pool based on rmt-miningcore - Pool wallet to miners wallet tranfer fails
I have setup miningcore solo mining pool on docker container that is working fine. Though facing following issue:
As per above image pending shares transferred to pending balance but these are not transferring to PAID balance with following error.
[2024-12-31 03:45:02.9038] [I] [khaaaicpu] [Bitcoin Payout Handler] Paying 110 CPU to 1 addresses
[2024-12-31 03:45:03.1267] [E] [khaaaicpu] [Bitcoin Payout Handler] sendmany returned error: Fee estimation failed. Fallbackfee is disabled. Wait a few blocks or enable -fallbackfee. code -6
Any advise how what to change in config.json file.
thanks
Hi there,
We use GitHub issues as a place to track bugs and other development-related issues.
Please see the link below to our dedicated support line:
Help Center : Click Here
Ticket ID: WR240
Note: Click on the live chat icon at the bottom corner of the page to start a conversation.
@KHPak2023 Dear user
I understand that you have been experiencing some issues and I want to ensure you get the support you need as quickly as possible by using our live chat option.
Support : Support
1. Access the Live Chat:
• Navigate to our website’s home page.
• Look for the chat icon located at the bottom right corner of the screen and start a 💬.
Hello,
I have no idea which coin you are trying to mine here, it would have been a good idea to at least provide its name.
The error which is returned by the node, seems to actually offer a good options, run your node with the option -fallbackfee
Hello @KHPak2023 A support ticket has been opened for you. Kindly Talk to Support on the Live support page so that your issue can be looked into. CLICK HERE
Hello,
I have no idea which coin you are trying to mine here, it would have been a good idea to at least provide its name.
The error which is returned by the node, seems to actually offer a good option, run your node with the option -fallbackfee
Thank you very much for response. The coin is cpuchain
Thanks it is done by adding fallback=0.00001 to coin.conf file.
Thanks for prompt response and help
| gharchive/issue | 2024-12-31T09:55:58 | 2025-04-01T04:33:38.758029 | {
"authors": [
"Justin-skies",
"KHPak2023",
"MatthewDEVo",
"blackmennewstyle",
"mark8700-bit"
],
"repo": "blackmennewstyle/miningcore",
"url": "https://github.com/blackmennewstyle/miningcore/issues/205",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1481587176 | Quick edit YAML properties from dataview table (like in database folder)
Is your feature request related to a problem? Please describe.
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
I'm always frustrated when I create a dataview table & need to visit each individual file & use metaedit to change YAML properties.
Describe the solution you'd like
A clear and concise description of what you want to happen.
Ability to change YAML properties for files in dataview table in a similar fashion to that in the database folder plugin i.e. with an assigned & appropriate data type (dropdowns, date & time, etc).
Describe alternatives you've considered
A clear and concise description of any alternative solutions or features you've considered.
Using database folder - issue is that it can't be embedded properly into other docs.
Using metaedit plugin - issue is that it still requires navigating away from the file to change the frontmatter & also does not support ease of use data types with drop downs, dates & times etc.
Hello,
this feature request is already created multiple times, like in #749 or #116. Closing this as duplicate in favor of the existing ones.
| gharchive/issue | 2022-12-07T10:35:31 | 2025-04-01T04:33:38.762364 | {
"authors": [
"danrfletcher",
"s-blu"
],
"repo": "blacksmithgu/obsidian-dataview",
"url": "https://github.com/blacksmithgu/obsidian-dataview/issues/1644",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1110383347 | Typo: Task instead of Calendar in a heading
Please provide a link to the documentation page and section
https://blacksmithgu.github.io/obsidian-dataview/query/queries/?query=title#task-queries_1
Describe the problem
The says "Task Query", even though is should be "Calendar Query"
Thanks! Fixed for 0.4.23.
| gharchive/issue | 2022-01-21T11:52:07 | 2025-04-01T04:33:38.764477 | {
"authors": [
"blacksmithgu",
"chrisgrieser"
],
"repo": "blacksmithgu/obsidian-dataview",
"url": "https://github.com/blacksmithgu/obsidian-dataview/issues/799",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1228716821 | 🛑 Discord Akary Bot is down
In db19567, Discord Akary Bot (https://discord-bot-akary.blackveriun.repl.co) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Discord Akary Bot is back up in 36f8c16.
| gharchive/issue | 2022-05-07T20:31:24 | 2025-04-01T04:33:38.771255 | {
"authors": [
"blackveriun"
],
"repo": "blackveriun/status",
"url": "https://github.com/blackveriun/status/issues/135",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1931807108 | 🛑 WSMM Bitbucket is down
In fe34e52, WSMM Bitbucket (https://scm.wsmm.de) was down:
HTTP code: 503
Response time: 447 ms
Resolved: WSMM Bitbucket is back up in cdfdf0c after 10 minutes.
| gharchive/issue | 2023-10-08T13:31:49 | 2025-04-01T04:33:38.773543 | {
"authors": [
"blacky1707"
],
"repo": "blacky1707/uptime",
"url": "https://github.com/blacky1707/uptime/issues/127",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1334184486 | Revert "Restore JSON decode for ws_api (#329)"
Related card PR: https://github.com/dermotduffy/frigate-hass-card/pull/796
Thanks @uvjustin !
| gharchive/pull-request | 2022-08-10T07:14:44 | 2025-04-01T04:33:38.779543 | {
"authors": [
"dermotduffy",
"uvjustin"
],
"repo": "blakeblackshear/frigate-hass-integration",
"url": "https://github.com/blakeblackshear/frigate-hass-integration/pull/330",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
356346797 | JS Modules export?
Hi @blakeembrey!
This module is really popular! It's also nice and small, but I was wondering if you'd be open to providing a JS Modules version in addition to CommonJS? I'd be happy to PR. Let me know if you have a preference for tooling/etc.
@developit Happy to accept a PR 👍
Added with https://github.com/blakeembrey/array-flatten/releases/tag/v3.0.0.
@blakeembrey This breaks existing usage code for the old CommonJS pattern, just as an FYI.
Where var flatten = require('array-flatten'); used to return the function directly, it now returns an object with the function defined as the object's flatten property. Existing calls to the flatten function directly need to be changed to call the method on the returned object instead. Otherwise seeing "Warning: flatten is not a function" error.
Not sure if there is a way to accommodate both usage patterns, or if packages using the old version need to either update or manually specify the older version of this package as a dependency, but I think a lot of people will have builds that break due to this. (grunt-assemble is affected)
I know, that’s why this was released as V3. If it affects anything, they’re using dependencies incorrectly.
I also don’t see this used in grunt assemble either.
@blakeembrey Thanks for taking the time to reply to my comment. I tracked the issue to this package from grunt-assemble > resolve-dep > arrayify-compact > array-flatten
Admittedly some of the versions of those packages that were in use on my affected project are 4+ years old and it's likely that there are issues with dependency management somewhere in that tree. I was able to fix the issue on my end by adding version 2.1.2 of your module as a dependency in my package.json, even though I'm not using it directly in my build.
arrayify-compact lists version 1.0.1 as a dependency but appears to have been pulling in the latest version instead.
Didn't mean to imply there was an issue on your end, just wanted to point that out in case anyone else came across the same problem. Thanks again!
| gharchive/issue | 2018-09-03T03:25:22 | 2025-04-01T04:33:38.829787 | {
"authors": [
"Lohdro",
"blakeembrey",
"developit"
],
"repo": "blakeembrey/array-flatten",
"url": "https://github.com/blakeembrey/array-flatten/issues/19",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
243085814 | clear ng2-datepicker after form submit
I want to clear the date field after form submit. I tried this.expire_date = null; where my model is expire_date. But it doesn't work. How can I do it? Obviously I don't want to manually click the clear button in the ui.
This will solve your problem.
delete this.expire_date['formatted'];
tried delete this.expire_date['formatted'];
and it does not work.
this works for me!
<ng2-datepicker [options]="options" [inputEvents]="validFormdatepickerInputEvents" formControlName="valid_from" [(ngModel)]="package.valid_from" class="text_area">
this.package.valid_from = new DateModel();
if binded to [(ngModel)] you reset it just like any other variable in the form.
You can easily clear selected date using ElementRef:
<ng-datepicker [(ngModel)]="empty_date" [options]="options1" #date1></ng-datepicker>
<button (click)="date1.displayValue=''">Clear Date</button>
OR
You can clear it from component also:
@ViewChild('date1') date1: ElementRef;
resetDate() {
this.date1['displayValue']='';
}
| gharchive/issue | 2017-07-14T19:10:41 | 2025-04-01T04:33:38.862022 | {
"authors": [
"ascuet",
"jkuri",
"kuldeeps1ngh",
"nilasissen",
"vaibhavpatil111",
"zjx"
],
"repo": "bleenco/ng2-datepicker",
"url": "https://github.com/bleenco/ng2-datepicker/issues/244",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1795512305 | Endpoints for LocalStateQuery via REST
The LocalStateQuery Ouroboros protocol has many client endpoints.
GetCurrentEra
GetSystemStart
GetChainBlockNo
GetChainPoint
GetEraHistory
GetEpochNo
GetCurrentProtocolParams
GetGenesisConfig
This depends on blinklabs-io/gouroboros#315 for underlying support for the query communication with the Cardano Node.
Merging ChainBlockNo, EpochNo, and ChainPoint into a single Tip to more closely match other tools and reduce the necessary API calls.
The first part of this is in #134 and the second part will be in another pull request.
The second part is in #143 with a caveat of a bug in gOuroboros around the Genesis Config.
| gharchive/issue | 2023-07-09T18:30:20 | 2025-04-01T04:33:38.877181 | {
"authors": [
"musik-c",
"wolf31o2"
],
"repo": "blinklabs-io/cardano-node-api",
"url": "https://github.com/blinklabs-io/cardano-node-api/issues/25",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
365159246 | Don't prompt for React Router if exists.
When using blix add the user is prompted with do you want to add React Router even if they already have react-router in their project.
Recommended solution - Do a quick check for react-router in the user's package.json prior to running the script.
I'll also go after the empty App folder bug
| gharchive/issue | 2018-09-29T19:13:02 | 2025-04-01T04:33:38.888397 | {
"authors": [
"DBULL7",
"alocke12992"
],
"repo": "blixjs/blix",
"url": "https://github.com/blixjs/blix/issues/338",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2628396201 | Fix validator functions
https://sepolia.etherscan.io/address/0xaa186e7c5483ed18cEd7C23eA3c60b3BC929F4d6#code
No sepolia! Straight base
| gharchive/pull-request | 2024-11-01T06:03:29 | 2025-04-01T04:33:38.891411 | {
"authors": [
"Mitchell139",
"bitcoinbrisbane"
],
"repo": "block52/poker-vm",
"url": "https://github.com/block52/poker-vm/pull/20",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
371246139 | Hide legacy proposals #1986
Closes #23
.I added new method hideProposals() for checking and hiding legacy proposals,
also I added checkbox and hideLegacyProposals state
@sschiessl-bcp Please review it , when you have time.
Please reference the corresponding issue in the PR
https://github.com/bitshares/bitshares-ui/issues/1986
I think the recognition is off.
Please adjust:
[ ] Proposals with almost same name and same payout account.
[ ] if both are not active, hide older one
[ ] if one is active, hide other one. Example: 1.14.74 and 1.14.73.
[ ] Add a manually defined list of ids that are to be hidden
[ ] If the start date of the proposal is more than one month in the past, and less than 10 million votes
if one is active, hide other one. Example: 1.14.74 and 1.14.73.
I am not sure that understood correctly. Should we display some active proposals when hide proposals is enabled?
if one is active, hide other one. Example: 1.14.74 and 1.14.73.
Consider the header of the proposal tab
Listed under "Proposed": Inactive
Listed under "Active": Active
I adjusted filter according to your`s comments. Please review it when you have time.
| gharchive/pull-request | 2018-10-17T20:07:01 | 2025-04-01T04:33:38.896724 | {
"authors": [
"VerevkinAlexander",
"sschiessl-bcp"
],
"repo": "blockchainprojects/bitshares-ui",
"url": "https://github.com/blockchainprojects/bitshares-ui/pull/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
200177893 | Fix output for insufficient funds to actually output the needed amount
The output prints the current balance instead of the fee
Looks good. Thanks for catching this!
| gharchive/pull-request | 2017-01-11T18:59:05 | 2025-04-01T04:33:38.940459 | {
"authors": [
"jcnelson",
"vbrandl"
],
"repo": "blockstack/blockstack-cli",
"url": "https://github.com/blockstack/blockstack-cli/pull/186",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
776892210 | Broken on recent versions of discourse
This plugin is broken in recent versions (true as of 2.5.6) of discourse.
The Stacks Foundation is looking for help with updating the plugin. (https://github.com/stacksgov/Stacks-Grants/issues/29)
Related issues/info:
https://github.com/blockstack/blockstack.org/issues/392
https://github.com/blockstack/blockstack.org/issues/1159
https://github.com/blockstack/omniauth-blockstack/blob/master/lib/omniauth/strategies/blockstack.rb#L39-L42
I spent a couple hours taking a look at this today.
@wileyj writes:
Pretty sure the issue is here:
https://github.com/blockstack/omniauth-blockstack/blob/master/lib/omniauth/strategies/blockstack.rb#L39-L42
One issue is here. What's supposed to happen is when /auth/blockstack?manifest=true is fetched by the authenticator, the manifest json is returned. It appears discourse changed behavior such that it instead returns an HTML page with a button the user has to click on before the JSON is returned by the omniauth strategy. I imagine this has something to do with the discourse folks trying to address cross-site scripting, but it breaks our authentication flow.
One short-term way to work around this to put the discourse installation behind a proxy and intercept requests to /auth/blockstack?manifest=true and return the manifest json.
A longer term solution would be really digging into the internals of discourse to figure a better way to do this - there might not be away to do it in a supported fashion.
A second issue was reported by @tvr132b here: https://github.com/blockstack/blockstack.org/issues/1159#issue-687503148
This issue is with the content security policy added by discourse. You need to disable to included content security policy on discourse or add the unsafe-inline keyword (unclear how to do this in on discourse in a supported way).
hmm, the proxy idea might be the path of least resistance here.
I think we can give it a shot, and see if it works
hmm, the proxy idea might be the path of least resistance here.
I think we can give it a shot, and see if it works
Cool! Let me know how it works!
Cool! Let me know how it works!
| gharchive/issue | 2020-12-31T09:32:20 | 2025-04-01T04:33:38.947835 | {
"authors": [
"larrysalibra",
"wileyj"
],
"repo": "blockstack/discourse-blockstack",
"url": "https://github.com/blockstack/discourse-blockstack/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
766373447 | How To Generate Tx And Sign offline?
Hello, how can I generate a transaction body and sign it offline? I want to complete a simple function of transferring STX. The content in the document is generated on the client side. I want to try the offline method. I see a similar method called makeUnsignedSTXTokenTransfer, but it is not very suitable . What I want is to first generate an unsigned transaction body, sign it by the sender, then serialize the signed transaction body, and broadcast the serialized result. Is there such a way? I'm not sure, looking forward to your reply, thanks
@Wangmmx Please use the forum.stacks.org or discord for these open questions and close this issue.
If the documentation at https://docs.blockstack.org/understand-stacks/transactions#serialization is not sufficient you can open an issue on github.com/blockstacks/docs
You don't understand what I mean. I need to confirm to the developer instead of the forum. There are many types of transaction bodies in the document. I want to know if there is a way to assemble transaction bodies and sign them offline. It is a simple one-signature account transfer: https://docs.blockstack.org/understand-stacks/transactions#stacks-token-transfer, here is to put the senderKey in the transaction body to build, is there any first to build How about the unsigned transaction body, senderKey to sign separately?
@Wangmmx makeUnsignedSTXTokenTransfer does exactly what you want, creating a transaction without the private key. What makes it not very suitable?
@Wangmmx makeUnsignedSTXTokenTransfer does exactly what you want, creating a transaction without the private key. What makes it not very suitable?
Because this method also uses generator to construct the transaction body, it returns a promise. I want to use a more normal method instead of promise. I have basically solved this problem and constructed a StackTransaction class to complete it. By the way, if you use the way you said, how to put the signature into the transaction body? In other words, what is the signature comes from? thank you for your reply
Can you explain why it's not possible to use a promise for your use case? If the nonce and fee are supplied as inputs to the transaction builder, it does not need to be async. Once you have an unsigned transaction, you can sign it using the TransactionSigner class.
It's my own problem. I try to avoid using asynchronous methods in my projects. If I quote directly, I need await or handle promise.
I have successfully constructed the transaction body and signed it, thank you for your reply, I responded too late.
I would like to ask how to broadcast transactions? I want to use the rpc interface /v2/transactions to broadcast. The rawTx parameter is the serialized transaction body, which looks like 8080000000..., but it keeps reporting errors: {Error: "Failed to deserialize posted transaction: Failed to parse transaction authorization : unrecognized auth flags 101"}, I used the api interface broadcastRawTransaction to be successful, may I ask why?
Your request body is probably not formatted correctly. Take a look at broadcastTransaction in @stacks/transactions
Okay, I made it, thank you very much! I forgot to add the url to the request, thanks again!
The successful broadcast transaction was not found on the browser. I found that the last broadcast was successful 10 days ago. Is the testnet unavailable now?
There were some issues with Testnet in the last 24 hours. It should be resolved now.
| gharchive/issue | 2020-12-14T12:24:11 | 2025-04-01T04:33:38.956836 | {
"authors": [
"Wangmmx",
"friedger",
"yknl"
],
"repo": "blockstack/stacks.js",
"url": "https://github.com/blockstack/stacks.js/issues/884",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
708897829 | Provide option for developers to require browser extension for app authentication
During a conversation with @radicleart, he expressed interest in requiring that all users of his apps authenticate with the Blockstack browser extension instead of the hosted authenticator.
This would ensure that those users get habituated during both onboarding and sign in to using an installation-based authenticator instead of a dynamically loaded, centrally hosted one, resolving concerns that the latter's use could lead to greater phishing or hacking exposure as discussed on the forum.
Providing an option for developers to require the browser extension for authentication in their apps would equip them to make their own choice in regards to the relevant tradeoffs.
I'm closing this out since our designs for Connect 2.0 will make the extension option primary for all new users.
| gharchive/issue | 2020-09-25T12:29:53 | 2025-04-01T04:33:38.959348 | {
"authors": [
"markmhx"
],
"repo": "blockstack/ux",
"url": "https://github.com/blockstack/ux/issues/607",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2595428797 | 🛑 Maxthon API is down
In 88b8215, Maxthon API (https://api.maxthon.com) was down:
HTTP code: 500
Response time: 218 ms
Resolved: Maxthon API is back up in df2965b after 13 minutes.
| gharchive/issue | 2024-10-17T18:19:49 | 2025-04-01T04:33:38.961762 | {
"authors": [
"bloodchen"
],
"repo": "bloodchen/upptime",
"url": "https://github.com/bloodchen/upptime/issues/1614",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1010053225 | Add atomWithDangerouslyAllowMutability signatures
Context
Signatures in Recoil.resi interface file were missing for Recoil.atomWithDangerouslyAllowMutability
Related issues: #30 #24
Related PRs: #29
What does this PR do?
It adds proper signatures to the interface file (Recoil.res)
Nice catch! Thanks for your PR
Released in v2.1.1
| gharchive/pull-request | 2021-09-28T18:04:21 | 2025-04-01T04:33:38.963705 | {
"authors": [
"bloodyowl",
"searleser97"
],
"repo": "bloodyowl/rescript-recoil",
"url": "https://github.com/bloodyowl/rescript-recoil/pull/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
386269075 | Drop diamond
Remove disk monitoring in graphite since the disk metrics are supplied
by diamond.
Diamond not only provides disk metrics, but also [CPU|https://github.com/bloomberg/chef-bach/pull/1317/files#diff-9abfd340ffd8489bd5f812ee7e0b10feL186], memory, and network.
Some of the alarms we currently have in zabbix is based on memory.
I would suggest disabling disk metrics collector rather than drop diamond as a whole.
dropped the memory queries in zabbix too.
We don't have zabbix configured to alert on CPU or network so it's safe to drop as well.
| gharchive/pull-request | 2018-11-30T16:55:06 | 2025-04-01T04:33:38.974345 | {
"authors": [
"aespinosa",
"leochen4891"
],
"repo": "bloomberg/chef-bach",
"url": "https://github.com/bloomberg/chef-bach/pull/1317",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2601360392 | When i try to join a game it says roblox failed to launch
Acknowledgement of preliminary instructions
[X] I have read the preliminary instructions, and I am certain that my problem has not already been addressed.
[X] I have thoroughly looked through the available Wiki articles and could not find a solution to my problem.
[X] I am using the latest version of Bloxstrap.
[X] I did not answer truthfully to all the above checkboxes.
Bloxstrap Version
v2.8.0
What problem did you encounter?
When i try to join a game it says roblox failed to launch and takes me to the help website and i dont know why
Bloxstrap Log
N/A
Runs fine, but is encounted as bug.
re: i have also been encountering this, yet roblox launches just fine
dunno why it's happening
I have the same issue, anyone have a solution for this?
That's a bug which is gonna be fixed in the new bloxstrap update
Addressed in Wiki
| gharchive/issue | 2024-10-21T05:20:06 | 2025-04-01T04:33:38.979898 | {
"authors": [
"FallenMxru",
"Spoonus",
"Vad1mCSO2",
"lukaku112",
"pizzaboxer",
"unfunny1"
],
"repo": "bloxstraplabs/bloxstrap",
"url": "https://github.com/bloxstraplabs/bloxstrap/issues/3381",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2668146372 | [BUG] the bloxstrap is running but it isn't displaying
Acknowledgement of preliminary instructions
[X] I have read the preliminary instructions, and I am certain that my problem has not already been addressed.
[X] I have thoroughly looked through the available Wiki articles and could not find a solution to my problem.
[X] I am using the latest version of Bloxstrap.
[ ] I did not answer truthfully to all the above checkboxes.
Bloxstrap Version
v2.8.1
What problem did you encounter?
I open Bloxstrap, and it's displayed on Task Manager but it doesn't display on the screen
Bloxstrap Log
N/A
Client related issues are no longer accepted.
| gharchive/issue | 2024-11-18T11:02:16 | 2025-04-01T04:33:38.982530 | {
"authors": [
"TranBao0707",
"bluepilledgreat"
],
"repo": "bloxstraplabs/bloxstrap",
"url": "https://github.com/bloxstraplabs/bloxstrap/issues/3780",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1037622954 | Constant latency spikes
Hello,
I am seeing a pattern with my AWS speed tests showing jitter/latency spikes up to roughly 120ms. The jitter amount is from about 60ms, since that is my average.
I have a new internet installation and was using AWS Speed test to gauge the reliability. I was wondering if this was normal. Screenshot attached.
Same here
Currently AWS latency is measured by performing a http get on a storage blob file in target region. The spikes might be something specific to AWS storage access as same issue does not occur on same code deployed for Azure latency test does.
I will dig why this could happen, I am also thinking to change average to pencential so the result would make more sense.
Same here:
@blrchen
I am also thinking to change average to pencential so the result would make more sense.
I like this idea.
You asked for help. I would like to assist, but not sure how I can assist you.
| gharchive/issue | 2021-10-27T16:28:24 | 2025-04-01T04:33:38.985898 | {
"authors": [
"Yuvix25",
"axljones",
"blrchen",
"cdp401",
"pwprince"
],
"repo": "blrchen/aws-speed-test",
"url": "https://github.com/blrchen/aws-speed-test/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1174336021 | Support for image file formats WebP and AVIF
See the following closed issue:
https://github.com/bludit/bludit/issues/1168
Supported in Bludit v3.14.1.
| gharchive/issue | 2022-03-19T19:17:13 | 2025-04-01T04:33:38.989736 | {
"authors": [
"clickwork-git",
"dignajar"
],
"repo": "bludit/bludit",
"url": "https://github.com/bludit/bludit/issues/1414",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
163836761 | Test on OS X
Update documentation accordingly.
Work and instructions here: https://github.com/blue-jay/blueprint/issues/36
| gharchive/issue | 2016-07-05T12:15:54 | 2025-04-01T04:33:38.990778 | {
"authors": [
"josephspurrier"
],
"repo": "blue-jay/jay",
"url": "https://github.com/blue-jay/jay/issues/18",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
180180587 | Console Warnings in Bluekit Example
When I launch the Bluekit Example (and also with my app) I get the following console warnings. Should I be concern?
warning.js:36 Warning: Unknown prop 'styles' on <div> tag. Remove this prop from the element. For details, see https://fb.me/react-unknown-prop
in div (created by PropsTable)
in div (created by PropsTable)
in div (created by PropsTable)
in PropsTable (created by PropsSidebar)
in div (created by PropsSidebar)
in div (created by PropsSidebar)
in PropsSidebar (created by Page)
in div (created by Page)
in div (created by StyleRoot)
in StyleRoot (created by Page)
in Page (created by Content)
in div (created by Content)
in Content (created by Page)
in div (created by Page)
in div (created by StyleRoot)
in StyleRoot (created by Page)
in Page (created by StateProvider)
in StateProvider (created by App)
in App printWarning @ warning.js:36
warning.js:36 Warning: Failed prop type: Invalid prop 'name' of type 'string' supplied to 'SourceCode', expected 'object'.
in SourceCode (created by Preview)
in Preview (created by Page)
in div (created by Page)
in div (created by StyleRoot)
in StyleRoot (created by Page)
in Page (created by Content)
in div (created by Content)
in Content (created by Page)
in div (created by Page)
in div (created by StyleRoot)
in StyleRoot (created by Page)
in Page (created by StateProvider)
in StateProvider (created by App)
in App printWarning @ warning.js:36
@andrerecinto These warnings are directly in BlueKit or BlueKit example components. So you shouldn't be concerned. We hope to get rid of them next week in a new release. Thanks for the feedback!
Ok good to know. Thank you.
| gharchive/issue | 2016-09-29T22:36:34 | 2025-04-01T04:33:38.992916 | {
"authors": [
"andrerecinto",
"imtoo"
],
"repo": "blueberryapps/react-bluekit",
"url": "https://github.com/blueberryapps/react-bluekit/issues/69",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
138345769 | Stop using bluebox gem mirror for now
It's under a lot of pressure and needs some time off to sort through its
life.
:+1:
| gharchive/pull-request | 2016-03-04T01:15:28 | 2025-04-01T04:33:39.006667 | {
"authors": [
"j2sol",
"pwplusnick"
],
"repo": "blueboxgroup/ursula",
"url": "https://github.com/blueboxgroup/ursula/pull/1652",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2044152428 | Work only first 11 seconds
device-2023-12-15-205512.webm
Fix PR: https://github.com/blueboy-tm/flutter_v2ray/pull/31
| gharchive/issue | 2023-12-15T18:01:40 | 2025-04-01T04:33:39.008238 | {
"authors": [
"zikwall"
],
"repo": "blueboy-tm/flutter_v2ray",
"url": "https://github.com/blueboy-tm/flutter_v2ray/issues/30",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
499563030 | IndexError: List Index out of range
Traceback (most recent call last):
File "/usr/local/lib64/python3.6/site-packages/tornado/web.py", line 1699, in _execute result = await result File "/usr/local/lib/python3.6/site-packages/samlauthenticator/samlauthenticator.py", line 713, in get login_handler_self)
File "/usr/local/lib/python3.6/site-packages/samlauthenticator/samlauthenticator.py", line 634, in _get_redirect_from_metadata_and_redirect handler_self.redirect(redirect_link_getter(saml_metadata_etree)[0], permanent=False)
IndexError: list index out of range
I am using the simple configuration where:
c.JupyterHub.authenticator_class = 'samlauthenticator.SAMLAuthenticator'
Where the SAML IdP's metadata is stored.
c.SAMLAuthenticator.metadata_filepath = '/PATH/TO/MYFILE.xml'
Hi @jameholme! This looks like an issue with how the authenticator interfaces with your Metadata xml. Would it be possible to post the metadata xml file? If not, could you check what value your metadata xml has at the following XPath?
//md:SingleSignOnService[@Binding='urn:oasis:names:tc:SAML:2.0:bindings:HTTP-Redirect']/@Location
How I'm walking through this issue:
It looks like the error is being thrown here, and I'm not entirely clear what the XPath is since it's dynamically defined above that line. So reading up the code (since there isn't much down), we find that we are being called from here. It looks like that line sets the element to md:SingleSignOnService, and then we can more-or-less reconstruct the XPath from there. If you want to double-check my work, I would encourage you to turn on debug logging for the JupyterHub server - I'm not sure how to do this, but it seems like it should be straightforward?
<md:SingleSignOnService Binding="urn:oasis:names:tc:SAML:2.0:bindings:HTTP-POST" Location="URL REMOVED FOR REASONS"/>
Thanks for the help, this is my first rodeo
@jameholme if this is your first SAML Rodeo, welcome! There's some really cool stuff about the tech - for example, the metadata xml is usually publicly available (here's a link to one of the metadata xmls from the services that I usually use, it should download a file). Just like it's ok to spread public keys around to whoever needs them, it's fine to spread the xml metadata around unless you have a company policy against it.
Looking at the XML snippet you posted, it looks like we're using different SAML bindings. It looks like you're using an HTTP-POST binding, and I'm looking for an HTTP-Redirect binding. The authenticator might work if you alter your metadata xml to say HTTP-Redirect where it currently says HTTP-POST, but this is a fairly off-the-rails configuration that we're setting up.
I think the best course of action would be to leave this issue open so that we know that we need to add HTTP-POST binding in the future, and let me know if you aren't able to get going with the advice in this post. Sound good?
I have changed it to redirect, but it did not change anything and I still received those same errors.
Below is the XML with the stuff I was told to remove before I could send it to you (again thanks for your help!):
<?xml version="1.0"?>
-<md:EntityDescriptor xmlns:md="urn:oasis:names:tc:SAML:2.0:metadata" entityID="COMPANYURL.COM" cacheDuration="DURATION" ID="ID">
-<ds:Signature xmlns:ds="http://www.w3.org/2000/09/xmldsig#">
-<ds:SignedInfo>
<ds:CanonicalizationMethod Algorithm="http://www.w3.org/2001/10/xml-exc-c14n#"/>
<ds:SignatureMethod Algorithm="http://www.w3.org/2001/04/xmldsig-more#rsa-sha256"/>
-<ds:Reference URI="URI">
-<ds:Transforms>
<ds:Transform Algorithm="http://www.w3.org/2000/09/xmldsig#enveloped-signature"/>
<ds:Transform Algorithm="http://www.w3.org/2001/10/xml-exc-c14n#"/>
</ds:Transforms>
<ds:DigestMethod Algorithm="http://www.w3.org/2001/04/xmlenc#sha256"/>
<ds:DigestValue> VALUE </ds:DigestValue>
</ds:Reference>
</ds:SignedInfo>
<ds:SignatureValue> CERT </ds:SignatureValue>
-<ds:KeyInfo>
-<ds:X509Data>
<ds:X509Certificate> CERT </ds:X509Certificate>
</ds:X509Data>
</ds:KeyInfo>
</ds:Signature>
-<md:IDPSSODescriptor WantAuthnRequestsSigned="true" protocolSupportEnumeration="urn:oasis:names:tc:SAML:2.0:protocol">
-<md:KeyDescriptor use="signing">
-<ds:KeyInfo xmlns:ds="http://www.w3.org/2000/09/xmldsig#">
-<ds:X509Data>
<ds:X509Certificate>CERT</ds:X509Certificate>
</ds:X509Data>
</ds:KeyInfo>
</md:KeyDescriptor>
<md:NameIDFormat>urn:oasis:names:tc:SAML:1.1:nameid-format:unspecified</md:NameIDFormat>
<md:SingleSignOnService Location="COMPANYURL.COM" Binding="urn:oasis:names:tc:SAML:2.0:bindings:HTTP-REDIRECT"/>
<saml:Attribute xmlns:saml="urn:oasis:names:tc:SAML:2.0:assertion" NameFormat="urn:oasis:names:tc:SAML:2.0:attrname-format:basic" Name="employeeID"/>
<saml:Attribute xmlns:saml="urn:oasis:names:tc:SAML:2.0:assertion" NameFormat="urn:oasis:names:tc:SAML:2.0:attrname-format:basic" Name="memberOf"/>
<saml:Attribute xmlns:saml="urn:oasis:names:tc:SAML:2.0:assertion" NameFormat="urn:oasis:names:tc:SAML:2.0:attrname-format:basic" Name="mail"/>
<saml:Attribute xmlns:saml="urn:oasis:names:tc:SAML:2.0:assertion" NameFormat="urn:oasis:names:tc:SAML:2.0:attrname-format:basic" Name="title"/>
<saml:Attribute xmlns:saml="urn:oasis:names:tc:SAML:2.0:assertion" NameFormat="urn:oasis:names:tc:SAML:2.0:attrname-format:basic" Name="displayName"/>
</md:IDPSSODescriptor>
<md:ContactPerson contactType="administrative"/>
</md:EntityDescriptor>
Ok, two things -
I'm not 100% sure how Python's XPath works, but it may be that the HTTP-Redirect value has to be case-matched exactly. Check this pdf around line 1026 to see the exact values for this field.
Bad news - currently we don't support signing the SAML Request. We can put that into the backlog, but I'm not sure when I'm going to get to working on it. This is a great opportunity for a pull request :)
| gharchive/issue | 2019-09-27T17:31:26 | 2025-04-01T04:33:39.017108 | {
"authors": [
"distortedsignal",
"jameholme"
],
"repo": "bluedatainc/jupyterhub-samlauthenticator",
"url": "https://github.com/bluedatainc/jupyterhub-samlauthenticator/issues/38",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1748606 | Error Insert file name to database
Error Insert file name to database
// this is the database configuration you need to setup before being able to use a database
$db_host = "localhost"; // usually localhost
$db_user = "user_name"; // the user you created when you create a database or a table, i don't remember
$db_pwd = "password"; // that user's password
$db_database= "database_name"; // the name of the database you created
// this starts the database update script
$mysqli = new mysqli($db_host, $db_user, $db_pwd, $db_database);
// this is a mysqli prepared statement which does the update to the database with the filename of the uploaded image
// add this at the end of the handle_file_upload method, before the line: } else { $file->error = $error;
if($stmt = $mysqli->prepare("INSERT INTO uploaded_images (id, uploaded_image) VALUES (NULL, ?) ")) {
$stmt->bind_param('s', $file_name);
$stmt->execute();
$stmt->close();
}else{
// this removes the image, in case there's an error with the updating of the database,
// so you don't stay with uploaded images that are not in the database
$file_path = $this->options['upload_dir'].$file_name;
$success = is_file($file_path) && $file_name[0] !== '.' && unlink($file_path);
if ($success) {
foreach($this->options['image_versions'] as $version => $options) {
$file = $options['upload_dir'].$file_name;
if (is_file($file)) {
unlink($file);
}
}
}
}
$mysqli->close();
Error Message:
0003.JPG 34.93 KB Error: SyntaxError: JSON.parse Delete
How to solve ?
Best regards
I tried your code and it gave me:
SyntaxError: JSON.parse: unexpected character
Have you succeeded in getting the file name into the DB?
| gharchive/issue | 2011-09-27T03:20:51 | 2025-04-01T04:33:39.023777 | {
"authors": [
"cadnet",
"cnodom"
],
"repo": "blueimp/jQuery-File-Upload",
"url": "https://github.com/blueimp/jQuery-File-Upload/issues/650",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
190507191 | Set font size for buttons to a reasonable size
This only applies to users with javascript disabled as the styling is otherwise overwritten by bootstrap.
200px size was introduced in 76ca40b044b7660f17a77ff82f7fcf14b51878a8 which mentions #2635 but the issue appears to be no longer available so I don't know what the point of the incredibly large size was.
For reference, this is how the demo looks with JS disabled:
Thanks @tribut.
There's a separate file where I fixed this issue.
| gharchive/pull-request | 2016-11-19T17:04:36 | 2025-04-01T04:33:39.025780 | {
"authors": [
"blueimp",
"tribut"
],
"repo": "blueimp/jQuery-File-Upload",
"url": "https://github.com/blueimp/jQuery-File-Upload/pull/3464",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1974851431 | HLS.js usage: bugfix, stability improvements and resilience improvement
I propose the following changes to the HLS.js:
BUGFIX: pass "video" element to create() on restart triggering fatal-error
STABILITY: await MEDIA_ATTACHED event before performing "loadSource" on HLS
STABILITY: await MANIFEST_PARSED event before performing "play" on video element
RESILIENCE: on "MEDIA_ERROR" event perform "recoverMediaError" on HLS
Code was now adjusted.
merged, thanks
| gharchive/pull-request | 2023-11-02T18:44:16 | 2025-04-01T04:33:39.047489 | {
"authors": [
"aler9",
"rse"
],
"repo": "bluenviron/mediamtx",
"url": "https://github.com/bluenviron/mediamtx/pull/2631",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2652629079 | Guideline for use of "tag" versus "label"
Steps to Reproduce
There are references to both "label" as well as "tag" in the English translation.
Are these different concepts? If so, what are their semantics? If not, can they be merged to one word?
In Dutch, a "tag" is often translated as "label", as is the English "label".
Sample of "Label"
Sample of "Tag"
Attachments
No response
What platform(s) does this occur on?
iOS, Android, Web (Desktop), Web (Mobile)
Device Info
No response
What version of the app are you using?
main
Additional Information
No response
They're distinct concepts.
Labels are annotations on users and content. They can be used to hide, warn, and categorize the network.
"Labels" are mainly a moderation tool, like "Rude" or "Intolerance", but a few can also be self-applied to warn about adult and graphic media (which then is hidden or behind a click-through by default). There are also positive labels for example by https://bsky.app/profile/yardcrow.com.
Hashtags are "tags", but "tags" is a bit more general, since the protocol also allows you to add tags to posts outside of the text and there (formally) without a #. Bluesky doesn't yet have UI for the latter though, and they may well end up with a purely visual "#" since they more or less do behave just like hashtags.
Thank you! Labels are for labeling, tags are hastags and for content (text). I will add it to the glossary.
| gharchive/issue | 2024-11-12T16:02:59 | 2025-04-01T04:33:39.058827 | {
"authors": [
"Tamschi",
"monty241"
],
"repo": "bluesky-social/social-app",
"url": "https://github.com/bluesky-social/social-app/issues/6262",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1926004184 | flaky tests
Sometimes CI tests fail. Flaky tests are bad and we should aim to not have any. Here is an example of such a test failing on one instance, but passing when rerun:
https://github.com/bluesky/ophyd-async/actions/runs/10741400139/job/29791703029
FAILED tests/epics/adsimdetector/test_sim.py::test_detector_writes_to_file - bluesky.utils.FailedStatus: <AsyncStatus, task: <coroutine object StandardDetector.trigger at 0x7fecaa70fe60>, errored: TimeoutError()>
I haven't seen flaky tests for several months so going to close this. If they come back, I've wondered about detecting them with infrequent use of pytest-repeat
Flaky tests have re-appeared, I'm going to re-open this ticket.
pytest plugin for generating CTRF report for the above https://pypi.org/project/pytest-json-ctrf/
I think all runs, why not?
Identifying flaky tests vs. tests that were temporarily broken by unfinished PRs?
@callumforrester I've enabled on this draft PR (scroll down to summary section): it hasn't marked any tests as flaky yet, and doesn't provide any easier method of seeing what's flaky/broken.
@coretl for comment too
https://github.com/bluesky/ophyd-async/actions/runs/10809455998?pr=567
Tango tests are apparently flaky because of Tango test infrastructure and there is nothing we can do to fix that. I suggest we report test failures on this ticket and make PRs that @pytest.mark.flaky(reruns=2) individual flaky tests until the CI is reliable again.
The first PR that does this needs to add pytest-rerunfailures to the dev dependencies to get the marker.
| gharchive/issue | 2023-10-04T11:29:19 | 2025-04-01T04:33:39.063311 | {
"authors": [
"DiamondJoseph",
"callumforrester",
"coretl",
"rosesyrett"
],
"repo": "bluesky/ophyd-async",
"url": "https://github.com/bluesky/ophyd-async/issues/23",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
623633980 | set serde derive to use serde untagged
Think the way serde is derived is a non-intuitive way. As I think most people do not want the left and right tags for the values.
Suggest adding untagged to the serde derive. However this would be a breaking change. I do not belive many people are using the serde fetaure as its now implemented.
This will allow us to do something like:
type ComplicatedType = HashMap<String, Either<Vec<String>, String>>
Some json that would be parsed now would be
{
"some_key": "hello",
"some_other_random_key": ["hello","world!"]
}
Either is just one two-case enum, and if you have a particular use case, you should of course make an enum that works for you.
It is not obvious if tagged or untagged should be the default for a general two-case enum, but I would lean towards tagged.
Think Either with serde is usefull when working with rapid prototyping. You get so much for free, like Display... That beeing sed, I agree that Either with conflicting types makes it hard, thow I do not see people using either with simular types for (de)serializing.
FWIW, this was also discussed in #22.
I opened this issue as I thought tagged was a strange default. Do also think that the serde-untagged feature is looking like a good solution.
Ok, it will certainly not be a feature for the reasons clarified in #22.
Closing as I think this is done the right way, reading further throught it once more.
| gharchive/issue | 2020-05-23T10:56:11 | 2025-04-01T04:33:39.068809 | {
"authors": [
"bluss",
"cuviper",
"meltinglava"
],
"repo": "bluss/either",
"url": "https://github.com/bluss/either/issues/46",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
312657587 | Use mongo state to update timestamp
The timestamp method should show when underlying database has changed. Currently mongo driver only watches changes that it makes, but if two arachne servers point at the same mongo db, they won't recognize when the db has changed. Need to use something like https://docs.mongodb.com/manual/changeStreams/ to update timestamp
This same problem also exists for elasticsearch
Will https://stackoverflow.com/questions/27255596/elasticsearch-index-last-update-time work for elasticsearch ?
| gharchive/issue | 2018-04-09T19:32:06 | 2025-04-01T04:33:39.078486 | {
"authors": [
"adamstruck",
"kellrott"
],
"repo": "bmeg/arachne",
"url": "https://github.com/bmeg/arachne/issues/88",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.