id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1878865498
infinite recursion when self assigned to option The following flake leads to an infinite recursion: { outputs = inputs@{flake-parts, nixpkgs, self, ...}: flake-parts.lib.mkFlake {inherit inputs;} { nixosConfigurations.default = nixpkgs.lib.nixosSystem { modules = [ { # assigning self to a non existent option triggers the infinite recursion foo.repoRoot = self; } ]; }; }; } I think this issue is import to fix. For example this could as well happen with the treefmt.repoRoot option. If it ever gets deprecated, users will be puzzled by a hard to debug issue. In my case this lead to a stack overflow with no trace at all. This does only happen when using flake-parts and not with vanilla flakes. @DavHau which version and platform did that happen? When I run it on x86_64-linux, I get infinite recursions instead, which do have a trace, although the quality of the trace varies between Nix versions. 2.13.3 seems best. Good thing the latest two or three releases have test infrastructure to catch such regressions - wish we had it sooner. Seems like a serious problem indeed. I think we should add sourceInfo directly to inputs in Nix to solve this without removing source locations for all the other errors when they occur in the anonymous "root" module that is the mkFlake {} argument. This solution may also help with https://github.com/hercules-ci/flake-parts/issues/148 In my case this lead to a stack overflow with no trace at all. The example above results in an infinite recursion for me as well. I got a stack overflow error in dream2nix, but after stripping it down to a minimal reproducer it became an infinite recursion. Not sure if that change in behavior was due to complexity or due to library versions. Let me know if I should publish the dream2nix expression that lead to a stack overflow. Could you try with this? https://github.com/NixOS/nix/pull/8879 PR description shows how to nix run that nix. You might be able to tell what's the difference between your original problem and reproducer with it. Blocked on https://github.com/NixOS/nix/pull/8908 Does #192 help? With it I get: $ nix eval . --override-input flake-parts github:hercules-ci/flake-parts/refs/pull/192/head warning: not writing modified lock file of flake 'path:/home/user/h/issue-flake-parts-185': • Updated input 'flake-parts': 'github:hercules-ci/flake-parts/7f53fdb7bdc5bb237da7fefef12d099e4fd611ca' (2023-09-01) → 'github:hercules-ci/flake-parts/0effb5db5ccc46f8787c98ca91ec64cc9721c121' (2023-10-13) error: The option `nixosConfigurations' does not exist. Definition values: - In `<unknown-file>' (use '--show-trace' to show detailed location information) Still an error, as expected, but actionable. Fixing it up a bit: { outputs = inputs@{flake-parts, nixpkgs, self, ...}: flake-parts.lib.mkFlake {inherit inputs;} { systems = [ "x86_64-linux" ]; flake.nixosConfigurations.default = nixpkgs.lib.nixosSystem { modules = [ { # assigning self to a non existent option triggers the infinite recursion foo.repoRoot = self; } ]; }; }; } I then get this, simulating nixos-rebuild a bit: $ nix eval .#nixosConfigurations.default.config.system.build.toplevel.drvPath --override-input flake-parts github:hercules-ci/flake-parts/refs/pull/192/head warning: not writing modified lock file of flake 'path:/home/user/h/issue-flake-parts-185': • Updated input 'flake-parts': 'github:hercules-ci/flake-parts/7f53fdb7bdc5bb237da7fefef12d099e4fd611ca' (2023-09-01) → 'github:hercules-ci/flake-parts/0effb5db5ccc46f8787c98ca91ec64cc9721c121' (2023-10-13) error: Neither nixpkgs.hostPlatform nor the legacy option nixpkgs.system has been set. You can set nixpkgs.hostPlatform in hardware-configuration.nix by re-running a recent version of nixos-generate-config. The option nixpkgs.system is still fully supported for NixOS 22.05 interoperability, but will be deprecated in the future, so we recommend to set nixpkgs.hostPlatform. (use '--show-trace' to show detailed location information) Fixed in #192
gharchive/issue
2023-09-02T22:29:53
2025-04-01T06:44:26.173759
{ "authors": [ "DavHau", "roberth" ], "repo": "hercules-ci/flake-parts", "url": "https://github.com/hercules-ci/flake-parts/issues/185", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
963168860
docker build failed in step 6 Hi, I have tried the installation with in a Ubuntu18.04 VM.But I met some error in STEP 6 Step 6/21 : RUN git clone --recursive https://hub.fastgit.org/open5gs/open5gs && cd open5gs && git checkout main && meson build --prefix=`pwd`/install && ninja -C build && cd build && ninja install ---> Running in 7b364147c2b9 Cloning into 'open5gs'... Already on 'main' Your branch is up to date with 'origin/main'. The Meson build system Version: 0.53.2 Source dir: /open5gs Build dir: /open5gs/build Build type: native build Project name: open5gs Project version: 2.3.2 C compiler for the host machine: cc (gcc 9.3.0 "cc (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0") C linker for the host machine: cc ld.bfd 2.34 C++ compiler for the host machine: c++ (gcc 9.3.0 "c++ (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0") C++ linker for the host machine: c++ ld.bfd 2.34 Host machine cpu family: x86_64 Host machine cpu: x86_64 Program git found: YES (/usr/bin/git) Program python3 found: YES (/usr/bin/python3) Program /usr/bin/python3 found: YES (/usr/bin/python3) Compiler for C supports arguments -Wextra: YES Compiler for C supports arguments -Wlogical-op: YES Compiler for C supports arguments -Werror=missing-include-dirs: YES Compiler for C supports arguments -Werror=pointer-arith: YES Compiler for C supports arguments -Werror=init-self: YES Compiler for C supports arguments -Wfloat-equal: YES Compiler for C supports arguments -Wsuggest-attribute=noreturn: YES Compiler for C supports arguments -Werror=missing-prototypes: YES Compiler for C supports arguments -Werror=missing-declarations: YES Compiler for C supports arguments -Werror=implicit-function-declaration: YES Compiler for C supports arguments -Werror=return-type: YES Compiler for C supports arguments -Werror=incompatible-pointer-types: YES Compiler for C supports arguments -Werror=format=2: YES Compiler for C supports arguments -Wstrict-prototypes: YES Compiler for C supports arguments -Wredundant-decls: YES Compiler for C supports arguments -Wimplicit-fallthrough=5: YES Compiler for C supports arguments -Wendif-labels: YES Compiler for C supports arguments -Wstrict-aliasing=3: YES Compiler for C supports arguments -Wwrite-strings: YES Compiler for C supports arguments -Werror=overflow: YES Compiler for C supports arguments -Werror=shift-count-overflow: YES Compiler for C supports arguments -Werror=shift-overflow=2: YES Compiler for C supports arguments -Wdate-time: YES Compiler for C supports arguments -Wnested-externs: YES Compiler for C supports arguments -Wunused: YES Compiler for C supports arguments -Wduplicated-branches: YES Compiler for C supports arguments -Wmisleading-indentation: YES Compiler for C supports arguments -Wno-sign-compare -Wsign-compare: YES Compiler for C supports arguments -Wno-unused-parameter -Wunused-parameter: YES Compiler for C supports arguments -ffast-math: YES Compiler for C supports arguments -fdiagnostics-show-option: YES Compiler for C supports arguments -fstack-protector: YES Compiler for C supports arguments -fstack-protector-strong: YES Compiler for C supports arguments --param=ssp-buffer-size=4: YES meson.build:108: WARNING: Consider using the built-in warning_level option instead of using "-Wextra". Configuring sample.yaml using configuration Configuring 310014.yaml using configuration Configuring csfb.yaml using configuration Configuring volte.yaml using configuration Configuring vonr.yaml using configuration Configuring slice.yaml using configuration Configuring srslte.yaml using configuration WARNING: Output file "configs/sample.yaml" for configure_file() at configs/meson.build:49 overwrites configure_file() output at configs/meson.build:49 Configuring sample.yaml using configuration Configuring non3gpp.yaml using configuration Program /usr/bin/python3 found: YES (/usr/bin/python3) Configuring mme.yaml using configuration Configuring sgwc.yaml using configuration Configuring sgwu.yaml using configuration Configuring smf.yaml using configuration Configuring amf.yaml using configuration Configuring upf.yaml using configuration Configuring hss.yaml using configuration Configuring pcrf.yaml using configuration Configuring nrf.yaml using configuration Configuring ausf.yaml using configuration Configuring udm.yaml using configuration Configuring udr.yaml using configuration Configuring pcf.yaml using configuration Configuring nssf.yaml using configuration Configuring bsf.yaml using configuration Program /usr/bin/python3 found: YES (/usr/bin/python3) Configuring mme.conf using configuration Configuring hss.conf using configuration Configuring smf.conf using configuration Configuring pcrf.conf using configuration Configuring cacert.pem using configuration Configuring mme.cert.pem using configuration Configuring mme.key.pem using configuration Configuring hss.cert.pem using configuration Configuring hss.key.pem using configuration Configuring smf.cert.pem using configuration Configuring smf.key.pem using configuration Configuring pcrf.cert.pem using configuration Configuring pcrf.key.pem using configuration Configuring open5gs-mmed.service using configuration Configuring open5gs-sgwcd.service using configuration Configuring open5gs-smfd.service using configuration Configuring open5gs-amfd.service using configuration Configuring open5gs-sgwud.service using configuration Configuring open5gs-upfd.service using configuration Configuring open5gs-hssd.service using configuration Configuring open5gs-pcrfd.service using configuration Configuring open5gs-nrfd.service using configuration Configuring open5gs-ausfd.service using configuration Configuring open5gs-udmd.service using configuration Configuring open5gs-pcfd.service using configuration Configuring open5gs-nssfd.service using configuration Configuring open5gs-bsfd.service using configuration Configuring open5gs-udrd.service using configuration Configuring 99-open5gs.netdev using configuration Configuring 99-open5gs.network using configuration Configuring open5gs using configuration Configuring open5gs.conf using configuration Has header "arpa/inet.h" : YES Has header "ctype.h" : YES Has header "errno.h" : YES Has header "execinfo.h" : YES Has header "fcntl.h" : YES Has header "ifaddrs.h" : YES Has header "netdb.h" : YES Has header "pthread.h" : YES Has header "signal.h" : YES Has header "stdarg.h" : YES Has header "stddef.h" : YES Has header "stdio.h" : YES Has header "stdint.h" : YES Has header "stdbool.h" : YES Has header "stdlib.h" : YES Has header "string.h" : YES Has header "strings.h" : YES Has header "time.h" : YES Has header "sys/time.h" : YES Has header "unistd.h" : YES Has header "net/if.h" : YES Has header "netinet/in.h" : YES Has header "netinet/in_systm.h" : YES Has header "netinet/udp.h" : YES Has header "netinet/tcp.h" : YES Has header "sys/ioctl.h" : YES Has header "sys/param.h" : YES Has header "sys/random.h" : YES Has header "sys/socket.h" : YES Has header "sys/stat.h" : YES Has header "limits.h" : YES Has header "sys/syslimits.h" : NO Has header "sys/types.h" : YES Has header "sys/wait.h" : YES Has header "sys/uio.h" : YES Checking for function "arc4random" : NO Checking for function "arc4random_buf" : NO Checking for function "getrandom" : YES Checking for function "localtime_r" : YES Checking for function "getifaddrs" : YES Checking for function "getenv" : YES Checking for function "putenv" : YES Checking for function "setenv" : YES Checking for function "unsetenv" : YES Checking for function "strerror_r" : YES Checking for function "sigaction" : YES Checking for function "sigwait" : YES Checking for function "sigsuspend" : YES Checking for function "eventfd" : YES Checking for function "kqueue" : NO Checking for function "epoll_ctl" : YES Run-time dependency threads found: YES Header <pthread.h> has symbol "pthread_barrier_wait" : YES Header <signal.h> has symbol "sys_siglist" : YES Checking if "strerror_r() returns char *" compiles: YES Library execinfo found: NO Checking for function "backtrace" : YES Checking if "clock_gettime()" links: YES Checking if "eventfd(2) system call" links: YES Library socket found: NO Checking if "socket()" links: YES Configuring core-config-private.h using configuration Configuring core-config.h using configuration Compiler for C supports arguments -Wno-shift-negative-value -Wshift-negative-value: YES Compiler for C supports arguments -Wno-unused-but-set-variable -Wunused-but-set-variable: YES Compiler for C supports arguments -Wno-unknown-warning-option -Wunknown-warning-option: NO Cloning into 'freeDiameter'... fatal: unable to access 'https://github.com/open5gs/freeDiameter.git/': GnuTLS recv error (-110): The TLS connection was non-properly terminated. Traceback (most recent call last): File "/usr/lib/python3/dist-packages/mesonbuild/mesonmain.py", line 129, in run return options.run_func(options) File "/usr/lib/python3/dist-packages/mesonbuild/msetup.py", line 245, in run app.generate() File "/usr/lib/python3/dist-packages/mesonbuild/msetup.py", line 159, in generate self._generate(env) File "/usr/lib/python3/dist-packages/mesonbuild/msetup.py", line 192, in _generate intr.run() File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 4167, in run super().run() File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 412, in run self.evaluate_codeblock(self.ast, start=1) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 436, in evaluate_codeblock raise e File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 430, in evaluate_codeblock self.evaluate_statement(cur) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 441, in evaluate_statement return self.function_call(cur) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 788, in function_call return func(node, posargs, kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 285, in wrapped return f(*wrapped_args, **wrapped_kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 174, in wrapped return f(*wrapped_args, **wrapped_kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 3689, in func_subdir self.evaluate_codeblock(codeblock) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 436, in evaluate_codeblock raise e File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 430, in evaluate_codeblock self.evaluate_statement(cur) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 441, in evaluate_statement return self.function_call(cur) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 788, in function_call return func(node, posargs, kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 285, in wrapped return f(*wrapped_args, **wrapped_kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 174, in wrapped return f(*wrapped_args, **wrapped_kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 3689, in func_subdir self.evaluate_codeblock(codeblock) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 436, in evaluate_codeblock raise e File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 430, in evaluate_codeblock self.evaluate_statement(cur) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 441, in evaluate_statement return self.function_call(cur) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 788, in function_call return func(node, posargs, kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 285, in wrapped return f(*wrapped_args, **wrapped_kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 174, in wrapped return f(*wrapped_args, **wrapped_kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 3689, in func_subdir self.evaluate_codeblock(codeblock) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 436, in evaluate_codeblock raise e File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 430, in evaluate_codeblock self.evaluate_statement(cur) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 443, in evaluate_statement return self.assignment(cur) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 1064, in assignment value = self.evaluate_statement(node.value) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 441, in evaluate_statement return self.function_call(cur) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 788, in function_call return func(node, posargs, kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 285, in wrapped return f(*wrapped_args, **wrapped_kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 174, in wrapped return f(*wrapped_args, **wrapped_kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 143, in wrapped return f(*wrapped_args, **wrapped_kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 2540, in func_subproject return self.do_subproject(dirname, 'meson', kwargs) File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 2582, in do_subproject resolved = r.resolve(dirname, method) File "/usr/lib/python3/dist-packages/mesonbuild/wrap/wrap.py", line 187, in resolve self.get_git() File "/usr/lib/python3/dist-packages/mesonbuild/wrap/wrap.py", line 282, in get_git verbose_git(['clone', self.wrap.get('url'), self.directory], self.subdir_root, check=True) File "/usr/lib/python3/dist-packages/mesonbuild/wrap/wrap.py", line 62, in verbose_git return git(cmd, workingdir, check=check).returncode == 0 File "/usr/lib/python3/dist-packages/mesonbuild/mesonlib.py", line 61, in git pc = subprocess.run([GIT, '-C', workingdir] + cmd, File "/usr/lib/python3.8/subprocess.py", line 516, in run raise CalledProcessError(retcode, process.args, subprocess.CalledProcessError: Command '['/usr/bin/git', '-C', '/open5gs/subprojects', 'clone', 'https://github.com/open5gs/freeDiameter.git', 'freeDiameter']' returned non-zero exit status 128. Compiler for C supports arguments -Wno-missing-prototypes -Wmissing-prototypes: YES Compiler for C supports arguments -Wno-missing-declarations -Wmissing-declarations: YES Compiler for C supports arguments -Wno-discarded-qualifiers -Wdiscarded-qualifiers: YES Compiler for C supports arguments -Wno-redundant-decls -Wredundant-decls: YES Compiler for C supports arguments -Wno-shift-overflow -Wshift-overflow: YES Compiler for C supports arguments -Wno-float-equal -Wfloat-equal: YES Compiler for C supports arguments -Wno-implicit-fallthrough -Wimplicit-fallthrough: YES Compiler for C supports arguments -Wno-incompatible-pointer-types-discards-qualifiers -Wincompatible-pointer-types-discards-qualifiers: NO Compiler for C supports arguments -Wno-format-nonliteral -Wformat-nonliteral: YES Compiler for C supports arguments -Wno-cpp -Wcpp: YES Found pkg-config: /usr/bin/pkg-config (0.29.1) Run-time dependency yaml-0.1 found: YES 0.2.2 Has header "netinet/sctp.h" : YES Library sctp found: YES Configuring sctp-config.h using configuration Run-time dependency libmongoc-1.0 found: YES 1.16.1 Removing intermediate container 7b364147c2b9 The command '/bin/sh -c git clone --recursive https://hub.fastgit.org/open5gs/open5gs && cd open5gs && git checkout main && meson build --prefix=`pwd`/install && ninja -C build && cd build && ninja install' returned a non-zero code: 2 root@user:/home/user/volte/docker_open5gs/base# It seems git clone error,I use https://hub.fastgit.org/***.git to download faster than https://github.com/***.git for geographical reasons.So how should I replace this address https://github.com/open5gs/freeDiameter.git? Thanks in advance! Update I use VPN to improve the network environment,and execute git clone https://github.com/open5gs/freeDiameter.git in VM,it works fine. root@user:/home/user# git clone https://github.com/open5gs/freeDiameter Cloning into 'freeDiameter'... remote: Enumerating objects: 749, done. remote: Counting objects: 100% (749/749), done. remote: Compressing objects: 100% (462/462), done. remote: Total 749 (delta 288), reused 727 (delta 266), pack-reused 0 Receiving objects: 100% (749/749), 1.24 MiB | 1.86 MiB/s, done. Resolving deltas: 100% (288/288), done. But when I execute docker build --no-cache --force-rm -t docker_open5gs . it failed in STEP 6 once agian.. Hi I resolved this issue. I find a solution for ERROR GnuTLS recv error (-110),that is add 3 commands in Dockfile as below: RUN apt-get install gnutls-bin RUN git config --global http.sslVerify false RUN git config --global http.postBuffer 1048576000 Then the error disappeared.I will continue with the installation and there may be other problems, but this issuecan be closed,thanks.
gharchive/issue
2021-08-07T06:38:21
2025-04-01T06:44:26.191757
{ "authors": [ "myonlystarWang" ], "repo": "herlesupreeth/docker_open5gs", "url": "https://github.com/herlesupreeth/docker_open5gs/issues/49", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
625577262
Cleanup optimizes imports, replaces assert with assert_eq and fixes a linter warning @stlankes Review required
gharchive/pull-request
2020-05-27T10:32:51
2025-04-01T06:44:26.193533
{ "authors": [ "jschwe" ], "repo": "hermitcore/rusty-loader", "url": "https://github.com/hermitcore/rusty-loader/pull/6", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
82578927
Allow client to cancel by transfer name Allow pg:backups cancel to take an optional transfer name, and change the default for when no name provided to pick the newest active backup, rather than relying on order from api. @uhoh-itsmaciek this seem ok? Sure, looks good.
gharchive/pull-request
2015-05-29T20:40:52
2025-04-01T06:44:26.202805
{ "authors": [ "tef", "uhoh-itsmaciek" ], "repo": "heroku/heroku", "url": "https://github.com/heroku/heroku/pull/1595", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1488945434
🛑 Auth-Bridge - Test 1 is down In 09108a2, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 3796002.
gharchive/issue
2022-12-10T20:48:24
2025-04-01T06:44:26.214628
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/10083", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1490263016
🛑 Software Center - Test 1 is down In 7c72fa6, Software Center - Test 1 ($SOFTWARECENTER_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Software Center - Test 1 is back up in 1cb8a54.
gharchive/issue
2022-12-11T17:40:17
2025-04-01T06:44:26.216897
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/10127", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1672433725
🛑 Auth-Bridge - Test 1 is down In a5f4970, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 8e172bc.
gharchive/issue
2023-04-18T06:24:56
2025-04-01T06:44:26.219117
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/14082", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1693508779
🛑 Auth-Bridge - Test 1 is down In 05b13d1, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 16a91da.
gharchive/issue
2023-05-03T06:57:19
2025-04-01T06:44:26.221583
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/14821", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1726905578
🛑 Auth-Bridge - Test 1 is down In 248ce6f, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 482de49.
gharchive/issue
2023-05-26T04:55:29
2025-04-01T06:44:26.223759
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/15965", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1819754374
🛑 Auth-Bridge - Test 1 is down In b5d427f, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 64becde.
gharchive/issue
2023-07-25T07:49:48
2025-04-01T06:44:26.226009
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/18806", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1184839634
🛑 Auth-Bridge - Test 1 is down In 69b64aa, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 8e28a09.
gharchive/issue
2022-03-29T13:03:34
2025-04-01T06:44:26.228215
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/1907", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1903585607
🛑 Auth-Bridge - Test 1 is down In ea21d79, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 353a124 after 21 minutes.
gharchive/issue
2023-09-19T19:13:03
2025-04-01T06:44:26.230498
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/21729", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1204653624
🛑 Auth-Bridge - Test 1 is down In 50d4256, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 21bd4f8.
gharchive/issue
2022-04-14T15:12:45
2025-04-01T06:44:26.232920
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/2538", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2034279120
🛑 Software Center - Test 1 is down In a5f7a71, Software Center - Test 1 ($SOFTWARECENTER_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Software Center - Test 1 is back up in b7356ce after 28 minutes.
gharchive/issue
2023-12-10T08:55:16
2025-04-01T06:44:26.235093
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/25781", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2081528860
🛑 Auth-Bridge - Test 1 is down In 54e0885, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 60635c4 after 54 minutes.
gharchive/issue
2024-01-15T08:46:27
2025-04-01T06:44:26.237319
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/27544", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2100776793
🛑 Software Center - Test 1 is down In b8c3a47, Software Center - Test 1 ($SOFTWARECENTER_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Software Center - Test 1 is back up in 2088002 after 28 minutes.
gharchive/issue
2024-01-25T16:42:38
2025-04-01T06:44:26.239512
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/28032", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2266792185
🛑 Software Center - Test 1 is down In 39d25b5, Software Center - Test 1 ($SOFTWARECENTER_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Software Center - Test 1 is back up in 0ac47fb after 43 minutes.
gharchive/issue
2024-04-27T03:33:46
2025-04-01T06:44:26.241889
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/32697", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1292451476
🛑 Auth-Bridge - Test 1 is down In b0d97ef, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 7820a0c.
gharchive/issue
2022-07-03T23:42:15
2025-04-01T06:44:26.244325
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/5315", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1349541800
🛑 Auth-Bridge - Test 1 is down In 714970b, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in ba24fa4.
gharchive/issue
2022-08-24T14:30:50
2025-04-01T06:44:26.246519
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/6996", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1414982913
🛑 Auth-Bridge - Test 1 is down In dc85dde, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down: HTTP code: 0 Response time: 0 ms Resolved: Auth-Bridge - Test 1 is back up in 6f6c4f0.
gharchive/issue
2022-10-19T13:30:56
2025-04-01T06:44:26.248754
{ "authors": [ "herrphon" ], "repo": "herrphon/upptime", "url": "https://github.com/herrphon/upptime/issues/8337", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
126673485
Fahrauftrag ausdrucken -> Fehler! Erwartet wird: Bereitstellen -> Fahrauftrag -> Drucken Fahrauftrag wird auf Printer gedruckt! Ausgegeben wird: Guten morgen Josi Ist kein Fehler, nur ein undeutliche Fehlermeldung. Diese Funktion ist noch nicht implementiert. Ich bin seit letzten Freitag genau hier am bauen, bald sollte Bereitstellen->Fahrauftrag->Drucken auch möglich sein. Martin Jonasse Seemattstrasse 38 6333 Hünenberg See Privat: 041 780 76 12 Mobile: 079 832 69 10 Am 14.01.2016 um 16:08 schrieb Josi Conrad notifications@github.com: Erwartet wird: Bereitstellen -> Fahrauftrag -> Drucken Fahrauftrag wird auf Printer gedruckt! Ausgegeben wird: — Reply to this email directly or view it on GitHub. This is "work in progress". Implementiert: Fahrauftrag -> Drucken erstellt eine PDF Datei für den Download bereit, gedruckt wird in Adobe Acrobat Reader.
gharchive/issue
2016-01-14T15:08:17
2025-04-01T06:44:26.262565
{ "authors": [ "Josi-Conrad", "Martin-Jonasse" ], "repo": "hertus/sfitixi", "url": "https://github.com/hertus/sfitixi/issues/153", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
428919890
"login" route is buggy Pointing the browser directly to the "login" route, or reloading the page while the "login" route is currently active, does not work. In both cases there is a redirect to "gamesinprogress" which leads to an authentication error. Possibly related: When pointing the browser to the "/" route, there is redirect to "gamesinprogress" before there is a redirect to "login".
gharchive/issue
2019-04-03T18:47:32
2025-04-01T06:44:26.266429
{ "authors": [ "herzbube" ], "repo": "herzbube/littlego-web", "url": "https://github.com/herzbube/littlego-web/issues/11", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
153965801
Specify a recursivity depth Hi ! Thanks for this great plugin for django rest framework. Is there any way to specify the depth of the recursion ? I tried using the class Meta depth attribute but I don't think the field uses it. What would be the best approach to this ? Best regards, This is not currently supported. I would suggest pruning the dataset before you try to serialize it. If you are deserializing something, then you could deserialize it and then prune. I'm using it on a ModelViewset so I don't know how I could prune it ? There is no queryset or get_queryset method in the arguments of the field... Can you set the queryset in the ModelViewSet? No... My datastructure is really simple: # models.py class Node(models.Model): parent = models.ForeignKey('self', related_name='children') # api.py class NodeSerializer(serializers.HyperlinkedModelSerializer): parent = serializers.HyperlinkedRelatedField( view_name='api:parameter-detail', queryset=Node.objects.all(), ) children = RecursiveField(many=True, allow_null=True) class Meta: model = Node fields = ('parent', 'children') depth = 2 class NodeViewSet(viewsets.ModelViewSet): queryset = Node.objects.all() I cannot prune or filter the queryset. It will depend on which initial node is requested. I need to limit the depth of the RecursiveField only... Or maybe I'm missing something ? You could look in to the django treebeard package for a tree with more features. Or django mptt. I believe there are probably others as well. Or I suppose you could extend the recursive field to add the functionality your desire I'll just extend the recursive field to add the functionality I need. Both mptt and treabeard have limitations or defaults that don't work with my project. Thanks.
gharchive/issue
2016-05-10T09:47:06
2025-04-01T06:44:26.380731
{ "authors": [ "achedeuzot", "heywbj" ], "repo": "heywbj/django-rest-framework-recursive", "url": "https://github.com/heywbj/django-rest-framework-recursive/issues/11", "license": "isc", "license_type": "permissive", "license_source": "bigquery" }
736686875
Possibly another method? Hey there, There's another elevated COM based method I'd like to share. Are you willing to take a look? Thanks. Hello, sure, I always welcome anything new. Awesome, I ask because I don't want to pester you if you're not free at the time. Anyways, this method is three steps. It uses environment variables abuse/modification, shell protocol handler hijack, and lastly the elevated COM interface IFwCplLua. Basically IFwCplLua::LaunchAdvancedUI() uses ShellExecuteExW() call with %WinDir%\System32\WF.msc. What this method does is change WinDir to custom location, and launch custom WF.msc. The custom WF.msc launches a custom protocol which in turn opens cmd.exe as admin. Here is the code: https://github.com/AzAgarampur/byeintegrity4-uac/ This method actually already was in UACMe as 42, https://github.com/hfiref0x/UACME/blob/v3.2.x/Source/Akagi/methods/hybrids.c#L2392 except it abused mscfile handler hijack without touching environment variables or using fake msc snap-in file. Starting from RS4 it produced mixed results - was working and not working at same time and later was set as fixed and removed. What does this custom msc btw? It just run something with that protocol specified with help of shockwave flash object? Also I will be able to test this only on Saturday as I'm away of main PC. It just run something with that protocol specified with help of shockwave flash object? Basically. The exploit creates a URL association called protocol-byeintegrity4, and the shockwave object uses the link to launch the desired process. <String ID="3" Refs="1">protocol-byeintegrity4:</String> Well, I can confirm it works (tested on 19042). I will look how I can integrate this into UACMe and post update here, presumably next week. Thanks for sharing. Awesome! Thanks for letting me know. Pushed into the dev branch. Currently it was tested only on Win10 1809. I've done testing this on Windows 7 SP1, Windows 8.1 full patch. Tested on Windows 10 21H1 (20241) so I can assume it will work on previous Win10 versions too. If no critical bugs were found I will release this later this week, as I still want to do some other additions not related to this particular method. As usual, thanks for contribution, good work! Hey, thanks! I appreciate the time you take to look at my work. Do you think we can drop the registry flush calls? I don't think it'll change anything if we remove them. Yes, sure. This is unrelated, but the link that has the bs explanation from MS about why it works should be updated to this: https://devblogs.microsoft.com/oldnewthing/20160816-00/?p=94105 Hopefully you can include it in the readme for the next release. Done.
gharchive/issue
2020-11-05T07:31:05
2025-04-01T06:44:26.390454
{ "authors": [ "AzAgarampur", "hfiref0x" ], "repo": "hfiref0x/UACME", "url": "https://github.com/hfiref0x/UACME/issues/88", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
418615997
[ bug ]多次刷新SQL工单详情时,工单内容出现重复 仅v1.4.0以前产生的旧数据会出现该情况,新数据不会 重现步骤 打开工单详情:http://139.199.0.191/detail/6/ 多次刷新,详情的SQL列表出现重复 截图 错误日志 无,仅前端展示异常 版本信息 应用版本 v1.4.3 部署方式 Docker 出现原因是 https://github.com/hhyo/archery/blob/master/sql/engines/models.py#L42 使用了空列表作为默认参数初始化对象属性, 并且在代码中使用了原地修改的方法修改对象属性, 这就等同于修改了对象的默认值。 解决方案是将对象属性的默认值改为None, 然后在init中判断,如果是None就置为空列表[] 这样的话,这个空列表是新生成的空列表, 而不是默认参数的那个列表,也就不会影响到之后生成的心对象 参考: https://stackoverflow.com/questions/366422/what-is-the-pythonic-way-to-avoid-default-parameters-that-are-empty-lists 经测试已修复. http://139.199.0.191/detail/6/
gharchive/issue
2019-03-08T03:28:32
2025-04-01T06:44:26.424570
{ "authors": [ "LeoQuote", "hhyo" ], "repo": "hhyo/archery", "url": "https://github.com/hhyo/archery/issues/63", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
185070058
HHH-11144 - Add test for issue https://hibernate.atlassian.net/browse/HHH-11144 Applied upstream, thanks Applied upstream, thanks
gharchive/pull-request
2016-10-25T09:53:28
2025-04-01T06:44:26.425935
{ "authors": [ "dreab8", "vladmihalcea" ], "repo": "hibernate/hibernate-orm", "url": "https://github.com/hibernate/hibernate-orm/pull/1607", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
40343443
[4.3] HHH-9337 Region.destroy() attempts to remove a cache listener, but regio... ...n class is not annotated with @Listener https://hibernate.atlassian.net/browse/HHH-9337 Cherry-picked and pushed. Thanks! Gail
gharchive/pull-request
2014-08-15T12:15:06
2025-04-01T06:44:26.427363
{ "authors": [ "gbadner", "pferraro" ], "repo": "hibernate/hibernate-orm", "url": "https://github.com/hibernate/hibernate-orm/pull/784", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2069031169
🛑 kiwifarms.st is down In d2a3703, kiwifarms.st (https://kiwifarms.st) was down: HTTP code: 0 Response time: 0 ms Resolved: kiwifarms.st is back up in 768fc56 after 33 minutes.
gharchive/issue
2024-01-07T07:56:41
2025-04-01T06:44:26.431842
{ "authors": [ "hickoryhouse" ], "repo": "hickoryhouse/kf", "url": "https://github.com/hickoryhouse/kf/issues/2869", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2209686011
🛑 Mad at the Internet is down In e53cdb5, Mad at the Internet (https://madattheinternet.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Mad at the Internet is back up in 8c76a38 after 10 minutes.
gharchive/issue
2024-03-27T02:27:30
2025-04-01T06:44:26.434543
{ "authors": [ "hickoryhouse" ], "repo": "hickoryhouse/kf", "url": "https://github.com/hickoryhouse/kf/issues/3261", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1510209097
🛑 Kiwi Farms Forum is down In d41e9a8, Kiwi Farms Forum (https://kiwifarms.net) was down: HTTP code: 0 Response time: 0 ms Resolved: Kiwi Farms Forum is back up in 5841932.
gharchive/issue
2022-12-24T23:38:15
2025-04-01T06:44:26.436912
{ "authors": [ "hickoryhouse" ], "repo": "hickoryhouse/kf", "url": "https://github.com/hickoryhouse/kf/issues/84", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
500722317
Project settings A solution for #42. Please refer to #37 . I'm sorry but I don't understand your objections to my patch. the super is the problem? Why? If I'm overriding a method from a class, calling super is a standard way how use the original method. Of course, you can reject the patch but the plugins will remain in conflict then. I agree that using super is the standard way to override class methods. However, since ProjectsHelper is a module, I think it is incorrect to use super. @maeda-m I think there should be a Module#prepend instead of ProjectsController.send :helper ? here'a an article about the difference https://www.justinweiss.com/articles/rails-5-module-number-prepend-and-the-end-of-alias-method-chain/ Thank you. I learned from you. Released version 1.5.0 !
gharchive/pull-request
2019-10-01T07:45:44
2025-04-01T06:44:26.470076
{ "authors": [ "ahorek", "maeda-m", "picman" ], "repo": "hidakatsuya/redmine_default_custom_query", "url": "https://github.com/hidakatsuya/redmine_default_custom_query/pull/43", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2410141209
Incorrect match in the source code results in inaccurate matching results.I don't know if this is a bug... this code const [_, commentSyntax, searchPhrase, commentSyntaxEnd] = match; in matchSearchPhrase.ts 。The suggestions for this extension are as follows: The result obtained by the regular expression is 4 groups. In addition to the first item, there should be 4 items in the match to fully indicate it. The searchPhrase in the source code only represents find, not find {question}, so one item should be added, as shown below: const [_, commentSyntax, searchSymbol, searchPhrase, commentSyntaxEnd] = match;,The output after debugging is as follows: After the modification, the result of debugging this extension is as follows: You're right. I've pushed an update for this issue https://github.com/hieunc229/copilot-clone/commit/1edf91fd0b8dcfcf4b27a4d2a6d86d5c8e865de7 and will make an update to the extension marketplace Thanks for spotting it @mengyangz86
gharchive/issue
2024-07-16T03:52:11
2025-04-01T06:44:26.483786
{ "authors": [ "hieunc229", "mengyangz86" ], "repo": "hieunc229/copilot-clone", "url": "https://github.com/hieunc229/copilot-clone/issues/88", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
113817628
AssertionViewメソッド用のbuilderを追加 スクリーンショットの取得、検証であるassertViewは以下のAPIとなっている。 assertView(String screenshotId, CompareTarget[] compareTargets,DomSelector[] hiddenElementsSelectors) CompareTargetやDomSelector内も複雑なオブジェクトとなっているため、 ビルダーを用意して簡略化できるようにする 下記コミットで対応しました。 c7e1eeb6ee4b0c4655e6188d1c7745e11c8332cd ce9339c81883c0e3efb1e6e76266f3e90509b2d9 55fb8110b1a73559a342f7e0bcc5c44dc10835d0 796cd8ed2eb17bce472137f74caeb9be87fb2392 78a4ba12cb1ef437a9cf3039220bbf79fc061db3 db7d67290832e298441c21b0252596305a4637ae 6379925d124ec4df3e231d68284e9f76284db7a1 13b67cab600479ce9401843979057bf8c9b5b01a a760f32f939f43ef3a96c68fb678870c42a56035 781490698fa4070b19bedf4c98fcb2918af6de43 ビルダー用の説明ページを追加しました https://www.htmlhifive.com/conts/web/view/pitalium-reference/screenshot-argument
gharchive/issue
2015-10-28T13:09:08
2025-04-01T06:44:26.486798
{ "authors": [ "tkashi" ], "repo": "hifive/hifive-pitalium", "url": "https://github.com/hifive/hifive-pitalium/issues/33", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
199260799
JS: Expressions inside template literals don't display correctly res = `start ${tags.map(tag => `pre ${tag} post`).join('\n ')}end`; Displays like: pre and tags.map don't have the right color Fiddle: https://jsfiddle.net/ug3tLcf6/ Not sure I see any issue. What are you expecting? Or perhaps this has been fixed. https://jsfiddle.net/ajoshguy/ym41ukjd/ Partly fixed (pre is coloured now). Partly as expected (tags.map is indeed also not coloured outside the template literal, I expected it to be coloured). Closing this. Thanks for responding after 3y instead of closing it after x months due to inactivity 😊! Glad to help!
gharchive/issue
2017-01-06T18:46:37
2025-04-01T06:44:26.609542
{ "authors": [ "teameh", "yyyc514" ], "repo": "highlightjs/highlight.js", "url": "https://github.com/highlightjs/highlight.js/issues/1405", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
246110508
Scrollbar not working on certain lines of code inside Hello, I am using the latest version of highlight.js through a Wordpress plugin, and am having problems. When the code stretches out longer than the maximum width of the container the scroll bar appears BUT in certain code you can't click and move it. It seems frozen. You can get inside the box and highlight the code from within, but the scroll bar is totally frozen! Here is are 2 line of code that DO NOT have a working scroll bar: <pre><code class="shell">sudo su -c 'dnf -y install https://download1.rpmfusion.org/free/fedora/rpmfusion-free-release-$(rpm -E %fedora).noarch.rpm https://download1.rpmfusion.org/nonfree/fedora/rpmfusion-nonfree-release-$(rpm -E %fedora).noarch.rpm' sudo rpm --import https://packages.microsoft.com/keys/microsoft.asc sudo sh -c 'echo -e "[code]\nname=Visual Studio Code\nbaseurl=https://packages.microsoft.com/yumrepos/vscode\nenabled=1\ngpgcheck=1\ngpgkey=https://packages.microsoft.com/keys/microsoft.asc" > /etc/yum.repos.d/vscode.repo'</code></pre> ![non-working-scrollbar](https://user-images.githubusercontent.com/12832280/28683201-9f90eb54-72f7-11e7-8a0e-280a13984299.png) <pre><code class="shell">echo fs.inotify.max_user_watches=524288 | sudo tee -a /etc/sysctl.conf sudo sysctl -p</code></pre> and here is one that works perfectly: <pre><code class="shell">sudo sed -i "s/User apache/User $USERNAME/g" /etc/httpd/conf/httpd.conf</code></pre> If I get rid of the class="shell" then it still has the exact same problem. Any idea? I attached a screen shot of a not working scrollbar. THANKS! I’m closing this issue because it has been inactive for over a year. This probably means that it is not reproducible or it has been fixed in a newer version. Please reopen if you still encounter this issue with the latest stable version. Thank you!
gharchive/issue
2017-07-27T17:18:18
2025-04-01T06:44:26.613289
{ "authors": [ "David-Else" ], "repo": "highlightjs/highlight.js", "url": "https://github.com/highlightjs/highlight.js/issues/1577", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
189009499
Allow starts to be defined as 'self' much like one can do in contains I used this for a weird linking system. Colon separated values as { end: /:/, endsWithParent: true, starts: 'self', ... } It's not a necessary thing, but it's consistent with the feature for contains. I'm confused when you would use this exactly? Since this was so long ago I hardly remember but going off of "colon separated values" I think I used it for like 1 : 1 : ... but where 1 is a complex type (like could be a number or a string or whatever) It's in here, and while the grammar sorta works, it's slow in complex cases so I kind of have up. Seems like a good way to get stuck in an infinite loop. How does it break out if after end it always starts self? And typically you'd do this with a parent and then inside your contains you'd just have you matcher... so it would run over and over and over again as many times as needed until you left the mode... no need for this start hack. It breaks out when the parent ends, as the example says. :p ... contains: [ { begin: /\d\s:\s?/, endWithParent: true } ] Doesn't that do exactly the same thing? I don't know, I'm out of touch with the system. I'm pretty sure I only did it for reducing redundancy. I have no complaints if this is just closed. Ok, closing. It would have needed extra documentation and I had some issues with the code also... and generally this type of feature we shouldn't just add "abstractly"... if there was a grammar included with it that required the new functionality that would make a much better case for why it's necessary/useful, etc...
gharchive/pull-request
2016-11-14T01:09:32
2025-04-01T06:44:26.616376
{ "authors": [ "logicplace", "yyyc514" ], "repo": "highlightjs/highlight.js", "url": "https://github.com/highlightjs/highlight.js/pull/1348", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
202914482
add postcss-hocus plugin postcss-hocus lets you type a:hocus instead of a:hover, a:focus. Short and simple :) 🎩 ✨ 🔮
gharchive/pull-request
2017-01-24T19:22:06
2025-04-01T06:44:26.628338
{ "authors": [ "Kilian", "himynameisdave" ], "repo": "himynameisdave/postcss-plugins", "url": "https://github.com/himynameisdave/postcss-plugins/pull/169", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
721690022
Uploading video annotations from Premiere Pro I tried to make a video project in Audiannotate and neither the video link nor the annotations were recognized by the program. To Reproduce Steps to reproduce the behavior: Go to http://audiannotate.brumfieldlabs.com/ Click on My Projects and edit project (in this case, Camille 1921) Add link and info Upload annotations Several errors, one when I tried to add the duration of the video (1:09:30 or 69:30, tried both) and another when I tried to upload my annotations, first without editing them and then editing them to have only four collumns Expected behavior I expected the link to work and the annotations to link to the timestamps with the duration Screenshots Got this error when I tried to correct the duration of the video: And these errors when I tried to upload the annotations: Additional context I think it has something to do with the format of my annotations exported from premiere, which I tried to fix by removing two collumns (so it is now just marker name/description/time in/time out) and changing the timecodes to a 00:00:00 format rather than 00;00;00 Oh also, these are my annotations as a .txt - Camille 1921 annotations.txt This is a really gnarly one; fortunately our next feature (#125) should finally get us to where we need to be to support Premiere. We think we're also going to hard-code support for Premiere upload annotations earlier to un-block research work.
gharchive/issue
2020-10-14T18:36:30
2025-04-01T06:44:26.641570
{ "authors": [ "benwbrum", "jreinschmidt" ], "repo": "hipstas/AudiAnnotate", "url": "https://github.com/hipstas/AudiAnnotate/issues/122", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1727045925
建议识别后导出Excel表格数据,并对应图片超链接,更加实用 建议识别后导出Excel表格数据,表格数据第一列是图片名称,第二列识别图片的内容,第三列对应图片超链接,这样可以很快找到需要搜索的内容图片。 OK,已有计划 邮件已经收到 期待你的更新,还有就是要对批量提取的字段,如果可以自定义就好了。 Copied to clipboard! ------------------ 原始邮件 ------------------ 发件人: @.>; 发送时间: 2023年5月27日(星期六) 下午5:58 收件人: @.>; 抄送: @.>; @.>; 主题: Re: [hiroi-sora/Umi-OCR] 建议识别后导出Excel表格数据,并对应图片超链接,更加实用 (Issue #148) OK,已有计划 — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you authored the thread.Message ID: @.***> 期待大佬的更新 V2预览版 已支持输出csv格式,可导入excel。 第一列图片名称、第二列识别内容、第三列图片路径。 邮件已经收到
gharchive/issue
2023-05-26T07:19:07
2025-04-01T06:44:26.646896
{ "authors": [ "WhoIAmm", "csq4017", "hiroi-sora" ], "repo": "hiroi-sora/Umi-OCR", "url": "https://github.com/hiroi-sora/Umi-OCR/issues/148", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1769743374
i cannot confirm my testnet requested since 3days now Describe the bug A clear and concise description of what the bug is. Transaction ID Address Block# Time stamp To Reproduce Steps to reproduce the behavior: Go to '...' Click on '....' Scroll down to '....' See error Expected behavior A clear and concise description of what you expected to happen. Screenshots If applicable, add screenshots or consol.log to help explain your problem. Desktop (please complete the following information): OS: [e.g. iOS] Browser [e.g. chrome, safari] Version [e.g. 22] Smartphone (please complete the following information): Device: [e.g. iPhone6] OS: [e.g. iOS8.1] Browser [e.g. stock browser, safari] Version [e.g. 22] Additional context Add any other context about the problem here: Please check the testnet status here https://status.hiro.so/
gharchive/issue
2023-06-22T13:58:11
2025-04-01T06:44:26.652656
{ "authors": [ "Christianogbonnaya", "andresgalante" ], "repo": "hirosystems/explorer", "url": "https://github.com/hirosystems/explorer/issues/1192", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1803880958
STX Token Transfer SENT which I did NOT do Here's the TRX id: 0x6829fa018ff0642fbc90255a8d3901e38f7f763c89c8fad629a2cbec01efaa63 Where did my tokens go? Why did that happen? Please HELP! URGENT! Sender address: SP2HEDH4SXP1A34M9KDY00PJEM6YP1QDMVD2BFYDH Recipient address: SP2KW0M6MBSSAV1BFDKH56VFNZK73Z36C0N369K9M Link: https://explorer.hiro.so/txid/0x6829fa018ff0642fbc90255a8d3901e38f7f763c89c8fad629a2cbec01efaa63?chain=mainnet Where did my STX go? Is it still within STX / Hiro system or network? Please help clarify what happened. This isn't an Explorer issue. Please request community support on the #support channel in Discord.
gharchive/issue
2023-07-13T23:09:11
2025-04-01T06:44:26.655363
{ "authors": [ "STX-Stargem", "andresgalante" ], "repo": "hirosystems/explorer", "url": "https://github.com/hirosystems/explorer/issues/1234", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
558727765
Integrate the 'argparse' library; add -l command-line option Implements the rest of https://github.com/hishamhm/tl/issues/35. Note that running tl script.tl will now result in an error. Also, it would be nice if we had tests for the CLI tool :) Also, it would be nice if we had tests for the CLI tool :) @pdesaulniers Added them! :grin: Awesome! I added some tests for the -l argument too. We'll need some tests for tlconfig.lua as well. To do this, I think we'll need to add a -p <path to directory with tlconfig.lua> argument to the CLI. I'll do this in a later PR. I think this is good to go — could you rebase this PR and fix the conflicts so it can be merged? Like this? :)
gharchive/pull-request
2020-02-02T17:22:56
2025-04-01T06:44:26.668344
{ "authors": [ "hishamhm", "pdesaulniers" ], "repo": "hishamhm/tl", "url": "https://github.com/hishamhm/tl/pull/42", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1854652362
Upgrade cryptography to fix vulnerability issues Description https://github.com/histolab/histolab/security/dependabot/24 Types of Changes [ ] Core [ ] Bugfix [ ] New feature [ ] Enhancement/optimization [ ] Documentation Issues Fixed or Closed by This PR Fixes: Checklist [ ] My code follows the code style of this project. [ ] My change requires a change to the documentation. [ ] I have updated the documentation accordingly. [ ] I have read the CONTRIBUTING document. [ ] I have added tests to cover my changes. [ ] I have tested the changes and verified that they work and don't break anything (as well as I can manage). fixed by #633
gharchive/pull-request
2023-08-17T09:45:56
2025-04-01T06:44:26.673103
{ "authors": [ "alessiamarcolini" ], "repo": "histolab/histolab", "url": "https://github.com/histolab/histolab/pull/619", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1816851539
Merge a QLoRA to base Llama 2 I used python src/train_web.py to train a QLoRA. How do i merge the QLoRA to the base model Llama 2? Please visit the export tab of the Web Tuner to merge the LoRA weights. Please visit the export tab of the Web Tuner to merge the LoRA weights. I selected and load QLoRa Checkpoints and base and got error: 07/24/2023 20:59:02 - WARNING - llmtuner.tuner.core.parser - Please specify `prompt_template` if you are using other pre-trained models. [INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,572 >> loading file tokenizer.model [INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file added_tokens.json [INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file special_tokens_map.json [INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file tokenizer_config.json [INFO|configuration_utils.py:710] 2023-07-24 20:59:02,580 >> loading configuration file C:\LLaMA-Efficient-Tuning\Llama-2-13B-Chat-fp16\config.json [INFO|configuration_utils.py:768] 2023-07-24 20:59:02,581 >> Model config LlamaConfig { "_name_or_path": "C:\\LLaMA-Efficient-Tuning\\Llama-2-13B-Chat-fp16", "architectures": [ "LlamaForCausalLM" ], "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_length": 4096, "max_position_embeddings": 4096, "model_type": "llama", "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "tie_word_embeddings": false, "torch_dtype": "float16", "transformers_version": "4.31.0", "use_cache": true, "vocab_size": 32000 } [INFO|modeling_utils.py:2600] 2023-07-24 20:59:02,582 >> loading weights file C:\LLaMA-Efficient-Tuning\Llama-2-13B-Chat-fp16\pytorch_model.bin.index.json [INFO|modeling_utils.py:1172] 2023-07-24 20:59:02,583 >> Instantiating LlamaForCausalLM model under default dtype torch.float16. [INFO|configuration_utils.py:599] 2023-07-24 20:59:02,583 >> Generate config GenerationConfig { "_from_model_config": true, "bos_token_id": 1, "eos_token_id": 2, "max_length": 4096, "pad_token_id": 0, "transformers_version": "4.31.0" } Traceback (most recent call last): File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\routes.py", line 442, in run_predict output = await app.get_blocks().process_api( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1389, in process_api result = await self.call_function( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1108, in call_function prediction = await utils.async_iteration(iterator) File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 346, in async_iteration return await iterator.__anext__() File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 339, in __anext__ return await anyio.to_thread.run_sync( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\to_thread.py", line 33, in run_sync return await get_asynclib().run_sync_in_worker_thread( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 877, in run_sync_in_worker_thread return await future File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 807, in run result = context.run(func, *args) File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 322, in run_sync_iterator_async return next(iterator) File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 691, in gen_wrapper yield from f(*args, **kwargs) File "C:\LLaMA-Efficient-Tuning\src\llmtuner\webui\utils.py", line 122, in export_model model, tokenizer = load_model_and_tokenizer(model_args, finetuning_args) File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\core\loader.py", line 105, in load_model_and_tokenizer model = AutoModelForCausalLM.from_pretrained( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\models\auto\auto_factory.py", line 493, in from_pretrained return model_class.from_pretrained( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\modeling_utils.py", line 2903, in from_pretrained ) = cls._load_pretrained_model( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\modeling_utils.py", line 3002, in _load_pretrained_model raise ValueError( ValueError: The current `device_map` had weights offloaded to the disk. Please provide an `offload_folder` for them. Alternatively, make sure you have `safetensors` installed if the model you are using offers the weights in this format. Please visit the export tab of the Web Tuner to merge the LoRA weights. I selected and load QLoRa Checkpoints and base and got error: 07/24/2023 20:59:02 - WARNING - llmtuner.tuner.core.parser - Please specify `prompt_template` if you are using other pre-trained models. [INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,572 >> loading file tokenizer.model [INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file added_tokens.json [INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file special_tokens_map.json [INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file tokenizer_config.json [INFO|configuration_utils.py:710] 2023-07-24 20:59:02,580 >> loading configuration file C:\LLaMA-Efficient-Tuning\Llama-2-13B-Chat-fp16\config.json [INFO|configuration_utils.py:768] 2023-07-24 20:59:02,581 >> Model config LlamaConfig { "_name_or_path": "C:\\LLaMA-Efficient-Tuning\\Llama-2-13B-Chat-fp16", "architectures": [ "LlamaForCausalLM" ], "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_length": 4096, "max_position_embeddings": 4096, "model_type": "llama", "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "tie_word_embeddings": false, "torch_dtype": "float16", "transformers_version": "4.31.0", "use_cache": true, "vocab_size": 32000 } [INFO|modeling_utils.py:2600] 2023-07-24 20:59:02,582 >> loading weights file C:\LLaMA-Efficient-Tuning\Llama-2-13B-Chat-fp16\pytorch_model.bin.index.json [INFO|modeling_utils.py:1172] 2023-07-24 20:59:02,583 >> Instantiating LlamaForCausalLM model under default dtype torch.float16. [INFO|configuration_utils.py:599] 2023-07-24 20:59:02,583 >> Generate config GenerationConfig { "_from_model_config": true, "bos_token_id": 1, "eos_token_id": 2, "max_length": 4096, "pad_token_id": 0, "transformers_version": "4.31.0" } Traceback (most recent call last): File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\routes.py", line 442, in run_predict output = await app.get_blocks().process_api( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1389, in process_api result = await self.call_function( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1108, in call_function prediction = await utils.async_iteration(iterator) File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 346, in async_iteration return await iterator.__anext__() File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 339, in __anext__ return await anyio.to_thread.run_sync( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\to_thread.py", line 33, in run_sync return await get_asynclib().run_sync_in_worker_thread( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 877, in run_sync_in_worker_thread return await future File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 807, in run result = context.run(func, *args) File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 322, in run_sync_iterator_async return next(iterator) File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 691, in gen_wrapper yield from f(*args, **kwargs) File "C:\LLaMA-Efficient-Tuning\src\llmtuner\webui\utils.py", line 122, in export_model model, tokenizer = load_model_and_tokenizer(model_args, finetuning_args) File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\core\loader.py", line 105, in load_model_and_tokenizer model = AutoModelForCausalLM.from_pretrained( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\models\auto\auto_factory.py", line 493, in from_pretrained return model_class.from_pretrained( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\modeling_utils.py", line 2903, in from_pretrained ) = cls._load_pretrained_model( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\modeling_utils.py", line 3002, in _load_pretrained_model raise ValueError( ValueError: The current `device_map` had weights offloaded to the disk. Please provide an `offload_folder` for them. Alternatively, make sure you have `safetensors` installed if the model you are using offers the weights in this format. Try removing the following lines and checking if it works. https://github.com/hiyouga/LLaMA-Efficient-Tuning/blob/182b42504399d2755897b9737db1d36655a0fa50/src/llmtuner/tuner/core/loader.py#L96-L97 如何启用QLoRA?在量化处选择4bit就可以吗? 如何启用QLoRA?在量化处选择4bit就可以吗? 是的 In our experiments, the performance of QLoRA is close to LoRA's one. In addition of WeChat, possibly add public discord servers? @hiyouga after update to last version and use export tab with a selected Prompt template, got error: Traceback (most recent call last): File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\routes.py", line 442, in run_predict output = await app.get_blocks().process_api( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1389, in process_api result = await self.call_function( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1108, in call_function prediction = await utils.async_iteration(iterator) File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 346, in async_iteration return await iterator.__anext__() File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 339, in __anext__ return await anyio.to_thread.run_sync( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\to_thread.py", line 33, in run_sync return await get_asynclib().run_sync_in_worker_thread( File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 877, in run_sync_in_worker_thread return await future File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 807, in run result = context.run(func, *args) File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 322, in run_sync_iterator_async return next(iterator) File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 691, in gen_wrapper yield from f(*args, **kwargs) File "C:\LLaMA-Efficient-Tuning\src\llmtuner\webui\utils.py", line 125, in save_model export_model(args, max_shard_size="{}GB".format(max_shard_size)) File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\tune.py", line 29, in export_model model_args, _, training_args, finetuning_args, _ = get_train_args(args) File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\core\parser.py", line 54, in get_train_args model_args, data_args, training_args, finetuning_args, general_args = parse_train_args(args) File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\core\parser.py", line 39, in parse_train_args return _parse_args(parser, args) File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\core\parser.py", line 24, in _parse_args return parser.parse_dict(args) File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\hf_argparser.py", line 373, in parse_dict obj = dtype(**inputs) TypeError: DataArguments.__init__() missing 1 required positional argument: 'template' @Katehuuh Fixed
gharchive/issue
2023-07-22T18:14:13
2025-04-01T06:44:26.693770
{ "authors": [ "DumoeDss", "Katehuuh", "PsychoSmiley", "hiyouga" ], "repo": "hiyouga/LLaMA-Efficient-Tuning", "url": "https://github.com/hiyouga/LLaMA-Efficient-Tuning/issues/223", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2197854829
🛑 HJStrauss is down In 92a7383, HJStrauss (https://www.hjstrauss.de) was down: HTTP code: 0 Response time: 0 ms Resolved: HJStrauss is back up in cfd8a72 after 9 minutes.
gharchive/issue
2024-03-20T15:36:19
2025-04-01T06:44:26.719872
{ "authors": [ "hjstrauss" ], "repo": "hjstrauss/MonitorMySites", "url": "https://github.com/hjstrauss/MonitorMySites/issues/237", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2286779397
🛑 Reime ohne Sinn und Verstand is down In 0ba04e4, Reime ohne Sinn und Verstand (https://www.reimeohnesinnundverstand.de) was down: HTTP code: 0 Response time: 0 ms Resolved: Reime ohne Sinn und Verstand is back up in f1d673a after 9 minutes.
gharchive/issue
2024-05-09T02:34:37
2025-04-01T06:44:26.722465
{ "authors": [ "hjstrauss" ], "repo": "hjstrauss/MonitorMySites", "url": "https://github.com/hjstrauss/MonitorMySites/issues/3193", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2190771210
🛑 LM Bogen WSV is down In 9974047, LM Bogen WSV (https://www.lmbogenwsv.de) was down: HTTP code: 0 Response time: 0 ms Resolved: LM Bogen WSV is back up in 5e1829d after 8 minutes.
gharchive/issue
2024-03-17T17:17:57
2025-04-01T06:44:26.724929
{ "authors": [ "hjstrauss" ], "repo": "hjstrauss/MonitorMySites", "url": "https://github.com/hjstrauss/MonitorMySites/issues/48", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2382260447
🛑 Reime ohne Sinn und Verstand is down In 61035ef, Reime ohne Sinn und Verstand (https://www.reimeohnesinnundverstand.de) was down: HTTP code: 0 Response time: 0 ms Resolved: Reime ohne Sinn und Verstand is back up in 4d411a4 after 8 minutes.
gharchive/issue
2024-06-30T13:40:42
2025-04-01T06:44:26.727351
{ "authors": [ "hjstrauss" ], "repo": "hjstrauss/MonitorMySites", "url": "https://github.com/hjstrauss/MonitorMySites/issues/5975", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1038456956
API available FYI, Fuji released a SDK in April 2021 https://fujifilm-x.com/global/special/camera-control-sdk/ hmph
gharchive/issue
2021-10-28T12:27:37
2025-04-01T06:44:26.747560
{ "authors": [ "sahver", "tensiondriven" ], "repo": "hkr/fuji-cam-wifi-tool", "url": "https://github.com/hkr/fuji-cam-wifi-tool/issues/20", "license": "unlicense", "license_type": "permissive", "license_source": "bigquery" }
738299072
fixed send_line function closes #172 Thanks a lot for the PR and for helping keep iron working :slightly_smiling_face: Best regards, Henry
gharchive/pull-request
2020-11-07T18:55:44
2025-04-01T06:44:26.752533
{ "authors": [ "doubleloop", "hkupty" ], "repo": "hkupty/iron.nvim", "url": "https://github.com/hkupty/iron.nvim/pull/173", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
435354071
error build with compile-private ➜ .emacs.d (develop) ✗ make compile-private (develop|✚2… Warning: byte compiling is for advanced users. It will interfere with your efforts to debug issues. It is not recommended you do it if you frequently tinker with your Emacs config. Alternatively, use `bin/doom compile :core` instead to byte-compile only the Doom core files, as these don't change often. If you have issues, please make sure byte-compilation isn't the cause by using `bin/doom clean` to clear out your *.elc files. Byte-compile anyway? (y or n) y Note, built-in variable ‘overflow-newline-into-fringe’ not bound ⚠ Ignored ~/.config/doom/packages.el ⚠ Ignored ~/.config/doom/lang.el ✓ Compiled ~/.config/doom/init.el simple profile: ON Eager macro-expansion failure: (error "featurep! couldn’t detect what module its in! (in /Users/jiya/.config/doom/config.el)") [Treemacs] Warning: coudn’t find default background color for icons, falling back on #2d2d31. [Treemacs] Warning: couldn’t find hl-line-mode’s background color for icons, falling back on #2d2d31. ✓ Compiled ~/.config/doom/config.el get an error message Eager macro-expansion failure: (error "featurep! couldn’t detect what module its in! (in /Users/jiya/.config/doom/config.el)"). Can you show me how you are using featurep! in your config? ./lang.el:31:;; (when (featurep! :ui doom) ./+ui.el:7:(when (featurep! :ui doom) ./+ui.el:115: (if (featurep! :ui doom) ./+ui.el:368:;; (unless (featurep! :ui doom-quit) ./config.el:152: :when (featurep! :tools flycheck) This error should only occur if you use featurep! with only one argument; which none of those are doing, so it's unclear to me why this would happen. Can you remove them one by one and see which one is causing it? remove all featurep! but still error. and in my config featurep! like (and (featurep! +lsp) (not (featurep! +tabnine)) (and (featurep! +lsp) (not (featurep! +tabnine)) This is the kind of code I was expecting. It is incorrect to use featurep! this way from your private config. You must use the full module path, e.g. (and (featurep! :tools lsp) (not (featurep! :completion company +tabnine)) Otherwise, it doesn't know where to look for the +lsp or +tabnine flags. I will assume this issue has been resolved. If that isn't the case, feel free to follow up here.
gharchive/issue
2019-04-20T04:02:54
2025-04-01T06:44:26.767936
{ "authors": [ "hlissner", "shuxiao9058" ], "repo": "hlissner/doom-emacs", "url": "https://github.com/hlissner/doom-emacs/issues/1346", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
779814932
Our Github issue tracker is temporarily locked down Hey folks! You may have noticed that our Github issue tracker is locked down; only previous contributors can post new issues/PRs or respond to old ones. This is temporary and will elapse in 7 days, while I clean up our issue tracker and transition to Discourse. As some of you know, I'm working on a Discourse server for Doom Emacs and I have prepared for it to go live some time next week. It will replace Discord as our primary platform for user support and replace Github as our issue tracker. There will be a post on the Discourse explaining the rationale for this change and what it means for our users, but going forward, our Github issue tracker will be reserved strictly for internal project management and issues escalated from the Discourse. You may have also noticed that two of our project boards have been closed. They were also moved to the Discourse. I realize this is an inconvenience for folks with issues to report, but please bare with me for a short while. I'll keep you posted! The issue tracker is open once again. I've collected what I needed from it to move to Discourse and will soon publish it; when I do I'll begin diverting traffic there. The issue tracker is open once again. I've collected what I needed from it to move to Discourse and will soon publish it; when I do I'll begin diverting traffic there.
gharchive/issue
2021-01-06T00:59:11
2025-04-01T06:44:26.770954
{ "authors": [ "hlissner" ], "repo": "hlissner/doom-emacs", "url": "https://github.com/hlissner/doom-emacs/issues/4510", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
797392276
Remap describe-symbol to counsel-describe-symbol . "SPC h o" now uses counsel-describe-symbol instead of plain helpful-symbol (maybe with completion) to get input. This allows us to fully take the advantage of ivy's superior completion system: [I]: info and [d]: definition hydra heads in ivy-dispatching-done Quickly describe symbols with "C-M-n/p" (ivy-next(or previous)-line-and-call or with the calling hydra head on Additionally the counsel one is prettier when icons are enabled. Users who have the +icons flag with ivy are expected to notice the change visually when invoking "o" in help-map as icons now appear there. But it's an arguably good change anyway since originally "f1 f" and "f1 v" have icons but "f1 o" doesn't. Excuse me, I didn't know that helpful-symbol asks an additional question when the symbol refers to both a command and a variable. Guess I have to personally stick to the plain old describe-symbol instead to use the Counsel version since "C-M-n/p" seems far too useful to me. Excuse me, I didn't know that helpful-symbol asks an additional question when the symbol refers to both a command and a variable. Guess I have to personally stick to the plain old describe-symbol instead to use the Counsel version since "C-M-n/p" seems far too useful to me. Thanks! Thanks!
gharchive/pull-request
2021-01-30T11:04:53
2025-04-01T06:44:26.775971
{ "authors": [ "datool", "hlissner" ], "repo": "hlissner/doom-emacs", "url": "https://github.com/hlissner/doom-emacs/pull/4593", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1644330956
Add Community Listing Block Already resolved by #118
gharchive/issue
2023-03-28T16:50:33
2025-04-01T06:44:26.777115
{ "authors": [ "badvision", "bstopp" ], "repo": "hlxsites/hsf-commonmoves", "url": "https://github.com/hlxsites/hsf-commonmoves/issues/23", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1704529248
Fixed aria attributes and link crawling Issue Fixed Part all images that are linked, that don't have linked text or an alt tag, should have aria-label added to describe the function of the link Check for links that are not crawlable, such as Close: Test URLs Before: https://main--merative2--hlxsites.hlx.page/company After: https://aria-attributes--merative2--hlxsites.hlx.page/company Description Is there a GitHub issue this PR can be linked to?
gharchive/pull-request
2023-05-10T19:43:01
2025-04-01T06:44:26.780043
{ "authors": [ "VladimirZaets", "arumsey" ], "repo": "hlxsites/merative2", "url": "https://github.com/hlxsites/merative2/pull/134", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
953772324
Pay 4882 testing Before creating a pull request make sure that: [ ] commit messages are meaningful and follow good commit message guidelines [ ] README and other documentation has been updated / added (if needed) [ ] tests have been updated / new tests has been added (if needed) Please remove this line and everything above and fill the following sections: JIRA link (if applicable) Change description Does this PR introduce a breaking change? (check one with "x") [ ] Yes [ ] No Closing Pull as we have some merge issues in the pipeline....
gharchive/pull-request
2021-07-27T11:13:17
2025-04-01T06:44:26.954207
{ "authors": [ "pats-john" ], "repo": "hmcts/ccpay-refunds-app", "url": "https://github.com/hmcts/ccpay-refunds-app/pull/19", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2413931308
CIV-14290 added publish instructions so WA JIRA link (if applicable) CIV-14290 Change description CCD changes for WA part Does this PR introduce a breaking change? (check one with "x") [ ] Yes [X] No Is Create_SDO_DJ missing here? Definitely right, STANDARD_DIRECTION_ORDER_DJ was missing
gharchive/pull-request
2024-07-17T15:31:11
2025-04-01T06:44:26.955365
{ "authors": [ "dtortolaV1" ], "repo": "hmcts/civil-ccd-definition", "url": "https://github.com/hmcts/civil-ccd-definition/pull/4639", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2383305276
CIV-13828 remove all notifications when claim is struck out JIRA link (if applicable) CIV-13828 Change description Before adding the new notifications, allow "*" to remove all existing notifications from both Claimant and Defendant Does this PR introduce a breaking change? (check one with "x") [ ] Yes [X] No Replaced by https://github.com/hmcts/civil-service/pull/5051
gharchive/pull-request
2024-07-01T09:20:07
2025-04-01T06:44:26.956623
{ "authors": [ "dtortolaV1" ], "repo": "hmcts/civil-service", "url": "https://github.com/hmcts/civil-service/pull/5015", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2488783629
DTSSTCI1086 - removing E2E's and all references Change description The old E2E tests are causing a repository failure - these need to be removed from Case-API as they have been replaced by the separate E2E repository. JIRA link DTSSTCI-1086 Before merging a pull request make sure that: [x] tests have been updated / new tests has been added (if needed) [x] README and other documentation has been updated / added (if needed) [x] enable-e2e-tests label can be used to run the e2e tests before QA handover and before release (required) If this ticket will have any visible impact on users and is not behind a feature toggle, make sure that: [x] this ticket has been reviewed by QA [x] the user story has been signed off by the PO Note that bug fixes, dependency updates and technical tasks do not directly impact the user experience and can be merged without QA and PO review. If this user story cannot be immediately merged find a way to put it behind a feature toggle and get it merged. Nightly-dev pipeline passing
gharchive/pull-request
2024-08-27T09:28:29
2025-04-01T06:44:26.958997
{ "authors": [ "JakeCG" ], "repo": "hmcts/sptribs-case-api", "url": "https://github.com/hmcts/sptribs-case-api/pull/1886", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
497890119
Re-enable --shell-escape for LaTeX Similar to f7bb61b5166e, but for main wscript. Was there any reason to disable it that I cannot remember? Leaving it on is crucial for the Stata template because the only thing Stata consistently produces are eps-files and they need conversion to pdf before being useful to pdflatex. we disabled it for azure, is it an option to save the stuff as png instead of eps, or directly as pdf (do we need shell escape then?) No, at least older versions of Stata relied on some library to generate pngs that was only available as GUI... Maybe they took the graph they displayed and made a screenshot :-) So we should leave the comment where it is unless we are using the Stata example. i am pretty sure, stata 13 at least already supports png, do we really need to care for older versions? Yes, but to my best recollection only the GUI version of Stata, not the command line version, which is what we need. fixed in #65 Just for reference: https://www.stata.com/statalist/archive/2011-07/msg00536.html
gharchive/issue
2019-09-24T19:49:52
2025-04-01T06:44:26.960807
{ "authors": [ "hmgaudecker", "raholler" ], "repo": "hmgaudecker/econ-project-templates", "url": "https://github.com/hmgaudecker/econ-project-templates/issues/64", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
730108621
pdfcrop --xetex does not give correct output Here is the code to reproduce the issue. generate a pdf page (named as abc.pdf) from the following code: \documentclass{standalone} \begin{document} \Huge A sample text. \end{document} use pdfcrop to remove its margins (actually it does not have white margins) pdfcrop --xetex abc.pdf you will see the output have two pages, and both are empty. If replacing standalone with article, the output also have two pages, but the first page is correct, and the second page is also empty. pdfcrop --pdftex abc.pdf pdfcrop --luatex abc.pdf both could give correct results. I have run that on my Windows 10 system with TeXLive 2020 full installed and updated. The pdfcrop version is PDFCROP 1.40, 2020/06/06. Best, Jinsong Well the second page is because of the special for the pdf version. It can probably be moved into the box. The text seems to disappear with standalone as the paper size is too small now: with the 1in displacement inserted by the code nothing is in the box, and then it gets dropped. This is perhaps a xetex bug but imho nothing pdfcrop can solve. With the example file: \documentclass{standalone} \begin{document} \Huge A sample text. \end{document} and running: $ xelatex test.tex $ pdfcrop --xetexcmd xelatex test.pdf I get the correct output file: One interesting thing I have noticed here is the following. When running pdfcrop --xetex test.pdf the line executed by pdfcrop is as follows: * XeTeX call: xetex -no-shell-escape -interaction=nonstopmode tmp-pdfcrop-9248 i.e., it uses plain XeTeX, which has an effect on the fonts of the file: $ pdffonts test.pdf name type encoding emb sub uni object ID ------------------------------------ ----------------- ---------------- --- --- --- --------- SYMVRB+LMRoman17-Regular-Identity-H CID Type 0C Identity-H yes yes yes 4 0 $ pdffonts test.pdf name type encoding emb sub uni object ID ------------------------------------ ----------------- ---------------- --- --- --- --------- SYMVRB+LMRoman17-Regular-Identity-H CID Type 0C Identity-H yes yes yes 4 0 $ pdfcrop --xetex test.pdf PDFCROP 1.40, 2020/06/06 - Copyright (c) 2002-2020 by Heiko Oberdiek, Oberdiek Package Support Group. ==> 1 page written on `test-crop.pdf'. $ pdffonts test-crop.pdf name type encoding emb sub uni object ID ------------------------------------ ----------------- ---------------- --- --- --- --------- SYMVRB+LMRoman17-Regular-Identity-H CID Type 0C Identity-H yes yes yes 4 0 DLSGEY+CMR10 Type 1C Builtin yes yes yes 16 0 Add the font CMR10 and that affects the generated pdf-crop file, but when executing the line: $ pdffonts test.pdf name type encoding emb sub uni object ID ------------------------------------ ----------------- ---------------- --- --- --- --------- SYMVRB+LMRoman17-Regular-Identity-H CID Type 0C Identity-H yes yes yes 4 0 $ pdfcrop --xetexcmd xelatex test.pdf PDFCROP 1.40, 2020/06/06 - Copyright (c) 2002-2020 by Heiko Oberdiek, Oberdiek Package Support Group. ==> 1 page written on `test-crop.pdf'. $ pdffonts test-crop.pdf name type encoding emb sub uni object ID ------------------------------------ ----------------- ---------------- --- --- --- --------- SYMVRB+LMRoman17-Regular-Identity-H CID Type 0C Identity-H yes yes yes 8 0 everything works OK. I hope this is helpful. Saludos fixed in v1.41
gharchive/issue
2020-10-27T05:01:05
2025-04-01T06:44:27.002844
{ "authors": [ "davidcarlisle", "jszhao", "pablgonz", "u-fischer" ], "repo": "ho-tex/pdfcrop", "url": "https://github.com/ho-tex/pdfcrop/issues/3", "license": "LPPL-1.3c", "license_type": "permissive", "license_source": "github-api" }
345034483
Beispiel-Flow überwachung Syslog erstellen /var/log/messages überwachen und Häufigkeit von ReGa-Fehlermeldungen ermitteln, bei überschreiten einer Häufigkeit x per Email benachrichtigen Ich weiß nicht was dir vorschwebt, aber so könnte es aussehen: [{"id":"50104e94.8152b8","type":"tab","label":"Error Monitoring","disabled":false,"info":""},{"id":"5d48d65c.e24b48","type":"tail","z":"50104e94.8152b8","name":"","filetype":"text","split":true,"filename":"/var/log/messages","x":130,"y":160,"wires":[["6595e35d.530f24"]]},{"id":"6595e35d.530f24","type":"switch","z":"50104e94.8152b8","name":"Filter Errors","property":"payload","propertyType":"msg","rules":[{"t":"regex","v":"^.*Error\\:.*$","vt":"str","case":false}],"checkall":"true","repair":false,"outputs":1,"x":330,"y":160,"wires":[["e9d593c5.3deb1"]]},{"id":"e9d593c5.3deb1","type":"counter","z":"50104e94.8152b8","name":"Count Errors","init":"0","step":"1","lower":"","upper":"","mode":"increment","outputs":2,"x":510,"y":160,"wires":[["253f1e33.ae4aea"],["ba691feb.d30d08","bd901793.899028"]]},{"id":"253f1e33.ae4aea","type":"ui_gauge","z":"50104e94.8152b8","name":"Show Error Count","group":"d79a8eea.a905c","order":0,"width":"6","height":"4","gtype":"gage","title":"Anzahl Fehler","label":"Fehler","format":"{{value}}","min":0,"max":"100","colors":["#00b500","#e6e600","#ca3838"],"seg1":"","seg2":"","x":710,"y":100,"wires":[]},{"id":"ba691feb.d30d08","type":"function","z":"50104e94.8152b8","name":"Rotate Entries","func":"var dashboardLog = context.get('dashboardLog')|| [];\n \ndashboardLog.push(msg);\nif (dashboardLog.length > 20) {\n // Delete oldest message if &gt; 20\n dashboardLog.shift();\n dashboardLog.length = 20;\n} \n \n// store the value back\ncontext.set('dashboardLog',dashboardLog);\n \n// make it part of the outgoing msg object\nmsg = {};\nmsg.payload = dashboardLog;\nreturn msg;","outputs":1,"noerr":0,"x":700,"y":220,"wires":[["f4ef2a5d.f01c18"]]},{"id":"bd901793.899028","type":"join","z":"50104e94.8152b8","name":"Prepare Message","mode":"custom","build":"string","property":"payload","propertyType":"msg","key":"topic","joiner":"\\r\\n\\r\\n","joinerType":"str","accumulate":false,"timeout":"","count":"5","reduceRight":false,"reduceExp":"","reduceInit":"","reduceInitType":"num","reduceFixup":"","x":710,"y":160,"wires":[["2e0650.3ce4e9b"]]},{"id":"f4ef2a5d.f01c18","type":"ui_template","z":"50104e94.8152b8","group":"2bf72e51.a156aa","name":"Event Log","order":0,"width":"6","height":"4","format":"<ul>\n <li ng-repeat=\"x in msg.payload\">\n <font color=\"red\">{{x.topic}}</font>\n <ul>\n <li>{{x.payload}}</li>\n </ul>\n </li>\n</ul>","storeOutMessages":true,"fwdInMessages":true,"templateScope":"local","x":900,"y":220,"wires":[[]]},{"id":"2e0650.3ce4e9b","type":"e-mail","z":"50104e94.8152b8","server":"smtp.gmail.com","port":"465","secure":true,"name":"","dname":"Send Email","x":910,"y":160,"wires":[]},{"id":"b8d3f654.c03e9","type":"comment","z":"50104e94.8152b8","name":"Fehlerüberwachung","info":"Überwacht /var/log/messages\nZeigt die letzten 20 Fehler in einem Log-Fenster\nZeigt die Gesamtzahl der Fehler\nSchickt Email nach 5 Fehlern mit Fehlerprotokoll","x":130,"y":80,"wires":[]},{"id":"d79a8eea.a905c","type":"ui_group","z":"","name":"Monitoring","tab":"a5b5c3f2.ffeeb8","disp":true,"width":"6","collapse":false},{"id":"2bf72e51.a156aa","type":"ui_group","z":"","name":"EventLog","tab":"a5b5c3f2.ffeeb8","disp":true,"width":"6","collapse":false},{"id":"a5b5c3f2.ffeeb8","type":"ui_tab","z":"","name":"Monitoring","icon":"dashboard"}] Zusätzliche Node: https://www.npmjs.com/package/node-red-contrib-counter Hey vielen Dank! Werde bei Gelegenheit ein bebilderten kommentierten Wiki Artikel draus machen, außer Du kommst mir zuvor ;-) Kann ich gerne übernehmen. Deine Zeit ist besser in der Entwicklung investiert ;-) https://github.com/hobbyquaker/RedMatic/wiki/Fehlerüberwachung-der-CCU Du müsstest mir bitte noch den Screenshot Ordner hochladen. errorlog.zip
gharchive/issue
2018-07-26T23:03:14
2025-04-01T06:44:27.024585
{ "authors": [ "Sineos", "hobbyquaker" ], "repo": "hobbyquaker/RedMatic", "url": "https://github.com/hobbyquaker/RedMatic/issues/44", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2367363669
Add in pull request guidelines and fix 'stolen'/copied code. Hi @hoeken I'm concerned that nothing has been done in regards to pull request #89 I spent a lot of time researching, writing, and documenting my TemplatePrinter and the pull request was ready to merge for months. The user who submitted #89 copied my code, not merged my commits, and then submitted my work in their own commits, which means that there is no link to the actual author of this work. I create and submit code for open source projects to benefit others, and more importantly build a portfolio of my work that other projects and future employers can view. By allowing another user to submit my work, I have no proof that I wrote that code, and others cannot find any reference to it from my profile. I really must ask that you do what is right and reverse that change and submit my PR (#56) if you want the TemplatePrinter in your project. As my code is unrelated to #89 it should not have been included there. That PR should be reverted and they should resubmit their own work only. This will not only make your repo cleaner, but people will be able to find the PR and justification behind its implementation. I also recommend you add in some guidelines to prevent this from happening in the future. Hi @Chris--A I agree 100% and I'm sorry I messed this one up. This one is definitely my fault. I was just going through all the PRs and trying to clear the queue out. I saw that the #89 pr included a bunch of others and my brain said that it was the easiest way. It didn't cross my mind that you would want the original commit merged and that was an oversight. Last week and this week are both pretty hectic for me, but as soon as I get some free time I will unwind this and merge everything in properly. Your contributions are very much appreciated. It wasn't an intentional slight or anything so hopefully no hard feelings. :) @Chris--A I re-opened and merged your code properly + added attribution in the upcoming changelog. @hoeken : it is also possible to activate this option to force contributors to accept that the PR they submit is in accordance with the projet's license: When contributing a PR in an LGPL project, attribution is not a "right" because the work should have been submitted under the project's license. Most projects are using SPDX generated headers so there is no attribution on any file / header but just the SPDX header. Submitting a PR with a work requiring attribution would mean to change the license of this specific file only, and depending on the project this can be an issue, for example someone might want to submit PR with some GPL additions then preventing the library in LGPL to be used commercially. That's why contributors should / have to accept to PR their work under a projet's license, and often, attribution is not automatic. This is very nice from @hoeken to maintain such list. yeah that's a good idea.
gharchive/issue
2024-06-21T23:46:09
2025-04-01T06:44:27.041316
{ "authors": [ "Chris--A", "hoeken", "mathieucarbou" ], "repo": "hoeken/PsychicHttp", "url": "https://github.com/hoeken/PsychicHttp/issues/100", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
164577143
Build Failures $ perl6 -v This is Rakudo version 2016.06-10-g7cff429 built on MoarVM version 2016.06 implementing Perl 6.c. $ zef install Linenoise ===> Searching for: Linenoise ===> Searching for missing dependencies: Native::Resources ===> Fetching: Linenoise ===> Fetching: Native::Resources ===> Building: Linenoise:ver('0.1.0'):auth('Rob Hoelz') `build-depends` is missing entries. Attemping to mimick missing dependencies... Cannot invoke this object (REPR: Null; VMNull) in sub get-vars at /home/zoffix/.rakudobrew/moar-nom/install/share/perl6/site/sources/D57EF38F5287179FB7D408925E155FB671060388 (LibraryMake) line 95 in sub get-vars at /home/zoffix/.zef/store/p6-native-resources.git/b05fd025d5a3541e88b5acc266446d3814d79cc0/lib/Native/Resources/Build.pm (Native::Resources::Build) line 13 in sub make at /home/zoffix/.zef/store/p6-native-resources.git/b05fd025d5a3541e88b5acc266446d3814d79cc0/lib/Native/Resources/Build.pm (Native::Resources::Build) line 27 in method build at /home/zoffix/.zef/store/p6-linenoise.git/46f9534b1568e68a2a077ee80713914738b16e28/Build.pm (Build) line 13 in block <unit> at -e line 1 ===> Building [FAIL]: Linenoise:ver('0.1.0'):auth('Rob Hoelz') ===SORRY!=== Aborting due to build failure: Linenoise:ver('0.1.0'):auth('Rob Hoelz')(use --force to override) @zoffixznet Would you mind trying this again? I just tried installing Linenoise with both panda and zef, and both worked for me. Tried it on a fresh build and everything worked fine. ¯\_(ツ)_/¯ maybe it's just my other box is busted. cpan@perlbuild2:~$ zef install Linenoise ===> Searching for: Linenoise ===> Found: Linenoise:ver('0.1.0'):auth('Rob Hoelz') ===> Dependencies: Native::Resources ===> Searching for missing dependencies: Native::Resources ===> Found dependencies: Native::Resources:ver('0.1.0'):auth('Rob Hoelz') ===> Dependencies: LibraryMake ===> Searching for missing dependencies: LibraryMake ===> Found dependencies: LibraryMake:ver('1.0.0'):auth('github:retupmoca') ===> Dependencies: Shell::Command ===> Searching for missing dependencies: Shell::Command ===> Found dependencies: Shell::Command ===> Dependencies: File::Which, File::Find, Test ===> Searching for missing dependencies: File::Which, File::Find ===> Found dependencies: File::Find:ver('0.1'), File::Which ===> Fetching: Linenoise Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Fetched: Linenoise to /home/cpan/.zef/tmp/p6-linenoise.git ===> Extracting: Linenoise Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Extracted: Linenoise to /home/cpan/.zef/store/p6-linenoise.git/46f9534b1568e68a2a077ee80713914738b16e28 ===> Fetching: Native::Resources Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Fetched: Native::Resources to /home/cpan/.zef/tmp/p6-native-resources.git ===> Extracting: Native::Resources Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Extracted: Native::Resources to /home/cpan/.zef/store/p6-native-resources.git/b05fd025d5a3541e88b5acc266446d3814d79cc0 ===> Fetching: LibraryMake Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Fetched: LibraryMake to /home/cpan/.zef/tmp/P6-LibraryMake.git ===> Extracting: LibraryMake Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Extracted: LibraryMake to /home/cpan/.zef/store/P6-LibraryMake.git/8ab4b23a981ebfd63a74c11d1398750a472195aa ===> Fetching: Shell::Command Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Fetched: Shell::Command to /home/cpan/.zef/tmp/Shell-Command.git ===> Extracting: Shell::Command Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Extracted: Shell::Command to /home/cpan/.zef/store/Shell-Command.git/3d21c0ffbc467f121935a3ec74e54c23b651836e ===> Fetching: File::Find Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Fetched: File::Find to /home/cpan/.zef/tmp/File-Find.git ===> Extracting: File::Find Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Extracted: File::Find to /home/cpan/.zef/store/File-Find.git/d3e2be709aeb2479dc02b154fdc179fe7a077d83 ===> Fetching: File::Which Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Fetched: File::Which to /home/cpan/.zef/tmp/perl6-file-which.git ===> Extracting: File::Which Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>} ===> Extracted: File::Which to /home/cpan/.zef/store/perl6-file-which.git/d1d9dbef35751e2383bec4ef6d3fb9aa2146eacf ===> Filtering: Linenoise:ver('0.1.0'):auth('Rob Hoelz') ===> Filtering [OK] for Linenoise:ver('0.1.0'):auth('Rob Hoelz') ===> Filtering: Native::Resources:ver('0.1.0'):auth('Rob Hoelz') ===> Filtering [OK] for Native::Resources:ver('0.1.0'):auth('Rob Hoelz') ===> Filtering: LibraryMake:ver('1.0.0'):auth('github:retupmoca') ===> Filtering [OK] for LibraryMake:ver('1.0.0'):auth('github:retupmoca') ===> Filtering: Shell::Command ===> Filtering [OK] for Shell::Command ===> Filtering: File::Find:ver('0.1') ===> Filtering [OK] for File::Find:ver('0.1') ===> Filtering: File::Which ===> Filtering [OK] for File::Which ===> # SKIP: No Build.pm for File::Which ===> # SKIP: No Build.pm for File::Find:ver('0.1') ===> # SKIP: No Build.pm for Shell::Command ===> # SKIP: No Build.pm for LibraryMake:ver('1.0.0'):auth('github:retupmoca') ===> # SKIP: No Build.pm for Native::Resources:ver('0.1.0'):auth('Rob Hoelz') ===> Building: Linenoise:ver('0.1.0'):auth('Rob Hoelz') `build-depends` is missing entries. Attemping to mimick missing dependencies... Command: /home/cpan/.rakudobrew/bin/../moar-nom/install/bin/perl6 -Ilib -I. -MBuild -I/home/cpan/.zef/store/p6-native-resources.git/b05fd025d5a3541e88b5acc266446d3814d79cc0/lib -I/home/cpan/.zef/store/P6-LibraryMake.git/8ab4b23a981ebfd63a74c11d1398750a472195aa/lib -I/home/cpan/.zef/store/Shell-Command.git/3d21c0ffbc467f121935a3ec74e54c23b651836e/lib -I/home/cpan/.zef/store/perl6-file-which.git/d1d9dbef35751e2383bec4ef6d3fb9aa2146eacf/lib -I/home/cpan/.zef/store/File-Find.git/d3e2be709aeb2479dc02b154fdc179fe7a077d83/lib -e ::('Build').new.build('/home/cpan/.zef/store/p6-linenoise.git/46f9534b1568e68a2a077ee80713914738b16e28'); exit(0); gcc -c -fPIC -Wdeclaration-after-statement -Werror=declaration-after-statement -O3 -DNDEBUG -D_REENTRANT -D_FILE_OFFSET_BITS=64 -fPIC -o linenoise.o linenoise.c perl6 -e "mkdir 'resources'; mkdir 'resources/lib'" gcc -shared -fPIC -O3 -DNDEBUG -Wl,-rpath,"//home/cpan/.rakudobrew/moar-nom/install/lib" -Wl,-rpath,"/home/cpan/.rakudobrew/moar-nom/install/share/perl6/site/lib" -lm -lpthread -lrt -ldl -o resources/lib/liblinenoise.so linenoise.o perl6 -e "print ''" > resources/lib/liblinenoise.dll perl6 -e "print ''" > resources/lib/liblinenoise.dylib gcc -o constant-helper -Wdeclaration-after-statement -Werror=declaration-after-statement -O3 -DNDEBUG -D_REENTRANT -D_FILE_OFFSET_BITS=64 -fPIC constant-helper.c perl6 fill-constants.pl < lib/Linenoise.pm.in > lib/Linenoise.pm ===> Building [OK] for Linenoise:ver('0.1.0'):auth('Rob Hoelz') ===> Testing: File::Which Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>} t/01-which.t ... ok t/02-win32.t ... ok t/03-export.t .. ok t/04-simple.t .. ok t/05-all.t ..... ok All tests successful. Files=5, Tests=29, 9 wallclock secs ( 0.02 usr 0.00 sys + 8.29 cusr 0.44 csys = 8.75 CPU) Result: PASS # Found perl6 at '/home/cpan/.rakudobrew/bin/perl6' # Found perl6 at '/home/cpan/.rakudobrew/bin/perl6' using whence ===> Testing [OK] for File::Which ===> Testing: File::Find:ver('0.1') Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>} t/01-file-find.t .. ok All tests successful. Files=1, Tests=11, 1 wallclock secs ( 0.02 usr 0.00 sys + 0.93 cusr 0.05 csys = 1.00 CPU) Result: PASS ===> Testing [OK] for File::Find:ver('0.1') ===> Testing: Shell::Command Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>} t/02-shell-command.t .. ok All tests successful. Files=1, Tests=16, 1 wallclock secs ( 0.02 usr 0.00 sys + 1.35 cusr 0.11 csys = 1.48 CPU) Result: PASS ===> Testing [OK] for Shell::Command ===> Testing: LibraryMake:ver('1.0.0'):auth('github:retupmoca') Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>} t/01-basic.t .. ok All tests successful. Files=1, Tests=7, 2 wallclock secs ( 0.01 usr 0.00 sys + 1.78 cusr 0.14 csys = 1.93 CPU) Result: PASS ===> Testing [OK] for LibraryMake:ver('1.0.0'):auth('github:retupmoca') ===> Testing: Native::Resources:ver('0.1.0'):auth('Rob Hoelz') Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>} ===> Testing [OK] for Native::Resources:ver('0.1.0'):auth('Rob Hoelz') ===> Testing: Linenoise:ver('0.1.0'):auth('Rob Hoelz') Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>} ===> Testing [OK] for Linenoise:ver('0.1.0'):auth('Rob Hoelz') ===> Installing: File::Which ===> Install [OK] for File::Which ===> Installing: File::Find:ver('0.1') ===> Install [OK] for File::Find:ver('0.1') ===> Installing: Shell::Command ===> Install [OK] for Shell::Command ===> Installing: LibraryMake:ver('1.0.0'):auth('github:retupmoca') ===> Install [OK] for LibraryMake:ver('1.0.0'):auth('github:retupmoca') ===> Installing: Native::Resources:ver('0.1.0'):auth('Rob Hoelz') ===> Install [OK] for Native::Resources:ver('0.1.0'):auth('Rob Hoelz') ===> Installing: Linenoise:ver('0.1.0'):auth('Rob Hoelz') ===> Install [OK] for Linenoise:ver('0.1.0'):auth('Rob Hoelz') cpan@perlbuild2:~$
gharchive/issue
2016-07-08T17:22:50
2025-04-01T06:44:27.051234
{ "authors": [ "hoelzro", "zoffixznet" ], "repo": "hoelzro/p6-linenoise", "url": "https://github.com/hoelzro/p6-linenoise/issues/15", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1678839731
Consider handling toRow/fromRow We'd rewrite to createSerializer and createDeserializer. Fixed :)
gharchive/issue
2023-04-21T17:10:10
2025-04-01T06:44:27.063049
{ "authors": [ "holdenk" ], "repo": "holdenk/spark-upgrade", "url": "https://github.com/holdenk/spark-upgrade/issues/59", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2655945215
fix empty queue in last read Fix for corner case produced in the next situation: async _read (cb) { // this._queue.length is 1 if (this._queue.length === 0) { this.push(null) return cb(null) } try { while (this._queue.length > 0) { const key = this._queue.shift() // this.module.has(key) is true if (this.modules.has(key)) continue const data = await this._addOnce(key) this.modules.set(key, data) this._pending.delete(key) if (this.push(data) === false) break } } catch (err) { return cb(err) } cb(null) } During the stream read: const dependencyStream = new DependencyStream(drive, { entrypoint }) for await (const dep of dependencyStream) { // doesnt do the last read (when queue.length is 0 } } IMPORTANT: I have been able to repro this issue only in Bare, the same test works for Node, 2.2.4
gharchive/pull-request
2024-11-13T15:46:23
2025-04-01T06:44:27.074686
{ "authors": [ "mafintosh", "rafapaezbas" ], "repo": "holepunchto/dependency-stream", "url": "https://github.com/holepunchto/dependency-stream/pull/5", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
569387814
Generate a build.gradle.kts file Switch the gradle file generated by the --idea option to use Kotlin. Closes https://github.com/holgerbrandl/kscript/issues/222 Merged at the same time as https://github.com/holgerbrandl/kscript/pull/258
gharchive/pull-request
2020-02-22T20:17:40
2025-04-01T06:44:27.077996
{ "authors": [ "martinbonnin" ], "repo": "holgerbrandl/kscript", "url": "https://github.com/holgerbrandl/kscript/pull/257", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2029685064
Map more Python types to SQLite types Fixes https://github.com/holoviz/holonote/issues/75 but likely in an unsatisfactory way. I combed through the pandas library trying to find the optimal solution for converting python/pandas/numpy dtypes into corresponding SQL types--each with its own trade-offs. In the end, since I don't knowing too much of HoloNote's internals, specifically whether I can update SpecItem, or how much of HoloNote I should change (can we completely remove SQLiteDB(Connector) from connector.py and use SQLAlchemy?), I chose the most compatible path forward. However, I list my notes below: # 1 doesn't touch internal method, but needs to parse the CREATE TABLE string to get schema, not sure where to inject inside HoloNote # Outputs: `CREATE TABLE "test" (\n"index" INTEGER,\n "a" INTEGER,\n "b" REAL,\n "c" TEXT,\n "d"...`` table = pd.io.sql.SQLiteTable("test", None, frame=df) table.sql_schema() # 2 touches internal method, and not sure where to inject inside HoloNote # however, for sqlalchemy, table._sql_type_name is _sqlalchemy_type table = pd.io.sql.SQLiteTable("test", None, frame=df) print(table._get_column_names_and_types(table._sql_type_name)) # 3 still need to use mapper, and not sure where to inject inside HoloNote for col in df.columns: try: print(col, type(df[col][0]), pd.api.types.pandas_dtype(type(df[col][0]))) except Exception as e: print(col, type(df[col][0]), "FAILED") continue # 4 still needs to use mapper and not sure where to inject inside HoloNote for col in df.columns: print(pd.api.types.infer_dtype(df[col])) # 5 include SQLAlchemy dependency # 6 include pyarrow dependency I was also wondering why we need to have the connector SQLiteDB.create_table define the types inside? Why not have df.to_sql() figure it out? Also, do we need a Connector class or can we depend on SQLAlchemy's classes? Thanks for fixing the tests! I think the suggested change is good and will merge it. can we completely remove SQLiteDB(Connector) from connector.py and use SQLAlchemy I don't think we can ever completely remove the need for a Connector, but it should be as thin as possible layer to the database. I see the class the same way as pd.io.sql.SQLiteTable, not handling any of the communication but giving a simple interface for us to send and get data from a database. We want to add an SQLAlchemy connector see https://github.com/holoviz/holonote/pull/65.
gharchive/pull-request
2023-12-07T01:35:52
2025-04-01T06:44:27.083910
{ "authors": [ "Hoxbro", "ahuang11" ], "repo": "holoviz/holonote", "url": "https://github.com/holoviz/holonote/pull/84", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1332047695
Document paths plot I'm trying to improve the docstrings of hvplot. I cannot find any documentation for hvPlot.paths. https://hvplot.holoviz.org/search.html?q=paths I believe something should be available in the Reference Gallery https://hvplot.holoviz.org/reference/index.html hvplot.paths are mentioned in https://hvplot.holoviz.org/user_guide/Geographic_Data.html but indeed this plot type doesn't have a reference gallery, and deserves one.
gharchive/issue
2022-08-08T15:34:47
2025-04-01T06:44:27.086485
{ "authors": [ "MarcSkovMadsen", "maximlt" ], "repo": "holoviz/hvplot", "url": "https://github.com/holoviz/hvplot/issues/828", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
710249047
Cant fill in particular config value in OZW addon The problem I have some Neo coolcam PIR PD02Z sensors and are unable to change parameter 99 (ambient light intensity calibration) . In the OZW addon I cant fill in / change a number at the entry of parameter 99. When I try this via the ozw.set_config_parameter I get some errors in HA log and OZW addon log. The ozw log doesn't show no errors / warnings. The strange thing is that the initial value is 1000. When I click on that value to change it, it changes automatically to 1. Then I cant change it anymore. I checked the xml file but i dont see something strange Environment Add-on with the issue: OpenZwave Add-on release with the issue: 0.5.2 Last working add-on release (if known): Operating environment (OS/Supervised): Supervised Aeotec gen 5 stick + NEO coolccam PIR V2 (PD02Z) Problem-relevant configuration device: /dev/serial/by-id/usb-0658_0200-if00 network_key: >- xxxx xxxx xxxx Traceback/Error logs HA message when using calling ozw.set_config_parameter Kan service ozw/set_config_parameter niet aanroepen 'NoneType' object has no attribute 'type' Message in OZW addon log [20200928 14:57:02.654 CEST] [default] [debug]: QVariant(int, 5) [20200928 14:57:02.656 CEST] [default] [debug]: QVariant(int, 14) [20200928 14:57:02.656 CEST] [default] [debug]: QVariant(int, 1) [20200928 14:57:02.657 CEST] [default] [debug]: QVariant(int, 3) [20200928 14:57:02.658 CEST] [default] [debug]: QVariant(int, 12) [20200928 14:57:02.659 CEST] [default] [debug]: QVariant(int, 4) [20200928 14:57:02.660 CEST] [default] [debug]: QVariant(int, 16) [20200928 14:57:02.661 CEST] [default] [debug]: QVariant(int, 17) Additional information There are various problems setting values in ozw-admin, for example: https://github.com/OpenZWave/ozw-admin/issues/33. The problem with set_config_parameter and short values was fixed in HA 0.115. What version are you using? Hi, I am using version 0.115.6 What is the service data you are using? Another question, what Stage is your node in? You can find this in the UI, Configuration -> Devices. Select this device and the Device Info will list the Stage. The stage of this node is complete and this is the service data: node_id: 14 parameter: 99 value: 1000 Kan service ozw/set_config_parameter niet aanroepen 'NoneType' object has no attribute 'type' I am guessing that you are running into the same problem that this PR was attempting to fix. Until that problem is fixed in the upstream library you won't be able to use Home Assistant to change the value. There also looks to be a bug in the device XML definition for parameter 99. This error is likely causing the ozw-admin GUI to be unable to configure the value correctly. <Value type="short" genre="config" instance="1" index="99" label="Ambient Light Intensity Calibration" value="1000" min="1" max="65536" size="2"> The type is a "short" with size "2", but the maximum value is set 65536. OZW represents "short" values as signed 16-bit integers. The maximum value is in fact 32767. A value of 65536 when casted to a short maps to 0, which is why ozw-admin won't let you change the value. This would also affect HA since it checks the min and max. If the device really does allow a value up to 65535 (0xffff), then to allow the full range of values, you'd need to modify the XML file to: <Value type="short" genre="config" instance="1" index="99" label="Ambient Light Intensity Calibration" value="1000" min="-32768" max="32767" size="2"> You cannot exclude the value 0. But this configuration parameter is undocumented, and there appear to be different versions of the product, so who knows what it's supposed to be. You can bypass HA and ozw-admin and set the value directly via MQTT using the setvalue command. Use MQTT Explorer to find the value id key. Use MQTT explorer to publish the command. topic: OpenZWave/1/command/setvalue/ payload: { "ValueIDKey": 123456789, "Value": 300 } Replace 123456789 with your key. If the device really does allow a value up to 65535 (0xffff), then to allow the full range of values, you'd need to modify the XML file to: <Value type="short" genre="config" instance="1" index="99" label="Ambient Light Intensity Calibration" value="1000" min="-32768" max="32767" size="2"> thanks @kpine that did the trick. Now I can change the value in the OZW gui. Very useful program, mqtt explorer. I did not know that. You can bypass HA and ozw-admin and set the value directly via MQTT using the setvalue command. Use MQTT Explorer to find the value id key. Use MQTT explorer to publish the command. topic: OpenZWave/1/command/setvalue/ payload: { "ValueIDKey": 123456789, "Value": 300 } Replace 123456789 with your key. How do I find the ValueIDKey for parameters 111-113 in MQTT explorer please? Plus I also can not see OpenZWave//command/ ? Background: I have the same issue as nelbs, but with two Aeotec Smart Switch 6's (ZW096). Parameters 111 - 113 are the reporting intervals for the switch (in seconds), but OpenZWave WebUI will not set them to anything but zero - which disables reporting. Running HA 118.3 and OpenZWave add-on 0.7.1 Thanks...
gharchive/issue
2020-09-28T13:06:58
2025-04-01T06:44:27.617034
{ "authors": [ "jasonpstokes", "kpine", "nelbs" ], "repo": "home-assistant/hassio-addons", "url": "https://github.com/home-assistant/hassio-addons/issues/1600", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
305877804
Platform xiaomi_miio not ready yet warning Make sure you are running the latest version of Home Assistant before reporting an issue. You should only file an issue if you found a bug. Feature and enhancement requests should go in the Feature Requests section of our community forum: Home Assistant release (hass --version): 0.65.5 Python release (python3 --version): 3.6 Component/platform: xiaomi_miio Description of problem: Xiaomi Smart WiFi Socket cannot be discovered. 2018-03-16 20:44:34 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 30 seconds. 2018-03-16 20:45:05 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 60 seconds. 2018-03-16 20:46:06 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 90 seconds. 2018-03-16 20:47:37 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 120 seconds. 2018-03-16 20:49:38 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 150 seconds. 2018-03-16 20:52:09 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 180 seconds. Expected: Xiaomi WIFI switchs are to be discovered as usual. Problem-relevant configuration.yaml entries and steps to reproduce: Traceback (if applicable): Additional info: I have the same problem. These log entries flood log: 8-03-16 09:25:22 ERROR (Thread-6) [miio.device] Got error when receiving: timed out 2018-03-16 09:25:22 WARNING (Thread-6) [miio.device] Retrying with incremented id, retries left: 3 2018-03-16 09:25:22 ERROR (Thread-2) [miio.device] Got error when receiving: timed out 2018-03-16 09:25:22 WARNING (Thread-2) [miio.device] Retrying with incremented id, retries left: 3 2018-03-16 09:25:26 WARNING (MainThread) [homeassistant.components.light] Setup of platform xiaomi_miio is taking over 10 seconds. 2018-03-16 09:43:05 ERROR (Thread-6) [miio.device] Got error when receiving: timed out 2018-03-16 09:43:05 WARNING (Thread-6) [miio.device] Retrying with incremented id, retries left: 3 2018-03-16 10:07:52 ERROR (Thread-4) [miio.device] Got error when receiving: timed out 2018-03-16 10:07:52 WARNING (Thread-4) [miio.device] Retrying with incremented id, retries left: 3 2018-03-16 10:25:46 ERROR (MainThread) [homeassistant.core] Timer got out of sync. Resetting Please enable debug output logger: default: warn logs: homeassistant.components.switch.xiaomi_miio: debug miio: debug and provide some more details. May be this thread provides a solution: https://community.home-assistant.io/t/xiaomi-miio-switch-not-initializing/46572/11 In my case the issue is with lights not switch. This is my debug output: Mar 16 13:46:27 hassbian hass[17048]: 2018-03-16 13:46:27 INFO (MainThread) [homeassistant.components.light.xiaomi_miio] Initializing with host philips-light-bulb_miio61806213.sozkara (token 263aa...) Mar 16 13:46:27 hassbian hass[17048]: 2018-03-16 13:46:27 DEBUG (MainThread) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:27 hassbian hass[17048]: 2018-03-16 13:46:27 DEBUG (MainThread) [miio.device] Got a response: Container: Mar 16 13:46:27 hassbian hass[17048]: data = Container: Mar 16 13:46:27 hassbian hass[17048]: length = 0 Mar 16 13:46:27 hassbian hass[17048]: value = (total 0) Mar 16 13:46:27 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:27 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:27 hassbian hass[17048]: data = (total 0) Mar 16 13:46:27 hassbian hass[17048]: header = Container: Mar 16 13:46:27 hassbian hass[17048]: length = 16 Mar 16 13:46:27 hassbian hass[17048]: value = Container: Mar 16 13:46:27 hassbian hass[17048]: length = 32 Mar 16 13:46:27 hassbian hass[17048]: unknown = 0 Mar 16 13:46:27 hassbian hass[17048]: device_id = \x03\xaf\x16\x85 (total 4) Mar 16 13:46:27 hassbian hass[17048]: ts = 1970-01-12 13:01:08 Mar 16 13:46:27 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:27 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:27 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\xaf\x16\x85\x00\x0f7\x94 (total 16) Mar 16 13:46:27 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:27 hassbian hass[17048]: 2018-03-16 13:46:27 DEBUG (MainThread) [miio.device] Discovered 03af1685 with ts: 1970-01-12 13:01:08, token: b'00000000000000000000000000000000' Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (MainThread) [miio.device] philips-light-bulb_miio61806213.sozkara:54321 >>: {'id': 1, 'method': 'miIO.info', 'params': []} Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (MainThread) [miio.device] philips-light-bulb_miio61806213.sozkara:54321 (ts: 1970-01-12 13:01:10, id: 1) << {'id': 1, 'result': {'life': 997270, 'uid': 1791780321, 'ap': {'rssi': -33, 'bssid': 'E8:DE:27:83:53:B5', 'ssid': 'sozkara 2.4ghz'}, 'hw_ver': 'ESP8266', 'wifi_fw_ver': '1.5.0-dev(7f7a714)', 'mmfree': 5264, 'token': '263aa92a412406f1fa33437abf35bffc', 'netif': {'localIp': '192.168.1.118', 'gw': '192.168.1.1', 'mask': '255.255.255.0'}, 'model': 'philips.light.bulb', 'fw_ver': '1.3.0_0033', 'mac': '34:CE:00:C5:CF:9D'}} Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 INFO (MainThread) [homeassistant.components.light.xiaomi_miio] philips.light.bulb 1.3.0_0033 ESP8266 detected Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 WARNING (MainThread) [homeassistant.setup] Setup of recorder is taking over 10 seconds. Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 INFO (MainThread) [homeassistant.components.light.xiaomi_miio] Initializing with host philips-light-bulb_miio60394599.sozkara (token abb00...) Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (Thread-2) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (MainThread) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (Thread-2) [miio.device] Got a response: Container: Mar 16 13:46:29 hassbian hass[17048]: data = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 0 Mar 16 13:46:29 hassbian hass[17048]: value = (total 0) Mar 16 13:46:29 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:29 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:29 hassbian hass[17048]: data = (total 0) Mar 16 13:46:29 hassbian hass[17048]: header = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 16 Mar 16 13:46:29 hassbian hass[17048]: value = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 32 Mar 16 13:46:29 hassbian hass[17048]: unknown = 0 Mar 16 13:46:29 hassbian hass[17048]: device_id = \x03\xaf\x16\x85 (total 4) Mar 16 13:46:29 hassbian hass[17048]: ts = 1970-01-12 13:01:10 Mar 16 13:46:29 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:29 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:29 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\xaf\x16\x85\x00\x0f7\x96 (total 16) Mar 16 13:46:29 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (MainThread) [miio.device] Got a response: Container: Mar 16 13:46:29 hassbian hass[17048]: data = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 0 Mar 16 13:46:29 hassbian hass[17048]: value = (total 0) Mar 16 13:46:29 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:29 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:29 hassbian hass[17048]: data = (total 0) Mar 16 13:46:29 hassbian hass[17048]: header = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 16 Mar 16 13:46:29 hassbian hass[17048]: value = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 32 Mar 16 13:46:29 hassbian hass[17048]: unknown = 0 Mar 16 13:46:29 hassbian hass[17048]: device_id = \x03\x99\x8cg (total 4) Mar 16 13:46:29 hassbian hass[17048]: ts = 1970-01-12 13:01:14 Mar 16 13:46:29 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:29 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:29 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\x99\x8cg\x00\x0f7\x9a (total 16) Mar 16 13:46:29 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (MainThread) [miio.device] Discovered 03998c67 with ts: 1970-01-12 13:01:14, token: b'00000000000000000000000000000000' Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (Thread-2) [miio.device] Discovered 03af1685 with ts: 1970-01-12 13:01:10, token: b'00000000000000000000000000000000' Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (MainThread) [miio.device] philips-light-bulb_miio60394599.sozkara:54321 >>: {'id': 1, 'method': 'miIO.info', 'params': []} Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (Thread-2) [miio.device] philips-light-bulb_miio61806213.sozkara:54321 >>: {'id': 1, 'method': 'get_prop', 'params': ['power', 'bright', 'cct', 'snm', 'dv']} Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (MainThread) [miio.device] philips-light-bulb_miio60394599.sozkara:54321 (ts: 1970-01-12 13:01:15, id: 1) << {'id': 1, 'result': {'life': 997275, 'uid': 1791780321, 'ap': {'rssi': -46, 'bssid': 'E8:DE:27:83:53:B5', 'ssid': 'sozkara 2.4ghz'}, 'hw_ver': 'ESP8266', 'wifi_fw_ver': '1.5.0-dev(7f7a714)', 'mmfree': 5408, 'token': 'abb00a4eec3a71c8ed18283820a0b3ca', 'netif': {'localIp': '192.168.1.129', 'gw': '192.168.1.1', 'mask': '255.255.255.0'}, 'model': 'philips.light.bulb', 'fw_ver': '1.3.0_0033', 'mac': '34:CE:00:97:18:42'}} Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 INFO (MainThread) [homeassistant.components.light.xiaomi_miio] philips.light.bulb 1.3.0_0033 ESP8266 detected Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (Thread-5) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (Thread-5) [miio.device] Got a response: Container: Mar 16 13:46:30 hassbian hass[17048]: data = Container: Mar 16 13:46:30 hassbian hass[17048]: length = 0 Mar 16 13:46:30 hassbian hass[17048]: value = (total 0) Mar 16 13:46:30 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:30 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:30 hassbian hass[17048]: data = (total 0) Mar 16 13:46:30 hassbian hass[17048]: header = Container: Mar 16 13:46:30 hassbian hass[17048]: length = 16 Mar 16 13:46:30 hassbian hass[17048]: value = Container: Mar 16 13:46:30 hassbian hass[17048]: length = 32 Mar 16 13:46:30 hassbian hass[17048]: unknown = 0 Mar 16 13:46:30 hassbian hass[17048]: device_id = \x03\x99\x8cg (total 4) Mar 16 13:46:30 hassbian hass[17048]: ts = 1970-01-12 13:01:15 Mar 16 13:46:30 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:30 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:30 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\x99\x8cg\x00\x0f7\x9b (total 16) Mar 16 13:46:30 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (Thread-5) [miio.device] Discovered 03998c67 with ts: 1970-01-12 13:01:15, token: b'00000000000000000000000000000000' Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (Thread-5) [miio.device] philips-light-bulb_miio60394599.sozkara:54321 >>: {'id': 1, 'method': 'get_prop', 'params': ['power', 'bright', 'cct', 'snm', 'dv']} Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 ERROR (Thread-2) [miio.device] Got error when receiving: timed out Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 WARNING (Thread-2) [miio.device] Retrying with incremented id, retries left: 3 Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 DEBUG (Thread-2) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 DEBUG (Thread-2) [miio.device] Got a response: Container: Mar 16 13:46:35 hassbian hass[17048]: data = Container: Mar 16 13:46:35 hassbian hass[17048]: length = 0 Mar 16 13:46:35 hassbian hass[17048]: value = (total 0) Mar 16 13:46:35 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:35 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:35 hassbian hass[17048]: data = (total 0) Mar 16 13:46:35 hassbian hass[17048]: header = Container: Mar 16 13:46:35 hassbian hass[17048]: length = 16 Mar 16 13:46:35 hassbian hass[17048]: value = Container: Mar 16 13:46:35 hassbian hass[17048]: length = 32 Mar 16 13:46:35 hassbian hass[17048]: unknown = 0 Mar 16 13:46:35 hassbian hass[17048]: device_id = \x03\xaf\x16\x85 (total 4) Mar 16 13:46:35 hassbian hass[17048]: ts = 1970-01-12 13:01:16 Mar 16 13:46:35 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:35 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:35 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\xaf\x16\x85\x00\x0f7\x9c (total 16) Mar 16 13:46:35 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 DEBUG (Thread-2) [miio.device] Discovered 03af1685 with ts: 1970-01-12 13:01:16, token: b'00000000000000000000000000000000' Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 DEBUG (Thread-2) [miio.device] philips-light-bulb_miio61806213.sozkara:54321 >>: {'id': 102, 'method': 'get_prop', 'params': ['power', 'bright', 'cct', 'snm', 'dv']} Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 DEBUG (Thread-2) [miio.device] philips-light-bulb_miio61806213.sozkara:54321 (ts: 1970-01-12 13:01:16, id: 102) << {'id': 102, 'result': ['off', 80, 1, 0, 0]} Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 ERROR (Thread-5) [miio.device] Got error when receiving: timed out Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 WARNING (Thread-5) [miio.device] Retrying with incremented id, retries left: 3 Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 DEBUG (Thread-5) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 DEBUG (Thread-5) [miio.device] Got a response: Container: Mar 16 13:46:36 hassbian hass[17048]: data = Container: Mar 16 13:46:36 hassbian hass[17048]: length = 0 Mar 16 13:46:36 hassbian hass[17048]: value = (total 0) Mar 16 13:46:36 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:36 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:36 hassbian hass[17048]: data = (total 0) Mar 16 13:46:36 hassbian hass[17048]: header = Container: Mar 16 13:46:36 hassbian hass[17048]: length = 16 Mar 16 13:46:36 hassbian hass[17048]: value = Container: Mar 16 13:46:36 hassbian hass[17048]: length = 32 Mar 16 13:46:36 hassbian hass[17048]: unknown = 0 Mar 16 13:46:36 hassbian hass[17048]: device_id = \x03\x99\x8cg (total 4) Mar 16 13:46:36 hassbian hass[17048]: ts = 1970-01-12 13:01:21 Mar 16 13:46:36 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:36 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:36 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\x99\x8cg\x00\x0f7\xa1 (total 16) Mar 16 13:46:36 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 DEBUG (Thread-5) [miio.device] Discovered 03998c67 with ts: 1970-01-12 13:01:21, token: b'00000000000000000000000000000000' Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 DEBUG (Thread-5) [miio.device] philips-light-bulb_miio60394599.sozkara:54321 >>: {'id': 102, 'method': 'get_prop', 'params': ['power', 'bright', 'cct', 'snm', 'dv']} Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 DEBUG (Thread-5) [miio.device] philips-light-bulb_miio60394599.sozkara:54321 (ts: 1970-01-12 13:01:21, id: 102) << {'id': 102, 'result': ['off', 80, 1, 0, 0]} Mar 16 13:46:39 hassbian hass[17048]: 2018-03-16 13:46:39 WARNING (MainThread) [homeassistant.components.light] Setup of platform xiaomi_miio is taking over 10 seconds. Mar 16 13:46:39 hassbian hass[17048]: 2018-03-16 13:46:39 DEBUG (MainThread) [homeassistant.components.light.xiaomi_miio] Got new state: <PhilipsBulbStatus power=off, brightness=80, color_temperature=1, scene=0, delay_off_countdown=0> @serhatozkara Your device works fine. The error messages 2018-03-16 10:07:52 ERROR (Thread-4) [miio.device] Got error when receiving: timed out 2018-03-16 10:07:52 WARNING (Thread-4) [miio.device] Retrying with incremented id, retries left: 3 will be suppressed with HA 0.66 (python-miio 0.3.8): https://github.com/rytilahti/python-miio/commit/8f4cb5d6203b36a9fc45fa617b7f9cc51c4c53ab It seems that API token has changed somehow. I have obtained the new key and it solved the issue. Thank you so much for your help syssi.
gharchive/issue
2018-03-16T10:34:23
2025-04-01T06:44:27.658342
{ "authors": [ "envinur", "serhatozkara", "syssi" ], "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/issues/13259", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
487004555
Foscam FI992x motion detection problem Home Assistant release with the issue: all Last working Home Assistant release (if known): none Operating environment (Hass.io/Docker/Windows/etc.): Hass.io (but all env are affected) Component/platform: https://www.home-assistant.io/components/foscam/ Description of problem: The enable/disable motion detection service do not work with the new foscam cameras because the HTTP API CGIProxy command getMotionDetectConfig (and setMotionDetectConfig) do not work anymore and has been replaced with the new getMotionDetectConfig1 (and setMotionDetectConfig1), please note the number 1 at the end. Additional information: I have inspected the code for the component (homeassistant/components/foscam/camera.py) and I have see that the python code used by the component (libpyfoscam) has two new methods to enable/disable motion detection: enable_motion_detection1 disable_motion_detection1 So can be enough to add a configuration option to use inside the configuration.yaml of the camera and select the correct libpyfoscam method based on the camera version. I am sorry but I do not have the time to fork, code, test and create a merge request ... I have made some test using the last version of lipyfoscam (1.1) and it is working well. Regards to all Hi, I have the same problem with my FI9900P. I can support with testing. I solved it with a workaround until Foscam component is updated.. switch: platform: command_line switches: cam1_motiondetection: command_on: 'curl -k "http://IPADDRESS:PORT/cgi-bin/CGIProxy.fcgi?usr=CAMUSER&pwd=CAMPASSWORD&cmd=setMotionDetectConfig1&isEnable=1&linkage=136&schedule0=281474976710655&schedule1=281474976710655&schedule2=281474976710655&schedule3=281474976710655&schedule4=281474976710655&schedule5=281474976710655&schedule6=281474976710655&width1=10000&height1=10000&valid1=1"' command_off: 'curl -k "http://IPADDRESS:PORT/cgi-bin/CGIProxy.fcgi?usr=CAMUSER&pwd=CAMPASSWORD&cmd=setMotionDetectConfig1&isEnable=0&linkage=136&schedule0=281474976710655&schedule1=281474976710655&schedule2=281474976710655&schedule3=281474976710655&schedule4=281474976710655&schedule5=281474976710655&schedule6=281474976710655&width1=10000&height1=10000&valid1=1"' command_state: 'curl -k --silent "http://IPADDRESS:PORT/cgi-bin/CGIProxy.fcgi?usr=CAMUSER&pwd=CAMPASSWORD&cmd=getMotionDetectConfig1" | grep -oP "(?<=isEnable>).*?(?=)"' value_template: '{{ value == "1" }}'
gharchive/issue
2019-08-29T14:46:55
2025-04-01T06:44:27.667905
{ "authors": [ "RicoRehn", "danilo-massa" ], "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/issues/26265", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
218700355
Invalid config? Make sure you are running the latest version of Home Assistant before reporting an issue. You should only file an issue if you found a bug. Feature and enhancement requests should go in the Feature Requests section of our community forum: Home Assistant release (hass --version): Latest dev since a couple of days. Python release (python3 --version): Python 3.4.2 Component/platform: hass Description of problem: HA doesnt start. Website remains unavailable. This is since I updated from the dev git a couple of days ago. startup gives many invalid config errors while nothing has changed on that side. Problem-relevant configuration.yaml entries and steps to reproduce: https://hastebin.com/erozapuyej.yaml Traceback (if applicable): http://i.imgur.com/Rs0crgx.png Did you run script/setup after upgrade the dev version? @Danielhiversen Maybe a stupid question, but how should I do that? I've always been following these commands: A check config gives me the following output: https://hastebin.com/ikuqocawol.txt Somehow fixed after a apt-get update and upgrade + reboot of the pi.
gharchive/issue
2017-04-01T14:45:10
2025-04-01T06:44:27.673170
{ "authors": [ "Danielhiversen", "f0nt4" ], "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/issues/6879", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
266231828
Errors in verisure :( I have HomeAssistant 0.55.2 I config verisure: verisure: username: !secret verisure_username password: !secret verisure_password And get this errors 2017-10-17 20:14:54 ERROR (MainThread) [homeassistant.setup] Error during setup of component verisure Traceback (most recent call last): File "/usr/src/app/homeassistant/setup.py", line 194, in _async_setup_component component.setup, hass, processed_config) File "/usr/local/lib/python3.6/asyncio/futures.py", line 332, in __iter__ yield self # This tells Task to wait for completion. File "/usr/local/lib/python3.6/asyncio/tasks.py", line 250, in _wakeup future.result() File "/usr/local/lib/python3.6/asyncio/futures.py", line 245, in result raise self._exception File "/usr/local/lib/python3.6/concurrent/futures/thread.py", line 56, in run result = self.fn(*self.args, **self.kwargs) File "/usr/src/app/homeassistant/components/verisure.py", line 69, in setup if not HUB.login(): File "/usr/src/app/homeassistant/components/verisure.py", line 119, in login self.session.login() File "/usr/local/lib/python3.6/site-packages/verisure/session.py", line 90, in login self._vid = json.loads(response.text)['cookie'] KeyError: 'cookie' I use latest docker image Python release Python 3.6.3 Any idea? Im from spain thanks! Not compatible with verisure spain https://github.com/persandstrom/python-verisure/issues/88
gharchive/issue
2017-10-17T18:31:06
2025-04-01T06:44:27.676104
{ "authors": [ "Nitrosito" ], "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/issues/9926", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
309359684
Fix Insteon Leak Sensor Description: The Insteon Leak Sensor device was not reporting dry/wet state correctly. This update bumps the underlying library to 0.8.3 which correctly reports dry/wet state and updates the binary_sensor/insteon_plm component to use the updated insteonplm module. Related issue (if applicable): fixes # Pull request in home-assistant.github.io with documentation (if applicable): home-assistant/home-assistant.github.io#<home-assistant.github.io PR number goes here> Example entry for configuration.yaml (if applicable): Checklist: [x] The code change is tested and works locally. [x] Local tests pass with tox. Your PR cannot be merged unless tests pass If user exposed functionality or configuration variables are added/changed: [ ] Documentation added/updated in home-assistant.github.io If the code communicates with devices, web services, or third-party tools: [x] New dependencies have been added to the REQUIREMENTS variable (example). [x] New dependencies are only imported inside functions that use them (example). [x] New dependencies have been added to requirements_all.txt by running script/gen_requirements_all.py. [N/A] New files were added to .coveragerc. If the code does not interact with devices: [ ] Tests have been added to verify that the new code works. HELP!!! CI tests keep failing on script/gen_requirements_all.py. Every time I run the script locally it changes my base module requirement from 0.8.3 back to 0.7.5. I have no idea why it would be doing that but I have tried everything to make it work and cannot figure it out. I cannot see where the old version is coming from. If you look at my code clearly the requirement is 0.8.3. homeassistant/components/insteon_plm.py Nevermind. Figured it out. The system I was working on at the time had a very old version of HA installed. The script was pulling from the installed version not the development version. Will update and repost.
gharchive/pull-request
2018-03-28T13:08:47
2025-04-01T06:44:27.684216
{ "authors": [ "teharris1" ], "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/pull/13515", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
417162188
Android TV integration Description: Allow for configuring Android TV devices as media players in Home Assistant. This is almost identical to the Fire TV component. This builds off the work by @a1ex4 in https://github.com/home-assistant/home-assistant/pull/19157. Pull request in home-assistant.io with documentation (if applicable): https://github.com/home-assistant/home-assistant.io/pull/8829 Example entry for configuration.yaml (if applicable): # Example configuration.yaml entry media_player: # Use the Python ADB implementation without authentication - platform: androidtv name: Android TV 1 host: 192.168.0.111 # Use the Python ADB implementation with authentication - platform: androidtv name: Android TV 2 host: 192.168.0.222 adbkey: "/config/android/adbkey" # Use an ADB server for sending ADB commands - platform: androidtv name: Android TV 3 host: 192.168.0.123 adb_server_ip: 127.0.0.1 Checklist: [ ] The code change is tested and works locally. [ ] Local tests pass with tox. Your PR cannot be merged unless tests pass [x] There is no commented out code in this PR. If user exposed functionality or configuration variables are added/changed: [x] Documentation added/updated in home-assistant.io If the code communicates with devices, web services, or third-party tools: [x] New dependencies have been added to the REQUIREMENTS variable (example). [x] New dependencies are only imported inside functions that use them (example). [x] New or updated dependencies have been added to requirements_all.txt by running script/gen_requirements_all.py. [x] New files were added to .coveragerc. If the code does not interact with devices: [ ] Tests have been added to verify that the new code works. I tested this with my Fire TV stick and it works (although the Fire TV component is more tailored for Fire TV devices, so this was just for testing purposes). Just tested this with my Shield and it works well :+1: Confirmed working on my Shield as well. OK, "WIP" removed from the title! It may be helpful to the reviewer(s) to look at the commit diffs. This really is almost identical to the Fire TV component, with "fire" replaced by "android" and some small changes. Change 'firetv' to 'androidtv' and 'ftv' to 'atv' More Fire TV -> Android TV substitutions Modify Fire TV component for Android TV I have been following your project on your github page and always keeping tabs on the your latest commits. No real issues along with way other than when MiBox drops network connection with its deep sleep issue / bug (not your component's fault). I have 4 MiBox that I use this with. Going to bump up to this latest commit. Is there a way to add to the known apps it displays? @glassbase This is not a support thread. Please discuss support in the forums. @arsaboo I think the known apps is hard coded in the component, so I would say it leads toward the development of this component to ask how. @glassbase check the linked docs in the PR @dshokouhi Thanks It would be great if we could get this merged in before tomorrow's release! @MartinHjelmare, since you reviewed some of my Fire TV pull requests, any chance you'd be willing to review this? This is a new feature so it won't make this week's release anyhow. Would you mind pointing out the major differences between this and the firetv (considering you wrote they it is basically identical to that) in the PR summary? Would it be feasible to merge these into one, as I don't think the maintenance burden of keeping two copies in sync (bug fixes in the future, keeping up the feature parity in case of new features are introduced, ..) is to be underestimated. Would you mind pointing out the major differences between this and the firetv (considering you wrote they it is basically identical to that) in the PR summary? Would it be feasible to merge these into one, as I don't think the maintenance burden of keeping two copies in sync (bug fixes in the future, keeping up the feature parity in case of new features are introduced, ..) is to be underestimated. From memory: The Android TV component has a unique_id property and supports volume control. The Fire TV component lists the running apps as sources, and you can use media_player.select_source to start and stop apps. The update methods are essentially the same, except that different attributes get assigned to. They could be combined. I think the way to do it would be to create an ADBMediaPlayer class from which the AndroidTVDevice and FireTVDevice classes both inherit. Okay, can the device type be autodetected somehow? I think it would be for the best to simply combine it into one and use a flag/model type variable per device type to tell homeassistant which features are supported and to adjust the queries where necessary. Or are there some pros to keep them separate? I don't know how to auto-detect the device type. But it would be easy to have a configuration entry where the user specifies either firetv or androidtv. It's probably for the best to combine them. What would the integration be called? androidmediaplayer? Both are Android devices after all Here's an initial attempt to combine the platforms. The build wouldn't pass -- the requirements_all.txt and .coverage files need to be updated and the androidtv‘ and firetv` packages need some minor modifications. But let me know what you think. https://github.com/JeffLIrion/home-assistant/blob/androidtv-firetv-combined/homeassistant/components/androidtv/media_player.py I don't know how to auto-detect the device type. If you can use adb at this stage, adb shell getprop ro.product.model could do the trick. I don't know how to auto-detect the device type. If you can use adb at this stage, adb shell getprop ro.product.model could do the trick. Thanks! I took a look at the output of adb shell getprop. I think adb shell getprop ro.product.manufacturer might be even easier -- if the output is "Amazon" it's a Fire TV device, otherwise it's an Android TV device. Of course, the question is whether a particular approach is correct 100% of the time. We could always allow for an optional parameter so the user can specify the device type. I've tested this version of the component and the same ADB debug messages are dumped to the HA console as in @a1ex4's version: https://github.com/home-assistant/home-assistant/pull/19157#issuecomment-469067661. Can this please be fixed in the dependencies before release? A device type parameter that defaults to automatic detection sounds great :) @jjlawren I am using this component myself and I do not get those errors in my logs, console is not the same as logs. In my console every single platform is sending data to the console, but that does not show up in the logs which is what logger is for. Also, let us not use this thread for support/feature requests to minimize the noise. This is not a support/feature request. No component should pass through debug messages to the console or logger. With logger set to default: warn your logs and console should be relatively silent. My attempt to fix the console debug logging: https://github.com/Swind/pure-python-adb/pull/14 @JeffLIrion here's what I propose after having a chat with @arsaboo: Convert this PR to be a modification of firetv component by adding the changes necessary to support non-firetv devices. Update firetv documentation accordingly. This way we can keep the existing git history. The follow-up steps (I'd prefer to have it in a separate PR just for the sake of git diffs and logs) would be: Rename firetv to androidtv if that's the final naming of the component Rename firetv in documentation, add forward from firetv to androidtv. This way there will be no confusion of having multiple components supporting similar devices, and renaming will be just a single breaking change. @JeffLIrion here's what I propose after having a chat with @arsaboo: Convert this PR to be a modification of firetv component by adding the changes necessary to support non-firetv devices. Update firetv documentation accordingly. This way we can keep the existing git history. The follow-up steps (I'd prefer to have it in a separate PR just for the sake of git diffs and logs) would be: Rename firetv to androidtv if that's the final naming of the component. Rename firetv in documentation, add forward from firetv to androidtv. This way there will be no confusion of having multiple components supporting similar devices, and renaming will be just a single breaking change. I thought about it myself and came to the same conclusion. The new updated media_player.py will look very similar to my initial attempt at a combined component that I posted above. To allow for automatically determining the device type, I'm going to add the firetv package functionality to the androidtv package. I'll define a setup function in that package that will return either an AndroidTV instance or a FireTV instance. For the HA component, I'll add a 'device_class' (CONF_DEVICE_CLASS) config entry that can be auto (default), androidtv, or firetv. Question: can I use type(aftv).__name__ to determine which type of object I'm working with? It seems unnecessary to do from androidtv.androidtv import AndroidTV just so I can do isinstance(aftv, AndroidTV). For the HA component, I'll add a 'device_class' (CONF_DEVICE_CLASS) config entry that can be auto (default), androidtv, or firetv. That sounds good. Question: can I use type(aftv).__name__ to determine which type of object I'm working with? It seems unnecessary to do from androidtv.androidtv import AndroidTV just so I can do isinstance(aftv, AndroidTV). Well, you could, but I recommend against that as I have experienced how painful it is to create tests when such structures are used (as mock objects are then not types, but objects). The better way is to inherit both types of classes from a same base class (so that they shared functionality can be put in it) which implements some sort of device_type() function to return an (enum) type. Well, you could, but I recommend against using isinstance or that as I have experienced how painful it is to create tests when such structures are used (as mock objects of that class do not match, as they are mock objects instead of types). The better way is to inherit both types of classes from a same base class (so that they shared functionality can be put in it) which implements some sort of device_type() function to return an (enum) type. I think you misunderstood the question. Or I misunderstood your answer! In the androidtv package, I've got a base class BaseTV from which AndroidTV and FireTV inherit. I have a function called setup that returns either an AndroidTV instance or a FireTV instance. My question is, in Home Assistant, can I use type(aftv).__name__ to determine whether I should create an AndroidTVDevice object or a FireTVDevice object? It doesn't seem worthwhile to do an extra import in HA or to define an attribute/method/property in the androidtv package when this simple command will suffice. Here's the current setup function that I've got: https://github.com/JeffLIrion/python-androidtv/blob/firetv/androidtv/init.py Is that setup() going to be a part of homeassistant's code base? If not, shouldn't it be? But anyway, you will not want to use type() or any variants of isinstance() checks in your code if that's somehow possible, as it will make a mess when writing tests on that functionality later on. I'm closing this in favor of https://github.com/home-assistant/home-assistant/pull/21872.
gharchive/pull-request
2019-03-05T07:51:10
2025-04-01T06:44:27.719639
{ "authors": [ "JeffLIrion", "MartinHjelmare", "arsaboo", "dshokouhi", "glassbase", "jjlawren", "outadoc", "rytilahti" ], "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/pull/21674", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
187459359
Refactory nest component/platforms Description: #4210 I have no nest device, please testing. Tested with a thermostat, looks good :dolphin: Latest version 0.32.1 causes Nest endpoint to respond with http status 429 (too many requests), see: https://github.com/home-assistant/home-assistant/issues/4241 Receiving this error on 34.2: 16-12-04 23:45:02 homeassistant.helpers.entity: Updating state for sensor.hallway_thermostat_temperature took 2.006 seconds. Please report platform to the developers at https://goo.gl/Nvioub 0.34.4 16-12-06 23:27:02 homeassistant.helpers.entity: Updating state for sensor.main_floor_thermostat_temperature took 1.488 seconds. Please report platform to the developers at https://goo.gl/Nvioub @bagobones Report it at the given link instead.
gharchive/pull-request
2016-11-04T23:27:33
2025-04-01T06:44:27.724362
{ "authors": [ "Landrash", "RealGandy", "bagobones", "captainnapalm", "pvizeli", "w1ll1am23" ], "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/pull/4219", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
225270631
light.sensehat: plugin to control the 8x8 LED matrix on a Sense hat Description: Plugin for Raspberry Pi Sense Hat which exposes the onboard 8x8 RGB LED matrix as a light with brightness and RGB color control. Pull request in home-assistant.github.io with documentation (if applicable): home-assistant/home-assistant.github.io#2530 Example entry for configuration.yaml (if applicable): light: - platform: sensehat Checklist: If user exposed functionality or configuration variables are added/changed: [x] Documentation added/updated in home-assistant.github.io If the code communicates with devices, web services, or third-party tools: [x] Local tests with tox run successfully. Your PR cannot be merged unless tests pass [x] New dependencies have been added to the REQUIREMENTS variable (example). [x] New dependencies are only imported inside functions that use them (example). [x] New dependencies have been added to requirements_all.txt by running script/gen_requirements_all.py. [x] New files were added to .coveragerc. Hi @imrehg, It seems you haven't yet signed a CLA. Please do so here. Once you do that we will be able to review and accept this pull request. Thanks! Awesome! 🐬 Realized after merging that you want to add the assumed state property: @property def assumed_state(self) -> bool: """Return True if unable to access real state of the entity.""" return True Added that in cf664e42cca80546f1f85963f58c96942ac62138 @balloob thanks for the comments! I was working from the piglow example, but the hardware is not the same, so was good to know more. Have a question thinking about your addition: the "brightness" is just emulated, so if properly done, should this platform really just be "RGB" only? Or in general this pretending of having brightness control is useful for actual use? (that was my experience when testing it out) If it was RGB only and no pretence was done, though, we could get the the real state of the entity (the displayed color) with get_pixels. What do you think? Nah, I think the addition of brightness still makes sense as it's actually showing the color in a specific brightness. Getting the real color would be cool but I guess you would have to scan all pixels and take an average? That sounds quite expensive. Like this is probably fine. @balloob sounds good, then this is great. :) The color query would be just a single function call btw, as all the pixels set together, so getting one's value would get the value of the whole light. But it's indeed not necessary, just wanted to ask, cheers!
gharchive/pull-request
2017-04-29T17:03:23
2025-04-01T06:44:27.734150
{ "authors": [ "balloob", "homeassistant", "imrehg" ], "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/pull/7365", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
267307232
Wink external action tracking Description: This adds a new attribute to all non-sensor Wink devices called "external_change" this attribute is calculated by checking the last update time of the device (set by HA) and if the state change was was greater than 15 seconds ago, and the state change came from PubNub the attribute will be set to true. So what this means is, if the user changes a device via the official Wink app, or turns on a switch by the physical switch, this attribute will be set to true. Any time a state is changed via HA the attribute is set back to false. This could be useful in making complex automations around how a device was changed. If my front door was manually unlocked don't autolock it, but if it was unlocked via an automation automatically lock it again" My use case is, I have a light that gets turned on automatically, and it gets turned back off automatically via an HA automation. If I turn that light on manually I don't want the delayed script to get executed and make the light automatically turn off again. So I can check this attribute in the automation and if it isn't false don't execute. This doesn't work for all Wink devices. It doesn't seem to work for powerstrips/outlets because to wink they are all one device. It also doesn't seem to work on a generic zwave switch I have that doesn't report it's state quick enough. Everything else seems to work as expected... This also fixes some Wink attributes that weren't getting pulled from the base Wink device. (Climate and Alarms) Pull request in home-assistant.github.io with documentation (if applicable): home-assistant/home-assistant.github.io#<home-assistant.github.io PR number goes here> Checklist: If user exposed functionality or configuration variables are added/changed: [ ] Documentation added/updated in home-assistant.github.io If the code communicates with devices, web services, or third-party tools: [ ] Local tests with tox run successfully. Your PR cannot be merged unless tests pass I'm not sure if we should do that. First Home-Assistant are only interfaces that are implemented by class and the only logic inside hass is for mapping device stuff to home-assistant stuff. With this code you add logic into home-assistant and have logic inside library. This doesn't work for all Wink devices. That sounds wired. @balloob ? @pvizeli Seems reasonable, I could add this logic into python-wink I think? The "doesn't work for all Wink devices" is because of legacy Wink devices they they still support. These devices, like power strips, are sent back via the Wink API inside of one JSON object, but are broke out into 3 objects. The problem with this, is that the updates are sent in response and then sent out to HA to all three devices. There is no way to determine where the update came from/which device it was meant for. This means changing one will update all three. Seems like a great feature to me. I can see the a few use cases for myself. I apologize if I’m overstepping, but can this be done globally against the core state changes? If from component Do this If from Hass Do this Such an attribute cannot be trusted. If we send a command to Wink but someone will also make a change via physical button/app, we will still attribute that change to Home Assistant. Why 15 seconds? Why not 5? 20? None of these values will make real sense. It's just guess work. And since it's guess work, we should not present it in Home Assistant as a fact. If you want to make this use case work, try using the input_datetime component to keep track of when you make calls. We should not add this to the Wink component. @balloob thanks for the feedback. I agree with your assessment. I think there may still be an opportunity here but maybe via python-wink. I'll have to take a look, but I think it might be possible to remove the "guess work" out of this if it is performed at a lower level.
gharchive/pull-request
2017-10-20T21:15:41
2025-04-01T06:44:27.743704
{ "authors": [ "balloob", "darthsebulba04", "pvizeli", "rbflurry", "w1ll1am23" ], "repo": "home-assistant/home-assistant", "url": "https://github.com/home-assistant/home-assistant/pull/9995", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1638143955
[NL] Round 3 binary_sensor any/all/which/how many presence power Closed for now, will take this on again if I find some time.
gharchive/pull-request
2023-03-23T19:10:04
2025-04-01T06:44:27.745218
{ "authors": [ "TheFes" ], "repo": "home-assistant/intents", "url": "https://github.com/home-assistant/intents/pull/1258", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
853541979
Error when launching the camera view Hi, I do have one Simplicam in my setup. The preview works and updates correctly, but as soon as I tap the preview to launch the live feed, nothing happens and when I check the logs I can see the following error Anyone knows what this is and how to solve this? Everything else in the plugin works great and it's generally amazing! Hi @dallombra . As the bot suggested, some basic info is missing here like your OS, versions etc. but that error spawn unknown system error -8 sounds like something about your setup.
gharchive/issue
2021-04-08T14:36:34
2025-04-01T06:44:27.793518
{ "authors": [ "dallombra", "shamoon" ], "repo": "homebridge-simplisafe3/homebridge-simplisafe3", "url": "https://github.com/homebridge-simplisafe3/homebridge-simplisafe3/issues/209", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2528841256
🛑 HomeRun CZ ~ Auth is down In a1a7b15, HomeRun CZ ~ Auth (https://id.homerun.co.cz/health_check) was down: HTTP code: 0 Response time: 0 ms Resolved: HomeRun CZ ~ Auth is back up in 811aa00 after 7 minutes.
gharchive/issue
2024-09-16T15:50:08
2025-04-01T06:44:27.806961
{ "authors": [ "sibest19" ], "repo": "homeruntech/upptime", "url": "https://github.com/homeruntech/upptime/issues/48", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1433529876
How to create a Github Action and download large binaries Plugin Version or Commit ID v0.10.1 Unity Version 2021.3.0f1 Your Host OS All Target Platform iOS Description The iOS binary for MediaPipe is over 100 MB and can thus not be uploaded to GitHub. I'm trying to create a GitHub action that will either Generate the binaries with the release or Download the binaries from another location Without access to the whole repo (Build scripts and source files) when using only the .unitypackage and .gitignore's to prevent the large binary files from being submitted to GitHub, How could I configure my project to generate these binary files? Is there a guide for how to accomplish this or is it currently not possible yet? Code to Reproduce the issue No response Additional Context No response Sorry, I'm afraid I don't get what you mean precisely. Can you describe your issue, distinguishing between what it is and what you would like it to be? At least, once you fork the repository, I think you can run workflows and build a package with your favorite options. If I want to use this plugin in a unity project. How do I create a GitHub action that contains the mediapipe libraries. The iOS library is over 100 mb so it can’t be uploaded to GitHub. One solution I’ve thought about is downloading your release and unpacking it in the Packages during a git action. My question is. Do you have any official solution to this? At least, I think you can build libraries by copying published GitHub workflows. The iOS library is over 100 mb so it can’t be uploaded to GitHub. You can reduce the size by specifying --no-apple_bitcode and --linkopt="-s" (cf. https://github.com/homuler/MediaPipeUnityPlugin/issues/724#issuecomment-1228353665). If you want to keep the embedded bitcode, you may also want to look into Git LFS (cf. https://docs.github.com/en/repositories/working-with-files/managing-large-files/about-large-files-on-github#file-size-limits). One solution I’ve thought about is downloading your release and unpacking it in the Packages during a git action. Sorry, I don't know what is the problem. Why don't you do so? I solved it! Thank you. The commands to make it smaller also helped! By creating a fork of the project and then making a custom release I'm now downloading that release in my github actions :+1: :D
gharchive/issue
2022-11-02T17:31:55
2025-04-01T06:44:27.821598
{ "authors": [ "Kariaro", "homuler" ], "repo": "homuler/MediaPipeUnityPlugin", "url": "https://github.com/homuler/MediaPipeUnityPlugin/issues/779", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2185274206
Update gui.py 解决页面启动时候 检查更新异常 闪退问题 解决页面启动时候 检查更新异常 闪退问题
gharchive/pull-request
2024-03-14T02:52:14
2025-04-01T06:44:27.834112
{ "authors": [ "arkizat" ], "repo": "honmashironeko/sqlmap-gui", "url": "https://github.com/honmashironeko/sqlmap-gui/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1830683743
Return empty 200 Response as default Like itty-router, just return an empty 200 Response as default if there's no response specified. http methods include GET/POST/PUT/PATCH/DELETE. Most of the time only the GET method return a details response, other methods return an empty 200 Response. So that, by implementing this feature, we can remove redundant return c.body() from the routes. Hi @chientrm, I don't think it's a good idea to return a 200 Response as the default. There is possibility that users don't desire it to be 200, but 404 or throwing the error instead. And, basically, we designed it so that the "handler should return a Response object." @usualoma @metrue @ThatOneBro or others: What do you think about it? Hi @chientrm @yusukebe If a handler returns undefined, it means "the handler did not process anything" and not "success, but the response is empty." This is a necessary specification for the routing fallback mechanism. Also, in general, it is preferable to return a meaningful response from even other than GET. Thus, I think no need to return an empty 200 response. I agree with @usualoma, it can lead to unexpected behavior and false positives that a route is working. Probably better not to default to a 200 response. We could close this.
gharchive/issue
2023-08-01T07:58:30
2025-04-01T06:44:27.840821
{ "authors": [ "ThatOneBro", "chientrm", "usualoma", "yusukebe" ], "repo": "honojs/hono", "url": "https://github.com/honojs/hono/issues/1280", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2070255596
fix: streamSSE's doesn't trigger onAbort correctly Author should do the followings, if applicable [x] Add tests [x] Run tests [x] yarn denoify to generate files for Deno #1902 Hi, @watany-dev. cc: @yusukebe I have changed streamSSE from the traditional implementation method to something similar to the original stream helper one, can you review it once as I am not familiar with this implementation? @sor4chi Thank you. I was also able to confirm that onAbort stops normally when I actually run it. I think this is the way to go. @watany-dev Thanks for reviewing! @yusukebe This PR is review ready, please take a look this! Hi @sor4chi Looks good! I've left one comment. Check it please! fixed! @sor4chi @watany-dev Thanks! Merging now.
gharchive/pull-request
2024-01-08T11:27:19
2025-04-01T06:44:27.844594
{ "authors": [ "sor4chi", "watany-dev", "yusukebe" ], "repo": "honojs/hono", "url": "https://github.com/honojs/hono/pull/1920", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
970053265
Add moonfly theme Hello, Author of moonfly and nightfly speaking. I notice that the nightly theme is already supported (thanks). I thought it would be nice to also add lualine support for my other theme (moonfly). Cheers. Great . But since you're the author of the colorscheme I think it'll be best if you put the theme in the colorscheme repo .That way you can tweek it later easily if you change some colors in the colorscheme without coming here for change. Since you're already providing lightline & airline themes it'll be pretty much the same . You'll just have to put the same moonfly.lua file in the same folder /lua/lualine/themes/ in your repo and lualine will pick it up :) Sure, I can do that quite easily. But, I believe that should also happen with the nightfly theme then as well, adding nightfly.lua to my nightfly repo /lua/lualine/themes directory. Will that be a problem with the nightly.lua that exists here? But, I believe that should also happen with the nightfly theme then as well, adding nightfly.lua to my nightfly repo /lua/lualine/themes directory. Will that be a problem with the nightly.lua that exists here? You can put that in your repo too. Though same name in two places may cause conflict . If you want I can remove the nightfly theme from from shadmansaleh/lualine.nvim That's the one I'm currently maintaining . I'll probably add something lualine loads theme from external plugin instead in from itself when there's a conflict . Cool, I will add both themes to my two repos. The nightfly.lua I add will be exactly the same as yours. Cheers. Is your lualine the one I should reference in my repositories or should it be this one? Is this project abandoned whilst yours is mantained? For now I can't contact @hoob3rt. I haven't hard forked it yet . The changes and fixes that I've made is currently a pr to this repo #311 . I'll give @hoob3rt some more time . If he doesn't contact back I'll just hard fork this. So for now you can reference the one you like . moonfly & nightfly no support lualine. I will keep an eye on #311 to see if you hard fork, if you do I will change my README files to reference your version (if that becomes the maintained version). Thanks. @bluz71 after https://github.com/shadmansaleh/lualine.nvim/commit/32727039a97ece7ee4eb6d889be0b3e50b570daa nightfly theme from your repo will be loaded when the colorschme is installed instead of the one bundled with lualine :) Thanks, that is good to know. It is a nice feeling having both my themes now supporting lualine along with airline/lightline.
gharchive/pull-request
2021-08-13T04:31:54
2025-04-01T06:44:27.856279
{ "authors": [ "bluz71", "shadmansaleh" ], "repo": "hoob3rt/lualine.nvim", "url": "https://github.com/hoob3rt/lualine.nvim/pull/318", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
205316780
请问我是在application目录下还是在根目录下 compose ? 能否提供几个案例,协助一下。谢谢 根目录下composer application\extra目录下需要wechat.php文件吗
gharchive/issue
2017-02-04T02:39:04
2025-04-01T06:44:27.871558
{ "authors": [ "banchlor", "hooklife", "rcj654020445" ], "repo": "hooklife/thinkphp5-wechat", "url": "https://github.com/hooklife/thinkphp5-wechat/issues/3", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
371981906
[HOPS-683], [HOPS-684], [HOPS-686], [HOPS-687], [HOPS-689], [HOPS-698], [HOPS-699], [HOPS-700], [HOPS-701], [HOPS-702], [HOPS-703], [HOPS-704], [HOPS-705], [HOPS-706], [HOPS-707], [HOPS-708], [HOPS-709], [HOPS-710], [HOPS-712], [HOPS-713] Make sure there is no duplicate PR for this issue Please check if the PR fulfills these requirements [ ] Tests for the changes have been added and passed (for bug fixes / features) [x] HOPS JIRA issue has been opened for this PR [x] All commits have been squashed down to a single commit [x] The commit message has the following format: [HOPS-XXX] message Post a link to the associated JIRA issue https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-683 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-684 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-686 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-687 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-689 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-698 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-699 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-700 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-701 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-702 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-703 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-704 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-705 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-706 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-707 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-708 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-709 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-710 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-712 https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-713 What kind of change does this PR introduce? (Bug fix, feature, docs update, ...) What is the new behavior (if this is a feature change)? Does this PR introduce a breaking change? (What changes might users need to make in their application due to this PR?) Other information: Jenkins test this please. Jenkins test this please. Jenkins test this please. Jenkins test this please. Jenkins test this please. Jenkins test this please. Jenkins test this please. Jenkins test this please.
gharchive/pull-request
2018-10-19T14:28:18
2025-04-01T06:44:27.898334
{ "authors": [ "berthoug" ], "repo": "hopshadoop/hops", "url": "https://github.com/hopshadoop/hops/pull/554", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
434706836
Handle Dust Transactions Look into how the dust transactions should be handled. That is to prevent the wallet user from leaking privacy by not using those UTXOs. May apply same logic to OP_return types Detailed task list will be provided below upon further research OP_return types OP_return types are not spendable , so no privacy leak
gharchive/issue
2019-04-18T10:46:58
2025-04-01T06:44:27.900095
{ "authors": [ "esengulov" ], "repo": "horizontalsystems/bank-wallet-android", "url": "https://github.com/horizontalsystems/bank-wallet-android/issues/828", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1516120557
Herzlichen Glückwunsch im Ranking der Top German Github Aktivitäten nun auf Platz 1 LINK Wow. :) Thank you for the info! Kein problem, sehr sehr gut gemacht! BTW you have a very good Github Profile. :) Excellent work! @horsicq Vielen Dank Was tun Sie hier? -> https://github.com/Community-s-Lab Ich werde über diese Organisation Anwendungen entwickeln die speziell für Linux geeignet sind und über python Programmieren. Dazu suche ich nun fähige Entwickler die dieses Projekt mit ihrem Wissen unterstützen und diesem beitreten. Es sieht gut aus. Ich könnte mich später anschließen. Würde mich sehr freuen,,, Danke
gharchive/issue
2023-01-02T07:58:06
2025-04-01T06:44:27.904094
{ "authors": [ "bylickilabs", "horsicq" ], "repo": "horsicq/Translate_all", "url": "https://github.com/horsicq/Translate_all/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1314781478
CB-17837 - Roll forward sync fails because some 'p' ending cloudera-m… …anager-server versions are named differently in repository and image catalog This commit: introduces a package pattern that removes unwanted characters from end of pattern See detailed description in the commit message. is this really required to merge into 2.60? is this really required to merge into 2.60? Yes, we intend to do a hotfix. It is causing upgrade failures once retry is used. We've seen several failures in prod already.
gharchive/pull-request
2022-07-22T09:43:24
2025-04-01T06:44:27.907122
{ "authors": [ "keyki", "lturcsanyi", "pdarvasi" ], "repo": "hortonworks/cloudbreak", "url": "https://github.com/hortonworks/cloudbreak/pull/13105", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
142694229
Remove grantResults from onActivityResult see #120 LGTM I'll check soon.
gharchive/pull-request
2016-03-22T15:55:11
2025-04-01T06:44:27.923242
{ "authors": [ "aurae", "hotchemi", "shiraji" ], "repo": "hotchemi/PermissionsDispatcher", "url": "https://github.com/hotchemi/PermissionsDispatcher/pull/121", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
807713334
Add request interceptor The problem The modern server-side embedded Shopify apps are built with JWT session tokens which are obtained on the client-side and then passed to every server request. Turbo works pretty well with it since it allows to navigate application without doing page reloading on every click or page submission. The biggest issue is passing JWT token to every Turbo request. The solution This PR introduces request interceptor support. It allows to insert async function between every fetch request. Inside this function you can do some preparation for request and add custom headers. Example usage Turbo.setRequestInterceptor(async (request) => { const token = await getSessionToken(window.app); request.addHeader("Authorization", `Bearer ${token}`); }); Could this same seam for modifying headers be achieved with a turbo:before-fetch-request event listener? addEventListener("turbo:before-fetch-request", ({ detail: { fetchOptions }) => fetchOptions.headers["..."] = "...")) Changing headers with turbo:before-fetch-request event listener will work fine for the simpler case when you need to add some static header to Turbo requests. Unfortunately, with Shopify embedded apps you need to retrieve the JWT token before every request because the token lifetime is short and we don't know exactly when it's expired inside the Turbo app. JWT token retrieval is async and we need to pause Turbo's fetch request invocation until it's done. So async interceptor works best for this complex case. I used Axios interceptors as an inspiration for this approach. @kirillplatonov Nice work! Great that we won't need to fetch JWT every x seconds with your setRequestInterceptor solution and rather ask for them just when we need them. 👍 Are there additional pros using Turbo.setRequestInterceptor VS "turbo:before-fetch-request? Is there any downside in pausing Turbo's fetching request invocation until it's done? The alternative as I see it is using turbo:before-fetch-request for Shopify Apps and a keepRetrievingToken function or similar which I assume is what @seanpdoyle was referring to above? You solution for sure looks better than the code below but would love to hear more from you regarding this and why you think it's not a good option to use keepRetrievingToken. Is another reason that the JWT might be expired when the call is made even though it was fetched just a few seconds ago? document.addEventListener("turbo:before-fetch-request", function (event) { var headers = event.detail.fetchOptions.headers; headers["Authorization"] = "Bearer " + window.sessionToken; }); We can use keepRetrievingToken or similar to fetch new tokens every x seconds: async function retrieveToken(app) { var SessionToken = window["app-bridge"].actions.SessionToken app.dispatch( SessionToken.request(), ); // Save a session token for future requests window.sessionToken = await new Promise((resolve) => { app.subscribe(SessionToken.ActionType.RESPOND, (data) => { resolve(data.sessionToken || ""); }); }); } function keepRetrievingToken(app) { setInterval(() => { retrieveToken(app); }, SESSION_TOKEN_REFRESH_INTERVAL); } @kirillplatonov Nice work on this PR! Great that we won't need to fetch JWT every x seconds with your setRequestInterceptor solution and rather ask for them just when we need them. 👍 Until it's merged I published my fork to npm: yarn add turbo-edge import * as Turbo from "turbo-edge" yarn add turbo-rails-edge import { Turbo } from "turbo-rails-edge" Great stuff @kirillplatonov. Could you add some documentation for this on https://turbo.hotwire.dev/reference/drive? It's in the turbo-site repo. Thanks! Thanks for getting this through @kirillplatonov @dhh - useful for JWT / Shopify apps! @kirillplatonov Upon further reflection, I think it might be worth exploring an approach similar to https://github.com/hotwired/turbo/pull/28#issuecomment-863986177. Which basically let's the turbo:before-fetch-request pause the request. Then we could do it all in a callback instead. @dhh having the same pausable API for both turbo:before-render and turbo:before-fetch-request would be great. I will explore this.
gharchive/pull-request
2021-02-13T09:24:12
2025-04-01T06:44:27.949166
{ "authors": [ "awd", "dhh", "kirillplatonov", "lonroth", "seanpdoyle" ], "repo": "hotwired/turbo", "url": "https://github.com/hotwired/turbo/pull/177", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1352901346
Scaling the Spectrum Hello! While I was plotting the SEDs, I realized that the spectrum was scaling differently with and without photometry points. I was wondering about the mechanism that was used to scale the spectrum. Also, the following image is a comparison between the plot that I got and the plot from Brewster. The orange line (SED that I created) should line up with the green line, but I have no idea why it doesn't align with it in the mir spectra. I've tried adding and dropping WISE photometry points, but the SED still doesn't match up. Is there a reason why the mir spectra is being plotted in this way? Thanks. Hi @graceparkk ! I'll take a look at this next week but I'm not too sure as of yet what the issue is.
gharchive/issue
2022-08-27T01:40:59
2025-04-01T06:44:27.968225
{ "authors": [ "graceparkk", "hover2pi" ], "repo": "hover2pi/sedkit", "url": "https://github.com/hover2pi/sedkit/issues/79", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
282850566
startConversation not waiting for response Hey guys I'm using the basic convo example included in the starter repo controller.hears(['color'], 'message_received', function (bot, message) { bot.startConversation(message, function (err, convo) { convo.say('This is an example of using convo.ask with a single callback.'); convo.ask('What is your favorite color?', function (response, convo) { convo.say('Cool, I like ' + response.text + ' too!'); convo.next(); }); }); }); My bot does not wait for a response and just goes straight to 'Cool, I like (blank) too!' I am using a Huawei p9 lite and hosting on Heroku. Any ideas? Thanks You should post general Botkit questions over on the main repo: https://github.com/howdyai/botkit/issues or try asking on our Public Slack in the #botkit channel: https://community.botkit.ai/
gharchive/issue
2017-12-18T11:42:08
2025-04-01T06:44:27.979169
{ "authors": [ "dale-french", "peterswimm" ], "repo": "howdyai/botkit-starter-facebook", "url": "https://github.com/howdyai/botkit-starter-facebook/issues/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2168242139
Access to huggingface vqvae Thanks for open-sourcing this incredible repo! I found that if specifying 'vqvae' for --compressor argument in train.py, it requires access to the pretrained model on Huggingface. Could you please provide the access to that model? Best Hi @quantumiracle We've opened it. Thanks. https://huggingface.co/hpcai-tech/vqvae
gharchive/issue
2024-03-05T03:51:11
2025-04-01T06:44:28.007097
{ "authors": [ "binmakeswell", "quantumiracle" ], "repo": "hpcaitech/Open-Sora", "url": "https://github.com/hpcaitech/Open-Sora/issues/29", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
801740449
Implement raw frame cleanup Make the default behaviour that the Exposure() class empties raw frames after CDS is performed, to save on memory (and include a parameter to turn this behaviour on and off e.g. cleanup=True/False) On initialisation, Exposure() takes the optional parameter cleanup (defaults to True). If cleanup is True, function cleanup_frames() is called, which empties out the raw_frames attribute.
gharchive/issue
2021-02-05T00:32:00
2025-04-01T06:44:28.182576
{ "authors": [ "hpearnshaw" ], "repo": "hpearnshaw/cmost-analysis", "url": "https://github.com/hpearnshaw/cmost-analysis/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
874466422
Use GitHub Artifact for tdlib build Do not download from some random DropBox, but keep the artifact in this repository and possibly automatically put them in persistent storage somewhere. I'm unsure what is meant with "artifact" in the descripton of this issue. However it seems the TelegramClient is trying to download libtdjson.so.1.6.0 from a Google drive location if it is not found in the current directory. If this is indeed what this issue is about , the TDLib library could be installed by external means such as by operating system specific commands to install the shared library from a repository of the operating system .... As an example, it can be done on some systems with: pkg install telegramdb and then cp /usr/lib/amd64/libtdjson.so.1.6.0 . to copy to the current directory. see #526
gharchive/issue
2021-05-03T10:55:19
2025-04-01T06:44:28.191230
{ "authors": [ "antonykamp", "cstes", "phkeese" ], "repo": "hpi-swa-teaching/TelegramClient", "url": "https://github.com/hpi-swa-teaching/TelegramClient/issues/324", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
129377886
Überschriften einzelner Seiten Jede Seite sollte eine Überschrift haben. Diese sollte nicht zu groß sein (nicht h1!). Die Überschrift soll dem Nutzer Kenntnis darüber geben, wo er sich im System gerade befindet. Ist er zum Beispiel gerade auf dem Dashboard? Sieht er sich gerade die Detailseite eines Urlaubsantrags an? Wenn der Nutzer selbst die Details seiner Anträge einsieht, so reicht eine Überschrift, die darüber informiert, welchen Antrag er gerade einsieht. Sieht der FG-Beauftragte die Details für einen Antrag eines Wimis ein, ist es notwendig, auch zu sehen, von wem der Antrag gestellt worden ist. Solved in #311
gharchive/issue
2016-01-28T08:30:58
2025-04-01T06:44:28.192969
{ "authors": [ "arneboockmeyer", "mandyklingbeil" ], "repo": "hpi-swt2/wimi-portal", "url": "https://github.com/hpi-swt2/wimi-portal/issues/276", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }