id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1878865498 | infinite recursion when self assigned to option
The following flake leads to an infinite recursion:
{
outputs = inputs@{flake-parts, nixpkgs, self, ...}:
flake-parts.lib.mkFlake {inherit inputs;} {
nixosConfigurations.default = nixpkgs.lib.nixosSystem {
modules = [
{
# assigning self to a non existent option triggers the infinite recursion
foo.repoRoot = self;
}
];
};
};
}
I think this issue is import to fix.
For example this could as well happen with the treefmt.repoRoot option. If it ever gets deprecated, users will be puzzled by a hard to debug issue.
In my case this lead to a stack overflow with no trace at all.
This does only happen when using flake-parts and not with vanilla flakes.
@DavHau which version and platform did that happen?
When I run it on x86_64-linux, I get infinite recursions instead, which do have a trace, although the quality of the trace varies between Nix versions. 2.13.3 seems best.
Good thing the latest two or three releases have test infrastructure to catch such regressions - wish we had it sooner.
Seems like a serious problem indeed. I think we should add sourceInfo directly to inputs in Nix to solve this without removing source locations for all the other errors when they occur in the anonymous "root" module that is the mkFlake {} argument.
This solution may also help with
https://github.com/hercules-ci/flake-parts/issues/148
In my case this lead to a stack overflow with no trace at all.
The example above results in an infinite recursion for me as well.
I got a stack overflow error in dream2nix, but after stripping it down to a minimal reproducer it became an infinite recursion.
Not sure if that change in behavior was due to complexity or due to library versions.
Let me know if I should publish the dream2nix expression that lead to a stack overflow.
Could you try with this?
https://github.com/NixOS/nix/pull/8879
PR description shows how to nix run that nix.
You might be able to tell what's the difference between your original problem and reproducer with it.
Blocked on https://github.com/NixOS/nix/pull/8908
Does #192 help?
With it I get:
$ nix eval . --override-input flake-parts github:hercules-ci/flake-parts/refs/pull/192/head
warning: not writing modified lock file of flake 'path:/home/user/h/issue-flake-parts-185':
• Updated input 'flake-parts':
'github:hercules-ci/flake-parts/7f53fdb7bdc5bb237da7fefef12d099e4fd611ca' (2023-09-01)
→ 'github:hercules-ci/flake-parts/0effb5db5ccc46f8787c98ca91ec64cc9721c121' (2023-10-13)
error: The option `nixosConfigurations' does not exist. Definition values:
- In `<unknown-file>'
(use '--show-trace' to show detailed location information)
Still an error, as expected, but actionable.
Fixing it up a bit:
{
outputs = inputs@{flake-parts, nixpkgs, self, ...}:
flake-parts.lib.mkFlake {inherit inputs;} {
systems = [ "x86_64-linux" ];
flake.nixosConfigurations.default = nixpkgs.lib.nixosSystem {
modules = [
{
# assigning self to a non existent option triggers the infinite recursion
foo.repoRoot = self;
}
];
};
};
}
I then get this, simulating nixos-rebuild a bit:
$ nix eval .#nixosConfigurations.default.config.system.build.toplevel.drvPath --override-input flake-parts github:hercules-ci/flake-parts/refs/pull/192/head
warning: not writing modified lock file of flake 'path:/home/user/h/issue-flake-parts-185':
• Updated input 'flake-parts':
'github:hercules-ci/flake-parts/7f53fdb7bdc5bb237da7fefef12d099e4fd611ca' (2023-09-01)
→ 'github:hercules-ci/flake-parts/0effb5db5ccc46f8787c98ca91ec64cc9721c121' (2023-10-13)
error: Neither nixpkgs.hostPlatform nor the legacy option nixpkgs.system has been set.
You can set nixpkgs.hostPlatform in hardware-configuration.nix by re-running
a recent version of nixos-generate-config.
The option nixpkgs.system is still fully supported for NixOS 22.05 interoperability,
but will be deprecated in the future, so we recommend to set nixpkgs.hostPlatform.
(use '--show-trace' to show detailed location information)
Fixed in #192
| gharchive/issue | 2023-09-02T22:29:53 | 2025-04-01T06:44:26.173759 | {
"authors": [
"DavHau",
"roberth"
],
"repo": "hercules-ci/flake-parts",
"url": "https://github.com/hercules-ci/flake-parts/issues/185",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
963168860 | docker build failed in step 6
Hi,
I have tried the installation with in a Ubuntu18.04 VM.But I met some error in STEP 6
Step 6/21 : RUN git clone --recursive https://hub.fastgit.org/open5gs/open5gs && cd open5gs && git checkout main && meson build --prefix=`pwd`/install && ninja -C build && cd build && ninja install
---> Running in 7b364147c2b9
Cloning into 'open5gs'...
Already on 'main'
Your branch is up to date with 'origin/main'.
The Meson build system
Version: 0.53.2
Source dir: /open5gs
Build dir: /open5gs/build
Build type: native build
Project name: open5gs
Project version: 2.3.2
C compiler for the host machine: cc (gcc 9.3.0 "cc (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0")
C linker for the host machine: cc ld.bfd 2.34
C++ compiler for the host machine: c++ (gcc 9.3.0 "c++ (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0")
C++ linker for the host machine: c++ ld.bfd 2.34
Host machine cpu family: x86_64
Host machine cpu: x86_64
Program git found: YES (/usr/bin/git)
Program python3 found: YES (/usr/bin/python3)
Program /usr/bin/python3 found: YES (/usr/bin/python3)
Compiler for C supports arguments -Wextra: YES
Compiler for C supports arguments -Wlogical-op: YES
Compiler for C supports arguments -Werror=missing-include-dirs: YES
Compiler for C supports arguments -Werror=pointer-arith: YES
Compiler for C supports arguments -Werror=init-self: YES
Compiler for C supports arguments -Wfloat-equal: YES
Compiler for C supports arguments -Wsuggest-attribute=noreturn: YES
Compiler for C supports arguments -Werror=missing-prototypes: YES
Compiler for C supports arguments -Werror=missing-declarations: YES
Compiler for C supports arguments -Werror=implicit-function-declaration: YES
Compiler for C supports arguments -Werror=return-type: YES
Compiler for C supports arguments -Werror=incompatible-pointer-types: YES
Compiler for C supports arguments -Werror=format=2: YES
Compiler for C supports arguments -Wstrict-prototypes: YES
Compiler for C supports arguments -Wredundant-decls: YES
Compiler for C supports arguments -Wimplicit-fallthrough=5: YES
Compiler for C supports arguments -Wendif-labels: YES
Compiler for C supports arguments -Wstrict-aliasing=3: YES
Compiler for C supports arguments -Wwrite-strings: YES
Compiler for C supports arguments -Werror=overflow: YES
Compiler for C supports arguments -Werror=shift-count-overflow: YES
Compiler for C supports arguments -Werror=shift-overflow=2: YES
Compiler for C supports arguments -Wdate-time: YES
Compiler for C supports arguments -Wnested-externs: YES
Compiler for C supports arguments -Wunused: YES
Compiler for C supports arguments -Wduplicated-branches: YES
Compiler for C supports arguments -Wmisleading-indentation: YES
Compiler for C supports arguments -Wno-sign-compare -Wsign-compare: YES
Compiler for C supports arguments -Wno-unused-parameter -Wunused-parameter: YES
Compiler for C supports arguments -ffast-math: YES
Compiler for C supports arguments -fdiagnostics-show-option: YES
Compiler for C supports arguments -fstack-protector: YES
Compiler for C supports arguments -fstack-protector-strong: YES
Compiler for C supports arguments --param=ssp-buffer-size=4: YES
meson.build:108: WARNING: Consider using the built-in warning_level option instead of using "-Wextra".
Configuring sample.yaml using configuration
Configuring 310014.yaml using configuration
Configuring csfb.yaml using configuration
Configuring volte.yaml using configuration
Configuring vonr.yaml using configuration
Configuring slice.yaml using configuration
Configuring srslte.yaml using configuration
WARNING: Output file "configs/sample.yaml" for configure_file() at configs/meson.build:49 overwrites configure_file() output at configs/meson.build:49
Configuring sample.yaml using configuration
Configuring non3gpp.yaml using configuration
Program /usr/bin/python3 found: YES (/usr/bin/python3)
Configuring mme.yaml using configuration
Configuring sgwc.yaml using configuration
Configuring sgwu.yaml using configuration
Configuring smf.yaml using configuration
Configuring amf.yaml using configuration
Configuring upf.yaml using configuration
Configuring hss.yaml using configuration
Configuring pcrf.yaml using configuration
Configuring nrf.yaml using configuration
Configuring ausf.yaml using configuration
Configuring udm.yaml using configuration
Configuring udr.yaml using configuration
Configuring pcf.yaml using configuration
Configuring nssf.yaml using configuration
Configuring bsf.yaml using configuration
Program /usr/bin/python3 found: YES (/usr/bin/python3)
Configuring mme.conf using configuration
Configuring hss.conf using configuration
Configuring smf.conf using configuration
Configuring pcrf.conf using configuration
Configuring cacert.pem using configuration
Configuring mme.cert.pem using configuration
Configuring mme.key.pem using configuration
Configuring hss.cert.pem using configuration
Configuring hss.key.pem using configuration
Configuring smf.cert.pem using configuration
Configuring smf.key.pem using configuration
Configuring pcrf.cert.pem using configuration
Configuring pcrf.key.pem using configuration
Configuring open5gs-mmed.service using configuration
Configuring open5gs-sgwcd.service using configuration
Configuring open5gs-smfd.service using configuration
Configuring open5gs-amfd.service using configuration
Configuring open5gs-sgwud.service using configuration
Configuring open5gs-upfd.service using configuration
Configuring open5gs-hssd.service using configuration
Configuring open5gs-pcrfd.service using configuration
Configuring open5gs-nrfd.service using configuration
Configuring open5gs-ausfd.service using configuration
Configuring open5gs-udmd.service using configuration
Configuring open5gs-pcfd.service using configuration
Configuring open5gs-nssfd.service using configuration
Configuring open5gs-bsfd.service using configuration
Configuring open5gs-udrd.service using configuration
Configuring 99-open5gs.netdev using configuration
Configuring 99-open5gs.network using configuration
Configuring open5gs using configuration
Configuring open5gs.conf using configuration
Has header "arpa/inet.h" : YES
Has header "ctype.h" : YES
Has header "errno.h" : YES
Has header "execinfo.h" : YES
Has header "fcntl.h" : YES
Has header "ifaddrs.h" : YES
Has header "netdb.h" : YES
Has header "pthread.h" : YES
Has header "signal.h" : YES
Has header "stdarg.h" : YES
Has header "stddef.h" : YES
Has header "stdio.h" : YES
Has header "stdint.h" : YES
Has header "stdbool.h" : YES
Has header "stdlib.h" : YES
Has header "string.h" : YES
Has header "strings.h" : YES
Has header "time.h" : YES
Has header "sys/time.h" : YES
Has header "unistd.h" : YES
Has header "net/if.h" : YES
Has header "netinet/in.h" : YES
Has header "netinet/in_systm.h" : YES
Has header "netinet/udp.h" : YES
Has header "netinet/tcp.h" : YES
Has header "sys/ioctl.h" : YES
Has header "sys/param.h" : YES
Has header "sys/random.h" : YES
Has header "sys/socket.h" : YES
Has header "sys/stat.h" : YES
Has header "limits.h" : YES
Has header "sys/syslimits.h" : NO
Has header "sys/types.h" : YES
Has header "sys/wait.h" : YES
Has header "sys/uio.h" : YES
Checking for function "arc4random" : NO
Checking for function "arc4random_buf" : NO
Checking for function "getrandom" : YES
Checking for function "localtime_r" : YES
Checking for function "getifaddrs" : YES
Checking for function "getenv" : YES
Checking for function "putenv" : YES
Checking for function "setenv" : YES
Checking for function "unsetenv" : YES
Checking for function "strerror_r" : YES
Checking for function "sigaction" : YES
Checking for function "sigwait" : YES
Checking for function "sigsuspend" : YES
Checking for function "eventfd" : YES
Checking for function "kqueue" : NO
Checking for function "epoll_ctl" : YES
Run-time dependency threads found: YES
Header <pthread.h> has symbol "pthread_barrier_wait" : YES
Header <signal.h> has symbol "sys_siglist" : YES
Checking if "strerror_r() returns char *" compiles: YES
Library execinfo found: NO
Checking for function "backtrace" : YES
Checking if "clock_gettime()" links: YES
Checking if "eventfd(2) system call" links: YES
Library socket found: NO
Checking if "socket()" links: YES
Configuring core-config-private.h using configuration
Configuring core-config.h using configuration
Compiler for C supports arguments -Wno-shift-negative-value -Wshift-negative-value: YES
Compiler for C supports arguments -Wno-unused-but-set-variable -Wunused-but-set-variable: YES
Compiler for C supports arguments -Wno-unknown-warning-option -Wunknown-warning-option: NO
Cloning into 'freeDiameter'...
fatal: unable to access 'https://github.com/open5gs/freeDiameter.git/': GnuTLS recv error (-110): The TLS connection was non-properly terminated.
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/mesonbuild/mesonmain.py", line 129, in run
return options.run_func(options)
File "/usr/lib/python3/dist-packages/mesonbuild/msetup.py", line 245, in run
app.generate()
File "/usr/lib/python3/dist-packages/mesonbuild/msetup.py", line 159, in generate
self._generate(env)
File "/usr/lib/python3/dist-packages/mesonbuild/msetup.py", line 192, in _generate
intr.run()
File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 4167, in run
super().run()
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 412, in run
self.evaluate_codeblock(self.ast, start=1)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 436, in evaluate_codeblock
raise e
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 430, in evaluate_codeblock
self.evaluate_statement(cur)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 441, in evaluate_statement
return self.function_call(cur)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 788, in function_call
return func(node, posargs, kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 285, in wrapped
return f(*wrapped_args, **wrapped_kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 174, in wrapped
return f(*wrapped_args, **wrapped_kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 3689, in func_subdir
self.evaluate_codeblock(codeblock)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 436, in evaluate_codeblock
raise e
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 430, in evaluate_codeblock
self.evaluate_statement(cur)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 441, in evaluate_statement
return self.function_call(cur)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 788, in function_call
return func(node, posargs, kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 285, in wrapped
return f(*wrapped_args, **wrapped_kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 174, in wrapped
return f(*wrapped_args, **wrapped_kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 3689, in func_subdir
self.evaluate_codeblock(codeblock)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 436, in evaluate_codeblock
raise e
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 430, in evaluate_codeblock
self.evaluate_statement(cur)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 441, in evaluate_statement
return self.function_call(cur)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 788, in function_call
return func(node, posargs, kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 285, in wrapped
return f(*wrapped_args, **wrapped_kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 174, in wrapped
return f(*wrapped_args, **wrapped_kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 3689, in func_subdir
self.evaluate_codeblock(codeblock)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 436, in evaluate_codeblock
raise e
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 430, in evaluate_codeblock
self.evaluate_statement(cur)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 443, in evaluate_statement
return self.assignment(cur)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 1064, in assignment
value = self.evaluate_statement(node.value)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 441, in evaluate_statement
return self.function_call(cur)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 788, in function_call
return func(node, posargs, kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 285, in wrapped
return f(*wrapped_args, **wrapped_kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 174, in wrapped
return f(*wrapped_args, **wrapped_kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreterbase.py", line 143, in wrapped
return f(*wrapped_args, **wrapped_kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 2540, in func_subproject
return self.do_subproject(dirname, 'meson', kwargs)
File "/usr/lib/python3/dist-packages/mesonbuild/interpreter.py", line 2582, in do_subproject
resolved = r.resolve(dirname, method)
File "/usr/lib/python3/dist-packages/mesonbuild/wrap/wrap.py", line 187, in resolve
self.get_git()
File "/usr/lib/python3/dist-packages/mesonbuild/wrap/wrap.py", line 282, in get_git
verbose_git(['clone', self.wrap.get('url'), self.directory], self.subdir_root, check=True)
File "/usr/lib/python3/dist-packages/mesonbuild/wrap/wrap.py", line 62, in verbose_git
return git(cmd, workingdir, check=check).returncode == 0
File "/usr/lib/python3/dist-packages/mesonbuild/mesonlib.py", line 61, in git
pc = subprocess.run([GIT, '-C', workingdir] + cmd,
File "/usr/lib/python3.8/subprocess.py", line 516, in run
raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['/usr/bin/git', '-C', '/open5gs/subprojects', 'clone', 'https://github.com/open5gs/freeDiameter.git', 'freeDiameter']' returned non-zero exit status 128.
Compiler for C supports arguments -Wno-missing-prototypes -Wmissing-prototypes: YES
Compiler for C supports arguments -Wno-missing-declarations -Wmissing-declarations: YES
Compiler for C supports arguments -Wno-discarded-qualifiers -Wdiscarded-qualifiers: YES
Compiler for C supports arguments -Wno-redundant-decls -Wredundant-decls: YES
Compiler for C supports arguments -Wno-shift-overflow -Wshift-overflow: YES
Compiler for C supports arguments -Wno-float-equal -Wfloat-equal: YES
Compiler for C supports arguments -Wno-implicit-fallthrough -Wimplicit-fallthrough: YES
Compiler for C supports arguments -Wno-incompatible-pointer-types-discards-qualifiers -Wincompatible-pointer-types-discards-qualifiers: NO
Compiler for C supports arguments -Wno-format-nonliteral -Wformat-nonliteral: YES
Compiler for C supports arguments -Wno-cpp -Wcpp: YES
Found pkg-config: /usr/bin/pkg-config (0.29.1)
Run-time dependency yaml-0.1 found: YES 0.2.2
Has header "netinet/sctp.h" : YES
Library sctp found: YES
Configuring sctp-config.h using configuration
Run-time dependency libmongoc-1.0 found: YES 1.16.1
Removing intermediate container 7b364147c2b9
The command '/bin/sh -c git clone --recursive https://hub.fastgit.org/open5gs/open5gs && cd open5gs && git checkout main && meson build --prefix=`pwd`/install && ninja -C build && cd build && ninja install' returned a non-zero code: 2
root@user:/home/user/volte/docker_open5gs/base#
It seems git clone error,I use https://hub.fastgit.org/***.git to download faster than https://github.com/***.git for geographical reasons.So how should I replace this address https://github.com/open5gs/freeDiameter.git?
Thanks in advance!
Update
I use VPN to improve the network environment,and execute git clone https://github.com/open5gs/freeDiameter.git in VM,it works fine.
root@user:/home/user# git clone https://github.com/open5gs/freeDiameter
Cloning into 'freeDiameter'...
remote: Enumerating objects: 749, done.
remote: Counting objects: 100% (749/749), done.
remote: Compressing objects: 100% (462/462), done.
remote: Total 749 (delta 288), reused 727 (delta 266), pack-reused 0
Receiving objects: 100% (749/749), 1.24 MiB | 1.86 MiB/s, done.
Resolving deltas: 100% (288/288), done.
But when I execute docker build --no-cache --force-rm -t docker_open5gs . it failed in STEP 6 once agian..
Hi
I resolved this issue.
I find a solution for ERROR GnuTLS recv error (-110),that is add 3 commands in Dockfile as below:
RUN apt-get install gnutls-bin
RUN git config --global http.sslVerify false
RUN git config --global http.postBuffer 1048576000
Then the error disappeared.I will continue with the installation and there may be other problems, but this issuecan be closed,thanks.
| gharchive/issue | 2021-08-07T06:38:21 | 2025-04-01T06:44:26.191757 | {
"authors": [
"myonlystarWang"
],
"repo": "herlesupreeth/docker_open5gs",
"url": "https://github.com/herlesupreeth/docker_open5gs/issues/49",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
625577262 | Cleanup
optimizes imports, replaces assert with assert_eq and fixes a linter warning
@stlankes Review required
| gharchive/pull-request | 2020-05-27T10:32:51 | 2025-04-01T06:44:26.193533 | {
"authors": [
"jschwe"
],
"repo": "hermitcore/rusty-loader",
"url": "https://github.com/hermitcore/rusty-loader/pull/6",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
82578927 | Allow client to cancel by transfer name
Allow pg:backups cancel to take an optional transfer name, and change the default for when no name provided to pick the newest active backup, rather than relying on order from api.
@uhoh-itsmaciek this seem ok?
Sure, looks good.
| gharchive/pull-request | 2015-05-29T20:40:52 | 2025-04-01T06:44:26.202805 | {
"authors": [
"tef",
"uhoh-itsmaciek"
],
"repo": "heroku/heroku",
"url": "https://github.com/heroku/heroku/pull/1595",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1488945434 | 🛑 Auth-Bridge - Test 1 is down
In 09108a2, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 3796002.
| gharchive/issue | 2022-12-10T20:48:24 | 2025-04-01T06:44:26.214628 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/10083",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1490263016 | 🛑 Software Center - Test 1 is down
In 7c72fa6, Software Center - Test 1 ($SOFTWARECENTER_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Software Center - Test 1 is back up in 1cb8a54.
| gharchive/issue | 2022-12-11T17:40:17 | 2025-04-01T06:44:26.216897 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/10127",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1672433725 | 🛑 Auth-Bridge - Test 1 is down
In a5f4970, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 8e172bc.
| gharchive/issue | 2023-04-18T06:24:56 | 2025-04-01T06:44:26.219117 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/14082",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1693508779 | 🛑 Auth-Bridge - Test 1 is down
In 05b13d1, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 16a91da.
| gharchive/issue | 2023-05-03T06:57:19 | 2025-04-01T06:44:26.221583 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/14821",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1726905578 | 🛑 Auth-Bridge - Test 1 is down
In 248ce6f, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 482de49.
| gharchive/issue | 2023-05-26T04:55:29 | 2025-04-01T06:44:26.223759 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/15965",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1819754374 | 🛑 Auth-Bridge - Test 1 is down
In b5d427f, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 64becde.
| gharchive/issue | 2023-07-25T07:49:48 | 2025-04-01T06:44:26.226009 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/18806",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1184839634 | 🛑 Auth-Bridge - Test 1 is down
In 69b64aa, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 8e28a09.
| gharchive/issue | 2022-03-29T13:03:34 | 2025-04-01T06:44:26.228215 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/1907",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1903585607 | 🛑 Auth-Bridge - Test 1 is down
In ea21d79, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 353a124 after 21 minutes.
| gharchive/issue | 2023-09-19T19:13:03 | 2025-04-01T06:44:26.230498 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/21729",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1204653624 | 🛑 Auth-Bridge - Test 1 is down
In 50d4256, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 21bd4f8.
| gharchive/issue | 2022-04-14T15:12:45 | 2025-04-01T06:44:26.232920 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/2538",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2034279120 | 🛑 Software Center - Test 1 is down
In a5f7a71, Software Center - Test 1 ($SOFTWARECENTER_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Software Center - Test 1 is back up in b7356ce after 28 minutes.
| gharchive/issue | 2023-12-10T08:55:16 | 2025-04-01T06:44:26.235093 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/25781",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2081528860 | 🛑 Auth-Bridge - Test 1 is down
In 54e0885, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 60635c4 after 54 minutes.
| gharchive/issue | 2024-01-15T08:46:27 | 2025-04-01T06:44:26.237319 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/27544",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2100776793 | 🛑 Software Center - Test 1 is down
In b8c3a47, Software Center - Test 1 ($SOFTWARECENTER_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Software Center - Test 1 is back up in 2088002 after 28 minutes.
| gharchive/issue | 2024-01-25T16:42:38 | 2025-04-01T06:44:26.239512 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/28032",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2266792185 | 🛑 Software Center - Test 1 is down
In 39d25b5, Software Center - Test 1 ($SOFTWARECENTER_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Software Center - Test 1 is back up in 0ac47fb after 43 minutes.
| gharchive/issue | 2024-04-27T03:33:46 | 2025-04-01T06:44:26.241889 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/32697",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1292451476 | 🛑 Auth-Bridge - Test 1 is down
In b0d97ef, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 7820a0c.
| gharchive/issue | 2022-07-03T23:42:15 | 2025-04-01T06:44:26.244325 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/5315",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1349541800 | 🛑 Auth-Bridge - Test 1 is down
In 714970b, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in ba24fa4.
| gharchive/issue | 2022-08-24T14:30:50 | 2025-04-01T06:44:26.246519 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/6996",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1414982913 | 🛑 Auth-Bridge - Test 1 is down
In dc85dde, Auth-Bridge - Test 1 ($AUTH_BRIDGE_TEST_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Auth-Bridge - Test 1 is back up in 6f6c4f0.
| gharchive/issue | 2022-10-19T13:30:56 | 2025-04-01T06:44:26.248754 | {
"authors": [
"herrphon"
],
"repo": "herrphon/upptime",
"url": "https://github.com/herrphon/upptime/issues/8337",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
126673485 | Fahrauftrag ausdrucken -> Fehler!
Erwartet wird:
Bereitstellen -> Fahrauftrag -> Drucken
Fahrauftrag wird auf Printer gedruckt!
Ausgegeben wird:
Guten morgen Josi
Ist kein Fehler, nur ein undeutliche Fehlermeldung. Diese Funktion ist noch nicht implementiert.
Ich bin seit letzten Freitag genau hier am bauen, bald sollte Bereitstellen->Fahrauftrag->Drucken auch möglich sein.
Martin Jonasse
Seemattstrasse 38
6333 Hünenberg See
Privat: 041 780 76 12
Mobile: 079 832 69 10
Am 14.01.2016 um 16:08 schrieb Josi Conrad notifications@github.com:
Erwartet wird:
Bereitstellen -> Fahrauftrag -> Drucken
Fahrauftrag wird auf Printer gedruckt!
Ausgegeben wird:
—
Reply to this email directly or view it on GitHub.
This is "work in progress".
Implementiert: Fahrauftrag -> Drucken erstellt eine PDF Datei für den Download bereit, gedruckt wird in Adobe Acrobat Reader.
| gharchive/issue | 2016-01-14T15:08:17 | 2025-04-01T06:44:26.262565 | {
"authors": [
"Josi-Conrad",
"Martin-Jonasse"
],
"repo": "hertus/sfitixi",
"url": "https://github.com/hertus/sfitixi/issues/153",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
428919890 | "login" route is buggy
Pointing the browser directly to the "login" route, or reloading the page while the "login" route is currently active, does not work. In both cases there is a redirect to "gamesinprogress" which leads to an authentication error.
Possibly related: When pointing the browser to the "/" route, there is redirect to "gamesinprogress" before there is a redirect to "login".
| gharchive/issue | 2019-04-03T18:47:32 | 2025-04-01T06:44:26.266429 | {
"authors": [
"herzbube"
],
"repo": "herzbube/littlego-web",
"url": "https://github.com/herzbube/littlego-web/issues/11",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
153965801 | Specify a recursivity depth
Hi !
Thanks for this great plugin for django rest framework. Is there any way to specify the depth of the recursion ?
I tried using the class Meta depth attribute but I don't think the field uses it. What would be the best approach to this ?
Best regards,
This is not currently supported. I would suggest pruning the dataset before you try to serialize it. If you are deserializing something, then you could deserialize it and then prune.
I'm using it on a ModelViewset so I don't know how I could prune it ?
There is no queryset or get_queryset method in the arguments of the field...
Can you set the queryset in the ModelViewSet?
No... My datastructure is really simple:
# models.py
class Node(models.Model):
parent = models.ForeignKey('self', related_name='children')
# api.py
class NodeSerializer(serializers.HyperlinkedModelSerializer):
parent = serializers.HyperlinkedRelatedField(
view_name='api:parameter-detail',
queryset=Node.objects.all(),
)
children = RecursiveField(many=True, allow_null=True)
class Meta:
model = Node
fields = ('parent', 'children')
depth = 2
class NodeViewSet(viewsets.ModelViewSet):
queryset = Node.objects.all()
I cannot prune or filter the queryset. It will depend on which initial node is requested. I need to limit the depth of the RecursiveField only... Or maybe I'm missing something ?
You could look in to the django treebeard package for a tree with more features. Or django mptt. I believe there are probably others as well.
Or I suppose you could extend the recursive field to add the functionality your desire
I'll just extend the recursive field to add the functionality I need. Both mptt and treabeard have limitations or defaults that don't work with my project.
Thanks.
| gharchive/issue | 2016-05-10T09:47:06 | 2025-04-01T06:44:26.380731 | {
"authors": [
"achedeuzot",
"heywbj"
],
"repo": "heywbj/django-rest-framework-recursive",
"url": "https://github.com/heywbj/django-rest-framework-recursive/issues/11",
"license": "isc",
"license_type": "permissive",
"license_source": "bigquery"
} |
736686875 | Possibly another method?
Hey there,
There's another elevated COM based method I'd like to share. Are you willing to take a look?
Thanks.
Hello,
sure, I always welcome anything new.
Awesome, I ask because I don't want to pester you if you're not free at the time.
Anyways, this method is three steps. It uses environment variables abuse/modification, shell protocol handler hijack, and lastly the elevated COM interface IFwCplLua. Basically IFwCplLua::LaunchAdvancedUI() uses ShellExecuteExW() call with %WinDir%\System32\WF.msc. What this method does is change WinDir to custom location, and launch custom WF.msc. The custom WF.msc launches a custom protocol which in turn opens cmd.exe as admin. Here is the code: https://github.com/AzAgarampur/byeintegrity4-uac/
This method actually already was in UACMe as 42, https://github.com/hfiref0x/UACME/blob/v3.2.x/Source/Akagi/methods/hybrids.c#L2392 except it abused mscfile handler hijack without touching environment variables or using fake msc snap-in file. Starting from RS4 it produced mixed results - was working and not working at same time and later was set as fixed and removed.
What does this custom msc btw? It just run something with that protocol specified with help of shockwave flash object?
Also I will be able to test this only on Saturday as I'm away of main PC.
It just run something with that protocol specified with help of shockwave flash object?
Basically. The exploit creates a URL association called protocol-byeintegrity4, and the shockwave object uses the link to launch the desired process. <String ID="3" Refs="1">protocol-byeintegrity4:</String>
Well, I can confirm it works (tested on 19042). I will look how I can integrate this into UACMe and post update here, presumably next week. Thanks for sharing.
Awesome! Thanks for letting me know.
Pushed into the dev branch. Currently it was tested only on Win10 1809.
I've done testing this on Windows 7 SP1, Windows 8.1 full patch. Tested on Windows 10 21H1 (20241) so I can assume it will work on previous Win10 versions too. If no critical bugs were found I will release this later this week, as I still want to do some other additions not related to this particular method. As usual, thanks for contribution, good work!
Hey, thanks! I appreciate the time you take to look at my work.
Do you think we can drop the registry flush calls? I don't think it'll change anything if we remove them.
Yes, sure.
This is unrelated, but the link that has the bs explanation from MS about why it works should be updated to this: https://devblogs.microsoft.com/oldnewthing/20160816-00/?p=94105
Hopefully you can include it in the readme for the next release.
Done.
| gharchive/issue | 2020-11-05T07:31:05 | 2025-04-01T06:44:26.390454 | {
"authors": [
"AzAgarampur",
"hfiref0x"
],
"repo": "hfiref0x/UACME",
"url": "https://github.com/hfiref0x/UACME/issues/88",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
418615997 | [ bug ]多次刷新SQL工单详情时,工单内容出现重复
仅v1.4.0以前产生的旧数据会出现该情况,新数据不会
重现步骤
打开工单详情:http://139.199.0.191/detail/6/
多次刷新,详情的SQL列表出现重复
截图
错误日志
无,仅前端展示异常
版本信息
应用版本 v1.4.3
部署方式 Docker
出现原因是 https://github.com/hhyo/archery/blob/master/sql/engines/models.py#L42 使用了空列表作为默认参数初始化对象属性, 并且在代码中使用了原地修改的方法修改对象属性, 这就等同于修改了对象的默认值。
解决方案是将对象属性的默认值改为None, 然后在init中判断,如果是None就置为空列表[]
这样的话,这个空列表是新生成的空列表, 而不是默认参数的那个列表,也就不会影响到之后生成的心对象
参考:
https://stackoverflow.com/questions/366422/what-is-the-pythonic-way-to-avoid-default-parameters-that-are-empty-lists
经测试已修复. http://139.199.0.191/detail/6/
| gharchive/issue | 2019-03-08T03:28:32 | 2025-04-01T06:44:26.424570 | {
"authors": [
"LeoQuote",
"hhyo"
],
"repo": "hhyo/archery",
"url": "https://github.com/hhyo/archery/issues/63",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
185070058 | HHH-11144 - Add test for issue
https://hibernate.atlassian.net/browse/HHH-11144
Applied upstream, thanks
Applied upstream, thanks
| gharchive/pull-request | 2016-10-25T09:53:28 | 2025-04-01T06:44:26.425935 | {
"authors": [
"dreab8",
"vladmihalcea"
],
"repo": "hibernate/hibernate-orm",
"url": "https://github.com/hibernate/hibernate-orm/pull/1607",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
40343443 | [4.3] HHH-9337 Region.destroy() attempts to remove a cache listener, but regio...
...n class is not annotated with @Listener
https://hibernate.atlassian.net/browse/HHH-9337
Cherry-picked and pushed.
Thanks!
Gail
| gharchive/pull-request | 2014-08-15T12:15:06 | 2025-04-01T06:44:26.427363 | {
"authors": [
"gbadner",
"pferraro"
],
"repo": "hibernate/hibernate-orm",
"url": "https://github.com/hibernate/hibernate-orm/pull/784",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2069031169 | 🛑 kiwifarms.st is down
In d2a3703, kiwifarms.st (https://kiwifarms.st) was down:
HTTP code: 0
Response time: 0 ms
Resolved: kiwifarms.st is back up in 768fc56 after 33 minutes.
| gharchive/issue | 2024-01-07T07:56:41 | 2025-04-01T06:44:26.431842 | {
"authors": [
"hickoryhouse"
],
"repo": "hickoryhouse/kf",
"url": "https://github.com/hickoryhouse/kf/issues/2869",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2209686011 | 🛑 Mad at the Internet is down
In e53cdb5, Mad at the Internet (https://madattheinternet.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Mad at the Internet is back up in 8c76a38 after 10 minutes.
| gharchive/issue | 2024-03-27T02:27:30 | 2025-04-01T06:44:26.434543 | {
"authors": [
"hickoryhouse"
],
"repo": "hickoryhouse/kf",
"url": "https://github.com/hickoryhouse/kf/issues/3261",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1510209097 | 🛑 Kiwi Farms Forum is down
In d41e9a8, Kiwi Farms Forum (https://kiwifarms.net) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Kiwi Farms Forum is back up in 5841932.
| gharchive/issue | 2022-12-24T23:38:15 | 2025-04-01T06:44:26.436912 | {
"authors": [
"hickoryhouse"
],
"repo": "hickoryhouse/kf",
"url": "https://github.com/hickoryhouse/kf/issues/84",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
500722317 | Project settings
A solution for #42.
Please refer to #37 .
I'm sorry but I don't understand your objections to my patch. the super is the problem? Why? If I'm overriding a method from a class, calling super is a standard way how use the original method.
Of course, you can reject the patch but the plugins will remain in conflict then.
I agree that using super is the standard way to override class methods.
However, since ProjectsHelper is a module, I think it is incorrect to use super.
@maeda-m I think there should be a Module#prepend instead of ProjectsController.send :helper ?
here'a an article about the difference
https://www.justinweiss.com/articles/rails-5-module-number-prepend-and-the-end-of-alias-method-chain/
Thank you. I learned from you.
Released version 1.5.0 !
| gharchive/pull-request | 2019-10-01T07:45:44 | 2025-04-01T06:44:26.470076 | {
"authors": [
"ahorek",
"maeda-m",
"picman"
],
"repo": "hidakatsuya/redmine_default_custom_query",
"url": "https://github.com/hidakatsuya/redmine_default_custom_query/pull/43",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2410141209 | Incorrect match in the source code results in inaccurate matching results.I don't know if this is a bug...
this code const [_, commentSyntax, searchPhrase, commentSyntaxEnd] = match; in matchSearchPhrase.ts 。The suggestions for this extension are as follows:
The result obtained by the regular expression is 4 groups. In addition to the first item, there should be 4 items in the match to fully indicate it. The searchPhrase in the source code only represents find, not find {question}, so one item should be added, as shown below:
const [_, commentSyntax, searchSymbol, searchPhrase, commentSyntaxEnd] = match;,The output after debugging is as follows:
After the modification, the result of debugging this extension is as follows:
You're right. I've pushed an update for this issue https://github.com/hieunc229/copilot-clone/commit/1edf91fd0b8dcfcf4b27a4d2a6d86d5c8e865de7 and will make an update to the extension marketplace
Thanks for spotting it @mengyangz86
| gharchive/issue | 2024-07-16T03:52:11 | 2025-04-01T06:44:26.483786 | {
"authors": [
"hieunc229",
"mengyangz86"
],
"repo": "hieunc229/copilot-clone",
"url": "https://github.com/hieunc229/copilot-clone/issues/88",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
113817628 | AssertionViewメソッド用のbuilderを追加
スクリーンショットの取得、検証であるassertViewは以下のAPIとなっている。
assertView(String screenshotId, CompareTarget[] compareTargets,DomSelector[] hiddenElementsSelectors)
CompareTargetやDomSelector内も複雑なオブジェクトとなっているため、
ビルダーを用意して簡略化できるようにする
下記コミットで対応しました。
c7e1eeb6ee4b0c4655e6188d1c7745e11c8332cd
ce9339c81883c0e3efb1e6e76266f3e90509b2d9
55fb8110b1a73559a342f7e0bcc5c44dc10835d0
796cd8ed2eb17bce472137f74caeb9be87fb2392
78a4ba12cb1ef437a9cf3039220bbf79fc061db3
db7d67290832e298441c21b0252596305a4637ae
6379925d124ec4df3e231d68284e9f76284db7a1
13b67cab600479ce9401843979057bf8c9b5b01a
a760f32f939f43ef3a96c68fb678870c42a56035
781490698fa4070b19bedf4c98fcb2918af6de43
ビルダー用の説明ページを追加しました
https://www.htmlhifive.com/conts/web/view/pitalium-reference/screenshot-argument
| gharchive/issue | 2015-10-28T13:09:08 | 2025-04-01T06:44:26.486798 | {
"authors": [
"tkashi"
],
"repo": "hifive/hifive-pitalium",
"url": "https://github.com/hifive/hifive-pitalium/issues/33",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
199260799 | JS: Expressions inside template literals don't display correctly
res = `start ${tags.map(tag => `pre ${tag} post`).join('\n ')}end`;
Displays like:
pre and tags.map don't have the right color
Fiddle: https://jsfiddle.net/ug3tLcf6/
Not sure I see any issue. What are you expecting? Or perhaps this has been fixed.
https://jsfiddle.net/ajoshguy/ym41ukjd/
Partly fixed (pre is coloured now). Partly as expected (tags.map is indeed also not coloured outside the template literal, I expected it to be coloured).
Closing this.
Thanks for responding after 3y instead of closing it after x months due to inactivity 😊!
Glad to help!
| gharchive/issue | 2017-01-06T18:46:37 | 2025-04-01T06:44:26.609542 | {
"authors": [
"teameh",
"yyyc514"
],
"repo": "highlightjs/highlight.js",
"url": "https://github.com/highlightjs/highlight.js/issues/1405",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
246110508 | Scrollbar not working on certain lines of code inside
Hello,
I am using the latest version of highlight.js through a Wordpress plugin, and am having problems.
When the code stretches out longer than the maximum width of the container the scroll bar appears BUT in certain code you can't click and move it. It seems frozen. You can get inside the box and highlight the code from within, but the scroll bar is totally frozen!
Here is are 2 line of code that DO NOT have a working scroll bar:
<pre><code class="shell">sudo su -c 'dnf -y install https://download1.rpmfusion.org/free/fedora/rpmfusion-free-release-$(rpm -E %fedora).noarch.rpm https://download1.rpmfusion.org/nonfree/fedora/rpmfusion-nonfree-release-$(rpm -E %fedora).noarch.rpm'
sudo rpm --import https://packages.microsoft.com/keys/microsoft.asc
sudo sh -c 'echo -e "[code]\nname=Visual Studio Code\nbaseurl=https://packages.microsoft.com/yumrepos/vscode\nenabled=1\ngpgcheck=1\ngpgkey=https://packages.microsoft.com/keys/microsoft.asc" > /etc/yum.repos.d/vscode.repo'</code></pre>

<pre><code class="shell">echo fs.inotify.max_user_watches=524288 | sudo tee -a /etc/sysctl.conf sudo sysctl -p</code></pre>
and here is one that works perfectly:
<pre><code class="shell">sudo sed -i "s/User apache/User $USERNAME/g" /etc/httpd/conf/httpd.conf</code></pre>
If I get rid of the class="shell" then it still has the exact same problem. Any idea? I attached a screen shot of a not working scrollbar. THANKS!
I’m closing this issue because it has been inactive for over a year. This probably means that it is not reproducible or it has been fixed in a newer version.
Please reopen if you still encounter this issue with the latest stable version.
Thank you!
| gharchive/issue | 2017-07-27T17:18:18 | 2025-04-01T06:44:26.613289 | {
"authors": [
"David-Else"
],
"repo": "highlightjs/highlight.js",
"url": "https://github.com/highlightjs/highlight.js/issues/1577",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
189009499 | Allow starts to be defined as 'self' much like one can do in contains
I used this for a weird linking system. Colon separated values as { end: /:/, endsWithParent: true, starts: 'self', ... }
It's not a necessary thing, but it's consistent with the feature for contains.
I'm confused when you would use this exactly?
Since this was so long ago I hardly remember but going off of "colon separated values" I think I used it for like 1 : 1 : ... but where 1 is a complex type (like could be a number or a string or whatever)
It's in here, and while the grammar sorta works, it's slow in complex cases so I kind of have up.
Seems like a good way to get stuck in an infinite loop. How does it break out if after end it always starts self?
And typically you'd do this with a parent and then inside your contains you'd just have you matcher... so it would run over and over and over again as many times as needed until you left the mode... no need for this start hack.
It breaks out when the parent ends, as the example says. :p
...
contains: [
{
begin: /\d\s:\s?/,
endWithParent: true
}
]
Doesn't that do exactly the same thing?
I don't know, I'm out of touch with the system. I'm pretty sure I only did it for reducing redundancy. I have no complaints if this is just closed.
Ok, closing. It would have needed extra documentation and I had some issues with the code also... and generally this type of feature we shouldn't just add "abstractly"... if there was a grammar included with it that required the new functionality that would make a much better case for why it's necessary/useful, etc...
| gharchive/pull-request | 2016-11-14T01:09:32 | 2025-04-01T06:44:26.616376 | {
"authors": [
"logicplace",
"yyyc514"
],
"repo": "highlightjs/highlight.js",
"url": "https://github.com/highlightjs/highlight.js/pull/1348",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
202914482 | add postcss-hocus plugin
postcss-hocus lets you type a:hocus instead of a:hover, a:focus. Short and simple :)
🎩 ✨ 🔮
| gharchive/pull-request | 2017-01-24T19:22:06 | 2025-04-01T06:44:26.628338 | {
"authors": [
"Kilian",
"himynameisdave"
],
"repo": "himynameisdave/postcss-plugins",
"url": "https://github.com/himynameisdave/postcss-plugins/pull/169",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
721690022 | Uploading video annotations from Premiere Pro
I tried to make a video project in Audiannotate and neither the video link nor the annotations were recognized by the program.
To Reproduce
Steps to reproduce the behavior:
Go to http://audiannotate.brumfieldlabs.com/
Click on My Projects and edit project (in this case, Camille 1921)
Add link and info
Upload annotations
Several errors, one when I tried to add the duration of the video (1:09:30 or 69:30, tried both) and another when I tried to upload my annotations, first without editing them and then editing them to have only four collumns
Expected behavior
I expected the link to work and the annotations to link to the timestamps with the duration
Screenshots
Got this error when I tried to correct the duration of the video:
And these errors when I tried to upload the annotations:
Additional context
I think it has something to do with the format of my annotations exported from premiere, which I tried to fix by removing two collumns (so it is now just marker name/description/time in/time out) and changing the timecodes to a 00:00:00 format rather than 00;00;00
Oh also, these are my annotations as a .txt -
Camille 1921 annotations.txt
This is a really gnarly one; fortunately our next feature (#125) should finally get us to where we need to be to support Premiere.
We think we're also going to hard-code support for Premiere upload annotations earlier to un-block research work.
| gharchive/issue | 2020-10-14T18:36:30 | 2025-04-01T06:44:26.641570 | {
"authors": [
"benwbrum",
"jreinschmidt"
],
"repo": "hipstas/AudiAnnotate",
"url": "https://github.com/hipstas/AudiAnnotate/issues/122",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1727045925 | 建议识别后导出Excel表格数据,并对应图片超链接,更加实用
建议识别后导出Excel表格数据,表格数据第一列是图片名称,第二列识别图片的内容,第三列对应图片超链接,这样可以很快找到需要搜索的内容图片。
OK,已有计划
邮件已经收到
期待你的更新,还有就是要对批量提取的字段,如果可以自定义就好了。
Copied to clipboard!
------------------ 原始邮件 ------------------
发件人: @.>;
发送时间: 2023年5月27日(星期六) 下午5:58
收件人: @.>;
抄送: @.>; @.>;
主题: Re: [hiroi-sora/Umi-OCR] 建议识别后导出Excel表格数据,并对应图片超链接,更加实用 (Issue #148)
OK,已有计划
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: @.***>
期待大佬的更新
V2预览版 已支持输出csv格式,可导入excel。
第一列图片名称、第二列识别内容、第三列图片路径。
邮件已经收到
| gharchive/issue | 2023-05-26T07:19:07 | 2025-04-01T06:44:26.646896 | {
"authors": [
"WhoIAmm",
"csq4017",
"hiroi-sora"
],
"repo": "hiroi-sora/Umi-OCR",
"url": "https://github.com/hiroi-sora/Umi-OCR/issues/148",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1769743374 | i cannot confirm my testnet requested since 3days now
Describe the bug
A clear and concise description of what the bug is.
Transaction ID
Address
Block#
Time stamp
To Reproduce
Steps to reproduce the behavior:
Go to '...'
Click on '....'
Scroll down to '....'
See error
Expected behavior
A clear and concise description of what you expected to happen.
Screenshots
If applicable, add screenshots or consol.log to help explain your problem.
Desktop (please complete the following information):
OS: [e.g. iOS]
Browser [e.g. chrome, safari]
Version [e.g. 22]
Smartphone (please complete the following information):
Device: [e.g. iPhone6]
OS: [e.g. iOS8.1]
Browser [e.g. stock browser, safari]
Version [e.g. 22]
Additional context
Add any other context about the problem here:
Please check the testnet status here https://status.hiro.so/
| gharchive/issue | 2023-06-22T13:58:11 | 2025-04-01T06:44:26.652656 | {
"authors": [
"Christianogbonnaya",
"andresgalante"
],
"repo": "hirosystems/explorer",
"url": "https://github.com/hirosystems/explorer/issues/1192",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1803880958 | STX Token Transfer SENT which I did NOT do
Here's the TRX id: 0x6829fa018ff0642fbc90255a8d3901e38f7f763c89c8fad629a2cbec01efaa63
Where did my tokens go? Why did that happen?
Please HELP! URGENT!
Sender address: SP2HEDH4SXP1A34M9KDY00PJEM6YP1QDMVD2BFYDH
Recipient address: SP2KW0M6MBSSAV1BFDKH56VFNZK73Z36C0N369K9M
Link: https://explorer.hiro.so/txid/0x6829fa018ff0642fbc90255a8d3901e38f7f763c89c8fad629a2cbec01efaa63?chain=mainnet
Where did my STX go? Is it still within STX / Hiro system or network? Please help clarify what happened.
This isn't an Explorer issue.
Please request community support on the #support channel in Discord.
| gharchive/issue | 2023-07-13T23:09:11 | 2025-04-01T06:44:26.655363 | {
"authors": [
"STX-Stargem",
"andresgalante"
],
"repo": "hirosystems/explorer",
"url": "https://github.com/hirosystems/explorer/issues/1234",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
558727765 | Integrate the 'argparse' library; add -l command-line option
Implements the rest of https://github.com/hishamhm/tl/issues/35.
Note that running tl script.tl will now result in an error.
Also, it would be nice if we had tests for the CLI tool :)
Also, it would be nice if we had tests for the CLI tool :)
@pdesaulniers Added them! :grin:
Awesome! I added some tests for the -l argument too.
We'll need some tests for tlconfig.lua as well. To do this, I think we'll need to add a -p <path to directory with tlconfig.lua> argument to the CLI. I'll do this in a later PR.
I think this is good to go — could you rebase this PR and fix the conflicts so it can be merged?
Like this? :)
| gharchive/pull-request | 2020-02-02T17:22:56 | 2025-04-01T06:44:26.668344 | {
"authors": [
"hishamhm",
"pdesaulniers"
],
"repo": "hishamhm/tl",
"url": "https://github.com/hishamhm/tl/pull/42",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1854652362 | Upgrade cryptography to fix vulnerability issues
Description
https://github.com/histolab/histolab/security/dependabot/24
Types of Changes
[ ] Core
[ ] Bugfix
[ ] New feature
[ ] Enhancement/optimization
[ ] Documentation
Issues Fixed or Closed by This PR
Fixes:
Checklist
[ ] My code follows the code style of this project.
[ ] My change requires a change to the documentation.
[ ] I have updated the documentation accordingly.
[ ] I have read the CONTRIBUTING document.
[ ] I have added tests to cover my changes.
[ ] I have tested the changes and verified that they work and don't break anything (as well as I can manage).
fixed by #633
| gharchive/pull-request | 2023-08-17T09:45:56 | 2025-04-01T06:44:26.673103 | {
"authors": [
"alessiamarcolini"
],
"repo": "histolab/histolab",
"url": "https://github.com/histolab/histolab/pull/619",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1816851539 | Merge a QLoRA to base Llama 2
I used python src/train_web.py to train a QLoRA. How do i merge the QLoRA to the base model Llama 2?
Please visit the export tab of the Web Tuner to merge the LoRA weights.
Please visit the export tab of the Web Tuner to merge the LoRA weights.
I selected and load QLoRa Checkpoints and base and got error:
07/24/2023 20:59:02 - WARNING - llmtuner.tuner.core.parser - Please specify `prompt_template` if you are using other pre-trained models.
[INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,572 >> loading file tokenizer.model
[INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file added_tokens.json
[INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file special_tokens_map.json
[INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file tokenizer_config.json
[INFO|configuration_utils.py:710] 2023-07-24 20:59:02,580 >> loading configuration file C:\LLaMA-Efficient-Tuning\Llama-2-13B-Chat-fp16\config.json
[INFO|configuration_utils.py:768] 2023-07-24 20:59:02,581 >> Model config LlamaConfig {
"_name_or_path": "C:\\LLaMA-Efficient-Tuning\\Llama-2-13B-Chat-fp16",
"architectures": [
"LlamaForCausalLM"
],
"bos_token_id": 1,
"eos_token_id": 2,
"hidden_act": "silu",
"hidden_size": 5120,
"initializer_range": 0.02,
"intermediate_size": 13824,
"max_length": 4096,
"max_position_embeddings": 4096,
"model_type": "llama",
"num_attention_heads": 40,
"num_hidden_layers": 40,
"num_key_value_heads": 40,
"pad_token_id": 0,
"pretraining_tp": 1,
"rms_norm_eps": 1e-05,
"rope_scaling": null,
"tie_word_embeddings": false,
"torch_dtype": "float16",
"transformers_version": "4.31.0",
"use_cache": true,
"vocab_size": 32000
}
[INFO|modeling_utils.py:2600] 2023-07-24 20:59:02,582 >> loading weights file C:\LLaMA-Efficient-Tuning\Llama-2-13B-Chat-fp16\pytorch_model.bin.index.json
[INFO|modeling_utils.py:1172] 2023-07-24 20:59:02,583 >> Instantiating LlamaForCausalLM model under default dtype torch.float16.
[INFO|configuration_utils.py:599] 2023-07-24 20:59:02,583 >> Generate config GenerationConfig {
"_from_model_config": true,
"bos_token_id": 1,
"eos_token_id": 2,
"max_length": 4096,
"pad_token_id": 0,
"transformers_version": "4.31.0"
}
Traceback (most recent call last):
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\routes.py", line 442, in run_predict
output = await app.get_blocks().process_api(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1389, in process_api
result = await self.call_function(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1108, in call_function
prediction = await utils.async_iteration(iterator)
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 346, in async_iteration
return await iterator.__anext__()
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 339, in __anext__
return await anyio.to_thread.run_sync(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 807, in run
result = context.run(func, *args)
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 322, in run_sync_iterator_async
return next(iterator)
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 691, in gen_wrapper
yield from f(*args, **kwargs)
File "C:\LLaMA-Efficient-Tuning\src\llmtuner\webui\utils.py", line 122, in export_model
model, tokenizer = load_model_and_tokenizer(model_args, finetuning_args)
File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\core\loader.py", line 105, in load_model_and_tokenizer
model = AutoModelForCausalLM.from_pretrained(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\models\auto\auto_factory.py", line 493, in from_pretrained
return model_class.from_pretrained(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\modeling_utils.py", line 2903, in from_pretrained
) = cls._load_pretrained_model(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\modeling_utils.py", line 3002, in _load_pretrained_model
raise ValueError(
ValueError: The current `device_map` had weights offloaded to the disk. Please provide an `offload_folder` for them. Alternatively, make sure you have `safetensors` installed if the model you are using offers the weights in this format.
Please visit the export tab of the Web Tuner to merge the LoRA weights.
I selected and load QLoRa Checkpoints and base and got error:
07/24/2023 20:59:02 - WARNING - llmtuner.tuner.core.parser - Please specify `prompt_template` if you are using other pre-trained models.
[INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,572 >> loading file tokenizer.model
[INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file added_tokens.json
[INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file special_tokens_map.json
[INFO|tokenization_utils_base.py:1837] 2023-07-24 20:59:02,573 >> loading file tokenizer_config.json
[INFO|configuration_utils.py:710] 2023-07-24 20:59:02,580 >> loading configuration file C:\LLaMA-Efficient-Tuning\Llama-2-13B-Chat-fp16\config.json
[INFO|configuration_utils.py:768] 2023-07-24 20:59:02,581 >> Model config LlamaConfig {
"_name_or_path": "C:\\LLaMA-Efficient-Tuning\\Llama-2-13B-Chat-fp16",
"architectures": [
"LlamaForCausalLM"
],
"bos_token_id": 1,
"eos_token_id": 2,
"hidden_act": "silu",
"hidden_size": 5120,
"initializer_range": 0.02,
"intermediate_size": 13824,
"max_length": 4096,
"max_position_embeddings": 4096,
"model_type": "llama",
"num_attention_heads": 40,
"num_hidden_layers": 40,
"num_key_value_heads": 40,
"pad_token_id": 0,
"pretraining_tp": 1,
"rms_norm_eps": 1e-05,
"rope_scaling": null,
"tie_word_embeddings": false,
"torch_dtype": "float16",
"transformers_version": "4.31.0",
"use_cache": true,
"vocab_size": 32000
}
[INFO|modeling_utils.py:2600] 2023-07-24 20:59:02,582 >> loading weights file C:\LLaMA-Efficient-Tuning\Llama-2-13B-Chat-fp16\pytorch_model.bin.index.json
[INFO|modeling_utils.py:1172] 2023-07-24 20:59:02,583 >> Instantiating LlamaForCausalLM model under default dtype torch.float16.
[INFO|configuration_utils.py:599] 2023-07-24 20:59:02,583 >> Generate config GenerationConfig {
"_from_model_config": true,
"bos_token_id": 1,
"eos_token_id": 2,
"max_length": 4096,
"pad_token_id": 0,
"transformers_version": "4.31.0"
}
Traceback (most recent call last):
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\routes.py", line 442, in run_predict
output = await app.get_blocks().process_api(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1389, in process_api
result = await self.call_function(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1108, in call_function
prediction = await utils.async_iteration(iterator)
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 346, in async_iteration
return await iterator.__anext__()
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 339, in __anext__
return await anyio.to_thread.run_sync(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 807, in run
result = context.run(func, *args)
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 322, in run_sync_iterator_async
return next(iterator)
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 691, in gen_wrapper
yield from f(*args, **kwargs)
File "C:\LLaMA-Efficient-Tuning\src\llmtuner\webui\utils.py", line 122, in export_model
model, tokenizer = load_model_and_tokenizer(model_args, finetuning_args)
File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\core\loader.py", line 105, in load_model_and_tokenizer
model = AutoModelForCausalLM.from_pretrained(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\models\auto\auto_factory.py", line 493, in from_pretrained
return model_class.from_pretrained(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\modeling_utils.py", line 2903, in from_pretrained
) = cls._load_pretrained_model(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\modeling_utils.py", line 3002, in _load_pretrained_model
raise ValueError(
ValueError: The current `device_map` had weights offloaded to the disk. Please provide an `offload_folder` for them. Alternatively, make sure you have `safetensors` installed if the model you are using offers the weights in this format.
Try removing the following lines and checking if it works.
https://github.com/hiyouga/LLaMA-Efficient-Tuning/blob/182b42504399d2755897b9737db1d36655a0fa50/src/llmtuner/tuner/core/loader.py#L96-L97
如何启用QLoRA?在量化处选择4bit就可以吗?
如何启用QLoRA?在量化处选择4bit就可以吗?
是的
In our experiments, the performance of QLoRA is close to LoRA's one.
In addition of WeChat, possibly add public discord servers?
@hiyouga after update to last version and use export tab with a selected Prompt template, got error:
Traceback (most recent call last):
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\routes.py", line 442, in run_predict
output = await app.get_blocks().process_api(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1389, in process_api
result = await self.call_function(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\blocks.py", line 1108, in call_function
prediction = await utils.async_iteration(iterator)
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 346, in async_iteration
return await iterator.__anext__()
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 339, in __anext__
return await anyio.to_thread.run_sync(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\anyio\_backends\_asyncio.py", line 807, in run
result = context.run(func, *args)
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 322, in run_sync_iterator_async
return next(iterator)
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\gradio\utils.py", line 691, in gen_wrapper
yield from f(*args, **kwargs)
File "C:\LLaMA-Efficient-Tuning\src\llmtuner\webui\utils.py", line 125, in save_model
export_model(args, max_shard_size="{}GB".format(max_shard_size))
File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\tune.py", line 29, in export_model
model_args, _, training_args, finetuning_args, _ = get_train_args(args)
File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\core\parser.py", line 54, in get_train_args
model_args, data_args, training_args, finetuning_args, general_args = parse_train_args(args)
File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\core\parser.py", line 39, in parse_train_args
return _parse_args(parser, args)
File "C:\LLaMA-Efficient-Tuning\src\llmtuner\tuner\core\parser.py", line 24, in _parse_args
return parser.parse_dict(args)
File "C:\LLaMA-Efficient-Tuning\venv\lib\site-packages\transformers\hf_argparser.py", line 373, in parse_dict
obj = dtype(**inputs)
TypeError: DataArguments.__init__() missing 1 required positional argument: 'template'
@Katehuuh Fixed
| gharchive/issue | 2023-07-22T18:14:13 | 2025-04-01T06:44:26.693770 | {
"authors": [
"DumoeDss",
"Katehuuh",
"PsychoSmiley",
"hiyouga"
],
"repo": "hiyouga/LLaMA-Efficient-Tuning",
"url": "https://github.com/hiyouga/LLaMA-Efficient-Tuning/issues/223",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2197854829 | 🛑 HJStrauss is down
In 92a7383, HJStrauss (https://www.hjstrauss.de) was down:
HTTP code: 0
Response time: 0 ms
Resolved: HJStrauss is back up in cfd8a72 after 9 minutes.
| gharchive/issue | 2024-03-20T15:36:19 | 2025-04-01T06:44:26.719872 | {
"authors": [
"hjstrauss"
],
"repo": "hjstrauss/MonitorMySites",
"url": "https://github.com/hjstrauss/MonitorMySites/issues/237",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2286779397 | 🛑 Reime ohne Sinn und Verstand is down
In 0ba04e4, Reime ohne Sinn und Verstand (https://www.reimeohnesinnundverstand.de) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Reime ohne Sinn und Verstand is back up in f1d673a after 9 minutes.
| gharchive/issue | 2024-05-09T02:34:37 | 2025-04-01T06:44:26.722465 | {
"authors": [
"hjstrauss"
],
"repo": "hjstrauss/MonitorMySites",
"url": "https://github.com/hjstrauss/MonitorMySites/issues/3193",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2190771210 | 🛑 LM Bogen WSV is down
In 9974047, LM Bogen WSV (https://www.lmbogenwsv.de) was down:
HTTP code: 0
Response time: 0 ms
Resolved: LM Bogen WSV is back up in 5e1829d after 8 minutes.
| gharchive/issue | 2024-03-17T17:17:57 | 2025-04-01T06:44:26.724929 | {
"authors": [
"hjstrauss"
],
"repo": "hjstrauss/MonitorMySites",
"url": "https://github.com/hjstrauss/MonitorMySites/issues/48",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2382260447 | 🛑 Reime ohne Sinn und Verstand is down
In 61035ef, Reime ohne Sinn und Verstand (https://www.reimeohnesinnundverstand.de) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Reime ohne Sinn und Verstand is back up in 4d411a4 after 8 minutes.
| gharchive/issue | 2024-06-30T13:40:42 | 2025-04-01T06:44:26.727351 | {
"authors": [
"hjstrauss"
],
"repo": "hjstrauss/MonitorMySites",
"url": "https://github.com/hjstrauss/MonitorMySites/issues/5975",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1038456956 | API available
FYI, Fuji released a SDK in April 2021
https://fujifilm-x.com/global/special/camera-control-sdk/
hmph
| gharchive/issue | 2021-10-28T12:27:37 | 2025-04-01T06:44:26.747560 | {
"authors": [
"sahver",
"tensiondriven"
],
"repo": "hkr/fuji-cam-wifi-tool",
"url": "https://github.com/hkr/fuji-cam-wifi-tool/issues/20",
"license": "unlicense",
"license_type": "permissive",
"license_source": "bigquery"
} |
738299072 | fixed send_line function
closes #172
Thanks a lot for the PR and for helping keep iron working :slightly_smiling_face:
Best regards,
Henry
| gharchive/pull-request | 2020-11-07T18:55:44 | 2025-04-01T06:44:26.752533 | {
"authors": [
"doubleloop",
"hkupty"
],
"repo": "hkupty/iron.nvim",
"url": "https://github.com/hkupty/iron.nvim/pull/173",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
435354071 | error build with compile-private
➜ .emacs.d (develop) ✗ make compile-private (develop|✚2…
Warning: byte compiling is for advanced users. It will interfere with your
efforts to debug issues. It is not recommended you do it if you frequently
tinker with your Emacs config.
Alternatively, use `bin/doom compile :core` instead to byte-compile only the
Doom core files, as these don't change often.
If you have issues, please make sure byte-compilation isn't the cause by using
`bin/doom clean` to clear out your *.elc files.
Byte-compile anyway? (y or n) y
Note, built-in variable ‘overflow-newline-into-fringe’ not bound
⚠ Ignored ~/.config/doom/packages.el
⚠ Ignored ~/.config/doom/lang.el
✓ Compiled ~/.config/doom/init.el
simple profile: ON
Eager macro-expansion failure: (error "featurep! couldn’t detect what module its in! (in /Users/jiya/.config/doom/config.el)")
[Treemacs] Warning: coudn’t find default background color for icons, falling back on #2d2d31.
[Treemacs] Warning: couldn’t find hl-line-mode’s background color for icons, falling back on #2d2d31.
✓ Compiled ~/.config/doom/config.el
get an error message Eager macro-expansion failure: (error "featurep! couldn’t detect what module its in! (in /Users/jiya/.config/doom/config.el)").
Can you show me how you are using featurep! in your config?
./lang.el:31:;; (when (featurep! :ui doom)
./+ui.el:7:(when (featurep! :ui doom)
./+ui.el:115: (if (featurep! :ui doom)
./+ui.el:368:;; (unless (featurep! :ui doom-quit)
./config.el:152: :when (featurep! :tools flycheck)
This error should only occur if you use featurep! with only one argument; which none of those are doing, so it's unclear to me why this would happen.
Can you remove them one by one and see which one is causing it?
remove all featurep! but still error.
and in my config featurep! like
(and (featurep! +lsp) (not (featurep! +tabnine))
(and (featurep! +lsp) (not (featurep! +tabnine))
This is the kind of code I was expecting. It is incorrect to use featurep! this way from your private config. You must use the full module path, e.g.
(and (featurep! :tools lsp)
(not (featurep! :completion company +tabnine))
Otherwise, it doesn't know where to look for the +lsp or +tabnine flags.
I will assume this issue has been resolved. If that isn't the case, feel free to follow up here.
| gharchive/issue | 2019-04-20T04:02:54 | 2025-04-01T06:44:26.767936 | {
"authors": [
"hlissner",
"shuxiao9058"
],
"repo": "hlissner/doom-emacs",
"url": "https://github.com/hlissner/doom-emacs/issues/1346",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
779814932 | Our Github issue tracker is temporarily locked down
Hey folks! You may have noticed that our Github issue tracker is locked down; only previous contributors can post new issues/PRs or respond to old ones. This is temporary and will elapse in 7 days, while I clean up our issue tracker and transition to Discourse.
As some of you know, I'm working on a Discourse server for Doom Emacs and I have prepared for it to go live some time next week. It will replace Discord as our primary platform for user support and replace Github as our issue tracker. There will be a post on the Discourse explaining the rationale for this change and what it means for our users, but going forward, our Github issue tracker will be reserved strictly for internal project management and issues escalated from the Discourse.
You may have also noticed that two of our project boards have been closed. They were also moved to the Discourse.
I realize this is an inconvenience for folks with issues to report, but please bare with me for a short while. I'll keep you posted!
The issue tracker is open once again. I've collected what I needed from it to move to Discourse and will soon publish it; when I do I'll begin diverting traffic there.
The issue tracker is open once again. I've collected what I needed from it to move to Discourse and will soon publish it; when I do I'll begin diverting traffic there.
| gharchive/issue | 2021-01-06T00:59:11 | 2025-04-01T06:44:26.770954 | {
"authors": [
"hlissner"
],
"repo": "hlissner/doom-emacs",
"url": "https://github.com/hlissner/doom-emacs/issues/4510",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
797392276 | Remap describe-symbol to counsel-describe-symbol .
"SPC h o" now uses counsel-describe-symbol instead of plain helpful-symbol (maybe with completion) to get input.
This allows us to fully take the advantage of ivy's superior completion system:
[I]: info and [d]: definition hydra heads in ivy-dispatching-done
Quickly describe symbols with "C-M-n/p" (ivy-next(or previous)-line-and-call or with the calling hydra head on
Additionally the counsel one is prettier when icons are enabled.
Users who have the +icons flag with ivy are expected to notice the change visually when invoking "o" in help-map as icons now appear there. But it's an arguably good change anyway since originally "f1 f" and "f1 v" have icons but "f1 o" doesn't.
Excuse me, I didn't know that helpful-symbol asks an additional question when the symbol refers to both a command and a variable.
Guess I have to personally stick to the plain old describe-symbol instead to use the Counsel version since "C-M-n/p" seems far too useful to me.
Excuse me, I didn't know that helpful-symbol asks an additional question when the symbol refers to both a command and a variable.
Guess I have to personally stick to the plain old describe-symbol instead to use the Counsel version since "C-M-n/p" seems far too useful to me.
Thanks!
Thanks!
| gharchive/pull-request | 2021-01-30T11:04:53 | 2025-04-01T06:44:26.775971 | {
"authors": [
"datool",
"hlissner"
],
"repo": "hlissner/doom-emacs",
"url": "https://github.com/hlissner/doom-emacs/pull/4593",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1644330956 | Add Community Listing Block
Already resolved by #118
| gharchive/issue | 2023-03-28T16:50:33 | 2025-04-01T06:44:26.777115 | {
"authors": [
"badvision",
"bstopp"
],
"repo": "hlxsites/hsf-commonmoves",
"url": "https://github.com/hlxsites/hsf-commonmoves/issues/23",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1704529248 | Fixed aria attributes and link crawling
Issue
Fixed Part
all images that are linked, that don't have linked text or an alt tag, should have aria-label added to describe the function of the link
Check for links that are not crawlable, such as Close:
Test URLs
Before: https://main--merative2--hlxsites.hlx.page/company
After: https://aria-attributes--merative2--hlxsites.hlx.page/company
Description
Is there a GitHub issue this PR can be linked to?
| gharchive/pull-request | 2023-05-10T19:43:01 | 2025-04-01T06:44:26.780043 | {
"authors": [
"VladimirZaets",
"arumsey"
],
"repo": "hlxsites/merative2",
"url": "https://github.com/hlxsites/merative2/pull/134",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
953772324 | Pay 4882 testing
Before creating a pull request make sure that:
[ ] commit messages are meaningful and follow good commit message guidelines
[ ] README and other documentation has been updated / added (if needed)
[ ] tests have been updated / new tests has been added (if needed)
Please remove this line and everything above and fill the following sections:
JIRA link (if applicable)
Change description
Does this PR introduce a breaking change? (check one with "x")
[ ] Yes
[ ] No
Closing Pull as we have some merge issues in the pipeline....
| gharchive/pull-request | 2021-07-27T11:13:17 | 2025-04-01T06:44:26.954207 | {
"authors": [
"pats-john"
],
"repo": "hmcts/ccpay-refunds-app",
"url": "https://github.com/hmcts/ccpay-refunds-app/pull/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2413931308 | CIV-14290 added publish instructions so WA
JIRA link (if applicable)
CIV-14290
Change description
CCD changes for WA part
Does this PR introduce a breaking change? (check one with "x")
[ ] Yes
[X] No
Is Create_SDO_DJ missing here?
Definitely right, STANDARD_DIRECTION_ORDER_DJ was missing
| gharchive/pull-request | 2024-07-17T15:31:11 | 2025-04-01T06:44:26.955365 | {
"authors": [
"dtortolaV1"
],
"repo": "hmcts/civil-ccd-definition",
"url": "https://github.com/hmcts/civil-ccd-definition/pull/4639",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2383305276 | CIV-13828 remove all notifications when claim is struck out
JIRA link (if applicable)
CIV-13828
Change description
Before adding the new notifications, allow "*" to remove all existing notifications from both Claimant and Defendant
Does this PR introduce a breaking change? (check one with "x")
[ ] Yes
[X] No
Replaced by https://github.com/hmcts/civil-service/pull/5051
| gharchive/pull-request | 2024-07-01T09:20:07 | 2025-04-01T06:44:26.956623 | {
"authors": [
"dtortolaV1"
],
"repo": "hmcts/civil-service",
"url": "https://github.com/hmcts/civil-service/pull/5015",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2488783629 | DTSSTCI1086 - removing E2E's and all references
Change description
The old E2E tests are causing a repository failure - these need to be removed from Case-API as they have been replaced by the separate E2E repository.
JIRA link
DTSSTCI-1086
Before merging a pull request make sure that:
[x] tests have been updated / new tests has been added (if needed)
[x] README and other documentation has been updated / added (if needed)
[x] enable-e2e-tests label can be used to run the e2e tests before QA handover and before release (required)
If this ticket will have any visible impact on users and is not behind a feature toggle, make sure that:
[x] this ticket has been reviewed by QA
[x] the user story has been signed off by the PO
Note that bug fixes, dependency updates and technical tasks do not directly impact the user experience and can be merged without QA and PO review.
If this user story cannot be immediately merged find a way to put it behind a feature toggle and get it merged.
Nightly-dev pipeline passing
| gharchive/pull-request | 2024-08-27T09:28:29 | 2025-04-01T06:44:26.958997 | {
"authors": [
"JakeCG"
],
"repo": "hmcts/sptribs-case-api",
"url": "https://github.com/hmcts/sptribs-case-api/pull/1886",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
497890119 | Re-enable --shell-escape for LaTeX
Similar to f7bb61b5166e, but for main wscript. Was there any reason to disable it that I cannot remember? Leaving it on is crucial for the Stata template because the only thing Stata consistently produces are eps-files and they need conversion to pdf before being useful to pdflatex.
we disabled it for azure, is it an option to save the stuff as png instead of eps, or directly as pdf (do we need shell escape then?)
No, at least older versions of Stata relied on some library to generate pngs that was only available as GUI... Maybe they took the graph they displayed and made a screenshot :-)
So we should leave the comment where it is unless we are using the Stata example.
i am pretty sure, stata 13 at least already supports png, do we really need to care for older versions?
Yes, but to my best recollection only the GUI version of Stata, not the command line version, which is what we need.
fixed in #65
Just for reference: https://www.stata.com/statalist/archive/2011-07/msg00536.html
| gharchive/issue | 2019-09-24T19:49:52 | 2025-04-01T06:44:26.960807 | {
"authors": [
"hmgaudecker",
"raholler"
],
"repo": "hmgaudecker/econ-project-templates",
"url": "https://github.com/hmgaudecker/econ-project-templates/issues/64",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
730108621 | pdfcrop --xetex does not give correct output
Here is the code to reproduce the issue.
generate a pdf page (named as abc.pdf) from the following code:
\documentclass{standalone}
\begin{document}
\Huge A sample text.
\end{document}
use pdfcrop to remove its margins (actually it does not have white margins)
pdfcrop --xetex abc.pdf
you will see the output have two pages, and both are empty.
If replacing standalone with article, the output also have two pages, but the first page is correct, and the second page is also empty.
pdfcrop --pdftex abc.pdf
pdfcrop --luatex abc.pdf
both could give correct results.
I have run that on my Windows 10 system with TeXLive 2020 full installed and updated. The pdfcrop version is PDFCROP 1.40, 2020/06/06.
Best,
Jinsong
Well the second page is because of the special for the pdf version. It can probably be moved into the box.
The text seems to disappear with standalone as the paper size is too small now: with the 1in displacement inserted by the code nothing is in the box, and then it gets dropped. This is perhaps a xetex bug but imho nothing pdfcrop can solve.
With the example file:
\documentclass{standalone}
\begin{document}
\Huge A sample text.
\end{document}
and running:
$ xelatex test.tex
$ pdfcrop --xetexcmd xelatex test.pdf
I get the correct output file:
One interesting thing I have noticed here is the following. When running pdfcrop --xetex test.pdf the line executed by pdfcrop is as follows:
* XeTeX call: xetex -no-shell-escape -interaction=nonstopmode tmp-pdfcrop-9248
i.e., it uses plain XeTeX, which has an effect on the fonts of the file:
$ pdffonts test.pdf
name type encoding emb sub uni object ID
------------------------------------ ----------------- ---------------- --- --- --- ---------
SYMVRB+LMRoman17-Regular-Identity-H CID Type 0C Identity-H yes yes yes 4 0
$ pdffonts test.pdf
name type encoding emb sub uni object ID
------------------------------------ ----------------- ---------------- --- --- --- ---------
SYMVRB+LMRoman17-Regular-Identity-H CID Type 0C Identity-H yes yes yes 4 0
$ pdfcrop --xetex test.pdf
PDFCROP 1.40, 2020/06/06 - Copyright (c) 2002-2020 by Heiko Oberdiek, Oberdiek Package Support Group.
==> 1 page written on `test-crop.pdf'.
$ pdffonts test-crop.pdf
name type encoding emb sub uni object ID
------------------------------------ ----------------- ---------------- --- --- --- ---------
SYMVRB+LMRoman17-Regular-Identity-H CID Type 0C Identity-H yes yes yes 4 0
DLSGEY+CMR10 Type 1C Builtin yes yes yes 16 0
Add the font CMR10 and that affects the generated pdf-crop file, but when executing the line:
$ pdffonts test.pdf
name type encoding emb sub uni object ID
------------------------------------ ----------------- ---------------- --- --- --- ---------
SYMVRB+LMRoman17-Regular-Identity-H CID Type 0C Identity-H yes yes yes 4 0
$ pdfcrop --xetexcmd xelatex test.pdf
PDFCROP 1.40, 2020/06/06 - Copyright (c) 2002-2020 by Heiko Oberdiek, Oberdiek Package Support Group.
==> 1 page written on `test-crop.pdf'.
$ pdffonts test-crop.pdf
name type encoding emb sub uni object ID
------------------------------------ ----------------- ---------------- --- --- --- ---------
SYMVRB+LMRoman17-Regular-Identity-H CID Type 0C Identity-H yes yes yes 8 0
everything works OK.
I hope this is helpful.
Saludos
fixed in v1.41
| gharchive/issue | 2020-10-27T05:01:05 | 2025-04-01T06:44:27.002844 | {
"authors": [
"davidcarlisle",
"jszhao",
"pablgonz",
"u-fischer"
],
"repo": "ho-tex/pdfcrop",
"url": "https://github.com/ho-tex/pdfcrop/issues/3",
"license": "LPPL-1.3c",
"license_type": "permissive",
"license_source": "github-api"
} |
345034483 | Beispiel-Flow überwachung Syslog erstellen
/var/log/messages überwachen und Häufigkeit von ReGa-Fehlermeldungen ermitteln, bei überschreiten einer Häufigkeit x per Email benachrichtigen
Ich weiß nicht was dir vorschwebt, aber so könnte es aussehen:
[{"id":"50104e94.8152b8","type":"tab","label":"Error Monitoring","disabled":false,"info":""},{"id":"5d48d65c.e24b48","type":"tail","z":"50104e94.8152b8","name":"","filetype":"text","split":true,"filename":"/var/log/messages","x":130,"y":160,"wires":[["6595e35d.530f24"]]},{"id":"6595e35d.530f24","type":"switch","z":"50104e94.8152b8","name":"Filter Errors","property":"payload","propertyType":"msg","rules":[{"t":"regex","v":"^.*Error\\:.*$","vt":"str","case":false}],"checkall":"true","repair":false,"outputs":1,"x":330,"y":160,"wires":[["e9d593c5.3deb1"]]},{"id":"e9d593c5.3deb1","type":"counter","z":"50104e94.8152b8","name":"Count Errors","init":"0","step":"1","lower":"","upper":"","mode":"increment","outputs":2,"x":510,"y":160,"wires":[["253f1e33.ae4aea"],["ba691feb.d30d08","bd901793.899028"]]},{"id":"253f1e33.ae4aea","type":"ui_gauge","z":"50104e94.8152b8","name":"Show Error Count","group":"d79a8eea.a905c","order":0,"width":"6","height":"4","gtype":"gage","title":"Anzahl Fehler","label":"Fehler","format":"{{value}}","min":0,"max":"100","colors":["#00b500","#e6e600","#ca3838"],"seg1":"","seg2":"","x":710,"y":100,"wires":[]},{"id":"ba691feb.d30d08","type":"function","z":"50104e94.8152b8","name":"Rotate Entries","func":"var dashboardLog = context.get('dashboardLog')|| [];\n \ndashboardLog.push(msg);\nif (dashboardLog.length > 20) {\n // Delete oldest message if > 20\n dashboardLog.shift();\n dashboardLog.length = 20;\n} \n \n// store the value back\ncontext.set('dashboardLog',dashboardLog);\n \n// make it part of the outgoing msg object\nmsg = {};\nmsg.payload = dashboardLog;\nreturn msg;","outputs":1,"noerr":0,"x":700,"y":220,"wires":[["f4ef2a5d.f01c18"]]},{"id":"bd901793.899028","type":"join","z":"50104e94.8152b8","name":"Prepare Message","mode":"custom","build":"string","property":"payload","propertyType":"msg","key":"topic","joiner":"\\r\\n\\r\\n","joinerType":"str","accumulate":false,"timeout":"","count":"5","reduceRight":false,"reduceExp":"","reduceInit":"","reduceInitType":"num","reduceFixup":"","x":710,"y":160,"wires":[["2e0650.3ce4e9b"]]},{"id":"f4ef2a5d.f01c18","type":"ui_template","z":"50104e94.8152b8","group":"2bf72e51.a156aa","name":"Event Log","order":0,"width":"6","height":"4","format":"<ul>\n <li ng-repeat=\"x in msg.payload\">\n <font color=\"red\">{{x.topic}}</font>\n <ul>\n <li>{{x.payload}}</li>\n </ul>\n </li>\n</ul>","storeOutMessages":true,"fwdInMessages":true,"templateScope":"local","x":900,"y":220,"wires":[[]]},{"id":"2e0650.3ce4e9b","type":"e-mail","z":"50104e94.8152b8","server":"smtp.gmail.com","port":"465","secure":true,"name":"","dname":"Send Email","x":910,"y":160,"wires":[]},{"id":"b8d3f654.c03e9","type":"comment","z":"50104e94.8152b8","name":"Fehlerüberwachung","info":"Überwacht /var/log/messages\nZeigt die letzten 20 Fehler in einem Log-Fenster\nZeigt die Gesamtzahl der Fehler\nSchickt Email nach 5 Fehlern mit Fehlerprotokoll","x":130,"y":80,"wires":[]},{"id":"d79a8eea.a905c","type":"ui_group","z":"","name":"Monitoring","tab":"a5b5c3f2.ffeeb8","disp":true,"width":"6","collapse":false},{"id":"2bf72e51.a156aa","type":"ui_group","z":"","name":"EventLog","tab":"a5b5c3f2.ffeeb8","disp":true,"width":"6","collapse":false},{"id":"a5b5c3f2.ffeeb8","type":"ui_tab","z":"","name":"Monitoring","icon":"dashboard"}]
Zusätzliche Node: https://www.npmjs.com/package/node-red-contrib-counter
Hey vielen Dank! Werde bei Gelegenheit ein bebilderten kommentierten Wiki Artikel draus machen, außer Du kommst mir zuvor ;-)
Kann ich gerne übernehmen. Deine Zeit ist besser in der Entwicklung investiert ;-)
https://github.com/hobbyquaker/RedMatic/wiki/Fehlerüberwachung-der-CCU
Du müsstest mir bitte noch den Screenshot Ordner hochladen.
errorlog.zip
| gharchive/issue | 2018-07-26T23:03:14 | 2025-04-01T06:44:27.024585 | {
"authors": [
"Sineos",
"hobbyquaker"
],
"repo": "hobbyquaker/RedMatic",
"url": "https://github.com/hobbyquaker/RedMatic/issues/44",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2367363669 | Add in pull request guidelines and fix 'stolen'/copied code.
Hi @hoeken
I'm concerned that nothing has been done in regards to pull request #89
I spent a lot of time researching, writing, and documenting my TemplatePrinter and the pull request was ready to merge for months.
The user who submitted #89 copied my code, not merged my commits, and then submitted my work in their own commits, which means that there is no link to the actual author of this work.
I create and submit code for open source projects to benefit others, and more importantly build a portfolio of my work that other projects and future employers can view. By allowing another user to submit my work, I have no proof that I wrote that code, and others cannot find any reference to it from my profile.
I really must ask that you do what is right and reverse that change and submit my PR (#56) if you want the TemplatePrinter in your project.
As my code is unrelated to #89 it should not have been included there. That PR should be reverted and they should resubmit their own work only. This will not only make your repo cleaner, but people will be able to find the PR and justification behind its implementation.
I also recommend you add in some guidelines to prevent this from happening in the future.
Hi @Chris--A
I agree 100% and I'm sorry I messed this one up. This one is definitely my fault. I was just going through all the PRs and trying to clear the queue out. I saw that the #89 pr included a bunch of others and my brain said that it was the easiest way. It didn't cross my mind that you would want the original commit merged and that was an oversight.
Last week and this week are both pretty hectic for me, but as soon as I get some free time I will unwind this and merge everything in properly. Your contributions are very much appreciated. It wasn't an intentional slight or anything so hopefully no hard feelings. :)
@Chris--A I re-opened and merged your code properly + added attribution in the upcoming changelog.
@hoeken : it is also possible to activate this option to force contributors to accept that the PR they submit is in accordance with the projet's license:
When contributing a PR in an LGPL project, attribution is not a "right" because the work should have been submitted under the project's license.
Most projects are using SPDX generated headers so there is no attribution on any file / header but just the SPDX header.
Submitting a PR with a work requiring attribution would mean to change the license of this specific file only, and depending on the project this can be an issue, for example someone might want to submit PR with some GPL additions then preventing the library in LGPL to be used commercially.
That's why contributors should / have to accept to PR their work under a projet's license, and often, attribution is not automatic.
This is very nice from @hoeken to maintain such list.
yeah that's a good idea.
| gharchive/issue | 2024-06-21T23:46:09 | 2025-04-01T06:44:27.041316 | {
"authors": [
"Chris--A",
"hoeken",
"mathieucarbou"
],
"repo": "hoeken/PsychicHttp",
"url": "https://github.com/hoeken/PsychicHttp/issues/100",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
164577143 | Build Failures
$ perl6 -v
This is Rakudo version 2016.06-10-g7cff429 built on MoarVM version 2016.06
implementing Perl 6.c.
$ zef install Linenoise
===> Searching for: Linenoise
===> Searching for missing dependencies: Native::Resources
===> Fetching: Linenoise
===> Fetching: Native::Resources
===> Building: Linenoise:ver('0.1.0'):auth('Rob Hoelz')
`build-depends` is missing entries. Attemping to mimick missing dependencies...
Cannot invoke this object (REPR: Null; VMNull)
in sub get-vars at /home/zoffix/.rakudobrew/moar-nom/install/share/perl6/site/sources/D57EF38F5287179FB7D408925E155FB671060388 (LibraryMake) line 95
in sub get-vars at /home/zoffix/.zef/store/p6-native-resources.git/b05fd025d5a3541e88b5acc266446d3814d79cc0/lib/Native/Resources/Build.pm (Native::Resources::Build) line 13
in sub make at /home/zoffix/.zef/store/p6-native-resources.git/b05fd025d5a3541e88b5acc266446d3814d79cc0/lib/Native/Resources/Build.pm (Native::Resources::Build) line 27
in method build at /home/zoffix/.zef/store/p6-linenoise.git/46f9534b1568e68a2a077ee80713914738b16e28/Build.pm (Build) line 13
in block <unit> at -e line 1
===> Building [FAIL]: Linenoise:ver('0.1.0'):auth('Rob Hoelz')
===SORRY!===
Aborting due to build failure: Linenoise:ver('0.1.0'):auth('Rob Hoelz')(use --force to override)
@zoffixznet Would you mind trying this again? I just tried installing Linenoise with both panda and zef, and both worked for me.
Tried it on a fresh build and everything worked fine.
¯\_(ツ)_/¯ maybe it's just my other box is busted.
cpan@perlbuild2:~$ zef install Linenoise
===> Searching for: Linenoise
===> Found: Linenoise:ver('0.1.0'):auth('Rob Hoelz')
===> Dependencies: Native::Resources
===> Searching for missing dependencies: Native::Resources
===> Found dependencies: Native::Resources:ver('0.1.0'):auth('Rob Hoelz')
===> Dependencies: LibraryMake
===> Searching for missing dependencies: LibraryMake
===> Found dependencies: LibraryMake:ver('1.0.0'):auth('github:retupmoca')
===> Dependencies: Shell::Command
===> Searching for missing dependencies: Shell::Command
===> Found dependencies: Shell::Command
===> Dependencies: File::Which, File::Find, Test
===> Searching for missing dependencies: File::Which, File::Find
===> Found dependencies: File::Find:ver('0.1'), File::Which
===> Fetching: Linenoise
Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Fetched: Linenoise to /home/cpan/.zef/tmp/p6-linenoise.git
===> Extracting: Linenoise
Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Extracted: Linenoise to /home/cpan/.zef/store/p6-linenoise.git/46f9534b1568e68a2a077ee80713914738b16e28
===> Fetching: Native::Resources
Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Fetched: Native::Resources to /home/cpan/.zef/tmp/p6-native-resources.git
===> Extracting: Native::Resources
Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Extracted: Native::Resources to /home/cpan/.zef/store/p6-native-resources.git/b05fd025d5a3541e88b5acc266446d3814d79cc0
===> Fetching: LibraryMake
Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Fetched: LibraryMake to /home/cpan/.zef/tmp/P6-LibraryMake.git
===> Extracting: LibraryMake
Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Extracted: LibraryMake to /home/cpan/.zef/store/P6-LibraryMake.git/8ab4b23a981ebfd63a74c11d1398750a472195aa
===> Fetching: Shell::Command
Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Fetched: Shell::Command to /home/cpan/.zef/tmp/Shell-Command.git
===> Extracting: Shell::Command
Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Extracted: Shell::Command to /home/cpan/.zef/store/Shell-Command.git/3d21c0ffbc467f121935a3ec74e54c23b651836e
===> Fetching: File::Find
Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Fetched: File::Find to /home/cpan/.zef/tmp/File-Find.git
===> Extracting: File::Find
Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Extracted: File::Find to /home/cpan/.zef/store/File-Find.git/d3e2be709aeb2479dc02b154fdc179fe7a077d83
===> Fetching: File::Which
Fetching with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Fetched: File::Which to /home/cpan/.zef/tmp/perl6-file-which.git
===> Extracting: File::Which
Extracting with plugin: Zef::Service::Shell::git+{<anon|83899248>}
===> Extracted: File::Which to /home/cpan/.zef/store/perl6-file-which.git/d1d9dbef35751e2383bec4ef6d3fb9aa2146eacf
===> Filtering: Linenoise:ver('0.1.0'):auth('Rob Hoelz')
===> Filtering [OK] for Linenoise:ver('0.1.0'):auth('Rob Hoelz')
===> Filtering: Native::Resources:ver('0.1.0'):auth('Rob Hoelz')
===> Filtering [OK] for Native::Resources:ver('0.1.0'):auth('Rob Hoelz')
===> Filtering: LibraryMake:ver('1.0.0'):auth('github:retupmoca')
===> Filtering [OK] for LibraryMake:ver('1.0.0'):auth('github:retupmoca')
===> Filtering: Shell::Command
===> Filtering [OK] for Shell::Command
===> Filtering: File::Find:ver('0.1')
===> Filtering [OK] for File::Find:ver('0.1')
===> Filtering: File::Which
===> Filtering [OK] for File::Which
===> # SKIP: No Build.pm for File::Which
===> # SKIP: No Build.pm for File::Find:ver('0.1')
===> # SKIP: No Build.pm for Shell::Command
===> # SKIP: No Build.pm for LibraryMake:ver('1.0.0'):auth('github:retupmoca')
===> # SKIP: No Build.pm for Native::Resources:ver('0.1.0'):auth('Rob Hoelz')
===> Building: Linenoise:ver('0.1.0'):auth('Rob Hoelz')
`build-depends` is missing entries. Attemping to mimick missing dependencies...
Command: /home/cpan/.rakudobrew/bin/../moar-nom/install/bin/perl6 -Ilib -I. -MBuild -I/home/cpan/.zef/store/p6-native-resources.git/b05fd025d5a3541e88b5acc266446d3814d79cc0/lib -I/home/cpan/.zef/store/P6-LibraryMake.git/8ab4b23a981ebfd63a74c11d1398750a472195aa/lib -I/home/cpan/.zef/store/Shell-Command.git/3d21c0ffbc467f121935a3ec74e54c23b651836e/lib -I/home/cpan/.zef/store/perl6-file-which.git/d1d9dbef35751e2383bec4ef6d3fb9aa2146eacf/lib -I/home/cpan/.zef/store/File-Find.git/d3e2be709aeb2479dc02b154fdc179fe7a077d83/lib -e ::('Build').new.build('/home/cpan/.zef/store/p6-linenoise.git/46f9534b1568e68a2a077ee80713914738b16e28'); exit(0);
gcc -c -fPIC -Wdeclaration-after-statement -Werror=declaration-after-statement -O3 -DNDEBUG -D_REENTRANT -D_FILE_OFFSET_BITS=64 -fPIC -o linenoise.o linenoise.c
perl6 -e "mkdir 'resources'; mkdir 'resources/lib'"
gcc -shared -fPIC -O3 -DNDEBUG -Wl,-rpath,"//home/cpan/.rakudobrew/moar-nom/install/lib" -Wl,-rpath,"/home/cpan/.rakudobrew/moar-nom/install/share/perl6/site/lib" -lm -lpthread -lrt -ldl -o resources/lib/liblinenoise.so linenoise.o
perl6 -e "print ''" > resources/lib/liblinenoise.dll
perl6 -e "print ''" > resources/lib/liblinenoise.dylib
gcc -o constant-helper -Wdeclaration-after-statement -Werror=declaration-after-statement -O3 -DNDEBUG -D_REENTRANT -D_FILE_OFFSET_BITS=64 -fPIC constant-helper.c
perl6 fill-constants.pl < lib/Linenoise.pm.in > lib/Linenoise.pm
===> Building [OK] for Linenoise:ver('0.1.0'):auth('Rob Hoelz')
===> Testing: File::Which
Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>}
t/01-which.t ... ok
t/02-win32.t ... ok
t/03-export.t .. ok
t/04-simple.t .. ok
t/05-all.t ..... ok
All tests successful.
Files=5, Tests=29, 9 wallclock secs ( 0.02 usr 0.00 sys + 8.29 cusr 0.44 csys = 8.75 CPU)
Result: PASS
# Found perl6 at '/home/cpan/.rakudobrew/bin/perl6'
# Found perl6 at '/home/cpan/.rakudobrew/bin/perl6' using whence
===> Testing [OK] for File::Which
===> Testing: File::Find:ver('0.1')
Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>}
t/01-file-find.t .. ok
All tests successful.
Files=1, Tests=11, 1 wallclock secs ( 0.02 usr 0.00 sys + 0.93 cusr 0.05 csys = 1.00 CPU)
Result: PASS
===> Testing [OK] for File::Find:ver('0.1')
===> Testing: Shell::Command
Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>}
t/02-shell-command.t .. ok
All tests successful.
Files=1, Tests=16, 1 wallclock secs ( 0.02 usr 0.00 sys + 1.35 cusr 0.11 csys = 1.48 CPU)
Result: PASS
===> Testing [OK] for Shell::Command
===> Testing: LibraryMake:ver('1.0.0'):auth('github:retupmoca')
Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>}
t/01-basic.t .. ok
All tests successful.
Files=1, Tests=7, 2 wallclock secs ( 0.01 usr 0.00 sys + 1.78 cusr 0.14 csys = 1.93 CPU)
Result: PASS
===> Testing [OK] for LibraryMake:ver('1.0.0'):auth('github:retupmoca')
===> Testing: Native::Resources:ver('0.1.0'):auth('Rob Hoelz')
Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>}
===> Testing [OK] for Native::Resources:ver('0.1.0'):auth('Rob Hoelz')
===> Testing: Linenoise:ver('0.1.0'):auth('Rob Hoelz')
Testing with plugin: Zef::Service::Shell::prove+{<anon|83899248>}
===> Testing [OK] for Linenoise:ver('0.1.0'):auth('Rob Hoelz')
===> Installing: File::Which
===> Install [OK] for File::Which
===> Installing: File::Find:ver('0.1')
===> Install [OK] for File::Find:ver('0.1')
===> Installing: Shell::Command
===> Install [OK] for Shell::Command
===> Installing: LibraryMake:ver('1.0.0'):auth('github:retupmoca')
===> Install [OK] for LibraryMake:ver('1.0.0'):auth('github:retupmoca')
===> Installing: Native::Resources:ver('0.1.0'):auth('Rob Hoelz')
===> Install [OK] for Native::Resources:ver('0.1.0'):auth('Rob Hoelz')
===> Installing: Linenoise:ver('0.1.0'):auth('Rob Hoelz')
===> Install [OK] for Linenoise:ver('0.1.0'):auth('Rob Hoelz')
cpan@perlbuild2:~$
| gharchive/issue | 2016-07-08T17:22:50 | 2025-04-01T06:44:27.051234 | {
"authors": [
"hoelzro",
"zoffixznet"
],
"repo": "hoelzro/p6-linenoise",
"url": "https://github.com/hoelzro/p6-linenoise/issues/15",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1678839731 | Consider handling toRow/fromRow
We'd rewrite to createSerializer and createDeserializer.
Fixed :)
| gharchive/issue | 2023-04-21T17:10:10 | 2025-04-01T06:44:27.063049 | {
"authors": [
"holdenk"
],
"repo": "holdenk/spark-upgrade",
"url": "https://github.com/holdenk/spark-upgrade/issues/59",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2655945215 | fix empty queue in last read
Fix for corner case produced in the next situation:
async _read (cb) {
// this._queue.length is 1
if (this._queue.length === 0) {
this.push(null)
return cb(null)
}
try {
while (this._queue.length > 0) {
const key = this._queue.shift()
// this.module.has(key) is true
if (this.modules.has(key)) continue
const data = await this._addOnce(key)
this.modules.set(key, data)
this._pending.delete(key)
if (this.push(data) === false) break
}
} catch (err) {
return cb(err)
}
cb(null)
}
During the stream read:
const dependencyStream = new DependencyStream(drive, { entrypoint })
for await (const dep of dependencyStream) {
// doesnt do the last read (when queue.length is 0
}
}
IMPORTANT: I have been able to repro this issue only in Bare, the same test works for Node,
2.2.4
| gharchive/pull-request | 2024-11-13T15:46:23 | 2025-04-01T06:44:27.074686 | {
"authors": [
"mafintosh",
"rafapaezbas"
],
"repo": "holepunchto/dependency-stream",
"url": "https://github.com/holepunchto/dependency-stream/pull/5",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
569387814 | Generate a build.gradle.kts file
Switch the gradle file generated by the --idea option to use Kotlin.
Closes https://github.com/holgerbrandl/kscript/issues/222
Merged at the same time as https://github.com/holgerbrandl/kscript/pull/258
| gharchive/pull-request | 2020-02-22T20:17:40 | 2025-04-01T06:44:27.077996 | {
"authors": [
"martinbonnin"
],
"repo": "holgerbrandl/kscript",
"url": "https://github.com/holgerbrandl/kscript/pull/257",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2029685064 | Map more Python types to SQLite types
Fixes https://github.com/holoviz/holonote/issues/75 but likely in an unsatisfactory way.
I combed through the pandas library trying to find the optimal solution for converting python/pandas/numpy dtypes into corresponding SQL types--each with its own trade-offs.
In the end, since I don't knowing too much of HoloNote's internals, specifically whether I can update SpecItem, or how much of HoloNote I should change (can we completely remove SQLiteDB(Connector) from connector.py and use SQLAlchemy?), I chose the most compatible path forward.
However, I list my notes below:
# 1 doesn't touch internal method, but needs to parse the CREATE TABLE string to get schema, not sure where to inject inside HoloNote
# Outputs: `CREATE TABLE "test" (\n"index" INTEGER,\n "a" INTEGER,\n "b" REAL,\n "c" TEXT,\n "d"...``
table = pd.io.sql.SQLiteTable("test", None, frame=df)
table.sql_schema()
# 2 touches internal method, and not sure where to inject inside HoloNote
# however, for sqlalchemy, table._sql_type_name is _sqlalchemy_type
table = pd.io.sql.SQLiteTable("test", None, frame=df)
print(table._get_column_names_and_types(table._sql_type_name))
# 3 still need to use mapper, and not sure where to inject inside HoloNote
for col in df.columns:
try:
print(col, type(df[col][0]), pd.api.types.pandas_dtype(type(df[col][0])))
except Exception as e:
print(col, type(df[col][0]), "FAILED")
continue
# 4 still needs to use mapper and not sure where to inject inside HoloNote
for col in df.columns:
print(pd.api.types.infer_dtype(df[col]))
# 5 include SQLAlchemy dependency
# 6 include pyarrow dependency
I was also wondering why we need to have the connector SQLiteDB.create_table define the types inside? Why not have df.to_sql() figure it out?
Also, do we need a Connector class or can we depend on SQLAlchemy's classes?
Thanks for fixing the tests!
I think the suggested change is good and will merge it.
can we completely remove SQLiteDB(Connector) from connector.py and use SQLAlchemy
I don't think we can ever completely remove the need for a Connector, but it should be as thin as possible layer to the database. I see the class the same way as pd.io.sql.SQLiteTable, not handling any of the communication but giving a simple interface for us to send and get data from a database.
We want to add an SQLAlchemy connector see https://github.com/holoviz/holonote/pull/65.
| gharchive/pull-request | 2023-12-07T01:35:52 | 2025-04-01T06:44:27.083910 | {
"authors": [
"Hoxbro",
"ahuang11"
],
"repo": "holoviz/holonote",
"url": "https://github.com/holoviz/holonote/pull/84",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1332047695 | Document paths plot
I'm trying to improve the docstrings of hvplot. I cannot find any documentation for hvPlot.paths. https://hvplot.holoviz.org/search.html?q=paths
I believe something should be available in the Reference Gallery https://hvplot.holoviz.org/reference/index.html
hvplot.paths are mentioned in https://hvplot.holoviz.org/user_guide/Geographic_Data.html but indeed this plot type doesn't have a reference gallery, and deserves one.
| gharchive/issue | 2022-08-08T15:34:47 | 2025-04-01T06:44:27.086485 | {
"authors": [
"MarcSkovMadsen",
"maximlt"
],
"repo": "holoviz/hvplot",
"url": "https://github.com/holoviz/hvplot/issues/828",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
710249047 | Cant fill in particular config value in OZW addon
The problem
I have some Neo coolcam PIR PD02Z sensors and are unable to change parameter 99 (ambient light intensity calibration) . In the OZW addon I cant fill in / change a number at the entry of parameter 99. When I try this via the ozw.set_config_parameter I get some errors in HA log and OZW addon log.
The ozw log doesn't show no errors / warnings.
The strange thing is that the initial value is 1000. When I click on that value to change it, it changes automatically to 1. Then I cant change it anymore.
I checked the xml file but i dont see something strange
Environment
Add-on with the issue: OpenZwave
Add-on release with the issue: 0.5.2
Last working add-on release (if known):
Operating environment (OS/Supervised): Supervised
Aeotec gen 5 stick + NEO coolccam PIR V2 (PD02Z)
Problem-relevant configuration
device: /dev/serial/by-id/usb-0658_0200-if00
network_key: >-
xxxx xxxx xxxx
Traceback/Error logs
HA message when using calling ozw.set_config_parameter
Kan service ozw/set_config_parameter niet aanroepen 'NoneType' object has no attribute 'type'
Message in OZW addon log
[20200928 14:57:02.654 CEST] [default] [debug]: QVariant(int, 5)
[20200928 14:57:02.656 CEST] [default] [debug]: QVariant(int, 14)
[20200928 14:57:02.656 CEST] [default] [debug]: QVariant(int, 1)
[20200928 14:57:02.657 CEST] [default] [debug]: QVariant(int, 3)
[20200928 14:57:02.658 CEST] [default] [debug]: QVariant(int, 12)
[20200928 14:57:02.659 CEST] [default] [debug]: QVariant(int, 4)
[20200928 14:57:02.660 CEST] [default] [debug]: QVariant(int, 16)
[20200928 14:57:02.661 CEST] [default] [debug]: QVariant(int, 17)
Additional information
There are various problems setting values in ozw-admin, for example: https://github.com/OpenZWave/ozw-admin/issues/33.
The problem with set_config_parameter and short values was fixed in HA 0.115. What version are you using?
Hi, I am using version 0.115.6
What is the service data you are using?
Another question, what Stage is your node in? You can find this in the UI, Configuration -> Devices. Select this device and the Device Info will list the Stage.
The stage of this node is complete and this is the service data:
node_id: 14
parameter: 99
value: 1000
Kan service ozw/set_config_parameter niet aanroepen 'NoneType' object has no attribute 'type'
I am guessing that you are running into the same problem that this PR was attempting to fix. Until that problem is fixed in the upstream library you won't be able to use Home Assistant to change the value.
There also looks to be a bug in the device XML definition for parameter 99. This error is likely causing the ozw-admin GUI to be unable to configure the value correctly.
<Value type="short" genre="config" instance="1" index="99" label="Ambient Light Intensity Calibration" value="1000" min="1" max="65536" size="2">
The type is a "short" with size "2", but the maximum value is set 65536. OZW represents "short" values as signed 16-bit integers. The maximum value is in fact 32767. A value of 65536 when casted to a short maps to 0, which is why ozw-admin won't let you change the value. This would also affect HA since it checks the min and max.
If the device really does allow a value up to 65535 (0xffff), then to allow the full range of values, you'd need to modify the XML file to:
<Value type="short" genre="config" instance="1" index="99" label="Ambient Light Intensity Calibration" value="1000" min="-32768" max="32767" size="2">
You cannot exclude the value 0. But this configuration parameter is undocumented, and there appear to be different versions of the product, so who knows what it's supposed to be.
You can bypass HA and ozw-admin and set the value directly via MQTT using the setvalue command. Use MQTT Explorer to find the value id key. Use MQTT explorer to publish the command.
topic: OpenZWave/1/command/setvalue/
payload:
{
"ValueIDKey": 123456789,
"Value": 300
}
Replace 123456789 with your key.
If the device really does allow a value up to 65535 (0xffff), then to allow the full range of values, you'd need to modify the XML file to:
<Value type="short" genre="config" instance="1" index="99" label="Ambient Light Intensity Calibration" value="1000" min="-32768" max="32767" size="2">
thanks @kpine that did the trick. Now I can change the value in the OZW gui.
Very useful program, mqtt explorer. I did not know that.
You can bypass HA and ozw-admin and set the value directly via MQTT using the setvalue command. Use MQTT Explorer to find the value id key. Use MQTT explorer to publish the command.
topic: OpenZWave/1/command/setvalue/
payload:
{
"ValueIDKey": 123456789,
"Value": 300
}
Replace 123456789 with your key.
How do I find the ValueIDKey for parameters 111-113 in MQTT explorer please?
Plus I also can not see OpenZWave//command/ ?
Background: I have the same issue as nelbs, but with two Aeotec Smart Switch 6's (ZW096). Parameters 111 - 113 are the reporting intervals for the switch (in seconds), but OpenZWave WebUI will not set them to anything but zero - which disables reporting. Running HA 118.3 and OpenZWave add-on 0.7.1
Thanks...
| gharchive/issue | 2020-09-28T13:06:58 | 2025-04-01T06:44:27.617034 | {
"authors": [
"jasonpstokes",
"kpine",
"nelbs"
],
"repo": "home-assistant/hassio-addons",
"url": "https://github.com/home-assistant/hassio-addons/issues/1600",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
305877804 | Platform xiaomi_miio not ready yet warning
Make sure you are running the latest version of Home Assistant before reporting an issue.
You should only file an issue if you found a bug. Feature and enhancement requests should go in the Feature Requests section of our community forum:
Home Assistant release (hass --version):
0.65.5
Python release (python3 --version):
3.6
Component/platform:
xiaomi_miio
Description of problem:
Xiaomi Smart WiFi Socket cannot be discovered.
2018-03-16 20:44:34 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 30 seconds.
2018-03-16 20:45:05 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 60 seconds.
2018-03-16 20:46:06 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 90 seconds.
2018-03-16 20:47:37 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 120 seconds.
2018-03-16 20:49:38 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 150 seconds.
2018-03-16 20:52:09 WARNING (MainThread) [homeassistant.components.switch] Platform xiaomi_miio not ready yet. Retrying in 180 seconds.
Expected:
Xiaomi WIFI switchs are to be discovered as usual.
Problem-relevant configuration.yaml entries and steps to reproduce:
Traceback (if applicable):
Additional info:
I have the same problem. These log entries flood log:
8-03-16 09:25:22 ERROR (Thread-6) [miio.device] Got error when receiving: timed out
2018-03-16 09:25:22 WARNING (Thread-6) [miio.device] Retrying with incremented id, retries left: 3
2018-03-16 09:25:22 ERROR (Thread-2) [miio.device] Got error when receiving: timed out
2018-03-16 09:25:22 WARNING (Thread-2) [miio.device] Retrying with incremented id, retries left: 3
2018-03-16 09:25:26 WARNING (MainThread) [homeassistant.components.light] Setup of platform xiaomi_miio is taking over 10 seconds.
2018-03-16 09:43:05 ERROR (Thread-6) [miio.device] Got error when receiving: timed out
2018-03-16 09:43:05 WARNING (Thread-6) [miio.device] Retrying with incremented id, retries left: 3
2018-03-16 10:07:52 ERROR (Thread-4) [miio.device] Got error when receiving: timed out
2018-03-16 10:07:52 WARNING (Thread-4) [miio.device] Retrying with incremented id, retries left: 3
2018-03-16 10:25:46 ERROR (MainThread) [homeassistant.core] Timer got out of sync. Resetting
Please enable debug output
logger:
default: warn
logs:
homeassistant.components.switch.xiaomi_miio: debug
miio: debug
and provide some more details. May be this thread provides a solution:
https://community.home-assistant.io/t/xiaomi-miio-switch-not-initializing/46572/11
In my case the issue is with lights not switch. This is my debug output:
Mar 16 13:46:27 hassbian hass[17048]: 2018-03-16 13:46:27 INFO (MainThread) [homeassistant.components.light.xiaomi_miio] Initializing with host philips-light-bulb_miio61806213.sozkara (token 263aa...) Mar 16 13:46:27 hassbian hass[17048]: 2018-03-16 13:46:27 DEBUG (MainThread) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:27 hassbian hass[17048]: 2018-03-16 13:46:27 DEBUG (MainThread) [miio.device] Got a response: Container: Mar 16 13:46:27 hassbian hass[17048]: data = Container: Mar 16 13:46:27 hassbian hass[17048]: length = 0 Mar 16 13:46:27 hassbian hass[17048]: value = (total 0) Mar 16 13:46:27 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:27 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:27 hassbian hass[17048]: data = (total 0) Mar 16 13:46:27 hassbian hass[17048]: header = Container: Mar 16 13:46:27 hassbian hass[17048]: length = 16 Mar 16 13:46:27 hassbian hass[17048]: value = Container: Mar 16 13:46:27 hassbian hass[17048]: length = 32 Mar 16 13:46:27 hassbian hass[17048]: unknown = 0 Mar 16 13:46:27 hassbian hass[17048]: device_id = \x03\xaf\x16\x85 (total 4) Mar 16 13:46:27 hassbian hass[17048]: ts = 1970-01-12 13:01:08 Mar 16 13:46:27 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:27 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:27 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\xaf\x16\x85\x00\x0f7\x94 (total 16) Mar 16 13:46:27 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:27 hassbian hass[17048]: 2018-03-16 13:46:27 DEBUG (MainThread) [miio.device] Discovered 03af1685 with ts: 1970-01-12 13:01:08, token: b'00000000000000000000000000000000' Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (MainThread) [miio.device] philips-light-bulb_miio61806213.sozkara:54321 >>: {'id': 1, 'method': 'miIO.info', 'params': []} Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (MainThread) [miio.device] philips-light-bulb_miio61806213.sozkara:54321 (ts: 1970-01-12 13:01:10, id: 1) << {'id': 1, 'result': {'life': 997270, 'uid': 1791780321, 'ap': {'rssi': -33, 'bssid': 'E8:DE:27:83:53:B5', 'ssid': 'sozkara 2.4ghz'}, 'hw_ver': 'ESP8266', 'wifi_fw_ver': '1.5.0-dev(7f7a714)', 'mmfree': 5264, 'token': '263aa92a412406f1fa33437abf35bffc', 'netif': {'localIp': '192.168.1.118', 'gw': '192.168.1.1', 'mask': '255.255.255.0'}, 'model': 'philips.light.bulb', 'fw_ver': '1.3.0_0033', 'mac': '34:CE:00:C5:CF:9D'}} Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 INFO (MainThread) [homeassistant.components.light.xiaomi_miio] philips.light.bulb 1.3.0_0033 ESP8266 detected Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 WARNING (MainThread) [homeassistant.setup] Setup of recorder is taking over 10 seconds. Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 INFO (MainThread) [homeassistant.components.light.xiaomi_miio] Initializing with host philips-light-bulb_miio60394599.sozkara (token abb00...) Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (Thread-2) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (MainThread) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (Thread-2) [miio.device] Got a response: Container: Mar 16 13:46:29 hassbian hass[17048]: data = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 0 Mar 16 13:46:29 hassbian hass[17048]: value = (total 0) Mar 16 13:46:29 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:29 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:29 hassbian hass[17048]: data = (total 0) Mar 16 13:46:29 hassbian hass[17048]: header = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 16 Mar 16 13:46:29 hassbian hass[17048]: value = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 32 Mar 16 13:46:29 hassbian hass[17048]: unknown = 0 Mar 16 13:46:29 hassbian hass[17048]: device_id = \x03\xaf\x16\x85 (total 4) Mar 16 13:46:29 hassbian hass[17048]: ts = 1970-01-12 13:01:10 Mar 16 13:46:29 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:29 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:29 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\xaf\x16\x85\x00\x0f7\x96 (total 16) Mar 16 13:46:29 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:29 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (MainThread) [miio.device] Got a response: Container: Mar 16 13:46:29 hassbian hass[17048]: data = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 0 Mar 16 13:46:29 hassbian hass[17048]: value = (total 0) Mar 16 13:46:29 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:29 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:29 hassbian hass[17048]: data = (total 0) Mar 16 13:46:29 hassbian hass[17048]: header = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 16 Mar 16 13:46:29 hassbian hass[17048]: value = Container: Mar 16 13:46:29 hassbian hass[17048]: length = 32 Mar 16 13:46:29 hassbian hass[17048]: unknown = 0 Mar 16 13:46:29 hassbian hass[17048]: device_id = \x03\x99\x8cg (total 4) Mar 16 13:46:29 hassbian hass[17048]: ts = 1970-01-12 13:01:14 Mar 16 13:46:29 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:29 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:29 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\x99\x8cg\x00\x0f7\x9a (total 16) Mar 16 13:46:29 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (MainThread) [miio.device] Discovered 03998c67 with ts: 1970-01-12 13:01:14, token: b'00000000000000000000000000000000' Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:29 DEBUG (Thread-2) [miio.device] Discovered 03af1685 with ts: 1970-01-12 13:01:10, token: b'00000000000000000000000000000000' Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (MainThread) [miio.device] philips-light-bulb_miio60394599.sozkara:54321 >>: {'id': 1, 'method': 'miIO.info', 'params': []} Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (Thread-2) [miio.device] philips-light-bulb_miio61806213.sozkara:54321 >>: {'id': 1, 'method': 'get_prop', 'params': ['power', 'bright', 'cct', 'snm', 'dv']} Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (MainThread) [miio.device] philips-light-bulb_miio60394599.sozkara:54321 (ts: 1970-01-12 13:01:15, id: 1) << {'id': 1, 'result': {'life': 997275, 'uid': 1791780321, 'ap': {'rssi': -46, 'bssid': 'E8:DE:27:83:53:B5', 'ssid': 'sozkara 2.4ghz'}, 'hw_ver': 'ESP8266', 'wifi_fw_ver': '1.5.0-dev(7f7a714)', 'mmfree': 5408, 'token': 'abb00a4eec3a71c8ed18283820a0b3ca', 'netif': {'localIp': '192.168.1.129', 'gw': '192.168.1.1', 'mask': '255.255.255.0'}, 'model': 'philips.light.bulb', 'fw_ver': '1.3.0_0033', 'mac': '34:CE:00:97:18:42'}} Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 INFO (MainThread) [homeassistant.components.light.xiaomi_miio] philips.light.bulb 1.3.0_0033 ESP8266 detected Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (Thread-5) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (Thread-5) [miio.device] Got a response: Container: Mar 16 13:46:30 hassbian hass[17048]: data = Container: Mar 16 13:46:30 hassbian hass[17048]: length = 0 Mar 16 13:46:30 hassbian hass[17048]: value = (total 0) Mar 16 13:46:30 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:30 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:30 hassbian hass[17048]: data = (total 0) Mar 16 13:46:30 hassbian hass[17048]: header = Container: Mar 16 13:46:30 hassbian hass[17048]: length = 16 Mar 16 13:46:30 hassbian hass[17048]: value = Container: Mar 16 13:46:30 hassbian hass[17048]: length = 32 Mar 16 13:46:30 hassbian hass[17048]: unknown = 0 Mar 16 13:46:30 hassbian hass[17048]: device_id = \x03\x99\x8cg (total 4) Mar 16 13:46:30 hassbian hass[17048]: ts = 1970-01-12 13:01:15 Mar 16 13:46:30 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:30 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:30 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\x99\x8cg\x00\x0f7\x9b (total 16) Mar 16 13:46:30 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (Thread-5) [miio.device] Discovered 03998c67 with ts: 1970-01-12 13:01:15, token: b'00000000000000000000000000000000' Mar 16 13:46:30 hassbian hass[17048]: 2018-03-16 13:46:30 DEBUG (Thread-5) [miio.device] philips-light-bulb_miio60394599.sozkara:54321 >>: {'id': 1, 'method': 'get_prop', 'params': ['power', 'bright', 'cct', 'snm', 'dv']} Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 ERROR (Thread-2) [miio.device] Got error when receiving: timed out Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 WARNING (Thread-2) [miio.device] Retrying with incremented id, retries left: 3 Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 DEBUG (Thread-2) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 DEBUG (Thread-2) [miio.device] Got a response: Container: Mar 16 13:46:35 hassbian hass[17048]: data = Container: Mar 16 13:46:35 hassbian hass[17048]: length = 0 Mar 16 13:46:35 hassbian hass[17048]: value = (total 0) Mar 16 13:46:35 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:35 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:35 hassbian hass[17048]: data = (total 0) Mar 16 13:46:35 hassbian hass[17048]: header = Container: Mar 16 13:46:35 hassbian hass[17048]: length = 16 Mar 16 13:46:35 hassbian hass[17048]: value = Container: Mar 16 13:46:35 hassbian hass[17048]: length = 32 Mar 16 13:46:35 hassbian hass[17048]: unknown = 0 Mar 16 13:46:35 hassbian hass[17048]: device_id = \x03\xaf\x16\x85 (total 4) Mar 16 13:46:35 hassbian hass[17048]: ts = 1970-01-12 13:01:16 Mar 16 13:46:35 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:35 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:35 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\xaf\x16\x85\x00\x0f7\x9c (total 16) Mar 16 13:46:35 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 DEBUG (Thread-2) [miio.device] Discovered 03af1685 with ts: 1970-01-12 13:01:16, token: b'00000000000000000000000000000000' Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 DEBUG (Thread-2) [miio.device] philips-light-bulb_miio61806213.sozkara:54321 >>: {'id': 102, 'method': 'get_prop', 'params': ['power', 'bright', 'cct', 'snm', 'dv']} Mar 16 13:46:35 hassbian hass[17048]: 2018-03-16 13:46:35 DEBUG (Thread-2) [miio.device] philips-light-bulb_miio61806213.sozkara:54321 (ts: 1970-01-12 13:01:16, id: 102) << {'id': 102, 'result': ['off', 80, 1, 0, 0]} Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 ERROR (Thread-5) [miio.device] Got error when receiving: timed out Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 WARNING (Thread-5) [miio.device] Retrying with incremented id, retries left: 3 Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 DEBUG (Thread-5) [miio.protocol] Unable to decrypt, returning raw bytes: b'' Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 DEBUG (Thread-5) [miio.device] Got a response: Container: Mar 16 13:46:36 hassbian hass[17048]: data = Container: Mar 16 13:46:36 hassbian hass[17048]: length = 0 Mar 16 13:46:36 hassbian hass[17048]: value = (total 0) Mar 16 13:46:36 hassbian hass[17048]: offset2 = 32 Mar 16 13:46:36 hassbian hass[17048]: offset1 = 32 Mar 16 13:46:36 hassbian hass[17048]: data = (total 0) Mar 16 13:46:36 hassbian hass[17048]: header = Container: Mar 16 13:46:36 hassbian hass[17048]: length = 16 Mar 16 13:46:36 hassbian hass[17048]: value = Container: Mar 16 13:46:36 hassbian hass[17048]: length = 32 Mar 16 13:46:36 hassbian hass[17048]: unknown = 0 Mar 16 13:46:36 hassbian hass[17048]: device_id = \x03\x99\x8cg (total 4) Mar 16 13:46:36 hassbian hass[17048]: ts = 1970-01-12 13:01:21 Mar 16 13:46:36 hassbian hass[17048]: offset2 = 16 Mar 16 13:46:36 hassbian hass[17048]: offset1 = 0 Mar 16 13:46:36 hassbian hass[17048]: data = !1\x00 \x00\x00\x00\x00\x03\x99\x8cg\x00\x0f7\xa1 (total 16) Mar 16 13:46:36 hassbian hass[17048]: checksum = \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00 (total 16) Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 DEBUG (Thread-5) [miio.device] Discovered 03998c67 with ts: 1970-01-12 13:01:21, token: b'00000000000000000000000000000000' Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 DEBUG (Thread-5) [miio.device] philips-light-bulb_miio60394599.sozkara:54321 >>: {'id': 102, 'method': 'get_prop', 'params': ['power', 'bright', 'cct', 'snm', 'dv']} Mar 16 13:46:36 hassbian hass[17048]: 2018-03-16 13:46:36 DEBUG (Thread-5) [miio.device] philips-light-bulb_miio60394599.sozkara:54321 (ts: 1970-01-12 13:01:21, id: 102) << {'id': 102, 'result': ['off', 80, 1, 0, 0]} Mar 16 13:46:39 hassbian hass[17048]: 2018-03-16 13:46:39 WARNING (MainThread) [homeassistant.components.light] Setup of platform xiaomi_miio is taking over 10 seconds. Mar 16 13:46:39 hassbian hass[17048]: 2018-03-16 13:46:39 DEBUG (MainThread) [homeassistant.components.light.xiaomi_miio] Got new state: <PhilipsBulbStatus power=off, brightness=80, color_temperature=1, scene=0, delay_off_countdown=0>
@serhatozkara Your device works fine. The error messages
2018-03-16 10:07:52 ERROR (Thread-4) [miio.device] Got error when receiving: timed out
2018-03-16 10:07:52 WARNING (Thread-4) [miio.device] Retrying with incremented id, retries left: 3
will be suppressed with HA 0.66 (python-miio 0.3.8):
https://github.com/rytilahti/python-miio/commit/8f4cb5d6203b36a9fc45fa617b7f9cc51c4c53ab
It seems that API token has changed somehow. I have obtained the new key and it solved the issue. Thank you so much for your help syssi.
| gharchive/issue | 2018-03-16T10:34:23 | 2025-04-01T06:44:27.658342 | {
"authors": [
"envinur",
"serhatozkara",
"syssi"
],
"repo": "home-assistant/home-assistant",
"url": "https://github.com/home-assistant/home-assistant/issues/13259",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
487004555 | Foscam FI992x motion detection problem
Home Assistant release with the issue: all
Last working Home Assistant release (if known): none
Operating environment (Hass.io/Docker/Windows/etc.): Hass.io (but all env are affected)
Component/platform: https://www.home-assistant.io/components/foscam/
Description of problem:
The enable/disable motion detection service do not work with the new foscam cameras because the HTTP API CGIProxy command getMotionDetectConfig (and setMotionDetectConfig) do not work anymore and has been replaced with the new getMotionDetectConfig1 (and setMotionDetectConfig1), please note the number 1 at the end.
Additional information:
I have inspected the code for the component (homeassistant/components/foscam/camera.py) and I have see that the python code used by the component (libpyfoscam) has two new methods to enable/disable motion detection:
enable_motion_detection1
disable_motion_detection1
So can be enough to add a configuration option to use inside the configuration.yaml of the camera and select the correct libpyfoscam method based on the camera version.
I am sorry but I do not have the time to fork, code, test and create a merge request ... I have made some test using the last version of lipyfoscam (1.1) and it is working well.
Regards to all
Hi,
I have the same problem with my FI9900P.
I can support with testing.
I solved it with a workaround until Foscam component is updated..
switch:
platform: command_line
switches:
cam1_motiondetection:
command_on: 'curl -k "http://IPADDRESS:PORT/cgi-bin/CGIProxy.fcgi?usr=CAMUSER&pwd=CAMPASSWORD&cmd=setMotionDetectConfig1&isEnable=1&linkage=136&schedule0=281474976710655&schedule1=281474976710655&schedule2=281474976710655&schedule3=281474976710655&schedule4=281474976710655&schedule5=281474976710655&schedule6=281474976710655&width1=10000&height1=10000&valid1=1"'
command_off: 'curl -k "http://IPADDRESS:PORT/cgi-bin/CGIProxy.fcgi?usr=CAMUSER&pwd=CAMPASSWORD&cmd=setMotionDetectConfig1&isEnable=0&linkage=136&schedule0=281474976710655&schedule1=281474976710655&schedule2=281474976710655&schedule3=281474976710655&schedule4=281474976710655&schedule5=281474976710655&schedule6=281474976710655&width1=10000&height1=10000&valid1=1"'
command_state: 'curl -k --silent "http://IPADDRESS:PORT/cgi-bin/CGIProxy.fcgi?usr=CAMUSER&pwd=CAMPASSWORD&cmd=getMotionDetectConfig1" | grep -oP "(?<=isEnable>).*?(?=)"'
value_template: '{{ value == "1" }}'
| gharchive/issue | 2019-08-29T14:46:55 | 2025-04-01T06:44:27.667905 | {
"authors": [
"RicoRehn",
"danilo-massa"
],
"repo": "home-assistant/home-assistant",
"url": "https://github.com/home-assistant/home-assistant/issues/26265",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
218700355 | Invalid config?
Make sure you are running the latest version of Home Assistant before reporting an issue.
You should only file an issue if you found a bug. Feature and enhancement requests should go in the Feature Requests section of our community forum:
Home Assistant release (hass --version):
Latest dev since a couple of days.
Python release (python3 --version):
Python 3.4.2
Component/platform:
hass
Description of problem:
HA doesnt start. Website remains unavailable. This is since I updated from the dev git a couple of days ago. startup gives many invalid config errors while nothing has changed on that side.
Problem-relevant configuration.yaml entries and steps to reproduce:
https://hastebin.com/erozapuyej.yaml
Traceback (if applicable):
http://i.imgur.com/Rs0crgx.png
Did you run script/setup after upgrade the dev version?
@Danielhiversen Maybe a stupid question, but how should I do that?
I've always been following these commands:
A check config gives me the following output:
https://hastebin.com/ikuqocawol.txt
Somehow fixed after a apt-get update and upgrade + reboot of the pi.
| gharchive/issue | 2017-04-01T14:45:10 | 2025-04-01T06:44:27.673170 | {
"authors": [
"Danielhiversen",
"f0nt4"
],
"repo": "home-assistant/home-assistant",
"url": "https://github.com/home-assistant/home-assistant/issues/6879",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
266231828 | Errors in verisure :(
I have HomeAssistant 0.55.2
I config verisure:
verisure:
username: !secret verisure_username
password: !secret verisure_password
And get this errors
2017-10-17 20:14:54 ERROR (MainThread) [homeassistant.setup] Error during setup of component verisure
Traceback (most recent call last):
File "/usr/src/app/homeassistant/setup.py", line 194, in _async_setup_component
component.setup, hass, processed_config)
File "/usr/local/lib/python3.6/asyncio/futures.py", line 332, in __iter__
yield self # This tells Task to wait for completion.
File "/usr/local/lib/python3.6/asyncio/tasks.py", line 250, in _wakeup
future.result()
File "/usr/local/lib/python3.6/asyncio/futures.py", line 245, in result
raise self._exception
File "/usr/local/lib/python3.6/concurrent/futures/thread.py", line 56, in run
result = self.fn(*self.args, **self.kwargs)
File "/usr/src/app/homeassistant/components/verisure.py", line 69, in setup
if not HUB.login():
File "/usr/src/app/homeassistant/components/verisure.py", line 119, in login
self.session.login()
File "/usr/local/lib/python3.6/site-packages/verisure/session.py", line 90, in login
self._vid = json.loads(response.text)['cookie']
KeyError: 'cookie'
I use latest docker image
Python release Python 3.6.3
Any idea? Im from spain
thanks!
Not compatible with verisure spain
https://github.com/persandstrom/python-verisure/issues/88
| gharchive/issue | 2017-10-17T18:31:06 | 2025-04-01T06:44:27.676104 | {
"authors": [
"Nitrosito"
],
"repo": "home-assistant/home-assistant",
"url": "https://github.com/home-assistant/home-assistant/issues/9926",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
309359684 | Fix Insteon Leak Sensor
Description:
The Insteon Leak Sensor device was not reporting dry/wet state correctly. This update bumps the underlying library to 0.8.3 which correctly reports dry/wet state and updates the binary_sensor/insteon_plm component to use the updated insteonplm module.
Related issue (if applicable): fixes #
Pull request in home-assistant.github.io with documentation (if applicable): home-assistant/home-assistant.github.io#<home-assistant.github.io PR number goes here>
Example entry for configuration.yaml (if applicable):
Checklist:
[x] The code change is tested and works locally.
[x] Local tests pass with tox. Your PR cannot be merged unless tests pass
If user exposed functionality or configuration variables are added/changed:
[ ] Documentation added/updated in home-assistant.github.io
If the code communicates with devices, web services, or third-party tools:
[x] New dependencies have been added to the REQUIREMENTS variable (example).
[x] New dependencies are only imported inside functions that use them (example).
[x] New dependencies have been added to requirements_all.txt by running script/gen_requirements_all.py.
[N/A] New files were added to .coveragerc.
If the code does not interact with devices:
[ ] Tests have been added to verify that the new code works.
HELP!!! CI tests keep failing on script/gen_requirements_all.py. Every time I run the script locally it changes my base module requirement from 0.8.3 back to 0.7.5. I have no idea why it would be doing that but I have tried everything to make it work and cannot figure it out. I cannot see where the old version is coming from. If you look at my code clearly the requirement is 0.8.3.
homeassistant/components/insteon_plm.py
Nevermind. Figured it out. The system I was working on at the time had a very old version of HA installed. The script was pulling from the installed version not the development version. Will update and repost.
| gharchive/pull-request | 2018-03-28T13:08:47 | 2025-04-01T06:44:27.684216 | {
"authors": [
"teharris1"
],
"repo": "home-assistant/home-assistant",
"url": "https://github.com/home-assistant/home-assistant/pull/13515",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
417162188 | Android TV integration
Description:
Allow for configuring Android TV devices as media players in Home Assistant. This is almost identical to the Fire TV component.
This builds off the work by @a1ex4 in https://github.com/home-assistant/home-assistant/pull/19157.
Pull request in home-assistant.io with documentation (if applicable): https://github.com/home-assistant/home-assistant.io/pull/8829
Example entry for configuration.yaml (if applicable):
# Example configuration.yaml entry
media_player:
# Use the Python ADB implementation without authentication
- platform: androidtv
name: Android TV 1
host: 192.168.0.111
# Use the Python ADB implementation with authentication
- platform: androidtv
name: Android TV 2
host: 192.168.0.222
adbkey: "/config/android/adbkey"
# Use an ADB server for sending ADB commands
- platform: androidtv
name: Android TV 3
host: 192.168.0.123
adb_server_ip: 127.0.0.1
Checklist:
[ ] The code change is tested and works locally.
[ ] Local tests pass with tox. Your PR cannot be merged unless tests pass
[x] There is no commented out code in this PR.
If user exposed functionality or configuration variables are added/changed:
[x] Documentation added/updated in home-assistant.io
If the code communicates with devices, web services, or third-party tools:
[x] New dependencies have been added to the REQUIREMENTS variable (example).
[x] New dependencies are only imported inside functions that use them (example).
[x] New or updated dependencies have been added to requirements_all.txt by running script/gen_requirements_all.py.
[x] New files were added to .coveragerc.
If the code does not interact with devices:
[ ] Tests have been added to verify that the new code works.
I tested this with my Fire TV stick and it works (although the Fire TV component is more tailored for Fire TV devices, so this was just for testing purposes).
Just tested this with my Shield and it works well :+1:
Confirmed working on my Shield as well.
OK, "WIP" removed from the title!
It may be helpful to the reviewer(s) to look at the commit diffs. This really is almost identical to the Fire TV component, with "fire" replaced by "android" and some small changes.
Change 'firetv' to 'androidtv' and 'ftv' to 'atv'
More Fire TV -> Android TV substitutions
Modify Fire TV component for Android TV
I have been following your project on your github page and always keeping tabs on the your latest commits. No real issues along with way other than when MiBox drops network connection with its deep sleep issue / bug (not your component's fault). I have 4 MiBox that I use this with.
Going to bump up to this latest commit.
Is there a way to add to the known apps it displays?
@glassbase This is not a support thread. Please discuss support in the forums.
@arsaboo I think the known apps is hard coded in the component, so I would say it leads toward the development of this component to ask how.
@glassbase check the linked docs in the PR
@dshokouhi Thanks
It would be great if we could get this merged in before tomorrow's release! @MartinHjelmare, since you reviewed some of my Fire TV pull requests, any chance you'd be willing to review this?
This is a new feature so it won't make this week's release anyhow.
Would you mind pointing out the major differences between this and the firetv (considering you wrote they it is basically identical to that) in the PR summary? Would it be feasible to merge these into one, as I don't think the maintenance burden of keeping two copies in sync (bug fixes in the future, keeping up the feature parity in case of new features are introduced, ..) is to be underestimated.
Would you mind pointing out the major differences between this and the firetv (considering you wrote they it is basically identical to that) in the PR summary? Would it be feasible to merge these into one, as I don't think the maintenance burden of keeping two copies in sync (bug fixes in the future, keeping up the feature parity in case of new features are introduced, ..) is to be underestimated.
From memory:
The Android TV component has a unique_id property and supports volume control.
The Fire TV component lists the running apps as sources, and you can use media_player.select_source to start and stop apps.
The update methods are essentially the same, except that different attributes get assigned to.
They could be combined. I think the way to do it would be to create an ADBMediaPlayer class from which the AndroidTVDevice and FireTVDevice classes both inherit.
Okay, can the device type be autodetected somehow? I think it would be for the best to simply combine it into one and use a flag/model type variable per device type to tell homeassistant which features are supported and to adjust the queries where necessary. Or are there some pros to keep them separate?
I don't know how to auto-detect the device type. But it would be easy to have a configuration entry where the user specifies either firetv or androidtv. It's probably for the best to combine them. What would the integration be called?
androidmediaplayer? Both are Android devices after all
Here's an initial attempt to combine the platforms. The build wouldn't pass -- the requirements_all.txt and .coverage files need to be updated and the androidtv‘ and firetv` packages need some minor modifications. But let me know what you think.
https://github.com/JeffLIrion/home-assistant/blob/androidtv-firetv-combined/homeassistant/components/androidtv/media_player.py
I don't know how to auto-detect the device type.
If you can use adb at this stage, adb shell getprop ro.product.model could do the trick.
I don't know how to auto-detect the device type.
If you can use adb at this stage, adb shell getprop ro.product.model could do the trick.
Thanks! I took a look at the output of adb shell getprop. I think adb shell getprop ro.product.manufacturer might be even easier -- if the output is "Amazon" it's a Fire TV device, otherwise it's an Android TV device.
Of course, the question is whether a particular approach is correct 100% of the time. We could always allow for an optional parameter so the user can specify the device type.
I've tested this version of the component and the same ADB debug messages are dumped to the HA console as in @a1ex4's version: https://github.com/home-assistant/home-assistant/pull/19157#issuecomment-469067661. Can this please be fixed in the dependencies before release?
A device type parameter that defaults to automatic detection sounds great :)
@jjlawren I am using this component myself and I do not get those errors in my logs, console is not the same as logs. In my console every single platform is sending data to the console, but that does not show up in the logs which is what logger is for.
Also, let us not use this thread for support/feature requests to minimize the noise.
This is not a support/feature request. No component should pass through debug messages to the console or logger. With logger set to default: warn your logs and console should be relatively silent.
My attempt to fix the console debug logging: https://github.com/Swind/pure-python-adb/pull/14
@JeffLIrion here's what I propose after having a chat with @arsaboo:
Convert this PR to be a modification of firetv component by adding the changes necessary to support non-firetv devices.
Update firetv documentation accordingly.
This way we can keep the existing git history. The follow-up steps (I'd prefer to have it in a separate PR just for the sake of git diffs and logs) would be:
Rename firetv to androidtv if that's the final naming of the component
Rename firetv in documentation, add forward from firetv to androidtv.
This way there will be no confusion of having multiple components supporting similar devices, and renaming will be just a single breaking change.
@JeffLIrion here's what I propose after having a chat with @arsaboo:
Convert this PR to be a modification of firetv component by adding the changes necessary to support non-firetv devices.
Update firetv documentation accordingly.
This way we can keep the existing git history. The follow-up steps (I'd prefer to have it in a separate PR just for the sake of git diffs and logs) would be:
Rename firetv to androidtv if that's the final naming of the component.
Rename firetv in documentation, add forward from firetv to androidtv.
This way there will be no confusion of having multiple components supporting similar devices, and renaming will be just a single breaking change.
I thought about it myself and came to the same conclusion. The new updated media_player.py will look very similar to my initial attempt at a combined component that I posted above. To allow for automatically determining the device type, I'm going to add the firetv package functionality to the androidtv package. I'll define a setup function in that package that will return either an AndroidTV instance or a FireTV instance. For the HA component, I'll add a 'device_class' (CONF_DEVICE_CLASS) config entry that can be auto (default), androidtv, or firetv.
Question: can I use type(aftv).__name__ to determine which type of object I'm working with? It seems unnecessary to do from androidtv.androidtv import AndroidTV just so I can do isinstance(aftv, AndroidTV).
For the HA component, I'll add a 'device_class' (CONF_DEVICE_CLASS) config entry that can be auto (default), androidtv, or firetv.
That sounds good.
Question: can I use type(aftv).__name__ to determine which type of object I'm working with? It seems unnecessary to do from androidtv.androidtv import AndroidTV just so I can do isinstance(aftv, AndroidTV).
Well, you could, but I recommend against that as I have experienced how painful it is to create tests when such structures are used (as mock objects are then not types, but objects). The better way is to inherit both types of classes from a same base class (so that they shared functionality can be put in it) which implements some sort of device_type() function to return an (enum) type.
Well, you could, but I recommend against using isinstance or that as I have experienced how painful it is to create tests when such structures are used (as mock objects of that class do not match, as they are mock objects instead of types).
The better way is to inherit both types of classes from a same base class (so that they shared functionality can be put in it) which implements some sort of device_type() function to return an (enum) type.
I think you misunderstood the question. Or I misunderstood your answer!
In the androidtv package, I've got a base class BaseTV from which AndroidTV and FireTV inherit. I have a function called setup that returns either an AndroidTV instance or a FireTV instance. My question is, in Home Assistant, can I use type(aftv).__name__ to determine whether I should create an AndroidTVDevice object or a FireTVDevice object? It doesn't seem worthwhile to do an extra import in HA or to define an attribute/method/property in the androidtv package when this simple command will suffice.
Here's the current setup function that I've got: https://github.com/JeffLIrion/python-androidtv/blob/firetv/androidtv/init.py
Is that setup() going to be a part of homeassistant's code base? If not, shouldn't it be?
But anyway, you will not want to use type() or any variants of isinstance() checks in your code if that's somehow possible, as it will make a mess when writing tests on that functionality later on.
I'm closing this in favor of https://github.com/home-assistant/home-assistant/pull/21872.
| gharchive/pull-request | 2019-03-05T07:51:10 | 2025-04-01T06:44:27.719639 | {
"authors": [
"JeffLIrion",
"MartinHjelmare",
"arsaboo",
"dshokouhi",
"glassbase",
"jjlawren",
"outadoc",
"rytilahti"
],
"repo": "home-assistant/home-assistant",
"url": "https://github.com/home-assistant/home-assistant/pull/21674",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
187459359 | Refactory nest component/platforms
Description:
#4210
I have no nest device, please testing.
Tested with a thermostat, looks good :dolphin:
Latest version 0.32.1 causes Nest endpoint to respond with http status 429 (too many requests), see:
https://github.com/home-assistant/home-assistant/issues/4241
Receiving this error on 34.2:
16-12-04 23:45:02 homeassistant.helpers.entity: Updating state for sensor.hallway_thermostat_temperature took 2.006 seconds. Please report platform to the developers at https://goo.gl/Nvioub
0.34.4
16-12-06 23:27:02 homeassistant.helpers.entity: Updating state for sensor.main_floor_thermostat_temperature took 1.488 seconds. Please report platform to the developers at https://goo.gl/Nvioub
@bagobones Report it at the given link instead.
| gharchive/pull-request | 2016-11-04T23:27:33 | 2025-04-01T06:44:27.724362 | {
"authors": [
"Landrash",
"RealGandy",
"bagobones",
"captainnapalm",
"pvizeli",
"w1ll1am23"
],
"repo": "home-assistant/home-assistant",
"url": "https://github.com/home-assistant/home-assistant/pull/4219",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
225270631 | light.sensehat: plugin to control the 8x8 LED matrix on a Sense hat
Description:
Plugin for Raspberry Pi Sense Hat which exposes the onboard 8x8 RGB LED matrix as a light with brightness and RGB color control.
Pull request in home-assistant.github.io with documentation (if applicable): home-assistant/home-assistant.github.io#2530
Example entry for configuration.yaml (if applicable):
light:
- platform: sensehat
Checklist:
If user exposed functionality or configuration variables are added/changed:
[x] Documentation added/updated in home-assistant.github.io
If the code communicates with devices, web services, or third-party tools:
[x] Local tests with tox run successfully. Your PR cannot be merged unless tests pass
[x] New dependencies have been added to the REQUIREMENTS variable (example).
[x] New dependencies are only imported inside functions that use them (example).
[x] New dependencies have been added to requirements_all.txt by running script/gen_requirements_all.py.
[x] New files were added to .coveragerc.
Hi @imrehg,
It seems you haven't yet signed a CLA. Please do so here.
Once you do that we will be able to review and accept this pull request.
Thanks!
Awesome! 🐬
Realized after merging that you want to add the assumed state property:
@property
def assumed_state(self) -> bool:
"""Return True if unable to access real state of the entity."""
return True
Added that in cf664e42cca80546f1f85963f58c96942ac62138
@balloob thanks for the comments! I was working from the piglow example, but the hardware is not the same, so was good to know more.
Have a question thinking about your addition: the "brightness" is just emulated, so if properly done, should this platform really just be "RGB" only? Or in general this pretending of having brightness control is useful for actual use? (that was my experience when testing it out)
If it was RGB only and no pretence was done, though, we could get the the real state of the entity (the displayed color) with get_pixels.
What do you think?
Nah, I think the addition of brightness still makes sense as it's actually showing the color in a specific brightness.
Getting the real color would be cool but I guess you would have to scan all pixels and take an average? That sounds quite expensive. Like this is probably fine.
@balloob sounds good, then this is great. :) The color query would be just a single function call btw, as all the pixels set together, so getting one's value would get the value of the whole light. But it's indeed not necessary, just wanted to ask, cheers!
| gharchive/pull-request | 2017-04-29T17:03:23 | 2025-04-01T06:44:27.734150 | {
"authors": [
"balloob",
"homeassistant",
"imrehg"
],
"repo": "home-assistant/home-assistant",
"url": "https://github.com/home-assistant/home-assistant/pull/7365",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
267307232 | Wink external action tracking
Description:
This adds a new attribute to all non-sensor Wink devices called "external_change" this attribute is calculated by checking the last update time of the device (set by HA) and if the state change was was greater than 15 seconds ago, and the state change came from PubNub the attribute will be set to true.
So what this means is, if the user changes a device via the official Wink app, or turns on a switch by the physical switch, this attribute will be set to true. Any time a state is changed via HA the attribute is set back to false. This could be useful in making complex automations around how a device was changed. If my front door was manually unlocked don't autolock it, but if it was unlocked via an automation automatically lock it again"
My use case is, I have a light that gets turned on automatically, and it gets turned back off automatically via an HA automation. If I turn that light on manually I don't want the delayed script to get executed and make the light automatically turn off again. So I can check this attribute in the automation and if it isn't false don't execute.
This doesn't work for all Wink devices. It doesn't seem to work for powerstrips/outlets because to wink they are all one device. It also doesn't seem to work on a generic zwave switch I have that doesn't report it's state quick enough. Everything else seems to work as expected...
This also fixes some Wink attributes that weren't getting pulled from the base Wink device. (Climate and Alarms)
Pull request in home-assistant.github.io with documentation (if applicable): home-assistant/home-assistant.github.io#<home-assistant.github.io PR number goes here>
Checklist:
If user exposed functionality or configuration variables are added/changed:
[ ] Documentation added/updated in home-assistant.github.io
If the code communicates with devices, web services, or third-party tools:
[ ] Local tests with tox run successfully. Your PR cannot be merged unless tests pass
I'm not sure if we should do that. First Home-Assistant are only interfaces that are implemented by class and the only logic inside hass is for mapping device stuff to home-assistant stuff. With this code you add logic into home-assistant and have logic inside library.
This doesn't work for all Wink devices.
That sounds wired.
@balloob ?
@pvizeli Seems reasonable, I could add this logic into python-wink I think?
The "doesn't work for all Wink devices" is because of legacy Wink devices they they still support. These devices, like power strips, are sent back via the Wink API inside of one JSON object, but are broke out into 3 objects. The problem with this, is that the updates are sent in response and then sent out to HA to all three devices. There is no way to determine where the update came from/which device it was meant for. This means changing one will update all three.
Seems like a great feature to me. I can see the a few use cases for myself.
I apologize if I’m overstepping, but can this be done globally against the core state changes?
If from component
Do this
If from Hass
Do this
Such an attribute cannot be trusted. If we send a command to Wink but someone will also make a change via physical button/app, we will still attribute that change to Home Assistant. Why 15 seconds? Why not 5? 20? None of these values will make real sense. It's just guess work.
And since it's guess work, we should not present it in Home Assistant as a fact.
If you want to make this use case work, try using the input_datetime component to keep track of when you make calls. We should not add this to the Wink component.
@balloob thanks for the feedback. I agree with your assessment. I think there may still be an opportunity here but maybe via python-wink. I'll have to take a look, but I think it might be possible to remove the "guess work" out of this if it is performed at a lower level.
| gharchive/pull-request | 2017-10-20T21:15:41 | 2025-04-01T06:44:27.743704 | {
"authors": [
"balloob",
"darthsebulba04",
"pvizeli",
"rbflurry",
"w1ll1am23"
],
"repo": "home-assistant/home-assistant",
"url": "https://github.com/home-assistant/home-assistant/pull/9995",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1638143955 | [NL] Round 3 binary_sensor any/all/which/how many
presence
power
Closed for now, will take this on again if I find some time.
| gharchive/pull-request | 2023-03-23T19:10:04 | 2025-04-01T06:44:27.745218 | {
"authors": [
"TheFes"
],
"repo": "home-assistant/intents",
"url": "https://github.com/home-assistant/intents/pull/1258",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
853541979 | Error when launching the camera view
Hi, I do have one Simplicam in my setup. The preview works and updates correctly, but as soon as I tap the preview to launch the live feed, nothing happens and when I check the logs I can see the following error
Anyone knows what this is and how to solve this?
Everything else in the plugin works great and it's generally amazing!
Hi @dallombra . As the bot suggested, some basic info is missing here like your OS, versions etc. but that error spawn unknown system error -8 sounds like something about your setup.
| gharchive/issue | 2021-04-08T14:36:34 | 2025-04-01T06:44:27.793518 | {
"authors": [
"dallombra",
"shamoon"
],
"repo": "homebridge-simplisafe3/homebridge-simplisafe3",
"url": "https://github.com/homebridge-simplisafe3/homebridge-simplisafe3/issues/209",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2528841256 | 🛑 HomeRun CZ ~ Auth is down
In a1a7b15, HomeRun CZ ~ Auth (https://id.homerun.co.cz/health_check) was down:
HTTP code: 0
Response time: 0 ms
Resolved: HomeRun CZ ~ Auth is back up in 811aa00 after 7 minutes.
| gharchive/issue | 2024-09-16T15:50:08 | 2025-04-01T06:44:27.806961 | {
"authors": [
"sibest19"
],
"repo": "homeruntech/upptime",
"url": "https://github.com/homeruntech/upptime/issues/48",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1433529876 | How to create a Github Action and download large binaries
Plugin Version or Commit ID
v0.10.1
Unity Version
2021.3.0f1
Your Host OS
All
Target Platform
iOS
Description
The iOS binary for MediaPipe is over 100 MB and can thus not be uploaded to GitHub.
I'm trying to create a GitHub action that will either Generate the binaries with the release or Download the binaries from another location
Without access to the whole repo (Build scripts and source files) when using only the .unitypackage and .gitignore's to prevent the large binary files from being submitted to GitHub, How could I configure my project to generate these binary files?
Is there a guide for how to accomplish this or is it currently not possible yet?
Code to Reproduce the issue
No response
Additional Context
No response
Sorry, I'm afraid I don't get what you mean precisely.
Can you describe your issue, distinguishing between what it is and what you would like it to be?
At least, once you fork the repository, I think you can run workflows and build a package with your favorite options.
If I want to use this plugin in a unity project. How do I create a GitHub action that contains the mediapipe libraries.
The iOS library is over 100 mb so it can’t be uploaded to GitHub.
One solution I’ve thought about is downloading your release and unpacking it in the Packages during a git action.
My question is. Do you have any official solution to this?
At least, I think you can build libraries by copying published GitHub workflows.
The iOS library is over 100 mb so it can’t be uploaded to GitHub.
You can reduce the size by specifying --no-apple_bitcode and --linkopt="-s" (cf. https://github.com/homuler/MediaPipeUnityPlugin/issues/724#issuecomment-1228353665).
If you want to keep the embedded bitcode, you may also want to look into Git LFS (cf. https://docs.github.com/en/repositories/working-with-files/managing-large-files/about-large-files-on-github#file-size-limits).
One solution I’ve thought about is downloading your release and unpacking it in the Packages during a git action.
Sorry, I don't know what is the problem. Why don't you do so?
I solved it! Thank you.
The commands to make it smaller also helped!
By creating a fork of the project and then making a custom release I'm now downloading that release in my github actions :+1:
:D
| gharchive/issue | 2022-11-02T17:31:55 | 2025-04-01T06:44:27.821598 | {
"authors": [
"Kariaro",
"homuler"
],
"repo": "homuler/MediaPipeUnityPlugin",
"url": "https://github.com/homuler/MediaPipeUnityPlugin/issues/779",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2185274206 | Update gui.py 解决页面启动时候 检查更新异常 闪退问题
解决页面启动时候 检查更新异常 闪退问题
| gharchive/pull-request | 2024-03-14T02:52:14 | 2025-04-01T06:44:27.834112 | {
"authors": [
"arkizat"
],
"repo": "honmashironeko/sqlmap-gui",
"url": "https://github.com/honmashironeko/sqlmap-gui/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1830683743 | Return empty 200 Response as default
Like itty-router, just return an empty 200 Response as default if there's no response specified.
http methods include GET/POST/PUT/PATCH/DELETE. Most of the time only the GET method return a details response, other methods return an empty 200 Response.
So that, by implementing this feature, we can remove redundant return c.body() from the routes.
Hi @chientrm,
I don't think it's a good idea to return a 200 Response as the default. There is possibility that users don't desire it to be 200, but 404 or throwing the error instead. And, basically, we designed it so that the "handler should return a Response object."
@usualoma @metrue @ThatOneBro or others:
What do you think about it?
Hi @chientrm @yusukebe
If a handler returns undefined, it means "the handler did not process anything" and not "success, but the response is empty." This is a necessary specification for the routing fallback mechanism.
Also, in general, it is preferable to return a meaningful response from even other than GET.
Thus, I think no need to return an empty 200 response.
I agree with @usualoma, it can lead to unexpected behavior and false positives that a route is working. Probably better not to default to a 200 response.
We could close this.
| gharchive/issue | 2023-08-01T07:58:30 | 2025-04-01T06:44:27.840821 | {
"authors": [
"ThatOneBro",
"chientrm",
"usualoma",
"yusukebe"
],
"repo": "honojs/hono",
"url": "https://github.com/honojs/hono/issues/1280",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2070255596 | fix: streamSSE's doesn't trigger onAbort correctly
Author should do the followings, if applicable
[x] Add tests
[x] Run tests
[x] yarn denoify to generate files for Deno
#1902
Hi, @watany-dev. cc: @yusukebe
I have changed streamSSE from the traditional implementation method to something similar to the original stream helper one, can you review it once as I am not familiar with this implementation?
@sor4chi
Thank you. I was also able to confirm that onAbort stops normally when I actually run it. I think this is the way to go.
@watany-dev Thanks for reviewing!
@yusukebe This PR is review ready, please take a look this!
Hi @sor4chi
Looks good! I've left one comment. Check it please!
fixed!
@sor4chi @watany-dev
Thanks! Merging now.
| gharchive/pull-request | 2024-01-08T11:27:19 | 2025-04-01T06:44:27.844594 | {
"authors": [
"sor4chi",
"watany-dev",
"yusukebe"
],
"repo": "honojs/hono",
"url": "https://github.com/honojs/hono/pull/1920",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
970053265 | Add moonfly theme
Hello,
Author of moonfly and nightfly speaking.
I notice that the nightly theme is already supported (thanks).
I thought it would be nice to also add lualine support for my other theme (moonfly).
Cheers.
Great . But since you're the author of the colorscheme I think it'll be best if you put the theme in the colorscheme repo .That way you can tweek it later easily if you change some colors in the colorscheme without coming here for change.
Since you're already providing lightline & airline themes it'll be pretty much the same .
You'll just have to put the same moonfly.lua file in the same folder /lua/lualine/themes/ in your repo and lualine will pick it up :)
Sure, I can do that quite easily.
But, I believe that should also happen with the nightfly theme then as well, adding nightfly.lua to my nightfly repo /lua/lualine/themes directory. Will that be a problem with the nightly.lua that exists here?
But, I believe that should also happen with the nightfly theme then as well, adding nightfly.lua to my nightfly repo /lua/lualine/themes directory. Will that be a problem with the nightly.lua that exists here?
You can put that in your repo too. Though same name in two places may cause conflict . If you want I can remove the nightfly theme from from shadmansaleh/lualine.nvim That's the one I'm currently maintaining . I'll probably add something lualine loads theme from external plugin instead in from itself when there's a conflict .
Cool, I will add both themes to my two repos.
The nightfly.lua I add will be exactly the same as yours.
Cheers.
Is your lualine the one I should reference in my repositories or should it be this one?
Is this project abandoned whilst yours is mantained?
For now I can't contact @hoob3rt. I haven't hard forked it yet . The changes and fixes that I've made is currently a pr to this repo #311 . I'll give @hoob3rt some more time . If he doesn't contact back I'll just hard fork this. So for now you can reference the one you like .
moonfly & nightfly no support lualine.
I will keep an eye on #311 to see if you hard fork, if you do I will change my README files to reference your version (if that becomes the maintained version).
Thanks.
@bluz71 after https://github.com/shadmansaleh/lualine.nvim/commit/32727039a97ece7ee4eb6d889be0b3e50b570daa
nightfly theme from your repo will be loaded when the colorschme is installed instead of the one bundled with lualine :)
Thanks, that is good to know.
It is a nice feeling having both my themes now supporting lualine along with airline/lightline.
| gharchive/pull-request | 2021-08-13T04:31:54 | 2025-04-01T06:44:27.856279 | {
"authors": [
"bluz71",
"shadmansaleh"
],
"repo": "hoob3rt/lualine.nvim",
"url": "https://github.com/hoob3rt/lualine.nvim/pull/318",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
205316780 | 请问我是在application目录下还是在根目录下 compose ?
能否提供几个案例,协助一下。谢谢
根目录下composer
application\extra目录下需要wechat.php文件吗
| gharchive/issue | 2017-02-04T02:39:04 | 2025-04-01T06:44:27.871558 | {
"authors": [
"banchlor",
"hooklife",
"rcj654020445"
],
"repo": "hooklife/thinkphp5-wechat",
"url": "https://github.com/hooklife/thinkphp5-wechat/issues/3",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
371981906 | [HOPS-683], [HOPS-684], [HOPS-686], [HOPS-687], [HOPS-689], [HOPS-698], [HOPS-699], [HOPS-700], [HOPS-701], [HOPS-702], [HOPS-703], [HOPS-704], [HOPS-705], [HOPS-706], [HOPS-707], [HOPS-708], [HOPS-709], [HOPS-710], [HOPS-712], [HOPS-713]
Make sure there is no duplicate PR for this issue
Please check if the PR fulfills these requirements
[ ] Tests for the changes have been added and passed (for bug fixes / features)
[x] HOPS JIRA issue has been opened for this PR
[x] All commits have been squashed down to a single commit
[x] The commit message has the following format: [HOPS-XXX] message
Post a link to the associated JIRA issue
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-683
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-684
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-686
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-687
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-689
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-698
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-699
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-700
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-701
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-702
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-703
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-704
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-705
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-706
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-707
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-708
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-709
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-710
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-712
https://hopshadoop.atlassian.net/projects/HOPS/issues/HOPS-713
What kind of change does this PR introduce? (Bug fix, feature, docs update, ...)
What is the new behavior (if this is a feature change)?
Does this PR introduce a breaking change? (What changes might users need to make in their application due to this PR?)
Other information:
Jenkins test this please.
Jenkins test this please.
Jenkins test this please.
Jenkins test this please.
Jenkins test this please.
Jenkins test this please.
Jenkins test this please.
Jenkins test this please.
| gharchive/pull-request | 2018-10-19T14:28:18 | 2025-04-01T06:44:27.898334 | {
"authors": [
"berthoug"
],
"repo": "hopshadoop/hops",
"url": "https://github.com/hopshadoop/hops/pull/554",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
434706836 | Handle Dust Transactions
Look into how the dust transactions should be handled.
That is to prevent the wallet user from leaking privacy by not using those UTXOs. May apply same logic to OP_return types
Detailed task list will be provided below upon further research
OP_return types
OP_return types are not spendable , so no privacy leak
| gharchive/issue | 2019-04-18T10:46:58 | 2025-04-01T06:44:27.900095 | {
"authors": [
"esengulov"
],
"repo": "horizontalsystems/bank-wallet-android",
"url": "https://github.com/horizontalsystems/bank-wallet-android/issues/828",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1516120557 | Herzlichen Glückwunsch
im Ranking der Top German Github Aktivitäten nun auf Platz 1
LINK
Wow. :) Thank you for the info!
Kein problem, sehr sehr gut gemacht!
BTW you have a very good Github Profile. :) Excellent work!
@horsicq
Vielen Dank
Was tun Sie hier? -> https://github.com/Community-s-Lab
Ich werde über diese Organisation Anwendungen entwickeln die speziell für Linux geeignet sind und über python Programmieren.
Dazu suche ich nun fähige Entwickler die dieses Projekt mit ihrem Wissen unterstützen und diesem beitreten.
Es sieht gut aus. Ich könnte mich später anschließen.
Würde mich sehr freuen,,, Danke
| gharchive/issue | 2023-01-02T07:58:06 | 2025-04-01T06:44:27.904094 | {
"authors": [
"bylickilabs",
"horsicq"
],
"repo": "horsicq/Translate_all",
"url": "https://github.com/horsicq/Translate_all/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1314781478 | CB-17837 - Roll forward sync fails because some 'p' ending cloudera-m…
…anager-server versions are named differently in repository and image catalog
This commit:
introduces a package pattern that removes unwanted characters from
end of pattern
See detailed description in the commit message.
is this really required to merge into 2.60?
is this really required to merge into 2.60?
Yes, we intend to do a hotfix. It is causing upgrade failures once retry is used. We've seen several failures in prod already.
| gharchive/pull-request | 2022-07-22T09:43:24 | 2025-04-01T06:44:27.907122 | {
"authors": [
"keyki",
"lturcsanyi",
"pdarvasi"
],
"repo": "hortonworks/cloudbreak",
"url": "https://github.com/hortonworks/cloudbreak/pull/13105",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
142694229 | Remove grantResults from onActivityResult
see #120
LGTM
I'll check soon.
| gharchive/pull-request | 2016-03-22T15:55:11 | 2025-04-01T06:44:27.923242 | {
"authors": [
"aurae",
"hotchemi",
"shiraji"
],
"repo": "hotchemi/PermissionsDispatcher",
"url": "https://github.com/hotchemi/PermissionsDispatcher/pull/121",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
807713334 | Add request interceptor
The problem
The modern server-side embedded Shopify apps are built with JWT session tokens which are obtained on the client-side and then passed to every server request. Turbo works pretty well with it since it allows to navigate application without doing page reloading on every click or page submission. The biggest issue is passing JWT token to every Turbo request.
The solution
This PR introduces request interceptor support. It allows to insert async function between every fetch request. Inside this function you can do some preparation for request and add custom headers.
Example usage
Turbo.setRequestInterceptor(async (request) => {
const token = await getSessionToken(window.app);
request.addHeader("Authorization", `Bearer ${token}`);
});
Could this same seam for modifying headers be achieved with a turbo:before-fetch-request event listener?
addEventListener("turbo:before-fetch-request", ({ detail: { fetchOptions }) => fetchOptions.headers["..."] = "..."))
Changing headers with turbo:before-fetch-request event listener will work fine for the simpler case when you need to add some static header to Turbo requests. Unfortunately, with Shopify embedded apps you need to retrieve the JWT token before every request because the token lifetime is short and we don't know exactly when it's expired inside the Turbo app. JWT token retrieval is async and we need to pause Turbo's fetch request invocation until it's done. So async interceptor works best for this complex case. I used Axios interceptors as an inspiration for this approach.
@kirillplatonov Nice work! Great that we won't need to fetch JWT every x seconds with your setRequestInterceptor solution and rather ask for them just when we need them. 👍
Are there additional pros using Turbo.setRequestInterceptor VS "turbo:before-fetch-request? Is there any downside in pausing Turbo's fetching request invocation until it's done?
The alternative as I see it is using turbo:before-fetch-request for Shopify Apps and a keepRetrievingToken function or similar which I assume is what @seanpdoyle was referring to above? You solution for sure looks better than the code below but would love to hear more from you regarding this and why you think it's not a good option to use keepRetrievingToken. Is another reason that the JWT might be expired when the call is made even though it was fetched just a few seconds ago?
document.addEventListener("turbo:before-fetch-request", function (event) {
var headers = event.detail.fetchOptions.headers;
headers["Authorization"] = "Bearer " + window.sessionToken;
});
We can use keepRetrievingToken or similar to fetch new tokens every x seconds:
async function retrieveToken(app) {
var SessionToken = window["app-bridge"].actions.SessionToken
app.dispatch(
SessionToken.request(),
);
// Save a session token for future requests
window.sessionToken = await new Promise((resolve) => {
app.subscribe(SessionToken.ActionType.RESPOND, (data) => {
resolve(data.sessionToken || "");
});
});
}
function keepRetrievingToken(app) {
setInterval(() => {
retrieveToken(app);
}, SESSION_TOKEN_REFRESH_INTERVAL);
}
@kirillplatonov Nice work on this PR! Great that we won't need to fetch JWT every x seconds with your setRequestInterceptor solution and rather ask for them just when we need them. 👍
Until it's merged I published my fork to npm:
yarn add turbo-edge
import * as Turbo from "turbo-edge"
yarn add turbo-rails-edge
import { Turbo } from "turbo-rails-edge"
Great stuff @kirillplatonov. Could you add some documentation for this on https://turbo.hotwire.dev/reference/drive? It's in the turbo-site repo. Thanks!
Thanks for getting this through @kirillplatonov @dhh - useful for JWT / Shopify apps!
@kirillplatonov Upon further reflection, I think it might be worth exploring an approach similar to https://github.com/hotwired/turbo/pull/28#issuecomment-863986177. Which basically let's the turbo:before-fetch-request pause the request. Then we could do it all in a callback instead.
@dhh having the same pausable API for both turbo:before-render and turbo:before-fetch-request would be great. I will explore this.
| gharchive/pull-request | 2021-02-13T09:24:12 | 2025-04-01T06:44:27.949166 | {
"authors": [
"awd",
"dhh",
"kirillplatonov",
"lonroth",
"seanpdoyle"
],
"repo": "hotwired/turbo",
"url": "https://github.com/hotwired/turbo/pull/177",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1352901346 | Scaling the Spectrum
Hello! While I was plotting the SEDs, I realized that the spectrum was scaling differently with and without photometry points. I was wondering about the mechanism that was used to scale the spectrum.
Also, the following image is a comparison between the plot that I got and the plot from Brewster. The orange line (SED that I created) should line up with the green line, but I have no idea why it doesn't align with it in the mir spectra.
I've tried adding and dropping WISE photometry points, but the SED still doesn't match up.
Is there a reason why the mir spectra is being plotted in this way?
Thanks.
Hi @graceparkk ! I'll take a look at this next week but I'm not too sure as of yet what the issue is.
| gharchive/issue | 2022-08-27T01:40:59 | 2025-04-01T06:44:27.968225 | {
"authors": [
"graceparkk",
"hover2pi"
],
"repo": "hover2pi/sedkit",
"url": "https://github.com/hover2pi/sedkit/issues/79",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
282850566 | startConversation not waiting for response
Hey guys
I'm using the basic convo example included in the starter repo
controller.hears(['color'], 'message_received', function (bot, message) {
bot.startConversation(message, function (err, convo) {
convo.say('This is an example of using convo.ask with a single callback.');
convo.ask('What is your favorite color?', function (response, convo) {
convo.say('Cool, I like ' + response.text + ' too!');
convo.next();
});
});
});
My bot does not wait for a response and just goes straight to 'Cool, I like (blank) too!'
I am using a Huawei p9 lite and hosting on Heroku.
Any ideas?
Thanks
You should post general Botkit questions over on the main repo:
https://github.com/howdyai/botkit/issues
or try asking on our Public Slack in the #botkit channel:
https://community.botkit.ai/
| gharchive/issue | 2017-12-18T11:42:08 | 2025-04-01T06:44:27.979169 | {
"authors": [
"dale-french",
"peterswimm"
],
"repo": "howdyai/botkit-starter-facebook",
"url": "https://github.com/howdyai/botkit-starter-facebook/issues/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2168242139 | Access to huggingface vqvae
Thanks for open-sourcing this incredible repo!
I found that if specifying 'vqvae' for --compressor argument in train.py, it requires access to the pretrained model on Huggingface. Could you please provide the access to that model?
Best
Hi @quantumiracle We've opened it. Thanks.
https://huggingface.co/hpcai-tech/vqvae
| gharchive/issue | 2024-03-05T03:51:11 | 2025-04-01T06:44:28.007097 | {
"authors": [
"binmakeswell",
"quantumiracle"
],
"repo": "hpcaitech/Open-Sora",
"url": "https://github.com/hpcaitech/Open-Sora/issues/29",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
801740449 | Implement raw frame cleanup
Make the default behaviour that the Exposure() class empties raw frames after CDS is performed, to save on memory (and include a parameter to turn this behaviour on and off e.g. cleanup=True/False)
On initialisation, Exposure() takes the optional parameter cleanup (defaults to True). If cleanup is True, function cleanup_frames() is called, which empties out the raw_frames attribute.
| gharchive/issue | 2021-02-05T00:32:00 | 2025-04-01T06:44:28.182576 | {
"authors": [
"hpearnshaw"
],
"repo": "hpearnshaw/cmost-analysis",
"url": "https://github.com/hpearnshaw/cmost-analysis/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
874466422 | Use GitHub Artifact for tdlib build
Do not download from some random DropBox, but keep the artifact in this repository and possibly automatically put them in persistent storage somewhere.
I'm unsure what is meant with "artifact" in the descripton of this issue. However it seems the TelegramClient is trying to download libtdjson.so.1.6.0 from a Google drive location if it is not found in the current directory. If this is indeed what this issue is about , the TDLib library could be installed by external means such as by operating system specific commands to install the shared library from a repository of the operating system .... As an example, it can be done on some systems with:
pkg install telegramdb
and then
cp /usr/lib/amd64/libtdjson.so.1.6.0 .
to copy to the current directory.
see #526
| gharchive/issue | 2021-05-03T10:55:19 | 2025-04-01T06:44:28.191230 | {
"authors": [
"antonykamp",
"cstes",
"phkeese"
],
"repo": "hpi-swa-teaching/TelegramClient",
"url": "https://github.com/hpi-swa-teaching/TelegramClient/issues/324",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
129377886 | Überschriften einzelner Seiten
Jede Seite sollte eine Überschrift haben. Diese sollte nicht zu groß sein (nicht h1!).
Die Überschrift soll dem Nutzer Kenntnis darüber geben, wo er sich im System gerade befindet. Ist er zum Beispiel gerade auf dem Dashboard? Sieht er sich gerade die Detailseite eines Urlaubsantrags an?
Wenn der Nutzer selbst die Details seiner Anträge einsieht, so reicht eine Überschrift, die darüber informiert, welchen Antrag er gerade einsieht. Sieht der FG-Beauftragte die Details für einen Antrag eines Wimis ein, ist es notwendig, auch zu sehen, von wem der Antrag gestellt worden ist.
Solved in #311
| gharchive/issue | 2016-01-28T08:30:58 | 2025-04-01T06:44:28.192969 | {
"authors": [
"arneboockmeyer",
"mandyklingbeil"
],
"repo": "hpi-swt2/wimi-portal",
"url": "https://github.com/hpi-swt2/wimi-portal/issues/276",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.