added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:39:46.766863
| 2024-12-10T07:56:20
|
2729316766
|
{
"authors": [
"EunjuYang",
"myungjoo"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8977",
"repo": "nnstreamer/nntrainer",
"url": "https://github.com/nnstreamer/nntrainer/pull/2823"
}
|
gharchive/pull-request
|
[ Meson ] BCQTensor dependency is added for Android build
This commit add BiQGEMM path to nntrainer_inc_abs to support Android build with enable-biqgemm option.
The path for BiQGEMM corresponds to the one in the top meson.build
Self evaluation:
Build test: [X]Passed [ ]Failed [ ]Skipped
Run test: [X]Passed [ ]Failed [ ]Skipped
LGTM! One minor suggestion: How about adding it to the tensor/meson.build for consistency?
https://github.com/nnstreamer/nntrainer/blob/cd17a66b8eec45a2e91e543a884722763b5e608e/nntrainer/tensor/meson.build#L79-L83
Following the suggestion from @djeong20, I moved the code to nntrainer/tensor/meson.build.
Since the existence check for BiQGEMM folder is completed at the precedent meson file, the duplicate check code is removed.
meson.source_root() / '..' / 'BiQGEMM'
DO NOT HARDCODE EXTERNAL PATHS IN MESON SCRIPT!
If pkgconfig or cmake is not ready with BiQGEMM,
get such information via meson_options and specify such "default paths" at meson_options.
And..
if get_option('enable-biqgemm')
# check if BiQGEMM directory exist. otherwise, throw an error
fs = import('fs')
if fs.is_dir('../BiQGEMM')
extra_defines += '-DENABLE_BIQGEMM=1'
biqgemm_inc = include_directories('../BiQGEMM')
else
error ('BiQGEMM cannot be enabled without BiQGEMM library.')
endif
endif
is completely non-sense. Other developers (e.g., nntrainer users in other departments) won't understand what's going on here.
You should try
get biqgemm info from pkgconfig/cmake.
(if 1 fails) try to load it from common path (/usr/include, /usr/include/biqgemm)
(if 2 fails) try to load it from user-defined path (meson_option)
(if 3 fails) try to load it from such hardcoded path. but specify such external hardcoded path at meson_options.txt, not in this script.
@myungjoo, Thank you for the detailed comment and guidance :) I will make additional PR to make it feasible following your suggestion. Sincerely.
|
2025-04-01T06:39:46.792216
| 2023-10-13T11:10:26
|
1941742213
|
{
"authors": [
"Rookie764",
"noahzn"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8978",
"repo": "noahzn/Lite-Mono",
"url": "https://github.com/noahzn/Lite-Mono/issues/75"
}
|
gharchive/issue
|
Dataset
hello,I want to know the information about how to split the kitti dataset,in the paper ,there are 39180 for training ,but monodepth2 has 39810 for training,will the difference affect results?
Hi, I think this should be a typo in our paper. We used the same training dataset as Monodepth2.
|
2025-04-01T06:39:46.844159
| 2024-05-21T18:46:00
|
2308899991
|
{
"authors": [
"jsetoain",
"newling",
"yzhang93"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8979",
"repo": "nod-ai/iree-amd-aie",
"url": "https://github.com/nod-ai/iree-amd-aie/issues/363"
}
|
gharchive/issue
|
matmul-elementwise fusion failed to legalize operation 'vector.contract'
With the support of fusing consumer (tensor.unpack) into the loop, I've modified the matmul-elementwise fusion with pack-peel pipeline. Now i32 test compiles and has correct outputs on hardware. However, bf16 test failed at AIE stage with error:
error: failed to legalize operation 'vector.contract' that was explicitly marked illegal
I've attached the dump IR here
afterall_aie.txt
Already talked to @erwei-xilinx about the issue, but would like to have more eyes on it, since this seems to be related to the vectorization problem. @newling @jsetoain
I've taking a look at the first few and they all seem valid, I'll take a closer look tomorrow.
The vector.contract is all bf16. What is the input IR, does the elementwise operation happen in f32 or bf16?
Currently we don't support matmul with result type bf16. I've been meaning to improve the error diagnostic error in the case where the matmul is bf16->bf16 as this issue has been encountered at least twice before.
The vector.contract is all bf16. What is the input IR, does the elementwise operation happen in f32 or bf16?
Currently we don't support matmul with result type bf16. I've been meaning to improve the error diagnostic error in the case where the matmul is bf16->bf16 as this issue has been encountered at least twice before.
Okay, I see. The output type of this test is bf16. Let me try another test with elementwise/output operation in f32.
Then here comes another question... Do you know what is the data type of elementwise op in tres leches bf16 model?
The vector.contract is all bf16. What is the input IR, does the elementwise operation happen in f32 or bf16?
Currently we don't support matmul with result type bf16. I've been meaning to improve the error diagnostic error in the case where the matmul is bf16->bf16 as this issue has been encountered at least twice before.
Okay, I see. The output type of this test is bf16. Let me try another test with elementwise/output operation in f32.
Then here comes another question... Do you know what is the data type of elementwise op in tres leches bf16 model?
Yes, I just confirmed that if I change the matmul output and elementwise type to f32 it works without problem! @newling Thanks for pointing out the issue.
|
2025-04-01T06:39:47.135983
| 2018-03-12T15:58:19
|
304432755
|
{
"authors": [
"Flarna",
"kkoopa"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8980",
"repo": "nodejs/nan",
"url": "https://github.com/nodejs/nan/pull/754"
}
|
gharchive/pull-request
|
avoid unneeded call to Isolate::GetCurrent() in converters
I have seen an unneeded call to Isolate::GetCurrent() in converters during stepping around. Even the overhead of this is small it's unneeded.
Thank you. Seems nobody has noticed this in over a year.
|
2025-04-01T06:39:47.491619
| 2016-04-05T09:28:55
|
145934926
|
{
"authors": [
"AdrianRossouw"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8981",
"repo": "nodezoo/nodezoo-github",
"url": "https://github.com/nodezoo/nodezoo-github/pull/47"
}
|
gharchive/pull-request
|
less generic token.
fixes #3
we are going to have multiple tokens configured in this app, so we have a shared namespace.
merged but github broke during
|
2025-04-01T06:39:47.495183
| 2016-12-27T12:14:01
|
197694652
|
{
"authors": [
"mattecapu",
"nodkz"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8982",
"repo": "nodkz/react-relay-network-layer",
"url": "https://github.com/nodkz/react-relay-network-layer/issues/30"
}
|
gharchive/issue
|
Ejecting express-middleware to its own module
It doesn't really belong to the network layer. It's just a mean to support the "batching protocol" that should be separated from the rest.
Not everybody use express. I use koa and I made koa-graphql-batch to use this network layer, thus I don't need express-middleware.
If one doesn't bother to do properly minification/tree-shaking, the middleware code will end up client side where it's all but needed.
I could take care of this if @nodkz agrees.
@mattecapu 👍
I agree with you. Express-middleware should be extracted from RRNL. In the next month I'll start implementing subscriptions and bump new major version.
Before this moment let keep untouched this extra 1.6 kb of express-middleware for semver compatibility.
BTW. Please add a link to your koa-graphql-batch to readme. Thanks!
|
2025-04-01T06:39:47.496181
| 2022-08-11T08:52:49
|
1335661244
|
{
"authors": [
"Barragek0",
"mrhappyasthma"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8983",
"repo": "nofatigue/runelite-profit-tracker",
"url": "https://github.com/nofatigue/runelite-profit-tracker/pull/32"
}
|
gharchive/pull-request
|
Remove parallel streaming
Parallel streams are run on a seperate thread causing AssertionError: must be run on client thread when calling ItemManager.getItemPrice
I confirmed this fixes the crash. Can we get this merged in?
|
2025-04-01T06:39:47.499696
| 2024-05-23T16:02:37
|
2454632995
|
{
"authors": [
"jon-nfc"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8984",
"repo": "nofusscomputing/centurion_erp",
"url": "https://github.com/nofusscomputing/centurion_erp/issues/26"
}
|
gharchive/issue
|
docker log format
the docker log format does not include the source IP address. correct so that it does (including using the proxy forward-for header)
added 2m of time spent
closing as release v1.3 #371 made changes forcing this issue now to be stale.
|
2025-04-01T06:39:47.509193
| 2022-09-04T12:04:20
|
1361128609
|
{
"authors": [
"MurdeRM3L0DY",
"noib3"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8985",
"repo": "noib3/nvim-completion",
"url": "https://github.com/noib3/nvim-completion/issues/19"
}
|
gharchive/issue
|
segfaults when calling setup
Hi. Would you mind sharing what versions of neovim, nvim-oxi and rustc you use to test nvim-completion?
I've tried with rustc 1.65.0-nightly/1.63.0, neovim (NVIM v0.8.0-dev-1021-g24fbda04b
Build type: Release), nvim-oxi latest master. Could you also share your config options for nvim-completion? Thanks.
rustc: rustc 1.65.0-nightly (e1b28cd2f 2022-08-19);
nvim: NVIM v0.7.2;
nvim-oxi: latest master, updated in (d6f48b06a448b73135dafbfbc771ce9dd61efead);
Config:
-- config.lua
local completion = require("nvim_completion")
completion.setup({
sources = {
lipsum = { enable = function(_buf) return true end },
lsp = { enable = true },
},
})
Starting Neovim w/ nvim --clean -u ./config.lua.
Then you need to open a new buffer for the sources to attach to via :e <somefilename>.
Can you share your config? It shouldn't segfault. Does the segfault also happen if you use Neovim 0.7.2 instead of nightly?
Yup, just confirmed it only segfaults with neovim latest master.
That's not a segfault, that's just a panic (which I'm aware of).
Yup, I figured you would.
|
2025-04-01T06:39:47.522163
| 2022-12-05T00:30:08
|
1475339859
|
{
"authors": [
"jmcph4",
"kevaundray"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8986",
"repo": "noir-lang/noir",
"url": "https://github.com/noir-lang/noir/issues/556"
}
|
gharchive/issue
|
noirup fails due to Barretenberg build failure
Description
Aim
Successful installation of the nargo tool in order to perform Noir development.
Expected behavior
For invocation of noirup to successfully install nargo. Something along the lines of:
$ noirup
# ...
$ echo $?
0
$ nargo
# ...
$ echo $?
0
Bug
$ sudo apt-get install libomp-dev cmake
$ curl -L https://raw.githubusercontent.com/noir-lang/noir/master/noirup/install | bash
$ noirup
HEAD is now at e239f415 Handle predicate operator during inlining (#544)
Installing nargo v0.1.0 (/home/jmcph4/.nargo/noir-lang/noir/crates/nargo)
Updating crates.io index
Compiling psm v0.1.21
Compiling barretenberg_wrapper v0.1.0 (https://github.com/AztecProtocol/barretenberg?rev=804c7dcf21111acd1302a768a8fa2f453dcec50f#804c7dcf)
error: failed to run custom build command for `barretenberg_wrapper v0.1.0 (https://github.com/AztecProtocol/barretenberg?rev=804c7dcf21111acd1302a768a8fa2f453dcec50f#804c7dcf)`
Caused by:
process didn't exit successfully: `/home/jmcph4/.nargo/noir-lang/noir/target/release/build/barretenberg_wrapper-6e0ec251307286f5/build-script-build` (exit status: 101)
--- stdout
CMAKE_TOOLCHAIN_FILE_x86_64-unknown-linux-gnu = None
CMAKE_TOOLCHAIN_FILE_x86_64_unknown_linux_gnu = None
HOST_CMAKE_TOOLCHAIN_FILE = None
CMAKE_TOOLCHAIN_FILE = None
CMAKE_GENERATOR_x86_64-unknown-linux-gnu = None
CMAKE_GENERATOR_x86_64_unknown_linux_gnu = None
HOST_CMAKE_GENERATOR = None
CMAKE_GENERATOR = None
CMAKE_PREFIX_PATH_x86_64-unknown-linux-gnu = None
CMAKE_PREFIX_PATH_x86_64_unknown_linux_gnu = None
HOST_CMAKE_PREFIX_PATH = None
CMAKE_PREFIX_PATH = None
CMAKE_x86_64-unknown-linux-gnu = None
CMAKE_x86_64_unknown_linux_gnu = None
HOST_CMAKE = None
CMAKE = None
running: "cmake" "/home/jmcph4/.cargo/git/checkouts/barretenberg-6ce83dfea69613eb/804c7dc/barretenberg_wrapper/../barretenberg" "-DCMAKE_INSTALL_PREFIX=/home/jmcph4/.nargo/noir-lang/noir/target/release/build/barretenberg_wrapper-24257eaf3fa79116/out" "-DCMAKE_C_FLAGS= -ffunction-sections -fdata-sections -fPIC -m64" "-DCMAKE_C_COMPILER=/usr/bin/cc" "-DCMAKE_CXX_FLAGS= -ffunction-sections -fdata-sections -fPIC -m64" "-DCMAKE_CXX_COMPILER=/usr/bin/c++" "-DCMAKE_ASM_FLAGS= -ffunction-sections -fdata-sections -fPIC -m64" "-DCMAKE_ASM_COMPILER=/usr/bin/cc" "-DCMAKE_BUILD_TYPE=Release"
-- The CXX compiler identification is GNU 12.2.0
-- The C compiler identification is GNU 12.2.0
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working C compiler: /usr/bin/cc - skipped
-- Detecting C compile features
-- Detecting C compile features - done
-- Found OpenMP_C: -fopenmp (found version "4.5")
-- Found OpenMP_CXX: -fopenmp (found version "4.5")
-- Found OpenMP: TRUE (found version "4.5")
-- Multithreading is enabled.
-- Found Python: /home/jmcph4/.pyenv/shims/python3 (found version "3.9.0") found components: Interpreter
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success
-- Found Threads: TRUE
-- Failed to find LLVM FileCheck
-- Found Git: /usr/bin/git (found version "2.37.2")
-- git version: v1.6.1 normalized to 1.6.1
-- Version: 1.6.1
-- Looking for shm_open in rt
-- Looking for shm_open in rt - found
-- Performing Test HAVE_CXX_FLAG_STD_CXX11
-- Performing Test HAVE_CXX_FLAG_STD_CXX11 - Success
-- Performing Test HAVE_CXX_FLAG_WALL
-- Performing Test HAVE_CXX_FLAG_WALL - Success
-- Performing Test HAVE_CXX_FLAG_WEXTRA
-- Performing Test HAVE_CXX_FLAG_WEXTRA - Success
-- Performing Test HAVE_CXX_FLAG_WSHADOW
-- Performing Test HAVE_CXX_FLAG_WSHADOW - Success
-- Performing Test HAVE_CXX_FLAG_WERROR
-- Performing Test HAVE_CXX_FLAG_WERROR - Success
-- Performing Test HAVE_CXX_FLAG_WSUGGEST_OVERRIDE
-- Performing Test HAVE_CXX_FLAG_WSUGGEST_OVERRIDE - Success
-- Performing Test HAVE_CXX_FLAG_PEDANTIC
-- Performing Test HAVE_CXX_FLAG_PEDANTIC - Success
-- Performing Test HAVE_CXX_FLAG_PEDANTIC_ERRORS
-- Performing Test HAVE_CXX_FLAG_PEDANTIC_ERRORS - Success
-- Performing Test HAVE_CXX_FLAG_WSHORTEN_64_TO_32
-- Performing Test HAVE_CXX_FLAG_WSHORTEN_64_TO_32 - Failed
-- Performing Test HAVE_CXX_FLAG_FSTRICT_ALIASING
-- Performing Test HAVE_CXX_FLAG_FSTRICT_ALIASING - Success
-- Performing Test HAVE_CXX_FLAG_WNO_DEPRECATED_DECLARATIONS
-- Performing Test HAVE_CXX_FLAG_WNO_DEPRECATED_DECLARATIONS - Success
-- Performing Test HAVE_CXX_FLAG_WNO_DEPRECATED
-- Performing Test HAVE_CXX_FLAG_WNO_DEPRECATED - Success
-- Performing Test HAVE_CXX_FLAG_WSTRICT_ALIASING
-- Performing Test HAVE_CXX_FLAG_WSTRICT_ALIASING - Success
-- Performing Test HAVE_CXX_FLAG_WD654
-- Performing Test HAVE_CXX_FLAG_WD654 - Failed
-- Performing Test HAVE_CXX_FLAG_WTHREAD_SAFETY
-- Performing Test HAVE_CXX_FLAG_WTHREAD_SAFETY - Failed
-- Performing Test HAVE_CXX_FLAG_COVERAGE
-- Performing Test HAVE_CXX_FLAG_COVERAGE - Success
-- Performing Test HAVE_STD_REGEX
-- Performing Test HAVE_STD_REGEX
-- Performing Test HAVE_STD_REGEX -- success
-- Performing Test HAVE_GNU_POSIX_REGEX
-- Performing Test HAVE_GNU_POSIX_REGEX
-- Performing Test HAVE_GNU_POSIX_REGEX -- failed to compile
-- Performing Test HAVE_POSIX_REGEX
-- Performing Test HAVE_POSIX_REGEX
-- Performing Test HAVE_POSIX_REGEX -- success
-- Performing Test HAVE_STEADY_CLOCK
-- Performing Test HAVE_STEADY_CLOCK
-- Performing Test HAVE_STEADY_CLOCK -- success
-- Using optimized assembly for field arithmetic.
-- Looking for unistd.h
-- Looking for unistd.h - found
-- Looking for crc32c_value in crc32c
-- Looking for crc32c_value in crc32c - not found
-- Looking for snappy_compress in snappy
-- Looking for snappy_compress in snappy - not found
-- Looking for malloc in tcmalloc
-- Looking for malloc in tcmalloc - not found
-- Looking for fdatasync
-- Looking for fdatasync - found
-- Looking for F_FULLFSYNC
-- Looking for F_FULLFSYNC - not found
-- Performing Test HAVE_CLANG_THREAD_SAFETY
-- Performing Test HAVE_CLANG_THREAD_SAFETY - Failed
-- Performing Test HAVE_CXX17_HAS_INCLUDE
-- Performing Test HAVE_CXX17_HAS_INCLUDE - Success
-- Looking for sqlite3_open in sqlite3
-- Looking for sqlite3_open in sqlite3 - found
-- Performing Test HAVE_KYOTOCABINET
-- Performing Test HAVE_KYOTOCABINET - Failed
-- Configuring done
-- Generating done
-- Build files have been written to: /home/jmcph4/.nargo/noir-lang/noir/target/release/build/barretenberg_wrapper-24257eaf3fa79116/out/build
running: "cmake" "--build" "." "--target" "install" "--config" "Release"
[ 0%] Building CXX object src/aztec/env/CMakeFiles/env_objects.dir/logstr.cpp.o
[ 0%] Built target env_objects
[ 1%] Building CXX object src/aztec/numeric/CMakeFiles/numeric_objects.dir/random/engine.cpp.o
[ 1%] Built target numeric_objects
[ 2%] Building CXX object _deps/googletest-build/googletest/CMakeFiles/gtest.dir/src/gtest-all.cc.o
--- stderr
CMake Warning:
Manually-specified variables were not used by the project:
CMAKE_ASM_COMPILER
CMAKE_ASM_FLAGS
In file included from /usr/include/c++/12/ios:40,
from /usr/include/c++/12/ostream:38,
from /usr/include/c++/12/bits/unique_ptr.h:41,
from /usr/include/c++/12/memory:76,
from /home/jmcph4/.nargo/noir-lang/noir/target/release/build/barretenberg_wrapper-24257eaf3fa79116/out/build/_deps/googletest-src/googletest/include/gtest/gtest.h:57,
from /home/jmcph4/.nargo/noir-lang/noir/target/release/build/barretenberg_wrapper-24257eaf3fa79116/out/build/_deps/googletest-src/googletest/src/gtest-all.cc:38:
In static member function ‘static constexpr std::char_traits<char>::char_type* std::char_traits<char>::copy(char_type*, const char_type*, std::size_t)’,
inlined from ‘static constexpr void std::__cxx11::basic_string<_CharT, _Traits, _Alloc>::_S_copy(_CharT*, const _CharT*, size_type) [with _CharT = char; _Traits = std::char_traits<char>; _Alloc = std::allocator<char>]’ at /usr/include/c++/12/bits/basic_string.h:423:21,
inlined from ‘constexpr std::__cxx11::basic_string<_CharT, _Traits, _Allocator>& std::__cxx11::basic_string<_CharT, _Traits, _Alloc>::_M_replace(size_type, size_type, const _CharT*, size_type) [with _CharT = char; _Traits = std::char_traits<char>; _Alloc = std::allocator<char>]’ at /usr/include/c++/12/bits/basic_string.tcc:532:22,
inlined from ‘constexpr std::__cxx11::basic_string<_CharT, _Traits, _Alloc>& std::__cxx11::basic_string<_CharT, _Traits, _Alloc>::replace(size_type, size_type, const _CharT*, size_type) [with _CharT = char; _Traits = std::char_traits<char>; _Alloc = std::allocator<char>]’ at /usr/include/c++/12/bits/basic_string.h:2171:19,
inlined from ‘constexpr std::__cxx11::basic_string<_CharT, _Traits, _Alloc>& std::__cxx11::basic_string<_CharT, _Traits, _Alloc>::insert(size_type, const _CharT*) [with _CharT = char; _Traits = std::char_traits<char>; _Alloc = std::allocator<char>]’ at /usr/include/c++/12/bits/basic_string.h:1928:22,
inlined from ‘constexpr std::__cxx11::basic_string<_CharT, _Traits, _Allocator> std::operator+(const _CharT*, __cxx11::basic_string<_CharT, _Traits, _Allocator>&&) [with _CharT = char; _Traits = char_traits<char>; _Alloc = allocator<char>]’ at /usr/include/c++/12/bits/basic_string.h:3541:36,
inlined from ‘static std::string testing::internal::StreamingListener::UrlEncode(const char*)’ at /home/jmcph4/.nargo/noir-lang/noir/target/release/build/barretenberg_wrapper-24257eaf3fa79116/out/build/_deps/googletest-src/googletest/src/gtest.cc:4882:27:
/usr/include/c++/12/bits/char_traits.h:431:56: error: ‘void* __builtin_memcpy(void*, const void*, long unsigned int)’ accessing<PHONE_NUMBER>854775810 or more bytes at offsets [2,<PHONE_NUMBER>854775807] and 1 may overlap up to<PHONE_NUMBER>854775813 bytes at offset -3 [-Werror=restrict]
431 | return static_cast<char_type*>(__builtin_memcpy(__s1, __s2, __n));
| ~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~
cc1plus: all warnings being treated as errors
gmake[2]: *** [_deps/googletest-build/googletest/CMakeFiles/gtest.dir/build.make:76: _deps/googletest-build/googletest/CMakeFiles/gtest.dir/src/gtest-all.cc.o] Error 1
gmake[1]: *** [CMakeFiles/Makefile2:1185: _deps/googletest-build/googletest/CMakeFiles/gtest.dir/all] Error 2
gmake: *** [Makefile:146: all] Error 2
thread 'main' panicked at '
command did not execute successfully, got: exit status: 2
build script failed, must exit now', /home/jmcph4/.cargo/registry/src/github.com-1ecc6299db9ec823/cmake-0.1.49/src/lib.rs:1104:5
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
error: failed to compile `nargo v0.1.0 (/home/jmcph4/.nargo/noir-lang/noir/crates/nargo)`, intermediate artifacts can be found at `/home/jmcph4/.nargo/noir-lang/noir/target`
noirup: command failed: cargo install --path ./crates/nargo --bins --locked --force --root /home/jmcph4/.nargo
To reproduce
Install noirup (as per https://github.com/noir-lang/noir/tree/1d1b592039fa73ce561d6b6886c36b02b36ada53/noirup#readme)
Invoke noirup
Environment
$ uname --all
Linux foobar 5.19.0-2-amd64 #1 SMP PREEMPT_DYNAMIC Debian 5.19.11-1 (2022-09-24) x86_64 GNU/Linux
$ cat /etc/os-release
PRETTY_NAME="Debian GNU/Linux bookworm/sid"
NAME="Debian GNU/Linux"
ID=debian
HOME_URL="https://www.debian.org/"
SUPPORT_URL="https://www.debian.org/support"
BUG_REPORT_URL="https://bugs.debian.org/"
Additional context
I initially attempted to follow the official documentation, but this failed for other reasons. I then found noirup and decided it was the more graceful approach.
The build appears to be failing during the Barretenberg build section. I appear to satisfy the Barretenberg dependences.
Since moving to nix, this issue has been solved. Feel free to reopen if you encounter it again!
|
2025-04-01T06:39:47.552117
| 2023-05-26T19:59:24
|
1728166166
|
{
"authors": [
"jfecher",
"kevaundray"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8987",
"repo": "noir-lang/noir",
"url": "https://github.com/noir-lang/noir/pull/1418"
}
|
gharchive/pull-request
|
chore(ssa refactor): Adds basic program with empty body to experimental-ssa test corpus
Description
This adds the program from #1403 into the test corpus.
Problem*
Resolves
Summary*
This PR sets out to
Example
Before:
After:
Documentation
[ ] This PR requires documentation updates when merged.
[ ] I will submit a noir-lang/docs PR.
[ ] I will request for and support Dev Rel's help in documenting this PR.
Additional Context
PR Checklist*
[ ] I have tested the changes locally.
[ ] I have formatted the changes with Prettier and/or cargo fmt on default settings.
Copying the CI output:
---- tests::noir_integration_ssa_refactor stdout ----
Running test simple_program_no_body
Initial SSA:
fn main f0 {
b0(v0: Field, v1: Field):
return unit 0
}
After Inlining:
fn main f1 {
b0(v0: Field, v1: Field):
return unit 0
}
After Unrolling:
fn main f1 {
b0(v0: Field, v1: Field):
return unit 0
}
After Simplifying:
fn main f1 {
b0(v0: Field, v1: Field):
return unit 0
}
After Flattening:
fn main f1 {
b0(v0: Field, v1: Field):
return unit 0
}
After Mem2Reg:
fn main f1 {
b0(v0: Field, v1: Field):
return unit 0
}
thread 'main' panicked at 'ICE: Return of value not yet encountered', crates/noirc_evaluator/src/ssa_refactor/acir_gen/mod.rs:113:28
Error is occuring because when we try to do acir_gen for the return instruction, we:
Get the return values
Check to see if we've acir_gen'd this return value and panic if not
The second point assumes that the return value will always be a variable that was encountered before, whereas return unit 0 is a constant that the acir_gen pass has never seen before
Looks good to me but we should avoid merging until the issue in acir-gen is fixed
ssa-gen.
Yep will debug this -- I think it would also trigger, if this returns any ValueId which was not used in another part of the program, so I imagine just returning a constant would also trigger this bug
Seems there is another error in acir_gen:
CompiledProgram {
circuit: current witness index : 3
public parameters indices : [1, 2]
return value indices : [3]
EXPR [ (-1, _3) 0 ]
,
abi: Abi {
parameters: [
AbiParameter {
name: "_x",
typ: Field,
visibility: Private,
},
AbiParameter {
name: "_y",
typ: Field,
visibility: Public,
},
],
param_witnesses: {
"_x": [
Witness(
1,
),
],
"_y": [
Witness(
2,
),
],
},
return_type: None,
return_witnesses: [
Witness(
3,
),
],
},
}
Serialized transcript does not contain the required number of bytes
Serialized transcript does not contain the required number of bytes
This error really only happens in the barretenberg when there is a discrepancy between the circuit or proving system. Since the proving system has not changed, then we can then deduce that something is malformed about the circuit.
In particular, that error arises when there is a mismatch between public inputs.
To fix:
If the return is unit 0 then we should treat this as not having a return parameter. We should then see return value indices : [] ie no returns. This should also get rid of EXPR [ (-1, _3) 0 ].
The next issue is public parameters indices : [1, 2] we should have ``public parameters indices : [2]` indicating that there is one public parameter and it is the second parameter which should have index 2
Also going to block this PR until we check for:
fn main() -> pub Field {
return <constant>;
}
To make sure we don't mess something up there
Tested this with a constant return and that also passes :) This passes for me locally so just waiting for it to pass on CI
|
2025-04-01T06:39:47.559404
| 2022-12-17T15:08:54
|
1501432885
|
{
"authors": [
"noirello",
"senfomat"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8988",
"repo": "noirello/bonsai",
"url": "https://github.com/noirello/bonsai/issues/75"
}
|
gharchive/issue
|
Raise specific Exceptions when LDAP-connections fail due to SSL-problems
Issue:
I recently ran multiple times into the problem, that LDAP-connection-attempts were unsuccessful, because the server-certificates could not got verified. This was due to a lack of the corresponding CA-Certificate in the central CA-system-store (in our case Ubuntu Linux).
When I was tracking down these connection-problems, I had to do trial&error to nail it down to this certificate-thing. Because bonsai just raises bonsai.errors.ConnectionError: Can't contact LDAP server. (unknown error code) (0xFFFF [-1]).
Is there a possibility to raise more specific exceptions on these SSL-related errors?
Setup:
Linux Ubuntu 20.04
Python 3.11.0
bonsai v1.5.1
Testscript:
#!/usr/bin/env python
import bonsai
from bonsai.errors import AuthenticationError, ConnectionError
bonsai.set_debug(True)
client = bonsai.LDAPClient("ldaps://ldap.domain.local")
client.set_credentials("SIMPLE", user="testuser", password="testpassword")
# client.set_cert_policy("allow")
try:
returnValue = client.connect()
except ConnectionError as err:
print(f"LDAP-Connection failed: {err}")
except AuthenticationError as err:
print(f"LDAP-Authentication failed: {err}")
except Exception as e:
print(f"Generic exception: {e}")
Output:
DBG: ldapconnection_new [self:0x7f7bcc9a35b0]
DBG: ldapconnection_init (self:0x7f7bcc9a35b0)
DBG: ldapconnection_open (self:0x7f7bcc9a35b0)
DBG: connecting (self:0x7f7bcc9a35b0)
DBG: create_conn_info (mech:SIMPLE, sock:-1, creds:0x7f7bcdf03f00)
DBG: ldapconnectiter_new [self:0x7f7bcc802740]
DBG: create_init_thread_data (client:0x7f7bcdc476d0, sock:-1)
DBG: create_init_thread (ld:0x24db110, info:0x24e7860, thread:0)
DBG: ldapconnection_result (self:0x7f7bcc9a35b0, args:0x7f7bcdd2ee80, kwds:(nil))[msgid:-1]
DBG: LDAPConnection_Result (self:0x7f7bcc9a35b0, msgid:-1, millisec:-1)
DBG: LDAPConnectIter_Next (self:0x7f7bcc802740, timeout:-1) [tls:0, state:0]
DBG: _ldap_finish_init_thread (async:0, thread:140169686472448, timeout:-1, misc:0x24db110)
DBG: _pthread_mutex_timedlock
DBG: ldap_init_thread_func (params:0x24db110)
DBG: set connecting async: 0
DBG: ldap_init_thread_func [retval:0]
DBG: LDAPConnectIter_Next (self:0x7f7bcc802740, timeout:-1) [tls:0, state:0]
DBG: _ldap_finish_init_thread (async:0, thread:140169686472448, timeout:-1, misc:0x24db110)
DBG: _pthread_mutex_timedlock
DBG: set_certificates (self:0x7f7bcc802740)
DBG: binding [state:3]
DBG: _ldap_bind (ld:0x7f7bc4000b60, info:0x24e7860, ppolicy:0, result:(nil), msgid:0)
DBG: ldapconnectiter_dealloc (self:0x7f7bcc802740)
DBG: dealloc_conn_info (info:0x24e7860)
LDAP-Connection failed: Can't contact LDAP server. (unknown error code) (0xFFFF [-1])
DBG: ldapconnection_dealloc (self:0x7f7bcc9a35b0)
(When I uncomment the line client.set_cert_policy("allow") in my code the connection gets successfully established.)
Unfortunately, If OpenLDAP doesn't set a specific return value or set a diagnostic message, then I don't think it's possible to raise a specific error.
The raised exception is based on the LDAP error code (returned by an LDAP function call or set to the LDAP structure's corresponding field), and if additional diagnostic message is provided, then it's concatenated to the exception's error message.
TLS related errors are usually only shown in libldap's trace logs. You can set trace level logging with bonsai.set_debug(True, -1).
Ok, I agree, that this is an openldap-Libraryissue. So I'll close the issue.
But thanks for the hint to the trace-logs. There I can see the real error:
[...]
DBG: _ldap_bind (ld:0x7f3764000b60, info:0x1943560, ppolicy:0, result:(nil), msgid:0)
ldap_sasl_bind
ldap_send_initial_request
ldap_new_connection 1 1 0
ldap_int_open_connection
ldap_connect_to_host: TCP ldap.domain.local:636
ldap_new_socket: 3
ldap_prepare_socket: 3
ldap_connect_to_host: Trying <IP_ADDRESS>:636
ldap_pvt_connect: fd: 3 tm: -1 async: 0
attempting to connect:
connect success
TLS: peer cert untrusted or revoked (0x42)
TLS: can't connect: (unknown error code).
ldap_msgfree
ldap_err2string
DBG: ldapconnectiter_dealloc (self:0x7f376a99a740)
[...]
|
2025-04-01T06:39:47.567072
| 2021-07-25T16:13:53
|
952306719
|
{
"authors": [
"nolanlawson"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8989",
"repo": "nolanlawson/emoji-picker-element",
"url": "https://github.com/nolanlawson/emoji-picker-element/pull/193"
}
|
gharchive/pull-request
|
fix: fix types for event listeners, remove typedoc
fixes #191
After looking at dom.d.ts a bit I figured out that I was missing some declarations for addEventListener/removeEventListener, which is why it would fail in strict mode. Unfortunately I couldn't figure out how to make these changes in the .ts files, but we can just put them directly in the .d.ts files and modify those directly, since I'm not really using Typedoc anymore so there's no need to keep it around.
This is the magic; you need these type: string declarations too:
https://github.com/nolanlawson/emoji-picker-element/blob/39e1ce51fb1f9efa7d5d292c2966be3dbcf19247/picker.d.ts#L20-L23
|
2025-04-01T06:39:47.568855
| 2017-07-07T13:59:16
|
241272538
|
{
"authors": [
"julmot",
"nolimits4web"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8990",
"repo": "nolimits4web/Swiper",
"url": "https://github.com/nolimits4web/Swiper/issues/2148"
}
|
gharchive/issue
|
Implement zoom methods and implement mousewheel support
This is a (multiple allowed):
[ ] bug
[x] enhancement
[ ] feature-discussion (RFC)
There should be zoomIn() zoomOut() and zoomReset() methods. The zoom should also have an option to work with mousewheel.
They are in v4
|
2025-04-01T06:39:47.570832
| 2014-08-14T14:13:53
|
40257872
|
{
"authors": [
"mnifakram",
"zapatoche"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8991",
"repo": "nolimits4web/Swiper",
"url": "https://github.com/nolimits4web/Swiper/issues/896"
}
|
gharchive/issue
|
width and height using percentage
Hey I want to make something that mix between responsive and partial, I set the swiper-container{ width:100%; height:100%} and the swiper-slide{width:90%; height:100%} but in the output I get the swiper-slide width 100% (the same as the container) how I can correct this.
Check the demo http://jsfiddle.net/fnhb137d/
But if we know the width of the parent (on page load) can't we define the width of the slider-wrapper depending one the number of item, then set their width in % of the slider-wrapper width?
I'm asking because I have a project with "fluid" typography, and when I increase the font size on the body, the slider breaks in IE10/11 (because the swipes width is in pixel).
|
2025-04-01T06:39:47.572474
| 2017-02-01T17:35:35
|
204652109
|
{
"authors": [
"klojniewski",
"nolimits4web"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8992",
"repo": "nolimits4web/Swiper",
"url": "https://github.com/nolimits4web/Swiper/pull/1999"
}
|
gharchive/pull-request
|
mousewheelControl param setting when updating this using MousewheelCo…
Fix for:
https://github.com/nolimits4web/Swiper/issues/1998
Always follow the contribution guidelines when submitting a pull request.
Merged, thanks
|
2025-04-01T06:39:47.578078
| 2021-04-19T15:46:25
|
861456826
|
{
"authors": [
"Bimmr",
"nolimits4web"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8993",
"repo": "nolimits4web/swiper",
"url": "https://github.com/nolimits4web/swiper/issues/4454"
}
|
gharchive/issue
|
Pagination renderBullet putting last bullet in front of other bullets
Swiper Version: Swiper 6.5.7
Platform/Target and Browser Versions: Windows, Chrome
What You Did
Using custom pagination renderBullet to use custom pagination so we can define buttons as data attributes in each slide item.
JSFiddle with Example
Expected Behavior
Show custom rendered bullets in correct order
Actual Behavior
Last custom rendered bullet is being shown first
How about more minimal example without extra logic (that could be a root of the problem/issue)?
How about more minimal example without extra logic (that could be a root of the problem/issue)?
I'll edit the JS Fiddle, and let you know when it's been minimized
So while updating the JSFiddle to a more minimal example I found the route of the cause.
While having loop enabled it looks like the custom bullet render is grabbing the duplicated slide and trying to create a button with that.
Everything is correct with pagination render. index number it returns in renderBullet is not an index of slide, it is an index of the bullet. And you try to use same index for slide which is wrong, as in loop mode there are duplicated slides added
|
2025-04-01T06:39:47.605725
| 2023-11-15T12:40:16
|
1994701350
|
{
"authors": [
"tangledbytes"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8995",
"repo": "noobaa/noobaa-core",
"url": "https://github.com/noobaa/noobaa-core/pull/7601"
}
|
gharchive/pull-request
|
[NC | NSFS] WAL based tape migrations and recalls
Explain the changes
This PR adds WAL based migrations and recalls to NooBaa. The PR adds 2 commands manage_nsfs:
tape recall - Uses eeadm to restore all the files that were requested for restore.
tape migrate - Uses eeadm to migrate all the files to the tape.
NOTE: There were many races but now I hope I am dealing with all of them. What are the races I accounted for?
Multiple nodes issuing the migrate/recall at the same time (maybe cron doing it?).
Same nodes issue migrate and recall at the same time such that both of them interleave which is not problematic on the surface as they don't share the log but it is problematic because eventually they all deal with the same S3 objects. Imagine, (1) Upload a file to GLACIER (2) Issue a restore-object (3) Cron schedules tape recall as well as tape migrate (4) a file gets "recalled" but then is migrated.
A single instance of migrate is running (probably because WAL was huge and eeadm is slow) while another one got scheduled. Same for recall.
NOTE: This solution does NOT allow running tape recall and tape migrate at the same time. It is intentional. One call will block another.
[ ] Doc added/updated
[ ] Tests added
Ah, there is still one possible race here:
P1 opens a file
P2 renames the file
Someone issues a tape migrate which gets exclusive access to the recently renamed file
P1 gets blocked as tape migrate as the exclusive access
tape migrate unlinks the file on completion
P1 gets unblocked, happily performs write (the file won't be deleted as there is an open FD).
P1 closes the FD upon swap, the file gets deleted
RESULT: Lost writes.
Working on it now...
|
2025-04-01T06:39:47.660373
| 2024-07-25T09:36:56
|
2429502384
|
{
"authors": [
"IanOvenden",
"akshaykalaskar1"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8996",
"repo": "norm-l/odx",
"url": "https://github.com/norm-l/odx/pull/641"
}
|
gharchive/pull-request
|
fix: BUG-8180/INC-B17940
Updated navigation to 'replace' so that ?code entry which in history with the intended new url without ?code
Hi @akshaykalaskar1,
So as I understand it, replace forces the page to reload and therefore running the process to remove code from the URL and load correctly. Is that right?
|
2025-04-01T06:39:47.681752
| 2016-04-07T15:54:54
|
146667484
|
{
"authors": [
"Khady",
"noseglid"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8997",
"repo": "noseglid/atom-build",
"url": "https://github.com/noseglid/atom-build/issues/384"
}
|
gharchive/issue
|
Append custom regex to errorMatch for a project
Hi,
I am using a Makefile to compile an ocaml project. This makefile actually calls ocamlbuild, which is a wrapper around ocamlc/ocamlopt (the ocaml compilers).
The errors that are provided by those compilers follow this schema:
File "foo/bar.ml", line 277, characters 12-13:
Warning 26: unused variable g.
File "foo/bar.ml", line 280, characters 6-64:
Warning 34: unused type t.
File "foo/bar.ml", line 279, characters 0-6:
Error: Syntax error
I have a regex to match the warnings and errors:
(?<file>[\\/0-9a-zA-Z\\._\\-]+)", line (?<line>\\d+), characters (?<col>\\d+)-(?<col_end>\\d+):\\n(?<message>.+)
For the moment, I have a fork of build-make in which I added my regex to the errorMatch array.
I wonder if there is a way with .atom-build.{js,json,cson,yml} to specify my regex rather than to modify build-make?
I'm thinking about using preBuild to append my regex to the existing array, but I'm not sure this is the clean way to do it. Also, I would have to create a custom build command for all the targets in my makefile. This is not really convenient, I would like to use this regex for all the targets.
Or should I create a build provider for ocamlc/ocamlopt and even if I launch a make task, the regex from my provider will be applied? As my workflow (Makefile + ocamlopt) is common in the ocaml world, I would be happy to provide a solution that works for everyone.
I have the problem for ocaml now, but will probably have the same one with C/C++ or other languages/tools later. I am curious to know what is the best solution.
Also, subsidiary questions:
is there a way to display the warnings and the errors in a different way? (I suspect it is not possible because of this line)
can I access col and col_end from postBuild? I'd like to modify their value.
Thanks
Can you not make a PR to build-make with your regex?
Sure, I will do it today.
Oh, also saw your other questions now:
There is no way to differentiate between errors and warnings right now. I wouldn't mind a way of doing this, I just don't know the best way yet.
You cannot access col or col_end from postBuild, not change them either. Before enabling things like this I want to verify that it makes sense in all (at least most) scenarios. I haven't thought too much about what one might want to achieve in postBuild, it's a quite new feature.
Ok, thanks for the answers and for your great tool :)
|
2025-04-01T06:39:47.683514
| 2018-05-23T09:28:12
|
325622538
|
{
"authors": [
"nosir",
"viktor-izettle"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8998",
"repo": "nosir/cleave.js",
"url": "https://github.com/nosir/cleave.js/issues/353"
}
|
gharchive/issue
|
Preact compatibility
Would you be open for a PR to make it compatible with preact?
Just checking..
Unfortunately we have limited resource to make it work and keep update with all those awesome libs, so this is not in the plan yet.
Also sorry for the late response.
Close it for now.
|
2025-04-01T06:39:47.689703
| 2021-05-26T21:33:17
|
902939635
|
{
"authors": [
"FoxxMD"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:8999",
"repo": "not-an-aardvark/snoowrap",
"url": "https://github.com/not-an-aardvark/snoowrap/issues/324"
}
|
gharchive/issue
|
How to check what moderator permissions client has?
I can see what permissions I (an account) has by GET to https://www.reddit.com/subreddits/mine/moderator/.json and looking at the property mod_permissions on each subreddit.
I don't see this available through getMe() or getModeratedSubreddits() in snoowrap. And the only place permissions are mentioned in typings is for inviteModerator() and setModeratorPermissions() as arguments.
How, using snoostorm, can I check that the authenticated account has permission to perform mod actions on a subreddit other than just trying and seeing if i get a 400/401/403 back?
It is done like this:
// options to getModerators is optional but will ensure index 0 of the returned array is the user you are looking for
const mods = snoowrap.getSubreddit('mySubreddit').getModerators({name: 'modName'});
const modPermissions = mods[0].mod_permissions;
|
2025-04-01T06:39:48.029590
| 2023-10-12T04:08:20
|
1939137735
|
{
"authors": [
"broofa"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9000",
"repo": "npmgraph/npmgraph",
"url": "https://github.com/npmgraph/npmgraph/pull/170"
}
|
gharchive/pull-request
|
feat: last-modified date colorizer
Adds ability to colorize by "last modified date".
Example: https://npmgraph-git-datecolorizer-npmgraph.vercel.app/?q=express#color=modified
I'm a little unsure whether to merge this. The "Last modified date" is last time the package record was modified in the NPM registry. That'll generally be the date of the last version that was published, but could also be something unrelated (e.g. when a version was unpublished, or an owner changed or... whatever.) I'm concerned this could be more confusing or misleading than actually helpful.
Thoughts?
Closing unmerged. This is going to be more confusing than it will be useful, I think.
|
2025-04-01T06:39:48.031481
| 2016-06-22T14:33:38
|
161699581
|
{
"authors": [
"TylerFisher",
"lindamood"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9001",
"repo": "nprapps/rockymountain",
"url": "https://github.com/nprapps/rockymountain/issues/100"
}
|
gharchive/issue
|
Bug: audio player and web vr icon aren't showing up on mobile by default
rotating the device causes the controls to show up.
Fixed.
|
2025-04-01T06:39:48.038923
| 2017-09-13T06:59:22
|
257276901
|
{
"authors": [
"jwatt",
"nrc",
"skade"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9002",
"repo": "nrc/find-work",
"url": "https://github.com/nrc/find-work/issues/8"
}
|
gharchive/issue
|
URL without www loads page, but has a CORS error for the data.
I.e., visiting https://rustaceans.org/findwork rather than https://www.rustaceans.org/findwork (and presumably the http versions) loads the website, but is stuck at loading... forever becaue fetching the data gives a CORS error.
Similarly, navigating to the http site (http://rustaceans.org/findwork) will lead you somewhere else entirely.
https://github.com/aturon/rfcs/blob/roadmap-2018/text/0000-roadmap-2018.md links to the broken page...
|
2025-04-01T06:39:48.044925
| 2015-06-18T22:12:15
|
89409950
|
{
"authors": [
"cassiersg",
"marcusklaas",
"nrc"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9003",
"repo": "nrc/rustfmt",
"url": "https://github.com/nrc/rustfmt/issues/110"
}
|
gharchive/issue
|
Rustfmt always recompiles
Rustfmt always recompiles on my machine, even if no source code has been altered. Is this happening for other people as well? If so, why does it happen and how can we prevent it?
It's also happening for me. It seems to be related to the build script. Commenting the line build=build.rs prevent the rebuild when nothing has changed. (I don't know why ?)
Interesting. Maybe it updates the last edit timestamp on default.toml in the root directory, which could trick Cargo into believing code may have changed?
Commenting the line std::fs::copy(in_file, out_file).unwrap(); doesn't change anything. It's maybe a default behavior of Cargo to rebuild everything when there is a build script... needs more investigation.
Weird -- commenting out that line does the trick here...
The source of trouble was a symlink rustfmt/rustfmt -> rustfmt/target/debug/rustfmt. Cargo seems to consider all files under rustfmt/ to decide when to rebuild.
Did you create this symlink by hand?
Unfortunately, https://github.com/nrc/rustfmt/pull/111 does not prevent recompilation one when of the test files in /tests/source/ or /tests/target/ changes... Any ideas on how we could fix that?
Yes, I created the symlink by hand (for tests, because cargo run recompiled always).
I don't know how to avoid rebuild when a non-source file is touched, but I have maybe an explaination why it occurs: the build script is intended to build external dependencies, but cargo doesn't which files the script builds, so it takes a conservative approach : rebuild if any file changed. If you remove the build script from Cargo.toml, a change in test files doesn't trigger rebuild anymore.
I think the good thing would be to remove the build script, and use something like a deploy script, but it doesn't seem to exist (yet ?). There is maybe another way : cargo is made to recieve build commands from the stdout of the build script, there are maybe commands to indicate dependencies (But at first sight there isn't anything like this in the docs.)
Removing the build script sounds good to me. Maybe we can add the default.toml to the root directory in the repository and also add it to .gitignore, so that editing it will not register as a change to the repository.
Might be worth asking on #cargo about this, seems like there ought to be a solution.
It's a known issue: https://github.com/rust-lang/cargo/issues/1162. No known workarounds at the moment. Maybe it's a nice issue to work on.
Build script was removed by #165.
|
2025-04-01T06:39:48.046709
| 2018-09-27T04:45:34
|
364303326
|
{
"authors": [
"Mierdin"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9004",
"repo": "nre-learning/antidote",
"url": "https://github.com/nre-learning/antidote/issues/41"
}
|
gharchive/issue
|
Reviewers Doc
Need a doc for reviewers. Not only syntax but the spirit of the project. Focusing on the workflow rather than just "automating the network". Being neat with diagrams and examples. Not only for reviewers but also so contributors know what to expect. Convey that our goal is to make this the first impression for automators, so it has to preserve that.
Closing in lieu of #17
|
2025-04-01T06:39:48.058231
| 2021-02-12T15:54:30
|
807344772
|
{
"authors": [
"EssBee59",
"bagage",
"mjaschen",
"nrenner"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9005",
"repo": "nrenner/brouter-web",
"url": "https://github.com/nrenner/brouter-web/issues/368"
}
|
gharchive/issue
|
Is it possible to get help to create a new pull request?
Hello,
I would like to create new pull requests (for #356 and #332), but I am not familar with the branch/fork and pull technologies....
Is it possible to get help? (or better, someone creates the pull reqests directly!)
My changes are already well tested, a documentation is also availlable.
Regards
Ess Bee
Hi!
did you make your changes on a cloned Git repository of brouter-web?
Hi @EssBee59, the procedure should be:
Click on Fork button on https://github.com/nrenner/brouter-web/ page:
Then clone your repository locally with: git clone git@github.com:EssBee59/brouter-web.git
Modify the code as you wish.
Commit with it with a brief explanation of the changes: git commit -am "Improve voice hint generation. Fix #356" for instance.
Push it: git push
Create a pull request by opening https://github.com/nrenner/brouter-web/compare/master...EssBee59:master
And voilà!
As I am not familiar with repositories I have a local installation on my PC.
So I make the changes locally, and test locally first.
When the tests are succesfull, I install the "dist" on an instance of brouter.de (brouter.de/essbee)
The changes are documented in both issues, by need I can deliver more documentation
Thank for your help and regards
Ess Bee
Maybe it's easiest to create an archive of your local BRouter installation and provide it as download or send it as mail attachment so that someone else can integrate the patches and create the pull-requests. I've some time tomorrow and would take a look if it's ok with you.
mjaschen,
Thank for your proposal, that I liked to accept!!
Please send me a mail addrees where I can send my changes
Reagrds
Ess bee
<EMAIL_ADDRESS>should work :-)
Thank´s all for your help!
Some more information:
GitHub used to have nice, short step-by-step guides that I can't find anymore. They seem to have rewritten the docs, and now I fear this is too much for starting:
Collaborating with issues and pull requests - GitHub Docs
From a quick search this seems to be a nice guide:
How to make your first pull request on GitHub
There is also GitHub Desktop if you prefer using a UI over command line (Windows and MacOS, also as Linux Fork):
Setting up GitHub Desktop and Git
Cloning and forking repositories from GitHub Desktop
Committing and reviewing changes to your project
Pushing changes to GitHub
Creating an issue or pull request (missing!?)
|
2025-04-01T06:39:48.110895
| 2024-05-07T22:09:56
|
2284309601
|
{
"authors": [
"markaj-nordic"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9006",
"repo": "nrfconnect/sdk-zephyr",
"url": "https://github.com/nrfconnect/sdk-zephyr/pull/1686"
}
|
gharchive/pull-request
|
[nrf fromtree] net: if: Extend the usage of rejoining the multicast groups
It may happen that the interface is up but not yet running when we issue the rejoin_ipv6_mcast_groups(). This can be fixed by calling this function again right after the iface is set to 'running' state in the notify_iface_up handler.
Signed-off-by: Marcin Kajor<EMAIL_ADDRESS>(cherry picked from commit b571e45d80923a786979460f218618e2000da8f3)
Green light on the corresponding sdk-nrf PR. We can merge this one I guess.
|
2025-04-01T06:39:48.121775
| 2022-07-22T11:30:06
|
1314889357
|
{
"authors": [
"codecov-commenter",
"jstarpl"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9007",
"repo": "nrkno/sofie-core",
"url": "https://github.com/nrkno/sofie-core/pull/748"
}
|
gharchive/pull-request
|
hotfix: add try-catch in setTimeout/setImmediate
What kind of change does this PR introduce? (Bug fix, feature, docs update, ...)
What is the current behavior? (You can also link to an open issue here)
What is the new behavior (if this is a feature change)?
Other information:
Status
[ ] Code documentation for the relevant parts in the code have been added/updated by the PR author
[ ] The functionality has been tested by the PR author
[ ] The functionality has been tested by NRK
Codecov Report
Merging #748 (379e4b3) into release41 (c80781a) will increase coverage by 0.02%.
The diff coverage is 16.66%.
@@ Coverage Diff @@
## release41 #748 +/- ##
=============================================
+ Coverage 69.86% 69.88% +0.02%
=============================================
Files 302 302
Lines 34428 34434 +6
Branches 4662 4664 +2
=============================================
+ Hits 24052 24065 +13
+ Misses 9925 9917 -8
- Partials 451 452 +1
Impacted Files
Coverage Δ
meteor/server/api/userActions.ts
20.97% <0.00%> (-0.28%)
:arrow_down:
meteor/lib/api/userActions.ts
100.00% <100.00%> (ø)
packages/job-worker/src/ingest/rundownInput.ts
78.33% <0.00%> (+2.07%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update c80781a...379e4b3. Read the comment docs.
|
2025-04-01T06:39:48.185103
| 2023-03-02T13:24:26
|
1606804457
|
{
"authors": [
"mechairoi",
"nsaunders"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9008",
"repo": "nsaunders/purescript-unique",
"url": "https://github.com/nsaunders/purescript-unique/issues/3"
}
|
gharchive/issue
|
Repository transfer
Hi @mechairoi, I got this updated in the PureScript Registry quicker than expected. :-)
At this point, I would like to migrate the repository to a different namespace, e.g. purescript-contrib, rowtype-yoga, or similar. This can distribute the maintenance burden and avoid either of us being a "single point of failure".
In the interim, though, would you please transfer this repository to my username?
Thanks again!
Thanks @nsaunders. I'm grad to hear that.
I have requested to transfer the repository to you. I think it's a good idea to migrate to a namespace like those.
Thanks again for your contribution!
Thank you @mechairoi!
|
2025-04-01T06:39:48.200791
| 2015-01-13T12:10:00
|
54184849
|
{
"authors": [
"nskvortsov"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9009",
"repo": "nskvortsov/deployer",
"url": "https://github.com/nskvortsov/deployer/issues/23"
}
|
gharchive/issue
|
SSH Deployer key based authentication not working
First things first: Thanks for this great plugin!
There seems to be a bug with SSH Deployer authentication when using SSH keys. Password authentication works fine, however, when selecting "Default Private Key" I can not enter a username with which to authenticate with the server (even if using a default private key, it's still mandatory for authentication to specify the user that should be used on the remote host).
It is also not possible to circumvent that problem by selecting "Custom Private Key" (with this it is possible to specify a user name) as the path to the "custom identity file" is incorrectly resolved. It always tries to find the path one specifies relative the to buildagent work dir instead of the home dir and it is not possible to specify an absolute path either as that always just get's appended to the current working dir. Hence, specifying something like ~/.ssh/id_rsa does not work. Using %agent.home.dir% doesn't work either (appends the home dir to the temp dir).
Bitbucket: https://bitbucket.org/nskvortsov/deployer/issue/23
Originally reported by: Sebastian Siemssen
Originally created at: 2014-01-07T14:29:06.547
fixed in this build
allow to enter username for default key
allow absolute paths to custom key
Original comment by: Nikita Skvortsov
|
2025-04-01T06:39:48.231973
| 2015-06-29T07:56:42
|
91729259
|
{
"authors": [
"boazin",
"nst"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9010",
"repo": "nst/STTwitter",
"url": "https://github.com/nst/STTwitter/issues/199"
}
|
gharchive/issue
|
verifyCredentialsWithSuccessBlock not returning
So here goes - I had an old code of STTwitter not via cocoapods.
Using pods now.
The problem:
verifyCredentialsWithSuccessBlock not returning - not success block and not errorBlock.
It started with the latest version (0.2.1) so I thought it has something to do with the deprecation of the function.
Reverted to 0.2.0 (with patching of the extern to get it to link)
Still same thing.
When a user already logged-in in the past I'm saving it in UserDefaults and retrying with it with the following code.
Any assistance on investigating will be appreciated
NSUserDefaults *prefs = [NSUserDefaults standardUserDefaults];
NSString *oauthAccessToken = [prefs stringForKey:@"twitterAccessToken"];
NSString *oauthAccessTokenSecret = [prefs stringForKey:@"twitterAccessTokenSecret"];
NSString *key = [[NSBundle mainBundle] objectForInfoDictionaryKey:@"twitterKey"];
NSString *secret = [[NSBundle mainBundle] objectForInfoDictionaryKey:@"twitterSecret"];
if (oauthAccessToken != nil) //we are logged in
{
STTwitterAPI *twitter = [STTwitterAPI twitterAPIWithOAuthConsumerKey:key consumerSecret:secret oauthToken:oauthAccessToken oauthTokenSecret:oauthAccessTokenSecret];
[twitter verifyCredentialsWithSuccessBlock:
^(NSString *username) {
//great... do stuff and login to the app
} errorBlock:^(NSError *error) {
NSLog(@"error verifying credentials: %@", error);
[self gotoLoginScreen];
}];
}
I updated the code, still not working.
From what I was able to dig in the code:
successBlock(username, userID); (ln 247 - in the success block of verifyCredentialsRemotelyWithSuccessBlock)
Is getting called twice, with nil in username, and userID,
And only afterwards the GET request from twitter returns (ln 246 in STTwitterOAuth.m) and nothing is done with the result
Not sure why :(
I'm confused. The following code works for me with the latest version of STTwitter. Can you double check that you're also using it? (by typing git pull)
NSString *oauthAccessToken = @"";
NSString *oauthAccessTokenSecret = @"";
NSString *key = @"";
NSString *secret = @"";
STTwitterAPI *twitter = [STTwitterAPI twitterAPIWithOAuthConsumerKey:key
consumerSecret:secret
oauthToken:oauthAccessToken
oauthTokenSecret:oauthAccessTokenSecret];
[twitter verifyCredentialsWithSuccessBlock:^(NSString *username) {
NSLog(@"-- username: %@", username);
} errorBlock:^(NSError *error) {
NSLog(@"-- error: %@", error);
}];
If it still doesn't work for you, can you please setup a minimal Xcode project, so that I'll be able to reproduce and fix the bug? Thank you for your help.
sorry for the noob question - how can I git pull if I'm using pod?
maybe you can git pull somewhere else, replace your curren STTwitter directory with the new one and run the project again
P.S. you can get a pod from GIT by doing:
pod 'STTwitter', :git => 'https://github.com/nst/STTwitter.git'
Anyway - a small project works, so it is not the code itself... I'll look for differences between the projects.
Do you think it is related to the fact that STTwitter used to be code in my project? (leftover frameworks?)
Well... after commenting out the verifyCredentialsLocallyWithSuccessBlock to see if it works and figuring out it is still gets executed I just removed all pods and podCache - reinstalled, and now everything works....
Well, thanks again! I believe we found a real bug in the process so it wasn't all for nothing...
Glad to hear that.
Don't hesitate to report any other issue you may find in the future.
|
2025-04-01T06:39:48.236933
| 2017-11-09T20:34:41
|
272715174
|
{
"authors": [
"mbuhot",
"nsweeting"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9011",
"repo": "nsweeting/authex",
"url": "https://github.com/nsweeting/authex/issues/1"
}
|
gharchive/issue
|
License
Thanks for creating this library, it looks nice and simple to use 👍
I notice the mix.exs file says the license is MIT, but I don't see a LICENSE file in the repo.
Could you add one just to clarify.
Thanks again!
Thanks for the heads up @mbuhot. Missing LICENSE has been added. Will be available in the next hex release.
|
2025-04-01T06:39:48.239286
| 2024-10-18T21:38:16
|
2598427125
|
{
"authors": [
"Abestanis",
"nt4f04uNd"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9012",
"repo": "nt4f04uNd/sweyer",
"url": "https://github.com/nt4f04uNd/sweyer/pull/161"
}
|
gharchive/pull-request
|
Add cyrillic letters to sort tests
Fixes https://github.com/nt4f04uNd/sweyer/issues/158
Thanks for adding the tests.
Thanks for making them =)
I've seen this CI failure before, but I can't replicate it locally. My guess is that we start the initialization in a test, and then the test ends but the while the initialization is waiting on an await point and we start a new test overwriting the queue controller. For now I think you can just re-run the test, but a real solution it to either find a way to cancel / await all un-awaited futures started or to not have singletons and instead pass in all controllers when creating the app.
Created an issue https://github.com/nt4f04uNd/sweyer/issues/162
|
2025-04-01T06:39:48.245510
| 2016-03-28T20:56:52
|
144081168
|
{
"authors": [
"lgeiger",
"n-riesco"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9013",
"repo": "nteract/hydrogen",
"url": "https://github.com/nteract/hydrogen/pull/232"
}
|
gharchive/pull-request
|
ResultView: Add action panel to text/html output
Closes #26
Just make the pre element clickable if it's a single line output.
I see a bug but I can't reproduce it reliably.
When it fails:
When it works:
The logs in the dev console look the same.
I will look at this more carefully. But I'm not sure if I'll be able to look at this tomorrow.
Thanks for taking a look.
I can't reproduce this. I always get the second output no matter how many times I try.
It happens more often when I use C-A-enter.
Strangely enough this doesn't happen on my machine.
Does it happen when you add back the previousText check at line 111 like this?
previousText = @getAllText()
container.appendChild htmlElement
if mimeType is 'text/plain'
text = @getAllText()
if previousText is '' and text.length < 50 and text.indexOf('\n')
@lgeiger I that was it. After adding the check, I don't it to fail.
It should work now.
I think it's easier for the user to be able to click anywhere container.
Another problem with the pre hack is that it breaks when a block with multiple stdouts is executed, e.g.:
print('A plot')
plt.plot(t, s)
OK I'll revert my changes except those directly related to #26.
All right, back to the basics :wink:
LGTM
|
2025-04-01T06:39:48.252590
| 2024-12-05T21:48:35
|
2721500635
|
{
"authors": [
"florianziemen",
"ntessore",
"philipc2"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9014",
"repo": "ntessore/healpix",
"url": "https://github.com/ntessore/healpix/issues/66"
}
|
gharchive/issue
|
Provide cell corners
Moin Nicolas,
First of all, thanks a lot for providing the healpix package!
To add HEALPix support to UXarray, an xarray extension for unstructured grids, we need a function that computes the vertices, and ideally also a mapping of the vertices to the cells.
I've played a bit with the healpix package, and found that the _uv functions will provide the corners if supplied with the correct arguments, as in this gist. From there, one can move on with unique functions as in this gist by @philipc2.
I'm sure this is not the most efficient way of generating the vertices, but I was wondering if you would be interested in supporting this operation in your library. I'd be more than happy to create a merge request if that helps.
Pinging @philipc2 and @erogluorhan, as they are core developers of UXarray.
Cheers
Flo
Hi Flo!
Thanks for getting in touch. It would be great if we added add a healpy-compatible function such as
def boundaries(nside, ipix, step=1, nest=False):
...
to the code. In the first instance, I think that using the pix2ang_uv functions at the Python layer, as in your examples, is perfectly fine. In a second step, it could make sense to add the boundaries() function directly to the C library. That might save a bit of overhead -- but it would anyway be good to have the Python functions in place to make the comparison.
How does that sound?
Thanks for pinging me @florianziemen!
We are excited to use this package to add support for HEALPix within UXarray. To add onto what Flo already mentioned:
UXarray is written around the UGRID conventions, which requires at least the vertices (node_lon, node_lat) and cell- vertex indicies(face_node_connectivity) to represent an arbitrary 2D unstructured grid. This would be for loading a HEALPix grid into UXarray.
Going the other way, converting from UGRID to HEALPix can be achieved using a nearest neighbor remap, however for conservative remapping, knowing the boundaries is important
Hi,
sorry for the late reply. I've been at the AGU conference and took a couple days off afterwards. I've looked into the boundaries function. In principle, I know what to do to replicate it. I'm not sure, we want to mimic exactly this function, as it kind of assumes a single point going in, and only goes to xyz (if I remember correctly).
Basic thought would be to provide the behavior of boundaries with a signature that's matching the original call for compatibility, and provide a second function, that produces angles in (theta/phi) and (lon/lat) notation.
I won't manage to do this before the holidays, but it's high on my todo list for early next year.
Cheers
Flo
|
2025-04-01T06:39:48.279506
| 2023-05-05T13:58:54
|
1697671314
|
{
"authors": [
"ancker010",
"massimocandela"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9015",
"repo": "nttgin/BGPalerter",
"url": "https://github.com/nttgin/BGPalerter/issues/1082"
}
|
gharchive/issue
|
Speed up generate function
Describe what you would like to achieve
Provide a mechanism to speed up generation of large numbers of prefixes or ASNs.
Describe why the current solution (if any) is not satisfactory
When passing a large number of ASNs or Prefixes to the generate function, the process takes several hours.
Provide an example
time ./bin/bgpalerter-macos-x64 generate -m -l .output/prefixes.txt -o output/prefixes-test.yml
real 245m27.073s
user 8m16.969s
sys 1m10.959s
Your information
Ryan Harden: AS11537
Hi @ancker010,
What does prefixes.txt contains and how many ASes are you monitoring?
prefixes.txt is a list of prefixes, one per line, currently 6887.
And approximately 1045 ASNs.
|
2025-04-01T06:39:48.301601
| 2017-09-21T05:53:46
|
259382873
|
{
"authors": [
"nuclearace",
"tellowkrinkle"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9016",
"repo": "nuclearace/SwiftDiscord",
"url": "https://github.com/nuclearace/SwiftDiscord/pull/51"
}
|
gharchive/pull-request
|
Switch to Encodable for JSONification
Now that Swift 4 is out, we have Codable! Yaaaay!
Except that it wasn't meant for situations where you use some structs and some random dictionaries. It really doesn't like trying to encode [String: Any]s or even [String: Encodable]s (trying to encode either with the default implementation throws a FatalError saying fatal error: Array<Encodable> does not conform to Encodable because Encodable does not conform to itself. You must use a concrete type to encode or decode.)
Is the new implementation hackish? A little. But I think it's a little less hackish than the one it's replacing. Also, we can finally follow the Swift API Guidelines on acronyms (I'm talking about you, Id and Url) without an overly complicated snakecase function (though that's not implemented here).
Yeah whenever conditional conformance arrive it should get rid of the hacky things we have to here since it'll let them handle these in the stdlib
@tellowkrinkle Have you toyed around with trying bring deserialization to the holding types? It would be fantastic if we could get rid of all that crufty stuff we have now.
Although thinking about it more I guess conditional conformance really won't help with the case where it's just Any... since it can't know what to encode. And toying around it doesn't look like you define a custom protocol that just declares Codable conformance to use, which is kind of a bummer.
I haven't looked into deserialization much yet. I do know that Codable doesn't support autogenerating with defaults, so we'll still need to manually implement init(from decoder:).
|
2025-04-01T06:39:48.335889
| 2024-01-02T18:02:43
|
2062764650
|
{
"authors": [
"nobkd",
"tipiirai"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9017",
"repo": "nuejs/nue",
"url": "https://github.com/nuejs/nue/pull/147"
}
|
gharchive/pull-request
|
fix: add type to favicon field
resolve #146
Adds the image type to the favicon field
Thank you!
|
2025-04-01T06:39:48.337213
| 2024-09-01T13:45:59
|
2499479263
|
{
"authors": [
"nobkd",
"tipiirai"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9018",
"repo": "nuejs/nue",
"url": "https://github.com/nuejs/nue/pull/338"
}
|
gharchive/pull-request
|
chore: exclude examples and docs from triggering test run
https://docs.github.com/en/actions/writing-workflows/workflow-syntax-for-github-actions#onpushpull_requestpull_request_targetpathspaths-ignore
Good change
|
2025-04-01T06:39:48.380373
| 2022-12-27T23:20:08
|
1512254002
|
{
"authors": [
"nullxception",
"rinme"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9026",
"repo": "nullxception/boorusphere",
"url": "https://github.com/nullxception/boorusphere/pull/69"
}
|
gharchive/pull-request
|
Add: data_backup, ime_incognito to Thai's i18n
Added "data_backup" and "ime_incognito" Translations to Thai's i18n
69th commit, Nice :)
In this pull request
Localization
Add: data_backup, ime_incognito to Thai's i18n
Fully Translated Japanese
Thank you!
|
2025-04-01T06:39:48.383055
| 2024-07-05T13:50:03
|
2392650512
|
{
"authors": [
"th0ger",
"vigith"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9027",
"repo": "numaproj/numaflow",
"url": "https://github.com/numaproj/numaflow/issues/1794"
}
|
gharchive/issue
|
Component Testing for Python SDK
Summary
What change needs making?
Use Cases
When would you use this?
Message from the maintainers:
If you wish to see this enhancement implemented please add a 👍 reaction to this issue! We often sort issues this way to know what to prioritize.
Yeah we need some better tooling and developer experience on testing.
But it's hard to know what's in the making when only issue titles are filled.
Please include community so we can help you grow. 🙂
Indeed very hard to tell from "test" labels whether is related to numaflow core tests (go/sdk), or test framework for users (pipelines/udf).
We usually tag area/sdk when it is SDK related.
|
2025-04-01T06:39:48.387016
| 2022-05-29T19:04:00
|
1251966632
|
{
"authors": [
"edlee2121",
"vigith"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9028",
"repo": "numaproj/numaflow",
"url": "https://github.com/numaproj/numaflow/issues/30"
}
|
gharchive/issue
|
AWS SQS source and sink
Is your feature request related to a problem? Please describe.
Would be useful to be able to source events from AWS SQS.
Describe the solution you'd like
A source that turns GCP Pub/Sub events into messages for further processing.
Describe alternatives you've considered
None
Additional context
None
Message from the maintainers:
If you wish to see this enhancement implemented please add a 👍 reaction to this issue! We often sort issues this way to know what to prioritize.
@shubhamdixit863 once you complete the issues listed in https://github.com/numaproj-contrib/aws-sqs-source-go/issues, please go ahead and close this.
Completed! SQS Source and SQS Sink
|
2025-04-01T06:39:48.391168
| 2021-05-04T12:29:35
|
875413156
|
{
"authors": [
"JSKenyon",
"stuartarchibald",
"vroomzel"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9029",
"repo": "numba/numba",
"url": "https://github.com/numba/numba/issues/6998"
}
|
gharchive/issue
|
np.linalg.solve unexpectedly returns an F_CONTIGUOUS array for C_CONTIGUOUS inputs
Reporting a bug
[X] I have tried using the latest released version of Numba (most recent is
visible in the change log (https://github.com/numba/numba/blob/master/CHANGE_LOG).
[X] I have included a self contained code sample to reproduce the problem.
i.e. it's possible to run as 'python bug.py'.
Reproducer
import numpy as np
from numba import njit
def f(x, y):
return np.linalg.solve(x, y)
if __name__ == "__main__":
x = np.array([[1, 2], [3, 5]], dtype=np.float64)
y = np.eye(2, dtype=np.float64)
g = njit()(f)
numpy_result = f(x, y)
numba_result = g(x, y)
print(numpy_result.flags)
print(numba_result.flags)
This prints the following:
C_CONTIGUOUS : True
F_CONTIGUOUS : False
OWNDATA : True
WRITEABLE : True
ALIGNED : True
WRITEBACKIFCOPY : False
UPDATEIFCOPY : False
C_CONTIGUOUS : False
F_CONTIGUOUS : True
OWNDATA : False
WRITEABLE : True
ALIGNED : True
WRITEBACKIFCOPY : False
UPDATEIFCOPY : False
I am not sure if this is the expected/desired behaviour. I encountered this when attempting to feed the results of np.linalg.solve into np.dot and saw:
NumbaPerformanceWarning: np.dot() is faster on contiguous arrays, called on (array(float64, 2d, A), array(float64, 1d, C))
Thanks for the report. Given what Numba's implementation of np.linalg.solve does this is not surprising, however it is a bug as it ought to match NumPy.
I actually encountered the same issue, but with np.linalg.qr function. Are there any workarounds for this?
|
2025-04-01T06:39:48.392453
| 2015-08-31T20:52:41
|
104132275
|
{
"authors": [
"pitrou",
"stefanseefeld"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9030",
"repo": "numba/numba",
"url": "https://github.com/numba/numba/pull/1401"
}
|
gharchive/pull-request
|
Fix typo(s).
For most cases in typeinfer.py, the word 'constrain' should actually be 'constraint'.
Yes, I've always been a bit confused by that :)
|
2025-04-01T06:39:48.401204
| 2015-12-09T11:44:53
|
121218796
|
{
"authors": [
"christianp"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9031",
"repo": "numbas/Numbas",
"url": "https://github.com/numbas/Numbas/issues/419"
}
|
gharchive/issue
|
Require student to specify units in number entry parts
There should be an option to make the student specify the units of their answer. For example, make them type "50g" instead of just "50".
Options:
name of unit (g,s,N, etc.)
allow SI orders of magnitude? (i.e., student can enter either "5000g" or "5kg" and be marked correctly)
which alternate names do you allow? (e.g. just "kg", or "kilos" or "kilograms" or "kilogrammes")
Will have to hardcode most units - need to know all the names for different SI units. Or, have presets that can be modified.
Maybe give a list of units and a scaling factor. For example:
(g or grammes or grams) * 1
(kg or kilogrammes or kilograms or kilos) / 1000
(mg or milligrammes or milligrams) * 1000
Some unit systems are more complicated!
Currencies go in front: "£3.50", except for
Geographical coordinates have several components: "66°33’ N", 1° = 60’ = 3600”.
Should extend the TNum type to have a "units" property, describing the units the number was given in.
While doing this, maybe we should also deal with different decimal notations - see https://en.wikipedia.org/wiki/Decimal_mark#Examples_of_use
If we hardcode the units, need to make it easy to add more for international users or new subjects.
How complicated do we want to get with combinations of units? Need to do at least things like "m/s ms^-1" and "m/s^2 = ms^-2". Can all units be expressed as a product or ratio of other units?
See http://www.boost.org/doc/libs/1_37_0/doc/html/boost_units/Dimensional_Analysis.html for an existing implementation.
Fields to add to the number entry part:
expected units (string, not necessarily directly shown to the student - complicated schemes like geographical coords will need a name like "latitude", which tells the part how to format the expected answer)
allow different orders of magnitude? (i.e., allow "kg" as well as "g")
partial credit if units not given, and a feedback message
allow alternate names? (i.e., allow "grams" as well as "g", or "l" as well as "cm^3") - maybe this should be a list of allowed units
Another point: if we're adding dimension information to TNum, should addition and subtraction fail when attempted on incompatible measurements, and how should they be combined for multiplication and division?
Rink is an existing calculator which might work as a reference https://github.com/tiffany352/rink-rs/wiki/Rink-Manual
The % symbol behaves a lot like a units symbol.
JS-quantities looks like a decent JS unit-handling library.
There's now a quantities extension.
|
2025-04-01T06:39:48.420843
| 2024-04-23T22:25:59
|
2259877943
|
{
"authors": [
"gwhitney"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9032",
"repo": "numberscope/frontscope",
"url": "https://github.com/numberscope/frontscope/issues/314"
}
|
gharchive/issue
|
Use multiple columns or something like that for contributor list
The contributor list is getting long. That's a good thing, but presented as a single-column list it's leaving a lot of whitespace on the About page. I recommend going to multiple columns, or just a comma-separated paragraph format, or something else space-efficient, if it grows much more. Certainly if each member of the Delft team is added individually to the list, that wold be the time to fix that formatting.
Should be incorporated in overhaul of docs for alpha.
Fixed in ui2 per #464, closing.
|
2025-04-01T06:39:48.554133
| 2023-04-02T01:22:10
|
1650764767
|
{
"authors": [
"nunomaduro",
"set0x"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9033",
"repo": "nunomaduro/collision",
"url": "https://github.com/nunomaduro/collision/pull/268"
}
|
gharchive/pull-request
|
Return testdox as test name
Return name generated for phpunit from testDox. This add possibility to rename test name with @testdox annotation.
No plans for changing this, sorry.
Hi, I currently use @testdox to describe my tests as it gives me a clearer interface. Is there any way to display it with the art test command. Currently if I set it as parameter art test --testdox the coverage functions do not work.
Example
art test --testdox --coverage --min=100.
This example does not work.
With the change I proposed it would work exactly the same for the current test suite and also the @testdox annotation would work.
How could I fix this?
Thanks.
Translated with www.DeepL.com/Translator (free version)
|
2025-04-01T06:39:48.648200
| 2024-11-07T06:36:50
|
2640024151
|
{
"authors": [
"blindFS",
"fdncred"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9034",
"repo": "nushell/tree-sitter-nu",
"url": "https://github.com/nushell/tree-sitter-nu/pull/134"
}
|
gharchive/pull-request
|
Fix unquoted command args with leading +
This PR fixes #131 by grouping "+/-" and digits in decimals.
Related test cases in cmd-026-unquoted-string-with-leading-plus.
Thanks!
|
2025-04-01T06:39:48.662638
| 2018-12-11T06:52:38
|
389620176
|
{
"authors": [
"CLAassistant",
"mcedica"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9035",
"repo": "nuxeo-sandbox/environment-jx-staging",
"url": "https://github.com/nuxeo-sandbox/environment-jx-staging/pull/8"
}
|
gharchive/pull-request
|
jenkinsx-quickstart-nuxeo-poc to 1.0.11
Promote jenkinsx-quickstart-nuxeo-poc to version 1.0.11
Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.jenkins-x-bot seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
|
2025-04-01T06:39:48.698830
| 2024-04-16T11:02:29
|
2245757984
|
{
"authors": [
"farnabaz"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9036",
"repo": "nuxt-hub/core",
"url": "https://github.com/nuxt-hub/core/pull/73"
}
|
gharchive/pull-request
|
feat: cache management
nuxt-hub/platform#152
@Atinux The tests were passed by recreating the pnpm-lock using pnpm@9.
|
2025-04-01T06:39:48.705883
| 2023-06-16T23:18:24
|
1761481344
|
{
"authors": [
"antlionguard",
"fabkho",
"simkuns"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9037",
"repo": "nuxt-modules/i18n",
"url": "https://github.com/nuxt-modules/i18n/issues/2168"
}
|
gharchive/issue
|
_ctx.$localePath is not a function
Environment
Operating System: Linux
Node Version: v16.14.2
Nuxt Version: 3.5.3
Nitro Version: 2.4.1
Package Manager<EMAIL_ADDRESS>Builder: vite
User Config: modules
Runtime Modules<EMAIL_ADDRESS>Build Modules: -
Reproduction
https://stackblitz.com/edit/github-6vqdsb?file=app.vue,package.json,nuxt.config.ts
Describe the bug
According to documentation available at https://v8.i18n.nuxtjs.org/guide/migrating#change-some-export-apis-name-on-nuxt-context it seems that API methods have to be prefixed with $, however it results in 500 error (_ctx.$localePath is not a function).
This works, but shows type error in VSCode
This does not work and results in 500 error
This does work
Additional context
No response
Logs
No response
any updates on this? Having the same problem...
As a workaround i use <nuxt-link-locale>, which just works.
Components · @nuxtjs/i18n
still same problem exist if $localePath used outside of setup.
|
2025-04-01T06:39:48.719399
| 2023-07-07T08:56:47
|
1793106760
|
{
"authors": [
"BobbieGoede",
"Tigriz",
"derHodrig",
"kazupon",
"productdevbook",
"thaikolja"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9038",
"repo": "nuxt-modules/i18n",
"url": "https://github.com/nuxt-modules/i18n/issues/2206"
}
|
gharchive/issue
|
nuxt 3 module i18n dont working
Environment
Operating System: Linux
Node Version: v18.16.0
Nuxt Version: 3.6.1
Nitro Version: 2.5.2
Package Manager<EMAIL_ADDRESS>Builder: vite
User Config: -
Runtime Modules: -
Build Modules: -
Reproduction
https://github.com/productdevbook/i18n-bugs-layer
Describe the bug
pnpm install
pnpm dev admin
✔ Nitro built in 364 ms nitro 11:55:20 AM
[intlify] Not found 'hello' key in 'en-US' locale messages. 11:55:21 AM
[intlify] Fall back to translate 'hello' key with 'en' locale. 11:55:21 AM
[intlify] Not found 'hello' key in 'en' locale messages. 11:55:21 AM
[intlify] Not found 'hello' key in 'en-US' locale messages. 11:55:21 AM
[intlify] Fall back to translate 'hello' key with 'en' locale. 11:55:21 AM
[intlify] Not found 'hello' key in 'en' locale messages. 11:55:21 AM
Additional context
pnpm install
pnpm dev admin
Logs
No response
Same here. All packages default (using Nuxt Content).
I got the same problem and tried nuxi upgrade --force once again (even though I updated the day 3.6.1 released) and it worked, dunno what happened tbh.
I got the same problem and tried nuxi upgrade --force once again (even though I updated the day 3.6.1 released) and it worked, dunno what happened tbh.
I updated the repo as you said, still the same problem persists.
Hi!
I’ve checked your minimal reproduction repo.
I’ve noticed that you need to set up two things to make nuxt i18n work with nuxt layer.
1. set your custom nuxt module at modules options
You need to set your nuxt module defined in the nuxt layer (layer dir) of the layer dir to nuxt.config.ts as follows
// https://nuxt.com/docs/api/configuration/nuxt-config
import MyModule from './module'
export default defineNuxtConfig({
modules: [
// https://i18n.nuxtjs.org/
MyModule,
'@nuxtjs/i18n',
],
devtools: { enabled: true }
})
It seems that the nuxt layer does not automatically install a custom nuxt module if you just define one.
configure i18n options
For nuxt application (admin dir) that extend the nuxt layer, the i18n options must be set as follows. The following is a case where lazy loading is used.
// https://nuxt.com/docs/api/configuration/nuxt-config
export default defineNuxtConfig({
extends: [
'../layer'
],
// see the docs: https://v8.i18n.nuxtjs.org/guide/layers
i18n: {
lazy: true,
langDir: 'lang', // need `lang` dir on `admin`
locales: [
{
code: 'en',
file: 'en.json',
},
{
code: 'fr',
file: 'fr.json',
},
]
}
})
If lazy loading is not used, at least the locales option must be defined; if locales does not have resource definitions in files, an empty array must be defined in files.
@BobbieGoede
If you have anything to add about the layer of nuxt i18n module, please comment. 🙏
Now that #2290 has been merged the configuration @kazupon described in step 2 are not necessary anymore. All you have to change in your reproduction is registering your module as mentioned.
I have changed your reproduction to demonstrate it works with the edge channel installed (I registered the module by putting it in a modules folder) check it out here.
Please let me know if you're still experiencing any issues even with the latest edge version installed!
@BobbieGoede
same repo nuxt 3.7.1 dont working
"@nuxtjs/i18n": "8.0.0-rc.4",
It is not working, if the Layer is installed as node dependency.
https://github.com/nuxt-modules/i18n/discussions/2388
Should we start a new Issue or reopen?
|
2025-04-01T06:39:48.724863
| 2022-09-02T03:02:58
|
1359646154
|
{
"authors": [
"FKonig",
"Mechse"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9039",
"repo": "nuxt-modules/strapi",
"url": "https://github.com/nuxt-modules/strapi/issues/278"
}
|
gharchive/issue
|
500 init is not a function
Version
@nuxtjs/strapi: v1.5.0
nuxt: v3.0.0-rc.4
node: v16.16.0
strapi: v4.3.6
Linux Pop_OS 20.04
Reproduction Link
https://github.com/FKonig/NuxtStrapi_Bug
Steps to reproduce
Normal minimal installation like in the linked repo.
Opening http://localhost:3000 -> 500 init is not a function
What is Expected?
Successful start of the application
What is actually happening?
Browser:
Terminal:
[nuxt] [request error] init is not a function
at Module.useState (./.nuxt/dist/server/server.mjs:993:26)
at Module.useStrapiUser (./.nuxt/dist/server/server.mjs:3184:51)
at Module.useStrapiAuth (./.nuxt/dist/server/server.mjs:3053:38)
at ./.nuxt/dist/server/server.mjs:3020:47
at fn (./.nuxt/dist/server/server.mjs:434:27)
at Object.callAsync (./node_modules/unctx/dist/index.mjs:42:19)
at callWithNuxt (./.nuxt/dist/server/server.mjs:436:23)
at applyPlugin (./.nuxt/dist/server/server.mjs:391:29)
at Module.applyPlugins (./.nuxt/dist/server/server.mjs:401:11)
at async createNuxtAppServer (./.nuxt/dist/server/server.mjs:46:7)
[nuxt] [request error] init is not a function
at Module.useState (./.nuxt/dist/server/server.mjs:993:26)
at Module.useStrapiUser (./.nuxt/dist/server/server.mjs:3184:51)
at Module.useStrapiAuth (./.nuxt/dist/server/server.mjs:3053:38)
at ./.nuxt/dist/server/server.mjs:3020:47
at fn (./.nuxt/dist/server/server.mjs:434:27)
at Object.callAsync (./node_modules/unctx/dist/index.mjs:42:19)
at callWithNuxt (./.nuxt/dist/server/server.mjs:436:23)
at applyPlugin (./.nuxt/dist/server/server.mjs:391:29)
at Module.applyPlugins (./.nuxt/dist/server/server.mjs:401:11)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
Hello FKonig,
I just cloned your reproduction repository.
Unfortunately it works fine on my machine.
If you use an functions from the @nuxtjs/strapi repo please consider this issue: #282 and there is an open pull request for this bug: #281
|
2025-04-01T06:39:48.727340
| 2023-08-02T15:09:27
|
1833376463
|
{
"authors": [
"CptJJ",
"larbish"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9040",
"repo": "nuxt-modules/supabase",
"url": "https://github.com/nuxt-modules/supabase/pull/222"
}
|
gharchive/pull-request
|
Release v1.0.0
Resolves #209
Resolves #185
Resolves #207
Resolves #163
Resolves #203
Resolves #153
Resolves #179
Resolves #137
Resolves #147
Resolves #80
Resolves #92
Resolves #208
Resolves #184
Resolves #165
Wooo! thanks a ton for this, works flawlessly. Im really curious why things seemed to implode over the last couple weeks for a ton of people? was it do to changes from the supabase team on their end?
The supabase libraries (mainly gotruejs) have evolved in a way that was not fiting our use in the module. A rewrite was needed !
|
2025-04-01T06:39:48.867043
| 2020-05-08T03:10:57
|
614472123
|
{
"authors": [
"Lustach",
"ahsanahmed321",
"alakameljebali",
"alexon1234",
"ankoe",
"daisuke-fukuda",
"ellaidevs",
"jcjp",
"petrovicz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9041",
"repo": "nuxt/create-nuxt-app",
"url": "https://github.com/nuxt/create-nuxt-app/issues/517"
}
|
gharchive/issue
|
Cannot create new project
C:\Users\Ala\Desktop\aa>npx create-nuxt-app dashboard
create-nuxt-app v2.15.0
✨ Generating Nuxt.js project in dashboard
? Project name dashboard
? Project description My kryptonian Nuxt.js project
? Author name
? Choose programming language JavaScript
? Choose the package manager Npm
? Choose UI framework Vuetify.js
? Choose custom server framework None (Recommended)
? Choose Nuxt.js modules Axios, Progressive Web App (PWA) Support, DotEnv
? Choose linting tools (Press to select, to toggle all, to invert selection)
? Choose test framework None
? Choose rendering mode Universal (SSR)
? Choose development tools jsconfig.json (Recommended for VS Code)
npm WARN deprecated<EMAIL_ADDRESS>core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to
the actual version of core-js@3.
npm WARN deprecated<EMAIL_ADDRESS>Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies.
npm WARN deprecated<EMAIL_ADDRESS>fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2.
npm WARN deprecated<EMAIL_ADDRESS>https://github.com/lydell/resolve-url#deprecated
npm WARN deprecated<EMAIL_ADDRESS>Please see https://github.com/lydell/urix#deprecated
npm WARN deprecated<EMAIL_ADDRESS>core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to
the actual version of core-js@3.
npm WARN deprecated<EMAIL_ADDRESS>Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies.
npm WARN deprecated<EMAIL_ADDRESS>fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2.
npm WARN deprecated<EMAIL_ADDRESS>https://github.com/lydell/resolve-url#deprecated
npm WARN deprecated<EMAIL_ADDRESS>Please see https://github.com/lydell/urix#deprecated
gyp ERR! find VS
gyp ERR! find VS msvs_version not set from command line or npm config
gyp ERR! find VS VCINSTALLDIR not set, not running in VS Command Prompt
gyp ERR! find VS could not use PowerShell to find Visual Studio 2017 or newer
gyp ERR! find VS looking for Visual Studio 2015
gyp ERR! find VS - not found
gyp ERR! find VS not looking for VS2013 as it is only supported up to Node.js 8
gyp ERR! find VS
gyp ERR! find VS **************************************************************
gyp ERR! find VS You need to install the latest version of Visual Studio
npm WARN deprecated<EMAIL_ADDRESS>core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to
the actual version of core-js@3.
npm WARN deprecated<EMAIL_ADDRESS>Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies.
npm WARN deprecated<EMAIL_ADDRESS>fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2.
npm WARN deprecated<EMAIL_ADDRESS>https://github.com/lydell/resolve-url#deprecated
npm WARN deprecated<EMAIL_ADDRESS>Please see https://github.com/lydell/urix#deprecated
gyp ERR! find VS
gyp ERR! find VS msvs_version not set from command line or npm config
gyp ERR! find VS VCINSTALLDIR not set, not running in VS Command Prompt
gyp ERR! find VS could not use PowerShell to find Visual Studio 2017 or newer
gyp ERR! find VS looking for Visual Studio 2015
gyp ERR! find VS - not found
gyp ERR! find VS not looking for VS2013 as it is only supported up to Node.js 8
gyp ERR! find VS
gyp ERR! find VS **************************************************************
gyp ERR! find VS You need to install the latest version of Visual Studio
npm WARN deprecated<EMAIL_ADDRESS>core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to
the actual version of core-js@3.
npm WARN deprecated<EMAIL_ADDRESS>Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies.
npm WARN deprecated<EMAIL_ADDRESS>fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2.
npm WARN deprecated<EMAIL_ADDRESS>https://github.com/lydell/resolve-url#deprecated
npm WARN deprecated<EMAIL_ADDRESS>Please see https://github.com/lydell/urix#deprecated
gyp ERR! find VS
gyp ERR! find VS msvs_version not set from command line or npm config
gyp ERR! find VS VCINSTALLDIR not set, not running in VS Command Prompt
gyp ERR! find VS could not use PowerShell to find Visual Studio 2017 or newer
gyp ERR! find VS looking for Visual Studio 2015
gyp ERR! find VS - not found
gyp ERR! find VS not looking for VS2013 as it is only supported up to Node.js 8
gyp ERR! find VS
gyp ERR! find VS **************************************************************
gyp ERR! find VS You need to install the latest version of Visual Studio
gyp ERR! find VS including the "Desktop development with C++" workload.
gyp ERR! find VS For more information consult the documentation at:
gyp ERR! find VS https://github.com/nodejs/node-gyp#on-windows
gyp ERR! find VS **************************************************************
gyp ERR! find VS
gyp ERR! configure error
gyp ERR! stack Error: Could not find any Visual Studio installation to use
gyp ERR! stack at VisualStudioFinder.fail (C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:121:47)
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:74:16
gyp ERR! stack at VisualStudioFinder.findVisualStudio2013 (C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:351:14)
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:70:14
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:372:16
npm WARN deprecated<EMAIL_ADDRESS>core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to
the actual version of core-js@3.
npm WARN deprecated<EMAIL_ADDRESS>Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies.
npm WARN deprecated<EMAIL_ADDRESS>fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2.
npm WARN deprecated<EMAIL_ADDRESS>https://github.com/lydell/resolve-url#deprecated
npm WARN deprecated<EMAIL_ADDRESS>Please see https://github.com/lydell/urix#deprecated
gyp ERR! find VS
gyp ERR! find VS msvs_version not set from command line or npm config
gyp ERR! find VS VCINSTALLDIR not set, not running in VS Command Prompt
gyp ERR! find VS could not use PowerShell to find Visual Studio 2017 or newer
gyp ERR! find VS looking for Visual Studio 2015
gyp ERR! find VS - not found
gyp ERR! find VS not looking for VS2013 as it is only supported up to Node.js 8
gyp ERR! find VS
gyp ERR! find VS **************************************************************
gyp ERR! find VS You need to install the latest version of Visual Studio
gyp ERR! find VS including the "Desktop development with C++" workload.
gyp ERR! find VS For more information consult the documentation at:
gyp ERR! find VS https://github.com/nodejs/node-gyp#on-windows
gyp ERR! find VS **************************************************************
gyp ERR! find VS
gyp ERR! configure error
gyp ERR! stack Error: Could not find any Visual Studio installation to use
gyp ERR! stack at VisualStudioFinder.fail (C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:121:47)
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:74:16
gyp ERR! stack at VisualStudioFinder.findVisualStudio2013 (C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:351:14)
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:70:14
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:372:16
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\util.js:54:7
npm WARN deprecated<EMAIL_ADDRESS>core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to
the actual version of core-js@3.
npm WARN deprecated<EMAIL_ADDRESS>Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies.
npm WARN deprecated<EMAIL_ADDRESS>fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2.
npm WARN deprecated<EMAIL_ADDRESS>https://github.com/lydell/resolve-url#deprecated
npm WARN deprecated<EMAIL_ADDRESS>Please see https://github.com/lydell/urix#deprecated
gyp ERR! find VS
gyp ERR! find VS msvs_version not set from command line or npm config
gyp ERR! find VS VCINSTALLDIR not set, not running in VS Command Prompt
gyp ERR! find VS could not use PowerShell to find Visual Studio 2017 or newer
gyp ERR! find VS looking for Visual Studio 2015
gyp ERR! find VS - not found
gyp ERR! find VS not looking for VS2013 as it is only supported up to Node.js 8
gyp ERR! find VS
gyp ERR! find VS **************************************************************
gyp ERR! find VS You need to install the latest version of Visual Studio
gyp ERR! find VS including the "Desktop development with C++" workload.
gyp ERR! find VS For more information consult the documentation at:
gyp ERR! find VS https://github.com/nodejs/node-gyp#on-windows
gyp ERR! find VS **************************************************************
gyp ERR! find VS
gyp ERR! configure error
gyp ERR! stack Error: Could not find any Visual Studio installation to use
gyp ERR! stack at VisualStudioFinder.fail (C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:121:47)
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:74:16
gyp ERR! stack at VisualStudioFinder.findVisualStudio2013 (C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:351:14)
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:70:14
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\find-visualstudio.js:372:16
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\util.js:54:7
gyp ERR! stack at C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\util.js:33:16
gyp ERR! stack at ChildProcess.exithandler (child_process.js:310:5)
gyp ERR! stack at ChildProcess.emit (events.js:315:20)
gyp ERR! stack at maybeClose (internal/child_process.js:1051:16)
gyp ERR! System Windows_NT 10.0.18363
gyp ERR! command "C:\Program Files\nodejs\node.exe" "C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\bin\node-gyp.js" "rebuild" "--release"
gyp ERR! cwd C:\Users\Ala\Desktop\aa\dashboard\node_modules\fibers
gyp ERR! node -v v14.2.0
gyp ERR! node-gyp -v v5.1.0
gyp ERR! not ok
node-gyp exited with code: 1
Please make sure you are using a supported platform and node version. If you
would like to compile fibers on this machine please make sure you have setup your
build environment--
Windows + OS X instructions here: https://github.com/nodejs/node-gyp
Ubuntu users please run: sudo apt-get install g++ build-essential
RHEL users please run: yum install gcc-c++ and yum groupinstall 'Development Tools'
Alpine users please run: sudo apk add python make g++
'nodejs' n'est pas reconnu en tant que commande interne
ou externe, un programme ex�cutable ou un fichier de commandes.
npm WARN optional SKIPPING OPTIONAL DEPENDENCY<EMAIL_ADDRESS>(node_modules\watchpack\node_modules\chokidar\node_modules\fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for<EMAIL_ADDRESS>wanted {"os":"darwin","arch":"any"} (current: {"os":"win32","arch":"x64"})
npm WARN optional SKIPPING OPTIONAL DEPENDENCY<EMAIL_ADDRESS>(node_modules\chokidar\node_modules\fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for<EMAIL_ADDRESS>wanted {"os":"darwin","arch":"any"} (current: {"os":"win32","arch":"x64"})
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR<EMAIL_ADDRESS>install: node build.js || nodejs build.js
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the<EMAIL_ADDRESS>install script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! C:\Users\Ala\AppData\Roaming\npm-cache_logs\2020-05-08T02_57_25_403Z-debug.log
<EMAIL_ADDRESS>install C:\Users\Ala\Desktop\aa\dashboard\node_modules\fibers
node build.js || nodejs build.js
C:\Users\Ala\Desktop\aa\dashboard\node_modules\fibers>if not defined npm_config_node_gyp (node "C:\Program Files\nodejs\node_modules\npm\node_modules\npm-lifecycle\nod
e-gyp-bin\....\node_modules\node-gyp\bin\node-gyp.js" rebuild --release ) else (node "C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\bin\node-gyp.js
" rebuild --release )
C:\Users\Ala\AppData\Roaming\npm-cache_npx\2584\node_modules\create-nuxt-app\node_modules\sao\lib\installPackages.js:108
throw new SAOError(Failed to install ${packageName} in ${cwd})
^
SAOError: Failed to install packages in C:\Users\Ala\Desktop\aa\dashboard
at ChildProcess. (C:\Users\Ala\AppData\Roaming\npm-cache_npx\2584\node_modules\create-nuxt-app\node_modules\sao\lib\installPackages.js:108:15)
at ChildProcess.emit (events.js:315:20)
at ChildProcess.cp.emit (C:\Users\Ala\AppData\Roaming\npm-cache_npx\2584\node_modules\create-nuxt-app\node_modules\sao\node_modules\cross-spawn\lib\enoent.js:34:29)
at maybeClose (internal/child_process.js:1051:16)
at Process.ChildProcess._handle.onexit (internal/child_process.js:287:5) {
__sao: true
}
I had the same issue and after updating node, the issue was fix.
I had the same problem, and solved.
See this line
Windows + OS X instructions here: https://github.com/nodejs/node-gyp
Ubuntu users please run: sudo apt-get install g++ build-essential
RHEL users please run: yum install gcc-c++ and yum groupinstall 'Development Tools'
Alpine users please run: sudo apk add python make g++
I used docker aipine image, so ran this.
apk update && apk add \
python\
make\
g++
Then solved.
@petrovicz try to clear your npm cache https://dev.to/rishiabee/npm-err-unexpected-end-of-json-input-while-parsing-near-743
@ankoe Unfortunately it didn't solve the issue for me
I had the same issue again in another project.
I needed to downgrade to use the node TLS version.
LTS versions are working for me also.
I tried clearing my npm cache and yarn cache both are failing for on the same error.
It seems that this problem only occurs when choosing Vuetify UI-framework
Same here ... When choosing vuetify
LTS version is the only solution for now
It seems that this problem only occurs when choosing Vuetify UI-framework UPD: u can use npm init nuxt-app
On my side, experiencing this after choosing tailwind css.
|
2025-04-01T06:39:48.900149
| 2017-11-25T06:42:15
|
276730313
|
{
"authors": [
"18717700273",
"clarkdo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9042",
"repo": "nuxt/nuxt.js",
"url": "https://github.com/nuxt/nuxt.js/issues/2236"
}
|
gharchive/issue
|
ERROR in common.83cd80e5c7fd8703ba74.js from UglifyJs
error
nuxt.config.js
package.json
why?
You should include the you dependency which use not supported es syntax yet in 'babel', pls have a look at: https://github.com/nuxt/nuxt.js/issues/1668#issuecomment-330510870
|
2025-04-01T06:39:48.907999
| 2018-08-23T05:08:26
|
353226629
|
{
"authors": [
"UltramarineLights",
"manniL",
"pbastowski"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9043",
"repo": "nuxt/nuxt.js",
"url": "https://github.com/nuxt/nuxt.js/issues/3790"
}
|
gharchive/issue
|
Major issue with namespaced Vuex modules
Version
v1.4.2
Reproduction link
https://vuex.vuejs.org/guide/modules.html
Steps to reproduce
Setup a module mode index.js in the store folder with code
WORKS
Copy the exact same code in a second-file.js in the store folder
Change all the helpers I'm using to reference the new namespace e.g. mapState(['stateName']) to mapState(['second-file/stateName'])
DOESN'T WORK
What is expected ?
I'm expecting Nuxt to normally access the module
What is actually happening?
Nuxt does not recognize any of the properties from the store module
Additional comments?
I want to highlight that this happens with a clean Nuxt install through Vue CLI and with very simple code in order to exclude mistakes of different nature. The two store files are exactly identical apart from one becoming namespaced due to Nuxt default behavior. The same issue persist if I try to register the modules in the classic Vuex mode, since also in this case Nuxt creates namespaced modules. Is there a way to disable the automatic namespacing and instead merge all the modules properties into the main store?
This bug report is available on Nuxt community (#c7619)
Have you tried nuxt-edge? I am using namespaced modules without any issues with nuxt-edge.
How do I migrate to nuxt-edge?
For me it was as simple as:
Uninstall nuxt
Install nuxt-edge
same problem with nuxt-edge, are you using the helpers?
I just tried as last resort to use the vuex namespace helper and now it works:
import { createNamespacedHelpers } from 'vuex'
const { mapState, mapActions } = createNamespacedHelpers('moduleName')
and then just the 'stateName' in the code, while without the namespaced helpers and just 'moduleName/stateName' it doesn't. It's a very strange issue
I kind of find a solution, which still doesn't justify to me why the other method doesn't work.
...mapState(['navigation/pages']) NOT WORKING
...mapState('navigation', ['pages']) WORKING
maybe it's just me not getting the obvious, anyways I'm very happy now that I managed to get it working, I was getting crazy
@ultramarinelights
I’m glad to read that you found a solution.
I do use helpers and they were working when I last checked. Perhaps I was using the second syntax, where you supply the namespace as the first argument. I’ll check when I’m back from vacation.
Closed in favor of https://github.com/nuxt/docs/issues/850
|
2025-04-01T06:39:48.914656
| 2020-04-24T11:03:09
|
606230581
|
{
"authors": [
"clarkdo",
"manniL",
"pi0",
"simllll"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9044",
"repo": "nuxt/nuxt.js",
"url": "https://github.com/nuxt/nuxt.js/issues/7268"
}
|
gharchive/issue
|
npm version of @nuxt/babel-preset-app is not up to date
Version
v2.12.2
Reproduction link
https://www.npmjs.com/package/@nuxt/babel-preset-app
Steps to reproduce
check out npm version of @nuxt/babel-preset-app
first the version there is higher than the one in master (2.12.2 vs. 2.12.1)
second the "bugfixes" option is not present in the docs (and also not in the source).
What is expected ?
correct version / source of babel-preset-app
What is actually happening?
old version is packed and published
This bug report is available on Nuxt community (#c10566)
cc @clarkdo
@pi0 It looks like there is an unexpected 2.12.2 packages, do you have any idea why they are published ?
@simllll The releases seem right, the release branch is 2.x not dev, we'll merge 2.x back to dev after release.
As bugfixs is a new feature, so it will be released 2.13.0.
Ah alright, wasn't aware of that. Thanks, looking forward to 2.13.0 then ;-)
Hi, @simllll release are correct. 2.12.1 and 2.12.2 released as hotfix to 2.x branch. For nuxt (stable) code reference, you should refer to 2.x branch not development.
New bugfixes option (#7144) added after 2.12.0 and will be available in<EMAIL_ADDRESS>(also currently nuxt-edge)
|
2025-04-01T06:39:48.919038
| 2020-11-16T11:05:15
|
743743609
|
{
"authors": [
"Zuckjet",
"pi0"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9045",
"repo": "nuxt/nuxt.js",
"url": "https://github.com/nuxt/nuxt.js/issues/8353"
}
|
gharchive/issue
|
content security policy header will lose When status code is 304
Versions
nuxt: v2.14.7
node: v12.16.3
Reproduction
Additional Details
Steps to reproduce
What is Expected?
What is actually happening?
I have a website built with nuxt.js, when I first visit, it has the correct csp header which I set in nuxt.config.js. When I visit this website again, it will lose csp header. And I found that when status code is 304, csp header will lose.
@pi0 @danielroe Would you please have a look on this issue ? Do you think it is a bug when http status code is 304, the CSP header will lose, if not, is there a configuration can make CSP header work as expected, if it is bug, do you have a plan to publish a patch because this can cause security problem because of the miss of CSP header. Thanks
This should be fixed in v2.14.8. Please reopen if not.
This should be fixed in v2.14.8. Please reopen if not.
Thanks. I have confirmed it was solved from v2.14.8
|
2025-04-01T06:39:48.929654
| 2021-02-17T07:46:21
|
809943266
|
{
"authors": [
"Soviut",
"pi0"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9046",
"repo": "nuxt/nuxt.js",
"url": "https://github.com/nuxt/nuxt.js/issues/8844"
}
|
gharchive/issue
|
Option to disable build progress bars in CLI even when TTY is enabled
Is your feature request related to a problem? Please describe.
I use Docker and wanted to enable TTY so I could see colours in Docker Compose logs. However, enabling TTY also renders the webpack progress bars during builds. The bars work but have some rendering issues, most likely due to Docker Compose prefixing each log line with with the container name.
For example, the Client and Server bars render twice for each update.
Describe the solution you'd like
A top level option, or documentation, on how to disable the progress bars in the CLI, even in a development environment would be very useful. I'd be able to keep TTY enabled without the rendering issues.
Ideally this would not disable the progress bars displayed in the browser during build.
Basically, I'd like the result I get with TTY disabled, but with colours.
and during an update...
Describe alternatives you've considered
For now, I run with TTY disabled in Docker Compose, however, I lose colours.
I tried configuring the plugin in the nuxt.config.js build extend() function but couldn't seem to get any options to override.
I tried removing the progress bar plugin from the config.plugins list in extend(config) but that broke a bunch of things Nuxt or webpack was relying on.
I could also cut a bug ticket on the https://github.com/nuxt-contrib/webpackbar repo to see if the rendering can be fixed inside a docker TTY with the line prefixing.
Additional context
Naturally, as soon as I posted this, I decided to try one last configuration that targeted the progress bar reporters.
// nuxt.config.js
export default {
// ...
build: {
extend: (config) => {
const bar = config.plugins.find((p) => p.constructor.name === 'WebpackBarPlugin')
bar.reporters = bar.reporters.filter((r) => r.constructor.name !== 'FancyReporter')
}
}
}
This worked!
By removing the FancyReporter from the WebpackBarPlugin's array of reporters it disabled the progress bars in the CLI but kept the progress bars in the browser.
It would appear that changing the WebpackBarPlugin.options.fancy = false doesn't work because the plugin already populated the reporters array based on that config (and environment) during setup.
I'll leave this here and keep the issue open since it may still be a good idea to streamline this process.
Some further investigation found that the previous example wasn't fully doing what I needed. It was removing the anonymous reporter that renders the progress bars in the browser. So instead of simply removing the FancyReporter, I wound up having to "clone" the WebpackBarPlugin, include the basic reporter and exclude the FancyReporter.
So here's the masterpiece
import isDocker from 'is-docker'
export default {
build: {
extend(config) {
// hide the build progress bars in the CLI (but not the browser)
// since they do not render correctly in docker-compose logs
if (isDocker()) {
const i = config.plugins.findIndex(
(p) => p.constructor.name === 'WebpackBarPlugin'
)
const plugin = config.plugins[i]
// replace existing bar plugins with clones that exclude the 'fancy'
// reporter and include a 'basic' reporter
config.plugins.splice(
i,
1,
new WebpackBarPlugin({
name: plugin.options.name,
reporters: [
'basic', // include reporter as string because we cannot import
...plugin.reporters.filter(
(p) => p.constructor.name !== 'FancyReporter'
),
],
})
)
}
}
}
}
For future me, it looks like here is where you'd have to start to implement this as a nuxt option
https://github.com/nuxt/nuxt.js/blob/ba44b0f9ca8955ddb884744f34192c831b2d1d16/packages/webpack/src/config/base.js#L445
Hi @Soviut. Have you tried MINIMAL=1 environment variable?
@pi0 I didn't know that was available but it seems to do the same thing. Is that environment variable controlling std-env to make env.minimalCLI?
Is that environment variable controlling std-env to make env.minimalCLI?
Yes :)
Thanks. I cut a ticket on the std-env repo to update the README with the environment variables. I'll see if I can contribute that documentation soon after.
https://github.com/nuxt-contrib/std-env/issues/9
|
2025-04-01T06:39:48.942922
| 2024-02-09T20:15:17
|
2127790239
|
{
"authors": [
"KresimirCosic",
"MickL",
"danielroe"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9047",
"repo": "nuxt/nuxt",
"url": "https://github.com/nuxt/nuxt/issues/25724"
}
|
gharchive/issue
|
3.10.x: nuxt build creates vue files in .nuxt/dist/server/_nuxt folder
Environment
- Operating System: Darwin
- Node Version: v21.2.0
- Nuxt Version: 3.10.0
- CLI Version: 3.10.0
- Nitro Version: 2.8.1
- Package Manager<EMAIL_ADDRESS>- Builder: -
- User Config: devtools, css, modules, runtimeConfig, image, build, i18n, ignore, nitro, routeRules, security, vite
- Runtime Modules<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS>- Build Modules: -
Reproduction
Describe the bug
With 3.10.0 and 3.10.1 I get an error at the end of nuxt build: ERROR RollupError: At least one <template> or <script> is required in a single file component, with 3.9.3 everything is fine.
Unfortunately I cant reproduce this in the sandbox but I already tried to debug this for hours. I nailed it down to one of my ~10 components: If it contains a <style> tag with ANY content (even just a comment). Weirdly when I delete the content of the <script> tag I can have styles:
Throws the error:
<script setup lang="ts">
// hello
</script>
<template>
hello
</template>
<style lang="scss">
// hello
</style>
Doesnt throw an error:
<script setup lang="ts">
// hello
</script>
<template>
hello
</template>
<style lang="scss"></style>
Doesnt throw an error:
<script setup lang="ts"></script>
<template>
hello
</template>
<style lang="scss">
// hello
</style>
It also only happens with this one component named overlay.vue. I feel like I am getting crazy or something but this seems to make absolutely no sense.
Full error:
ERROR RollupError: At least one <template> or <script> is required in a single file component. nitro 9:00:59 PM
undefined
ERROR At least one <template> or <script> is required in a single file component. 9:00:59 PM
at error (node_modules/rollup/dist/es/shared/parseAst.js:337:30)
at Object.error (node_modules/rollup/dist/es/shared/node-entry.js:18507:20)
at Object.error (node_modules/rollup/dist/es/shared/node-entry.js:17616:42)
at node_modules/rollup-plugin-vue/dist/sfc.js:22:49
at Array.forEach (<anonymous>)
at Object.transformSFCEntry (node_modules/rollup-plugin-vue/dist/sfc.js:22:16)
at Object.transform (node_modules/rollup-plugin-vue/dist/index.js:99:38)
at node_modules/rollup/dist/es/shared/node-entry.js:18692:40
ERROR At least one <template> or <script> is required in a single file component. 9:00:59 PM
error Command failed with exit code 1.
Additional context
No response
Logs
No response
This is unlikely to be an issue with Nuxt - rollup-plugin-vue is not used by Nuxt or Nitro and is what is throwing your error.
I am not 100% sure about that. Yes this plugin throws the error BUT I disabled the plugin and realized there is exactly one .vue file in the folder .nuxt/dist/server/_nuxt which is exactly the file I debugged to be the one creates the error. I guess there shouldnt be any vue file in this folder which one mean that there actually is a bug within Nuxt.
Interesting. Have you customised assetFileNames by any chance?
I did not
I probably can't look into this without a reproduction or more info.
I can try to dig more into this and make this reproducable.
@danielroe So it took me a few hours of removing more and more stuff from my app until I ended in an empty app and I realized:
This happens on a blank empty app created with nuxi init and just a single index page and a single component in it! If the component has a script with lang=ts and a style tag (both with ANY content) a vue file is in the .nuxt/dist/server folder!
Since StackBlitz doesnt seem to work, here is the repo: https://github.com/MickL/nuxt-bug
When you say that StackBlitz doesn't work, what do you mean?
I can't reproduce with that repo.
Here's a StackBlitz created from it - seemingly working fine: https://stackblitz.com/edit/github-iqamf7.
Could you share any more info about your setup?
I used bun install if that changed anything. My specs are at the very top. I think #25690 could be the same issue
When you say that StackBlitz doesn't work, what do you mean?
I meant that the links on the Nuxt website (the two I posted above) both dont start and produce errors.
So I double checked on another machine. I downloaded my repo, run npm install and then npm run buld. If I check the folder `` I see (as I described) there is a vue file which is probably supposed to be a js file:
You're quite right; it is reproducible there - I was checking .output, not .nuxt.
I am not sure if this is causing me problems now - I updated my packages to latest (all of them), and somehow I get an error:
cannot find module './stringify' (module in question is qs, which is only used by Strapi module as far as I can tell)
This only happens in dev mode, build is running fine. When I don't include it in the configuration everything works.
If I keep use the Nuxt ^3.10.1 version I have no problems. Does any of this ring a bell?
|
2025-04-01T06:39:48.955357
| 2024-02-19T18:28:27
|
2142944119
|
{
"authors": [
"benjamincanac",
"carlosvaldesweb"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9048",
"repo": "nuxt/ui",
"url": "https://github.com/nuxt/ui/issues/1382"
}
|
gharchive/issue
|
Horizontal landing hero items end not works
Environment
Operating System: Darwin
Node Version: v20.9.0
Nuxt Version: 3.10.2
CLI Version: 3.10.1
Nitro Version: 2.8.1
Package Manager<EMAIL_ADDRESS>Builder: -
User Config: css, extends, modules, stripe, eslint, image, runtimeConfig, imports, ui, devtools, experimental
Runtime Modules<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS>Build Modules: -
Version
"2.13.0",
Reproduction
In description
Description
Hello, i cannot override lg:items-center in uLandingHero when orientation is horizontal
<ULandingHero
:ui="{
container: 'lg:items-end sm:gap-0',
}"
orientation="horizontal"/>
Additional context
No response
Logs
No response
I've already transferred your issue to the @nuxt/ui-pro repository! This will be fixed in the next release.
|
2025-04-01T06:39:48.956659
| 2023-07-14T02:05:29
|
1804026910
|
{
"authors": [
"acidjazz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9049",
"repo": "nuxtlabs/ui",
"url": "https://github.com/nuxtlabs/ui/issues/412"
}
|
gharchive/issue
|
u-accordian - name and provide a type for items
import type { AccordianItems } from '@nuxtlabs/ui'
const items:AccordianItems = [
...
Is this in the edge? how do i import it?
import type { AccordianItem } from '@nuxthq/ui'
?
|
2025-04-01T06:39:48.959045
| 2024-08-12T21:18:17
|
2461912322
|
{
"authors": [
"dagardner-nv",
"mdemoret-nv"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9050",
"repo": "nv-morpheus/Morpheus",
"url": "https://github.com/nv-morpheus/Morpheus/pull/1840"
}
|
gharchive/pull-request
|
Fix usage of the C++ impl of write_df_to_file
Description
Fixes a bug when C++ mode is enabled, the file would be written twice.
By Submitting this PR I confirm:
I am familiar with the Contributing Guidelines.
When the PR is ready for review, new or existing tests cover these changes.
When the PR is ready for review, the documentation is up to date with these changes.
/merge
|
2025-04-01T06:39:49.007408
| 2020-02-29T23:15:46
|
573428705
|
{
"authors": [
"bearcatsandor",
"cjoshmartin"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9051",
"repo": "nvbn/thefuck",
"url": "https://github.com/nvbn/thefuck/issues/1058"
}
|
gharchive/issue
|
I just ran fuck, and it already has support for this feature
Ugh. I donno why i missed that. Probably because I'm in a neovim terminal and there was a delay of a second due to my keybindings
Thank you!
|
2025-04-01T06:39:49.008264
| 2016-02-18T17:19:37
|
134640989
|
{
"authors": [
"nvbn",
"scorphus"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9052",
"repo": "nvbn/thefuck",
"url": "https://github.com/nvbn/thefuck/pull/462"
}
|
gharchive/pull-request
|
Thanks!
|
2025-04-01T06:39:49.026865
| 2022-08-15T04:09:49
|
1338511872
|
{
"authors": [
"dannyvassallo",
"nvh95"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9057",
"repo": "nvh95/jest-preview",
"url": "https://github.com/nvh95/jest-preview/issues/235"
}
|
gharchive/issue
|
@media doesn't seem to compile with css / scss modules + more
The Problem
I'm using jest-preview to visualize screens using css modules. Some of the styling is dependent on media queries. None of the @media statements seem to get compiled into the embedded style tags. Other classes from the module make it in fine.
I also cannot import SCSS files into the setup file. If I compile them using gulp or webpack and then import the resulting .css file, it appears to work fine.
Additional issue but not related, I can't get auto preview to work at all.
What I've tried
So far I've tried compiling the SCSS myself and removing hashes from the class names. I've faced trouble using the advanced configuration to try and remove the hashes from the jest-preview transform source. This didn't seem to work either.
I've also tried using a custom transform with some additional postCss utilities but haven't had any success. If I render components with no media queries, this tool is absolutely wonderful.
What is the framework/ technology you want to integrate Jest Preview to?
"react": "^16.13.1"
"jest": "^26.6.3"
"postcss": "^8.4.6"
"sass": "^1.53.0"
"sass-loader": "^10.2.1"
Thank you
Thank you for your time and continued effort on this tool! If I can contribute, I'd love to if someone can point me in the right direction.
None of the @media statements seem to get compiled into the embedded style tags
Can you provide minimum reproduction. I added a commit to include @media query and it works fine. Please refer to this commit:
https://github.com/nvh95/jest-preview/commit/0e8a8273146a993b300500e30b2c5fc7e157bbd3#diff-650149a55a5b9fea04fefedba299bd2bf341e1a1c8b4d6e63e483e95669efbfcR4-R9
I also cannot import SCSS files into the setup file.
Please help to prepare a reproduction. In our demo, we can import CSS modules file (https://github.com/nvh95/jest-preview/blob/41019e629c89c48580f29841d6380d89d8cdb41d/demo/setupTests.js#L4). Thanks.
It looks like you are correct! I do think I might have an idea of what's happening. I think the class name property on the component changes based on the media or screen size and it doesnt seem to be updating properly. Regardless of the window or screen size I stub in jest, it falls back to either default or the mobile values. Maybe it's something with the way I'm handling my renders?
|
2025-04-01T06:39:49.039654
| 2024-09-16T03:45:31
|
2527438735
|
{
"authors": [
"Tonitum",
"lambtho12"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9058",
"repo": "nvim-telekasten/telekasten.nvim",
"url": "https://github.com/nvim-telekasten/telekasten.nvim/pull/343"
}
|
gharchive/pull-request
|
fix: remove hard-coded "weekly" in folder creation prompt
Proposed change
Replace hard-coded "Create weekly dir" in new folder creation prompt with more generic "Notes", resulting in the following:
Previous: "Create weekly dir folder /home//projects/second-brain/foobar/ does not exist! Shall I create it?"
New: "Notes folder /home/cameron/projects/second-brain/foobar/ does not exist! Shall I create it?"
Type of change
[x] Bugfix (non-breaking change which fixes an issue)
[ ] New feature (thank you!)
[x] Code quality improvements to existing code or addition of tests
[ ] Documentation update
Additional information
This PR fixes or closes issue: fixes #297
Checklist
[x] I am running the latest version of the plugin.
[x] The code change is tested and works locally.
[x] There is no commented out code in this PR.
[x] The code has been formatted using Stylua (a .stylua.toml file is provided)
[x] The code has been checked with luacheck (a .luacheckrc file is provided)
[ ] The README.md has been updated according to this change.
[ ] The doc/telekasten.txt helpfile has been updated according to this change.
Thank you for your contribution!
|
2025-04-01T06:39:49.041319
| 2021-02-12T18:00:06
|
807438977
|
{
"authors": [
"traysh"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9059",
"repo": "nvim-telescope/telescope-github.nvim",
"url": "https://github.com/nvim-telescope/telescope-github.nvim/issues/14"
}
|
gharchive/issue
|
pull_request diff view broken
in pull_request, when I try to switch to the "diff" mode, the preview shows only:
Usage: gh pr diff [<number> | <url> | <branch>] [flags]
Flags:
--color string Use color in diff output: {always|never|auto} (default "auto")
[Process exited 1]
fix: https://github.com/nvim-telescope/telescope-github.nvim/pull/16
|
2025-04-01T06:39:49.118816
| 2024-09-28T22:30:59
|
2554544034
|
{
"authors": [
"mklefrancois",
"tigrazone"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9060",
"repo": "nvpro-samples/nvpro_core",
"url": "https://github.com/nvpro-samples/nvpro_core/issues/71"
}
|
gharchive/issue
|
overbrights
I have a scene which rendered overbright.
Is it posible to autoexposure or use all lights power for automatically balance of lights in scene.
In sky mode all looks ok
But when I upload HDRI and also when I set intensity to 0 - I have overbright
Looks like an issue
one more case of this issue - sun disk scale = 0 or sun disk intensity = 0 make overbrights
one more scene with emissive textures
with HDRI on
HDRI off
Have you tried the tonemappers ?
tone mappers doesnt helps.
emissive area lights not calculated in MIS well.
in code processed hdris and punctual lights, not area lights
The scene has indeed very strong lights, which I believe doesn't follow the rule of the intensity parameter, which is lumens per steradian (lm/sr).
There are a few things which can be done:
Increase "Max Luminance", which is controlling the Firefly Filter and is cutting out incoming energy.
Reduce the Tonemapper exposure to compensate the over bright lights
Set the sun elevation to -90 deg, to remove any contribution from the Physical Sun & Sky model. Changing the Disk Scale is not physically correct, but a user control was added for "artistical effect", but only the 1.0 value is correct.
Same is true for the SciFi scene
big versions of images doesnt open (started from private-user-images.githubusercontent.com
thank you for explaination. is this parameters can be automatic processed?
for example, "auto exposure" which can measure max luminance and fit parameters?
An auto-exposure can be done, but to properly do this, we need the inverse of the tonemapper in use. We do not plan this feature in the short term.
|
2025-04-01T06:39:49.154420
| 2022-02-28T19:35:20
|
1154470623
|
{
"authors": [
"AgentEnder",
"YasserKharab"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9061",
"repo": "nx-dotnet/nx-dotnet",
"url": "https://github.com/nx-dotnet/nx-dotnet/issues/381"
}
|
gharchive/issue
|
Dependency Graph not being correctly generated
Just started using nx-dotnet and having issues with dep-graph
To Reproduce
Just created a new workspace with two dotnet libs (lib1 ref lib2)
generate the workspace: npx create-nx-workspace@latest
choose template: apps
install nxdotnet: npm i --save-dev @nx-dotnet/core
create lib1 and lib2
add reference from lib1 to lib2
build everything: npx nx run-many --target=build --all
generate dep-graph: nx dep-graph
result:
also the lib1.csproj inside .cashe\nx\nxdeps.json doesn't contain any "deps" node
Expected behavior
nxdeps.json should contain the deps for each csproj
Can you run nx report and also check that '@nx-dotnet/core' is listed as a plugin in nx.json?
Yes i do have "plugins": ["@nx-dotnet/core"] under nx.json
results for nx report
Node : 16.3.0
OS : win32 x64
npm : 7.18.1
nx : 13.8.3
@nrwl/angular : undefined
@nrwl/cli : 13.8.3
@nrwl/cypress : undefined
@nrwl/detox : undefined
@nrwl/devkit : 13.8.3
@nrwl/eslint-plugin-nx : undefined
@nrwl/express : undefined
@nrwl/jest : 13.8.3
@nrwl/js : undefined
@nrwl/linter : 13.8.3
@nrwl/nest : undefined
@nrwl/next : undefined
@nrwl/node : undefined
@nrwl/nx-cloud : 13.1.5
@nrwl/react : undefined
@nrwl/react-native : undefined
@nrwl/schematics : undefined
@nrwl/storybook : undefined
@nrwl/tao : 13.8.3
@nrwl/web : undefined
@nrwl/workspace : 13.8.3
typescript : 4.5.5
rxjs : 6.5.5
Community plugins:
@nx-dotnet/core: 1.9.2
|
2025-04-01T06:39:49.162382
| 2017-03-14T02:41:58
|
213953214
|
{
"authors": [
"formatme",
"neko"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9063",
"repo": "nya/cpomf",
"url": "https://github.com/nya/cpomf/issues/25"
}
|
gharchive/issue
|
shards install not grabbing md5
in lib/pg/src/pq/connection.cr:2: while requiring "crypto/md5": can't find file 'crypto/md5' relative to '/root/cpomf/lib/pg/src/pq'
require "crypto/md5"
the crypto/md5 is not under the lib folder compared to the rest of the things shards installed
running a new version of Crystal compared to Crystal 0.20.4
https://github.com/crystal-lang/crystal/tree/master/src/crypto
md5 is missing from the newest one so might be why
we're currently running Crystal 0.20.4 without any issues; but we'll likely try to update to a newer Crystal version in the near future.
in the meantime i'd recommend just running Crystal 0.20.4, or submitting a pull request.
Switched to Crystal 0.20.4 and had to get shards and it compiles now
|
2025-04-01T06:39:49.185842
| 2021-10-19T12:46:20
|
1030264007
|
{
"authors": [
"Agreon",
"ali-elamri"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9064",
"repo": "nykula/connect-typeorm",
"url": "https://github.com/nykula/connect-typeorm/issues/29"
}
|
gharchive/issue
|
Let users define the name of the expiredAt column
We use snake_case in our project, would be cool to just be able to pass this as an option
Not to revive an old comment, but if you/someone stumbles upon this, you can:
@Index()
@Column({ name: 'expired_at', type: 'bigint' })
expiredAt: number;
which is a built-in functionality in TypeORM :)
|
2025-04-01T06:39:49.191339
| 2022-08-25T19:27:33
|
1351357871
|
{
"authors": [
"AaronDDM",
"mrashed-dev"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9065",
"repo": "nylas/use-cases",
"url": "https://github.com/nylas/use-cases/pull/16"
}
|
gharchive/pull-request
|
Create vue demo for read-email use case
Description
This PR adds Vue support for read emails use cases. Functionally it works but the styling is a bit off.
License
I confirm that this contribution is made under the terms of the MIT license and that I have the authority necessary to make this contribution on behalf of its copyright owner.
@mrashed-dev is there some prettier plugin we can enable for vue?
|
2025-04-01T06:39:49.201934
| 2021-12-16T05:00:38
|
1081772931
|
{
"authors": [
"watarutmnh"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9066",
"repo": "nystudio107/craft-vite",
"url": "https://github.com/nystudio107/craft-vite/issues/19"
}
|
gharchive/issue
|
[Support] No assets output at production
My craft won't output asset links with {{ craft.vite.script("/src/app.js") }} at production.
Only the code below is output.
<script type="module">!function(){const e=document.createElement("link").relList;if(!(e&&e.supports&&e.supports("modulepreload"))){for(const e of document.querySelectorAll('link[rel="modulepreload"]'))r(e);new MutationObserver((e=>{for(const o of e)if("childList"===o.type)for(const e of o.addedNodes)if("LINK"===e.tagName&&"modulepreload"===e.rel)r(e);else if(e.querySelectorAll)for(const o of e.querySelectorAll("link[rel=modulepreload]"))r(o)})).observe(document,{childList:!0,subtree:!0})}function r(e){if(e.ep)return;e.ep=!0;const r=function(e){const r={};return e.integrity&&(r.integrity=e.integrity),e.referrerpolicy&&(r.referrerPolicy=e.referrerpolicy),"use-credentials"===e.crossorigin?r.credentials="include":"anonymous"===e.crossorigin?r.credentials="omit":r.credentials="same-origin",r}(e);fetch(e.href,r)}}();</script>
I couldn't figure out why but everything is fine at development mode.
craft 3.7.23
craft-vite 1.0.19
No error at debug bar at frontend.
Please help.
It seems be caused by some composer modules.
composer update will produce this issue.
I found that craft-plugin-vite 1.0.17 required by craft-vite 1.0.18 will cause the issue.
I created new issue for craft-plugin-vite and will close this.
https://github.com/nystudio107/craft-plugin-vite/issues/6
|
2025-04-01T06:39:49.273236
| 2021-11-07T18:53:00
|
1046822466
|
{
"authors": [
"hamster65",
"o0shojo0o"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9067",
"repo": "o0shojo0o/PixelIt",
"url": "https://github.com/o0shojo0o/PixelIt/issues/85"
}
|
gharchive/issue
|
Temperature/Humidity stopped working after update to 0.3.6
Hi,
since release 0.3.6 when support for BME280 was added, my sensor AM2320 returns "not installed" most of the time.
I tried to figure out what is wrong and discovered that it works after I reset the ESP (for example by klicking "save config" in the webinterface). It does not work after a power cycle.
Maybe a delay is needed after power up before initialization / detection? I guess the DHT22 does not need that delay which is why no one else seems to be having that issue :)
Hi,
I'll take a look, however I don't have an AM2320 here.
Hi,
I am testing right now. The sensor supports i2C and single bus mode. For single bus mode, which is the same as with DHT22, SCL (connected to GND) must be low for 500 ms. I guess reading the sensor earlier puts it into I2C mode. The datasheet also says that only a power cycle can change the mode. Page 5:
https://cdn-shop.adafruit.com/product-files/3721/AM2320.pdf
I tested a delay right before dht.setup. 400ms was too short, 500 worked most of the time, so I am using 600 ms now.
Hi, I think we can work with a delay here, the 600 ms should not be a problem.
Cool, thanks! 600ms works here without any problems so far.
fixed in v0.3.14
Hi, after flashing the latest version, my sensor stoppend working. I think this happens because it checks for other sensors on the same pin before the 600 ms delay.
I have not verified this because I am missing getSmoothedLux. How do I update the library? I expected vscode / platform.io to do this automatically.
Thanks!
Got it, I had to do a "clean all" to get the latest version of the LDR lib. I moved the delay up to where it says "// Init Temp Sensors" and it's working again.
Seems like a longer delay of 800ms is needed to make it work reliably.
The problem that the sensor could not be detected happend only in the morning after powering up my Pixelit. Power cycles during the day, leaving it off for some minutes, resulted in a working sensor. Perhaps it depends on the room temperature.
|
2025-04-01T06:39:49.275312
| 2016-12-20T19:10:03
|
196753845
|
{
"authors": [
"o355"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9068",
"repo": "o355/pyterm",
"url": "https://github.com/o355/pyterm/issues/33"
}
|
gharchive/issue
|
Check for updates program
Grabs a file, checks it's internal build number versus what the file provided is.
It'll have a method to disable itself on indev builds, and be put on GitHub (it's a text file...after all.)
Targeted to beta 3, but it may get pushed to beta 4.
Now implemented into 2.0.
I'll implement the update program into PyTerm 1.15.1 LTS, and stable 1.14.1 soon.
|
2025-04-01T06:39:49.321774
| 2021-11-23T18:54:32
|
1061610940
|
{
"authors": [
"FiniteStateGit",
"NiklasHenricson",
"sptramer"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9069",
"repo": "o3de/sig-docs-community",
"url": "https://github.com/o3de/sig-docs-community/issues/24"
}
|
gharchive/issue
|
SIG Elections 12/1 - 12/15 Nominations
SIG Docs-Community chair / co-chair elections for 2022
For the first year of O3DE, the D&C chair has been staffed as an interim position. It's time to hold some official elections, following some of the proposed guidance but with our own process due to the holiday season and in order to expedite the elections into next year.
The chair / co-chair roles
The chair and co-chair serve equivalent roles in the governance of the SIG and are only differentiated by title in that the highest vote-getter is the chair and the second-highest is the co-chair. The chair and co-chair are expected to govern together in an effective way and split their responsibilities to make sure that the SIG operates smoothly and has the availability of a chairperson at any time.
Unless distinctly required, the term "chairperson" refers to either/both of the chair and co-chair. If a chair or co-chair is required to perform a specific responsibility for the SIG they will always be addressed by their official role title.
In particular, if both chairpersons would be unavailable during a period of time, the chair is considered to be an on-call position during this period. As the higher vote-getter they theoretically represent more of the community and should perform in that capacity under extenuating circumstances. This means that if there is an emergency requiring immediate action from the Documentation & Community SIG, the chair will be called to perform a responsibility.
Responsibilities
Schedule and proctor regular SIG meetings on a cadence to be determined by the SIG.
Serve as a source of authority (and ideally wisdom) with regards to O3DE documentation. Chairpersons are the ultimate arbiters of many documentation standards, processes, and practices.
Participate in the SIG Docs-Community Discord channel and on the GitHub Discussion forums.
Serve as a representative of the broader O3DE community to all other SIGs, partners, the governing board, and the Linux Foundation.
Represent the SIG to O3DE partners, the governing board, and the Linux Foundation.
Coordinate with partners and the Linux Foundation regarding official community events.
Represent (or select/elect representatives) to maintain relationships with all other SIGs as well as the marketing committee.
Serve as an arbiter in SIG-related disputes.
Coordinate releases with SIG Release.
Assist contributors in finding resources and setting up official project or task infrastructure monitored/conducted by the SIG.
Long-term planning and strategy for the course of documentation for O3DE.
Maintain a release roadmap for the O3DE documentation.
Additionally, at this stage of the project, the SIG chairpersons are expected to act in the Maintainer role for review and merge purposes only, due to the lack of infrastructure and available reviewer/maintainer pool.
... And potentially more. Again, this is an early stage of the project and chair responsibilities have been determined more or less ad-hoc as new requirements and situations arise. In particular the community half of this SIG has been very lacking due to no infrastructural support, and a chairperson will ideally bring some of these skills.
Nomination
Nomination may either be by a community member or self-nomination. A nominee may withdraw from the election at any time for any reason until the election starts on 12/1.
Nomination requirements
For this election, nominees are required to have at minimum two merged submissions to o3de.org. This is to justify any temporary promotion to Maintainer as required by this term as chairperson. Submissions may be in-flight as of the nomination deadline (2021-12-01 12PM PT), but the nominee must meet the 2-merge requirement by the end of the election or they will be removed from the results.
Any elected chairperson who does not currently meet the Maintainer status will be required to work with contributors from the SIG to produce an appropriate number of accepted submissions by January 31, 2022 or they will be removed and another election will be held.
The only other nomination requirement is that the nominee agrees to be able to perform their required duties and has the availability to do so, taking into account the fact that another chairperson will always be available as a point of contact.
How to nominate
Nominate somebody (including yourself) by responding to this issue with:
A statement that the nominee should be nominated for a chair position in the Documentation & Community SIG. Nominees are required to provide a statement that they understand the responsibilities and requirements of the role, and promise to faithfully fulfill them and follow all contributor requirements for O3DE.
The name under which the nominee should be addressed. Nominees are allowed to contact the election proctor to have this name changed.
The GitHub username of the nominee (self-nominations need not include this; it's on your post.)
Nominee's Discord username (sorry, but you must be an active Discord user if you are a chairperson.)
Election process
The election will be conducted between 2021-12-01 12:00PM PT and 2021-12-15 12:00PM PT and held through an online poll. Votes will be anonymous and anyone invested in the direction of O3DE and its documentation may vote. If you choose to vote, we ask that you be familiar with the nominees.
The current interim chair (@sptramer) will announce the results in the sig-docs-community Discord and on the sig-docs-community O3DE mailing list no later than 2021-12-15 1:00PM PT. At that time if there is a dispute over the result or concern over vote tampering, voting information will be made public to the extent that it can be exported from the polling system and the SIG will conduct an independent audit under the guidance of a higher governing body in the foundation.
The elected chairpersons will begin serving their term on 2022-01-01 at 12AM PT. Tentatively SIG D&C chairs will be elected on a yearly basis. If you have concerns about wanting to replace chairs earlier, please discuss in the request for feedback on Governance.
I would like to nominate myself (Stephen Tramer) for a chairship. I understand the responsibilities and requirements of the role, having performed them for the last year, and will continue to fulfill them and all contributor requirements for O3DE.
Nominee name: Stephen Tramer
Discord: stramer#7057
I, Jonathan Capes, nominate myself and accept nomination for a chair position in the Docs-Community SIG, and will serve if elected. I affirm that I understand the requirements and responsibilities of the chairpersonship. I further affirm that I will fulfill the requirements and responsibilities of both the chairpersonship and of a O3DE contributor.
I would like to nominate myself (Stephen Tramer) for a chairship. I understand the responsibilities and requirements of the role, having performed them for the last year, and will continue to fulfill them and all contributor requirements for O3DE.
Nominee name: Stephen Tramer
Discord: stramer#7057
I 2nd Stephen's nomination. He has really good knowledge of the engine, the needs for software developers, he listens to others which is one of his strongest sides. Something I haven't experienced with other candidates. I'm there fore behind his decision to step up here.
Nominations have concluded. In light of the fact that there are only two nominees and one received a "second", as well as unanimous agreement among the chairpersons, Stephen Tramer (@sptramer) will serve as Chair and Jonathan Capes (@FiniteStateGit) will serve as Co-chair.
Congratulations!
|
2025-04-01T06:39:49.327849
| 2019-11-29T08:40:03
|
530220994
|
{
"authors": [
"pkliczewski"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9070",
"repo": "oVirt/ovirt-engine-sdk-go",
"url": "https://github.com/oVirt/ovirt-engine-sdk-go/pull/188"
}
|
gharchive/pull-request
|
Add disk download example
This PR contains reimplemention of python example how to download
disk.
@imjoey please take a look
@imjoey updated, please take a look.
|
2025-04-01T06:39:49.336945
| 2015-09-23T14:12:50
|
107929064
|
{
"authors": [
"dooremont",
"mrvisser",
"nicolaasmatthijs"
],
"license": "ECL-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9071",
"repo": "oaeproject/3akai-ux",
"url": "https://github.com/oaeproject/3akai-ux/issues/4067"
}
|
gharchive/issue
|
Problem with timeago in French
Assigning to @mrvisser for fix
thanks
This was probably seen as a result of a trailing space being removed after \. A crowdin sync appears to have added it back in, so this is no longer reproducable.
|
2025-04-01T06:39:49.347797
| 2021-12-31T03:15:53
|
1091394418
|
{
"authors": [
"barnettZQG",
"chwetion"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9072",
"repo": "oam-dev/kubevela",
"url": "https://github.com/oam-dev/kubevela/issues/3030"
}
|
gharchive/issue
|
[Feature] add the api that trigger deletion
Currently, support creation and query list API for the trigger. we need to support deletion.
https://github.com/oam-dev/kubevela/blob/master/pkg/apiserver/rest/webservice/application.go#L126
/assign chwetion
@chwetion thanks!
|
2025-04-01T06:39:49.354200
| 2022-01-05T03:04:11
|
1093932893
|
{
"authors": [
"StevenLeiZhang",
"wangyikewxgm",
"wonderflow"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9073",
"repo": "oam-dev/kubevela",
"url": "https://github.com/oam-dev/kubevela/issues/3042"
}
|
gharchive/issue
|
invalid addon repo url makes apiserver crash
Describe the bug
invalid addon repo url makes apiserver crash
To Reproduce
launch VelaUX
go to Addons
click button "addon registries"
click button "new"
set "Type" to be GIthub
input one invalid URL(means that is not a git repo follows "Addon Format") to "URL" textbox
click button "submit"
Expected behavior
VelaUx should works as before, but apiserver pod is crashed by null point exception.
➜ ~ kubectl -n vela-system logs -f apiserver-7dc97f7558-kqs92
{"level":"info","ts":1641302139.5081124,"caller":"apiserver/main.go:108","msg":"KubeVela information: version: undefined, gitRevision: undefined"}
I0104 13:15:40.315824 1 utils.go:143] find cluster gateway service vela-system/kubevela-cluster-gateway-service:9443
{"level":"info","ts":1641302140.3392406,"caller":"rest/rest_server.go:251","msg":"HTTP APIs are being served on: <IP_ADDRESS>:8000, ctx: context.Background.WithCancel"}
I0104 13:15:40.339524 1 leaderelection.go:248] attempting to acquire leader lease vela-system/apiserver-lock...
I0104 13:15:40.342548 1 rest_server.go:150] new leader elected: f28b817d-f8e7-45a3-b713-dbff80c9f15e
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x18 pc=0x22b1528]
goroutine 231 [running]:
github.com/oam-dev/kubevela/pkg/addon.(*Registry).ListAddonMeta(0xc00071df20, 0xc000e13330, 0x8, 0xc000e90dc0)
/workspace/pkg/addon/source.go:212 +0x68
github.com/oam-dev/kubevela/pkg/addon.(*Cache).discoverAndRefreshRegistry(0xc0003c4330)
/workspace/pkg/addon/cache.go:213 +0x127
github.com/oam-dev/kubevela/pkg/addon.(*Cache).DiscoverAndRefreshLoop(0xc0003c4330, 0x8bb2c97000)
/workspace/pkg/addon/cache.go:69 +0x6a
created by github.com/oam-dev/kubevela/pkg/apiserver/rest/usecase.NewAddonUsecase
/workspace/pkg/apiserver/rest/usecase/addon.go:104 +0x1b7
Screenshots
KubeVela Version
v1.2.0-RC2
Cluster information
Additional context
@StevenLeiZhang Thanks for your report.This bug has been fixed by #3026 by test. Please check again with the code with master branch or next 1.20 release . Let me close this issue firstly, feel free we can reopen this issue if the bug still exist.
I have question about this
fmt.Errorf("git type repository only support github for now")
why is Vela only suport github? I want to use one local private gitlab, this private gitlab also suport github openapi.
@StevenLeiZhang yes, we're planing to support gitlab, but we only tested github currently. You're very welcome to contribute if the gitlab is tested well. Thanks!
|
2025-04-01T06:39:49.362200
| 2019-03-29T14:08:47
|
427004624
|
{
"authors": [
"asilveir",
"centos-ci",
"pirat89"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9074",
"repo": "oamg/leapp",
"url": "https://github.com/oamg/leapp/pull/471"
}
|
gharchive/pull-request
|
WIP: Show information in debug mode about external commands
As in debug mode we have a lot of messages the user needs to be informed which external command has generated certain output.
This PR wants to solve this by adding this output:
leapp.stdlib.Run: External command is started: [/usr/bin/dnf rhel-upgrade]
*** external command output ***
leapp.stdlib.Run: External command is finished: [/usr/bin/dnf rhel-upgrade]
Can one of the admins verify this patch?
I afraid little about the readability. What about reword the solution like:
====== ext cmd started:
*** external command output ***
====== ext cmd finished:
Just idea that it is easy to find substring like =====. What do you think guys? @oamg/developers
|
2025-04-01T06:39:49.382864
| 2023-12-25T06:38:32
|
2055433608
|
{
"authors": [
"lgbo-ustc",
"zhanglistar"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9075",
"repo": "oap-project/gluten",
"url": "https://github.com/oap-project/gluten/pull/4179"
}
|
gharchive/pull-request
|
[GLUTEN-4178][CH] Reduce memory usage in aggregate operators
What changes were proposed in this pull request?
(Please fill in changes proposed in this fix)
Fixes: #4178
At previous version, we convert a hash table into a block list at once,bring a double memory peak usage. It's easy to cause OOM problems.
At this version, we try to convert the buckets in a two-level hash table one by one, and the bucket in the hash table is released immediately after it have been converted. This make memory peak usage more smoothly and smaller.
How was this patch tested?
(Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests)
unit tests
(If this patch involves UI changes, please attach a screenshot; otherwise, remove this)
Add some comments on key classes and functions.
|
2025-04-01T06:39:49.384096
| 2024-10-28T18:33:47
|
2619209892
|
{
"authors": [
"cosban"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9076",
"repo": "oapi-codegen/oapi-codegen",
"url": "https://github.com/oapi-codegen/oapi-codegen/pull/1815"
}
|
gharchive/pull-request
|
add literal colon support for gin and echo
Closes https://github.com/oapi-codegen/oapi-codegen/issues/1726
I see the test failures and will get to them in a bit.
|
2025-04-01T06:39:49.391366
| 2018-08-03T16:54:01
|
347467671
|
{
"authors": [
"clenk",
"rpiazza"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9077",
"repo": "oasis-open/cti-python-stix2",
"url": "https://github.com/oasis-open/cti-python-stix2/issues/205"
}
|
gharchive/issue
|
Need to "export" BooleanConstant
in the init.py file
Is this still a problem in the latest release? It looks like we already do this:
https://github.com/oasis-open/cti-python-stix2/blob/master/stix2/__init__.py#L36
|
2025-04-01T06:39:49.396927
| 2024-07-29T07:31:59
|
2434675978
|
{
"authors": [
"KnightChaser",
"chisholm"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9078",
"repo": "oasis-open/cti-stix-generator",
"url": "https://github.com/oasis-open/cti-stix-generator/issues/55"
}
|
gharchive/issue
|
Is there a concise way to directly export generated STIX data as JSON file?
Well, this package is straightforward at generating a random STIX data and I'm satisfied at this level of convenience. However, when I try to load this as JSON data and save it to the data like the code below,
import stix2generator
import json
stix2_generator = stix2generator.create_stix_generator()
stix = stix2_generator.generate()
stix_dump = json.dumps(stix, indent=4)
I get the errors below because some STIX SDO objects are not python default data types. (For example, the runtime exception log below says that the generated STIX data can't be converted to the JSON format since Note which is one of STIX SDO is not compatible.)
Traceback (most recent call last):
File "C:\Users\user\Documents\GitHub\nis-ems-client\test\fake_stix2_generator.py", line 6, in <module>
stix_dump = json.dumps(stix, indent=4)
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\user\.pyenv\pyenv-win\versions\3.11.9\Lib\json\__init__.py", line 238, in dumps
**kw).encode(obj)
^^^^^^^^^^^
File "C:\Users\user\.pyenv\pyenv-win\versions\3.11.9\Lib\json\encoder.py", line 202, in encode
chunks = list(chunks)
^^^^^^^^^^^^
File "C:\Users\user\.pyenv\pyenv-win\versions\3.11.9\Lib\json\encoder.py", line 432, in _iterencode
yield from _iterencode_dict(o, _current_indent_level)
File "C:\Users\user\.pyenv\pyenv-win\versions\3.11.9\Lib\json\encoder.py", line 406, in _iterencode_dict
yield from chunks
File "C:\Users\user\.pyenv\pyenv-win\versions\3.11.9\Lib\json\encoder.py", line 439, in _iterencode
o = _default(o)
^^^^^^^^^^^
File "C:\Users\user\.pyenv\pyenv-win\versions\3.11.9\Lib\json\encoder.py", line 180, in default
raise TypeError(f'Object of type {o.__class__.__name__} '
TypeError: Object of type Note is not JSON serializable
So, my question is, is there a way to cope with this issue? I'm almost new to STIX, I just read the STIX documentation from the official page today and I am about to use STIX data with my project. Thanks in advance.
It's possible to simply(maybe) parse the STIX2 generation result as a JSON dump with Python's JSON package(json) with STIX2's JSON serialization class object(stix2.serialization.STIXJSONEncoder). :)
Here I'll attach a working makeshift example
import stix2generator
import stix2validator
import stix2
import json
import uuid
from rich import print
# Generate the fake STIX2 data
stix2_generator = stix2generator.create_stix_generator()
stix = stix2_generator.generate()
# Convert the STIX data (Dictionary with STIX-proprietary objects) to pure JSON objects
def stix_to_json(stix_data):
stix_objects = []
for value in stix_data.values():
stix_objects.append(json.loads(value.serialize()))
return stix_objects
stix_objects = stix_to_json(stix)
# Create a STIX bundle with the generated objects
stix_bundle = {
"type": "bundle",
"id": f"bundle--{uuid.uuid4()}",
"objects": stix_objects
}
# Convert the STIX bundle to JSON
stix_bundle_json = json.dumps(stix_bundle, cls=stix2.serialization.STIXJSONEncoder, indent=4).encode('utf-8').decode('utf-8')
print(stix_bundle_json)
# Validate the generated STIX bundle
result = stix2validator.validate_string(str(stix_bundle_json))
stix2validator.print_results(result)
Not sure what you mean by repositioning the id property. Abstractly speaking, mappings/objects don't have an intrinsic order to their entries (some implementations can allow you to control some kinds of ordering, e.g. a traversal order, but that is an implementation detail). I don't think JSON-Schema allows you to express a required ordering on object entries (correct me if I am wrong). So, it seems to me an object entry ordering difference cannot cause a JSON-Schema validation failure.
A simpler way to generate and validate is (but does not directly use the json package):
import io
import sys
# ... and all the other imports
stix = stix2_generator.generate()
bundle = stix2.Bundle(list(stix.values()))
# A simple way to dump to stdout:
# (don't use pretty=True for large numbers of objects!)
bundle.fp_serialize(sys.stdout, pretty=True)
# A simple way to dump JSON to a memory text buffer
buf = io.StringIO()
bundle.fp_serialize(buf)
# Reposition for reading, and validate
buf.seek(0, io.SEEK_SET)
result = stix2validator.validate(buf)
stix2validator.print_results(result)
Or via shell:
# -b causes output to be wrapped in a bundle
generate_stix -b > ./bundle.json
stix2_validator ./bundle.json
|
2025-04-01T06:39:49.413273
| 2024-07-22T01:22:11
|
2421714526
|
{
"authors": [
"dteleguin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9079",
"repo": "oauth-wg/oauth-transaction-tokens",
"url": "https://github.com/oauth-wg/oauth-transaction-tokens/issues/113"
}
|
gharchive/issue
|
subject_token_type for Replacement Txn-Token Request
In 7.5.2:
To request a replacement Txn-Token, the requester makes a Txn-Token Request as described in Section 7.1 but includes the Txn-Token to be replaced as the value of the subject_token parameter.
Does this assume that subject_token_type should be urn:ietf:params:oauth:token-type:txn_token? Should we call it out explicitly?
Also, do other parameters (audience, scope, request_context) make sense in the context of TraT replacement flow? (I believe they don't, as they are meant to remain constant for the whole invocation chain.)
Should we call out that they must be ignored in the replacement flow?
|
2025-04-01T06:39:49.419685
| 2022-06-08T04:42:37
|
1264186168
|
{
"authors": [
"codecov-commenter",
"stanhu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9080",
"repo": "oauth-xx/oauth2",
"url": "https://github.com/oauth-xx/oauth2/pull/590"
}
|
gharchive/pull-request
|
Drop multi_json dependency
As discussed in https://github.com/oauth-xx/oauth2/issues/579#issuecomment-1084174454, the Ruby JSON stdlib module should be good enough.
Closes #579
Codecov Report
Merging #590 (08f7c75) into master (7e5cd6c) will increase coverage by 0.04%.
The diff coverage is 100.00%.
@@ Coverage Diff @@
## master #590 +/- ##
==========================================
+ Coverage 89.66% 89.70% +0.04%
==========================================
Files 15 15
Lines 445 447 +2
==========================================
+ Hits 399 401 +2
Misses 46 46
Impacted Files
Coverage Δ
lib/oauth2/response.rb
100.00% <100.00%> (ø)
:mega: Codecov can now indicate which changes are the most critical in Pull Requests. Learn more
Code coverage failure: SimpleCov failed with exit 2 due to a coverage related error
Rubocop failure: Will be fixed in 16345ec1e12c348fc752b851173c22a8c7b68a8f via https://github.com/oauth-xx/oauth2/pull/589
|
2025-04-01T06:39:49.481352
| 2015-03-24T18:19:51
|
64065366
|
{
"authors": [
"djwatson82",
"oblador"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9081",
"repo": "oblador/angular-scroll",
"url": "https://github.com/oblador/angular-scroll/issues/113"
}
|
gharchive/issue
|
Memory Leak in duScrollspy directive
In src/directives/scrollspy.js line number 58. Listener is being added to $rootScope, but is not being cleaned up when the directive is destroyed.
Well spotted, thanks!
|
2025-04-01T06:39:49.485127
| 2016-08-02T13:33:31
|
168886081
|
{
"authors": [
"Burak07",
"iRoachie"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9082",
"repo": "oblador/react-native-collapsible",
"url": "https://github.com/oblador/react-native-collapsible/issues/43"
}
|
gharchive/issue
|
How to handle dynamic accordion content
Thanks for this useful component,
I'm having a trouble while trying to create an accordion.
When user clicks a submenu under a menu, content of the accordion is changed and it is rerendered with a different content.
This is how i use activeSection in accordion
activeSection={this.state.activeSection}
In my header content every view's structure is
View
View (image)
View (text)
View (image)
View
assuming that if i have 4 rows in old content and i click 2nd item which is index 1 in new content 2nd head (index1)view's text is changed but images are not changed. I'm having this issue only with iOS. Code works well with Android images are changed properly.
note: images's sources and logs controlled several times.
Hey, this issue has been inactive for long time and will be closed. If the issue still persists feel free to tag me to reopen.
|
2025-04-01T06:39:49.486154
| 2017-12-27T16:22:00
|
284753459
|
{
"authors": [
"amirhosseinnazari",
"solsticedhiver"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9083",
"repo": "oblique/create_ap",
"url": "https://github.com/oblique/create_ap/issues/306"
}
|
gharchive/issue
|
DNS not resolve on macOs (high sierra)
we run this script with Additional hosts file(-e parameter) and additional hosts resolves in windows, linux, android but not resolve in macOS
I can not understand what you are trying to say. Could you please rephrase it and add more details so that we can understand you ?
|
2025-04-01T06:39:49.487675
| 2021-09-29T14:53:43
|
1011067710
|
{
"authors": [
"anitacaron",
"dosumis"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9084",
"repo": "obophenotype/CCF_tools",
"url": "https://github.com/obophenotype/CCF_tools/pull/25"
}
|
gharchive/pull-request
|
Data release OWL and Log
Please, merge first the PR #24.
I also deleted the log files from the previous PR #18 to not have the same file in different names and formats.
Fix #22
Looks good, but I think logs are in the wrong place. We should have a dedicated directory for these. We should also capture STDERR logs to files in the same directory as this captures other errors (we should find better ways to communicate these errors in future, but raw logs will do for now.
Ok, I'll create a dedicated directory for logs in this PR.
|
2025-04-01T06:39:49.492097
| 2021-02-09T20:03:37
|
804877852
|
{
"authors": [
"addiehl",
"dosumis"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9085",
"repo": "obophenotype/cell-ontology",
"url": "https://github.com/obophenotype/cell-ontology/pull/939"
}
|
gharchive/pull-request
|
B cell issues Fixes #935
Removes duplicate label and extra definition for 'B cell' and extra obo_namespace stanza.
OWL axiom in PR import!
|
2025-04-01T06:39:49.511505
| 2022-05-01T00:48:28
|
1221962873
|
{
"authors": [
"mbostock",
"reubano"
],
"license": "ISC",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9086",
"repo": "observablehq/plot",
"url": "https://github.com/observablehq/plot/issues/859"
}
|
gharchive/issue
|
Time scale labels don't adjust to font size.
As seen below, increasing the font size makes the time scale labels overlap.
style: {
fontFamily: "system-ui",
fontSize: "16px",
overflow: "visible"
},
style: {
fontFamily: "system-ui",
fontSize: "24px",
overflow: "visible"
},
Yes, that’s how it works. Plot does not consider text metrics when laying out axes.
So what's the work-a-round for this?
You can specify a different number of ticks using the scale.ticks option (e.g., x: {ticks: 4}) or you can rotate the ticks using scale.tickRotate option (e.g., x: {tickRotate: 90}) and then increasing marginBottom.
var graph = Plot.plot({
x: {
ticks: 4,
tickRotate: -45,
},
width: 640,
height: 400,
marginLeft: 100,
marginTop: 105,
marginBottom: 100,
});
|
2025-04-01T06:39:49.530101
| 2022-09-29T22:36:56
|
1391581028
|
{
"authors": [
"claremacrae",
"joetifa2003"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9087",
"repo": "obsidian-tasks-group/obsidian-tasks",
"url": "https://github.com/obsidian-tasks-group/obsidian-tasks/issues/1196"
}
|
gharchive/issue
|
Append completion date to tasks completed in the Kanban plugin
Expected Behavior
When marking a task in kanban as done it should add the ✅ emoji with the date
Current behaviour
It doesn;t add the ✅emoji if i mark the task as done in the kanban but if i opened the board as markdown and click manually it'll add it like normal
Steps to reproduce
Install kanban and make a new board
Add tasks and mark it as done, the ✅ emoji with the date will not show up
Which Operating Systems are you using?
[ ] Android
[ ] iPhone/iPad
[ ] Linux
[ ] macOS
[X] Windows
Obsidian Version
0.15.9
Tasks Plugin Version
1.11.0
Checks
[x] I have tried it with all other plugins disabled and the error still occurs
Possible solution
No response
Hi @joetifa2003, thanks for the write up.
I’m not sure that I understand - it sounds like you would like the KanBan plugin to provide support for Tasks emojis, is that correct?
The done date is would be really usefull if it automatically added but that's ok i'll find a workaround for now, thanks!
Just to follow up from the Discussions above, when a task is marked as done in another plugin, any addition of Due date or similar would be done by code in that plugin.
So it's not something within the scope of Tasks to implement.
|
2025-04-01T06:39:49.531144
| 2018-11-25T14:58:21
|
384086616
|
{
"authors": [
"luigy",
"mightybyte"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9088",
"repo": "obsidiansystems/obelisk",
"url": "https://github.com/obsidiansystems/obelisk/pull/321"
}
|
gharchive/pull-request
|
Add a FAQ to document common problems
I encountered this entitlements error and it took me awhile to find the problem. I figured this would be a good candidate for starting a FAQ.
Thanks!
|
2025-04-01T06:39:49.633501
| 2024-09-01T17:14:04
|
2499588303
|
{
"authors": [
"vthriller"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:9089",
"repo": "ocavue/prosemirror-virtual-cursor",
"url": "https://github.com/ocavue/prosemirror-virtual-cursor/issues/47"
}
|
gharchive/issue
|
incorrect cursor position and inconsistent scrolling behaviour in editor with overflow
Exhibit A
.ProseMirror {
word-wrap: normal;
overflow: auto;
}
abcdefghijklmnopqrstuvwxyzabcdefghijklmnopqrstuvwxyzabcdefghijklmnopqrstuvwxyzabcdefghijklmnopqrstuvwxyzabcdefghijklmnopqrstuvwxyzabcdefghijklmnopqrstuvwxyzabcdefghijklmnopqrstuvwxyzabcdefghijklmnopqrstuvwxyz
cursor in "abc┘def", scrolled outside of viewport
press →
cursor is not scrolled into view
press →
viewport is scrolled to the cursor position ("efghi..."), but no cursor visible
manually scroll to the left
cursor is blinking at the very edge of the element, in the padding area:
Exhibit B
.ProseMirror {
overflow: auto;
}
.ProseMirror pre {
white-space: pre;
}
console.warn(
`[prosemirror-virtual-cursor] Virtual cursor does not work well with marks that have inclusive set to false. Please consider removing the inclusive option from the "${mark}" mark or adding it to the "skipWarning" option.`,
);
place cursor anywhere in the code block
scroll horizontally
use arrow keys or mouse to change cursor position
cursor is rendered with offset (verify by typing in text)
Well, it is affected by vertical overflow as well.
.ProseMirror {
max-height: 10em;
overflow: auto;
}
lorem
ipsum
dolor
sit
amet
Cursor on the last line is rendered higher than it should be:
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.