added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:39:43.772229
2015-09-21T11:53:13
107495250
{ "authors": [ "KatharinaSt", "kdambekalns", "kitsunet" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8873", "repo": "neos/form", "url": "https://github.com/neos/form/issues/3" }
gharchive/issue
Fatal error: Class 'TYPO3\SwiftMailer\Message' not found This line in the EmailFinisher class does not ensure that the required Swiftmailer class is actually available (and causes a fatal error if you don't have Swiftmailer installed): $mail = new \TYPO3\SwiftMailer\Message(); I suggest to either include Swiftmailer as dependency in composer.json or to check for installed Swiftmailer and throw an informative exception instead. Thanks, good point! I guess we should at least suggest the switftmailer package if we don't want to create a hard dependency. Fixed with #23
2025-04-01T06:39:43.798150
2016-09-20T16:20:54
178107977
{ "authors": [ "aertmann", "albe", "bwaidelich" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8874", "repo": "neos/setup", "url": "https://github.com/neos/setup/pull/13" }
gharchive/pull-request
BUGFIX: Prevent Fatal Error in Setup This is a hotfix for a regression introduced with https://github.com/neos/flow-development-collection/pull/499. Background: The latest Http\RequestHandler of Flow internally uses the ComponentChain to determine the current Request/Response. This package uses a specialized handler which doesn't rely on the ComponentChain leading to a Fatal Error. needs to go into the 2.0 branch right? needs to go into the 2.0 branch right? I'm not sure, 2.0 is used in Neos 2.0+ and the fix is only required for Neos 2.3.x. I guess we should probably create a 2.3 branch in the setup and apply it there!? The evils of inheritance... Makes sense, but be aware that the $request and $response properties will also be removed with 4.0 - so maybe already copy them over be aware that the $request and $response properties will also be removed with 4.0 That's news to me. Is that documented somewhere or is there a PR for this? Anyways, I rather merge this now to fix the fatal. I'd suggest to merge it into master and then create a branch 2.3 and require that in the affected distributions. Does that make sense? Is that documented somewhere or is there a PR for this? https://github.com/neos/flow-development-collection/pull/500 @bwaidelich: I'd rather avoid a separate version if it can be done in backwards compatible way, which this seems to be.. I messed up the rebase, see #28
2025-04-01T06:39:43.948983
2021-12-17T03:50:42
1082856868
{ "authors": [ "jose-elias-alvarez", "kylo252", "mjlbach" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8875", "repo": "neovim/nvim-lspconfig", "url": "https://github.com/neovim/nvim-lspconfig/issues/1582" }
gharchive/issue
:LspRestart throws error when using non-supported language server Description When called without arguments in a buffer attached to a non-supported language server, :LspRestart throws the following error: Error executing vim.schedule lua callback: .../site/pack/packer/start/nvim-lspconfig/lua/lspconfig.lua:58: attempt to index a nil value stack traceback: .../site/pack/packer/start/nvim-lspconfig/lua/lspconfig.lua:58: in function '' vim.lua: in function '' vim.lua: in function <vim.lua:0> Neovim version NVIM v0.7.0-dev+732-g80210c189 Nvim-lspconfig version 64bd3b95345f2dbedc83a3d9a472662c48b3e940 Operating system and version macOS 12.1 Affected language servers All Steps to reproduce nvim --clean -u minimal_init.lua test.py :LspRestart Actual behavior The error from the description is thrown. Expected behavior No error is thrown. Minimal config local on_windows = vim.loop.os_uname().version:match("Windows") local function join_paths(...) local path_sep = on_windows and "\\" or "/" local result = table.concat({ ... }, path_sep) return result end vim.cmd([[set runtimepath=$VIMRUNTIME]]) local temp_dir if on_windows then temp_dir = vim.loop.os_getenv("TEMP") else temp_dir = "/tmp" end vim.cmd("set packpath=" .. join_paths(temp_dir, "nvim", "site")) local package_root = join_paths(temp_dir, "nvim", "site", "pack") local install_path = join_paths(package_root, "packer", "start", "packer.nvim") local compile_path = join_paths(install_path, "plugin", "packer_compiled.lua") local function load_plugins() require("packer").startup({ { "wbthomason/packer.nvim", "neovim/nvim-lspconfig", }, config = { package_root = package_root, compile_path = compile_path, }, }) end _G.attach_buffer = function() local client_id = vim.lsp.start_client({ cmd = { "pylsp" } }) -- cmd is irrelevant vim.lsp.buf_attach_client(vim.api.nvim_get_current_buf(), client_id) end _G.load_config = function() -- filetype is irrelevant but should be consistent with language server vim.cmd("autocmd FileType python lua attach_buffer()") end if vim.fn.isdirectory(install_path) == 0 then vim.fn.system({ "git", "clone", "https://github.com/wbthomason/packer.nvim", install_path }) load_plugins() require("packer").sync() vim.cmd([[autocmd User PackerComplete ++once lua load_config()]]) else load_plugins() require("packer").sync() _G.load_config() end Health check lspconfig: require("lspconfig.health").check() ======================================================================== - WARNING: Can't find any config. LSP log Probably not relevant The following trivial patch fixes the issue as well as a related issue where a config is defined but setup has not yet been called, meaning that launch is not yet defined. I didn't yet submit a PR in case I'm missing something but am happy to do so if this is the right fix. diff --git a/lua/lspconfig.lua b/lua/lspconfig.lua index 8404632..685b166 100644 --- a/lua/lspconfig.lua +++ b/lua/lspconfig.lua @@ -54,9 +54,11 @@ function M._root._setup() function(cmd_args) for _, client in ipairs(M.util.get_clients_from_cmd_args(cmd_args)) do client.stop() - vim.defer_fn(function() - configs[client.name].launch() - end, 500) + if configs[client.name] and configs[client.name].launch then + vim.defer_fn(function() + configs[client.name].launch() + end, 500) + end end end, '-nargs=? -complete=customlist,v:lua.lsp_get_active_client_ids', I think a cleaner "fix" would be to expose the table of client ids managed by lspconfig, and make these methods only restart those clients. Sure, that makes sense. Does this table already exist somewhere in the code, or is it something that would need to be created? I think a cleaner "fix" would be to expose the table of client ids managed by lspconfig, and make these methods only restart those clients. How about using the docs tag? or maybe add a special marker in the client config. Otherwise, how would you know which ones are managed? just because tsserver is active, doesn't mean it's managed by lspconfig. Because lspconfig tracks internally which clients it starts... Please read the code and evaluate the solution I suggested before theorycrafting alternatives. I took another look, and I can see that the active client IDs are available via manager.clients(). Does something like this make sense? local get_lspconfig_clients = function() local clients = {} for _, config in pairs(require("lspconfig.configs")) do if config.manager then vim.list_extend(clients, config.manager.clients()) end end return clients end If so, using that here instead of vim.lsp.get_active_clients() would solve this specific issue. Yup
2025-04-01T06:39:43.974423
2015-12-20T18:43:34
123177028
{ "authors": [ "soygul" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8876", "repo": "neptulon/neptulon", "url": "https://github.com/neptulon/neptulon/issues/49" }
gharchive/issue
WebSocket support Create interfaces for client/Client and neptulon/Listener objects and separate implementation as: client/TCPClient & client/WSClient neptulon/TCPListener & neptulon/WSListener We can even skip this step and keep everything all in one if things are as easy as what we did with ListenTLS. We could also change project description after this to: TCP/WebSocket framework for Go. Minimalist web framework for Go. Done.
2025-04-01T06:39:43.982032
2021-03-24T11:13:28
839614666
{ "authors": [ "alibner", "insectme", "thomaspeitz" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8877", "repo": "nerdswords/yet-another-cloudwatch-exporter", "url": "https://github.com/nerdswords/yet-another-cloudwatch-exporter/issues/344" }
gharchive/issue
Fix on China AWS support if run Yace on china AWS, there is error “Couldn't describe resources for region cn-northwest-1: InvalidClientTokenId: The security token included in the request is invalid.\n\tstatus code: 403,” and china AWS is in seperate infra. Session new need to specified region, we fixed code based on Release 0.26.3-alpha, it has been tested successfully in China. and in aws_cloudwatch.go ln50: func createStsSession(roleArn string) this part of code seems new adding, and region is unnecessary here. It has been tested. Was fixed and merged the same issue ; {"arn":"","error":"InvalidClientTokenId: The security token included in the request is invalid.\n\tstatus code: 403, request id: 46cbd89f-1a5d-49b4-996f-c2b616f4479c","job_type":"alb","level":"error","msg":"Couldn't get account Id","region":"cn-northwest-1","time":"2023-10-24T11:43:05Z"}.
2025-04-01T06:39:44.010334
2022-12-27T00:39:30
1511329507
{ "authors": [ "THE-COB", "brawlawnzo", "dooonnis", "grshennan", "tancik" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8878", "repo": "nerfstudio-project/nerfstudio", "url": "https://github.com/nerfstudio-project/nerfstudio/issues/1177" }
gharchive/issue
OSError: Unknown compute capability. Ensure PyTorch with CUDA support is installed. Installed nerfstudio with CUDA 11.3 as per setup page. Everything works fine. Followed the tutorial all the way to ns-train nerfacto --data data/nerfstudio/poster where I get this error: $ ns-train nerfacto --data data/nerfstudio/poster Traceback (most recent call last): File "C:\Users\glens\miniconda3\envs\ns\lib\runpy.py", line 194, in _run_module_as_main return _run_code(code, main_globals, None, File "C:\Users\glens\miniconda3\envs\ns\lib\runpy.py", line 87, in _run_code exec(code, run_globals) File "C:\Users\glens\miniconda3\envs\ns\Scripts\ns-train.exe\__main__.py", line 4, in <module> File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\scripts\train.py", line 50, in <module> from nerfstudio.configs.method_configs import AnnotatedBaseConfigUnion File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\nerfstudio\configs\method_configs.py", line 46, in <module> from nerfstudio.field_components.temporal_distortions import TemporalDistortionKind File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\nerfstudio\field_components\__init__.py", line 17, in <module> from .encodings import Encoding, ScalingAndOffset File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\nerfstudio\field_components\encodings.py", line 34, in <module> import tinycudann as tcnn File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\tinycudann\__init__.py", line 9, in <module> from tinycudann.modules import free_temporary_memory, NetworkWithInputEncoding, Network, Encoding File "C:\Users\glens\miniconda3\envs\ns\lib\site-packages\tinycudann\modules.py", line 16, in <module> raise EnvironmentError("Unknown compute capability. Ensure PyTorch with CUDA support is installed.") OSError: Unknown compute capability. Ensure PyTorch with CUDA support is installed. Removed CUDA 11.3, installed 11.7 and followed the install instructions on the Pytorch page, i.e. conda install pytorch torchvision torchaudio pytorch-cuda=11.7 -c pytorch -c nvidia. Everything works as per the pip install instructions. tinycudann even seemed to build a lot faster. Ended up with exactly the same error: OSError: Unknown compute capability. Ensure PyTorch with CUDA support is installed. CUDA variables are set as (not sure if this is relevant): I understand devs are busy and I don't want to bother anyone but I'm trying really hard to get this working and just keep getting pipped at the post. Let me know if there's a better forum. I ran into the same exact error today with conda installation of PyTorch with CUDA 11.7 (11.3 worked fine for me though). Seems like a problem with tinycuda rather than nerfstudio. I'm also using CUDA 11.3, did the tinycuda installation go entirely smoothly? @THE-COB I have both had tinycudann installation go smoothly and not work at all. I think @brawlawnzo is right, it's an issue with tinycuda. I just tried to install as per nerfstudio installation instructions on my work computer. Visual Studio 2019, CUDA 11.3, Python 3.8. Completely clean install. Got this error install tinycudann: pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch Collecting git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch Cloning https://github.com/NVlabs/tiny-cuda-nn/ to c:\users\glens\appdata\local\temp\pip-req-build-s1rw0r85 Running command git clone --filter=blob:none --quiet https://github.com/NVlabs/tiny-cuda-nn/ 'C:\Users\glens\AppData\Local\Temp\pip-req-build-s1rw0r85' Resolved https://github.com/NVlabs/tiny-cuda-nn/ to commit 14053e9a87ebf449d32bda335c0363dd4f5667a4 Running command git submodule update --init --recursive -q Preparing metadata (setup.py) ... done Building wheels for collected packages: tinycudann Building wheel for tinycudann (setup.py) ... error error: subprocess-exited-with-error × python setup.py bdist_wheel did not run successfully. │ exit code: 1 ╰─> [1327 lines of output] Building PyTorch extension for tiny-cuda-nn version 1.7 Obtained compute capability 86 from PyTorch running bdist_wheel C:\Users\glens\miniconda3\envs\ns\lib\site-packages\torch\utils\cpp_extension.py:411: UserWarning: Attempted to use ninja as the BuildExtension backend but we could not find ninja.. Falling back to using the slow distutils backend. warnings.warn(msg.format('we could not find ninja.')) running build running build_py creating build creating build\lib.win-amd64-cpython-38 creating build\lib.win-amd64-cpython-38\tinycudann copying tinycudann\modules.py -> build\lib.win-amd64-cpython-38\tinycudann copying tinycudann\__init__.py -> build\lib.win-amd64-cpython-38\tinycudann running egg_info creating tinycudann.egg-info writing tinycudann.egg-info\PKG-INFO writing dependency_links to tinycudann.egg-info\dependency_links.txt writing top-level names to tinycudann.egg-info\top_level.txt writing manifest file 'tinycudann.egg-info\SOURCES.txt' reading manifest file 'tinycudann.egg-info\SOURCES.txt' writing manifest file 'tinycudann.egg-info\SOURCES.txt' copying tinycudann\bindings.cpp -> build\lib.win-amd64-cpython-38\tinycudann running build_ext building 'tinycudann_bindings_86._C' extension creating build\dependencies creating build\dependencies\fmt creating build\dependencies\fmt\src creating build\src creating build\temp.win-amd64-cpython-38 creating build\temp.win-amd64-cpython-38\Release creating build\temp.win-amd64-cpython-38\Release\tinycudann "C:\Program Files (x86)\Microsoft Visual Studio\2019\Professional\VC\Tools\MSVC\14.29.30133\bin\HostX86\x64\cl.exe" /c /nologo /O2 /W3 /GL /DNDEBUG /MD -IC:\Users\gle... followed by thousands of lines of errors, and then: Error limit reached. 100 errors detected in the compilation of "../../src/common.cu". Compilation terminated. common.cu error: command 'C:/Program Files/NVIDIA GPU Computing Toolkit/CUDA\\v11.3\\bin\\nvcc.exe' failed with exit code 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: legacy-install-failure × Encountered error while trying to install package. ╰─> tinycudann @grshennan I think I might have run into this as well! It seems to resemble the issue in here: https://github.com/NVlabs/tiny-cuda-nn/issues/169 One of the things in that thread that I think got me passed that (and then to the next error lol) was conda installing ninja. If that doesn't work maybe try reinstalling visual studio (2019 should be alright, but I'm currently using 2017) @THE-COB I have VS 2019 installed but same idea, yes. I think it's the same as this issue: https://github.com/NVlabs/tiny-cuda-nn/issues/147 Tried installing ninja and got a different error. Now it's running build_ext error: [WinError 2] The system cannot find the file specified as per the tinycudann issue. What is your current error? @grshennan $ pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch Collecting git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch Cloning https://github.com/NVlabs/tiny-cuda-nn/ to c:\users\glens\appdata\local\temp\pip-req-build-6w2nnebl Running command git clone --filter=blob:none --quiet https://github.com/NVlabs/tiny-cuda-nn/ 'C:\Users\glens\AppData\Local\Temp\pip-req-build-6w2nnebl' Resolved https://github.com/NVlabs/tiny-cuda-nn/ to commit 14053e9a87ebf449d32bda335c0363dd4f5667a4 Running command git submodule update --init --recursive -q Preparing metadata (setup.py) ... done Building wheels for collected packages: tinycudann Building wheel for tinycudann (setup.py) ... error error: subprocess-exited-with-error × python setup.py bdist_wheel did not run successfully. │ exit code: 1 ╰─> [21 lines of output] Building PyTorch extension for tiny-cuda-nn version 1.7 Obtained compute capability 86 from PyTorch running bdist_wheel running build running build_py creating build creating build\lib.win-amd64-cpython-38 creating build\lib.win-amd64-cpython-38\tinycudann copying tinycudann\modules.py -> build\lib.win-amd64-cpython-38\tinycudann copying tinycudann\__init__.py -> build\lib.win-amd64-cpython-38\tinycudann running egg_info creating tinycudann.egg-info writing tinycudann.egg-info\PKG-INFO writing dependency_links to tinycudann.egg-info\dependency_links.txt writing top-level names to tinycudann.egg-info\top_level.txt writing manifest file 'tinycudann.egg-info\SOURCES.txt' reading manifest file 'tinycudann.egg-info\SOURCES.txt' writing manifest file 'tinycudann.egg-info\SOURCES.txt' copying tinycudann\bindings.cpp -> build\lib.win-amd64-cpython-38\tinycudann running build_ext error: [WinError 2] The system cannot find the file specified [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. ERROR: Failed building wheel for tinycudann Running setup.py clean for tinycudann Failed to build tinycudann Installing collected packages: tinycudann Running setup.py install for tinycudann ... error error: subprocess-exited-with-error × Running setup.py install for tinycudann did not run successfully. │ exit code: 1 ╰─> [15 lines of output] Building PyTorch extension for tiny-cuda-nn version 1.7 Obtained compute capability 86 from PyTorch running install C:\Users\glens\miniconda3\envs\ns\lib\site-packages\setuptools\command\install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools. warnings.warn( running build running build_py running egg_info writing tinycudann.egg-info\PKG-INFO writing dependency_links to tinycudann.egg-info\dependency_links.txt writing top-level names to tinycudann.egg-info\top_level.txt reading manifest file 'tinycudann.egg-info\SOURCES.txt' writing manifest file 'tinycudann.egg-info\SOURCES.txt' running build_ext error: [WinError 2] The system cannot find the file specified [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: legacy-install-failure × Encountered error while trying to install package. ╰─> tinycudann note: This is an issue with the package mentioned above, not pip. hint: See above for output from the failure. @grshennan that is tough, I think I might have run into something similar but I don't remember. If you want (this may or may not be worth it) try building tinycudann in the way suggested in this comment? https://github.com/NVlabs/tiny-cuda-nn/issues/169#issuecomment-1365984644 Sorry I wasn't able to give you concrete help I'm fairly new to this topic. Would you explain, where you have placed the tiny-cuda-nn files and in which directory you have started the cmake process? Is that necessary for everything to work? I am receiving this error: Traceback (most recent call last): File "F:\Programme\miniconda3\envs\nerfstudio\lib\runpy.py", line 194, in _run_module_as_main return _run_code(code, main_globals, None, File "F:\Programme\miniconda3\envs\nerfstudio\lib\runpy.py", line 87, in _run_code exec(code, run_globals) File "F:\Programme\miniconda3\envs\nerfstudio\Scripts\ns-train.exe\__main__.py", line 7, in <module> File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\scripts\train.py", line 247, in entrypoint main( File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\scripts\train.py", line 233, in main launch( File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\scripts\train.py", line 172, in launch main_func(local_rank=0, world_size=world_size, config=config) File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\scripts\train.py", line 86, in train_loop trainer.setup() File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\engine\trainer.py", line 145, in setup self.pipeline = self.config.pipeline.setup( File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\configs\base_config.py", line 57, in setup return self._target(self, **kwargs) File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\pipelines\base_pipeline.py", line 229, in __init__ self._model = config.model.setup( File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\configs\base_config.py", line 57, in setup return self._target(self, **kwargs) File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\models\base_model.py", line 82, in __init__ self.populate_modules() # populate the modules File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\models\nerfacto.py", line 135, in populate_modules self.field = TCNNNerfactoField( File "F:\Programme\miniconda3\envs\nerfstudio\lib\site-packages\nerfstudio\fields\nerfacto_field.py", line 134, in __init__ self.direction_encoding = tcnn.Encoding( NameError: name 'tcnn' is not defined It sounds like tiny-cuda-nn wasn't installed correctly. I would try running pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch again. Also try opening up a python prompt, python, then running import tinycudann to see what error you get. It sounds like tiny-cuda-nn wasn't installed correctly. I would try running pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch again. Also try opening up a python prompt, python, then running import tinycudann to see what error you get. Thank you very much! The reinstallation actually worked, although the pip installation method of Tiny CUDA Neural Networks has never worked before for me. To document what I have done: Installed CUDA 11.7 and VS 2019 and deinstalled all other versions Created a conda envoriment conda create --name nerfstudio -y python=3.8 conda activate nerfstudio python -m pip install --upgrade pip Installed pytorch pip install torch==1.12.1+cu113 torchvision==0.13.1+cu113 -f https://download.pytorch.org/whl/torch_stable.html Git cloned tiny-cuda-nn via the repo into miniconda3\envs\nerfstudio\Lib\site-packages git clone --recursive https://github.com/nvlabs/tiny-cuda-nn cd tiny-cuda-nn cmake . -B build cmake --build build --config RelWithDebInfo -j (cmake commands run in VS developer command prompt) Installed nerfstudio pip install nerfstudio Reinstalled tiny-cuda-nn via pip pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch I am not sure, why the installation via pip worked without errors during the last step, but nerfstudio is now fully functional.
2025-04-01T06:39:44.026004
2024-05-24T05:53:07
2314454244
{ "authors": [ "0pdd" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8879", "repo": "nergal-perm/java-decita", "url": "https://github.com/nergal-perm/java-decita/issues/120" }
gharchive/issue
Make templates resolutions the Coordinate's responsibility The puzzle 23-93ef7c09 from #23 has to be resolved: https://github.com/nergal-perm/java-decita/blob/18bd62953fdb5dfe351db2621e3a14e4e8db6144/java-decita/src/main/java/ru/ewc/commands/SimpleCommand.java#L68-L68 The puzzle was created by Евгений Терехов on 24-May-24. role: DEV. If you have any technical questions, don't ask me, submit new tickets instead. The task will be "done" when the problem is fixed and the text of the puzzle is removed from the source code. Here is more about PDD and about me. The puzzle 23-93ef7c09 has disappeared from the source code, that's why I closed this issue.
2025-04-01T06:39:44.044323
2024-09-24T07:01:10
2544526323
{ "authors": [ "DininduSenanayake" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8880", "repo": "nesi/APSIM-HPC", "url": "https://github.com/nesi/APSIM-HPC/issues/53" }
gharchive/issue
Snakefile 1 - First Production level Snakefile for 06-slurm-array/array_create_db_files.sl Proposed Snakefile Following Snakefile defines a configuration with the Apptainer settings and excluded files. It gets a list of .apsimx files, excluding the specified files. The all rule defines the final output files we want to create. The process_apsimx rule defines how to process each .apsimx file: It requires both the .apsimx file and the corresponding .db ( this is the placeholder .db) file as input. It creates a .processed file as output to mark completion. Perhaps we can modify this later to save file quota sets the resources (memory and time) and number of threads. The shell command loads the Apptainer module, sets the bind path, and runs the Models command. import os from glob import glob # Configuration config = { "apptainer_bind": "/agr/scratch,/agr/persist", "apptainer_image": "/agr/persist/projects/2024_apsim_improvements/apsim-simulations/container/apsim-2024.09.7579.0.aimg", "excluded_files": ["2023-10-09_MasterSoilApsimLibrary.apsimx", "LargerExample.apsimx"] } # Get list of .apsimx files apsimx_files = [f for f in glob("*.apsimx") if f not in config["excluded_files"]] rule all: input: expand("{file}.processed", file=[os.path.splitext(f)[0] for f in apsimx_files]) rule process_apsimx: input: apsimx = "{file}.apsimx", db = "{file}.db" output: "{file}.processed" resources: mem_mb = 8000, time = "00:10:00" threads: 12 shell: """ module load Apptainer export APPTAINER_BIND="{config[apptainer_bind]}" apptainer exec {config[apptainer_image]} Models {input.apsimx} touch {output} """ Configure Snakemake - These instructions are for snakemake > 8 Create a configuration directory for Snakemake profiles: profile_dir="${HOME}/.config/snakemake" mkdir -p "$profile_dir" Use cookiecutter to create the Slurm profile template: If cookiecutter isn't installed, load snakemake module and run pip install --user cookiecutter template="gh:Snakemake-Profiles/slurm" cookiecutter --output-dir "$profile_dir" "$template" During the cookiecutter process, you'll be prompted to set values for your profile. For example: profile_name [slurm]: slurm.my_account sbatch_defaults []: account=my_account no-requeue exclusive cluster_sidecar_help: [Use cluster sidecar. NB! Requires snakemake >= 7.0! Enter to continue...] Select cluster_sidecar: 1 - yes 2 - no Choose from 1, 2 : cluster_name []: After completing the prompts, the profile scripts and configuration file will be installed in the $profile_dir as profile_name/[2]. You can then use this profile when running Snakemake by adding the --profile flag: snakemake --profile slurm.my_account ... This will configure Snakemake to submit jobs to Slurm using the settings you specified in the profile[2]. Remember that the profile typically includes scripts for job submission, status checking, and a configuration file (config.yaml) that defines default values for Snakemake command line arguments. You may need to adjust these files to match your specific Slurm environment and requirements. Working Snakefile for this rule : https://github.com/nesi/APSIM-HPC/pull/54
2025-04-01T06:39:44.151240
2023-11-16T12:59:38
1996805575
{ "authors": [ "LasseSkogland", "scottt732" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8881", "repo": "net-daemon/netdaemon", "url": "https://github.com/net-daemon/netdaemon/pull/1003" }
gharchive/pull-request
Initial implementation of automatic MQTT config retrieval from HA Supervisor Breaking change Should be no breaking changes. This only touches the MQTT extension. Proposed change The MQTT configuration should be retrieved from HA Supervisor. I've explained the process in this issue: #1002 Type of change [ ] Dependency upgrade [ ] Bugfix (non-breaking change which fixes an issue) [x ] New feature (which adds functionality to an existing integration) [ ] Breaking change (fix/feature causing existing functionality to break) [ ] Code quality improvements to existing code or addition of tests Additional information This PR fixes or closes issue: fixes #1002 This PR is related to issue: #1002 Link to documentation pull request: Checklist [ ] The code change is tested and works locally. [ ] Local tests pass. Your PR cannot be merged unless tests pass [ ] There is no commented out code in this PR. [ ] I have followed the [development checklist][dev-checklist] [ ] The code compiles without warnings (code quality chek) [ ] Tests have been added to verify that the new code works. If user exposed functionality or configuration variables are added/changed: [ ] Documentation added/updated for www.home-assistant.io First draft, quickly thrown together. You may want to make some adjustments here: https://github.com/net-daemon/netdaemon/blob/main/Docker/rootfs/etc/services.d/netdaemon_addon/run#L14-L22 Also I couldn't find if/where you define a config.yaml file for the netdaemon. I think you would want to express that the app wants rather than needs the mqtt service. HA can only provide details for it if it's installed as an addon. In my case, I don't use run mosquitto as an addon (it runs on another machine). Ideally (I think) you'd want to collect optional MQTT config from users for people who use external brokers via config.yaml and pass it through to the right env vars if it's defined in the run command above. Consider rewrite, Mqtt env vars is already set på in run script.
2025-04-01T06:39:44.162806
2022-10-04T16:27:41
1396547767
{ "authors": [ "coveralls", "helto4real" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8882", "repo": "net-daemon/netdaemon_v2", "url": "https://github.com/net-daemon/netdaemon_v2/pull/71" }
gharchive/pull-request
Dependency upgrades Breaking change Proposed change Type of change [x] Dependency upgrade [ ] Bugfix (non-breaking change which fixes an issue) [ ] New feature (which adds functionality to an existing integration) [ ] Breaking change (fix/feature causing existing functionality to break) [ ] Code quality improvements to existing code or addition of tests Additional information This PR fixes or closes issue: fixes # This PR is related to issue: Link to documentation pull request: Checklist [x] The code change is tested and works locally. [ ] Local tests pass. Your PR cannot be merged unless tests pass [ ] There is no commented out code in this PR. [ ] I have followed the [development checklist][dev-checklist] [ ] The code compiles without warnings (code quality chek) [ ] Tests have been added to verify that the new code works. If user exposed functionality or configuration variables are added/changed: [ ] Documentation added/updated for www.home-assistant.io Pull Request Test Coverage Report for Build<PHONE_NUMBER> 0 of 0 changed or added relevant lines in 0 files are covered. 4 unchanged lines in 1 file lost coverage. Overall coverage decreased (-0.03%) to 63.208% Files with Coverage Reduction New Missed Lines % src/Daemon/NetDaemon.Daemon/Daemon/NetDaemonHost.cs 4 60.26% Totals Change from base Build<PHONE_NUMBER>: -0.03% Covered Lines: 4505 Relevant Lines: 6666 💛 - Coveralls
2025-04-01T06:39:44.175907
2023-10-21T03:00:34
1955255692
{ "authors": [ "Cs0o", "kriodoxis", "nReus", "net00-1" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8883", "repo": "net00-1/SW-DLT", "url": "https://github.com/net00-1/SW-DLT/issues/38" }
gharchive/issue
Ability to provide credentials When trying to download an Instgram video the following message appears: Provided the following link: https://www.instagram.com/reel/CybzFHyOcUo/?igshid=MzRlODBiNWFlZA== I'd like to add on to this issue. It appears that the frontend takes links from Gallery-dl before G-DL begins downloading, which causes it to download from sources without the credentials I've manually provided to G-DL, resulting in 403 forbidden errors being downloaded as html files instead. In my case, it was Pixiv that didn't work for all R18-tagged posts. And I don't think it would work on SW-DLT without major revision to how the frontend downloads galleries since Pixiv requires OAuth instead just providing username and password (G-DL uses an Android user-agent and Pixiv on Android browsers require OAuth afaik). For accounts that work with username & password, you can look in the [docs[(https://github.com/net00-1/SW-DLT/blob/master/Docs.md#authentication). There's a guide to setup a-shell to keep your credentials. I will keep this open to check for ways of other authentication types As a newbie, do I have this directory structure set up correctly? It keeps prompting "requires. authentication" [~group]$ ls -Ra . .. .com.apple.mobile_container_manager.metadata.plist .netrc Library SW_DLT.py gallery-dl yt-dlp ./Library: . Caches .. Preferences ./Library/Caches: . .. ./Library/Preferences: . .. group.AsheKube.a-Shell.plist ./gallery-dl: . config.json .. ./yt-dlp: . .. config [~group]$ cat yt-dlp/config --netrc [~group]$ cat gallery-dl/config.json { "extractor": { "twitter": { "username": "$$$$$$", "password": "$$$$$$" } }} Now much more convenient cookie-based authentication is available, you can check the updated docs about how to use it.
2025-04-01T06:39:44.227350
2021-11-10T11:06:38
1049688029
{ "authors": [ "maxkratz", "netcupClaudiaM", "raphi-0901" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8884", "repo": "netcup-community/community-tutorials", "url": "https://github.com/netcup-community/community-tutorials/pull/55" }
gharchive/pull-request
Add standalone Seafile Docker tutorial This PR adds a tutorial on how to set up a standalone Seafile server including a valid LetsEncrypt certificate for TLS/SSL secured HTTPS connections. Please notice that the linked tutorial on how to install Docker and Docker-Compose is not merged yet (https://github.com/netcup-community/community-tutorials/pull/17). I have read and understood the Contributor's Certificate of Origin at the end of the template and I hereby certify that I meet the contribution criteria described in it. Signed-off-by: Max Kratz<EMAIL_ADDRESS> Hi maxkratz, thank you for contributing! It's been a while since you opened this PR and it probably is an open secret we're currently overwhelmed with reviewing all the awesome tutorials we got. Please be patient - we're about to publish tutorials this and next week. Thank you! @netcupClaudiaM Thank you for the update! Take the time you need for all the reviews :). Hey @maxkratz, could you please upload all your pictures in english so that everything is consistent? Thanks! @raphi-0901 No problem, I've fixed the German screenshots. Thank you for the suggestion. Furthermore, I've added a configuration option for the services to automatically start in case the system had to be (re-)booted.
2025-04-01T06:39:44.292652
2022-07-26T17:06:25
1318530620
{ "authors": [ "giorio94" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8885", "repo": "netgroup-polito/CrownLabs", "url": "https://github.com/netgroup-polito/CrownLabs/pull/801" }
gharchive/pull-request
Infrastructure: bump velero version and add selector Description This PR bumps the version of velero, and introduces a label selector to filter out the resources originated from CrownLabs instances, in order to prevent the partial failures due to the kubevirt/velero issue. Fixes # (issue) How Has This Been Tested? Please describe the tests that you ran to verify your changes. Please also note any relevant details for your test configuration. [x] Deploying the updated manifest in the cluster, and checking that the backup completed correctly /merge
2025-04-01T06:39:44.333574
2020-08-17T10:41:38
680134328
{ "authors": [ "SteveALee", "erezrokah", "sarahetter" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8886", "repo": "netlify/cli", "url": "https://github.com/netlify/cli/issues/1116" }
gharchive/issue
invlide update available messge When running "netlify dev function:invoke" I got a message saying a new version of the cli was available and I should run npm -i -g netlify-cli -i is invalid and should be just i I can't find this in the source code so guess it's a message from somewhere else. Thanks @SteveALee, it is probably coming from https://github.com/yeoman/update-notifier. Can you please share the output of npx envinfo --system --binaries --npmPackages netlify-cli --npmGlobalPackages netlify-cli so we can try and reproduce using the same environment? Sadly I reinstalled the cli - after having to delete some files (ntl.psi and ntl.cmd) from ...AppData/Roaming/npm/, even after npm rm -g netlify-cli But anyway C:\Users\steve>npx envinfo --system --binaries --npmPackages netlify-cli --npmGlobalPackages netlify-cli npx: installed 1 in 1.639s System: OS: Windows 10 10.0.19041 CPU: (8) x64 Intel(R) Core(TM) i7-8550U CPU @ 1.80GHz Memory: 4.08 GB / 15.86 GB Binaries: Node: 14.6.0 - C:\Program Files\nodejs\node.EXE npm: 6.14.6 - C:\Program Files\nodejs\npm.CMD Did you expect netlify-cli there? C:\Users\steve>npm ls -g netlify-cli C:\Users\steve\AppData\Roaming\npm `--<EMAIL_ADDRESS> Yeah I was expecting to see see the CLI, but maybe we have some installation issues with Windows machines. Thanks for the information. yes the update notifier text looks correct. And the version bumped .1 Perhaps I imaged the -i ? Closing as stale. Please re-open if needed
2025-04-01T06:39:44.335250
2020-11-30T18:07:57
753690520
{ "authors": [ "jlengstorf", "tzmanics" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8887", "repo": "netlify/explorers", "url": "https://github.com/netlify/explorers/issues/430" }
gharchive/issue
add resource links to stages many stages refer to links and resources, so it might be a good idea to put those resources in a list Do you think we should separate this out in the UI, like make a text field Resource List in the stage data object? that's probably not the worst idea, yeah covered by #431 covered by #431
2025-04-01T06:39:44.341611
2018-06-05T09:23:22
329371462
{ "authors": [ "erquhart", "papandreou" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8888", "repo": "netlify/netlify-cms", "url": "https://github.com/netlify/netlify-cms/issues/1405" }
gharchive/issue
Hidden field in editor component is rendered - Do you want to request a feature or report a bug? bug - What is the current behavior? An editor component that declares a hidden field will have it rendered: - If the current behavior is a bug, please provide the steps to reproduce. Declare an editor component that has a field with widget: 'hidden': CMS.registerEditorComponent({ // Internal id of the component id: "youtube", // Visible label label: "Youtube", // Fields the user need to fill out when adding an instance of the component fields: [{ name: 'id', label: 'Youtube Video ID', widget: 'string' }, { name: 'myHiddenField', label: 'My hidden field', widget: 'hidden' }], // Pattern to identify a block as being an instance of this component pattern: /^youtube (\S+)$/, // Function to extract data elements from the regexp match fromBlock: function(match) { return { id: match[1] }; }, // Function to create a text block from an instance of this component toBlock: function(obj) { return 'youtube ' + obj.id; }, // Preview output for this component. Can either be a string or a React component // (component gives better render performance) toPreview: function(obj) { return ( '<img src="http://img.youtube.com/vi/' + obj.id + '/maxresdefault.jpg" alt="Youtube Video"/>' ); } }); Add an instance of the component to a markdown editor. Observe that the label for the hidden field is rendered with a No control for widget: 'hidden' message below it. The same thing happens if the hidden field has no label except that a small gray box is rendered instead of the label. - What is the expected behavior? That the field isn't rendered, similar to how hidden widgets work for collections. - Please mention your versions where applicable. Netlify CMS version: 1.8.4 Browser version: Chrome 67.0.3396.62/OSX Node.JS version: 9.5.0 Operating System: OSX - Please link or paste your config.yml below if applicable. Good catch!
2025-04-01T06:39:44.351415
2020-10-21T13:13:23
726469152
{ "authors": [ "clearstatus", "erezrokah", "thelynchpin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8889", "repo": "netlify/netlify-cms", "url": "https://github.com/netlify/netlify-cms/issues/4489" }
gharchive/issue
config should have required property 'options' - Clearstatus Netlify I get this error post a successful build and deploy on netlify. https://youthful-swartz-68b931.netlify.app/ Error loading the CMS configuration Config Errors: config should have required property 'options' Check your config.yml file. Hi @thelynchpin, one of your select widgets is missing the options configuration: fields: - label: "Name" name: "name" widget: string - label: "Group" name: "group" widget: select multiple: false required: false - label: "Description" name: "description" widget: string required: false - label: "Link" name: "link" widget: string required: false I'm new to this and have deployed the Clearstatus app through the deploy to netlify button on https://github.com/weeblrpress/clearstatus. Please specify what needs to be modified on the list. Welcome :) Please see the docs for the select widget here: https://www.netlifycms.org/docs/widgets/#select I suggest you open an issue on that repo and ask where is the config.yml file is located @thelynchpin Can you give it a try now, I think we fixed the problem. @erezrokah What seems to have happened is that previously Netlify CMS would accept that select without options (which depends on the user particular site configuration) but at some point this started to trigger an error. I'm now removing the entire select if no "groups" option is configured by user. @erezrokah What seems to have happened is that previously Netlify CMS would accept that select without options (which depends on the user particular site configuration) but at some point this started to trigger an error. I'm now removing the entire select if no "groups" option is configured by user. Yes, we added schema validations for widgets configurations. The widget used output an error div prior to that: https://github.com/netlify/netlify-cms/blob/519cb2d4c2db729d2643c9116f93656b6a9dba23/packages/netlify-cms-widget-select/src/SelectControl.js#L114 Thanks. It's working now. Regards, Wilbur On Fri, Nov 6, 2020 at 3:13 PM clearstatus<EMAIL_ADDRESS>wrote: @thelynchpin https://github.com/thelynchpin Can you give it a try now, I think we fixed the problem. @erezrokah https://github.com/erezrokah What seems to have happened is that previously Netlify CMS would accept that select without options (which depends on the user particular site configuration) but at some point this started to trigger an error. I'm now removing the entire select if no "groups" option is configured by user. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/netlify/netlify-cms/issues/4489#issuecomment-722983208, or unsubscribe https://github.com/notifications/unsubscribe-auth/ACEOACCGO2YD2AOUDIJEMOLSOPANZANCNFSM4SZYORPA .
2025-04-01T06:39:44.359853
2022-11-02T18:08:40
1433574428
{ "authors": [ "ascorbic", "jasonbarry", "serhalp" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8890", "repo": "netlify/next-runtime", "url": "https://github.com/netlify/next-runtime/issues/1736" }
gharchive/issue
[Bug]: fail build if output: 'standalone' config option is set Summary Those coming from the docker/k8s world are likely to follow this guidance from Next.js official docs, which instructs them to clone the with-docker example, whose config explicitly uses output: 'standalone': module.exports = { output: 'standalone', } This will create a folder at .next/standalone which can then be deployed on its own without installing node_modules. Additionally, a minimal server.js file is also output which can be used instead of next start. This minimal server does not copy the public or .next/static folders by default as these should ideally be handled by a CDN instead, although these folders can be copied to the standalone/public and standalone/.next/static folders manually, after which server.js file will serve these automatically. As this relies on an always-running node.js server, it's incompatible with our serverless lyfestyle. @ascorbic suggested that rather than changing the site's next.config.js file at build-time, we fail the build and output an error message: I think the most useful then would be for us to detect that setting and fail the build, with an appropriate message Having this error message will be immensely useful for those coming from Next.js wrapped in docker. Steps to reproduce Clone and deploy the with-docker example from Next.js examples A link to a reproduction repository No response Next Runtime version 4.28.3 More information about your build [ ] I am building using the CLI [ ] I am building using file-based configuration (netlify.toml) What OS are you using? No response Your netlify.toml file No response Your public/_redirects file No response Your next.config.js file `next.config.js` module.exports = { output: 'standalone', } Builds logs (or link to your logs) No response Function logs No response .next JSON files No response @ascorbic Is this still relevant with v5? It won't be, because it's forced to standalone anyway
2025-04-01T06:39:44.362708
2023-11-07T00:30:40
1980316070
{ "authors": [ "iKristy", "serhalp" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8891", "repo": "netlify/next-runtime", "url": "https://github.com/netlify/next-runtime/pull/2355" }
gharchive/pull-request
update README image with new logo Description Removes the old version of the README image and replaces it with one donning the new Netlify logo ✨ Documentation Tests Visit the README file See the change (updated Netlify logo) Relevant links (GitHub issues, etc.) or a picture of cute animal Superseded by v5. (I checked and it has the updated logo.)
2025-04-01T06:39:44.376200
2023-03-21T12:40:49
1633834592
{ "authors": [ "jotak", "jpinsonneau", "memodi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8892", "repo": "netobserv/network-observability-console-plugin", "url": "https://github.com/netobserv/network-observability-console-plugin/pull/310" }
gharchive/pull-request
NETOBSERV-857 After some time, it fails to retrieve flows Reload HOST token from file on every query /ok-to-test /lgtm up to you to do what I suggested above or not .. this isn't super critical /label qe-approved Verified no longer running into the cert issue with this change when authToken: HOST is set. @jpinsonneau are we good to merge? /approve
2025-04-01T06:39:44.433520
2016-08-24T17:50:21
173016493
{ "authors": [ "Scottmitch", "carl-mastrangelo", "nmittler" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8893", "repo": "netty/netty", "url": "https://github.com/netty/netty/pull/5746" }
gharchive/pull-request
HTTP/2 Hpack Encoder Cleanup Motivation: The HTTP/2 HPACK Encoder class has some code which is only used for test purposes. This code can be removed to reduce complexity and member variable count. Modifications: Remove test code and update unit tests Other minor cleanup Result: Test code is removed from operational code. this is a minor cleanup ... we loose some ability to control the encoder in tests however I'm not sure if this provides too much value. @nmittler @normanmaurer @carl-mastrangelo - FYI No red flags from my pov. LGTM. LGTM 4.1 (208893a)
2025-04-01T06:39:44.437335
2018-05-03T12:13:24
319895025
{ "authors": [ "jchambers", "normanmaurer" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8894", "repo": "netty/netty", "url": "https://github.com/netty/netty/pull/7903" }
gharchive/pull-request
Allow to disable native transport and native ssl support via system p… …roperty. Motivation: Sometimes it's useful to disable native transports / native ssl to debug a problem. We should allow to do so with a system property so people not need to adjust code for this. Modifications: Add system properties which allow to disable native transport and native ssl. Result: Easier to disable native code usage without code changes. @trustin done... will merge once ci pass. Awesome! This is, indeed, very useful for testing. Some notes for future readers who might also be interested in this change: This is dramatically easier to read with whitespace-only changes hidden The system property for disabling native SSL providers (i.e. netty-tcnative) is io.netty.handler.ssl.noOpenSsl (example: -Dio.netty.handler.ssl.noOpenSsl=true) The system property for disabling native transports (i.e. epoll and kqueue) is io.netty.transport.noNative (example: -Dio.netty.transport.noNative=true)
2025-04-01T06:39:44.472460
2011-11-08T18:39:31
2177681
{ "authors": [ "netzpirat", "oreoshake" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8895", "repo": "netzpirat/guard-cucumber", "url": "https://github.com/netzpirat/guard-cucumber/issues/24" }
gharchive/issue
Individual scenarios are no longer run if the entire feature is also sche Individual scenarios are no longer run if the entire feature is also scheduled to run for https://github.com/netzpirat/guard-cucumber/issues/22 Run Cucumber features features/password_client.feature -break test Run Cucumber features features/password_client.feature -fix test Run Cucumber features features/password_client.feature features/ password_client.feature:22 This is no longer the case Thanks a lot for this! I will release a bug fix version today.
2025-04-01T06:39:44.481910
2017-10-09T20:04:31
264010379
{ "authors": [ "Tornhoof", "firephantomassasin", "neuecc" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8896", "repo": "neuecc/Utf8Json", "url": "https://github.com/neuecc/Utf8Json/issues/17" }
gharchive/issue
Use JsonUtf8Encoding : Encoding Escaping string character is hurt of performance of JSON serialization. It is possible to reduce escape cost by creating custom UTF8 Encoding that includes JSON encoding/decoding. for invoke internal FastAllocateString, it is necessary to inherit Encoding. public class JsonUtf8Encoding : Encoding { // called from GetString. // return CharCount is \" ... \" unescaped. public override int GetCharCount(byte[] bytes, int index, int count) { throw new NotImplementedException(); } public override int GetChars(byte[] bytes, int byteIndex, int byteCount, char[] chars, int charIndex) { throw new NotImplementedException(); } public override int GetMaxByteCount(int charCount) { throw new NotImplementedException(); } public override int GetMaxCharCount(int byteCount) { throw new NotImplementedException(); } public override int GetByteCount(char[] chars, int index, int count) { throw new NotImplementedException(); } public override int GetBytes(char[] chars, int charIndex, int charCount, byte[] bytes, int byteIndex) { throw new NotImplementedException(); } } Also, it is necessary to implement efficient UTF 8 encoding/decoding. I found this article. http://bjoern.hoehrmann.de/utf-8/decoder/dfa/ If there are any other good examples, please let me know. @itn3000 is trying fast utf8 <-> utf16 utilities. https://github.com/itn3000/unicode-convert-utilities @ufcpp is building custom UTF8 decoder. https://github.com/ufcpp/Utf8Utils NStack is golang like new encoding system. https://github.com/migueldeicaza/NStack System.Text.Utf8String is span based new primitive. https://github.com/dotnet/corefxlab/tree/master/src/System.Text.Utf8String/System/Text Regarding utf-8: http://nullprogram.com/blog/2017/10/06/ https://news.ycombinator.com/item?id=15423674 and related from https://github.com/dotnet/corefxlab/issues/1831 hi @neuecc we badly need your help, do you have an idea or workaround for our issue? https://github.com/neuecc/Utf8Json/issues/224
2025-04-01T06:39:44.486088
2024-11-26T10:25:29
2694072059
{ "authors": [ "ElseFaisal", "Study-is-happy", "leonbaiyu" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8897", "repo": "neufieldrobotics/NeuFlow_v2", "url": "https://github.com/neufieldrobotics/NeuFlow_v2/issues/10" }
gharchive/issue
Feasibility and Capability of Inference on CPU Hi there, I have a usecase for your model on a spacecraft but the onboard system will only have a 1 GHz CPU and no dedicated GPU. I wanted to ask whether you believe it would be possible to implement your model on this level of hardware. I have attempted to and failed to carry out inference on the CPU also. I think you need to make changes to the model architecture code. Usually by replacing device variable with "cpu" instead of "0" or "cuda". I have tried this. There are challenges with regards to the half precision floating point data types. What exactly the problem is is not clear to me. Have you attempted to adjust the quantization settings of the model to prevent it from defaulting to half-precision (FP16)? I'll try to do the inference via a CPU and will let you know Hi leonbaiyu, you can change the datatype to FP32 instead of default FP16, by searching "amp" and make it False. Thank you Hi leonbaiyu, you can change the datatype to FP32 instead of default FP16, by searching "amp" and make it False. Thank you Thank you for this. This solved the problem and now I'm able to infer fairly quickly on CPU. Specifically, I removed any .cuda(), replaced 'cuda' device calls to 'cpu', replaced any .half() with float, and specified in model.init_bhwd for amp = False. No changes were required within the actual NeuFlow code, changes were only required in infer.py . The results are looking promising. We'll be in touch if we end up using your model in our applications.
2025-04-01T06:39:44.489776
2023-02-01T16:11:32
1566398607
{ "authors": [ "davidmezzetti" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8898", "repo": "neuml/txtai", "url": "https://github.com/neuml/txtai/issues/417" }
gharchive/issue
Extractor pipeline 2.0 This is major update on the path to Generative Semantic Search The extractor pipeline was one of the first components in txtai, going all the way back to 1.0. Since then, much has changed both with txtai and externally. This pipeline has a lot of potential but it needs a couple updates. Make the following upgrades to the Extractor pipeline. Ability to run embeddings searches. Given that content is supported, text can be retrieved from the embeddings instance. In addition to extractive qa, support text generation models, sequence to sequence models and custom pipelines Better detection of when a tokenizer should be used (word vector models only) These changes will enable a prompt-driven approach to question-answering with LLMs. This includes Hugging Face models and external services like OpenAI/Cohere. Services can be called directly or with another library like langchain. Custom pipelines only require a __call__ interface. This recent notebook should help: https://github.com/neuml/txtai/blob/master/examples/62_RAG_with_llama_cpp_and_external_API_services.ipynb
2025-04-01T06:39:44.502411
2023-11-23T19:57:39
2008737790
{ "authors": [ "coveralls", "surchs" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8899", "repo": "neurobagel/api", "url": "https://github.com/neurobagel/api/pull/227" }
gharchive/pull-request
[FIX] Remove --no-progress-meter Closes #213 Changes proposed in this pull request: replace the --no-progress-meter argument with -s -S which has the same effect but works on curl < 7.67.0 Checklist [ ] PR has an interpretable title with a prefix ([ENH], [FIX], [REF], [TST], [CI], [MNT], [INF], [MODEL], [DOC]) (see https://neurobagel.org/contributing/pull_requests for more info) [ ] PR links to GitHub issue with mention Closes #XXXX [ ] Tests pass [ ] Checks pass For new features: [ ] Tests have been added For bug fixes: [ ] There is at least one test that would fail under the original bug conditions. Pull Request Test Coverage Report for Build<PHONE_NUMBER> 0 of 0 changed or added relevant lines in 0 files are covered. No unchanged relevant lines lost coverage. Overall coverage remained the same at 99.732% Totals Change from base Build<PHONE_NUMBER>: 0.0% Covered Lines: 745 Relevant Lines: 747 💛 - Coveralls :crossed_fingers:
2025-04-01T06:39:44.634922
2023-05-01T21:47:01
1691452317
{ "authors": [ "AlexanderZhirov", "FirstLicz", "matt335672" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8900", "repo": "neutrinolabs/xrdp", "url": "https://github.com/neutrinolabs/xrdp/issues/2652" }
gharchive/issue
black screen xrdp in Mate It is not possible to connect to RDP in Mate environment on stock settings. Installed only xrdp and xorgxrdp packages. The connection goes through, but the session is not established. xrdp, version: 0.9.21 xorgxrdp, version: 0.9.19 xrdp.log: [20230501-16:16:08] [INFO ] address [<IP_ADDRESS>] port [3389] mode 1 [20230501-16:16:08] [INFO ] listening to port 3389 on <IP_ADDRESS> [20230501-16:16:08] [INFO ] xrdp_listen_pp done [20230501-16:16:08] [DEBUG] Closed socket 7 (AF_INET6 :: port 3389) [20230501-16:16:10] [INFO ] starting xrdp with pid 2455 [20230501-16:16:10] [INFO ] address [<IP_ADDRESS>] port [3389] mode 1 [20230501-16:16:10] [INFO ] listening to port 3389 on <IP_ADDRESS> [20230501-16:16:10] [INFO ] xrdp_listen_pp done [20230501-16:16:24] [INFO ] Socket 12: AF_INET6 connection received from ::ffff:<IP_ADDRESS> port 49236 [20230501-16:16:24] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:<IP_ADDRESS> port 3389) [20230501-16:16:24] [DEBUG] Closed socket 11 (AF_INET6 :: port 3389) [20230501-16:16:24] [DEBUG] item ini_version, value 1 [20230501-16:16:24] [DEBUG] item fork, value true [20230501-16:16:24] [DEBUG] item port, value 3389 [20230501-16:16:24] [DEBUG] item use_vsock, value false [20230501-16:16:24] [DEBUG] item tcp_nodelay, value true [20230501-16:16:24] [DEBUG] item tcp_keepalive, value true [20230501-16:16:24] [DEBUG] item security_layer, value negotiate [20230501-16:16:24] [DEBUG] item crypt_level, value high [20230501-16:16:24] [DEBUG] item certificate, value [20230501-16:16:24] [INFO ] Using default X.509 certificate: /etc/xrdp/cert.pem [20230501-16:16:24] [DEBUG] item key_file, value [20230501-16:16:24] [INFO ] Using default X.509 key file: /etc/xrdp/key.pem [20230501-16:16:24] [DEBUG] item ssl_protocols, value TLSv1.2, TLSv1.3 [20230501-16:16:24] [DEBUG] TLSv1.3 enabled [20230501-16:16:24] [DEBUG] TLSv1.2 enabled [20230501-16:16:24] [DEBUG] item autorun, value [20230501-16:16:24] [DEBUG] item allow_channels, value true [20230501-16:16:24] [DEBUG] item allow_multimon, value true [20230501-16:16:24] [DEBUG] item bitmap_cache, value true [20230501-16:16:24] [DEBUG] item bitmap_compression, value true [20230501-16:16:24] [DEBUG] item bulk_compression, value true [20230501-16:16:24] [DEBUG] item max_bpp, value 16 [20230501-16:16:24] [DEBUG] item new_cursors, value true [20230501-16:16:24] [DEBUG] item use_fastpath, value both [20230501-16:16:24] [DEBUG] item blue, value 009cb5 [20230501-16:16:24] [DEBUG] item grey, value dedede [20230501-16:16:24] [DEBUG] item ls_top_window_bg_color, value 009cb5 [20230501-16:16:24] [DEBUG] item ls_width, value 350 [20230501-16:16:24] [DEBUG] item ls_height, value 430 [20230501-16:16:24] [DEBUG] item ls_bg_color, value dedede [20230501-16:16:24] [DEBUG] item ls_logo_filename, value [20230501-16:16:24] [DEBUG] item ls_logo_x_pos, value 55 [20230501-16:16:24] [DEBUG] item ls_logo_y_pos, value 50 [20230501-16:16:24] [DEBUG] item ls_label_x_pos, value 30 [20230501-16:16:24] [DEBUG] item ls_label_width, value 65 [20230501-16:16:24] [DEBUG] item ls_input_x_pos, value 110 [20230501-16:16:24] [DEBUG] item ls_input_width, value 210 [20230501-16:16:24] [DEBUG] item ls_input_y_pos, value 220 [20230501-16:16:24] [DEBUG] item ls_btn_ok_x_pos, value 142 [20230501-16:16:24] [DEBUG] item ls_btn_ok_y_pos, value 370 [20230501-16:16:24] [DEBUG] item ls_btn_ok_width, value 85 [20230501-16:16:24] [DEBUG] item ls_btn_ok_height, value 30 [20230501-16:16:24] [DEBUG] item ls_btn_cancel_x_pos, value 237 [20230501-16:16:24] [DEBUG] item ls_btn_cancel_y_pos, value 370 [20230501-16:16:24] [DEBUG] item ls_btn_cancel_width, value 85 [20230501-16:16:24] [DEBUG] item ls_btn_cancel_height, value 30 [20230501-16:16:24] [INFO ] Security protocol: configured [SSL|RDP], requested [SSL|HYBRID|HYBRID_EX|RDP], selected [SSL] [20230501-16:16:24] [DEBUG] Using TLS security, and setting RDP security crypto to LEVEL_NONE and METHOD_NONE [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] receive connection request [20230501-16:16:24] [INFO ] Connected client computer name: iPhone (Але [20230501-16:16:24] [DEBUG] Client supports 40 bit encryption [20230501-16:16:24] [DEBUG] Client supports 128 bit encryption [20230501-16:16:24] [DEBUG] Client supports 56 bit encryption [20230501-16:16:24] [DEBUG] Client supports fips encryption [20230501-16:16:24] [DEBUG] The connection is using TLS, skipping RDP crypto negotiation [20230501-16:16:24] [DEBUG] Adding channel: name rdpdr, channel id 1004, flags 0x80800000 [20230501-16:16:24] [DEBUG] Adding channel: name rdpsnd, channel id 1005, flags 0xc0000000 [20230501-16:16:24] [DEBUG] Adding channel: name cliprdr, channel id 1006, flags 0xc0a00000 [20230501-16:16:24] [DEBUG] Adding channel: name drdynvc, channel id 1007, flags 0xc0800000 [20230501-16:16:24] [WARN ] Received [MS-RDPBCGR] TS_UD_HEADER type 0xc006 is unknown (ignored) [20230501-16:16:24] [WARN ] Received [MS-RDPBCGR] TS_UD_HEADER type 0xc00a is unknown (ignored) [20230501-16:16:24] [WARN ] Client requested 32 bpp color depth, but the server configuration is limited to 16 bpp. Downgrading the color depth to 16 bits-per-pixel. [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] construct connection response [20230501-16:16:24] [DEBUG] using no security [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] send connection response [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] receive erect domain request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] receive attach user request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence] send attach user confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230501-16:16:24] [DEBUG] [MCS Connection Sequence (TLS)] completed [20230501-16:16:24] [INFO ] xrdp_load_keyboard_layout: Keyboard information sent by the RDP client, keyboard_type:[0x04], keyboard_subtype:[0x00], keylayout:[0x00000409] [20230501-16:16:24] [DEBUG] keyboard_cfg_file /etc/xrdp/xrdp_keyboard.ini [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value 0x00000409 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value us [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item keyboard_type value 4 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item keyboard_subtype value 3 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item keyboard_type value 7 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item keyboard_subtype value 2 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item model value pc105 [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item rdp_layouts value default_rdp_layouts [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item layouts_map value default_layouts_map [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value us [20230501-16:16:24] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230501-16:16:24] [INFO ] xrdp_load_keyboard_layout: model [] variant [] layout [us] options [] [20230501-16:16:24] [INFO ] TLS connection established from ::ffff:<IP_ADDRESS> port 49236: TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 [20230501-16:16:24] [DEBUG] Client requested auto logon. [20230501-16:16:24] [DEBUG] Client requested compression enabled. [20230501-16:16:24] [DEBUG] Client supplied domain: [20230501-16:16:24] [DEBUG] Client supplied username: user [20230501-16:16:24] [DEBUG] Client supplied password: <omitted from log> [20230501-16:16:24] [DEBUG] Client supplied program: [20230501-16:16:24] [DEBUG] Client supplied directory: [20230501-16:16:24] [INFO ] xrdp_caps_process_pointer: client supports new(color) cursor [20230501-16:16:24] [INFO ] xrdp_process_offscreen_bmpcache: support level 0 cache size 0 MB cache entries 0 [20230501-16:16:24] [INFO ] xrdp_caps_process_codecs: nscodec, codec id 1, properties len 3 [20230501-16:16:24] [DEBUG] xrdp_000009a1_wm_login_state_event_00000001 [20230501-16:16:24] [INFO ] Loading keymap file /etc/xrdp/km-00000409.ini [20230501-16:16:24] [WARN ] local keymap file for 0x00000409 found and doesn't match built in keymap, using local keymap file [20230501-16:16:24] [DEBUG] Login state change request WMLS_RESET -> WMLS_RESET [20230501-16:16:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 0 [20230501-16:16:24] [DEBUG] Login state change request WMLS_RESET -> WMLS_USER_PROMPT [20230501-16:16:24] [DEBUG] in xrdp_wm_init: [20230501-16:16:24] [DEBUG] ini_version: 1 [20230501-16:16:24] [DEBUG] use_bitmap_cache: 1 [20230501-16:16:24] [DEBUG] use_bitmap_compression: 1 [20230501-16:16:24] [DEBUG] port: 3389 [20230501-16:16:24] [DEBUG] crypt_level: 3 [20230501-16:16:24] [DEBUG] allow_channels: 1 [20230501-16:16:24] [DEBUG] max_bpp: 16 [20230501-16:16:24] [DEBUG] fork: 1 [20230501-16:16:24] [DEBUG] tcp_nodelay: 1 [20230501-16:16:24] [DEBUG] tcp_keepalive: 1 [20230501-16:16:24] [DEBUG] tcp_send_buffer_bytes: 0 [20230501-16:16:24] [DEBUG] tcp_recv_buffer_bytes: 0 [20230501-16:16:24] [DEBUG] new_cursors: 1 [20230501-16:16:24] [DEBUG] allow_multimon: 1 [20230501-16:16:24] [DEBUG] grey: 14606046 [20230501-16:16:24] [DEBUG] black: 0 [20230501-16:16:24] [DEBUG] dark_grey: 0 [20230501-16:16:24] [DEBUG] blue: 40117 [20230501-16:16:24] [DEBUG] dark_blue: 0 [20230501-16:16:24] [DEBUG] white: 0 [20230501-16:16:24] [DEBUG] red: 0 [20230501-16:16:24] [DEBUG] green: 0 [20230501-16:16:24] [DEBUG] background: 0 [20230501-16:16:24] [DEBUG] autorun: [20230501-16:16:24] [DEBUG] hidelogwindow: 0 [20230501-16:16:24] [DEBUG] require_credentials: 0 [20230501-16:16:24] [DEBUG] bulk_compression: 1 [20230501-16:16:24] [DEBUG] new_cursors: 1 [20230501-16:16:24] [DEBUG] nego_sec_layer: 0 [20230501-16:16:24] [DEBUG] allow_multimon: 1 [20230501-16:16:24] [DEBUG] enable_token_login: 0 [20230501-16:16:24] [DEBUG] ls_top_window_bg_color: 4f6 [20230501-16:16:24] [DEBUG] ls_width: 350 [20230501-16:16:24] [DEBUG] ls_height: 430 [20230501-16:16:24] [DEBUG] ls_bg_color: defb [20230501-16:16:24] [DEBUG] ls_title: [20230501-16:16:24] [DEBUG] ls_logo_filename: [20230501-16:16:24] [DEBUG] ls_logo_x_pos: 55 [20230501-16:16:24] [DEBUG] ls_logo_y_pos: 50 [20230501-16:16:24] [DEBUG] ls_label_x_pos: 30 [20230501-16:16:24] [DEBUG] ls_label_width: 65 [20230501-16:16:24] [DEBUG] ls_input_x_pos: 110 [20230501-16:16:24] [DEBUG] ls_input_width: 210 [20230501-16:16:24] [DEBUG] ls_input_y_pos: 220 [20230501-16:16:24] [DEBUG] ls_btn_ok_x_pos: 142 [20230501-16:16:24] [DEBUG] ls_btn_ok_y_pos: 370 [20230501-16:16:24] [DEBUG] ls_btn_ok_width: 85 [20230501-16:16:24] [DEBUG] ls_btn_ok_height: 30 [20230501-16:16:24] [DEBUG] ls_btn_cancel_x_pos: 237 [20230501-16:16:24] [DEBUG] ls_btn_cancel_y_pos: 370 [20230501-16:16:24] [DEBUG] ls_btn_cancel_width: 85 [20230501-16:16:24] [DEBUG] ls_btn_cancel_height: 30 [20230501-16:16:24] [DEBUG] libxrdp_query_channel - Channel 0 name rdpdr [20230501-16:16:24] [DEBUG] xrdp_wm_init: channel rdpdr channel id 0 is enabled [20230501-16:16:24] [DEBUG] Enabling channel 1004 (rdpdr) [20230501-16:16:24] [DEBUG] libxrdp_query_channel - Channel 1 name rdpsnd [20230501-16:16:24] [DEBUG] xrdp_wm_init: channel rdpsnd channel id 1 is enabled [20230501-16:16:24] [DEBUG] Enabling channel 1005 (rdpsnd) [20230501-16:16:24] [DEBUG] libxrdp_query_channel - Channel 2 name cliprdr [20230501-16:16:24] [DEBUG] xrdp_wm_init: channel cliprdr channel id 2 is enabled [20230501-16:16:24] [DEBUG] Enabling channel 1006 (cliprdr) [20230501-16:16:24] [DEBUG] libxrdp_query_channel - Channel 3 name drdynvc [20230501-16:16:24] [DEBUG] xrdp_wm_init: channel drdynvc channel id 3 is enabled [20230501-16:16:24] [DEBUG] Enabling channel 1007 (drdynvc) [20230501-16:16:24] [DEBUG] Login state change request WMLS_USER_PROMPT -> WMLS_START_CONNECT [20230501-16:16:24] [DEBUG] out xrdp_wm_init: [20230501-16:16:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 2 [20230501-16:16:24] [DEBUG] Login state change request WMLS_START_CONNECT -> WMLS_CONNECT_IN_PROGRESS [20230501-16:16:24] [DEBUG] xrdp_wm_log_msg: connecting to sesman on <IP_ADDRESS>:3350 [20230501-16:16:24] [INFO ] connecting to sesman on <IP_ADDRESS>:3350 [20230501-16:16:24] [INFO ] xrdp_wm_log_msg: sesman connect ok [20230501-16:16:24] [INFO ] sesman connect ok [20230501-16:16:24] [DEBUG] xrdp_wm_log_msg: sending login info to session manager. Please wait... [20230501-16:16:24] [INFO ] sending login info to session manager. Please wait... [20230501-16:16:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 3 [20230501-16:16:24] [INFO ] xrdp_wm_log_msg: login successful for user user on display 10 [20230501-16:16:24] [INFO ] login successful for user user on display 10 [20230501-16:16:24] [INFO ] loaded module 'libxup.so' ok, interface size 10296, version 4 [20230501-16:16:24] [DEBUG] xrdp_wm_log_msg: started connecting [20230501-16:16:24] [INFO ] started connecting [20230501-16:16:24] [INFO ] lib_mod_connect: connecting via UNIX socket [20230501-16:16:28] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:31] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:35] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:38] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:42] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:45] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:49] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:52] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:56] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:16:59] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:03] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:06] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:10] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:13] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:17] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:20] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:24] [DEBUG] Closed socket 19 (AF_UNIX) [20230501-16:17:27] [DEBUG] Closed socket 19 (AF_UNIX) xrdp-sesman.log: [20230501-16:16:08] [INFO ] starting xrdp-sesman with pid 2453 [20230501-16:16:24] [INFO ] Socket 12: AF_INET6 connection received from ::1 port 39962 [20230501-16:16:24] [INFO ] Terminal Server Users group is disabled, allowing authentication [20230501-16:16:24] [INFO ] ++ created session (access granted): username user, ip ::ffff:<IP_ADDRESS>:49236 - socket: 12 [20230501-16:16:24] [INFO ] starting Xorg session... [20230501-16:16:24] [INFO ] Starting session: session_pid 2466, display :10.0, width 1456, height 818, bpp 24, client ip ::ffff:<IP_ADDRESS>:49236 - socket: 12, user name user [20230501-16:16:24] [INFO ] [session start] (display 10): calling auth_start_session from pid 2466 [20230501-16:16:24] [ERROR] sesman_data_in: scp_process_msg failed [20230501-16:16:24] [ERROR] sesman_main_loop: trans_check_wait_objs failed, removing trans [20230501-16:16:24] [INFO ] Starting X server on display 10: Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log [20230501-16:16:34] [WARN ] Timed out waiting for X server on display 10 to startup [20230501-16:16:34] [INFO ] Session started successfully for user user on display 10 [20230501-16:16:34] [INFO ] Session in progress on display 10, waiting until the window manager (pid 2467) exits to end the session [20230501-16:16:34] [INFO ] Starting the xrdp channel server for display 10 [20230501-16:16:34] [WARN ] Timed out waiting for X server on display 10 to startup [20230501-16:16:34] [ERROR] There is no X server active on display 10 [20230501-16:16:34] [ERROR] A fatal error has occurred attempting to start the window manager on display 10, aborting connection [20230501-16:16:34] [WARN ] Window manager (pid 2467, display 10) exited quickly (0 secs). This could indicate a window manager config problem [20230501-16:16:34] [INFO ] Calling auth_stop_session and auth_end from pid 2466 [20230501-16:16:34] [INFO ] Terminating X server (pid 2468) on display 10 [20230501-16:16:34] [INFO ] Terminating the xrdp channel server (pid 2476) on display 10 [20230501-16:16:34] [INFO ] X server on display 10 (pid 2468) returned exit code 1 and signal number 0 [20230501-16:16:34] [INFO ] xrdp channel server for display 10 (pid 2476) exit code 0 and signal number 0 [20230501-16:16:34] [INFO ] cleanup_sockets: [20230501-16:16:34] [INFO ] Process 2466 has exited [20230501-16:16:34] [INFO ] ++ terminated session: username user, display :10.0, session_pid 2466, ip ::ffff:<IP_ADDRESS>:49236 - socket: 12 Thanks for the logs. This line is interesting:- Starting X server on display 10: Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log The Xorg command should be qualified with a PATH, or you will run the wrapper script in /usr/bin/Xorg. On Mint I believe the command should be /usr/lib/xorg/Xorg Have a look in /etc/xrdp/sesman.ini and find the [Xorg] section. Add a path to the Xorg command, restart xrdp-sesman and see what happens. Now the log file gives this [20230502-18:00:30] [INFO ] Socket 12: AF_INET6 connection received from ::1 port 53048 [20230502-18:00:30] [DEBUG] session_get_bydata: search policy 0 U user W 1456 H 818 bpp 24 T 3 IP ::ffff:<IP_ADDRESS>:49323 - socket: 12 [20230502-18:00:30] [INFO ] Terminal Server Users group is disabled, allowing authentication [20230502-18:00:30] [INFO ] ++ created session (access granted): username user, ip ::ffff:<IP_ADDRESS>:49323 - socket: 12 [20230502-18:00:30] [INFO ] starting Xorg session... [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:30] [DEBUG] Closed socket 13 (AF_INET6 :: port 5910) [20230502-18:00:30] [DEBUG] Did not find a running X server at 5910 [20230502-18:00:30] [DEBUG] Closed socket 13 (AF_INET6 :: port 6010) [20230502-18:00:30] [DEBUG] Did not find a running X server at 6010 [20230502-18:00:30] [DEBUG] Closed socket 13 (AF_INET6 :: port 6210) [20230502-18:00:30] [DEBUG] Did not find a running X server at 6210 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_socket_10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_audio_out_socket_10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_audio_in_socket_10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdpapi_10 [20230502-18:00:30] [INFO ] Starting session: session_pid 1022, display :10.0, width 1456, height 818, bpp 24, client ip ::ffff:<IP_ADDRESS>:49323 - socket: 12, user name user [20230502-18:00:30] [INFO ] [session start] (display 10): calling auth_start_session from pid 1022 [20230502-18:00:30] [ERROR] sesman_data_in: scp_process_msg failed [20230502-18:00:30] [ERROR] sesman_main_loop: trans_check_wait_objs failed, removing trans [20230502-18:00:30] [DEBUG] Closed socket 11 (AF_INET6 ::1 port 3350) [20230502-18:00:30] [DEBUG] Closed socket 12 (AF_INET6 ::1 port 3350) [20230502-18:00:30] [DEBUG] Closed socket 12 (AF_INET6 ::1 port 3350) [20230502-18:00:30] [DEBUG] Waiting for X server to start on display 10 [20230502-18:00:30] [DEBUG] Waiting for X server to start on display 10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:30] [INFO ] Starting X server on display 10: Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log [20230502-18:00:30] [DEBUG] Calling exec (executable: Xorg, arguments: Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log) [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:35] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:36] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:37] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:38] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:39] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:40] [WARN ] Timed out waiting for X server on display 10 to startup [20230502-18:00:40] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:40] [INFO ] Session started successfully for user user on display 10 [20230502-18:00:40] [INFO ] Starting the xrdp channel server for display 10 [20230502-18:00:40] [INFO ] Session in progress on display 10, waiting until the window manager (pid 1023) exits to end the session [20230502-18:00:40] [DEBUG] waiting for pid 1023 to exit [20230502-18:00:40] [DEBUG] Calling exec (executable: /usr/sbin/xrdp-chansrv, arguments: /usr/sbin/xrdp-chansrv) [20230502-18:00:41] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:41] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:41] [WARN ] Timed out waiting for X server on display 10 to startup [20230502-18:00:41] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-18:00:41] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-18:00:41] [ERROR] There is no X server active on display 10 [20230502-18:00:41] [ERROR] A fatal error has occurred attempting to start the window manager on display 10, aborting connection [20230502-18:00:41] [WARN ] Window manager (pid 1023, display 10) exited quickly (1 secs). This could indicate a window manager config problem [20230502-18:00:41] [INFO ] Calling auth_stop_session and auth_end from pid 1022 [20230502-18:00:41] [INFO ] Terminating X server (pid 1024) on display 10 [20230502-18:00:41] [INFO ] Terminating the xrdp channel server (pid 1026) on display 10 [20230502-18:00:41] [DEBUG] waiting for pid 1024 to exit [20230502-18:00:41] [INFO ] X server on display 10 (pid 1024) returned exit code 1 and signal number 0 [20230502-18:00:41] [DEBUG] waiting for pid 1026 to exit [20230502-18:00:41] [INFO ] xrdp channel server for display 10 (pid 1026) exit code 0 and signal number 0 [20230502-18:00:41] [INFO ] cleanup_sockets: [20230502-18:00:41] [DEBUG] receiving SIGCHLD [20230502-18:00:41] [INFO ] Process 1022 has exited [20230502-18:00:41] [INFO ] ++ terminated session: username user, display :10.0, session_pid 1022, ip ::ffff:<IP_ADDRESS>:49323 - socket: 12 I checked a few more options. Here is the full result of my attempts: /etc/xrdp/xrdp.ini [Globals] ; xrdp.ini file version number ini_version=1 ; fork a new process for each incoming connection fork=true ; ports to listen on, number alone means listen on all interfaces ; <IP_ADDRESS> or :: if ipv6 is configured ; space between multiple occurrences ; ALL specified interfaces must be UP when xrdp starts, otherwise xrdp will fail to start ; ; Examples: ; port=3389 ; port=unix://./tmp/xrdp.socket ; port=tcp://.:3389 <IP_ADDRESS>:3389 ; port=tcp://:3389 *:3389 ; port=tcp://<any ipv4 format addr>:3389 <IP_ADDRESS>:3389 ; port=tcp6://.:3389 ::1:3389 ; port=tcp6://:3389 *:3389 ; port=tcp6://{<any ipv6 format addr>}:3389 {FC00:0:0:0:0:0:0:1}:3389 ; port=vsock://<cid>:<port> port=3389 ; 'port' above should be connected to with vsock instead of tcp ; use this only with number alone in port above ; prefer use vsock://<cid>:<port> above use_vsock=false ; regulate if the listening socket use socket option tcp_nodelay ; no buffering will be performed in the TCP stack tcp_nodelay=true ; regulate if the listening socket use socket option keepalive ; if the network connection disappear without close messages the connection will be closed tcp_keepalive=true ; set tcp send/recv buffer (for experts) #tcp_send_buffer_bytes=32768 #tcp_recv_buffer_bytes=32768 ; security layer can be 'tls', 'rdp' or 'negotiate' ; for client compatible layer security_layer=negotiate ; minimum security level allowed for client for classic RDP encryption ; use tls_ciphers to configure TLS encryption ; can be 'none', 'low', 'medium', 'high', 'fips' crypt_level=high #crypt_level=none ; X.509 certificate and private key ; openssl req -x509 -newkey rsa:2048 -nodes -keyout key.pem -out cert.pem -days 365 certificate= key_file= ; set SSL protocols ; can be comma separated list of 'SSLv3', 'TLSv1', 'TLSv1.1', 'TLSv1.2', 'TLSv1.3' ssl_protocols=TLSv1.2, TLSv1.3 ; set TLS cipher suites #tls_ciphers=HIGH ; concats the domain name to the user if set for authentication with the separator ; for example when the server is multi homed with SSSd #domain_user_separator=@ ; The following options will override the keyboard layout settings. ; These options are for DEBUG and are not recommended for regular use. #xrdp.override_keyboard_type=0x04 #xrdp.override_keyboard_subtype=0x01 #xrdp.override_keylayout=0x00000409 ; Section name to use for automatic login if the client sends username ; and password. If empty, the domain name sent by the client is used. ; If empty and no domain name is given, the first suitable section in ; this file will be used. autorun= allow_channels=true allow_multimon=true bitmap_cache=true bitmap_compression=true bulk_compression=true #hidelogwindow=true max_bpp=32 new_cursors=true ; fastpath - can be 'input', 'output', 'both', 'none' use_fastpath=both ; when true, userid/password *must* be passed on cmd line #require_credentials=true ; when true, the userid will be used to try to authenticate #enable_token_login=true ; You can set the PAM error text in a gateway setup (MAX 256 chars) #pamerrortxt=change your password according to policy at http://url ; ; colors used by windows in RGB format ; blue=009cb5 grey=dedede #black=000000 #dark_grey=808080 #blue=08246b #dark_blue=08246b #white=ffffff #red=ff0000 #green=00ff00 #background=626c72 ; ; configure login screen ; ; Login Screen Window Title #ls_title=My Login Title ; top level window background color in RGB format ls_top_window_bg_color=009cb5 ; width and height of login screen ; ; The default height allows for about 5 fields to be comfortably displayed ; above the buttons at the bottom. To display more fields, make <ls_height> ; larger, and also increase <ls_btn_ok_y_pos> and <ls_btn_cancel_y_pos> ; below ; ls_width=350 ls_height=430 ; login screen background color in RGB format ls_bg_color=dedede ; optional background image filename. BMP format is always supported, ; but other formats will be supported if xrdp is build with imlib2 ; The transform can be one of the following:- ; none : No transformation. Image is placed in bottom-right corner ; of the screen. ; scale : Image is scaled to the screen size. The image aspect ; ratio is not preserved. ; zoom : Image is scaled to the screen size. The image aspect ; ratio is preserved by clipping the image. #ls_background_image= #ls_background_transform=none ; logo ; full path to file or file in shared folder. BMP format is always supported, ; but other formats will be supported if xrdp is build with imlib2 ; For transform values, see 'ls_background_transform'. The logo width and ; logo height are ignored for a transform of 'none'. ls_logo_filename= #ls_logo_transform=none #ls_logo_width=240 #ls_logo_height=140 ls_logo_x_pos=55 ls_logo_y_pos=50 ; for positioning labels such as username, password etc ls_label_x_pos=30 ls_label_width=65 ; for positioning text and combo boxes next to above labels ls_input_x_pos=110 ls_input_width=210 ; y pos for first label and combo box ls_input_y_pos=220 ; OK button ls_btn_ok_x_pos=142 ls_btn_ok_y_pos=370 ls_btn_ok_width=85 ls_btn_ok_height=30 ; Cancel button ls_btn_cancel_x_pos=237 ls_btn_cancel_y_pos=370 ls_btn_cancel_width=85 ls_btn_cancel_height=30 [Logging] ; Note: Log levels can be any of: core, error, warning, info, debug, or trace LogFile=xrdp.log LogLevel=debug EnableSyslog=true #SyslogLevel=INFO #EnableConsole=false #ConsoleLevel=INFO #EnableProcessId=false [LoggingPerLogger] ; Note: per logger configuration is only used if xrdp is built with ; --enable-devel-logging #xrdp.c=INFO #main()=INFO [Channels] ; Channel names not listed here will be blocked by XRDP. ; You can block any channel by setting its value to false. ; IMPORTANT! All channels are not supported in all use ; cases even if you set all values to true. ; You can override these settings on each session type ; These settings are only used if allow_channels=true rdpdr=true rdpsnd=true drdynvc=true cliprdr=true rail=true xrdpvr=true tcutils=true ; for debugging xrdp, in section xrdp1, change port=-1 to this: #port=/tmp/.xrdp/xrdp_display_10 ; ; Session types ; ; Some session types such as Xorg, X11rdp and Xvnc start a display server. ; Startup command-line parameters for the display server are configured ; in sesman.ini. See and configure also sesman.ini. [Xorg] name=Xorg lib=libxup.so username=ask password=ask ip=<IP_ADDRESS> port=-1 code=20 [Xvnc] name=Xvnc lib=libvnc.so username=ask password=ask ip=<IP_ADDRESS> port=-1 #xserverbpp=24 #delay_ms=2000 ; Disable requested encodings to support buggy VNC servers ; (1 = ExtendedDesktopSize) #disabled_encodings_mask=0 ; Use this to connect to a chansrv instance created outside of sesman ; (e.g. as part of an x11vnc console session). Replace '0' with the ; display number of the session #chansrvport=DISPLAY(0) ; Generic VNC Proxy ; Tailor this to specific hosts and VNC instances by specifying an ip ; and port and setting a suitable name. [vnc-any] name=vnc-any lib=libvnc.so ip=ask port=ask5900 username=na password=ask #pamusername=asksame #pampassword=asksame #pamsessionmng=<IP_ADDRESS> #delay_ms=2000 ; Generic RDP proxy using NeutrinoRDP ; Tailor this to specific hosts by specifying an ip and port and setting ; a suitable name. [neutrinordp-any] name=neutrinordp-any ; To use this section, you should build xrdp with configure option ; --enable-neutrinordp. lib=libxrdpneutrinordp.so ip=ask port=ask3389 username=ask password=ask ; Uncomment the following lines to enable PAM authentication for proxy ; connections. #pamusername=ask #pampassword=ask #pamsessionmng=<IP_ADDRESS> ; Currently NeutrinoRDP doesn't support dynamic resizing. Uncomment ; this line if you're using a client which does. #enable_dynamic_resizing=false ; By default, performance settings requested by the RDP client are ignored ; and chosen by NeutrinoRDP. Uncomment this line to allow the user to ; select performance settings in the RDP client. #perf.allow_client_experiencesettings=true ; Override any experience setting by uncommenting one or more of the ; following lines. #perf.wallpaper=false #perf.font_smoothing=false #perf.desktop_composition=false #perf.full_window_drag=false #perf.menu_anims=false #perf.themes=false #perf.cursor_blink=false ; By default NeutrinoRDP supports cursor shadows. If this is giving ; you problems (e.g. cursor is a black rectangle) try disabling cursor ; shadows by uncommenting the following line. #perf.cursor_shadow=false ; By default, NeutrinoRDP uses the keyboard layout of the remote RDP Server. ; If you want to tell the remote the keyboard layout of the RDP Client, ; by uncommenting the following line. #neutrinordp.allow_client_keyboardLayout=true ; The following options will override the remote keyboard layout settings. ; These options are for DEBUG and are not recommended for regular use. #neutrinordp.override_keyboardLayout_mask=0x0000FFFF #neutrinordp.override_kbd_type=0x04 #neutrinordp.override_kbd_subtype=0x01 #neutrinordp.override_kbd_fn_keys=12 #neutrinordp.override_kbd_layout=0x00000409 ; You can override the common channel settings for each session type #channel.rdpdr=true #channel.rdpsnd=true #channel.drdynvc=true #channel.cliprdr=true #channel.rail=true #channel.xrdpvr=true /etc/xrdp/sesman.ini ;; See `man 5 sesman.ini` for details [Globals] ListenAddress=<IP_ADDRESS> ListenPort=3350 EnableUserWindowManager=true ; Give in relative path to user's home directory UserWindowManager=startwm.sh ; Give in full path or relative path to /etc/xrdp DefaultWindowManager=startwm.sh ; Give in full path or relative path to /etc/xrdp ReconnectScript=reconnectwm.sh [Security] AllowRootLogin=true MaxLoginRetry=4 TerminalServerUsers=tsusers TerminalServerAdmins=tsadmins ; When AlwaysGroupCheck=false access will be permitted ; if the group TerminalServerUsers is not defined. AlwaysGroupCheck=false ; When RestrictOutboundClipboard=all clipboard from the ; server is not pushed to the client. ; In addition, you can control text/file/image transfer restrictions ; respectively. It also accepts comma separated list such as text,file,image. ; To keep compatibility, some aliases are also available: ; true: an alias of all ; false: an alias of none ; yes: an alias of all RestrictOutboundClipboard=none ; When RestrictInboundClipboard=all clipboard from the ; client is not pushed to the server. ; In addition, you can control text/file/image transfer restrictions ; respectively. It also accepts comma separated list such as text,file,image. ; To keep compatibility, some aliases are also available: ; true: an alias of all ; false: an alias of none ; yes: an alias of all RestrictInboundClipboard=none [Sessions] ;; X11DisplayOffset - x11 display number offset ; Type: integer ; Default: 10 X11DisplayOffset=10 ;; MaxSessions - maximum number of connections to an xrdp server ; Type: integer ; Default: 0 MaxSessions=50 ;; KillDisconnected - kill disconnected sessions ; Type: boolean ; Default: false ; if 1, true, or yes, every session will be killed within DisconnectedTimeLimit ; seconds after the user disconnects KillDisconnected=false ;; DisconnectedTimeLimit (seconds) - wait before kill disconnected sessions ; Type: integer ; Default: 0 ; if KillDisconnected is set to false, this value is ignored DisconnectedTimeLimit=0 ;; IdleTimeLimit (seconds) - wait before disconnect idle sessions ; Type: integer ; Default: 0 ; Set to 0 to disable idle disconnection. IdleTimeLimit=0 ;; Policy - session allocation policy ; Type: enum [ "Default" | "UBD" | "UBI" | "UBC" | "UBDI" | "UBDC" ] ; "Default" session per <User,BitPerPixel> ; "UBD" session per <User,BitPerPixel,DisplaySize> ; "UBI" session per <User,BitPerPixel,IPAddr> ; "UBC" session per <User,BitPerPixel,Connection> ; "UBDI" session per <User,BitPerPixel,DisplaySize,IPAddr> ; "UBDC" session per <User,BitPerPixel,DisplaySize,Connection> Policy=Default [Logging] ; Note: Log levels can be any of: core, error, warning, info, debug, or trace LogFile=xrdp-sesman.log LogLevel=debug EnableSyslog=true #SyslogLevel=INFO #EnableConsole=false #ConsoleLevel=INFO #EnableProcessId=false [LoggingPerLogger] ; Note: per logger configuration is only used if xrdp is built with ; --enable-devel-logging #sesman.c=INFO #main()=INFO ; ; Session definitions - startup command-line parameters for each session type ; [Xorg] ; Specify the path of non-suid Xorg executable. It might differ depending ; on your distribution and version. Find out the appropriate path for your ; environment. The typical path is known as follows: ; ; Fedora 26 or later : param=/usr/libexec/Xorg ; Debian 9 or later : param=/usr/lib/xorg/Xorg ; Ubuntu 16.04 or later : param=/usr/lib/xorg/Xorg ; Arch Linux : param=/usr/lib/Xorg ; CentOS 7 : param=/usr/bin/Xorg or param=Xorg ; CentOS 8 : param=/usr/libexec/Xorg ; FreeBSD (from 2022Q4) : param=/usr/local/libexec/Xorg ; #param=Xorg param=/usr/bin/Xorg ; Leave the rest parameters as-is unless you understand what will happen. param=-config param=xrdp/xorg.conf param=-noreset param=-nolisten param=tcp param=-logfile param=.xorgxrdp.%s.log [Xvnc] param=Xvnc param=-bs param=-nolisten param=tcp param=-localhost param=-dpi param=96 [Chansrv] ; drive redirection ; See sesman.ini(5) for the format of this parameter #FuseMountName=/run/user/%u/thinclient_drives #FuseMountName=/media/thinclient_drives/%U/thinclient_drives FuseMountName=thinclient_drives ; this value allows only the user to access their own mapped drives. ; Make this more permissive (e.g. 022) if required. FileUmask=077 ; Can be used to disable FUSE functionality - see sesman.ini(5) #EnableFuseMount=false ; Uncomment this line only if you are using GNOME 3 versions 3.29.92 ; and up, and you wish to cut-paste files between Nautilus and Windows. Do ; not use this setting for GNOME 4, or other file managers #UseNautilus3FlistFormat=true [ChansrvLogging] ; Note: one log file is created per display and the LogFile config value ; is ignored. The channel server log file names follow the naming convention: ; xrdp-chansrv.${DISPLAY}.log ; ; Note: Log levels can be any of: core, error, warning, info, debug, or trace LogLevel=INFO EnableSyslog=true #SyslogLevel=INFO #EnableConsole=false #ConsoleLevel=INFO #EnableProcessId=false [ChansrvLoggingPerLogger] ; Note: per logger configuration is only used if xrdp is built with ; --enable-devel-logging #chansrv.c=INFO #main()=INFO [SessionVariables] PULSE_SCRIPT=/etc/xrdp/pulse/default.pa xrdp.log [20230502-22:32:22] [INFO ] Socket 12: AF_INET6 connection received from ::ffff:<IP_ADDRESS> port 46378 [20230502-22:32:22] [DEBUG] Closed socket 12 (AF_INET6 ::ffff:<IP_ADDRESS> port 3389) [20230502-22:32:22] [DEBUG] Closed socket 11 (AF_INET6 :: port 3389) [20230502-22:32:22] [DEBUG] item ini_version, value 1 [20230502-22:32:23] [DEBUG] item fork, value true [20230502-22:32:23] [DEBUG] item port, value 3389 [20230502-22:32:23] [DEBUG] item use_vsock, value false [20230502-22:32:23] [DEBUG] item tcp_nodelay, value true [20230502-22:32:23] [DEBUG] item tcp_keepalive, value true [20230502-22:32:23] [DEBUG] item security_layer, value negotiate [20230502-22:32:23] [DEBUG] item crypt_level, value high [20230502-22:32:23] [DEBUG] item certificate, value [20230502-22:32:23] [INFO ] Using default X.509 certificate: /etc/xrdp/cert.pem [20230502-22:32:23] [DEBUG] item key_file, value [20230502-22:32:23] [INFO ] Using default X.509 key file: /etc/xrdp/key.pem [20230502-22:32:23] [DEBUG] item ssl_protocols, value TLSv1.2, TLSv1.3 [20230502-22:32:23] [DEBUG] TLSv1.3 enabled [20230502-22:32:23] [DEBUG] TLSv1.2 enabled [20230502-22:32:23] [DEBUG] item autorun, value [20230502-22:32:23] [DEBUG] item allow_channels, value true [20230502-22:32:23] [DEBUG] item allow_multimon, value true [20230502-22:32:23] [DEBUG] item bitmap_cache, value true [20230502-22:32:23] [DEBUG] item bitmap_compression, value true [20230502-22:32:23] [DEBUG] item bulk_compression, value true [20230502-22:32:23] [DEBUG] item max_bpp, value 32 [20230502-22:32:23] [DEBUG] item new_cursors, value true [20230502-22:32:23] [DEBUG] item use_fastpath, value both [20230502-22:32:23] [DEBUG] item blue, value 009cb5 [20230502-22:32:23] [DEBUG] item grey, value dedede [20230502-22:32:23] [DEBUG] item ls_top_window_bg_color, value 009cb5 [20230502-22:32:23] [DEBUG] item ls_width, value 350 [20230502-22:32:23] [DEBUG] item ls_height, value 430 [20230502-22:32:23] [DEBUG] item ls_bg_color, value dedede [20230502-22:32:23] [DEBUG] item ls_logo_filename, value [20230502-22:32:23] [DEBUG] item ls_logo_x_pos, value 55 [20230502-22:32:23] [DEBUG] item ls_logo_y_pos, value 50 [20230502-22:32:23] [DEBUG] item ls_label_x_pos, value 30 [20230502-22:32:23] [DEBUG] item ls_label_width, value 65 [20230502-22:32:23] [DEBUG] item ls_input_x_pos, value 110 [20230502-22:32:23] [DEBUG] item ls_input_width, value 210 [20230502-22:32:23] [DEBUG] item ls_input_y_pos, value 220 [20230502-22:32:23] [DEBUG] item ls_btn_ok_x_pos, value 142 [20230502-22:32:23] [DEBUG] item ls_btn_ok_y_pos, value 370 [20230502-22:32:23] [DEBUG] item ls_btn_ok_width, value 85 [20230502-22:32:23] [DEBUG] item ls_btn_ok_height, value 30 [20230502-22:32:23] [DEBUG] item ls_btn_cancel_x_pos, value 237 [20230502-22:32:23] [DEBUG] item ls_btn_cancel_y_pos, value 370 [20230502-22:32:23] [DEBUG] item ls_btn_cancel_width, value 85 [20230502-22:32:23] [DEBUG] item ls_btn_cancel_height, value 30 [20230502-22:32:23] [INFO ] Security protocol: configured [SSL|RDP], requested [SSL|HYBRID|RDP], selected [SSL] [20230502-22:32:23] [DEBUG] Using TLS security, and setting RDP security crypto to LEVEL_NONE and METHOD_NONE [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] receive connection request [20230502-22:32:23] [INFO ] Connected client computer name: solus [20230502-22:32:23] [DEBUG] The connection is using TLS, skipping RDP crypto negotiation [20230502-22:32:23] [DEBUG] Adding channel: name rdpdr, channel id 1004, flags 0xc0800000 [20230502-22:32:23] [DEBUG] Adding channel: name rdpsnd, channel id 1005, flags 0xc0000000 [20230502-22:32:23] [DEBUG] Adding channel: name cliprdr, channel id 1006, flags 0xc0a00000 [20230502-22:32:23] [DEBUG] Adding channel: name drdynvc, channel id 1007, flags 0xc0800000 [20230502-22:32:23] [WARN ] Received [MS-RDPBCGR] TS_UD_HEADER type 0xc006 is unknown (ignored) [20230502-22:32:23] [WARN ] Received [MS-RDPBCGR] TS_UD_HEADER type 0xc00a is unknown (ignored) [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] construct connection response [20230502-22:32:23] [DEBUG] using no security [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] send connection response [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] receive erect domain request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] receive attach user request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence] send attach user confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] receive channel join request [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] send channel join confirm [20230502-22:32:23] [DEBUG] [MCS Connection Sequence (TLS)] completed [20230502-22:32:23] [INFO ] xrdp_load_keyboard_layout: Keyboard information sent by the RDP client, keyboard_type:[0x04], keyboard_subtype:[0x00], keylayout:[0x00000409] [20230502-22:32:23] [DEBUG] keyboard_cfg_file /etc/xrdp/xrdp_keyboard.ini [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value 0x00000409 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value us [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item keyboard_type value 4 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item keyboard_subtype value 3 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item keyboard_type value 7 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item keyboard_subtype value 2 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item model value pc105 [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item rdp_layouts value default_rdp_layouts [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item layouts_map value default_layouts_map [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: item rdp_layout_us value us [20230502-22:32:23] [DEBUG] xrdp_load_keyboard_layout: skipping configuration item - rdp_layout_us, continuing to next section [20230502-22:32:23] [INFO ] xrdp_load_keyboard_layout: model [] variant [] layout [us] options [] [20230502-22:32:23] [INFO ] TLS connection established from ::ffff:<IP_ADDRESS> port 46378: TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 [20230502-22:32:23] [DEBUG] Client requested auto logon. [20230502-22:32:23] [DEBUG] Client requested compression enabled. [20230502-22:32:23] [DEBUG] Client supplied domain: [20230502-22:32:23] [DEBUG] Client supplied username: alexander [20230502-22:32:23] [DEBUG] Client supplied password: <omitted from log> [20230502-22:32:23] [DEBUG] Client supplied program: [20230502-22:32:23] [DEBUG] Client supplied directory: [20230502-22:32:24] [INFO ] xrdp_caps_process_pointer: client supports new(color) cursor [20230502-22:32:24] [INFO ] xrdp_process_offscreen_bmpcache: support level 1 cache size 7864320 MB cache entries 2000 [20230502-22:32:24] [INFO ] xrdp_caps_process_codecs: RemoteFX, codec id 3, properties len 49 [20230502-22:32:24] [WARN ] Client Capability: not enough orders supported by client, client wants off screen bitmap but offscreen bitmaps disabled [20230502-22:32:24] [DEBUG] xrdp_00000738_wm_login_state_event_00000001 [20230502-22:32:24] [INFO ] Loading keymap file /etc/xrdp/km-00000409.ini [20230502-22:32:24] [WARN ] local keymap file for 0x00000409 found and doesn't match built in keymap, using local keymap file [20230502-22:32:24] [DEBUG] Login state change request WMLS_RESET -> WMLS_RESET [20230502-22:32:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 0 [20230502-22:32:24] [DEBUG] Login state change request WMLS_RESET -> WMLS_USER_PROMPT [20230502-22:32:24] [DEBUG] in xrdp_wm_init: [20230502-22:32:24] [DEBUG] ini_version: 1 [20230502-22:32:24] [DEBUG] use_bitmap_cache: 1 [20230502-22:32:24] [DEBUG] use_bitmap_compression: 1 [20230502-22:32:24] [DEBUG] port: 3389 [20230502-22:32:24] [DEBUG] crypt_level: 3 [20230502-22:32:24] [DEBUG] allow_channels: 1 [20230502-22:32:24] [DEBUG] max_bpp: 32 [20230502-22:32:24] [DEBUG] fork: 1 [20230502-22:32:24] [DEBUG] tcp_nodelay: 1 [20230502-22:32:24] [DEBUG] tcp_keepalive: 1 [20230502-22:32:24] [DEBUG] tcp_send_buffer_bytes: 0 [20230502-22:32:24] [DEBUG] tcp_recv_buffer_bytes: 0 [20230502-22:32:24] [DEBUG] new_cursors: 1 [20230502-22:32:24] [DEBUG] allow_multimon: 1 [20230502-22:32:24] [DEBUG] grey: 14606046 [20230502-22:32:24] [DEBUG] black: 0 [20230502-22:32:24] [DEBUG] dark_grey: 0 [20230502-22:32:24] [DEBUG] blue: 40117 [20230502-22:32:24] [DEBUG] dark_blue: 0 [20230502-22:32:24] [DEBUG] white: 0 [20230502-22:32:24] [DEBUG] red: 0 [20230502-22:32:24] [DEBUG] green: 0 [20230502-22:32:24] [DEBUG] background: 0 [20230502-22:32:24] [DEBUG] autorun: [20230502-22:32:24] [DEBUG] hidelogwindow: 0 [20230502-22:32:24] [DEBUG] require_credentials: 0 [20230502-22:32:24] [DEBUG] bulk_compression: 1 [20230502-22:32:24] [DEBUG] new_cursors: 1 [20230502-22:32:24] [DEBUG] nego_sec_layer: 0 [20230502-22:32:24] [DEBUG] allow_multimon: 1 [20230502-22:32:24] [DEBUG] enable_token_login: 0 [20230502-22:32:24] [DEBUG] ls_top_window_bg_color: b59c00 [20230502-22:32:24] [DEBUG] ls_width: 350 [20230502-22:32:24] [DEBUG] ls_height: 430 [20230502-22:32:24] [DEBUG] ls_bg_color: dedede [20230502-22:32:24] [DEBUG] ls_title: [20230502-22:32:24] [DEBUG] ls_logo_filename: [20230502-22:32:24] [DEBUG] ls_logo_x_pos: 55 [20230502-22:32:24] [DEBUG] ls_logo_y_pos: 50 [20230502-22:32:24] [DEBUG] ls_label_x_pos: 30 [20230502-22:32:24] [DEBUG] ls_label_width: 65 [20230502-22:32:24] [DEBUG] ls_input_x_pos: 110 [20230502-22:32:24] [DEBUG] ls_input_width: 210 [20230502-22:32:24] [DEBUG] ls_input_y_pos: 220 [20230502-22:32:24] [DEBUG] ls_btn_ok_x_pos: 142 [20230502-22:32:24] [DEBUG] ls_btn_ok_y_pos: 370 [20230502-22:32:24] [DEBUG] ls_btn_ok_width: 85 [20230502-22:32:24] [DEBUG] ls_btn_ok_height: 30 [20230502-22:32:24] [DEBUG] ls_btn_cancel_x_pos: 237 [20230502-22:32:24] [DEBUG] ls_btn_cancel_y_pos: 370 [20230502-22:32:24] [DEBUG] ls_btn_cancel_width: 85 [20230502-22:32:24] [DEBUG] ls_btn_cancel_height: 30 [20230502-22:32:24] [DEBUG] libxrdp_query_channel - Channel 0 name rdpdr [20230502-22:32:24] [DEBUG] xrdp_wm_init: channel rdpdr channel id 0 is enabled [20230502-22:32:24] [DEBUG] Enabling channel 1004 (rdpdr) [20230502-22:32:24] [DEBUG] libxrdp_query_channel - Channel 1 name rdpsnd [20230502-22:32:24] [DEBUG] xrdp_wm_init: channel rdpsnd channel id 1 is enabled [20230502-22:32:24] [DEBUG] Enabling channel 1005 (rdpsnd) [20230502-22:32:24] [DEBUG] libxrdp_query_channel - Channel 2 name cliprdr [20230502-22:32:24] [DEBUG] xrdp_wm_init: channel cliprdr channel id 2 is enabled [20230502-22:32:24] [DEBUG] Enabling channel 1006 (cliprdr) [20230502-22:32:24] [DEBUG] libxrdp_query_channel - Channel 3 name drdynvc [20230502-22:32:24] [DEBUG] xrdp_wm_init: channel drdynvc channel id 3 is enabled [20230502-22:32:24] [DEBUG] Enabling channel 1007 (drdynvc) [20230502-22:32:24] [DEBUG] Login state change request WMLS_USER_PROMPT -> WMLS_START_CONNECT [20230502-22:32:24] [DEBUG] out xrdp_wm_init: [20230502-22:32:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 2 [20230502-22:32:24] [DEBUG] Login state change request WMLS_START_CONNECT -> WMLS_CONNECT_IN_PROGRESS [20230502-22:32:24] [DEBUG] xrdp_wm_log_msg: connecting to sesman on <IP_ADDRESS>:3350 [20230502-22:32:24] [INFO ] connecting to sesman on <IP_ADDRESS>:3350 [20230502-22:32:24] [INFO ] xrdp_wm_log_msg: sesman connect ok [20230502-22:32:24] [INFO ] sesman connect ok [20230502-22:32:24] [DEBUG] xrdp_wm_log_msg: sending login info to session manager. Please wait... [20230502-22:32:24] [INFO ] sending login info to session manager. Please wait... [20230502-22:32:24] [DEBUG] xrdp_wm_login_mode_changed: login_mode is 3 [20230502-22:32:24] [INFO ] xrdp_wm_log_msg: login successful for user alexander on display 10 [20230502-22:32:24] [INFO ] login successful for user alexander on display 10 [20230502-22:32:24] [INFO ] loaded module 'libxup.so' ok, interface size 10296, version 4 [20230502-22:32:24] [DEBUG] xrdp_wm_log_msg: started connecting [20230502-22:32:24] [INFO ] started connecting [20230502-22:32:24] [INFO ] lib_mod_connect: connecting via UNIX socket [20230502-22:32:27] [DEBUG] Closed socket 19 (AF_UNIX) [20230502-22:32:31] [DEBUG] Closed socket 19 (AF_UNIX) [20230502-22:32:34] [DEBUG] Closed socket 19 (AF_UNIX) [20230502-22:32:38] [DEBUG] Closed socket 19 (AF_UNIX) xrdp-sesman.log [20230502-22:32:24] [INFO ] Socket 12: AF_INET6 connection received from ::1 port 55930 [20230502-22:32:24] [DEBUG] session_get_bydata: search policy 0 U alexander W 2516 H 984 bpp 24 T 3 IP ::ffff:<IP_ADDRESS>:46378 - socket: 12 [20230502-22:32:24] [INFO ] Terminal Server Users group is disabled, allowing authentication [20230502-22:32:24] [INFO ] ++ created session (access granted): username alexander, ip ::ffff:<IP_ADDRESS>:46378 - socket: 12 [20230502-22:32:24] [INFO ] starting Xorg session... [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Closed socket 13 (AF_INET6 :: port 5910) [20230502-22:32:24] [DEBUG] Did not find a running X server at 5910 [20230502-22:32:24] [DEBUG] Closed socket 13 (AF_INET6 :: port 6010) [20230502-22:32:24] [DEBUG] Did not find a running X server at 6010 [20230502-22:32:24] [DEBUG] Closed socket 13 (AF_INET6 :: port 6210) [20230502-22:32:24] [DEBUG] Did not find a running X server at 6210 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_socket_10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_audio_out_socket_10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdp_chansrv_audio_in_socket_10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.xrdp/xrdpapi_10 [20230502-22:32:24] [INFO ] Starting session: session_pid 1852, display :10.0, width 2516, height 984, bpp 24, client ip ::ffff:<IP_ADDRESS>:46378 - socket: 12, user name alexander [20230502-22:32:24] [INFO ] [session start] (display 10): calling auth_start_session from pid 1852 [20230502-22:32:24] [ERROR] sesman_data_in: scp_process_msg failed [20230502-22:32:24] [DEBUG] Closed socket 11 (AF_INET6 ::1 port 3350) [20230502-22:32:24] [ERROR] sesman_main_loop: trans_check_wait_objs failed, removing trans [20230502-22:32:24] [DEBUG] Closed socket 12 (AF_INET6 ::1 port 3350) [20230502-22:32:24] [DEBUG] Closed socket 12 (AF_INET6 ::1 port 3350) [20230502-22:32:24] [DEBUG] Waiting for X server to start on display 10 [20230502-22:32:24] [DEBUG] Waiting for X server to start on display 10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [INFO ] Starting X server on display 10: /usr/bin/Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log [20230502-22:32:24] [DEBUG] Calling exec (executable: /usr/bin/Xorg, arguments: /usr/bin/Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log) [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:24] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:25] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:26] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:27] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:28] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:29] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:30] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:31] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:32] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:33] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [WARN ] Timed out waiting for X server on display 10 to startup [20230502-22:32:34] [INFO ] Session started successfully for user alexander on display 10 [20230502-22:32:34] [INFO ] Session in progress on display 10, waiting until the window manager (pid 1853) exits to end the session [20230502-22:32:34] [INFO ] Starting the xrdp channel server for display 10 [20230502-22:32:34] [DEBUG] waiting for pid 1853 to exit [20230502-22:32:34] [DEBUG] Calling exec (executable: /usr/sbin/xrdp-chansrv, arguments: /usr/sbin/xrdp-chansrv) [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [WARN ] Timed out waiting for X server on display 10 to startup [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X11-unix/X10 [20230502-22:32:34] [DEBUG] Did not find a running X server at /tmp/.X10-lock [20230502-22:32:34] [ERROR] There is no X server active on display 10 [20230502-22:32:34] [ERROR] A fatal error has occurred attempting to start the window manager on display 10, aborting connection [20230502-22:32:34] [WARN ] Window manager (pid 1853, display 10) exited quickly (0 secs). This could indicate a window manager config problem [20230502-22:32:34] [INFO ] Calling auth_stop_session and auth_end from pid 1852 [20230502-22:32:34] [INFO ] Terminating X server (pid 1854) on display 10 [20230502-22:32:34] [INFO ] Terminating the xrdp channel server (pid 1879) on display 10 [20230502-22:32:34] [DEBUG] waiting for pid 1854 to exit [20230502-22:32:34] [INFO ] X server on display 10 (pid 1854) returned exit code 1 and signal number 0 [20230502-22:32:34] [DEBUG] waiting for pid 1879 to exit [20230502-22:32:34] [INFO ] xrdp channel server for display 10 (pid 1879) exit code 0 and signal number 0 [20230502-22:32:34] [INFO ] cleanup_sockets: [20230502-22:32:34] [DEBUG] receiving SIGCHLD [20230502-22:32:34] [INFO ] Process 1852 has exited [20230502-22:32:34] [INFO ] ++ terminated session: username alexander, display :10.0, session_pid 1852, ip ::ffff:<IP_ADDRESS>:46378 - socket: 12 Just in case, I made these settings: /etc/X11/xrdp/xrdp.conf Section "ServerLayout" Identifier "X11 Server" Screen "Screen (xrdpdev)" InputDevice "xrdpMouse" "CorePointer" InputDevice "xrdpKeyboard" "CoreKeyboard" EndSection Section "ServerFlags" # This line prevents "ServerLayout" sections in xorg.conf.d files # overriding the "X11 Server" layout (xrdp #1784) Option "DefaultServerLayout" "X11 Server" Option "DontVTSwitch" "on" Option "AutoAddDevices" "off" Option "AutoAddGPU" "off" EndSection Section "Module" Load "dbe" Load "ddc" Load "extmod" Load "glx" Load "int10" Load "record" Load "vbe" Load "xorgxrdp" Load "fb" EndSection Section "InputDevice" Identifier "xrdpKeyboard" Driver "xrdpkeyb" EndSection Section "InputDevice" Identifier "xrdpMouse" Driver "xrdpmouse" EndSection Section "Monitor" Identifier "Monitor" Option "DPMS" HorizSync 30-80 VertRefresh 60-75 ModeLine "1920x1080"<PHONE_NUMBER> 1968 2000 2080 1080 1083 1088 1111 +hsync -vsync ModeLine "1280x720" 74.25 1280 1720 1760 1980 720 725 730 750 +HSync +VSync Modeline "1368x768" 72.25 1368 1416 1448 1528 768 771 781 790 +hsync -vsync Modeline "1600x900" 119.00 1600 1696 1864 2128 900 901 904 932 -hsync +vsync EndSection Section "Device" Identifier "Video Card (xrdpdev)" Driver "xrdpdev" Option "DRMDevice" "/dev/dri/renderD128" Option "DRI3" "1" EndSection Section "Screen" Identifier "Screen (xrdpdev)" Device "Video Card (xrdpdev)" GPUDevice "" Monitor "Monitor" DefaultDepth 24 SubSection "Display" Depth 24 Modes "640x480" "800x600" "1024x768" "1280x720" "1280x1024" "1600x900" "1920x1080" EndSubSection EndSection But for some reason the file is ~/.xorgxrdp.10.log is missing I checked a few more options. Here is the full result of my attempts: /etc/xrdp/xrdp.ini /etc/xrdp/sesman.ini xrdp.log xrdp-sesman.log Just in case, I made these settings: /etc/X11/xrdp/xrdp.conf /etc/xrdp/startwm.sh But for some reason the file is ~/.xorgxrdp.10.log is missing Hm... This solution helped me. For some reason, I didn't find any mention of this configuration file. How true is everything in such a decision? A couple of things:- In your second log (this post), you've still not got a qualified path to Xorg. This probably means you didn't successfully restart sesman at the time. The path in sesman.ini is incorrect. Where you have param=/usr/bin/Xorg, you should have param=/usr/lib/xorg/Xorg. Adding the Xorg in the standard PATH won't work. To see why, try cat /usr/bin/Xorg. This can be seen in your 3rd sesman.log above. You should get to a state where you can see this in your log:- [INFO ] Starting X server on display 10: /usr/lib/xorg/Xorg :10 -auth .Xauthority -config xrdp/xorg.conf -noreset -nolisten tcp -logfile .xorgxrdp.%s.log Your linked solution doesn't seem relevant to the problem as you've posted it. @matt335672 I apologize for the very late response. I have now checked all my settings and did as you indicated and the problem has been solved for me. > $ cat /usr/bin/Xorg #!/bin/sh # # Execute Xorg.wrap if it exists otherwise execute Xorg directly. # This allows distros to put the suid wrapper in a separate package. basedir="/usr/lib64/xorg-server" if [ -x "$basedir"/Xorg.wrap ]; then exec "$basedir"/Xorg.wrap "$@" else exec "$basedir"/Xorg "$@" fi > $ ls /usr/lib64/xorg-server/Xorg /usr/lib64/xorg-server/Xorg And as a path, I specified the full path in sesman.ini: path=/usr/lib64/xorg-server/Xorg [root@iaas-kunpeng-bljcs-02 bwda]# cat .xorgxrdp.10.log [ 3160.803] X.Org X Server 1.20.8 X Protocol Version 11, Revision 0 [ 3160.803] Build Operating System: localhost 4.19.90-23.0.v2101.ky10.aarch64 [ 3160.803] Current Operating System: Linux iaas-kunpeng-bljcs-02 4.19.90-24.4.v2101.ky10.aarch64 #1 SMP Mon May 24 14:45:37 CST 2021 aarch64 [ 3160.803] Kernel command line: BOOT_IMAGE=/vmlinuz-4.19.90-24.4.v2101.ky10.aarch64 root=/dev/mapper/system-lv_root ro biosdevname=0 net.ifnames=0 console=ttyS0,115200n8 crashkernel=1024M,high smmu.bypassdev=0x1000:0x17 smmu.bypassdev=0x1000:0x15 video=efifb:off video=VGA-1:640x480-32@60me [ 3160.803] Build Date: 27 February 2021 05:07:25PM [ 3160.803] Build ID: xorg-x11-server 1.20.8-3.p01.ky10 [ 3160.803] Current version of pixman: 0.40.0 [ 3160.803] Before reporting problems, check http://wiki.x.org to make sure that you have the latest version. [ 3160.803] Markers: (--) probed, () from config file, (==) default setting, (++) from command line, (!!) notice, (II) informational, (WW) warning, (EE) error, (NI) not implemented, (??) unknown. [ 3160.803] (++) Log file: ".xorgxrdp.10.log", Time: Wed May 24 10:46:06 2023 [ 3160.803] (EE) Unable to locate/open config file: "xrdp/xorg.conf" [ 3160.803] (==) Using config directory: "/etc/X11/xorg.conf.d" [ 3160.803] (==) Using system config directory "/usr/share/X11/xorg.conf.d" [ 3160.804] (==) No Layout section. Using the first Screen section. [ 3160.804] (==) No screen section available. Using defaults. [ 3160.804] () |-->Screen "Default Screen Section" (0) [ 3160.804] (**) | |-->Monitor "" [ 3160.804] (==) No monitor specified for screen "Default Screen Section". Using a default monitor configuration. [ 3160.804] (==) Automatically adding devices [ 3160.804] (==) Automatically enabling devices [ 3160.804] (==) Automatically adding GPU devices [ 3160.804] (==) Automatically binding GPU devices [ 3160.804] (==) Max clients allowed: 256, resource mask: 0x1fffff [ 3160.804] (==) FontPath set to: catalogue:/etc/X11/fontpath.d, built-ins [ 3160.804] (==) ModulePath set to "/usr/lib64/xorg/modules" [ 3160.804] (II) The server relies on udev to provide the list of input devices. If no devices become available, reconfigure udev or disable AutoAddDevices. [ 3160.804] (II) Loader magic: 0xaaab76d30dd8 [ 3160.804] (II) Module ABI versions: [ 3160.804] X.Org ANSI C Emulation: 0.4 [ 3160.804] X.Org Video Driver: 24.1 [ 3160.804] X.Org XInput driver : 24.1 [ 3160.804] X.Org Server Extension : 10.0 [ 3160.805] (II) systemd-logind: took control of session /org/freedesktop/login1/session/c8 [ 3160.806] (II) xfree86: Adding drm device (/dev/dri/card0) [ 3160.806] (EE) systemd-logind: failed to take device /dev/dri/card0: Operation not permitted [ 3160.806] (EE) /dev/dri/card0: failed to set DRM interface version 1.4: Permission denied [ 3160.811] (--) PCI:*(6@0:0:0) 19e5:1711:19e5:1711 rev 1, Mem @ 0xe4000000/33554432, 0xe7200000/2097152, BIOS @ 0x????????/65536 [ 3160.811] (II) LoadModule: "glx" [ 3160.811] (II) Loading /usr/lib64/xorg/modules/extensions/libglx.so [ 3160.813] (II) Module glx: vendor="X.Org Foundation" [ 3160.813] compiled for 1.20.8, module version = 1.0.0 [ 3160.813] ABI class: X.Org Server Extension, version 10.0 [ 3160.813] (==) Matched modesetting as autoconfigured driver 0 [ 3160.813] (==) Matched fbdev as autoconfigured driver 1 [ 3160.813] (==) Assigned the driver to the xf86ConfigLayout [ 3160.813] (II) LoadModule: "modesetting" [ 3160.813] (II) Loading /usr/lib64/xorg/modules/drivers/modesetting_drv.so [ 3160.813] (II) Module modesetting: vendor="X.Org Foundation" [ 3160.813] compiled for 1.20.8, module version = 1.20.8 [ 3160.813] Module class: X.Org Video Driver [ 3160.813] ABI class: X.Org Video Driver, version 24.1 [ 3160.813] (II) LoadModule: "fbdev" [ 3160.813] (II) Loading /usr/lib64/xorg/modules/drivers/fbdev_drv.so [ 3160.813] (II) Module fbdev: vendor="X.Org Foundation" [ 3160.813] compiled for 1.20.8, module version = 0.5.0 [ 3160.813] Module class: X.Org Video Driver [ 3160.813] ABI class: X.Org Video Driver, version 24.1 [ 3160.813] (II) modesetting: Driver for Modesetting Kernel Drivers: kms [ 3160.813] (II) FBDEV: driver for framebuffer: fbdev [ 3160.813] (EE) Fatal server error: [ 3160.813] (EE) parse_vt_settings: Cannot open /dev/tty0 (Permission denied) [ 3160.813] (EE) [ 3160.813] (EE) Please consult the openEuler Project support at http://wiki.x.org for help. [ 3160.813] (EE) Please also check the log file at ".xorgxrdp.10.log" for additional information. [ 3160.813] (EE) [ 3160.813] (WW) xf86CloseConsole: KDSETMODE failed: Bad file descriptor [ 3160.813] (WW) xf86CloseConsole: VT_GETMODE failed: Bad file descriptor [ 3160.813] (EE) Server terminated with error (1). Closing log file.
2025-04-01T06:39:44.643966
2023-11-03T18:13:29
1976716590
{ "authors": [ "cpz", "es3n1n", "saul" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8901", "repo": "neverlosecc/source2gen", "url": "https://github.com/neverlosecc/source2gen/issues/19" }
gharchive/issue
Project license Hi - what is this code licensed under? I'd like to fork it for CS2 reverse engineering purposes. Many thanks Hey and sorry, we completely forgot about the license. 1063e42a9e25e4b290e52d0222bf33b3c6b4fb10 Many thanks for the quick update. One question - the underlying library that you forked from is MIT licensed. I believe it requires that license acknowledged in your copy (this is how I did it: https://github.com/saul/CS2SchemaGen/blob/main/ACKNOWLEDGEMENTS) Many thanks for the quick update. One question - the underlying library that you forked from is MIT licensed. I believe it requires that license acknowledged in your copy (this is how I did it: https://github.com/saul/CS2SchemaGen/blob/main/ACKNOWLEDGEMENTS) Thank you very much! We did what you said, it took a bit more time as I was doing some reverse for schema at the time.
2025-04-01T06:39:44.654283
2012-02-10T16:07:04
3173922
{ "authors": [ "nevir", "svnlto" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8902", "repo": "nevir/groc", "url": "https://github.com/nevir/groc/issues/22" }
gharchive/issue
use within other scripts hey there, I'm writing a grunt task that generates docs and I was thinking of using groc. I was wondering if it's possible to use groc within another script as opposed to on the CLI. I have a very rough API that you could use, though it's definitely going to be a moving target for a while. (I'm totally happy to work with you to solidify it over time though :)) Take a look at where the CLI code builds a Project object for the low level API. Alternatively, you can fake a command line call w/o the shell as an intermediary by including the CLI module directly, and calling it similar to the bin file - this interface is less likely to change, but is definitely more awkward. Thinking further on this, a better and probably relatively stable API would be to decouple the argument parsing, and have a call that takes the resultant options hash - easy to configure, and also plays nicely with the .groc.json file
2025-04-01T06:39:44.657116
2022-05-03T03:20:30
1223617008
{ "authors": [ "jmf61" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8903", "repo": "new-meta-incorporated/pokemon-showdown", "url": "https://github.com/new-meta-incorporated/pokemon-showdown/pull/35" }
gharchive/pull-request
Minor balance changes also cleaned duplicates in learnsets.ts I was removing duplicate moves. Round was a mistake though, so I added it back
2025-04-01T06:39:44.660653
2015-03-17T19:04:58
62487578
{ "authors": [ "Ryuno-Ki", "newfolder0" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8904", "repo": "newfolder0/chrome-clacks", "url": "https://github.com/newfolder0/chrome-clacks/issues/5" }
gharchive/issue
Firefox Add-On? Is there any idea/motivation for a Firefox AddOn? Would be my first, if nobody wants to take it … Someone has posted the following on reddit.com/r/GNUTerryPratchett but I don't use Firefox so I haven't tested it myself: https://addons.mozilla.org/en-US/firefox/addon/gnu_terry_pratchett/ I don't know anything about Firefox development, but then I didn't really know about Chrome dev before I tried this. My original motivation for this was really just for my own use (I use Chrome) and I released it after seeing how much interest there was!
2025-04-01T06:39:44.665272
2024-08-13T15:26:44
2463632448
{ "authors": [ "jsbnr" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8905", "repo": "newrelic-experimental/nr1-drop-rule-explorer", "url": "https://github.com/newrelic-experimental/nr1-drop-rule-explorer/issues/9" }
gharchive/issue
FR: Toggle Functionality for Drop Rule A functionality that allows for both individual and bulk enabling or disabling of drop rules. This flexibility is crucial for efficient rule management and operational agility. The nerdgraph API does not offer an option to disable a drop rule and therefore can not be added as a feature to this app. This would require a change to the product via a feature request. You could possibly get the effect of a disbaled rule by adding a filter to the where clause that would never match like where true=false - Im not sure this is ideal. Internal FR created FRB-00007240
2025-04-01T06:39:44.669381
2021-07-29T07:59:50
955569426
{ "authors": [ "CLAassistant", "cianBuckley" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8906", "repo": "newrelic/entity-definitions", "url": "https://github.com/newrelic/entity-definitions/pull/285" }
gharchive/pull-request
set aws region to be alertable Relevant information Describe what you have done and any details that you think are relevant or that you might want to discuss with us. Checklist [ ] I've read the guidelines and understand the acceptance criteria. [ ] The value of the attribute marked as identifier will be unique and valid. [ ] I've confirmed that my entity type wasn't already defined. If it is I'm providing an explanation above. Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.Cian Buckley seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
2025-04-01T06:39:44.679554
2020-10-21T00:29:45
726036127
{ "authors": [ "kaojiri", "thande" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8907", "repo": "newrelic/newrelic-kubernetes-operator", "url": "https://github.com/newrelic/newrelic-kubernetes-operator/pull/123" }
gharchive/pull-request
fix: namespace's prefix deleted(manager.yaml) fix: namespace's prefix deleted(manager.yaml) so when I ran this with my original version of kustomize (3.8.2) I was seeing a different $ diff master.yaml kaojiri-master.yaml 6c6 < name: newrelic-kubernetes-operator-system --- > name: system $ kustomize version {Version:3.8.2 GitCommit:e2973f6ecc9be6187cfd5ecf5e180f842249b3c6 BuildDate:2020-08-29T19:15:49+01:00 GoOs:darwin GoArch:amd64} However, I updated to the latest version of kustomize (3.8.5), no difference found $ diff master.yaml kaojiri-master.yaml $ kustomize version {Version:kustomize/v3.8.5 GitCommit:4052cd4fd8c76a17b5f64e32509f3fba9713fe75 BuildDate:2020-10-08T05:35:40+01:00 GoOs:darwin GoArch:amd64} I'm concerned that with merging this any users with older versions of kustomize may encounter unexpected issues. Let me check with some others and get a second opinion on this. after further internal discussion among users of this operator and in light of the fact that this is restoring previous behavior before it changed by kustomize, I'll go ahead and merge this.
2025-04-01T06:39:44.685937
2024-04-09T19:43:40
2234184037
{ "authors": [ "CLAassistant", "jcountsNR" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8908", "repo": "newrelic/newrelic-quickstarts", "url": "https://github.com/newrelic/newrelic-quickstarts/pull/2362" }
gharchive/pull-request
feat: New singlestore quickstart Summary Added a new quickstart that utilizes an example OTeL tool collector to get metrics from SingleStore's API and ingest it into NR. Pre merge checklist [x] Did you check you NRQL syntax? - Does it work? [x] Did you include a Data source and Documentation reference? [x] Are all documentation links publicly accessible? [x] Did you check your descriptive content for voice and tone? [x] Did you check your descriptive content for spelling and grammar errors? [x] Did you review your content with a subject matter expert? (e.g. a Browser agent quickstart is reviewed with a member of the Browser Agent team) Dashboards [ ] Does the PR contain a screenshot for each of your dashboards? [ ] Do your screenshots show data? [ ] Has the sanitization script been run on each dashboard? Alerts [ ] Did you check that your alerts actually work? Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you all sign our Contributor License Agreement before we can accept your contribution.0 out of 2 committers have signed the CLA.:x: jcountsNR:x: MichelLosierYou have signed the CLA already but the status is still pending? Let us recheck it.
2025-04-01T06:39:44.687903
2024-10-12T01:06:43
2582386023
{ "authors": [ "fallwith", "kaylareopelle" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8909", "repo": "newrelic/newrelic-ruby-agent", "url": "https://github.com/newrelic/newrelic-ruby-agent/issues/2909" }
gharchive/issue
Docs say that Grape, Padrino, and Sinatra aren't supported for Ruby 3.0+ but they certainly are Some old documentation that presumably predates Ruby v3.0 being commonplace and well supported by the agent still notes: Please note that Grape, Padrino, and Sinatra aren't supported for Ruby 3.0+. https://docs.newrelic.com/docs/apm/agents/ruby-agent/getting-started/ruby-agent-requirements-supported-frameworks/ Resolved by https://github.com/newrelic/docs-website/pull/19465
2025-04-01T06:39:44.716600
2020-05-08T21:22:59
614994335
{ "authors": [ "Minoru", "coveralls" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8910", "repo": "newsboat/newsboat", "url": "https://github.com/newsboat/newsboat/pull/925" }
gharchive/pull-request
A bit of CI upkeep This PR mixes a number of small changes, so I think it's fine to bundle them like this. First: back when we were only using Travis, we ran all our jobs on the same Ubuntu version. Official repos didn't contain all the C++ compilers we need, so we also added r/test PPA and apt.llvm.org. This made the .travis.yml slightly less complicated than it would've been otherwise. Now that we're using Cirrus CI and Docker, it's far simpler to run a couple different Ubuntu versions such that they cover the whole range of compilers we support. That's what I did here: Ubuntu 18.04 keeps shouldering most of the load, while 16.04 is used for GCC 4.9, and 20.04 is used for GCC 9, GCC 10, and Clang 10. Second: .travis.yml had some leftovers from my failed attempt to run builds with more warnings enabled, and run tests with more checks (like stack protector). This PR adds these jobs to Cirrus. I'm not yet sure how useful they will be; I made a note to re-evaluate in two months. Fixes #838 Fixes #922 Reviews welcome. Will merge in 24 hours. Coverage remained the same at 56.199% when pulling 275f8cba7acbd4bf49e67c863d6f7e8de5ddb2c7 on feature/dockerfiles-upkeep into b1884ff27f341c8d400432d50567ec4857374f0b on master.
2025-04-01T06:39:44.730298
2023-01-20T18:12:42
1551263381
{ "authors": [ "spookybeetle" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8911", "repo": "newtfire/textAnalysis-Hub", "url": "https://github.com/newtfire/textAnalysis-Hub/pull/102" }
gharchive/pull-request
adding chatGPT files FINALLY! I didn't click something when making my access token. I was able to do this on Friday last week, I just didn't update it. My bad.
2025-04-01T06:39:44.762427
2022-07-27T00:21:18
1318879766
{ "authors": [ "JonoYang" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8912", "repo": "nexB/scancode.io", "url": "https://github.com/nexB/scancode.io/issues/476" }
gharchive/issue
Package Resources are not associated to packages in application Packages scan step update In updating the application Packages scanning step in the scan_codebase pipeline for #447, I noticed that the CodebaseResources of a Package are not associated with the Packages that was scanned. Looking a little deeper, I see that the Package Resources in scancode.io cannot be properly associated to the Package they are from by using the .assemble() methods from packagedcode's Package handlers. This is because of how for_packages is implemented on the CodebaseResource model in scancode.io and the Resource model in commoncode. In scancode.io, for_packages on the CodebaseResource model is a property that collects the purls for Packages that have been related to that CodebaseResource. In scancode-toolkit/commoncode, for_packages on the Resource model is a list that contains package_uid strings for the Package the Resource is from. When you run the .assemble() methods from packagedcode Package handlers on a CodebaseResource object, it attempts to append package_uid strings to the for_packages field. This does not properly work since CodebaseResource.for_packages is a property, not an attribute that can be used the same way. This issue has been fixed, following an update to scancode-toolkit where the code has been modified to be able to accept different functions that adds a package to a resource. https://github.com/nexB/scancode-toolkit/pull/3035
2025-04-01T06:39:45.359124
2020-11-18T10:30:58
745535611
{ "authors": [ "IceCodeNew", "dmfrey", "rs" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8913", "repo": "nextdns/diag", "url": "https://github.com/nextdns/diag/issues/3" }
gharchive/issue
Script hangs in the middle of execution for up to an hour. time sh -c 'sh -c "$(curl -sL https://nextdns.io/diag)"' Welcome to NextDNS network diagnostic tool. This tool will download a small binary to capture latency and routing information regarding the connectivity of your network with NextDNS. In order to perform a traceroute, root permission is required. You may therefore be asked to provide your password for sudo. The source code of this tool is available at https://github.com/nextdns/diag Do you want to continue? (press enter to accept) Testing IPv6 connectivity available: true Fetching https://test.nextdns.io status: ok client: 193.110.******* protocol: DOH dest IP: <IP_ADDRESS> server: rix-hkg-1 Traceroute for primary IPv4 (<IP_ADDRESS>) 1 193.110.****** 0ms 0ms 0ms ^C real 72m1.162s user 0m0.419s sys 0m0.174s What is your platform/OS/version? What is your platform/OS/version? uname -a Linux ******** 5.8.0-29-generic #31~20.04.1-Ubuntu SMP Fri Nov 6 16:10:42 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux Traceroute is hanging when trying to run diag against 1.4.3 install. See this output where it's stuck: Testing IPv6 connectivity available: false Fetching https://test.nextdns.io status: ok client: <IP_ADDRESS> protocol: DOH dest IP: <IP_ADDRESS> server: anexia-ewr-1 Fetching PoP name for ultra low latency primary IPv4 (ipv4.dns1.nextdns.io) vultr-ewr: 19.638ms Fetching PoP name for ultra low latency secondary IPv4 (ipv4.dns2.nextdns.io) anexia-ewr: 13.054ms Fetching PoP name for anycast primary IPv4 (<IP_ADDRESS>) zepto-xrs: 26.782ms Fetching PoP name for anycast secondary IPv4 (<IP_ADDRESS>) anexia-ewr: 16.078ms Pinging PoPs anexia-ewr: 11.887ms hetzner-iad: 20.496ms smarthost-bos: 20.537ms vultr-ewr: 20.505ms anexia-mnz: 20.451ms zepto-xrs: 20.544ms zepto-iad: 20.465ms axcelx-bos: 20.506ms teraswitch-pit: 24.632ms cloudzy-pit: 52.182ms Traceroute for ultra low latency primary IPv4 (<IP_ADDRESS>) 1 <IP_ADDRESS> 13ms 5ms 7ms 2 <IP_ADDRESS> 9ms 8ms 7ms 3 <IP_ADDRESS> 13ms 14ms 13ms 4 <IP_ADDRESS> 13ms 14ms 13ms 5 <IP_ADDRESS> 14ms 15ms 12ms 6 <IP_ADDRESS> 74ms 44ms 48ms
2025-04-01T06:39:45.396471
2022-03-22T10:40:29
1176591207
{ "authors": [ "daslu", "mk" ], "license": "ISC", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8914", "repo": "nextjournal/clerk", "url": "https://github.com/nextjournal/clerk/issues/119" }
gharchive/issue
small typo: java.imageio java.imageio mentioned in the changelog should be javax.imageio: https://github.com/nextjournal/clerk/blob/fc297cb8b1cd879c888f573d1c7913dafd1d66dd/CHANGELOG.md#05346-2022-01-27 @daslu thanks, fixed!
2025-04-01T06:39:45.461107
2023-09-02T13:46:39
1878696221
{ "authors": [ "CHENFANGC", "abhinav700", "robodove" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8916", "repo": "nextui-org/nextui", "url": "https://github.com/nextui-org/nextui/issues/1548" }
gharchive/issue
[BUG] - Modal scrollBehavior="outside" breaks placement prop NextUI Version latest Describe the bug By using the scrollBehavior="outside" prop, placement is always at the top (with a my-16 margin) Cause: https://github.com/nextui-org/nextui/blob/main/packages/core/theme/src/components/modal.ts#L171 For now the only way to fix it on the user side is to not use scrollBehavior="outside" but enter <Modal classNames={{ wrapper: "items-end overflow-y-auto", base: "my-1", }} /> for example if one prefers placement="bottom" and scrollBehavior="outside" Your Example Website or App No response Steps to Reproduce the Bug or Issue Create a Add scrollBehaviour="bottom" Modal will be placed at top Expected behavior Modal should always respect the placement prop Screenshots or Videos No response Operating System Version any Browser Other (add additonal context) It might be due to incorrect usage of parameters. bottom is not a valid parameter for scrollBehavior. The valid parameters for scrollBehavior are normal | inside | outside https://nextui.org/docs/components/modal#api It might be due to incorrect usage of parameters. bottom is not a valid parameter for scrollBehavior. The valid parameters for scrollBehavior are normal | inside | outside https://nextui.org/docs/components/modal#api No. The bug is emerging when scrollBehavior=outside and placement=bottom is used together. Look at your screenshot of the docs page, the prop just below scrollBehavior scrollBehavior=“outside” Thank you for correcting the reproduction steps. https://codesandbox.io/p/sandbox/sleepy-hugle-d6tt88?file=/App.jsx:37,42&utm_medium=sandpack There are two scenarios: Scenario 1: When the text does not exceed the container. In this case, scrollBehavior='outside' should not be used. placement='top' placement='bottom' In this case, the placement property is effective. Scenario 2: When the text exceeds the container. In this case, the scrollBehavior property must be added. When scrollBehavior='outside' and placement='bottom' are used together, the behavior is such that the placement property does not take effect. Regardless of the placement value set, the behavior remains the same, with a distance from the top, and a scroll-down state, as shown in the image below. scrollBehavior='outside' and placement='top' In this case, the placement property does not take effect. Is it your expectation that when setting scrollBehavior='outside' and placement='bottom', the initial pop-up state is at the bottom of the scrollbar, as shown in the image below? @wingkwong Can I work on this issue and add the 3 cases suggested by @robodove ? @wingkwong This behavior is taking place because of the items-start class in scrollBehavior:outside variant and disabling it should provide the solution for OP. Should we allow the user to change the placement property when scrollBehavior = outside, or we should keep the things unchanged?
2025-04-01T06:39:45.613455
2023-03-03T09:07:27
1608228652
{ "authors": [ "skrakau", "tillenglert" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8917", "repo": "nf-core/metapep", "url": "https://github.com/nf-core/metapep/issues/48" }
gharchive/issue
Check supported models and handle warnings Description of feature Check if requested alleles and lengths combinations are supported by epitope prediction tools and handle warnings/errors properly! Given in MERGE_PREDICTIONS_BUFFER.out.ch_prediction_warnings, use and output directly! Another check for the module lengths is introduced in PR #75 The peptide length will be checked for an available model and if multiple alleles are given only the matching models are used and the user is informed that the prediction will happen at reduced peptide length (e.g. for mouse alleles H2-Db and H2-Ld: H2-Db is available at length 9 and H2-Ld is available at length 8 and 9. Only peptides of length 9 will be predicted) This will be implemented as extra process. @skrakau Should I check in this process if the alleles are available and remove it from check_samplesheet_create_tables? related to https://github.com/nf-core/metapep/issues/74
2025-04-01T06:39:45.626465
2019-04-15T11:55:25
433241086
{ "authors": [ "apeltzer", "hmenager" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8918", "repo": "nf-core/nf-co.re", "url": "https://github.com/nf-core/nf-co.re/issues/107" }
gharchive/issue
Usage docs page missing Hello nf-core team, https://nf-co.re/usage_docs answers with a 404, so I'm letting you know! Well, a few pages are not there yet Nextflow tutorial Helper tools Guidelines etc. so I guess it's a work in progress. Good luck! Where did you find that url ? If I go to the main homepage, the links on top work :-) Would be interesting so we can fix the urls ! At the bottom menus: Getting Started Using nf-core Available pipelines Helper tools Nextflow tutorial Amazing that you found these - I'll update them to point to the correct subpages :-) Fixed, thanks a bunch @hmenager
2025-04-01T06:39:45.632699
2024-12-10T11:03:27
2729793583
{ "authors": [ "grst" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8919", "repo": "nf-core/rnaseq", "url": "https://github.com/nf-core/rnaseq/issues/1465" }
gharchive/issue
Unify SummarizedExperiment output Description of feature Currently, the pipeline generates three SummarizedExperiment .rds files while it would be more convenient and less confusing to generate a single one with different assay layers. Current situation There are three rds files: All of which have a counts and abundance layer The abundance layer is the same in all three objects While the counts are different representing counts, length scaled counts and scaled counts, respectively. Proposed solution Create a single SummarizedExperiment with the following assays: tpm (what is currently in "abundance", but the name is more explicit) counts counts_scaled counts_length_scaled gene_lengths (what is currently in salmon.merged.gene_lengths.tsv) possibly, it could be worth adding a vst layer generated by DESeq2::vst(). Create another one for the transcripts with the same layers. @drpatelh would you accept a PR for this?
2025-04-01T06:39:45.644362
2018-01-31T21:24:52
293316817
{ "authors": [ "Hackashaq666", "Marsgames", "zackaweed" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8920", "repo": "nfarina/homebridge", "url": "https://github.com/nfarina/homebridge/issues/1728" }
gharchive/issue
iOS 11.3 Hey, do you now if there is an issue caused by iOS 11.3 ? It seems that my homebridge doesn't work anymore since I update my iPad and iPhone to 11.3, but I don't know if it's the cause works for me I've also been having a strange issue since I updated to iOS 11.3 and tvOS 11.3 . No problem at all with any of my homekit accessories when I'm on my home network where my pi is attached, but the minute I leave the network all homebridge accessories are showing "no response". No error or anything strange looking in the homebridge log, any ideas? I removed my bridge (so all my accessories) then reinstall it, and now it works perfectly
2025-04-01T06:39:45.645892
2016-10-19T11:42:12
183938427
{ "authors": [ "FrankC2000", "lagunacomputer" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8921", "repo": "nfarina/homebridge", "url": "https://github.com/nfarina/homebridge/issues/858" }
gharchive/issue
Callback? Hi there. How do my device "tells" homebridge some status changing? For instance, lights have been turned off directly. Could you be provide some example? Thanks. If you turn a lamp off manually, I don't see how homebridge would ever get the message, unless you had somesort of 'smartsocket' that could self monitor electrical usage. Then that would have to 'report back' via a Homebridge plugin https://www.npmjs.com/search?q=homebridge-plugin
2025-04-01T06:39:45.652779
2018-05-23T10:39:13
325647944
{ "authors": [ "Cumberbatch08", "EtoDemerzel0427", "nfmcclure" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8922", "repo": "nfmcclure/tensorflow_cookbook", "url": "https://github.com/nfmcclure/tensorflow_cookbook/issues/130" }
gharchive/issue
tf.contrib.learn.preprocessing will be deprecated since tf.contrib.learn.preprocessing will be deprecated in future version of TensorFlow, the code in chapter 7 about the bag-of-word should be updated by using tf.data or something else. Thanks @EtoDemerzel0427 for bringing this up. I'll see what I can do. who can give me a example?
2025-04-01T06:39:45.832126
2015-04-10T17:18:25
67641760
{ "authors": [ "Prindle19", "stephenrjones" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8923", "repo": "ngageoint/geoq", "url": "https://github.com/ngageoint/geoq/issues/308" }
gharchive/issue
WSGI.py (or other WSGI Confguration) I noticed that this repo had a wsgi.py which would be useful for deploying with wsgi on Apache https://github.com/jaycrossler/geoq-django Understand that the project may favor nginx, but would it be possible to supply a default wsgi configuration for multithreaded wsgi deployment on Apache? We actually have used Apache / wsgi for other implementations of the system. I'll see if we have a default configuration we can make available
2025-04-01T06:39:45.833326
2016-06-06T19:52:55
158760001
{ "authors": [ "rfecher" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8924", "repo": "ngageoint/geowave", "url": "https://github.com/ngageoint/geowave/pull/820" }
gharchive/pull-request
GEOWAVE-346: landsat8 utilities from commandline tool apologies for the fairly large diff - it includes general raster improvements working through landsat 8 as the use case this commit history got messed up...closing this and opening a new clean PR
2025-04-01T06:39:45.849010
2020-10-13T13:29:12
720253224
{ "authors": [ "paskal", "umputun", "vtoupet" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8925", "repo": "nginx-le/nginx-le", "url": "https://github.com/nginx-le/nginx-le/issues/45" }
gharchive/issue
postgresql secure upstream I am trying to secure a postgresql behind nginx. I've followed the documentation from here (https://docs.nginx.com/nginx/admin-guide/security-controls/securing-tcp-traffic-upstream/) the config file looks like this: stream { upstream postgres { server localhost:5432; } server { listen 443; server_name db.myservername.com; proxy_pass postgres; proxy_ssl on; ssl_certificate SSL_CERT; ssl_certificate_key SSL_KEY; ssl_trusted_certificate SSL_CHAIN_CERT; } } Nginx complains that conf.d is not the correct path for this kind of file. Could you tell me if that can even work with nginx-le ? nginx-le doesn't do any magic. All it does is LE certificate update and replacement of SSL_CERT, SSL_CERT and SSL_CHAIN_CERT with those cert paths. And the file you mapped to /etc/nginx/service.conf will be copied to /etc/nginx.d/conf/ and this about it. You can see what it made by yourself inside the container, i.e. docker exec -it nginx cat /etc/nginx/conf.d/service.conf. As long as your source service.conf valid the result should be valid as well, so whatever works with nginx will be working fine with nginx-le because nginx-le is the nginx. Thanks for your reply. With a bit of hacking I managed to make it work (but I finally realized that I could not do SSL termination for that TCP stream as it requires NGINX Plus (see https://docs.nginx.com/nginx/admin-guide/security-controls/terminating-ssl-tcp/). Can you please post what you did at least briefly? Otherwise somebody will find this thread and will not be able to learn about how to do what you did. Sure, in the docker-compose.yml, I add a volume like so: - ./path_to/service_db.conf:/etc/nginx/stream.d/service_db.conf and the content of service_db.conf (see the hardcoded value for ssl_certificate as nginx-le script does not take care of this) upstream postgres { server localhost:5432; } server { listen 5433; proxy_pass postgres; proxy_ssl on; ssl_certificate /etc/nginx/ssl/le-crt.pem; ssl_certificate_key /etc/nginx/ssl/le-key.pem; ssl_trusted_certificate /etc/nginx/ssl/le-chain-crt.pem; ssl_protocols TLSv1.2 TLSv1.3; ssl_ciphers 'ECDHE-ECDSA-CHACHA20-POLY1305:ECDHE-RSA-CHACHA20-POLY1305:ECDHE-ECDSA-AES128-GCM-SHA256:ECDHE-RSA-AES128-GCM-SHA256:ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES128-GCM-SHA256:DHE-RSA-AES256-GCM-SHA384:ECDHE-ECDSA-AES128-SHA256:ECDHE-RSA-AES128-SHA256:ECDHE-ECDSA-AES128-SHA:ECDHE-RSA-AES256-SHA384:ECDHE-RSA-AES128-SHA:ECDHE-ECDSA-AES256-SHA384:ECDHE-ECDSA-AES256-SHA:ECDHE-RSA-AES256-SHA:DHE-RSA-AES128-SHA256:DHE-RSA-AES128-SHA:DHE-RSA-AES256-SHA256:DHE-RSA-AES256-SHA:ECDHE-ECDSA-DES-CBC3-SHA:ECDHE-RSA-DES-CBC3-SHA:EDH-RSA-DES-CBC3-SHA:AES128-GCM-SHA256:AES256-GCM-SHA384:AES128-SHA256:AES256-SHA256:AES128-SHA:AES256-SHA:DES-CBC3-SHA:!DSS'; ssl_prefer_server_ciphers on; ssl_dhparam /etc/nginx/ssl/dhparams.pem; ssl_ecdh_curve secp384r1; ssl_session_timeout 10m; } Maybe one enhancement of neginx-le script could be to retrieve all files that look like stream_*.conf and replace certificate path the same way as service*.conf and move the files to stream.dfolder. @vtoupet thanks for the idea, will be done in #48. I believe it was proposed some time ago in another issue but we didn't pick it up that time.
2025-04-01T06:39:45.860757
2018-10-01T06:18:48
365322254
{ "authors": [ "nginyc", "nudles" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8926", "repo": "nginyc/rafiki", "url": "https://github.com/nginyc/rafiki/issues/50" }
gharchive/issue
Model developers to tune architecture With "Efficient Neural Architecture Search via Parameter Sharing" Planned major changes To better support architecture tuning with ENAS, I'm planning changes to Rafiki's current model training framework: Replacing budget option MODEL_TRIAL_COUNT with TIME_HOURS Context Currently, when application developers create model training jobs, they pass a budget like { 'GPU_COUNT': 1, 'MODEL_TRIAL_COUNT': 20 }, with MODEL_TRIAL_COUNT deciding the no. of trials to conduct for each model template. Change Replace MODEL_TRIAL_COUNT option with TIME_HOURS option, which specifies how long the train job should run for. It is a soft time target. At the same time, I'll be reworking the Advisor component (which proposes trials' knobs) such that it is additionally in charge of deciding how many trials to run, when to stop each worker, when to stop the train job, given the budget e.g. GPU_COUNT and TIME_HOURS. Reasons for change May not be intuitive to the application developer to specify no. of trials while creating a train job ("how many trials should I put as budget? how long do I need to wait?"), especially if they're not supposed to be familiar with details like how model are trained and tuned. In contrast, TIME_HOURS is more straightforward. Currently, different models & model tuning strategies would require different no. of trials to be effective. For example, the original ENAS tuning strategy requires maybe (301x150+10+1) trials for sufficient train-eval cycles. In the future, it gives more flexibility for model tuning strategies at the Advisor component - for example, I'll be adding a new type of tuning strategy that takes all the models with no hyperparameters (e.g. model's knob config only consists of fixed values) and just conducts a single trial (since there's nothing to tune). It's also possible that a new tuning strategy can situationally conduct more/fewer trials based on feedback from workers. Introducing PolicyKnob Motivation I have been integrating ENAS as a new model tuning strategy on Rafiki (e.g. at the Advisor component). If model templates want to do architecture tuning with ENAS, the model's training code needs to switch between different "modes": During the ENAS architecture search phase, the model needs to alternate between "train my parameters for 1 epoch" and "don't train my parameters; just evaluate on the validation dataset" At the end of the architecture search, the model needs to switch to training its parameters from scratch with a full-sized architecture stacked with more cells, and train for 310 epochs Similarly, when you think about a standard hyperparameter tuning procedure, you might want the model to do early-stopping for the first e.g. 100 trials, then conduct a final trial for a full e.g. 300 epochs. In both architecture tuning & hyperparameter tuning, the model needs to be configured by Rafiki somehow to switch between these "modes" on a trial-basis. Change We can model the configuration of a model template for different training "modes" with different model policies. For example, if a model is to engage in policy QUICK_TRAIN, it is to prematurely speed up its training step e.g. by either doing early-stopping or reducing the no. of epochs. The model communicates to Rafiki which policies it supports by adding PolicyKnob(policy_name) to its knob_config. On the other hand, Rafiki configures the activation of the model's policies on a trial-basis by realising the values of PolicyKnobs to either True (activated) or False (not activated). For example, here is a example knob config of a model which supports the policy QUICK_TRAIN: Whenever the model is to do early-stopping, Rafiki will pass quick_train=True as part of the model's knobs. Otherwise, the model defaults to full-length training. Here is my current documentation for PolicyKnob: ''' Knob type representing whether a certain policy should be activated, as a boolean. E.g. the `QUICK_TRAIN` policy knob decides whether the model should stop model training early, or not. Offering the ability to activate different policies can optimize hyperparameter search for your model. Activation of all policies default to false. ===================== ===================== **Policy** Description --------------------- --------------------- ``SHARE_PARAMS`` Whether model supports parameter sharing ``QUICK_TRAIN`` Whether model should stop training early in `train()`, e.g. with use of early stopping or reduced no. of epochs ``SKIP_TRAIN`` Whether model should skip training its parameters ``QUICK_EVAL`` Whether model should stop evaluation early in `evaluate()`, e.g. by evaluating on only a subset of the validation dataset ``DOWNSCALE`` Whether a smaller version of the model should be constructed e.g. with fewer layers ===================== ===================== ''' @nudles I have added some details & reasoning on the major changes I'm going to make for architecture tuning. Let me know if you have any comments & advice on them! In terms budget, can we let users to configure either hours or trials? rename QUICK_TRAIN to EARLY_STOP? when will DOWNSCALE be used? Okay, I will keep MODEL_TRIAL_COUNT for backward compatibility as well. Noted on the suggestion. DOWNSCALE is used in architecture search for both NAS and ENAS. During the architecture search phase, the model constructed has fewer layers (e.g. 6 layers) to speed up the search, and its performance is somewhat a proxy of actual performance. On the other hand, at the final train phase, the final model trained from scratched is full-sized (e.g. 15 layers), which will be the one that is going to give the best performance.
2025-04-01T06:39:45.867406
2019-10-06T21:19:49
503162231
{ "authors": [ "aurel-l", "garyo", "ppillot", "sto3psl" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8927", "repo": "nglviewer/ngl", "url": "https://github.com/nglviewer/ngl/pull/698" }
gharchive/pull-request
performance optimisation when stage has been disposed of I realised that there were some functions still being called recursively at every frame even after removing the scene. In hope of getting rid of as much as I could, I was removing everything by doing: stage.removeAllComponents(); stage.dispose(); stage.viewer.container.innerHTML = ""; stage = null; This might not be too much of a problem in some pages, but if you happen to add and remove stages a lot (popup/modal with ngl viewer, or single-page applications) then this can build up. After adding and removing a stage 20 times (using a modified version of /examples/embedded.html) This is what I would get on the performance tab of Chrome. Every single frame I would have 20 calls to mouseObserver._listen() and 20 to viewer.animate() that would amount to about 5ms of useless use of the main thread, and memory usage growing (even if it does get caught by the garbage collector eventually). Compare to this screenshot using the code in this PR Here with have absolutely nothing happening after doing the same steps, which is way better regarding resource usage. OK, so I had installed this LGTM analysis thing on my forked repo and force-pushed to retrigger a check but I don't think it's something that needs to be done from my side. Let me know if I have to do something in order to pass the check This looks basically OK, though I'm not expert on this code, but perhaps you should protect against accidentally calling window.cancelAnimationFrame with an undefined value, and set the request back to undefined once the request is canceled. (And type frameRequest to include undefined.) but perhaps you should protect against accidentally calling window.cancelAnimationFrame This is not strictly necessary. cancelAnimationFrame does nothing when called with undefined or any invalid value. Reviewed and tested after a rebase. This works as advertised. Thanks @aurel-l for your contribution! Reviewed and tested after a rebase. This works as advertised. Thanks @aurel-l for your contribution!
2025-04-01T06:39:45.881694
2023-12-29T08:13:57
2059184376
{ "authors": [ "Puzzle-Drops", "cdog2323", "ngraves95" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8928", "repo": "ngraves95/attacktimer", "url": "https://github.com/ngraves95/attacktimer/issues/31" }
gharchive/issue
Scythe off by one tick Here are my settings. It is always ready to attack when the bar is still one tick from full. Kicking and other attack cooldowns seem to work fine. Idk if it's just the red scythe or also the default scythe. I use the red scythe. This seems to be a new issue, I wasn't experiencing this when running tob last week but all this week it's been happening. It is still happening today. Scythe and Shadow are off by one tick. the bar doesn't quite fill up to the highest before you can attack again. It functions perfectly with Tbow though. If you need a video I can make one, but probably not necessary? Try mimicking the same check marks. LMK Doesn't work how I like it. I use it as an action bar that fills up and when the bar is full I am ready to attack. Here is a video showing what is wrong. Tbow attacks when bar is filled. (correct) Shadow and Scythe attack when bar is one tick away from filled. (incorrect) https://www.youtube.com/watch?v=birD-b951qA Well the magic animation starts for tumeken's at 3 seconds (and it matches the video timestamps too). Keep in mind these are projectiles. The magic animation for shadow and range animation (on rapid) for Twisted bow have different speeds, but if you don't move and stay in the same spot will end up hitting the same if the monster is right in front of you. Try this setting? I was using regular scythe and it worked for me where it would be fill to the max then empty. I will try that. I went back and looked and I could just be wrong here. I'll report back later, thanks. I tried it. It's still wrong by one tick for both shadow and scythe. I was butterflying Akkha and everytime I go to attack it's one tick away from full bar. I've been using this plugin for over a thousand raids and it's never had this issue until this last week. I'm not sure if this is wrong - can you get a video that includes xp drops from shadow and tbow? And just to confirm - you didn't change "Empties Before Attack", "Fills Before Attack", or "Attack Bar Fills or Drains" between the previous behavior or now? Two more things: Can you also enable the "Show Attack Cooldown Ticks" option? Can you confirm the attack animation IDs for the scythe and shadow are 8056 and 9493, respectively?
2025-04-01T06:39:45.900515
2023-11-11T00:50:43
1988616446
{ "authors": [ "ngushchin", "stepelu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8929", "repo": "ngushchin/EntropicOTBenchmark", "url": "https://github.com/ngushchin/EntropicOTBenchmark/issues/3" }
gharchive/issue
requirements.txt is missing The README suggests to pip install -r requirements.txt but the file seems not to have been included in this repository. The requirements.txt file has been added to the repository. Thank you, however I am getting the following error: ERROR: Cannot install -r requirements.txt (line 13), -r requirements.txt (line 4) and torch==2.0.0 because these package versions have conflicting dependencies. The conflict is caused by: The user requested torch==2.0.0 lightning 2.0.1.post0 depends on torch<4.0 and >=1.11.0 torchvision 0.15.2 depends on torch==2.0.1 On a side note (that might be an issue on my side), I am experiencing issue with some packages (pillow, scikit_learn, ...) that needs to be built from source (on Ubuntu 20.04.6 LTS) in the required versions as the wheel seems not to be available. How was the requirements.txt file generated? Maybe it would be worth updating to latest, if it can be verified that the results of the benchmark stays consistent. The requirements.txt was generated using pipreqs (https://pypi.org/project/pipreqs/). I fixed the dependency errors in the new requirements.txt and verified that all these requirements can be downloaded and installed using "pip install -r requirementx.txt". As I mentioned in the requirements.txt, "lightning" was only used for the lightning version of the ENOT baseline, which we ended up not using for the evaluation (we used the pure pytorch version). For clarity and to resolve dependency issues, I have removed this version of ENOT. In general, the benchmark code uses only basic math operations from torch and numpy. All benchmark parameters are downloaded from Google Drive. Therefore, it is not critical to have exactly the same version as long as numpy and torch perform math operations in the same way in all versions. There is plotting and metric code for some benchmark pairs in the mixtures_benchmark_visualization_eot.ipynb notebook. You can compare it with the output you get in your Python environment. The results should be the same if your torch/numpy version has the same random generators, and may be slightly different if not.
2025-04-01T06:39:45.906873
2019-02-01T11:47:35
405668282
{ "authors": [ "jamolkhon", "nhaarman" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8931", "repo": "nhaarman/Acorn", "url": "https://github.com/nhaarman/Acorn/issues/115" }
gharchive/issue
Documentation errors In Scenes section ‘started’ : The Scene is dormant, waiting to be started or to be destroyed. ‘stopped’ : The Scene is started. In Navigators section: During the lifetime of a Navigator it can go from ‘inactive’ to ‘inactive’ and vice versa multiple times, until it reaches the ‘destroyed’ state. In Usage section: Acorn is tactically divided in several modules to be able to separate different concerns from eachother. I'm sorry I cannot create a pull request at the moment. Thanks! No problem, I will update them :+1: Fixed!
2025-04-01T06:39:45.923440
2023-10-04T12:30:23
1926111087
{ "authors": [ "nhirschey", "smoothdeveloper" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8932", "repo": "nhirschey/FSharp.Data.Fred", "url": "https://github.com/nhirschey/FSharp.Data.Fred/pull/17" }
gharchive/pull-request
support for netstandard2.0 This should enable using the library with .NET framework targets as well. Went over the unit tests (with api key and developer mode) and built the doc again to verify things are still OK. remark: setting as work in progress as I'm facing some issue with transitive FSharp.Data dependency when running samples under fsianycpu.exe (the .NET Framework version of FSI). It is actually working, it is just necessary to reference the same version of FSharp.Data (4.2.4 as of now) in order for things to bind correctly. I think this can be merged, it works with .NET Framework FSI now. I'm facing some issue with transitive FSharp.Data dependency when running samples under fsianycpu.exe (the .NET Framework version of FSI). I don't recall if there is an issue using netstandard or not. There's weirdness using FSharp.Data as a Library for parsing Json as we are here. I thought targetting net5.0/6.0 was because of a prior FAKE dependency, but maybe there are issues. FYI I pushed a commit fixing the github action workflow workflow. I don't recall if there is an issue using netstandard or not. There's weirdness using FSharp.Data as a Library for parsing Json as we are here. I thought targetting net5.0/6.0 was because of a prior FAKE dependency, but maybe there are issues. I was thinking to make a separate PR to remove FAKE dependencies from paket.dependencies (and running paket install again), do you want me to adjust it in this PR? The error I'm facing is, I think, because FSharp.Data split the types in separate assembly, and as of 4.2.4 it still expects everything under a single assembly. We should focus on upgrading FSharp.Data to the earliest version that introduced the split, but this would require a bit of research. FYI I pushed a commit fixing the github action workflow workflow. No problem! it is just necessary to reference the same version of FSharp.Data (4.2.4 as of now) I think #r nuget doesn't pick up dependency restrictions properly (I've had this issue elsewhere). You have to manually specify versions. Anyway, thanks for putting this together! Confirming it works with 4.2.10, so it doesn't have to be the exact same version, just the types have to be in same assembly as the library is compiled for. The earlier version of FSharp.Data that bring the split seems to be 5.0.1: https://www.nuget.org/packages/FSharp.Data.Json.Core/#versions-body-tab So my suggestion, on a next major update is to update to 5.0.1, and people should be able to use any later release, it would simplify the `#r "nuget: FSharp.Data" to not have to bake the version. I was thinking to make a separate PR to remove FAKE dependencies from paket.dependencies (and running paket install again), do you want me to adjust it in this PR? It would be excellent if you removed them here. The error I'm facing is, I think, because FSharp.Data split the types in separate assembly, and as of 4.2.4 it still expects everything under a single assembly. We should focus on upgrading FSharp.Data to the earliest version that introduced the split, but this would require a bit of research. About that time they merged a breaking change to HTTP/Json parsing which is the problem. Phillip knew it was breaking when he merged it but our usage here is a bit of an edge case. I think 5.0.2 is the last compatible FSharp.Data before the break. Happy for you to update it in this PR if you'd like, but I'm happy to merge without this change. It seems non trivial to update it without breaking documentation generation for now, I'll try to pay a look at another time. 5.0.2 seems to be same as the 4.2.10 based on the release note, best for now is to keep it as it is. Thank you @smoothdeveloper for providing this improvement! It should show up on nuget soon as v0.2.0. Great, will be able to use it out of box off nuget then :) Thanks for the great package!
2025-04-01T06:39:45.940779
2018-09-10T20:10:29
358775402
{ "authors": [ "SteveSchreiner", "kyuwoo-choi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8933", "repo": "nhnent/tui.editor", "url": "https://github.com/nhnent/tui.editor/issues/295" }
gharchive/issue
How can I restrict the file types that a user chooses for uploading an image? How can I restrict the file types that a user chooses for uploading an image? Thanks! the feature is not yet supported. let us discuss the feature here.
2025-04-01T06:39:45.947628
2023-12-04T15:21:40
2024112270
{ "authors": [ "codecov-commenter", "onehassan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8934", "repo": "nhost/nhost", "url": "https://github.com/nhost/nhost/pull/2402" }
gharchive/pull-request
fix(ci): pin @nhost/nhost-js dep version in sveltekit quickstart This will ensure CI doesn't complain about trying to install an unpublished version of the @nhost/nhost-js package in the SvelteKit quickstart. Codecov Report Attention: 1 lines in your changes are missing coverage. Please review. Comparison is base (f7c2148) 86.84% compared to head (278a641) 86.80%. Report is 12 commits behind head on main. Files Patch % Lines packages/nhost-js/src/utils/helpers.ts 80.00% 1 Missing :warning: :exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality. Additional details and impacted files @@ Coverage Diff @@ ## main #2402 +/- ## ========================================== - Coverage 86.84% 86.80% -0.05% ========================================== Files 85 85 Lines 9353 9328 -25 Branches 495 489 -6 ========================================== - Hits 8123 8097 -26 - Misses 1230 1231 +1 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
2025-04-01T06:39:45.949154
2017-11-03T14:25:02
270995654
{ "authors": [ "KevinMayfield", "VictorHarris" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8935", "repo": "nhsconnect/careconnect-reference-implementation", "url": "https://github.com/nhsconnect/careconnect-reference-implementation/issues/28" }
gharchive/issue
Typo in capability statement. Documentation for the practitioner search in PractitionerRole has a typo. organation should be organization. In core hapi - see https://github.com/jamesagnew/hapi-fhir/search?utf8=✓&q=organation&type=
2025-04-01T06:39:45.956547
2023-01-24T09:27:26
1554618446
{ "authors": [ "sarawilcox" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8936", "repo": "nhsuk/nhsuk-service-manual", "url": "https://github.com/nhsuk/nhsuk-service-manual/issues/1860" }
gharchive/issue
Create 2 new components: tabs and character count To be released in February release - frontend and service manual Related issue: https://github.com/nhsuk/nhsuk-service-manual/issues/1857 Draft in branch review/feb-release. Published: https://service-manual.nhs.uk/design-system/components/character-count https://service-manual.nhs.uk/design-system/components/tabs
2025-04-01T06:39:45.957724
2017-07-18T08:25:38
243637538
{ "authors": [ "c2s-dev", "st3v3nhunt" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8937", "repo": "nhsuk/profiles-db", "url": "https://github.com/nhsuk/profiles-db/pull/71" }
gharchive/pull-request
Feature/upgrade ci scripts Supersedes #68 :rocket: deployment of nhsuk/profiles-db succeeded (http://profiles-db-pr-71.dev.beta.nhschoices.net)
2025-04-01T06:39:45.981633
2019-03-14T12:48:35
420999419
{ "authors": [ "eyecatchup", "fuzzy76" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8938", "repo": "nicegist/nicegist.github.io", "url": "https://github.com/nicegist/nicegist.github.io/issues/10" }
gharchive/issue
Embed original gist comments? This issue is to track discussion whether to render the comments on the original gist or not. You can see the current implementation here: https://nicegist.github.io/84aec347c6a1b90890dad8953d7e8c67#gist-comments The comment section is only shown for gists that have received comments. (Compare the link above to this Nicegist without comments) What I like about it is, that you can easily link to comments (since I create unique anchor links for each comment). See: https://nicegist.github.io/84aec347c6a1b90890dad8953d7e8c67#comment-1876170 Let me know what you think. I'm still not sure if I should keep it or if its too distracting (because, ultimately, the idea of Nicegist is to offer a clean representation for gists). Hmm... I'm torn. But that sort of also means I don't have a strong opinion either way. :) But design-wise I think they look nice. Closing, since it's not an open issue. Further discussion welcome. Pinned this issue.
2025-04-01T06:39:46.053654
2023-08-29T07:41:28
1871084288
{ "authors": [ "MalikaIhle" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8939", "repo": "nickhaf/r_tutorial", "url": "https://github.com/nickhaf/r_tutorial/issues/54" }
gharchive/issue
remove mention of summer school ? in title (probably in yaml file), in intro text on welcome page. we (you and I and others) will reuse this for regular workshops as well ;) in read me file: maybe you could say instead that it was initially created for the LMU OSC summer school 2023 and linking to the website https://malikaihle.github.io/OSC-Open-Research-Summer-School-2023/ for the title: could be simply "Introduction to R' ? also mention Note This chapter is optional. It is not necessary to follow the rest of the conference, but you will get startet with plotting in R. in https://nickhaf.github.io/r_tutorial/qmd/plotting/plotting.html can be removed I think. the title already says it's optional
2025-04-01T06:39:46.125771
2015-07-14T12:13:13
94931887
{ "authors": [ "50percentDave", "nicolaskruchten" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8940", "repo": "nicolaskruchten/pivottable", "url": "https://github.com/nicolaskruchten/pivottable/issues/355" }
gharchive/issue
Smoothing/ Running Averages Hi I'm interested in producing a running average &/or smoothing via the average of a definable number of previous values (in a date series). I could produce this via sql on the database that I'm getting the data from, but thought that this could be either a derived attribute or an aggregation function, but I'm a bit stuck on the best approach and also how to reference the values in previous cells. Any guidance gratefully received. Edit It's amazing that just after posting I came across #140 which has given me something to work from. I'll see what I can come up with. OK, let me know how it goes! I think there's enough info in that other ticket to put you on the right track :) Both of these are minor modifications of the code in #140 by philliproso There are probably more efficient and elegant ways of doing this, but I've never used coffeescript before, so apologies in advance. runningAverage takes account of whether some of the cells are empty or not,smoothedAverage does not In the code for smoothedAverage, the value in a cell is the average of the current cell and the previous 3 cell values, in the earlier cells it's the average of the current cell and the previous cells. runningAverage: (formatter=usFmt) -> ([attr]) -> (data, rowKey, colKey) -> sum: 0 push: (record) -> @sum += parseFloat(record[attr]) if not isNaN parseFloat(record[attr]) value: -> colKeys=data.getColKeys() counter = 0 flat_col_key=colKey.join(String.fromCharCode(0)) for item in colKeys flat_item=item.join(String.fromCharCode(0)) if flat_item is flat_col_key itter=counter counter++ prev_value = 0 if itter >0 denom = 1 for i in [1...itter+1] aggregator = data.getAggregator(rowKey, colKeys[itter-i]) if 'sum' of aggregator prev_value += aggregator.sum denom++ return (@sum + prev_value) / (denom - 1) format: formatter numInputs: 1 smoothedAverage: (formatter=usFmt) -> ([attr]) -> (data, rowKey, colKey) -> sum: 0 push: (record) -> @sum += parseFloat(record[attr]) if not isNaN parseFloat(record[attr]) value: -> colKeys=data.getColKeys() counter = 0 flat_col_key=colKey.join(String.fromCharCode(0)) for item in colKeys flat_item=item.join(String.fromCharCode(0)) if flat_item is flat_col_key itter=counter counter++ for i in [1...itter+1] prev_value = 0 if (itter -1) of colKeys aggregatorone = data.getAggregator(rowKey, colKeys[itter-i]) if 'sum' of aggregatorone prev_value += aggregatorone.sum if (itter -2) of colKeys aggregatortwo = data.getAggregator(rowKey, colKeys[itter-(i+1)]) if 'sum' of aggregatortwo prev_value += aggregatortwo.sum if (itter -3) of colKeys aggregatorthree = data.getAggregator(rowKey, colKeys[itter-(i+2)]) if 'sum' of aggregatorthree prev_value += aggregatorthree.sum if itter > 3 return ((@sum + prev_value) / 4) return ((@sum + prev_value) / (5 - (4 - itter))) format: formatter numInputs: 1 Both of these are minor modifications of the code by philliproso in #140 runningAverage calculates a running average of non-blank cells smoothedAverage calculates the average of the current cell and the 3 previous cells - it doesn't take account of blank cells, or if there have been fewer than 3 previous values, the average of the current cell and it's predecessors. There are probably lots of more efficient and elegant ways of doing this in coffeescript, but this was my first time using it, so please indulge my inefficiencies (I got part of the way there in coffeescript, hacked the resulting javascript, then tried to work out the necessary coffeescript to get that js. Anyway I've now got node.js / npm installed so may further refine things, but it's working sufficiently for me needs runningAverage: (formatter=usFmt) -> ([attr]) -> (data, rowKey, colKey) -> sum: 0 push: (record) -> @sum += parseFloat(record[attr]) if not isNaN parseFloat(record[attr]) value: -> colKeys=data.getColKeys() counter = 0 flat_col_key=colKey.join(String.fromCharCode(0)) for item in colKeys flat_item=item.join(String.fromCharCode(0)) if flat_item is flat_col_key itter=counter counter++ prev_value = 0 if itter >0 denom = 1 for i in [1...itter+1] aggregator = data.getAggregator(rowKey, colKeys[itter-i]) if 'sum' of aggregator prev_value += aggregator.sum denom++ return ((@sum + prev_value) / (denom - 1)) format: formatter numInputs: 1 smoothedAverage: (formatter=usFmt) -> ([attr]) -> (data, rowKey, colKey) -> sum: 0 push: (record) -> @sum += parseFloat(record[attr]) if not isNaN parseFloat(record[attr]) value: -> colKeys=data.getColKeys() counter = 0 flat_col_key=colKey.join(String.fromCharCode(0)) for item in colKeys flat_item=item.join(String.fromCharCode(0)) if flat_item is flat_col_key itter=counter counter++ for i in [1...itter+1] prev_value = 0 if itter-1 of colKeys aggregatorone = data.getAggregator(rowKey, colKeys[itter-1]) if 'sum' of aggregatorone prev_value += aggregatorone.sum if itter-2 of colKeys aggregatortwo = data.getAggregator(rowKey, colKeys[itter-2]) if 'sum' of aggregatortwo prev_value += aggregatortwo.sum if itter-3 of colKeys aggregatorthree = data.getAggregator(rowKey, colKeys[itter-3]) if 'sum' of aggregatorthree prev_value += aggregatorthree.sum if itter > 3 return ((@sum + prev_value) / 4) return ((@sum + prev_value) / (5 - (4 - itter))) format: formatter numInputs: 1 Again, another minor modification Percentage change from previous value percentageChange: (formatter=usFmt) -> ([attr]) -> (data, rowKey, colKey) -> sum: 0 push: (record) -> @sum += parseFloat(record[attr]) if not isNaN parseFloat(record[attr]) value: -> colKeys=data.getColKeys() counter = 0 flat_col_key=colKey.join(String.fromCharCode(0)) for item in colKeys flat_item=item.join(String.fromCharCode(0)) if flat_item is flat_col_key itter=counter counter++ if itter >0 aggregator = data.getAggregator(rowKey, colKeys[itter-1]) if 'sum' of aggregator return (((@sum - aggregator.sum)/aggregator.sum)) return format: formatter numInputs: 1 I'm going to close this as a duplicate of #140 and leave a link there to here.
2025-04-01T06:39:46.135764
2016-11-25T14:06:06
191709008
{ "authors": [ "chapt0011", "datnguyen0606", "ilyaguy", "nagarajanchinnasamy", "nicolaskruchten" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8941", "repo": "nicolaskruchten/pivottable", "url": "https://github.com/nicolaskruchten/pivottable/issues/584" }
gharchive/issue
How to deal with large data set getting from server side My situation is that I have more than 200k rows records which are gotten from server side. In order to use pivottable js, a client browser needs to download whole data, e.g via ajax request, it causes browser hang and not respond. I'm thinking that instead of processing pivot table in front-end, I would do it on back-end side. However I would love to use the js library for drag&drop UI. It means that when I manipulate pivot table, it would send a request to server. Is there any way I could achieve this result? Or any suggestion would prevent browser not respond while downloading a large data set for initializing the pivotable? Thanks I'm working now on the same issue. My idea is to load data by portions (may be 20k rows), then init pivotUI with this array. So, it means that pivottable will not reload all data every for each drag-n-drop move. @ilyaguy: yeah, that's one way I have tried, but one down side is that user needs to wait until client gets all portions. And server needs to query into database for every portion (20k rows), it will take time to load whole data. My idea is that we only load data from database one time and store it in cache or local file. For each drag&drop move, client sends request to server, server don't need to query again, just get data from cache and process the pivot table (e.g using Pandas python library to handle it) Hello guy's, I'had to work on an similar issue... In my case, the final user has to deal with several data set When then user logs in, for each data set he has subscribed to, i run an asynchronous ajax request that stores the result of the sql query in php session. During this asynchronous request, the front end shows a progressbar ( Loaded 1 of X...). When finished, you can easily reload the full page without doing any other request on the database. Your pivot table should easily deal with the 200k rows that are stored in the php session. @chapt0011: storing data in php session means that the data is in server memory, right? If data grows up quickly, server will be out of memory. Is there any way we store data in client local storage? so we don't affect server memory. I think you need to choose where the pivoting needs to happen!. If it happens on server side, it will help reducing the amount of data to be transferred to client... But then you wouldn't need a pivot table ui on the client side (pls correct me if i am wrong) On the othethand, if you want the user to have entire data and perform pivoting on it, then data volume to be transferred to client may go up!!!! Good point to discuss about.... Waiting to learn from this discussion :) @datnguyen0606 Sure, but assume data has been loaded in php session. Render the page on the server side does'nt meen you can't unset the session variable on the server side after it has been sent to client side. First: While loading data into php session variable //SQLRequest // $_SESSION['dataset']=SQLResult json formatted display loading progress! On loaded Redirect user to the following page: <?php session_start(); //if you don't use any template engine echo '<html> <head>[...]</head> <body> <script> $(function(){ var derivers = $.pivotUtilities.derivers; $("#output").pivotUI('.$_SESSION['dataset'].', //your pivotUI params here }); }); }); </script> <div id="output"></div> </body>'; //FINALLY unset the session variable which is already written! unset($_SESSION['data']); ?> Peform pivot on the server side is not a good idea. Anytime the final user wants to change rows or cols in the table, he will have to reload the dataset or the pivot result. Actually the data in pivotui is already stored on the client side. I think you are just looking for the best way to get them as fast as possible from the server.. 1st. you need to optimize your request ! if you are using mysql you can try to use a ramdisk... 2nd. Use Ajax asynchronously to perform your request, you don t really need php sessions . This way you won t get any javascript timeout My feeling is that the architecture of this library is incompatible with a server-side integration along these lines... See #150 @chapt0011 yes you can bring in data in smaller chunks. But the pivot result on the client will be based on the data so far loaded on the client. And, the performance will be based on the client machine configuration!!! @nicolaskruchten it would be nice if you can point us to references of such architectures that would address such requirements. Thanks :) https://github.com/nicolaskruchten/pivottable/wiki/Frequently-Asked-Questions#server-side-integration
2025-04-01T06:39:46.139865
2021-10-26T09:33:35
1036062966
{ "authors": [ "grimmer0125", "nicolaspanel", "rawpixel-vincent" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8942", "repo": "nicolaspanel/numjs", "url": "https://github.com/nicolaspanel/numjs/issues/106" }
gharchive/issue
not installable on mac M1 machine The output log is: gyp info spawn args [ 'BUILDTYPE=Release', '-C', 'build' ] TOUCH Release/obj.target/libvips-cpp.stamp CC(target) Release/obj.target/nothing/node_modules/node-addon-api/nothing.o LIBTOOL-STATIC Release/nothing.a warning: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: archive library: Release/nothing.a the table of contents is empty (no object file members in the library define global symbols) CXX(target) Release/obj.target/sharp/src/common.o ../src/common.cc:24:10: fatal error: 'vips/vips8' file not found #include <vips/vips8> ^~~~~~~~~~~~ 1 error generated. make: *** [Release/obj.target/sharp/src/common.o] Error 1 gyp ERR! build error gyp ERR! stack Error: `make` failed with exit code: 2 gyp ERR! stack at ChildProcess.onExit (/Users/grimmer/.nvm/versions/node/v15.14.0/lib/node_modules/npm/node_modules/node-gyp/lib/build.js:194:23) gyp ERR! stack at ChildProcess.emit (node:events:369:20) gyp ERR! stack at Process.ChildProcess._handle.onexit (node:internal/child_process:290:12) gyp ERR! System Darwin 20.6.0 gyp ERR! command "/Users/grimmer/.nvm/versions/node/v15.14.0/bin/node" "/Users/grimmer/.nvm/versions/node/v15.14.0/lib/node_modules/npm/node_modules/node-gyp/bin/node-gyp.js" "rebuild" gyp ERR! cwd /Users/grimmer/git/webpack-demo/node_modules/sharp gyp ERR! node -v v15.14.0 gyp ERR! node-gyp -v v7.1.2 gyp ERR! not ok @nicolaspanel , another related thing is, I forked this repo to mine and consider to release a TypeScript version as another npm package, one of the reasons is that I remove the image manipulation feature. The changes are here, https://github.com/grimmer0125/numjs/pull/4 and I do not change the algorithm part. Should you have any concerns and suggestions, please tell me. hi @grimmer0125 I just gave you write access to this repo => Feel free to improve it in any way 1 advice though: sticking to numpy api (or as close as possible) make it easier for numpy users Best regards see #125 that fixes this issue in the meantime, adding "resolutions": { "sharp": "0.30.7" // or 0.29.2 }, should fix the issue ref #110
2025-04-01T06:39:46.144414
2020-03-26T10:29:02
588313088
{ "authors": [ "delucis", "nicolodavis" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8943", "repo": "nicolodavis/boardgame.io", "url": "https://github.com/nicolodavis/boardgame.io/pull/581" }
gharchive/pull-request
refactor(db): Make log handling explicit in StorageAPI.setState Move the log concatenation logic out of setState and into a separate appendLog method. Add a deltalog parameter to the StorageAPI.setState method to make it more obvious that setState should also handle appending entries to the game log. Closes #577. Two quick details: This only uses appendLog during onUpdate in the game master. I think this is the only place that will produce deltalogs, but wanted to flag it just in case a setState call elsewhere could have been adding to the log. deltalog is still being included in the state object sent to setState. It might not be necessary to store it in state, but I wasn’t sure (and it’s usually a pretty small object in any case). deltalog resides in the state object primarily because reducer.ts passes around a state object. I'd love to separate it out if it's possible to do that in a clean way, but like you pointed out, it's a pretty small object. OK, I’ve stripped deltalog from the state that gets persisted. I’m going to refrain from stripping it out elsewhere for now, because I’m not very clear where it is needed (e.g. on the client). Thanks, and sorry about the additional work to revert some of the changes! No worries — it’s probably good to keep the API surface small too.
2025-04-01T06:39:46.160900
2017-05-16T14:27:37
229053472
{ "authors": [ "migueldemoura", "nielsenramon" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8944", "repo": "nielsenramon/chalk", "url": "https://github.com/nielsenramon/chalk/pull/85" }
gharchive/pull-request
Remove misplaced .gitignore entry OS-specific entries, like *.DS_Store, should reside in a global .gitignore, created by each user to fit their needs. Example: git config --global core.excludesfile '~/.gitignore' echo '*.DS_Store' >> ~/.gitignore @migueldemoura I agree but I do not want users having to do this separately. Adding this line to the .gitignore isn't really impacting anything anyway. If you have any additions for Windows users, please let me know! Fair enough. I'd suggest separating the two types of ignores with two new lines. As for other suggestions: Windows: Thumbs.db Desktop.ini If you want, I can push another commit to address this. @migueldemoura Please do! Thanks for the help :-)
2025-04-01T06:39:46.253333
2022-06-03T17:10:22
1260155897
{ "authors": [ "cdrake", "hanayik" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8946", "repo": "niivue/niivue", "url": "https://github.com/niivue/niivue/pull/333" }
gharchive/pull-request
Added multiuser support List of fixed issues (if they exist): Demos work on: [ ] iOS Safari [ ] iOS Firefox [ ] iOS Chrome [ ] Android Chrome [ ] Android Firefox [ ] macOS Safari [ ] macOS Firefox [ ] macOS Chrome [ ] Windows Chrome [ ] Windows Firefox @cdrake, is this ready to be merged in? looks like the checks have passed ok.
2025-04-01T06:39:46.256525
2018-02-14T13:44:55
297100369
{ "authors": [ "nikhita", "razasayed" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8947", "repo": "nikhita/tech-conferences-india", "url": "https://github.com/nikhita/tech-conferences-india/pull/67" }
gharchive/pull-request
Added deccan ruby conf 2018 Deccan Ruby Conf 2018 happening in Pune on 4th Aug 2018 2018 is almost over, so I'll close this. Thanks for creating the PR, much appreciated!
2025-04-01T06:39:46.268273
2024-11-23T17:07:07
2686279806
{ "authors": [ "nikolasibalic" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8948", "repo": "nikolasibalic/status", "url": "https://github.com/nikolasibalic/status/issues/201" }
gharchive/issue
⚠️ LabBricks has degraded performance In 64aac7e, LabBricks (https://labbricks.com) experienced degraded performance: HTTP code: 523 Response time: 3210 ms Resolved: LabBricks performance has improved in 2b9573f after 20 minutes.
2025-04-01T06:39:46.271441
2016-11-30T07:09:51
192491730
{ "authors": [ "brson", "nikomatsakis" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8949", "repo": "nikomatsakis/cargo-chrono", "url": "https://github.com/nikomatsakis/cargo-chrono/pull/1" }
gharchive/pull-request
Set up error-chain @nikomatsakis I fixed this up to use the error chain in this PR and got it compiling. I expanded the throw macro to take fmt arguments as well and it is indeed pretty cool. I don't know about the word 'throw', but I think this macro should be in error-chain. cc @Yamakaky this project contains an interesting throw! macro that removes some of the boilerplate of creating a new error. @nikomatsakis check out that second commit that uses chain_err. That's what this lib encourages when adding information to an error (instead of creating a new error by e.g. format!("some error happened: {}", inner_error). Then the third commit sets up main to print the chain of errors. @brson nice! did, this require any changes to error-chain? Regarding throw!, The original ? included a throw keyword roughly equivalent to the macro here -- we removed it but in the time since I've come to think such a thing would be useful. @nikomatsakis No, this code itself didn't require any error-chain patches.
2025-04-01T06:39:46.290369
2021-06-19T15:25:46
925415821
{ "authors": [ "ashuydv", "nilisha-jais" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8950", "repo": "nilisha-jais/Musicophilia", "url": "https://github.com/nilisha-jais/Musicophilia/pull/368" }
gharchive/pull-request
issue #335 solved Description I've updated about's card images Checklist [x] I am making a proper pull request, not a spam. [x] I've checked the issue list before deciding what to submit. Related Issues or Pull Requests (Write your answer here.) Add relevant screenshot or video (if any) @niloysikdar the conflicts are solved, plz check it ! @niloysikdar please review it !! Hey @ashuydv please check this does'nt seem fine #368 issue solved Make UI more engaging #337, Added a new loader , plz review it. @ashuydv the pull request is good to merge you just need to fix the footer. Please check this And the preloader looks great! on my localhost the loader looks like this, why its expanding ?? Check the css maybe something might have been overwritten Can U plz explain, what should I fix in footer ?? should I change the background or should i make it like this ?? Plz guide me The footer looks like this. Just make it of screen size Its working now, plz review it now @ashuydv your preloader is working fine now. Just resolve the conflicts I'll merge it. And extremely sorry for the delay. ok, will do sry, for the delay, review it for the last time index.html file is fine but it seems that you have made changes in the about page also. Please remove those changes. Check this .About page looks different from rest of the pages. I've added only cards, I haven't touched navbar in any of them , Seems fine now. Thank you!
2025-04-01T06:39:46.383985
2024-11-29T09:37:48
2704545700
{ "authors": [ "quang-ng" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8952", "repo": "nimh-dsst/dsst-etl", "url": "https://github.com/nimh-dsst/dsst-etl/pull/12" }
gharchive/pull-request
Rtransparent Publication Ingestion Working on issue https://github.com/nimh-dsst/dsst-etl/issues/10 alembic done, please help me check @leej3
2025-04-01T06:39:46.401070
2017-08-01T16:21:28
247125663
{ "authors": [ "iSDP", "ninsuo" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8953", "repo": "ninsuo/symfony-collection", "url": "https://github.com/ninsuo/symfony-collection/pull/78" }
gharchive/pull-request
Update jquery.collection.js Fixing label issue (#28) Hello Thank you for your contribution. Quickly read the problem, and wondering if you checked this page? https://symfony-collection.fuz.org/symfony3/troubleshoot/hide-form-labels Yeah, but that's just hiding the label. With my addition, it's now possible to actually use that label and show numbered "steps" or "items" if you will. Nice. Will check this closer within a couple of hours. Okay, so when adding fields, it works. But when loading in such a collection it gets rendered by twig, currently i'm solving it with adding {%- if name matches '/^\\d+$/' -%} {%- set name = 'item ' ~ (name|number_format + 1) -%} {%- endif -%} But users need to add this to their templates if they want labels to be displayed properly. And I think moving elements won't sort correctly those numbers. We may think of a better solution, like a selector to write position nbr somewhere. Position is being stored in a hidden field, though. Sure, but it may be more friendly to set a selector like span.position so you could use <label>Item #<span class="position"></span></label> to automatically fill Item #42 on your view without tricky hacks. Yeah, but keep in mind it wouldn't always be "Item" either ;)
2025-04-01T06:39:46.410973
2018-11-06T19:45:34
378005109
{ "authors": [ "JMccormick264", "nint8835" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8954", "repo": "nint8835/iTunesRichPresence", "url": "https://github.com/nint8835/iTunesRichPresence/issues/26" }
gharchive/issue
Build Error building the latest source. DiscordBridge.cs line 117. Missing ; Severity Code Description Project File Line Suppression State Error CS1002 ; expected iTunesRichPresence-Rewrite D:\Desktop_Stuff\Github_Clones\C#\iTunesRichPresense\iTunesRichPresence\DiscordBridge.cs 117 Active Here catch (EntryPointNotFoundException) { var newPresence = new DiscordRpc.RichPresence { largeImageKey = "itunes_logo_big", details = "No song playing", state = "Re-install iTunesRichPresence to clear this message" } } Adding a semi colon after the ending curly bracket let's the project build successfully. As you can tell, I put extensive testing into this fix. I'd notice the error when I went to build a patch, but thanks for pointing it out to me. Resolved in 06cfd81
2025-04-01T06:39:46.429832
2023-03-11T23:01:12
1620171758
{ "authors": [ "martinRenou", "nir", "westurner" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8955", "repo": "nir/jupylet", "url": "https://github.com/nir/jupylet/issues/38" }
gharchive/issue
ENH: JupyterLite, WebGL https://github.com/emscripten-forge/empack https://github.com/emscripten-forge/recipes https://www.google.com/search?q=jupyterlite+emscripten-forge https://jupyterlite.rtfd.io -> Lab https://jupyterlite.readthedocs.io/en/latest/howto/configure/simple_extensions.html#the-case-of-jupyter-widgets-and-custom-renderers https://jupyterlite.readthedocs.io/en/latest/howto/configure/rtc.html#enabling-rtc-in-jupyterlite https://www.google.com/search?q=emscripten-forge+webgl I don't think it can be made to work without a SIGNIFICANT amount of work Hopefully the compiler will do most of that work. Maybe @martinRenou knows of an example emscripten-forge empack package that has OpenGL/EGL code? ipygany does 3D in notebooks with VTK, which may or may not already be transpired to WASM IIUC https://github.com/QuantStack/ipygany https://emscripten.org/docs/optimizing/Optimizing-WebGL.html#which-gl-mode-to-target There may need to be a 'polyfill' like requests-wasm-polyfill? https://github.com/emscripten-forge/recipes/blob/main/recipes/recipes_emscripten/requests-wasm-polyfill/recipe.yaml https://www.google.com/search?q="glfw"+emscripten https://www.google.com/search?q="moderngl"+emscripten https://github.com/emscripten-core/emscripten/blob/main/system/include/GL/glfw.h https://gist.github.com/ousttrue/0f3a11d5d28e365b129fe08f18f4e141?permalink_comment_id=4484709#gistcomment-4484709 -sUSE_GLFW=3 option Students could easily develop and share STEM games with a WASM compilation of jupylet that works in JupyterLite or VSCode.dev Maybe @martinRenou knows of an example emscripten-forge empack package that has OpenGL/EGL code? I don't! ipygany does 3D in notebooks with VTK, which may or may not already be transpiled to WASM IIUC ipygany only uses VTK on the back-end for file loading, the rendering is done entirely using WebGL. From time to time I google to see if Kitware folks provide an easy way to compile VTK for WASM, thankfully today you triggered that search and it seems to be fruitful https://gitlab.kitware.com/vtk/vtk-wasm-docker (initial commit a month ago). Though I guess all of this VTK discussion is completely out of scope for jupylet, so sorry for the spamming. If you want, @westurner, I'd be happy to have this discussion under an ipygany issue. hi @martinRenou, you were one of the first people to star jupylet when it first came out. thanks! :) @westurner, I don't have the time to jump into such a project at the moment, but I see that jupyterlite supports numpy ipywidgets and ipyevents, and is using async just as classic jupyter, so actually it may be possible to put aside audio, 3d, and sharertoy support and just reimplement the sprites and labels modules and start with them and get an initial version of jupylet for jupyterlite and continue from there (that is how jupylet started anyway). Probably worth compiling to WASM and running the tests first This is the requests-wasm-polyfill, because actual requests is not built on the ES/JS fetch api: https://github.com/emscripten-forge/requests-wasm-polyfill/tree/main/requests an example emscripten-forge empack package that has OpenGL/EGL code? It looks like there's already glfw support in emscripten. Is VTK actually out of scope? There probably could be a VTK backend; IDK what the advantage would be. actually it may be possible to put aside audio, 3d, and sharertoy support W3C Web Audio API https://developer.mozilla.org/en-US/docs/Web/API/Web_Audio_API https://github.com/topics/web-audio-api W3C Gamepad API https://developer.mozilla.org/en-US/docs/Web/API/Gamepad_API https://github.com/pybricks/support/issues/995 W3C Sensor API https://developer.mozilla.org/en-US/docs/Web/API/Sensor_APIs Like JupyterLite (and the vscode pyodide extensions) pyscript is also built on pyodide: https://dev.jeff.glass/pyscript-audio/index.html https://realpython.com/pyscript-python-in-browser/#sensor-api 3d https://news.ycombinator.com/item?id=32657051 : [ ] ENH: SensorCraft: replace Pyglet (OpenGL) with an alternate WebGL/WebGPU implementation https://github.com/quobit/awesome-python-in-education/issues/50
2025-04-01T06:39:46.450910
2016-12-15T17:07:14
195860873
{ "authors": [ "NilaxSpaceo", "nisrulz", "primissus", "simonkarmy" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8956", "repo": "nisrulz/qreader", "url": "https://github.com/nisrulz/qreader/issues/29" }
gharchive/issue
Can i limit the reading area? Is there any way I can use the full screen camera area to display but limit the reading area to a subview? Yes. The reading area is your SurfaceView, so how you size it (be it full screen or a specific height and width) in your layout is on you. Next is where you display your read data into , again depends on where you show the read data. I may be able to answer more clearly if you gave me a specific case. Hi, I'm sorry I lated too much to answer but I've been a little busy. What I was trying to say is that I wanted to show the camera on the whole screen but limit the qr reading area to another inside view, like this: I ended up modifying your QRDataListener by adding another onDetected method that sends the whole Barcode object so i could get the Rect of the read data and compare it to the Rect of my inside view in order to limit the reading area. QRDataListener: public interface QRDataListener { /** * On detected. * * @param data * the data */ // Called from not main thread. Be careful void onDetected(final String data); // Allows to use the whole captured data void onDetected(final Barcode data); } QREader: @Override public void receiveDetections(Detector.Detections<Barcode> detections) { final SparseArray<Barcode> barcodes = detections.getDetectedItems(); if (barcodes.size() != 0 && qrDataListener != null) { qrDataListener.onDetected(barcodes.valueAt(0)); qrDataListener.onDetected(barcodes.valueAt(0).displayValue); } } My implementation: ... qReader = new QREader.Builder(ScanActivity.this, scanner, new QRDataListener() { @Override public void onDetected(final String data) { } @Override public void onDetected(Barcode barcode) { if(areaRect.contains(barcode.getBoundingBox())) { final String data = barcode.displayValue; qReader.stop(); ... } } }).facing(QREader.BACK_CAM) .enableAutofocus(true) .height(width) .width(height) .build(); ... I hope this could help somebody else and I thank you for your attention. @primissus if you added a feature, do consider sending a PR. I am facing a problem on the aspect ratio of the camera preview. Always displayed with bad aspect ration. What to do ?? @simonkarmy I am facing same issue too. did you get any solution for that?
2025-04-01T06:39:46.462240
2022-02-15T00:42:03
1138017564
{ "authors": [ "codecov-commenter", "davemooreuws" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8957", "repo": "nitrictech/react-animated-term", "url": "https://github.com/nitrictech/react-animated-term/pull/1" }
gharchive/pull-request
chore: add example deployment Deploys to vercel Codecov Report Merging #1 (aab750f) into develop (84c2fbf) will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## develop #1 +/- ## ======================================== Coverage 82.08% 82.08% ======================================== Files 1 1 Lines 67 67 Branches 21 21 ======================================== Hits 55 55 Misses 12 12 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 84c2fbf...aab750f. Read the comment docs.
2025-04-01T06:39:46.488519
2024-11-08T17:01:09
2644616408
{ "authors": [ "nyarly", "peterldowns" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8958", "repo": "nix-community/lorri", "url": "https://github.com/nix-community/lorri/issues/137" }
gharchive/issue
Lorri displays "error extending watch paths" warning during successful operation I recently upgraded my version of nix and lorri and began to encounter the following warning lines during lorri builds. It does this on every single new build, but not when loading from a cached build. Nov 08 16:51:03.439 WARN error extending watch paths:, paths: [Normal("/Users/pd/.config/nixpkgs/config.nix"), Normal("<nix/fetchurl.nix>"), Normal("/Users/pd/Desktop/lorri-init-path-errors/shell.nix"), Normal("<nix/derivation-internal.nix>")], error: Error { kind: Io(Os { code: 2, kind: NotFound, message: "No such file or directory" }), paths: [] }, nix_file: /Users/pd/Desktop/lorri-init-path-errors/shell.nix I was able to create a minimal reproduction using a brand new repository where the only files were created by lorri init, you should be able to check it out here: https://github.com/peterldowns/lorri-init-path-warnings Hopefully you can reproduce the behavior by checking out the repository and running: cd lorri-init-path-warnings direnv allow . lorri watch --once Expected behavior Lorri functions correctly, updates the current shell environment to add the hello binary to the $PATH. Actual behavior Lorri functions correctly, updates the current shell environment to add the hello binary to the $PATH, and prints the warning log visible above. Metadata $ lorri info --shell-file shell.nix Project Shell File: /Users/pd/Desktop/lorri-init-path-errors/shell.nix Project Garbage Collector Root: /Users/pd/Library/Caches/com.github.nix-community.lorri.lorri.lorri/gc_roots/bad6431c93487fe2821f8b802ac8a0d8/gc_root/shell_gc_root General: Lorri User GC Root Dir: /Users/pd/Library/Caches/com.github.nix-community.lorri.lorri.lorri/gc_roots Lorri Daemon Socket: /Users/pd/Library/Caches/com.github.nix-community.lorri.lorri.lorri/daemon.socket Lorri Daemon Status: `lorri daemon` is running $ uname -a Darwin pld-mbp-22 23.4.0 Darwin Kernel Version 23.4.0: Fri Mar 15 00:10:42 PDT 2024; root:xnu-10063.101.17~1/RELEASE_ARM64_T6000 arm64 arm Darwin Additional context macOS 14.4.1 Sonoma lorri 1.7.1 installed via service.lorri.enable = true; in my system flake system flake nixpkgs (where lorri comes from) is github:NixOS/nixpkgs/85f7e662eda4fa3a995556527c87b2524b691933?narHash=sha256-JwQZIGSYnRNOgDDoIgqKITrPVil%2BRMWHsZH1eE1VGN0%3D (2024-11-07 05:50:23) Interesting. Thanks for putting together a reproduction. I'll see if I can take a look over the weekend, but I'm on the road next week. I agree: the warning should describe which path can't be found, at the very least. Cool, thanks for the quick response. This isn't an urgent bug in any way, since the only problem is that there is a warning shown — lorri still works great. Thanks for maintaining and for the improvements you've been making, I'm looking forward to the improved flakes support!
2025-04-01T06:39:46.493527
2023-07-18T10:24:25
1809641283
{ "authors": [ "adrian-gierakowski", "exarkun", "smulikHakipod" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8959", "repo": "nix-community/terraform-nixos", "url": "https://github.com/nix-community/terraform-nixos/pull/76" }
gharchive/pull-request
Use Nix copy command instead of nix-closure-copy This relates to #63 I tested it, and its very fast. It only work fast on NisOS unstable though, as NixOS 23.05 contains Nix 2.13 which doesn't support that yet. On older nixos it just works slowly. I tested it, and its very fast. Can you provide some numbers to go with this? How does it deal with latency? @exarkun sure, here is a time terraform apply -auto-approve on some example terraform on AWS (with around 100ms latency) and sample nix config without the patch: terraform apply -auto-approve 14.52s user 4.95s system 3% cpu 8:26.58 total and here is with the patch (on completely new machine): terraform apply -auto-approve 7.47s user 1.55s system 14% cpu 1:00.89 total @exarkun i don’t have capacity to test this right now. I’d be happy to merge it if it was opt-in. Can we add a flag in the terraform module to switch between original implementation and this one? Thanks!
2025-04-01T06:39:46.496021
2023-03-02T22:56:40
1607650525
{ "authors": [ "benwis", "cstrahan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8960", "repo": "nix-community/tree-sitter-nix", "url": "https://github.com/nix-community/tree-sitter-nix/issues/36" }
gharchive/issue
Release tree-sitter-nix on Crates.io Hi, I'd love to include tree-sitter-nix in my markdown to html converter crate, but to do that I can't have git dependencies. Do you have any interest or plans on releasing the Rust adapter on crates.io? Oh yeah, that's a good idea! I'll see if I can push it up tonight or this weekend. @benwis I just now published the crate: https://crates.io/crates/tree-sitter-nix :tada: Lemme know if there's anything else I can do!
2025-04-01T06:39:46.500321
2017-03-06T04:51:23
212023668
{ "authors": [ "chaosagent", "kamalmarhubi" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8961", "repo": "nix-rust/nix", "url": "https://github.com/nix-rust/nix/issues/550" }
gharchive/issue
PTRACE_O_SYSGOOD flag is interpreted as part of a signal In sys::wait, if waitpid returns a status originating from ptrace and PTRACE_O_SYSGOOD is set, the seventh (0x80) bit of the status will be flipped. This causes status parsing to fail. The flag technically isn't a ptrace event either, but I feel like that would be the best place to put it. Alternatively, a SYSGOOD-status could be its own unique kind of status, or a boolean entry could be added in WaitStatus::Stopped to mark it. I've opened a pull request simply ignoring the bit for now (#549) @chaosagent thanks for the report. What's the platform / architecture this happens on? This only happens on Linux, and I can confirm it happens on x86_64, but the Linux man page says that it might not work on all platforms. The name of the flag is PTRACE_O_TRACESYSGOOD btw; I'll changethe issue title to reflect.
2025-04-01T06:39:46.587683
2014-12-22T21:39:33
52691391
{ "authors": [ "TheAifam5", "nlhans" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8962", "repo": "nlhans/ets2-sdk-plugin", "url": "https://github.com/nlhans/ets2-sdk-plugin/issues/4" }
gharchive/issue
Linux This is compatible with Linux ETS2? The current revisions are DLLs and only run on Windows. I am unaware of any support for SDK plugins in ETS2 for Linux neither. All SDK examples are for Windows, and they also use Windows-specific includes in their code. I don't run Linux as a gaming OS at this moment, so I haven't got ETS2 installed to try this out. If you do have any related links to the ETS2 SDK and Linux, please let me know.
2025-04-01T06:39:46.650655
2022-07-18T12:55:16
1307918060
{ "authors": [ "falbrechtskirchinger", "gregmarr", "nathanieltagg", "nlohmann" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8963", "repo": "nlohmann/json", "url": "https://github.com/nlohmann/json/issues/3588" }
gharchive/issue
Multiple versions causing conflict Description I'm using a 3rd party library that uses nlohmann/json under the hood. I am also using it in my code. I have found that unless these two versions are exactly the same, there are side effects that cause the code not to function in the 3rd party library. (The failure happens with 3.9.2 in the 3rd party library and 3.10.5 in mine.) The exact failure I can't pinpoint, because it happens deep in code I'm not famililar with, but it causes no build-time error or crash, but DOES cause incorrect behavior that appears to be null JSON objects. Keeping my version matched to the library gives me temporary respite, but is a pain to maintain, and becomes unsolvable if it happpens with yet another library. The problem is clearly a collision in object code resolution. Could we have it customizable to change from nlohmann to MYNAMESPACE::nlhomann? Wrapping the header file in a namespace doesn't work for a variety of reasons. I CAN do a very cheap hack by doing #define nlhomann my_nlhomann but I suspect that might not be very sustainable. Or perhaps there is another solution that I haven't found online? Reproduction steps The 3rd party library in question is depthai-core, in a dependency. Expected vs. actual results The exact throw is "cannot use at() with null", but this is due to a previous problem. Minimal code example No response Error messages No response Compiler and operating system Linux gcc Library version 3.9.2, 3.10.5 Validation [ ] The bug also occurs if the latest version from the develop branch is used. [ ] I can successfully compile and run the unit tests. What about a versioned inline namespace? (@nlohmann) I might try it later, but it'll need some support from your release scripts to bump the namespace version. But I agree; if the header file used namespace nlohmann_3_10_5 {...} namespace nlohmann = nlohman_3_10_5; then that would easily solve the problem forevermore. That's not quite what I'm suggesting. namespace nlohmann { inline namespace v3_10_5 { // ... } } Nothing changes for end users unless there's an ambiguity because both headers are included in one translation unit. See https://en.cppreference.com/w/cpp/language/namespace#Inline_namespaces for an explanation. Are you having this issue because two different versions of the library headers are being included in the same file (translation unit) or because two different files (translation units) with different versions are linked together? The former is now prevented by https://github.com/nlohmann/json/pull/3418 in the develop branch. Are you having this issue because two different versions of the library headers are being included in the same file (translation unit) or because two different files (translation units) with different versions are linked together? The former is now prevented by #3418 in the develop branch. Based on the proposed workaround, I'm assuming this is about linking different versions together. An inline namespace would result in different symbol names. JSON_DIAGNOSTICS would still pose a problem, but could be solved with an inline namespace as well, as mentioned during one of the last ABI-related discussions. FYI, we'll soon check whether incompatible versions are used together, see https://json.nlohmann.me/api/macros/json_skip_library_version_check/#runtime-assertions. That is, 3.11.0 will detect when it is used by earlier versions. @nlohmann That only helps with including two different versions of the header in the same file. It doesn't help with ODR violations that occur at link time when you use different versions in different files. There are things that can help with that too, but we're not doing any of them yet. https://docs.microsoft.com/en-us/cpp/preprocessor/detect-mismatch?view=msvc-170 This can also be used to detect one file built with diagnostics, and one without. It will only help going forward as it can only detect when the pragmas exist and are different between files. It won't prevent previous versions being used together or with the new version. @gregmarr The problem is specifically two object files that use diffent versions of the header-only json, so that the wrong code gets executed leading to bad data. The inlined versioning looks like it would work, and is much more elegant than my brute force #define
2025-04-01T06:39:46.656990
2017-05-13T15:22:13
228480051
{ "authors": [ "coveralls", "nlohmann", "zhaohuaxishi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8964", "repo": "nlohmann/json", "url": "https://github.com/nlohmann/json/pull/583" }
gharchive/pull-request
fix doxygen error of basic_json::get() @ref command do not seems to work in markdown code span. so i change `void from_json(const @ref basic_json&, ValueType&)` into `void from_json(const ` @ref basic_json`&, ValueType&)` are there better solution for this problem? The mentioned functions are not meant to be linked in the first place, because they are most likely in the client's code - it can be any from_json function based on ValueType. I'm not sure whether fixing this so that basic_json can be clicked brings too much value here. Coverage remained the same at 99.722% when pulling dfa371c436fa14926544eb11f3207d0439cd2c43 on zhaohuaxishi:doxygen_error into 9b764ee5d671b41255d390ab9089f12036b2d38a on nlohmann:develop. is that means the @ref command should just be removed then. Coverage remained the same at 99.722% when pulling b8dff3bc1674acb6a30cd733ecae545b93b1baeb on zhaohuaxishi:doxygen_error into 723c87560459eaacb3e3b3f0d4666bd2d02be317 on nlohmann:develop. Coverage remained the same at 99.722% when pulling b8dff3bc1674acb6a30cd733ecae545b93b1baeb on zhaohuaxishi:doxygen_error into 723c87560459eaacb3e3b3f0d4666bd2d02be317 on nlohmann:develop.
2025-04-01T06:39:46.665480
2020-06-21T08:26:02
642520078
{ "authors": [ "17839192463", "GuanNiPiShi123" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8965", "repo": "nlpyang/PreSumm", "url": "https://github.com/nlpyang/PreSumm/issues/179" }
gharchive/issue
how to understand use_bert_emb? I change the bert model to Albert model. But when use_bert_emb is set to true, the model reports an error: RuntimeError: The size of tensor a (128) must match the size of tensor b (768) at non-singleton dimension 2. if use_bert_emb is set to false, the model works normally. How can I solve this problem? I have the same problem with you. Have you solved it? Sorry, the problem has not been solved for the time being. | | Shuai Zhao | | <EMAIL_ADDRESS>| Signature is customized by Netease Mail Master On 08/21/2020 15:16, GuanNiPiShi123 wrote: I have the same problem with you. Have you solved it? — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub, or unsubscribe.
2025-04-01T06:39:46.667169
2015-10-27T21:59:33
113699096
{ "authors": [ "JonathanReeve", "alvations" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8966", "repo": "nltk/nltk", "url": "https://github.com/nltk/nltk/issues/1180" }
gharchive/issue
Is generate() coming back? Would really love to see this function back again. The book says it will be reinstated in a later version, but when? +1 , given the recent RNN generation from @karpathy (https://github.com/karpathy/char-rnn), it might be very easy to create a generate algorithm but support of any NN implementation is all very unstable API changes almost every 2-3 weeks =( Is there a reason why this function was removed? Just curious.
2025-04-01T06:39:46.668158
2015-08-28T21:06:05
103795622
{ "authors": [ "nlwillia" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8967", "repo": "nlwillia/bootstrap", "url": "https://github.com/nlwillia/bootstrap/pull/1" }
gharchive/pull-request
#2404 Modal open calls should complete in order This is a second PR attempt for issue #2404. (PR #2443 has rebase issues and will be closed.) Oops...PR'd to the fork instead of upstream.
2025-04-01T06:39:46.687332
2024-06-28T21:03:41
2381198262
{ "authors": [ "patudom" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8968", "repo": "nmearl/hubbleds", "url": "https://github.com/nmearl/hubbleds/pull/4" }
gharchive/pull-request
Stage 6 updates This is based on #3, so that should get merged first. This does the bare minimum to allow the user to go to Stage 6. Guidelines with mc or fr questions will not appear, so if you try to navigate past the first guideline, you will be stuck. I did not attempt to set up any gates, but I can work on that once the new mc & fr methods are set up. I did not attempt to wire up the layer toggle or hubble viewer. @johnarban or @Carifio24, perhaps one of you could work on that once you're finished with the other content you are updating. Thanks! Opened new PR at https://github.com/cosmicds/hubbleds/pull/436 to merge this to cosmicds main, so closing this. Reopening this and closing https://github.com/cosmicds/hubbleds/pull/436
2025-04-01T06:39:46.692869
2024-07-28T10:46:44
2433869434
{ "authors": [ "nmondal" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8969", "repo": "nmondal/cowj", "url": "https://github.com/nmondal/cowj/issues/111" }
gharchive/issue
Benefit of GRAAL VM integration Theoretical, mostly. Current situation with GRAAL, native-image does not run properly. Rhino is an order of magnitude faster than Graal's JS implementation. https://github.com/oracle/graaljs/issues/836 It is also evident that graal is slow, at least by 12% for our use case. More update: https://github.com/mozilla/rhino/issues/1555
2025-04-01T06:39:46.700369
2024-10-29T12:29:23
2621104257
{ "authors": [ "Milena-Czierlinski", "jkoenig134", "sebbi08" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8970", "repo": "nmshd/connector", "url": "https://github.com/nmshd/connector/pull/299" }
gharchive/pull-request
Re-add deprecated function deleteThirdPartyOwnedRelationshipAttributeAndNotifyPeer For backwards compatibility Readiness checklist [ ] I added/updated tests. [x] I ensured that the PR title is good enough for the changelog. [x] I labeled the PR. @Milena-Czierlinski as we are the only consumers we can keep this removed silently! :) @sebbi08 expressed his concern that the previous release introduced a breaking change for integrators. IMO as the API is unchanged and we don't know about any consumers of the js sdk than us we can ignore it. also I don't expect that someone is already using this functionality @sebbi08 as you requested this, any word from your side for or against this? If we are sure that no one uses this, I am also fine with just removing it. Otherwise, this would be a breaking change.
2025-04-01T06:39:46.710585
2020-06-15T19:04:41
639072185
{ "authors": [ "cathay4t", "ffmancera" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8971", "repo": "nmstate/nmstate", "url": "https://github.com/nmstate/nmstate/issues/1116" }
gharchive/issue
Release version 0.3.2 Creating a New Release Changelog [x] First, update the CHANGELOG file in the project root directory. Use the command git log --oneline v0.3.1..HEAD to get the changes since the last tag. Add an entry like the following: ## [X.Y.Z] - YYYY-MM-DD ### Breaking Changes - ... ### New Features - ... ### Bug Fixes - ... Tagging [x] Tag new release in git. [x] Ensure that the vdsm tests ran against this state. # Make sure your local git repo is sync with upstream. # The whole version string should be like `v0.0.3`. # Put strings like `nmstate 0.0.3 release` as commit message. git tag --sign v<version> git push upstream --tags [x] If you need to remove a tag because something needs to be fixed: # Remove local tag git tag -d <tag_name> # Remove upstream tag git push --delete upstream <tag_name> GitHub Release [x] Generate and sign the tarball. git clean -x -d -n # before running the next command check, that it is ok to remove the files git clean -x -d -f # Please remove python3-setuptools_scm, or it will all git files into tarbal. env --unset=PYTHONPATH python3 setup.py sdist gpg2 --armor --detach-sign dist/nmstate-<version>.tar.gz [x] Visit github draft release page. [x] Make sure you are in Release tab. [x] Choose the git tag just pushed. [x] Title should be like Version 0.0.3 release. [x] The content should be copied from the CHANGELOG file. [x] Click Attach binaries by dropping them here or selecting them. and upload the dist/nmstate-<version>.tar.gz and dist/nmstate-<version>.tar.gz.asc. [x] Download the tarball and the signature. [x] Check if the signature is correct. curl --silent https://www.nmstate.io/nmstate.gpg | gpg2 --import gpg2 --verify nmstate-<version>.tar.gz.asc nmstate-<version>.tar.gz [x] Check in a clean Fedora/centOS container if the package build and install correctly. podman run -d -it --name <name> docker.io/library/fedora:31 bash podman cp nmstate-<version>.tar.gz <name>:/home/nmstate-<version>.tar.gz podman exec -it <name> bash cd /home/ tar xzvf nmstate-<version>.tar.gz python3 setup.py build python3 setup.py install [x] Click Save draft and seek for review. [x] Click Publish release once approved. PyPi Release # Make sure you installed python package: wheel and twine. yum install twine python3-wheel rm -rf dist python3 setup.py sdist bdist_wheel # Upload to pypi test. python3 -m twine upload --repository-url https://test.pypi.org/legacy/ dist/* # Now, check https://test.pypi.org/project/nmstate/ # If it works, now upload to pypi. python3 -m twine upload dist/* Post Release [x] 1. Create a pull request with increased version number in the VERSION file and merge it before any other PR. This is necessary to ensure that the development RPMs are newer than the stable version in distributions. [x] 2. Update the SPEC files in Fedora, create new builds and updates as neccessary [x] 3. Rebuild Copr repositories for stable releases as necessary (this requires the SPEC files in Fedora to be updated, first) https://copr.fedorainfracloud.org/coprs/nmstate/ [x] 4. Send out a notification to the fedorahosted mailing list: <EMAIL_ADDRESS> All done, thanks!
2025-04-01T06:39:46.728870
2022-11-25T17:52:33
1464807763
{ "authors": [ "NaturalGradient", "codecov-commenter" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8972", "repo": "nnaisense/evotorch", "url": "https://github.com/nnaisense/evotorch/pull/41" }
gharchive/pull-request
Draft: Add torch-based CMA-ES implementation This pull request: Refactors the existing evotorch.algorithms.cmaes.CMAES, which is simply a wrapper for pycma to evotorch.algorithms.pycmaes.PyCMAES. Introduces a new, torch-based, implementation of CMAES under the name evotorch.algorithms.cmaes.CMAES. This aims to faithfully reimplement the most recent version of pycma, while also benefiting from the evotorch ecosystem e.g. vectorized, GPU based, working with torch tensors for compatibility with other features. Introduces evotorch.algorithms.restarters, which provides basic functionality for meta-algorithms that wrap around a SearchAlgorithm class + parameterization, and allow automatic restarting of the algorithm. This creates basic functionality for the eventual re-implementation e.g. of IPOP-CMA-ES and BIPOP-CMA-ES, and equivalent algorithms for XNES, SNES etc. Adds the is_terminated flag to all SearchAlgorithm classes, that will allow the user to define their own general functionality for detecting termination states that should trigger restarts. Currently, this value defaults to False, but in the future it is intended to add example termination states to the new CMAES implementation. Codecov Report Merging #41 (63f3cc0) into master (d3e3f0a) will decrease coverage by 2.14%. The diff coverage is 12.45%. :exclamation: Current head 63f3cc0 differs from pull request most recent head b1ec581. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
2025-04-01T06:39:46.745069
2021-01-09T18:44:22
782663638
{ "authors": [ "ericdallo", "nnichols" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8973", "repo": "nnichols/clojure-dependency-update-action", "url": "https://github.com/nnichols/clojure-dependency-update-action/issues/4" }
gharchive/issue
Add lib Changelog to the PR feature Feature Request I known this is not simple but if the lib has a CHANGELOG file on it's repo (if it's a github repo or something like that), add the changelog to the PR description would be great I'll take a bit to think through this one. I don't think it'll be too hard to write a generic header and the same info as is in the commit message, but I also suspect this would be best implemented with some configurable options around the CHANGELOG message. I'll take a bit to think through this one. I don't think it'll be too hard to write a generic header and the same info as is in the commit message, but I also suspect this would be best implemented with some configurable options around the CHANGELOG message. Yeah, at work (Nubank) we have a bot that opens PRs called Bumpito, it's pretty similar to your github action, but it's closed source, unfortunately... It opens PRs for our common libs bumps and in the description, it prints something like this: PR description: Bumpito has found new common libraries versions. Refer to the following changelogs to see what is new :newspaper_roll: Changelogs common-kafka 10.86.0 Remove jackson-core explicit dependency from common-kafka 10.85.1 remove consumer cron tick code common-i18n 4.5.0 Fix Federal Holidays for Mexico common-crypto 10.24.1 improve the way of communicating error getting s3 keys common-metrics 10.6.0 Add prometheus text format parser common-redis Could not parse changelog :( Yeah, at work (Nubank) we have a bot that opens PRs called Bumpito, it's pretty similar to your github action, but it's closed source, unfortunately... It opens PRs for our common libs bumps and in the description, it prints something like this: PR description: Bumpito has found new common libraries versions. Refer to the following changelogs to see what is new :newspaper_roll: Changelogs common-kafka 10.86.0 Remove jackson-core explicit dependency from common-kafka 10.85.1 remove consumer cron tick code common-i18n 4.5.0 Fix Federal Holidays for Mexico common-crypto 10.24.1 improve the way of communicating error getting s3 keys common-metrics 10.6.0 Add prometheus text format parser common-redis Could not parse changelog :( @ericdallo Following up on this, antq now has the ability to surface links to the GitHub diff between the old and the new version. It's slightly less ergonomic than a compiled list of changelog updates- but it is available for little extra effort. I've updated the commit messages to include these links in this PR: https://github.com/nnichols/clojure-dependency-update-action/pull/9 Would you find these sufficient for now? @michols I think worth the shot, better than nothing! thank you Closed by: https://github.com/nnichols/clojure-dependency-update-action/pull/9
2025-04-01T06:39:46.748559
2015-02-10T18:28:58
57215196
{ "authors": [ "derekperkins", "fulldecent", "mazubieta" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8974", "repo": "nnnick/Chart.js", "url": "https://github.com/nnnick/Chart.js/issues/930" }
gharchive/issue
Highlight Lines instead of Events in LineChart Hi, thanks for the nice library! I haven't come across if this is possible or not. Is it possible to highlight a line, instead of the event on mouse hover? The use case would be to be able to easily follow the history of a particular line, throughout all events. With anything more than a dozen or so lines, it becomes difficult to track the progress of a line. Thanks. Hello, I'm sorry this issue hasn't received much attention yet. It seems this is an implementation question, could you please post your question to StackOverflow with the Chart.js tag and include a link here. This will ensure your question reaches the largest audience. https://stackoverflow.com/questions/tagged/chart.js Huzzah! The first alpha of Chart.js 2.0 has landed and should fix this issue. Check out the release and try it out! We've got a lot of momentum right now, so please help us test so we can launch 2.0 Gold by the end of the month. https://github.com/nnnick/Chart.js/releases/tag/v2.0-alpha I'm closing this issue for now, but if you have implementation questions or find bugs, please create a jsfiddle and post the link here and we'll reopen this issue and get it fixed.
2025-04-01T06:39:46.751531
2024-08-30T08:26:24
2496573796
{ "authors": [ "anyj0527", "taos-ci" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8975", "repo": "nnstreamer/api", "url": "https://github.com/nnstreamer/api/pull/556" }
gharchive/pull-request
[android] Set envvar for Qualcomm related filters QNN / SNPE needs some additional tasks. This job has been done by nnstreamer's each sub-plugin. Let the task be done in nnstreamer-native-api.c https://github.com/nnstreamer/nnstreamer/pull/4563 and this PR should be merged together. :memo: TAOS-CI Version: 1.5.20200925. Thank you for submitting PR #556. Please a submit 1commit/1PR (one commit per one PR) policy to get comments quickly from reviewers. Your PR must pass all verificiation processes of cibot before starting a review process from reviewers. If you are new member to join this project, please read manuals in documentation folder and wiki page. In order to monitor a progress status of your PR in more detail, visit http://ci.nnstreamer.ai/.
2025-04-01T06:39:46.760792
2021-06-25T10:35:28
930054204
{ "authors": [ "taos-ci", "zhoonit" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8976", "repo": "nnstreamer/nntrainer", "url": "https://github.com/nnstreamer/nntrainer/pull/1340" }
gharchive/pull-request
[Wait for #1339] [Resnet] Connect the model with cifar100 [Pending commit: #1339] [Resnet] Connect the model with cifar100 **Changes proposed in this PR:** - Implement cifar100dataloader - Connect the data loader **Self evaluation:** 1. Build test: [X]Passed [ ]Failed [ ]Skipped 2. Run test: [X]Passed [ ]Failed [ ]Skipped Signed-off-by: Jihoon Lee<EMAIL_ADDRESS> :memo: TAOS-CI Version: 1.5.20200925. Thank you for submitting PR #1340. Please a submit 1commit/1PR (one commit per one PR) policy to get comments quickly from reviewers. Your PR must pass all verificiation processes of cibot before starting a review process from reviewers. If you are new member to join this project, please read manuals in documentation folder and wiki page. In order to monitor a progress status of your PR in more detail, visit http://nnstreamer.mooo.com/. :octocat: cibot: @zhoonit, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://nnstreamer.mooo.com/nntrainer/ci/repo-workers/pr-checker/1340-202106301147160.68477702140808-b72c6f2e4171fd48f8f7f3d062b38a56ca92f95f/. :octocat: cibot: @zhoonit, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://nnstreamer.mooo.com/nntrainer/ci/repo-workers/pr-checker/1340-202107051429260.86057209968567-39d732fa59d7212522da55b25c70736d7248e2ae/.