added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:37:19.096017
2024-05-27T13:26:26
2319159008
{ "authors": [ "r-ryantm" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2115", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/315086" }
gharchive/pull-request
nu_scripts: 0-unstable-2024-05-19 -> 0-unstable-2024-06-01 Automatic update generated by nixpkgs-update tools. This update was made based on information from passthru.updateScript. meta.description for nu_scripts is: A place to share Nushell scripts with each other meta.homepage for nu_scripts is: https://github.com/nushell/nu_scripts Updates performed Ran passthru.UpdateScript To inspect upstream changes Compare changes on GitHub Impact Checks done built on NixOS The tests defined in passthru.tests, if any, passed found 0-unstable-2024-05-26 in filename of file in /nix/store/sk4wljbnq1pnrh8hps19y6cy2hpl43vi-nu_scripts-0-unstable-2024-05-26 Rebuild report (if merged into master) (click to expand) 2 total rebuild path(s) 2 package rebuild(s) First fifty rebuilds by attrpath nu_scripts Instructions to test this update (click to expand) Either download from Cachix: nix-store -r /nix/store/sk4wljbnq1pnrh8hps19y6cy2hpl43vi-nu_scripts-0-unstable-2024-05-26 \ --option binary-caches 'https://cache.nixos.org/ https://nix-community.cachix.org/' \ --option trusted-public-keys ' nix-community.cachix.org-1:mB9FSh9qf2dCimDSUo8Zy7bkq5CX+/rkCWyvRCYg3Fs= cache.nixos.org-1:6NCHdD59X431o0gWypbMrAURkbJ16ZPMQFGspcDShjY= ' (The Cachix cache is only trusted for this store-path realization.) For the Cachix download to work, your user must be in the trusted-users list or you can use sudo since root is effectively trusted. Or, build yourself: nix-build -A nu_scripts https://github.com/r-ryantm/nixpkgs/archive/9916a250d26b6240e271502b7bb27ea48d5aab4f.tar.gz Or: nix build github:r-ryantm/nixpkgs/9916a250d26b6240e271502b7bb27ea48d5aab4f#nu_scripts After you've downloaded or built it, look at the files and if there are any, run the binaries: ls -la /nix/store/sk4wljbnq1pnrh8hps19y6cy2hpl43vi-nu_scripts-0-unstable-2024-05-26 ls -la /nix/store/sk4wljbnq1pnrh8hps19y6cy2hpl43vi-nu_scripts-0-unstable-2024-05-26/bin Pre-merge build results We have automatically built all packages that will get rebuilt due to this change. This gives evidence on whether the upgrade will break dependent packages. Note sometimes packages show up as failed to build independent of the change, simply because they are already broken on the target branch. Result of nixpkgs-review --extra-nixpkgs-config '{ allowInsecurePredicate = x: true; }' run on x86_64-linux 1 1 package built: nu_scripts Maintainer pings cc @CardboardTurkey for testing. Automatic update generated by nixpkgs-update tools. This update was made based on information from passthru.updateScript. meta.description for nu_scripts is: A place to share Nushell scripts with each other meta.homepage for nu_scripts is: https://github.com/nushell/nu_scripts Updates performed Ran passthru.UpdateScript To inspect upstream changes Compare changes on GitHub Impact Checks done built on NixOS The tests defined in passthru.tests, if any, passed found 0-unstable-2024-06-01 in filename of file in /nix/store/n156kvffwhyqh9azsazblhis6xhpmxwx-nu_scripts-0-unstable-2024-06-01 Rebuild report (if merged into master) (click to expand) 2 total rebuild path(s) 2 package rebuild(s) First fifty rebuilds by attrpath nu_scripts Instructions to test this update (click to expand) Either download from Cachix: nix-store -r /nix/store/n156kvffwhyqh9azsazblhis6xhpmxwx-nu_scripts-0-unstable-2024-06-01 \ --option binary-caches 'https://cache.nixos.org/ https://nix-community.cachix.org/' \ --option trusted-public-keys ' nix-community.cachix.org-1:mB9FSh9qf2dCimDSUo8Zy7bkq5CX+/rkCWyvRCYg3Fs= cache.nixos.org-1:6NCHdD59X431o0gWypbMrAURkbJ16ZPMQFGspcDShjY= ' (The Cachix cache is only trusted for this store-path realization.) For the Cachix download to work, your user must be in the trusted-users list or you can use sudo since root is effectively trusted. Or, build yourself: nix-build -A nu_scripts https://github.com/r-ryantm/nixpkgs/archive/bec6bcc79f8e7af0f98e2cb0c43e0445d9d15dc6.tar.gz Or: nix build github:r-ryantm/nixpkgs/bec6bcc79f8e7af0f98e2cb0c43e0445d9d15dc6#nu_scripts After you've downloaded or built it, look at the files and if there are any, run the binaries: ls -la /nix/store/n156kvffwhyqh9azsazblhis6xhpmxwx-nu_scripts-0-unstable-2024-06-01 ls -la /nix/store/n156kvffwhyqh9azsazblhis6xhpmxwx-nu_scripts-0-unstable-2024-06-01/bin Pre-merge build results We have automatically built all packages that will get rebuilt due to this change. This gives evidence on whether the upgrade will break dependent packages. Note sometimes packages show up as failed to build independent of the change, simply because they are already broken on the target branch. Result of nixpkgs-review --extra-nixpkgs-config '{ allowInsecurePredicate = x: true; }' run on x86_64-linux 1 1 package built: nu_scripts Maintainer pings cc @CardboardTurkey for testing.
2025-04-01T06:37:19.108118
2024-05-31T18:36:59
2328393031
{ "authors": [ "r-vdp" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2116", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/316225" }
gharchive/pull-request
Remove the revCount attribute from the generated flake registry Description of changes The revCount attribute is only present when the flake is an actual git repo and not when using a flake reference with an input scheme for a git forge that downloads tarballs instead (like github: or sourcehut:). Because of this, the hash of the nixos config changes depending on whether the config was built from a local git clone or from a flake ref using one of those input schemes. This leads to confusion (I lost quite some time figuring out what happened here) and unnecessary rebuilds. To my knowledge, there is no immediate benefit to including revCount in the flake registry entry. Things done Built on platform(s) [ ] x86_64-linux [ ] aarch64-linux [ ] x86_64-darwin [ ] aarch64-darwin For non-Linux: Is sandboxing enabled in nix.conf? (See Nix manual) [ ] sandbox = relaxed [ ] sandbox = true [ ] Tested, as applicable: NixOS test(s) (look inside nixos/tests) and/or package tests or, for functions and "core" functionality, tests in lib/tests or pkgs/test made sure NixOS tests are linked to the relevant packages [ ] Tested compilation of all packages that depend on this change using nix-shell -p nixpkgs-review --run "nixpkgs-review rev HEAD". Note: all changes have to be committed, also see nixpkgs-review usage [ ] Tested basic functionality of all binary files (usually in ./result/bin/) 24.11 Release Notes (or backporting 23.11 and 24.05 Release notes) [ ] (Package updates) Added a release notes entry if the change is major or breaking [ ] (Module updates) Added a release notes entry if the change is significant [ ] (Module addition) Added a release notes entry if adding a new NixOS module [ ] Fits CONTRIBUTING.md. Add a :+1: reaction to pull requests you find important. Well at least for sourcehut, I don't get a revCount when I build using a sourcehut: reference, but I do get one when I build from a local clone: Lix 2.90.0-beta.1 Type :? for help. nix-repl> :lf sourcehut:~r-vdp/nixos-config Added 21 variables. nix-repl> sourceInfo { lastModified =<PHONE_NUMBER>; lastModifiedDate = "20240601090531"; narHash = "sha256-zCzXuwkyGoQDRvCSkOi2uv1NMd+yZjufZG6lzGVP5AA="; outPath = "/nix/store/9ilkiqfx5m85riwmnc6b0p4qdrwy4s95-source"; rev = "23b24e1e7d993cb5ec005a3a22345524f9579c39"; shortRev = "23b24e1"; } nix-repl> :lf . Added 23 variables. nix-repl> sourceInfo { lastModified =<PHONE_NUMBER>; lastModifiedDate = "20240601090531"; narHash = "sha256-zCzXuwkyGoQDRvCSkOi2uv1NMd+yZjufZG6lzGVP5AA="; outPath = "/nix/store/9ilkiqfx5m85riwmnc6b0p4qdrwy4s95-source"; rev = "23b24e1e7d993cb5ec005a3a22345524f9579c39"; revCount = 2319; shortRev = "23b24e1"; submodules = false; } @lf- have you had the time to give this some more thought?
2025-04-01T06:37:19.122004
2024-06-09T08:05:10
2342088711
{ "authors": [ "fabaff" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2117", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/318471" }
gharchive/pull-request
python311Packages.aioshelly: 10.0.0 -> 10.0.1 Diff: https://github.com/home-assistant-libs/aioshelly/compare/refs/tags/10.0.0...10.0.1 Changelog: https://github.com/home-assistant-libs/aioshelly/releases/tag/10.0.1 Description of changes Things done Built on platform(s) [ ] x86_64-linux [ ] aarch64-linux [ ] x86_64-darwin [ ] aarch64-darwin For non-Linux: Is sandboxing enabled in nix.conf? (See Nix manual) [ ] sandbox = relaxed [ ] sandbox = true [ ] Tested, as applicable: NixOS test(s) (look inside nixos/tests) and/or package tests or, for functions and "core" functionality, tests in lib/tests or pkgs/test made sure NixOS tests are linked to the relevant packages [ ] Tested compilation of all packages that depend on this change using nix-shell -p nixpkgs-review --run "nixpkgs-review rev HEAD". Note: all changes have to be committed, also see nixpkgs-review usage [ ] Tested basic functionality of all binary files (usually in ./result/bin/) 24.11 Release Notes (or backporting 23.11 and 24.05 Release notes) [ ] (Package updates) Added a release notes entry if the change is major or breaking [ ] (Module updates) Added a release notes entry if the change is significant [ ] (Module addition) Added a release notes entry if adding a new NixOS module [ ] Fits CONTRIBUTING.md. Add a :+1: reaction to pull requests you find important. Result of nixpkgs-review pr 318471 run on x86_64-linux 1 41 packages built: home-assistant-component-tests.airthings_ble home-assistant-component-tests.aranet home-assistant-component-tests.bluemaestro home-assistant-component-tests.bluetooth_adapters home-assistant-component-tests.bluetooth_le_tracker home-assistant-component-tests.bthome home-assistant-component-tests.dormakaba_dkey home-assistant-component-tests.eufylife_ble home-assistant-component-tests.fjaraskupan home-assistant-component-tests.gardena_bluetooth home-assistant-component-tests.govee_ble home-assistant-component-tests.homekit_controller home-assistant-component-tests.ibeacon home-assistant-component-tests.inkbird home-assistant-component-tests.kegtron home-assistant-component-tests.keymitt_ble home-assistant-component-tests.lamarzocco home-assistant-component-tests.ld2410_ble home-assistant-component-tests.led_ble home-assistant-component-tests.moat home-assistant-component-tests.mopeka home-assistant-component-tests.oralb home-assistant-component-tests.private_ble_device home-assistant-component-tests.qingping home-assistant-component-tests.rapt_ble home-assistant-component-tests.ruuvitag_ble home-assistant-component-tests.sensirion_ble home-assistant-component-tests.sensorpro home-assistant-component-tests.sensorpush home-assistant-component-tests.shelly home-assistant-component-tests.snooz home-assistant-component-tests.switchbot home-assistant-component-tests.thermobeacon home-assistant-component-tests.thermopro home-assistant-component-tests.tilt_ble home-assistant-component-tests.xiaomi_ble home-assistant-component-tests.yalexs_ble python311Packages.aioshelly python311Packages.aioshelly.dist python312Packages.aioshelly python312Packages.aioshelly.dist
2025-04-01T06:37:19.192567
2024-07-02T19:22:48
2386990824
{ "authors": [ "Aleksanaa", "Atemu", "GaetanLepage", "GeoffreyFrogeye", "ISibboI", "Kernald", "NyCodeGHG", "PowerUser64", "RomeoV", "Scrumplex", "Titaniumtown", "aciceri", "bct", "crertel", "dav-wolff", "diogotcorreia", "dotlambda", "eclairevoyant", "gitwah", "h7x4", "happysalada", "inclyc", "jlbribeiro", "jpds", "jvanbruegge", "oddlama", "onny", "qbit", "rathmerdominik", "rhoriguchi", "xinyangli" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2118", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/324127" }
gharchive/pull-request
immich: init at 1.109.2; nixos/immich: init module Description of changes This adds the package, nixos module and nixos test for immich Closes #244803 Things done Built on platform(s) [x] x86_64-linux [ ] aarch64-linux [ ] x86_64-darwin [ ] aarch64-darwin For non-Linux: Is sandboxing enabled in nix.conf? (See Nix manual) [ ] sandbox = relaxed [ ] sandbox = true [ ] Tested, as applicable: NixOS test(s) (look inside nixos/tests) and/or package tests or, for functions and "core" functionality, tests in lib/tests or pkgs/test made sure NixOS tests are linked to the relevant packages [ ] Tested compilation of all packages that depend on this change using nix-shell -p nixpkgs-review --run "nixpkgs-review rev HEAD". Note: all changes have to be committed, also see nixpkgs-review usage [x] Tested basic functionality of all binary files (usually in ./result/bin/) 24.11 Release Notes (or backporting 23.11 and 24.05 Release notes) [ ] (Package updates) Added a release notes entry if the change is major or breaking [ ] (Module updates) Added a release notes entry if the change is significant [x] (Module addition) Added a release notes entry if adding a new NixOS module [x] Fits CONTRIBUTING.md. Add a :+1: reaction to pull requests you find important. Does this superseed #244803? Yes I occasionally see these messages in my logs: server[36665]: [Nest] 36665 - 07/05/2024, 12:00:00 AM ERROR [Microservices:JobService] Unable to run job handler (thumbnailGeneration/generate-preview): Error: Cannot find ffprobe server[36665]: [Nest] 36665 - 07/05/2024, 12:00:00 AM ERROR [Microservices:JobService] Error: Cannot find ffprobe server[8610]: [Nest] 8610 - 07/02/2024, 10:35:52 PM WARN [Api:ServerInfoRepository~dfqq7ca6] Failed to read build-lock.json ffprobe is provided by the ffmpeg the build already depends on. Perhaps it needs to be wrapped to have access to the binary at runtime. Just wanted to say I'm psyched to see this getting added. Thanks for the work! I have addressed most of the review comments now and rebased onto the current master. This broke albumentations, which I fixed in a seperate commit. We will still need to wait for the upstream PR to get the geonames without downloading the full docker container Ok, the albumentation fix got its own PR, I will rebase once that is merged: https://github.com/NixOS/nixpkgs/pull/325019/files Ok, the albumentation fix got its own PR, I will rebase once that is merged: #325019 Merged :) Yes, but we are waiting for a PR from upstream. When that is merged and a new release is cut, I will update this here accordingly We get this error in the Immich logs while indexing photos: Jul 09 23:26:18 server2 server[1740]: [Nest] 1740 - 07/09/2024, 11:26:18 PM ERROR [Microservices:JobService] Unable to run job handler (faceDetection/face-detection): Error: Machine learning request to "http://immich-machine-learning:3003" failed with ConnectTimeoutError: Connect Timeout Error Setting services.immich.environment.IMMICH_MACHINE_LEARNING_URL = "http://localhost:3003"; fixes the issue. Maybe we should have a default like IMMICH_MACHINE_LEARNING_URL = "${cfg.host}:3003"; or something. Also would it be possible to rename the environment option to settings? We get this error in the Immich logs while indexing photos: Jul 09 23:26:18 server2 server[1740]: [Nest] 1740 - 07/09/2024, 11:26:18 PM ERROR [Microservices:JobService] Unable to run job handler (faceDetection/face-detection): Error: Machine learning request to "http://immich-machine-learning:3003" failed with ConnectTimeoutError: Connect Timeout Error Setting services.immich.environment.IMMICH_MACHINE_LEARNING_URL = "http://localhost:3003"; fixes the issue. Maybe we should have a default like IMMICH_MACHINE_LEARNING_URL = "${cfg.host}:3003"; or something. I have no idea whether upstream actually supports that, but a Unix socket could be an even nicer option if they do. There's an issue with HEIC / HEIF images, they can't be viewed and theres no thumbnail generated. Might be related to the implementation of libheif? Error log of immich-server: Jul 14 11:24:59 server2 server[174045]: [Nest] 174045 - 07/14/2024, 11:24:59 AM ERROR [Microservices:JobService] Unable to run job handler (thumbnailGeneration/generate-preview): Error: /var/lib/immich/upload/8046b9e9-b0cf-46d1-9bc7-542fca747436/81/a3/81a3dc3a-db01-45ea-abe5-ed6996ced369.HEIC: bad seek to 1531357 Jul 14 11:24:59 server2 server[174045]: heif: Unsupported feature: Unsupported codec (4.3000) Jul 14 11:24:59 server2 server[174045]: [Nest] 174045 - 07/14/2024, 11:24:59 AM ERROR [Microservices:JobService] Error: /var/lib/immich/upload/8046b9e9-b0cf-46d1-9bc7-542fca747436/81/a3/81a3dc3a-db01-45ea-abe5-ed6996ced369.HEIC: bad seek to 1531357 Jul 14 11:24:59 server2 server[174045]: heif: Unsupported feature: Unsupported codec (4.3000) Jul 14 11:24:59 server2 server[174045]: at Sharp.toFile (/nix/store/k6sa9kikymqmbnpfbkvnamk1ci793d2p-immich-1.107.1/node_modules/sharp/lib/output.js:89:19) Jul 14 11:24:59 server2 server[174045]: at MediaRepository.generateThumbnail (/nix/store/k6sa9kikymqmbnpfbkvnamk1ci793d2p-immich-1.107.1/dist/repositories/media.repository.js:69:14) Jul 14 11:24:59 server2 server[174045]: at MediaService.generateThumbnail (/nix/store/k6sa9kikymqmbnpfbkvnamk1ci793d2p-immich-1.107.1/dist/services/media.service.js:158:48) Jul 14 11:24:59 server2 server[174045]: at process.processTicksAndRejections (node:internal/process/task_queues:95:5) Jul 14 11:24:59 server2 server[174045]: at async MediaService.handleGeneratePreview (/nix/store/k6sa9kikymqmbnpfbkvnamk1ci793d2p-immich-1.107.1/dist/services/media.service.js:135:29) Jul 14 11:24:59 server2 server[174045]: at async /nix/store/k6sa9kikymqmbnpfbkvnamk1ci793d2p-immich-1.107.1/dist/services/job.service.js:148:36 Jul 14 11:24:59 server2 server[174045]: at async Worker.processJob (/nix/store/k6sa9kikymqmbnpfbkvnamk1ci793d2p-immich-1.107.1/node_modules/bullmq/dist/cjs/classes/worker.js:394:28) Jul 14 11:24:59 server2 server[174045]: at async Worker.retryIfFailed (/nix/store/k6sa9kikymqmbnpfbkvnamk1ci793d2p-immich-1.107.1/node_modules/bullmq/dist/cjs/classes/worker.js:581:24) Jul 14 11:24:59 server2 server[174045]: [Nest] 174045 - 07/14/2024, 11:24:59 AM ERROR [Microservices:JobService] Object: Jul 14 11:24:59 server2 server[174045]: { Jul 14 11:24:59 server2 server[174045]: "id": "dd9353ed-3d2d-4099-b6cc-a5e8d5382519", Jul 14 11:24:59 server2 server[174045]: "source": "upload" Jul 14 11:24:59 server2 server[174045]: } Isn't that expected? It is not listed as one of the formats on the Sharp website: https://sharp.pixelplumbing.com/#formats I think the formatting check is broken. I do exactly what the CI says and nothing is changed any more but the check still fails CI uses precisely the commit mentioned here to pull nixfkt-rfc-style from: https://github.com/NixOS/nixpkgs/blob/master/ci/pinned-nixpkgs.json#L2 Using that should pass the check Yes, the missing rebase was the issue. I think I have now accounted for every bit of feedback, as well as updated to the latest version of immich I noticed that a few videos were not getting thumbnails, and traced it back to missing ffmpeg encoders. Switching to ffmpeg-full fixes that I noticed that a few videos were not getting thumbnails Do you have any specific examples? Container type, encoding, etc? iirc ffmpeg-full was discouraged due to the larger closure. I know that at least webp is missing from the normal ffmpeg Let's leave it as ffmpeg for now, I'll go through and see if we can enable some more codecs in the regular ffmpeg package without impacting closure too much. Related: https://github.com/lovell/sharp/issues/3844 https://github.com/immich-app/immich/issues/4899 https://github.com/libvips/libvips/issues/3380 Nice share, this motivated me to dig a bit more. So in the Docker setup they indeed add the -Dtiff=disable flag. I thought it would be a simple matter of overriding vips with that flag, however it doesn't cut it. It seems that the vips in buildInputs doesn't get used at all. The<EMAIL_ADDRESS>appears in the immich derivation regardless if it's included or not. Its ldd output is very small, which makes me think this is the default library included in the sharp npm package, just patchelf-ed. I managed to make it work by adding the below instructions to the derivation but surely there must be a better way to make sharp use the provided vips. rm<EMAIL_ADDRESS>ln -s ${vips.out}/lib/libvips-cpp.so.42.17.2<EMAIL_ADDRESS> Given the pace at which upstream development happens, we might want to warn users to stop the immich apps from auto-updating as we cannot ensure that we will be able to keep up in a timely manner. I thought it would be a simple matter of overriding vips with that flag, however it doesn't cut it. It seems that the vips in buildInputs doesn't get used at all. The<EMAIL_ADDRESS>appears in the immich derivation regardless if it's included or not. Its ldd output is very small, which makes me think this is the default library included in the sharp npm package, just patchelf-ed. That should not be the case. There's no patchelfing going on here. If we're downloading binaries, this should also be tagged with the appropriate sourceProvenance. That should not be the case. There's no patchelfing going on here. Ah yes indeed, I misinterpreted the output of ldd. It's the same file regardless of the OS. If we're downloading binaries, this should also be tagged with the appropriate sourceProvenance. I hope this won't be necessary as sharp supports building from the system vips. While investigating ad-hoc I noticed that it silently fails to build from source unless node-gyp and node-addon-api are installed before running npm install sharp. I suspect that the latter not being present as buildInput causing the issue, although as I haven't managed to package it I can't really be sure. But when it is, then there will be a ./node_modules/sharp/src/build/Release/sharp-linux-x64.node file which is linked against the provided vips. It still embedds the prebuilt<EMAIL_ADDRESS>for some reason but it has less priority so we should be able to delete it (and maybe the whole of ./node_modules/sharp/src/build/Release/sharp-linux-x64.node as upstream does. (I was testing the automatic review using nixf-tidy, comments are generated above). Seems like these issues got flagged by the new CI check: Editing title won't print them to 'Files changed' overview, to avoid duplicate. Once a push happens it will. Also checking escaping with is disabled now. Not sure why you're replying to me about that, I edited the title because it was inaccurate. And the nixf-tidy check did fire after I edited. In case you don't know why it behaved like that, nevermind Not sure if this is the correct place to ask, but I tried the Immich module from this PR (on commit 7290604) and it's not working on my server (however it does run on my main computer). I also tried running the official docker images and those worked, so I presume the issue is with the nix package. I set IMMICH_LOG_LEVEL to verbose, but still the log only shows this which is not enough information for me to gain any insight: `journalctl -b -u immich-server` Started Immich backend server (Self-hosted photo and video backup solution). immich-server.service: Main process exited, code=dumped, status=4/ILL immich-server.service: Failed with result 'core-dump'. immich-server.service: Consumed 27.786s CPU time, no IP traffic. immich-server.service: Scheduled restart job, restart counter is at 1. Started Immich backend server (Self-hosted photo and video backup solution). immich-server.service: Main process exited, code=dumped, status=4/ILL immich-server.service: Failed with result 'core-dump'. immich-server.service: Consumed 19.285s CPU time, no IP traffic. immich-server.service: Scheduled restart job, restart counter is at 2. Started Immich backend server (Self-hosted photo and video backup solution). immich-server.service: Main process exited, code=dumped, status=4/ILL immich-server.service: Failed with result 'core-dump'. immich-server.service: Consumed 18.356s CPU time, no IP traffic. immich-server.service: Scheduled restart job, restart counter is at 3. Started Immich backend server (Self-hosted photo and video backup solution). immich-server.service: Main process exited, code=dumped, status=4/ILL immich-server.service: Failed with result 'core-dump'. immich-server.service: Consumed 18.251s CPU time, no IP traffic. What kind of processor is this? You're running into illegal instructions and that should basically never happen on non-legacy hardware, especially not given that we target a generic x86_64-processor that should even run machines over a decade old. Yeah it's quite old. It's an Intel Atom CPU D525. Is this not supported by nixpkgs? I haven't run into any issues running NixOS on it so far. You'll definitely have to disable machine-learning. Unsure if that'd be sufficient though, that cpu only supports SSE3 I've been testing this module today, and discovered that generating thumbnails failed for heic files from my phone. Log Jul 29 19:33:34 weilite server[795]: [Nest] 795 - 07/29/2024, 7:33:34 PM ERROR [Microservices:JobService] Unable to run job handler (thumbnailGeneration/generate-preview): Error: /mnt/XinPhotos/immich/library/admin/2024/2024-05-14/20240514_161132.heic: bad seek to 7948144 Jul 29 19:33:34 weilite server[795]: heif: Unsupported feature: Unsupported codec (4.3000) Jul 29 19:33:34 weilite server[795]: [Nest] 795 - 07/29/2024, 7:33:34 PM ERROR [Microservices:JobService] Error: /mnt/XinPhotos/immich/library/admin/2024/2024-05-14/20240514_161132.heic: bad seek to 7948144 Jul 29 19:33:34 weilite server[795]: heif: Unsupported feature: Unsupported codec (4.3000) Jul 29 19:33:34 weilite server[795]: at Sharp.toFile (/nix/store/hjw0dx1ap99l7dqi06k1a5f2dmgl0akq-immich-1.109.2/node_modules/sharp/lib/output.js:89:19) Jul 29 19:33:34 weilite server[795]: at MediaRepository.generateThumbnail (/nix/store/hjw0dx1ap99l7dqi06k1a5f2dmgl0akq-immich-1.109.2/dist/repositories/media.repository.js:69:14) Jul 29 19:33:34 weilite server[795]: at MediaService.generateThumbnail (/nix/store/hjw0dx1ap99l7dqi06k1a5f2dmgl0akq-immich-1.109.2/dist/services/media.service.js:164:48) Jul 29 19:33:34 weilite server[795]: at process.processTicksAndRejections (node:internal/process/task_queues:95:5) Jul 29 19:33:34 weilite server[795]: at async MediaService.handleGeneratePreview (/nix/store/hjw0dx1ap99l7dqi06k1a5f2dmgl0akq-immich-1.109.2/dist/services/media.service.js:135:29) Jul 29 19:33:34 weilite server[795]: at async /nix/store/hjw0dx1ap99l7dqi06k1a5f2dmgl0akq-immich-1.109.2/dist/services/job.service.js:148:36 Jul 29 19:33:34 weilite server[795]: at async Worker.processJob (/nix/store/hjw0dx1ap99l7dqi06k1a5f2dmgl0akq-immich-1.109.2/node_modules/bullmq/dist/cjs/classes/worker.js:394:28) Jul 29 19:33:34 weilite server[795]: at async Worker.retryIfFailed (/nix/store/hjw0dx1ap99l7dqi06k1a5f2dmgl0akq-immich-1.109.2/node_modules/bullmq/dist/cjs/classes/worker.js:581:24) Jul 29 19:33:34 weilite server[795]: [Nest] 795 - 07/29/2024, 7:33:34 PM ERROR [Microservices:JobService] Object: Jul 29 19:33:34 weilite server[795]: { Jul 29 19:33:34 weilite server[795]: "id": "db366035-dce6-4d12-8127-67ca60a12fb3" Jul 29 19:33:34 weilite server[795]: } This might be related to lovell/sharp#2518, so I agree with @GeoffreyFrogeye that it seems like vips provided in buildInputs is not used in the final result. What's left here to fix? Looking forward to seeing this merged! See the unresolved review comments. I should have time this weekend to address the remaining comments Great. I think we all appreciate @jvanbruegge continued work, including pioneering the packaging of immich after other continued efforts have failed (e.g. #244803). Considering how this package already provides a lot of value, I think it would be great to merge this after e.g. this weekend, even if not all problems are resolved. We can then open issues for any remaining problems, e.g. with thumbnail creation, and others can step in to fix those problems. I think it should also generally not be seen as rude to simply take a seemingly stalled PR and continue work on it yourself. In the worst case where the original author does wish to continue working on it, they can simply pull back your changes into their branch. That's the beauty of FOSS and distributed version control. So, I've addressed all review comments, and updated immich to the latest version. I am now looking into rebuilding sharp. @Atemu adding the default environment to the environment config option causes the manual to fail, what should be done here? ok, that took a while, but I finally managed to debug the build failure for pgvecto-rs. Now I can actually test my new changes and investigate the sharp rebuild @jvanbruegge Maintainer of the pgvecto-rs package here. Thanks for fixing the build failure, I hadn't noticed it was failing! Judging by the date (29th July) it started failing, it seems to be releated to the upgrade to rustc 1.79. Question for more experienced maintainers: is there a way to get notified when hydra fails to build a package you maintain? adding the default environment to the environment config option causes the manual to fail, what should be done here? The usual cause is using options or config in a default value, which necessitates setting defaultText = lib.literalExpression "...", but details will depend on the actual cause if it comes up again. My only comment is around backups. I dont know immich at all, but do they have a backup solution included or do they just leave that to the user ? If they leave it to the user then it makes sense create a dedicated user and group for immich so that the backup user can just read immich directory as well. If they have a backup solution then maybe dynamic user might be a better fit. To answer another comment i dont know how to find out with hydra that your builds failed. In my experience you just have to be on unstable and update your system every week with your packages. Many people were involved in this PR, are we all ok to merge ? If they leave it to the user then it makes sense create a dedicated user and group for immich it already does? Immich does not have a dedicated backup feature so far Many people were involved in this PR, are we all ok to merge ? No, I am still trying to fix the thumbnails maybe dynamic user might be a better fit I think NixOS modules should always use a dynamic user if possible. I assume in this case it's impossible because multiple services need access to the same data? Many people were involved in this PR, are we all ok to merge ? The newest substantial changes were pushed just hours ago, please give people some time to review before calling for further actions. So, I tested if I really managed to fix the raw/heif issues and yes indeed :tada: Here are some samples that work on a Docker setup if you want to test for yourself. Thanks, for having a look at the issue, vips is built with -Dtiff=disable now. Please pull https://github.com/jvanbruegge/nixpkgs/pull/1 Btw, 1.112 came out eith support for pg-vector 0.3. We are waiting before merging 0.3 to nixpkgs for the update for 1.112 to be done. Take the time you need though . https://wiki.nixos.org/wiki/Nixpkgs/Reviewing_changes Thank you for this! Lookin forward for the merge Ok, I have updated immich to the latest version and addressed all review comments again 1.115.0 is running nice and smooth for me! Thanks for all the hard work! Yes, I also upgraded my deployment to the latest version of this PR and everything works Given the huge popularity of Immich, I'd like to take a bit of time to really thank all the people involved in packaging this. @jvanbruegge I'd like to thank you for bringing this to the finish line; I know it probably wasn't easy, given this PR was open almost 3 months ago. You started as a user requesting a package and ended up stepping up and sharing your work with the rest of the NixOS community, when it looked like the consensus was having a separate project. I'm sure the community will help you maintain both the package and the module up to date, given the frenetic rate at which Immich is being developed :) There are a couple of unsung heroes I'd like to mention: @oddlama, for taking one of the first stabs at this more than a year ago! @diogotcorreia, for having packaged one of the blockers, pgvecto.rs! You also started by sharing your work in a comment, and ended up making it available to the whole nixpkgs and unblocking @oddlama / @jvanbruegge efforts! Thank you! And a huge thanks to all the reviewers involved! Amazing work! How can I properly integrate this into my nix config? Does anyone have a public server config I can take a look at. Thanks! https://github.com/jvanbruegge/server-config/blob/86d8f6ff4c9e576cdfd94410dd4e35e58c6e8e5d/caladan/services.nix#L10 Thank you @jvanbruegge, just trying to figure out how to integrate this with the differing directories I have things stored in. Thanks again for all your hard work! Very much appreciated! back then i wrote this wiki page but not sure if its still up-to date https://wiki.nixos.org/wiki/Immich Thank you so much for finally bringing immich to nixos! I have one question: does the default setup support external libraries? Because I tried to add my local picture library on my server to immich as external library, but it cannot access any path I specify. It always gives EACCESS, even though the whole path is readable and executable by everyone. @ISibboI This is probably because of the service hardening I did in the first version. You will have to add your directory to a ReadWritePaths directive for the relevant systemd units to make it work @oddlama This is probably because of the service hardening I did in the first version. You will have to add your directory to a ReadWritePaths directive for the relevant systemd units to make it work This should be done automatically if it's not the default state directory imo. I don't think that's the issue, as I also do not use the standard path and have no issues Congratulations @jvanbruegge! This was one of the most through reviews I've seen, and you made it through! Thanks for all your hard work as well as the hard work of others to get this merged! I was able to migrate my Docker Compose deployment. An issue I ran into was that file paths in the database were relative to immich's working directory, meaning that everything was prefixed with upload/. I was able to remedy this by replacing upload/ with my mediaLocation (which is also what the UPLOAD_LOCATION variable from my Compose deployment pointed to). [!NOTE] This is what worked for my installation that is a few years old. I am not sure if newer installations behave differently. Some facts about my Docker Compose installation: The database is called immich The database user is called immich My UPLOAD_LOCATION (docker mount path) is set to /media/immich-library Make a backup of your current library and save a copy of the following database dump Enable NixOS Immich and switch. This will start an empty instance of Immich but will create the database and database user for us.1 In my case it looked like this: services.immich = { enable = true; mediaLocation = "/media/immich-library"; }; Stop NixOS Immich systemctl stop immich-server.service immich-machine-learning.service Stop Docker Immich. Leave your postgres container running Dump your database. The following command is going to dump it into Postgres' data directory, so it will be easy to access from your host system. docker compose exec database pg_dump --user immich -f /var/lib/postgresql/data/dump.sql immich Fix unset search_path variable.2 sed -i "s#SELECT pg_catalog.set_config('search_path', '', false);#SELECT pg_catalog.set_config('search_path', 'public', false);#" </path/to/immich/compose>/postgres/dump.sql Replace upload/ path prefix.3 In my case I replace it with /media/immich-library sed -i "s#upload/#/media/immich-library#g" </path/to/immich/compose>/postgres/dump.sql Prepare NixOS database sudo -u postgres psql immich immich=# <paste following codeblock and exit> DROP SCHEMA public CASCADE; DROP SCHEMA vectors CASCADE; CREATE SCHEMA public; CREATE EXTENSION IF NOT EXISTS unaccent; CREATE EXTENSION IF NOT EXISTS "uuid-ossp"; CREATE EXTENSION IF NOT EXISTS vectors; CREATE EXTENSION IF NOT EXISTS cube; CREATE EXTENSION IF NOT EXISTS earthdistance; CREATE EXTENSION IF NOT EXISTS pg_trgm; ALTER SCHEMA public OWNER TO immich; ALTER SCHEMA vectors OWNER TO immich; GRANT SELECT ON TABLE pg_vector_index_stat TO immich; ALTER EXTENSION vectors UPDATE; Import your dump. Make sure the file is readable. I moved it to a location that is readable by the postgres user. sudo -u postgres psql immich -f </path/to/immich/compose>/postgres/dump.sql Start NixOS Immich systemctl start immich-server.service immich-machine-learning.service Make sure you can actually view images on the web app. If any of these steps fail, you should have your original (unmodified!) database as well as a backup of your media location. 1: As a precaution, you could make your existing Immich library read-only (i.e. chown -R root:root /media/immich-library) to avoid any modifications done by this empty installation of Immich. 2: Restoring a dump that uses Postgres earthdistance seems to fail if search_path isn't set to the target schema. See this StackOverflow comment 3: Perhaps you can inspect your database dump for this first. In my case every file reference started with upload/ (i.e. upload/thumbs/ instead of just thumbs/) I hope this helps someone out there trying to get this working. @Scrumplex could you PR that into a migration doc and link it in the patch notes? Thank you @Scrumplex, I've been working on migration now, luckily I had a backup of my sql because I've broken it hahaha I have written more in the last hour than I did on average per day for my bachelor's thesis. :sweat_smile: See https://github.com/NixOS/nixpkgs/pull/344300 for my proposed NixOS manual documentation. I'll keep "Allow edits and access to secrets by maintainers" in case someone with commit access wants to commit changes into my PR. I will go to bed now! :zzz: Found an issue with the module, when running immich and changing the group services.immich.group = "test"; the redis service will have the wrong group and fail to start. The issue is this line, so it would make sense to just create the group if the name is immich. https://github.com/NixOS/nixpkgs/blob/23cbb250f3bf4f516a2d0bf03c51a30900848075/nixos/modules/services/databases/redis.nix#L356 > journalctl -r -u redis-immich.service Sep 25 04:42:53 HOSTNAME systemd[1]: Failed to start Redis Server - redis-immich. Sep 25 04:42:53 HOSTNAME systemd[1]: redis-immich.service: Failed with result 'exit-code'. Sep 25 04:42:53 HOSTNAME systemd[1]: redis-immich.service: Control process exited, code=exited, status=216/GROUP Sep 25 04:42:53 HOSTNAME (rep-conf)[13315]: redis-immich.service: Failed at step GROUP spawning /nix/store/338pjs37wgmhghid8wbai5qgyssihq2n-redis-immich-prep-conf: No such process Sep 25 04:42:53 HOSTNAME (rep-conf)[13315]: redis-immich.service: Failed to determine group credentials: No such process Sep 25 04:42:53 HOSTNAME systemd[1]: Starting Redis Server - redis-immich... If you use a non-socket postgres connection then services.immich.environment refers to cfg.database.port, but the module doesn't declare this option. Is this a bug, or am I missing something? @bct i believe that is a bug, I can also reproduce. Yes, that probably got lost somewhere during all of the rebasing for the reviews Anyone had success testing this on aarch64? It tries to build python3Packages.insightface which is broken on aarch64-linux, don't know if it would work removing it (without face recognition obviously). Moreover postgresqlPackages.pgvecto-rstoo is broken onaarch64-linux` and there may be other broken dependencies. Sorry to interject but please create separate issues rather than commenting on this PR unless it is specifically something that requires the attention of the people who were involved in this PR. A lot of people are subscribed and this code is in Nixpkgs now, so it should be treated like any other Nixpkgs code.
2025-04-01T06:37:19.205098
2024-07-02T20:06:04
2387064493
{ "authors": [ "SuperSandro2000" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2119", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/324140" }
gharchive/pull-request
fedifetcher: 7.1.1 -> 7.1.4 Diff: https://github.com/nanos/FediFetcher/compare/refs/tags/v7.1.1...v7.1.4 Description of changes Things done Built on platform(s) [x] x86_64-linux [ ] aarch64-linux [ ] x86_64-darwin [ ] aarch64-darwin For non-Linux: Is sandboxing enabled in nix.conf? (See Nix manual) [ ] sandbox = relaxed [ ] sandbox = true [ ] Tested, as applicable: NixOS test(s) (look inside nixos/tests) and/or package tests or, for functions and "core" functionality, tests in lib/tests or pkgs/test made sure NixOS tests are linked to the relevant packages [ ] Tested compilation of all packages that depend on this change using nix-shell -p nixpkgs-review --run "nixpkgs-review rev HEAD". Note: all changes have to be committed, also see nixpkgs-review usage [ ] Tested basic functionality of all binary files (usually in ./result/bin/) 24.11 Release Notes (or backporting 23.11 and 24.05 Release notes) [ ] (Package updates) Added a release notes entry if the change is major or breaking [ ] (Module updates) Added a release notes entry if the change is significant [ ] (Module addition) Added a release notes entry if adding a new NixOS module [x] Fits CONTRIBUTING.md. Add a :+1: reaction to pull requests you find important. @ofborg build fedifetcher
2025-04-01T06:37:19.215703
2024-07-28T16:58:13
2434025427
{ "authors": [ "fabaff" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2120", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/330689" }
gharchive/pull-request
python312Packages.cyclonedx-python-lib: refactor Description of changes Things done Built on platform(s) [ ] x86_64-linux [ ] aarch64-linux [ ] x86_64-darwin [ ] aarch64-darwin For non-Linux: Is sandboxing enabled in nix.conf? (See Nix manual) [ ] sandbox = relaxed [ ] sandbox = true [ ] Tested, as applicable: NixOS test(s) (look inside nixos/tests) and/or package tests or, for functions and "core" functionality, tests in lib/tests or pkgs/test made sure NixOS tests are linked to the relevant packages [ ] Tested compilation of all packages that depend on this change using nix-shell -p nixpkgs-review --run "nixpkgs-review rev HEAD". Note: all changes have to be committed, also see nixpkgs-review usage [ ] Tested basic functionality of all binary files (usually in ./result/bin/) 24.11 Release Notes (or backporting 23.11 and 24.05 Release notes) [ ] (Package updates) Added a release notes entry if the change is major or breaking [ ] (Module updates) Added a release notes entry if the change is significant [ ] (Module addition) Added a release notes entry if adding a new NixOS module [ ] Fits CONTRIBUTING.md. Add a :+1: reaction to pull requests you find important. Result of nixpkgs-review pr 330689 run on x86_64-linux 1 10 packages built: checkov checkov.dist cyclonedx-python cyclonedx-python.dist pip-audit pip-audit.dist python311Packages.cyclonedx-python-lib python311Packages.cyclonedx-python-lib.dist python312Packages.cyclonedx-python-lib python312Packages.cyclonedx-python-lib.dist
2025-04-01T06:37:19.225681
2024-10-04T12:12:50
2566229353
{ "authors": [ "Bot-wxt1221" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2121", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/346418" }
gharchive/pull-request
parpd: init 2.3.0 Fix https://github.com/NixOS/nixpkgs/issues/345461 Things done Built on platform(s) [x] x86_64-linux [ ] aarch64-linux [ ] x86_64-darwin [ ] aarch64-darwin For non-Linux: Is sandboxing enabled in nix.conf? (See Nix manual) [ ] sandbox = relaxed [ ] sandbox = true [x] Tested, as applicable: NixOS test(s) (look inside nixos/tests) and/or package tests or, for functions and "core" functionality, tests in lib/tests or pkgs/test made sure NixOS tests are linked to the relevant packages [x] Tested compilation of all packages that depend on this change using nix-shell -p nixpkgs-review --run "nixpkgs-review rev HEAD". Note: all changes have to be committed, also see nixpkgs-review usage [x] Tested basic functionality of all binary files (usually in ./result/bin/) 24.11 Release Notes (or backporting 23.11 and 24.05 Release notes) [ ] (Package updates) Added a release notes entry if the change is major or breaking [ ] (Module updates) Added a release notes entry if the change is significant [ ] (Module addition) Added a release notes entry if adding a new NixOS module [x] Fits CONTRIBUTING.md. Add a :+1: reaction to pull requests you find important. @Aleksanaa Done.
2025-04-01T06:37:19.236570
2024-10-05T23:41:29
2568369348
{ "authors": [ "otavio", "r-ryantm" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2122", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/346755" }
gharchive/pull-request
renode: 1.15.2 -> 1.15.3 Automatic update generated by nixpkgs-update tools. This update was made based on information from passthru.updateScript. meta.description for renode is: Virtual development framework for complex embedded systems meta.homepage for renode is: https://renode.io Updates performed Ran passthru.UpdateScript To inspect upstream changes Release on GitHub Compare changes on GitHub Impact Checks done built on NixOS The tests defined in passthru.tests, if any, passed found 1.15.3 in filename of file in /nix/store/dicby0bcpxqa3ws0mpbkz5jnz2w9rz8h-renode-1.15.3 Rebuild report (if merged into master) (click to expand) 2 total rebuild path(s) 2 package rebuild(s) First fifty rebuilds by attrpath renode Instructions to test this update (click to expand) Either download from Cachix: nix-store -r /nix/store/dicby0bcpxqa3ws0mpbkz5jnz2w9rz8h-renode-1.15.3 \ --option binary-caches 'https://cache.nixos.org/ https://nix-community.cachix.org/' \ --option trusted-public-keys ' nix-community.cachix.org-1:mB9FSh9qf2dCimDSUo8Zy7bkq5CX+/rkCWyvRCYg3Fs= cache.nixos.org-1:6NCHdD59X431o0gWypbMrAURkbJ16ZPMQFGspcDShjY= ' (The Cachix cache is only trusted for this store-path realization.) For the Cachix download to work, your user must be in the trusted-users list or you can use sudo since root is effectively trusted. Or, build yourself: nix-build -A renode https://github.com/r-ryantm/nixpkgs/archive/12bfac6b125be9b6ed323e9f6268566384cb2217.tar.gz Or: nix build github:r-ryantm/nixpkgs/12bfac6b125be9b6ed323e9f6268566384cb2217#renode After you've downloaded or built it, look at the files and if there are any, run the binaries: ls -la /nix/store/dicby0bcpxqa3ws0mpbkz5jnz2w9rz8h-renode-1.15.3 ls -la /nix/store/dicby0bcpxqa3ws0mpbkz5jnz2w9rz8h-renode-1.15.3/bin Pre-merge build results We have automatically built all packages that will get rebuilt due to this change. This gives evidence on whether the upgrade will break dependent packages. Note sometimes packages show up as failed to build independent of the change, simply because they are already broken on the target branch. Result of nixpkgs-review run on x86_64-linux 1 1 package built: renode Maintainer pings cc @otavio for testing. [!TIP] As a maintainer, if your package is located under pkgs/by-name/*, you can comment @NixOS/nixpkgs-merge-bot merge to automatically merge this update using the nixpkgs-merge-bot. Add a :+1: reaction to pull requests you find important. @NixOS/nixpkgs-merge-bot merge
2025-04-01T06:37:19.249690
2024-10-30T05:43:17
2622978151
{ "authors": [ "NickCao", "r-ryantm" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2123", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/352284" }
gharchive/pull-request
python312Packages.python-docs-theme: 2024.6 -> 2024.10 Automatic update generated by nixpkgs-update tools. This update was made based on information from passthru.updateScript. meta.description for python312Packages.python-docs-theme is: Sphinx theme for CPython project meta.homepage for python312Packages.python-docs-theme is: https://github.com/python/python-docs-theme meta.changelog for python312Packages.python-docs-theme is: https://github.com/python/python-docs-theme/blob/2024.10/CHANGELOG.rst Updates performed Ran passthru.UpdateScript To inspect upstream changes Impact Checks done built on NixOS The tests defined in passthru.tests, if any, passed found 2024.10 with grep in /nix/store/rs2qirl06p9myvkvz5z6m4zlzz83dids-python3.12-python-docs-theme-2024.10 found 2024.10 in filename of file in /nix/store/rs2qirl06p9myvkvz5z6m4zlzz83dids-python3.12-python-docs-theme-2024.10 Rebuild report (if merged into master) (click to expand) 3 total rebuild path(s) 3 package rebuild(s) First fifty rebuilds by attrpath python311Packages.python-docs-theme python312Packages.python-docs-theme Instructions to test this update (click to expand) Either download from Cachix: nix-store -r /nix/store/rs2qirl06p9myvkvz5z6m4zlzz83dids-python3.12-python-docs-theme-2024.10 \ --option binary-caches 'https://cache.nixos.org/ https://nix-community.cachix.org/' \ --option trusted-public-keys ' nix-community.cachix.org-1:mB9FSh9qf2dCimDSUo8Zy7bkq5CX+/rkCWyvRCYg3Fs= cache.nixos.org-1:6NCHdD59X431o0gWypbMrAURkbJ16ZPMQFGspcDShjY= ' (The Cachix cache is only trusted for this store-path realization.) For the Cachix download to work, your user must be in the trusted-users list or you can use sudo since root is effectively trusted. Or, build yourself: nix-build -A python312Packages.python-docs-theme https://github.com/r-ryantm/nixpkgs/archive/edece1aa0db54b5c144d7f7cc92a7b970026e91e.tar.gz Or: nix build github:r-ryantm/nixpkgs/edece1aa0db54b5c144d7f7cc92a7b970026e91e#python312Packages.python-docs-theme After you've downloaded or built it, look at the files and if there are any, run the binaries: ls -la /nix/store/rs2qirl06p9myvkvz5z6m4zlzz83dids-python3.12-python-docs-theme-2024.10 ls -la /nix/store/rs2qirl06p9myvkvz5z6m4zlzz83dids-python3.12-python-docs-theme-2024.10/bin Pre-merge build results We have automatically built all packages that will get rebuilt due to this change. This gives evidence on whether the upgrade will break dependent packages. Note sometimes packages show up as failed to build independent of the change, simply because they are already broken on the target branch. nixpkgs-review result Generated using nixpkgs-review. Command: nixpkgs-review x86_64-linux :white_check_mark: 4 packages built: python311Packages.python-docs-theme python311Packages.python-docs-theme.dist python312Packages.python-docs-theme python312Packages.python-docs-theme.dist Maintainer pings cc @KAction for testing. [!TIP] As a maintainer, if your package is located under pkgs/by-name/*, you can comment @NixOS/nixpkgs-merge-bot merge to automatically merge this update using the nixpkgs-merge-bot. Add a :+1: reaction to pull requests you find important. nixpkgs-review result Generated using nixpkgs-review. Command: nixpkgs-review pr 352284 x86_64-linux :white_check_mark: 4 packages built: python311Packages.python-docs-theme python311Packages.python-docs-theme.dist python312Packages.python-docs-theme python312Packages.python-docs-theme.dist
2025-04-01T06:37:19.283632
2018-07-08T11:50:16
339219592
{ "authors": [ "GrahamcOfBorg", "Infinisil", "mpickering", "nlewo", "xeji" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2124", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/43218" }
gharchive/pull-request
whitebox-tools: init at 0.9.0 Motivation for this change Things done [ ] Tested using sandboxing (nix.useSandbox on NixOS, or option sandbox in nix.conf on non-NixOS) Built on platform(s) [x] NixOS [ ] macOS [ ] other Linux distributions [ ] Tested via one or more NixOS test(s) if existing and applicable for the change (look inside nixos/tests) [ ] Tested compilation of all pkgs that depend on this change using nix-shell -p nox --run "nox-review wip" [x] Tested execution of all binary files (usually in ./result/bin/) [ ] Determined the impact on package closure size (by running nix path-info -S before and after) [ ] Fits CONTRIBUTING.md. @GrahamcOfBorg build whitebox-tools Failure on aarch64-linux (full log) Attempted: whitebox-tools Partial log (click to expand) cannot build derivation '/nix/store/1hgh2pg0gis0g29dkcc5aj5gh2a03ci8-rust_serde_derive-1.0.18.drv': 6 dependencies couldn't be built cannot build derivation '/nix/store/132j7mfl04zb7vlhdj9a6kgm6i3668bj-rust_crates-io-0.11.0.drv': 30 dependencies couldn't be built cannot build derivation '/nix/store/37lmq2kc9fwscasgjvwr5s7wpjgjd636-rust_docopt-0.8.1.drv': 19 dependencies couldn't be built cannot build derivation '/nix/store/6i9xgpvlp3i943v09n91p4yc97vhhcvc-rust_git2-curl-0.7.0.drv': 20 dependencies couldn't be built cannot build derivation '/nix/store/hdsnr2qc7ac19rnb9mrl3hj7w3107gzh-rust_ignore-0.2.2.drv': 18 dependencies couldn't be built cannot build derivation '/nix/store/sbg4sjdjg534ij7g4ki9vj41s5kxiisq-rust_cargo-0.22.0.drv': 78 dependencies couldn't be built cannot build derivation '/nix/store/vdaccxg8m8yjxk4vhv2m54ygykmhfabs-rust_cargo-vendor-0.1.13.drv': 79 dependencies couldn't be built cannot build derivation '/nix/store/szv1gqq935xrhsb9bsj76i268qg779fa-whitebox_tools-0.9.0-vendor.drv': 2 dependencies couldn't be built cannot build derivation '/nix/store/i03ar26jrqv7snss735w73wlcr3wpm4q-whitebox_tools-0.9.0.drv': 3 dependencies couldn't be built error: build of '/nix/store/i03ar26jrqv7snss735w73wlcr3wpm4q-whitebox_tools-0.9.0.drv' failed Failure on x86_64-linux (full log) Attempted: whitebox-tools Partial log (click to expand) cannot build derivation '/nix/store/gk7jma27balgcjhkm4r40zrb64b39hn5-rust_serde_derive-1.0.18.drv': 6 dependencies couldn't be built cannot build derivation '/nix/store/cb42b4wsk5531gfvw7nagwhk100dlh8z-rust_crates-io-0.11.0.drv': 30 dependencies couldn't be built cannot build derivation '/nix/store/sly01qb04j3i03qi3m61905v8v5c7lvb-rust_docopt-0.8.1.drv': 19 dependencies couldn't be built cannot build derivation '/nix/store/w9zzb5fdr7pb1afv7j2lc3ls2nz0yg26-rust_git2-curl-0.7.0.drv': 20 dependencies couldn't be built cannot build derivation '/nix/store/dz3j5lf77pd0pjv4jy971z8nnmp4qxhw-rust_ignore-0.2.2.drv': 18 dependencies couldn't be built cannot build derivation '/nix/store/53xkf9vyx0b7ipw0b383jr5dsz82bhjd-rust_cargo-0.22.0.drv': 78 dependencies couldn't be built cannot build derivation '/nix/store/yjvsqlbilv24ychxkd52qylf7qcqk93n-rust_cargo-vendor-0.1.13.drv': 79 dependencies couldn't be built cannot build derivation '/nix/store/4c1vdq45pl0cfx73m08pmm35if3cgyag-whitebox_tools-0.9.0-vendor.drv': 2 dependencies couldn't be built cannot build derivation '/nix/store/22mqd97n017cjr5p9nyiivr09vkzmp93-whitebox_tools-0.9.0.drv': 3 dependencies couldn't be built error: build of '/nix/store/22mqd97n017cjr5p9nyiivr09vkzmp93-whitebox_tools-0.9.0.drv' failed Looks like rustc-1.27.0 is broken. Success on x86_64-linux (full log) Attempted: whitebox-tools Partial log (click to expand) installing post-installation fixup shrinking RPATHs of ELF executables and libraries in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0 shrinking /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0/bin/whitebox_tools strip is /nix/store/4qvrxzxa535y8304mk195x50b6p9607d-binutils-2.30/bin/strip stripping (with command strip and flags -S) in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0/bin patching script interpreter paths in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0 checking for references to /build in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0... /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0 Failure on x86_64-darwin (full log) Attempted: whitebox-tools Partial log (click to expand) clang-5.0: error: linker command failed with exit code 1 (use -v to see invocation) error: aborting due to previous error error: Could not compile `whitebox_tools`. To learn more, run the command again with --verbose. builder for '/nix/store/z75nbfh8m4v0gazbzbja0w1nk3i71rc9-whitebox_tools-0.9.0.drv' failed with exit code 101 error: build of '/nix/store/z75nbfh8m4v0gazbzbja0w1nk3i71rc9-whitebox_tools-0.9.0.drv' failed Success on aarch64-linux (full log) Attempted: whitebox-tools Partial log (click to expand) installing post-installation fixup shrinking RPATHs of ELF executables and libraries in /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0 shrinking /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0/bin/whitebox_tools strip is /nix/store/0pjsgkxz0rp5baycq5sp2s72lrr5q9sg-binutils-2.30/bin/strip stripping (with command strip and flags -S) in /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0/bin patching script interpreter paths in /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0 checking for references to /build in /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0... /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0 I'll fix this on darwin. @GrahamcOfBorg build whitebox-tools Success on aarch64-linux (full log) Attempted: whitebox-tools Partial log (click to expand) /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0 Success on x86_64-linux (full log) Attempted: whitebox-tools Partial log (click to expand) installing post-installation fixup shrinking RPATHs of ELF executables and libraries in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0 shrinking /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0/bin/whitebox_tools strip is /nix/store/4qvrxzxa535y8304mk195x50b6p9607d-binutils-2.30/bin/strip stripping (with command strip and flags -S) in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0/bin patching script interpreter paths in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0 checking for references to /build in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0... /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0 to check darwin build: @GrahamcOfBorg build whitebox-tools Success on x86_64-linux (full log) Attempted: whitebox-tools Partial log (click to expand) installing post-installation fixup shrinking RPATHs of ELF executables and libraries in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0 shrinking /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0/bin/whitebox_tools strip is /nix/store/4qvrxzxa535y8304mk195x50b6p9607d-binutils-2.30/bin/strip stripping (with command strip and flags -S) in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0/bin patching script interpreter paths in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0 checking for references to /tmp/nix-build-whitebox_tools-0.9.0.drv-0 in /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0... /nix/store/bqyzzfda84cfamsqm57pcxd0a4yqa8f1-whitebox_tools-0.9.0 Success on x86_64-darwin (full log) Attempted: whitebox-tools Partial log (click to expand) running 0 tests test result: ok. 0 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out installing post-installation fixup strip is /nix/store/7ddbq63v97nk8gkbf7gcsfmby37h6gbl-cctools-binutils-darwin/bin/strip stripping (with command strip and flags -S) in /nix/store/yr4b8lpfjixkh97nn8a73qhrj3z5q10n-whitebox_tools-0.9.0/bin patching script interpreter paths in /nix/store/yr4b8lpfjixkh97nn8a73qhrj3z5q10n-whitebox_tools-0.9.0 /nix/store/yr4b8lpfjixkh97nn8a73qhrj3z5q10n-whitebox_tools-0.9.0 Success on aarch64-linux (full log) Attempted: whitebox-tools Partial log (click to expand) installing post-installation fixup shrinking RPATHs of ELF executables and libraries in /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0 shrinking /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0/bin/whitebox_tools strip is /nix/store/0pjsgkxz0rp5baycq5sp2s72lrr5q9sg-binutils-2.30/bin/strip stripping (with command strip and flags -S) in /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0/bin patching script interpreter paths in /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0 checking for references to /build in /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0... /nix/store/1q5xyrymspw126pzhlc8jzrnr5lq8d3i-whitebox_tools-0.9.0
2025-04-01T06:37:19.292699
2018-10-05T13:51:59
367221175
{ "authors": [ "GrahamcOfBorg", "timokau" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2125", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/47915" }
gharchive/pull-request
zn_poly: 0.9 -> 0.9.1 Motivation for this change Sage has taken over maintenance since the original author no longer maintains the project. The upgrade integrates various patches from sage. It also enables tuning by default, which now has to be disabled explicitly. Things done [ ] Tested using sandboxing (nix.useSandbox on NixOS, or option sandbox in nix.conf on non-NixOS) Built on platform(s) [ ] NixOS [ ] macOS [ ] other Linux distributions [ ] Tested via one or more NixOS test(s) if existing and applicable for the change (look inside nixos/tests) [ ] Tested compilation of all pkgs that depend on this change using nix-shell -p nox --run "nox-review wip" [ ] Tested execution of all binary files (usually in ./result/bin/) [ ] Determined the impact on package closure size (by running nix path-info -S before and after) [ ] Fits CONTRIBUTING.md. Success on aarch64-linux (full log) Attempted: zn_poly Partial log (click to expand) post-installation fixup shrinking RPATHs of ELF executables and libraries in /nix/store/ff4ya6fy78g5fzzym185sfk8fg3wr8kp-zn_poly-0.9.1 shrinking /nix/store/ff4ya6fy78g5fzzym185sfk8fg3wr8kp-zn_poly-0.9.1/lib/libzn_poly.so shrinking /nix/store/ff4ya6fy78g5fzzym185sfk8fg3wr8kp-zn_poly-0.9.1/lib/libzn_poly-0.9.so shrinking /nix/store/ff4ya6fy78g5fzzym185sfk8fg3wr8kp-zn_poly-0.9.1/lib/libzn_poly-0.9.1.so strip is /nix/store/428gs2z4b8f9byvghzlpbjwjb3a7jwxx-binutils-2.30/bin/strip stripping (with command strip and flags -S) in /nix/store/ff4ya6fy78g5fzzym185sfk8fg3wr8kp-zn_poly-0.9.1/lib patching script interpreter paths in /nix/store/ff4ya6fy78g5fzzym185sfk8fg3wr8kp-zn_poly-0.9.1 checking for references to /build in /nix/store/ff4ya6fy78g5fzzym185sfk8fg3wr8kp-zn_poly-0.9.1... /nix/store/ff4ya6fy78g5fzzym185sfk8fg3wr8kp-zn_poly-0.9.1 Success on x86_64-darwin (full log) Attempted: zn_poly Partial log (click to expand) zn_array_mul_fft_dft()... ok zn_array_invert()... ok All tests passed. installing post-installation fixup strip is /nix/store/9xjkb4xz0b5lmizij9ppxy7lkxdxhx6b-cctools-binutils-darwin/bin/strip stripping (with command strip and flags -S) in /nix/store/5rns1nmh7zkkljdsax34s6714q17wdkk-zn_poly-0.9.1/lib patching script interpreter paths in /nix/store/5rns1nmh7zkkljdsax34s6714q17wdkk-zn_poly-0.9.1 /nix/store/5rns1nmh7zkkljdsax34s6714q17wdkk-zn_poly-0.9.1 Success on x86_64-linux (full log) Attempted: zn_poly Partial log (click to expand) post-installation fixup shrinking RPATHs of ELF executables and libraries in /nix/store/85iam4zzihnkkrjrlanal487ykv03xa5-zn_poly-0.9.1 shrinking /nix/store/85iam4zzihnkkrjrlanal487ykv03xa5-zn_poly-0.9.1/lib/libzn_poly-0.9.1.so shrinking /nix/store/85iam4zzihnkkrjrlanal487ykv03xa5-zn_poly-0.9.1/lib/libzn_poly-0.9.so shrinking /nix/store/85iam4zzihnkkrjrlanal487ykv03xa5-zn_poly-0.9.1/lib/libzn_poly.so strip is /nix/store/dxf1m7dhc4qb655bdljc1fsd74v1nag3-binutils-2.30/bin/strip stripping (with command strip and flags -S) in /nix/store/85iam4zzihnkkrjrlanal487ykv03xa5-zn_poly-0.9.1/lib patching script interpreter paths in /nix/store/85iam4zzihnkkrjrlanal487ykv03xa5-zn_poly-0.9.1 checking for references to /build in /nix/store/85iam4zzihnkkrjrlanal487ykv03xa5-zn_poly-0.9.1... /nix/store/85iam4zzihnkkrjrlanal487ykv03xa5-zn_poly-0.9.1
2025-04-01T06:37:19.318548
2018-10-09T22:01:52
368416151
{ "authors": [ "ElvishJerricco", "GrahamcOfBorg", "andir", "c0bw3b", "copumpkin", "florianjacob", "xeji" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2126", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/48128" }
gharchive/pull-request
gitea: 1.5.1 -> 1.5.2 Motivation for this change https://blog.gitea.io/2018/10/gitea-1.5.2-is-released/ β€œThis is a smaller release, with only 12 merged PRs, but it contains several security related fixes and so we recommend upgrading whenever possible.” Should therefore be safe to backport to 18.09, and required to get the security fixes. Things done [ ] Tested using sandboxing (nix.useSandbox on NixOS, or option sandbox in nix.conf on non-NixOS) Built on platform(s) [x] NixOS [ ] macOS [ ] other Linux distributions [ ] Tested via one or more NixOS test(s) if existing and applicable for the change (look inside nixos/tests) [ ] Tested compilation of all pkgs that depend on this change using nix-shell -p nox --run "nox-review wip" [x] Tested execution of all binary files (usually in ./result/bin/) [ ] Determined the impact on package closure size (by running nix path-info -S before and after) [X] Fits CONTRIBUTING.md. Maintainer: @disassembler Failure on aarch64-linux (full log) Attempted: gitea Partial log (click to expand) code.gitea.io/gitea/models code.gitea.io/gitea/modules/notification code.gitea.io/gitea/modules/templates code.gitea.io/gitea/modules/auth code.gitea.io/gitea/modules/context code.gitea.io/gitea/modules/lfs code.gitea.io/gitea/routers/repo code.gitea.io/gitea/routers/user builder for '/nix/store/rz4y8f8pppi5k194gj7k89cskiapy8hf-gitea-1.5.2.drv' failed with exit code 47 error: build of '/nix/store/rz4y8f8pppi5k194gj7k89cskiapy8hf-gitea-1.5.2.drv' failed Unexpected error: command failed with exit code 1 on x86_64-darwin (full log) Attempted: gitea Partial log (click to expand) trying https://github.com/go-gitea/gitea/archive/v1.5.2.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 122 0 122 0 0 203 0 --:--:-- --:--:-- --:--:-- 202 100 17.9M 0 17.9M 0 0 4643k 0 --:--:-- 0:00:03 --:--:-- 6856k unpacking source archive /private/tmp/nix-build-source.drv-0/v1.5.2.tar.gz fixed-output derivation produced path '/nix/store/4gydlwklm9v77jk8zj3n03f7csmkq18z-source' with sha256 hash '1dr7pxb60y215625nndmisxl6n5zlq1aicaswsg3xfzlk7a93k7k' instead of the expected hash '049f4gs1lw14l0ybz6rx6sibwpqdihqrx1qgvs6ghy6asrc2blif' cannot build derivation '/nix/store/jsm9n5l3qsr8n7filwzmcbxvkvpgf3xa-gitea-1.5.2.drv': 1 dependencies couldn't be built error: build of '/nix/store/jsm9n5l3qsr8n7filwzmcbxvkvpgf3xa-gitea-1.5.2.drv' failed aarch64 failure looks like a builder issue, we can try again later. Please fix the hash mismatch on darwin. Success on x86_64-linux (full log) Attempted: gitea Partial log (click to expand) checking for references to /build in /nix/store/fz05imgv890j71v5jp8bc8j54wqvx9is-gitea-1.5.2-bin... shrinking RPATHs of ELF executables and libraries in /nix/store/hypb926rcw4dw6df3pf9pshrj54mnxvy-gitea-1.5.2 strip is /nix/store/dxf1m7dhc4qb655bdljc1fsd74v1nag3-binutils-2.30/bin/strip patching script interpreter paths in /nix/store/hypb926rcw4dw6df3pf9pshrj54mnxvy-gitea-1.5.2 checking for references to /build in /nix/store/hypb926rcw4dw6df3pf9pshrj54mnxvy-gitea-1.5.2... shrinking RPATHs of ELF executables and libraries in /nix/store/mj23a6lwgj105m2z58rh6llrn3mjymh8-gitea-1.5.2-data strip is /nix/store/dxf1m7dhc4qb655bdljc1fsd74v1nag3-binutils-2.30/bin/strip patching script interpreter paths in /nix/store/mj23a6lwgj105m2z58rh6llrn3mjymh8-gitea-1.5.2-data checking for references to /build in /nix/store/mj23a6lwgj105m2z58rh6llrn3mjymh8-gitea-1.5.2-data... /nix/store/fz05imgv890j71v5jp8bc8j54wqvx9is-gitea-1.5.2-bin @GrahamcOfBorg build gitea Unexpected error: command failed with exit code 1 on x86_64-darwin (full log) Attempted: gitea Partial log (click to expand) trying https://github.com/go-gitea/gitea/archive/v1.5.2.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 122 0 122 0 0 652 0 --:--:-- --:--:-- --:--:-- 655 100 17.9M 0 17.9M 0 0 5791k 0 --:--:-- 0:00:03 --:--:-- 6654k unpacking source archive /private/tmp/nix-build-source.drv-0/v1.5.2.tar.gz fixed-output derivation produced path '/nix/store/4gydlwklm9v77jk8zj3n03f7csmkq18z-source' with sha256 hash '1dr7pxb60y215625nndmisxl6n5zlq1aicaswsg3xfzlk7a93k7k' instead of the expected hash '049f4gs1lw14l0ybz6rx6sibwpqdihqrx1qgvs6ghy6asrc2blif' cannot build derivation '/nix/store/jsm9n5l3qsr8n7filwzmcbxvkvpgf3xa-gitea-1.5.2.drv': 1 dependencies couldn't be built error: build of '/nix/store/jsm9n5l3qsr8n7filwzmcbxvkvpgf3xa-gitea-1.5.2.drv' failed Success on x86_64-linux (full log) Attempted: gitea Partial log (click to expand) /nix/store/fz05imgv890j71v5jp8bc8j54wqvx9is-gitea-1.5.2-bin Failure on aarch64-linux (full log) Attempted: gitea Partial log (click to expand) code.gitea.io/gitea/routers/api/v1/repo code.gitea.io/gitea/routers/api/v1/user code.gitea.io/gitea/routers code.gitea.io/gitea/routers/api/v1/org code.gitea.io/gitea/routers/api/v1/admin code.gitea.io/gitea/routers/admin code.gitea.io/gitea/routers/api/v1 code.gitea.io/gitea/routers/routes builder for '/nix/store/rz4y8f8pppi5k194gj7k89cskiapy8hf-gitea-1.5.2.drv' failed with exit code 45 error: build of '/nix/store/rz4y8f8pppi5k194gj7k89cskiapy8hf-gitea-1.5.2.drv' failed builds fine on aarch64 when issued with --option cores 2. Probably some limitation with the default go routine settings. I vaguely recall some default thread limits. @NixOS/darwin-maintainers Any idea why fetchFromGitHub expects a different hash on darwin? @ElvishJerricco posted the following on IRC: $ nix-prefetch-url --unpack https://github.com/go-gitea/gitea/archive/v1.5.2.tar.gz unpacking... path is '/nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz' 049f4gs1lw14l0ybz6rx6sibwpqdihqrx1qgvs6ghy6asrc2blif $ ssh my-mac nix-prefetch-url --unpack https://github.com/go-gitea/gitea/archive/v1.5.2.tar.gz unpacking... path is '/nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz' 1dr7pxb60y215625nndmisxl6n5zlq1aicaswsg3xfzlk7a93k7k 

``` My first guess is that it's due to macOS's case sensitive file system, but I haven't confirmed that the repo actually has any collisions. Doesn't appear to have obvious case collisions on this macOS, so I'm not sure why it fails. Furthermore, I'm on macOS, and: $ nix-prefetch-url --unpack https://github.com/go-gitea/gitea/archive/v1.5.2.tar.gz unpacking... path is '/nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz' 049f4gs1lw14l0ybz6rx6sibwpqdihqrx1qgvs6ghy6asrc2blif I'm not running a case-sensitive filesystem so I'm puzzled. @ElvishJerricco I can't get the other .tar.gz, but perhaps you can unpack and diff the two trees on your linux box? @copumpkin: $ linux=/nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz $ mac=/nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz $ diff -r $linux $mac Only in /nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz/integrations/gitea-repositories-meta/user2/utf8.git/refs/heads: ブランチ Only in /nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz/integrations/gitea-repositories-meta/user2/utf8.git/refs/heads: ブランチ Only in /nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz/integrations/gitea-repositories-meta/user2/utf8.git/refs/heads: Grüßen Only in /nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz/integrations/gitea-repositories-meta/user2/utf8.git/refs/heads: Grüßen Only in /nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz/integrations/gitea-repositories-meta/user2/utf8.git/refs/tags: γ‚Ώγ‚° Only in /nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz/integrations/gitea-repositories-meta/user2/utf8.git/refs/tags: γ‚Ώγ‚―γ‚™ Only in /nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz/integrations/gitea-repositories-meta/user2/utf8.git/refs/tags: Π•Μˆ Only in /nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz/integrations/gitea-repositories-meta/user2/utf8.git/refs/tags: Ё diff: /nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz/vendor/github.com/Unknwon/cae/tz/testdata/testdir.lnk: No such file or directory diff: /nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz/vendor/github.com/Unknwon/cae/tz/testdata/testdir.lnk: No such file or directory diff: /nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz/vendor/github.com/Unknwon/cae/tz/testdata/test.lnk: No such file or directory diff: /nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz/vendor/github.com/Unknwon/cae/tz/testdata/test.lnk: No such file or directory diff: /nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz/vendor/github.com/Unknwon/cae/zip/testdata/testdir.lnk: No such file or directory diff: /nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz/vendor/github.com/Unknwon/cae/zip/testdata/testdir.lnk: No such file or directory diff: /nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz/vendor/github.com/Unknwon/cae/zip/testdata/test.lnk: No such file or directory diff: /nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz/vendor/github.com/Unknwon/cae/zip/testdata/test.lnk: No such file or directory diff: /nix/store/n3bg2vhxv8yfc631la9gfykglg3f55b3-v1.5.2.tar.gz/vendor/gopkg.in/macaron.v1/fixtures/symlink: No such file or directory diff: /nix/store/jq8p6y5zjzhaipsb5wg6bqziw7i45q7g-v1.5.2.tar.gz/vendor/gopkg.in/macaron.v1/fixtures/symlink: No such file or directory Maybe something unicode related? I'm on sierra, if that matters. Ah yes, some of the unicode normalization of combining pairs (whether to store them as two characters or one) behavior can be weird. Not really sure what to do about that, other than to add a postFetch step to our fetchFromGitHub that strips that junk out @copumpkin @ElvishJerricco does any of you have the time to implement test it? I do not have a Mac to test it and never used one. Would probably take me 10x just to figure that out with always going through ofBorg :confused: Here is a previous occurrence of the same issue. If subdirs vendor and integrations are not necessary to the build process, then you can do: src = fetchFromGitHub { owner = "go-gitea"; repo = "gitea"; rev = "v${version}"; sha256 = "049f4gs1lw14l0ybz6rx6sibwpqdihqrx1qgvs6ghy6asrc2blif"; extraPostFetch = "rm -rf $out/integrations && rm -rf $out/vendor"; }; @c0bw3b you'll need to update the sha256 there. Yes indeed that hash needs to be re-computed. I edited the code block. Thank you all for taking a look at this, I was really lost at how it could possibly be the case that MacOS computes a different checksum than on Linux. As one could expect from their names, I just verified that integrations contains integration tests and is not required for build, while vendor contains various third-party library dependencies and is indeed required, and needs selective manual / automatic filtering. For further testing / discussion, I pushed a new version with a extraPostFetch that deletes offending files and still seems to work for me. Some kind of automatic mechanism in the fetch function would still help though, so that this won't return on the next release, and nix-prefetch-git gives the correct sha256 sum. This would lead to hard-to-detect problems though if some project uses / starts to use affected unicode regions in file names that are relevant for the build process. Could someone with the appropriate permissions please trigger a @GrahamcOfBorg build gitea? @GrahamcOfBorg build gitea Failure on aarch64-linux (full log) Attempted: gitea Partial log (click to expand) code.gitea.io/gitea/vendor/github.com/Unknwon/cae/zip code.gitea.io/gitea/vendor/github.com/glycerine/go-unsnap-stream code.gitea.io/gitea/vendor/github.com/couchbase/vellum code.gitea.io/gitea/vendor/code.gitea.io/git code.gitea.io/gitea/vendor/github.com/Unknwon/i18n code.gitea.io/gitea/vendor/gopkg.in/editorconfig/editorconfig-core-go.v1 code.gitea.io/gitea/vendor/github.com/go-xorm/xorm code.gitea.io/gitea/vendor/gopkg.in/testfixtures.v2 builder for '/nix/store/8jijwi2w49hsyfcyhb0mgana4hxp6sbd-gitea-1.5.2.drv' failed with exit code 43 error: build of '/nix/store/8jijwi2w49hsyfcyhb0mgana4hxp6sbd-gitea-1.5.2.drv' failed Works fine on my mac. $ nix build -f https://github.com/NixOS/nixpkgs/archive/0a5c7cf512fcefb4b337186f0aee69ae8dd1df2c.tar.gz gitea.src [1 built, 55 copied (111.7 MiB), 37.0 MiB DL] /nix/store/cg1gbwbc68ilw81djc13hp78bhgfsqq1-source Success on x86_64-linux (full log) Attempted: gitea Partial log (click to expand) checking for references to /build in /nix/store/7jk9mp702a9ry6qnv2rc40vhl8pxrzl8-gitea-1.5.2-bin... shrinking RPATHs of ELF executables and libraries in /nix/store/w5h82yifgv8ifcpjm73v288k65n49766-gitea-1.5.2 strip is /nix/store/vcc4svb8gy29g4pam2zja6llkbcwsyiq-binutils-2.30/bin/strip patching script interpreter paths in /nix/store/w5h82yifgv8ifcpjm73v288k65n49766-gitea-1.5.2 checking for references to /build in /nix/store/w5h82yifgv8ifcpjm73v288k65n49766-gitea-1.5.2... shrinking RPATHs of ELF executables and libraries in /nix/store/h51jpsy5mwx25h73qx3jj4vbs6gfll6z-gitea-1.5.2-data strip is /nix/store/vcc4svb8gy29g4pam2zja6llkbcwsyiq-binutils-2.30/bin/strip patching script interpreter paths in /nix/store/h51jpsy5mwx25h73qx3jj4vbs6gfll6z-gitea-1.5.2-data checking for references to /build in /nix/store/h51jpsy5mwx25h73qx3jj4vbs6gfll6z-gitea-1.5.2-data... /nix/store/7jk9mp702a9ry6qnv2rc40vhl8pxrzl8-gitea-1.5.2-bin 18.03: 8ccf5b63b5786fbf1fdda05132df30e78260139b 18.09: 38f5b076d77ce204a237a5dc9238c3586e607707 Thanks! :smiley:
2025-04-01T06:37:19.324564
2019-03-25T21:48:57
425131118
{ "authors": [ "dywedir", "eadwu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2127", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/58296" }
gharchive/pull-request
linux_testing_bcachefs,bcachefs-tools: 2019-03-29 Motivation for this change Things done [x] Tested using sandboxing (nix.useSandbox on NixOS, or option sandbox in nix.conf on non-NixOS) Built on platform(s) [x] NixOS [ ] macOS [ ] other Linux distributions [ ] Tested via one or more NixOS test(s) if existing and applicable for the change (look inside nixos/tests) [ ] Tested compilation of all pkgs that depend on this change using nix-shell -p nix-review --run "nix-review wip" [ ] Tested execution of all binary files (usually in ./result/bin/) [ ] Determined the impact on package closure size (by running nix path-info -S before and after) [ ] Assured whether relevant documentation is up to date [x] Fits CONTRIBUTING.md. At least bcachefs-tools is broken on 2019-03-29. Thank you!
2025-04-01T06:37:19.327358
2015-02-22T19:37:47
58515055
{ "authors": [ "sfultong" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2128", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/6523" }
gharchive/pull-request
updating bitcoin and changing source location I propose we pull the source from bitcoin.org, since the sha256 sums that are GPG signed match the source packages on bitcoin.org, but not the packages on github. See: https://bitcoin.org/bin/bitcoin-core-0.10.0/SHA256SUMS.asc https://bitcoin.org/bin/bitcoin-core-0.10.0/bitcoin-0.10.0.tar.gz https://github.com/bitcoin/bitcoin/archive/v0.10.0.tar.gz Paging @roconnor @AndersonTorres since you are listed as maintainers.
2025-04-01T06:37:19.336109
2019-10-14T20:36:20
506861692
{ "authors": [ "FRidh", "francescocarzaniga", "teh" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2129", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/71141" }
gharchive/pull-request
dask: 2.2.0 -> 2.5.2, distributed: 2.4.0 -> 2.5.2, dask-ml: 0.11.0 -> 1.0.0, dask-jobqueue: 0.4.1 -> 0.7.0 Motivation for this change Dask and relative packages were quite outdated. Things done [ ] Tested using sandboxing (nix.useSandbox on NixOS, or option sandbox in nix.conf on non-NixOS) Built on platform(s) [ ] NixOS [ ] macOS [x] other Linux distributions [ ] Tested via one or more NixOS test(s) if existing and applicable for the change (look inside nixos/tests) [ ] Tested compilation of all pkgs that depend on this change using nix-shell -p nix-review --run "nix-review wip" [x] Tested execution of all binary files (usually in ./result/bin/) [ ] Determined the impact on package closure size (by running nix path-info -S before and after) [ ] Ensured that relevant documentation is up to date [ ] Fits CONTRIBUTING.md. Notify maintainers cc @FRidh @costrouc Thanks for the changes! Would you mind splitting out each update to its own commit? That makes it easier for tooling that uses the commit format, and also to selectively back out if something goes wrong. @teh No problem, there are now four commits. apologies, one more commit style change suggestion: distributed: 2.4.0->2.5.2 to distributed: 2.4.0 -> 2.5.2 (check the git commit log of nixpkgs for other examples). LGTM otherwise, thanks! @GrahamcOfBorg build python3.pkgs.dask python3.pkgs.distributed python3.pkgs.dask-ml python3.pkgs.dask-jobqueue apologies, one more commit style change suggestion: distributed: 2.4.0->2.5.2 to distributed: 2.4.0 -> 2.5.2 (check the git commit log of nixpkgs for other examples). LGTM otherwise, thanks! Even better, prefix with the package set so it is an actual attribute path. That way the bot can pick it up. So what do I have to do to get these packages updated? It's been ten days and I see that dask has already been replaced in master. @francescocarzaniga I think roughly: resolve conflict prefix packages e.g. pythonPackages.dask-ml ping @FRidh to merge Nothing happening so closing.
2025-04-01T06:37:19.341424
2019-10-19T00:58:33
509367825
{ "authors": [ "groodt", "jonringer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2130", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/71372" }
gharchive/pull-request
kubecfg: 0.12.5 -> 0.13.1 Motivation for this change Things done [ ] Tested using sandboxing (nix.useSandbox on NixOS, or option sandbox in nix.conf on non-NixOS) Built on platform(s) [ ] NixOS [x] macOS [ ] other Linux distributions [ ] Tested via one or more NixOS test(s) if existing and applicable for the change (look inside nixos/tests) [ ] Tested compilation of all pkgs that depend on this change using nix-shell -p nix-review --run "nix-review wip" [x] Tested execution of all binary files (usually in ./result/bin/) [ ] Determined the impact on package closure size (by running nix path-info -S before and after) [x] Ensured that relevant documentation is up to date [x] Fits CONTRIBUTING.md. Notify maintainers cc @ @GrahamcOfBorg build kubecfg
2025-04-01T06:37:19.382705
2020-03-16T19:03:41
582536852
{ "authors": [ "Infinisil", "dali99", "marsam", "pasqui23", "roberth", "vcunat" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2131", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/82743" }
gharchive/pull-request
Freeform modules Motivation for this change This implement freeform modules, which are like a combination between types.submodule and types.attrsOf: The options declared with the submodule are type-checked, while the others are combined using a freeform type. This has been a desired feature in multiple occasions now: This will simplify NixOS/rfcs#42 drastically, making the process of declaring options corresponding to settings values very conventional. Type checking for nixpkgs' config argument, as previously attempted in https://github.com/NixOS/nixpkgs/pull/57123 and https://github.com/NixOS/nixpkgs/pull/80582, will now be possible in a sane way. How it looks like (updated for latest version) (evaluate with nix-instantiate --eval --strict example.nix -A config) with import ./lib; evalModules { modules = [{ options.settings = mkOption { default = {}; type = with types; submodule { # All values that don't have an associated option have to be of this type freeformType = attrsOf int; options.port = mkOption { type = port; default = 80; description = "The port to use."; }; }; }; # Works, even though foo isn't declared as an option config.settings.foo = 10; # error: The option value `settings.port' in `example.nix' is not of type # `16 bit unsigned integer; between 0 and 65535 (both inclusive)'. #config.settings.port = "not-a-port"; ## Overrides default value config.settings.port = 8080; }]; } Ping @roberth @Ma27 @rycee @Profpatsch @oxij TODO [x] Write tests [x] Potentially separate orthogonal changes into a separate PR [x] Write docs [x] Check how this affects performance -> The current implementation is slower than it could be when many freeform values are defined, but since the number of them per module evaluation is expected to be rather low, I'll leave it at this for now. I now separated the orthogonal changes into https://github.com/NixOS/nixpkgs/pull/82751 I don't think attrValueType is necessary if you merge in one go at the root of the submodule and then recursiveUpdate the unstructured values back into config. This puts less burden on the types and, if we don't hardcode recursiveUpdate, it allows the module author to keep the unstructured attributes separate if desired. @roberth I don't think this would work without big changes, because currently it's not known upfront which values are unstructured and which aren't, and this could get complicated with nested submodules and attrsOf and such. @Infinisil mergeModules could return the option values and unstructured attrs separately. Doing so should also fix the laziness problem caused by generating fake option names for the unstructured definitions. Quote for when the commit is lost: TODO: This leads to infinite recursion in NixOS if it's always turned on, so some use cases won't be supported anymore because of unstructured attrs. Figure out what they are. One feature I suspect may be lost due to this change is the ability to check for the existence of option declarations before defining them, which I think will be important when modules move into flakes and need to adapt a bit at runtime to their now variable environment. I've used this to optionally define NixOps deployment keys in a NixOS-compatible module. I'll make sure we have a test for this laziness property. @roberth I think I see what you mean, this might be possible, I'll give it a try soon. Implementation would be a bit like: Do a recursive mergeModules' for declared options first. After that, determine which definitions haven't been matched with a declaration, and merge those with a single unstructured type. Then combine the results of the mergeModules' and the unstructured merge with something like recursiveUpdate (it will need to insert the fake option declarations). I don't think it needs to be configurable though, as it will essentially only allow unstructured values under option sets, which makes a lot of sense, and is also the behavior of the current implementation. @roberth Done that! Commits need some cleanup, and other things are still left to do, but now the code is looking much nicer. I also did some very sweet cleanups related to this. Also I'm now not generating fake options anymore: Only .config will now contain unstructured values, .options doesn't, which should then also not get in the way of what you mentioned in https://github.com/NixOS/nixpkgs/pull/82802. Unfortunately without attrValueType, a specific error message is significantly worse: With something like { config._module.unstructuredType = attrsOf int; options.a.b = mkOption {}; config.a.c = "foo"; } it now throws The option value `a' in `example.nix' is not of type `signed integer'. and if you change it to { config.a = 0; } you get (as expected) error: The option path `a' is an attribute set of options, but it is defined to not be an attribute set in `example.nix'. Did you define its value at the correct and complete path? Instead of the previous (code now in https://github.com/Infinisil/nixpkgs/tree/partially-typed-v2-old) The option `a.c' defined in `example.nix' does not exist and the unstructured type `signed integer' of `a' does not support attribute sets. I could get back the original error message if I add attrValueType again, but I'm not sure if it's worth it. Maybe that's better left for the future. error message This can probably be improved quite a bit by rethrowing the error with an extra message. "The module at ${prefix} allows undeclared definitions to be mixed with known options, but a definition did not satisfy the type for undeclared definitions: ${caughtMsg}. This may also be the result of mistyping an otherwise known option." @roberth Hm it does work with builtins.tryEval and a deepSeq on the unstructuredType.merge ..., but that's not very nice. I'd have to do some changes to how types throw errors to make it work without tryEval. I think I'll have to get the attrValueType thing back after all. Not only does it allow for much better error messages in some cases, but after doing a performance check, I also found that the whole definition coercion thing is really slow due to the many list operations. @roberth I think let's not do this attrValueType thing after all, for now. The current implementation here works well and has clean and simple code, the only problem is that it can be a bit slow if there are many unstructured attributes, but that will be a rare occurence. If this does becomes a problem in the future, it can still be changed then, though ideally this would also bring some module system improvements along with it so attrValueType can be implemented nicely (which it currently can't). So I think the only thing left to do here is to write some docs. Also I think the naming is a bit off, it's not "partially typed modules", because everything is still type checked, there's just a fallback type now for undeclared options. "unstructured attributes" is also a bit unfitting. Maybe the feature could be called "Fallback-typing" instead. I have now written some initial docs for this feature. Once the docs are finalized and reviewed, I think this is ready to be merged. The docs read like you're describing a change, like you would in an rfc or a commit message Yeah that's what I noticed too, will definitely try to improve on that Another use would be to use packages as submodules,like what I would like to do in https://github.com/NixOS/nixpkgs/pull/89943 @roberth What do you think would be a better term for this feature? Maybe one of: A fallback option type Declarationless definitions Optionless definitions Freeform definitions Freeform configuration Freeform modules The last one sounds pretty good imo Freeform modules The last one sounds pretty good imo :100: I'll wait until https://github.com/NixOS/nixpkgs/pull/75584 is merged so I can extend the "Options for Program Settings" section with this feature. Now updated to term this feature freeform modules instead and revised the documentation to be less RFC-y. Also fixed a bug with freeform modules. Once the documentation is reviewed, I think this PR is ready for merging. New section under "Writing modules": New section under the settings option docs (introduced in #75584): @roberth Thanks for all the suggestions, I implemented them now. After some thought, I think there's no reason not to introduce some syntactic sugar for freeformType, so I did that now. Was fairly easy to add even. The freeform type doesn't seem to be visible in the generated option docs. Could we generate a line for an "option" like "${somePrefix}.*" there? I think we've discussed this in an earlier pr, but I couldn't find it. We may want to use a freeform "option" rather than a freeform type for this reason, because that provides an opportunity to provide the description string and an example. @roberth With https://github.com/NixOS/nixpkgs/pull/82743/commits/2d45a62899d47c109a0b8ce4ca9d33265b8a1a37 the freeform type is shown as the submodules type. E.g. with a config like { lib, ... }: { options.fofofofo = lib.mkOption { type = lib.types.submodule { freeformType = lib.types.attrsOf lib.types.str; options.babababa = lib.mkOption {}; }; }; config = { documentation.nixos.includeAllModules = true; fileSystems."/".device = "ntest"; boot.loader.grub.device = "nodev"; }; } Building with nix-build '<nixpkgs/nixos>' --arg configuration ./config.nix -A config.system.build.manual.manualHTML gives Awesome! Thanks for all the reviewing and merging @roberth! This broke nix-darwin, bisected to https://github.com/NixOS/nixpkgs/commit/fd75dc876586bde8cdb683a6952a41132e8db166 $ darwin-rebuild build --dry-run building the system configuration... error: The option `nixpkgs.localSystem' defined in `<unknown-file>' does not exist. (use '--show-trace' to show detailed location information) full trace $ darwin-rebuild build --dry-run --show-trace building the system configuration... error: while evaluating the attribute 'activationScript' of the derivation 'darwin-system-20.09.git.78922ca+darwin3.47fe6db' at /Users/marsam/code/nix-config/darwin/modules/system/default.nix:74:7: while evaluating the attribute 'activationScripts.script.text' at undefined position: while evaluating 'g' at /Users/marsam/code/nix-config/nixpkgs/lib/attrsets.nix:276:19, called from undefined position: while evaluating anonymous function at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:98:72, called from /Users/marsam/code/nix-config/nixpkgs/lib/attrsets.nix:279:20: while evaluating the attribute 'value' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:450:9: while evaluating the option `system.activationScripts.script.text': while evaluating the attribute 'mergedValue' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:482:5: while evaluating the attribute 'values' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:476:9: while evaluating the attribute 'values' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:575:7: while evaluating anonymous function at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:462:28, called from /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:462:17: while evaluating definitions from `/Users/marsam/code/nix-config/darwin/modules/system/activation-scripts.nix': while evaluating 'dischargeProperties' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:534:25, called from /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:463:137: while evaluating the attribute 'value' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:327:44: while evaluating the attribute 'activationScripts.applications.text' at undefined position: while evaluating 'g' at /Users/marsam/code/nix-config/nixpkgs/lib/attrsets.nix:276:19, called from undefined position: while evaluating anonymous function at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:98:72, called from /Users/marsam/code/nix-config/nixpkgs/lib/attrsets.nix:279:20: while evaluating the attribute 'value' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:450:9: while evaluating the option `system.activationScripts.applications.text': while evaluating the attribute 'mergedValue' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:482:5: while evaluating the attribute 'values' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:476:9: while evaluating the attribute 'values' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:575:7: while evaluating anonymous function at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:462:28, called from /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:462:17: while evaluating definitions from `/Users/marsam/code/nix-config/darwin/modules/system/applications.nix': while evaluating 'dischargeProperties' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:534:25, called from /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:463:137: while evaluating the attribute 'value' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:327:44: while evaluating the attribute 'passAsFile' of the derivation 'system-applications' at /Users/marsam/code/nix-config/nixpkgs/pkgs/build-support/trivial-builders.nix:7:7: while evaluating the attribute 'buildCommand' of the derivation 'darwin-manpages' at /Users/marsam/code/nix-config/nixpkgs/pkgs/build-support/trivial-builders.nix:7:7: while evaluating the attribute 'buildCommand' of the derivation 'manual-olinkdb' at /Users/marsam/code/nix-config/nixpkgs/pkgs/build-support/trivial-builders.nix:7:7: while evaluating the attribute 'buildCommand' of the derivation 'darwin-manual-combined' at /Users/marsam/code/nix-config/nixpkgs/pkgs/build-support/trivial-builders.nix:7:7: while evaluating the attribute 'buildCommand' of the derivation 'generated-docbook' at /Users/marsam/code/nix-config/nixpkgs/pkgs/build-support/trivial-builders.nix:7:7: while evaluating the attribute 'buildCommand' of the derivation 'options-db.xml' at /Users/marsam/code/nix-config/nixpkgs/pkgs/build-support/trivial-builders.nix:7:7: while evaluating 'flip' at /Users/marsam/code/nix-config/nixpkgs/lib/trivial.nix:138:16, called from /Users/marsam/code/nix-config/darwin/doc/manual/default.nix:48:21: while evaluating 'optionAttrSetToDocList'' at /Users/marsam/code/nix-config/nixpkgs/lib/options.nix:140:37, called from /Users/marsam/code/nix-config/darwin/doc/manual/default.nix:9:72: while evaluating 'collect' at /Users/marsam/code/nix-config/nixpkgs/lib/attrsets.nix:178:19, called from /Users/marsam/code/nix-config/nixpkgs/lib/options.nix:162:67: while evaluating 'isType' at /Users/marsam/code/nix-config/nixpkgs/lib/types.nix:14:18, called from /Users/marsam/code/nix-config/nixpkgs/lib/attrsets.nix:179:8: while evaluating 'evalModules' at /Users/marsam/code/nix-config/nixpkgs/lib/modules.nix:21:17, called from /Users/marsam/code/nix-config/darwin/modules/documentation/default.nix:19:24: The option `nixpkgs.localSystem' defined in `<unknown-file>' does not exist. The only place nix-darwin uses nixpkgs.localSystem is in the manual https://github.com/LnL7/nix-darwin/blob/47fe6db9c9cd22c4ba57df3281f98f454880bea7/modules/documentation/default.nix#L20. I'm not familiar with modules internals, so I'm appreciate any hint to for a proper fix cc: @LnL7 @marsam I can't find a declaration for localSystem in nix-darwin, so the error message seems to be correct. Grepping only reveals the usage site you linked but no declaration. The file analogous to Nixpkgs where I would expect it to be defined is https://github.com/LnL7/nix-darwin/blob/master/modules/nix/nixpkgs.nix I don't think nix-darwin needs that option, so I've created a PR for the simple solution https://github.com/LnL7/nix-darwin/pull/216 NixOS-specific rant: If a module needs access to cross-compilation info, it should generally be taken from pkgs directly; either pkgs.stdenv.hostPlatform, which is the what the eventual machine runs or if you really do need to care about where the config is built: pkgs.stdenv.buildPlatform. Generally you care about pkgs.stdenv.hostPlatform which can not be assumed to be equivalent to config.nixpkgs.localSystem because of config.nixpkgs.crossSystem. local and cross only make sense in the context of nixos-rebuild --target-host, which uses incompatible terminology... TL;DR: I assume we'll revert this merge until the issues get resolved? This merge also broke Hydra's evaluation of trunk-combined. It's some kind of interaction with other conditions and I haven't managed to reproduce it locally so far though @samueldr did (maybe this just triggered some Nix bug)... but the merge commit of this PR (6d0a85fe52) won't evaluate on Hydra whereas the parent commit succeeded. The message is error: [json.exception.type_error.302] type must be string, but is null and it seems likely that it comes from evaluating nixos/release-combined.nix -A nixos.manual.x86_64-linux. Some context was on IRC yesterday. @roberth Nice detective work! Both the nix-darwin and the hydra failures seem to originate from https://github.com/NixOS/nixpkgs/commit/fd75dc876586bde8cdb683a6952a41132e8db166, which apparently checks options stricter, but I'm not quite sure yet why (I'll investigate this a bit more) Fyi, the hydra failure originates from this command now failing: $ nix-instantiate nixos/tests/boot.nix -A biosCdrom error: The option `virtualisation.qemu.consoles' defined in `/home/infinisil/src/nixpkgs/nixos/modules/testing/test-instrumentation.nix' does not exist. Which is indeed fixed by @roberth's PR. Similar for nix-darwin. TL;DR: I assume we'll revert this merge until the issues get resolved? Shouldn't be necessary. This change is merely exposing undeclared options that weren't discovered by the previous implementation. I've merged a fix for the one that was blocking release-small. trunk-combined should also be fixed now, because it doesn't seem to suffer from other errors of the same time. We'll see when evaluation is done. (I'm wondering, shouldn't TL;DR be a summary?) Ah figured it out: The previous implementation only checked option existence for config paths that are evaluated, while the new one checks all of them unconditionally. Here's an example that illustrates this: (import <nixpkgs/lib>).evalModules { modules = [({ lib, ... }: { options.root = lib.mkOption {}; options.foo.bar = lib.mkOption {}; config.root = "root"; config.foo.qux = "qux"; })]; } Evaluating with nix-instantiate --eval -A config.root previously worked because config.foo wasn't evaluated, so it would still fail with nix-instantiate --eval -A config.foo. With the new implementation, both commands fail. I think this change in behavior is justified, because there can be other reasons an option doesn't exist other than the config path not being evaluated (e.g. misspellings). In fact I think this change might even fix https://github.com/NixOS/nixpkgs/issues/36590. It looks like previously this was (ab)used to make evaluation not fail: https://github.com/NixOS/nixpkgs/blob/aeffd67cec5b76e327f244c7464a4254113836cc/nixos/modules/testing/test-instrumentation.nix#L11-L18 Thanks, that was really fast. (I'm wondering, shouldn't TL;DR be a summary?) I did consider that as something like summary/conclusion, explaining the details below, though I agree I typically see it used in a bit different way. @Infinisil I think @LnL7 has a good point here https://github.com/LnL7/nix-darwin/pull/216#issuecomment-674966068 that this is an api change. Even if the old api is undocumented and untested, it may have an important role in modular code bases. Do you think it's feasible to warn about the old behavior or to emulate it? If only warning about it is feasible, we can recommend to extract the option into a separate file, which may be a good idea. I don't think trying to emulate the old behavior is a good idea, because from what we've seen already, that behavior was very prone to mistakes (https://github.com/NixOS/nixpkgs/issues/36590 and https://github.com/LnL7/nix-darwin/pull/216). While https://github.com/NixOS/nixpkgs/pull/95584 wasn't a mistake per se, it still relied on a hack to make it work. I think it would be a good idea to mention this change in behavior though, probably in the breaking changes section in the release notes. There doesn't seem to be any way to define typing for an optional setting, I was looking to set default = unset in some way, since some configuration handle null and "not set" differently. @dali99 Indeed, that's something that could be improved in the future. See also https://github.com/NixOS/nixpkgs/pull/63553 which is related.
2025-04-01T06:37:19.390840
2020-03-27T21:24:13
589420525
{ "authors": [ "LnL7", "teto" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2132", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/83532" }
gharchive/pull-request
Revert "luaPackages.luv: 1.30.0-0 -> 1.34.1-1" This reverts commit eaa47284113cda19a037b7fcbaa892ddb596b813. With 81461cff5f540c92e5030f62b89ee7b64e85c6df luv also needs to be downgraded until luarocks can be updated to 3.3.1 again. Fixes #82727 Motivation for this change Things done [x] Tested using sandboxing (nix.useSandbox on NixOS, or option sandbox in nix.conf on non-NixOS linux) Built on platform(s) [x] NixOS [x] macOS [ ] other Linux distributions [ ] Tested via one or more NixOS test(s) if existing and applicable for the change (look inside nixos/tests) [ ] Tested compilation of all pkgs that depend on this change using nix-shell -p nixpkgs-review --run "nixpkgs-review wip" [x] Tested execution of all binary files (usually in ./result/bin/) [ ] Determined the impact on package closure size (by running nix path-info -S before and after) [ ] Ensured that relevant documentation is up to date [x] Fits CONTRIBUTING.md. On the luv issue, as we have no proper fix really (the current fix is just as complicated as the current one), I am fine with this. I have not understood yet if the issue on darwin was random or systematic (i.e., if it could be fixed by trying several times). For reference the darwin build worked fine with luarocks-3.3.1 + luv-1.34.1-1 (even without your changes). So once #80528 is resolved we just have to make sure to update both.
2025-04-01T06:37:19.397102
2020-06-11T00:46:40
636638482
{ "authors": [ "jonringer", "siriobalmelli" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2133", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/90108" }
gharchive/pull-request
pyspinel: fix python3 build Motivation for this change Update to reference an upstream fix to build on python >=3.3 Things done [ ] Tested using sandboxing (nix.useSandbox on NixOS, or option sandbox in nix.conf on non-NixOS linux) Built on platform(s) [ ] NixOS [x] macOS [x] other Linux distributions [ ] Tested via one or more NixOS test(s) if existing and applicable for the change (look inside nixos/tests) [x] Tested compilation of all pkgs that depend on this change using nix-shell -p nixpkgs-review --run "nixpkgs-review wip" [ ] Tested execution of all binary files (usually in ./result/bin/) [x] Determined the impact on package closure size (by running nix path-info -S before and after) [x] Ensured that relevant documentation is up to date [x] Fits CONTRIBUTING.md. sorry for all the round about trips. To comply with CONTRIBUTING.md please have the commit message name be of the format <pkg-name>: <subject-line> for more examples, please look at https://github.com/NixOS/nixpkgs/blob/master/.github/CONTRIBUTING.md#submitting-changes in your case, the commit message should be: python3Packages.pyspinel: fix build
2025-04-01T06:37:19.412288
2020-07-01T17:38:42
649156631
{ "authors": [ "datafoo", "flokli", "mweinelt" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2134", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/91960" }
gharchive/pull-request
nixos/networkd: update options Motivation for this change To fix missing option DNSDefaultRoute= (see #91761) and other missing options. Things done I have not tested all options that I touched. [ ] Tested using NixOps Built on platform(s) [x] NixOS [ ] macOS [ ] other Linux distributions [ ] Tested using sandboxing (nix.useSandbox on NixOS, or option sandbox in nix.conf on non-NixOS linux) Built on platform(s) [ ] NixOS [ ] macOS [ ] other Linux distributions [ ] Tested via one or more NixOS test(s) if existing and applicable for the change (look inside nixos/tests) [ ] Tested compilation of all pkgs that depend on this change using nix-shell -p nixpkgs-review --run "nixpkgs-review wip" [ ] Tested execution of all binary files (usually in ./result/bin/) [ ] Determined the impact on package closure size (by running nix path-info -S before and after) [ ] Ensured that relevant documentation is up to date [ ] Fits CONTRIBUTING.md. I will update the PR following this comment. This needs another rebase. This needs another rebase. Done This is the second time I rebased my PR on master after more recent PRs than mine were merged. What guides the order in which PRs are merged? Thanks! This is the second time I rebased my PR on master after more recent PRs than mine were merged. What guides the order in which PRs are merged? There's no specific order, but this PR is quite big, which is why it's harder to review. Networking-Tests are looking good. /nix/store/4y8k9isgglvhycvq59kg50hx62rw08kn-vm-test-run-Bond-Networking-Networkd /nix/store/0kmcb3mwlm63hjkbd50h075d61ncqf0i-vm-test-run-Bridge-Networking-Networkd /nix/store/dcz18v4qz82sx9ynr3wyqimzrmaigpgg-vm-test-run-OneInterfaceDHCP-Networking-Networkd /nix/store/1g9p0kgpd8nzbv6kqkkcq08nh8hmvmxi-vm-test-run-SimpleDHCP-Networking-Networkd /nix/store/3014l21l3qrg01s3x33ys08fgjqrs9lw-vm-test-run-Link-Networking-Networkd /nix/store/jllmqcvalzn1imsb6x0r57dfcdw2v7cz-vm-test-run-Loopback-Networking-Networkd /nix/store/431izfl3mb82crr6kaw143ky0r04c9yl-vm-test-run-MACVLAN-Networking-Networkd /nix/store/p7vj2zfcl228a9p5ri27ycrdm1yix6v9-vm-test-run-Privacy-Networking-Networkd /nix/store/fs1rk6x92h83gfm1f8pm4ps8158g17ph-vm-test-run-routes-Networking-Networkd /nix/store/wvh4cha8ssx3ypplwqp75wf1phiy46yx-vm-test-run-Sit-Networking-Networkd /nix/store/blbh606jdzsbgdadbavj8nl0ic00c4d0-vm-test-run-Static-Networking-Networkd /nix/store/mgv2b09h658glqn8lgz2s3akxdwzf7d7-vm-test-run-Virtual-Networking-Networkd /nix/store/8kfj6r2kh6ab2w5g6ix304fr21zlw2vz-vm-test-run-vlan-Networking-Networkd I tried reviewing this, but it's really hard to do. e9d13d3 added some commented-out checks due to Nix' <= 2.2 integer size, and 70407f0 commented them back in. Could this be squashed together, so we only add the checks in a single commit? As much as I could try, each of these 2 commits do one logical thing and one thing only, as described by the commit messages e9d13d3: update options for systemd 245. That commit synchronizes the code with the man page of networkd and, for homogeneity with the rest of the code, keeps avoiding the use of assertRange with 64bits integers. 70407f0: use assertRange with 64bits integers. This commit takes into account this comment. Perhaps it would better fit as a separate PR after this one would have been merged. Since the PR is about cleaning up the networkd module, I included this commit here. I apologize if my commit messages were not clear enough. Also, e9d13d3 does a bit too much at once. For example, it removes OriginalName. Is this intended? Again, that commit synchronizes the code with the man page of networkd. The best way to review that I would suggest is to read the man page. For the specifics of OriginalName, refer to man systemd.link. I also try to double check by contacting the author here, without success. However, you kind of confirmed it yourself here. To facilitate our work, let's discuss specific code issues/questions with the review feature of Github. Could you revise the commits, and make it these changes easier to review? Maybe something like this also deserves a separate commit. I do not think it is beneficial because the commit does one clear logical thing. If you disagree, please describe what you mean by "these changes" because, as you may have noted, this is not the only thing I removed in this commit e9d13d3. Thanks again for your patience! I'll try to react quicker on further updates. Thank you as well for reviewing this. Could we move on with this PR? I gave this another close look, and while it is pretty hard to review due to the amount of changes, this seems to be fine, and should make things more readable and understandable in the longer run. Thanks for the work you put in this, and your patience with me :-)
2025-04-01T06:37:19.421741
2015-08-27T02:02:26
103405721
{ "authors": [ "andrewrk", "cillianderoiste", "wkennington" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2135", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/9467" }
gharchive/pull-request
jack2: update from b5bceb to 460063 An especially important commit that this includes is https://github.com/jackaudio/jack2/commit/05216197b491d64036733a2d3be511456c1f65cd which changes the API to match jack1 and the official JACK specification. I tested this. Travis failure is due to a network failure. @andrewrk this does look like a nice improvement, but personally I'd rather wait for the next release. Does it fix any particular issue for you at the moment? @wkennington why did you upgrade jack2 to 460063d8d from the v1.9.10 release (you didn't provide any reason in your commit message)? Do we need to have a separate unstable version of jack2 in addition to the stable release, or do we really need to use git master as the default? @andrewrk this does look like a nice improvement, but personally I'd rather wait for the next release. Does it fix any particular issue for you at the moment? Yes. This commit modifies the project I am working on to be compliant against jack1 and the official JACK specification. However it causes my project to fail to build from source when compiling against jack2Full. This nixpkgs update allows me to compile my project against jack1, jack2, and be compliant with the spec, at the same time. Furthermore, the previous version of the package was not an official JACK release so I see no reason to prefer one arbitrary SHA1 over another. @andrewrk ah, so if we added an expression for jack2-git, that would be OK for your development work (or if you override it locally). I'd be happy with that, but I'm also curious to hear why we're not using the release at the moment. ah, so if we added an expression for jack2-git, that would be OK for your development work (or if you override it locally) It's not so bad working from my nixpkgs fork for now. But I'm also collaborating with people from other distros whose package managers have updated to include this fix. For example, debian: https://packages.debian.org/source/sid/jackd2 @goibhniu I usually only use git snapshots if the build is fubar for one of the releases or there is no release at all. I don't remember now what the issue was but there was something wrong with the stable version at the time. On Thu, Aug 27, 2015, 01:11 Andrew Kelley<EMAIL_ADDRESS>wrote: ah, so if we added an expression for jack2-git, that would be OK for your development work (or if you override it locally) It's not so bad working from my nixpkgs fork for now. But I'm also collaborating with people from other distros whose package managers have updated to include this fix. For example, debian: https://packages.debian.org/source/sid/jackd2 β€” Reply to this email directly or view it on GitHub https://github.com/NixOS/nixpkgs/pull/9467#issuecomment-135335582. @wkennington thanks. If nobody objects, I'd like to try to get the released version working again and use that as the default and then we could also add jack2-git for @andrewrk and others, since jack2 stable releases seem to only happen every year or two. I'll add a separate ticket for it but if you want to have something sooner @andrewrk, I'd be happy to merge a PR with a separate expression for jack2-git.
2025-04-01T06:37:19.428600
2020-09-23T16:20:32
707505654
{ "authors": [ "jonringer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2136", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/98573" }
gharchive/pull-request
[20.09] orangefs: 2.9.7 -> 2.9.8, fix build Motivation for this change backport #98525 ZHF: #97479 (cherry picked from commit aa25c6576611ebab7d7448fa30edad9eb1c1a066) Things done [x] Tested using sandboxing (nix.useSandbox on NixOS, or option sandbox in nix.conf on non-NixOS linux) Built on platform(s) [x] NixOS [ ] macOS [ ] other Linux distributions [ ] Tested via one or more NixOS test(s) if existing and applicable for the change (look inside nixos/tests) [x] Tested compilation of all pkgs that depend on this change using nix-shell -p nixpkgs-review --run "nixpkgs-review wip" [ ] Tested execution of all binary files (usually in ./result/bin/) [x] Determined the impact on package closure size (by running nix path-info -S before and after) [ ] Ensured that relevant documentation is up to date [x] Fits CONTRIBUTING.md. @GrahamcOfBorg test orangefs tests passed
2025-04-01T06:37:19.435045
2015-09-20T16:39:51
107403578
{ "authors": [ "edwtjo", "khumba", "matthiasbeyer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2137", "repo": "NixOS/nixpkgs", "url": "https://github.com/NixOS/nixpkgs/pull/9962" }
gharchive/pull-request
gitit: Remove staticDir and templatesDir as these cannot be overridden if specified here. My attempt to fix #9863. I don't know whether this is a step in the right direction or not. Ping to all haskellers I can find, because I have no clue who is responsible for the service file CC: @gebner @gridaphobe @jb55 @khumba @psibi @ttuegel (Disclaimer, I don't know about gitit.) I see that there are config options defined for those, and that the options are used in the generated config file and the systemd service, but strangely, they're hard-coded. It conflicts if you try to set it without your patch, right? If so, looks fine to me, but I'd recommend setting default attributes for staticDir and templatesDir to what they were previously fixed to. @khumba Good idea! As the author of the module I would like to have these options available, which I'm using in a private module. For nixpkgs, it was commit https://github.com/NixOS/nixpkgs/commit/99750d89dd962cc7011b02e487ed1a38c7842524 that essentially made them useless. @khumba that would require evaluation of the haskellPackages set. I would recommend setting them null by default and changing the type to nullOr. Also checking for null in the config inclusion. @edwtjo Sorry I do not understand your comment. Shall I set these options to null by default? What's funny: As far as I can see, the path for the static and template dir is hard coded to the store itself in the service definition. And there is no way to override a service, is there? So this package is rather useless as-is. @matthiasbeyer yes that is what I meant, but also modify configFile. Anyway these alterations will just be ugly hacks, let's avoid those on second thought. And there is no way to override a service, is there? So this package is rather useless as-is. I'm not exactly sure what you mean but no 99750d89dd962cc7011b02e487ed1a38c7842524 was not a very good idea. So since @edolstra don't want the computational and storage overhead with the module; The only correct way to move forward is to remove gitit.nix from the module list, reverting 99750d89dd962cc7011b02e487ed1a38c7842524 and instead requiring it to be manually added into your configuration.nix. Which was done in 122d5e497e6df55f4b2e5c353bb2224b79cfd7b0 and aeba3eabcdbfa6308d9af27b2a70e7864ec4d3e3
2025-04-01T06:37:19.468669
2020-04-07T11:28:55
595797653
{ "authors": [ "Noah2610" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2141", "repo": "Noah2610/deathfloor", "url": "https://github.com/Noah2610/deathfloor/issues/32" }
gharchive/issue
Add Knockback component(s) Similar to DealsDamage and TakesDamage. Knockback and KnockbackAffected. For now, we just configure the entity configs to have knockback behavior. Knockback components may be nice for a knockback state in the future, but has low priority for now. Could also help with #41
2025-04-01T06:37:19.696176
2017-06-24T12:59:33
238316018
{ "authors": [ "Muchiachio", "thewinner55" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2142", "repo": "NonFactors/MVC5.Grid", "url": "https://github.com/NonFactors/MVC5.Grid/issues/93" }
gharchive/issue
reload grid and clear Filter var deferred = $('.mvc-grid').mvcgrid({ requestType: 'post', // defaults to get reload: true, data: { ClearParameter: ClearFilter, SBU: $(".ddlSBU").val(), SegmentCode: $(".ddlSegment").val() }, reloadEnded: function (grid) { //$("#mvc-grid-AjaxSearch").val("param=value"); //ClearFilter = false; } }); What's your question? What did you try and what doesn't work? i want clear filter and sort after reload grid Clear it from the UI or server side? If you are using your own filter data to query the models it should already be reset. You can also disable them through grid configuration .Sortable(false) and .Filterable(false). Did you managed to clear it?
2025-04-01T06:37:19.700679
2021-12-21T17:16:08
1086048998
{ "authors": [ "bazelajoanna", "evulins", "franiakat" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2143", "repo": "Nontrivial-Things/thinking_classrooms_client", "url": "https://github.com/Nontrivial-Things/thinking_classrooms_client/pull/29" }
gharchive/pull-request
Fix Integration tests for Problem Index functionality Fix integration tests for Problem Index page. @evulins where now we should component inputWithTags? What do you mean? Where should we test it? :) What I meant is that I wrote two tests in the previous test file for the pseudo inputWithTags component. And because of our mechanism structure they didn't work properly. Right now they are missing :D And I wonder if it's on purpose and if yes should we test it somewhere else? @evulins oh, sure, I just forgot, I will paste them here, as because of the mechanics the also belong here. :) πŸ‘Œ Nice!
2025-04-01T06:37:19.731735
2016-06-19T11:49:39
161069186
{ "authors": [ "doaa-khaled", "essiembre" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2144", "repo": "Norconex/collector-http", "url": "https://github.com/Norconex/collector-http/issues/263" }
gharchive/issue
error when load config file in java I get that error when trying to load the config file `SEVERE: Servlet.service() for servlet [tasks] in context with path [/Crawler_NORCQNEX] threw exception [Servlet execution threw an exception] with root cause java.lang.NoClassDefFoundError: com/norconex/collector/core/CollectorConfigLoader at servlets.Tasks.doPost(Tasks.java:64) at javax.servlet.http.HttpServlet.service(HttpServlet.java:650) at javax.servlet.http.HttpServlet.service(HttpServlet.java:731) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:303) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208) at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:220) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:122) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:505) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:169) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:103) at org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:956) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:116) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:436) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1078) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:625) at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:318) at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) at java.lang.Thread.run(Unknown Source) ` and that is the code of it HttpCollectorConfig config = (HttpCollectorConfig) new CollectorConfigLoader(HttpCollectorConfig.class).loadCollectorConfig(configFile); HttpCollector collector = new HttpCollector(config); I make sure that jar of this class is imported!! Do you have that issue when you run it standalone? It seems your webapp could not find that class so it looks like a classpath/classloading issue with your webapp. As a side note, you may want to know the HTTP Collector is meant to be run in isolation/standalone (live and die with the JVM). The crawling/parsing of some files can be quite resource intensive sometimes and running it in a webapp may compromize the webapp stability and will likely not work as you expect. If you really want to run it in a web app, I recommend you launch the Collector as an external process. You are of course welcome to experiment. :-) Thank you for your reply, but what do you mean is that I can't use the library inside my code ? Of course not! You can use the library with your code. I was simply sharing what type of usage works best to save you trouble. The problem you have (NoClassDefFoundError) is likely not one of the HTTP Collector, but rather one of the library not being found by your Java app. You'll have to troubleshoot why your webapp (Tomcat) does not load some Jars/classes (assumming you've included all dependencies in your web app). It works now! I added these jars to deployment classpath.. thanks a lot
2025-04-01T06:37:19.741150
2019-05-30T05:49:07
450141780
{ "authors": [ "Wikiwide", "marcusramberg" ], "license": "Artistic-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2145", "repo": "Nordaaker/convos", "url": "https://github.com/Nordaaker/convos/pull/383" }
gharchive/pull-request
window.document.hasFocus instead of window.hasFocus? It could potentially be a setting. But I think that it's better to show notification, if possible. Because user may be at a different tab in the same window?.. Seems sane.
2025-04-01T06:37:19.800430
2024-03-07T11:23:08
2173632191
{ "authors": [ "MagicalWitchAshley", "NotNite" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2146", "repo": "NotNite/skindl", "url": "https://github.com/NotNite/skindl/issues/1" }
gharchive/issue
Error Decoding Response Body Not entirely sure what this is or what the issue is caused by, but Moruke's skin causes skindl to fail: Test case: https://gamebanana.com/mods/481891 Can only reproduce on the first file (moruke_omura_red_.zip), which returns a broken response from the GameBanana API: The other download links (which contain ID 481891 instead of 481890) work correctly, and properly contain the first file: This is a GameBanana bug, nothing I can do to fix it. Select one of the other 1-Click Install buttons to download the mod.
2025-04-01T06:37:19.810284
2021-11-25T20:29:00
1063904855
{ "authors": [ "dwertheimer", "jgclark", "m1well" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2147", "repo": "NotePlan/plugins", "url": "https://github.com/NotePlan/plugins/pull/122" }
gharchive/pull-request
dailyjournal idea question improvements i had an idea to improve the dailyreview questions. also I talked to @jgclark a few days ago about the ideas. here is now my PR with little improvements of the review questions. now you can add subheading, which aren't rendered as commandbar questions but then with ### in the final note additionally you can create bulletpoints just read my README adjustments - I hope you understand what I did :D i pushed the PR rework commit. you can hardly see this here in the GitHub conversation This sounds interesting, but will be a little hard for users to follow. Maybe you can make a short video of how this piece works and put that in the README? On Thu, Nov 25, 2021 at 1:23 PM, Michael Wellner < @.*** > wrote: i pushed the PR rework commit. you can hardly see this here in the GitHub conversation β€” You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub ( https://github.com/NotePlan/plugins/pull/122#issuecomment-979474432 ) , or unsubscribe ( https://github.com/notifications/unsubscribe-auth/ACEI6VEUR4OH3GVPZQ44HXLUN2SNXANCNFSM5IZIVZBQ ). This sounds interesting, but will be a little hard for users to follow. Maybe you can make a short video of how this piece works and put that in the README? … hm ok how can I create this video? never done that before :D I've tidied up the README a little more, removed the bold around the mood answer, and did some other changes ready for changes to the framework that you're probably not aware of yet. I will now publish this update. You can still add bold around the mood output in your personal settings, but I don't want to force it on everyone else. Look at Loom. It’s free. Here’s a @jgclark video. But yours could be super short and focus on how u use these new additions. https://youtu.be/IWjr0dy3JhM https://youtu.be/IWjr0dy3JhM?t=115 On Thu, Nov 25 2021 at 1:52 PM, Michael Wellner @.***> wrote: This sounds interesting, but will be a little hard for users to follow. Maybe you can make a short video of how this piece works and put that in the README? … <#> hm ok how can I create this video? never done that before :D β€” You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/NotePlan/plugins/pull/122#issuecomment-979483399, or unsubscribe https://github.com/notifications/unsubscribe-auth/ACEI6VBSFWYZTEV7NTGXVPDUN2VYDANCNFSM5IZIVZBQ .
2025-04-01T06:37:19.818869
2024-05-22T19:02:13
2311281117
{ "authors": [ "aaronpoweruser", "dwertheimer", "jgclark" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2148", "repo": "NotePlan/plugins", "url": "https://github.com/NotePlan/plugins/pull/572" }
gharchive/pull-request
Unlinked note finder Searches through notes and find unlinked references to other note plan notes. TODO Add an action log note. @aaronpoweruser This is great stuff. Thanks Aaron! Question...what would you think about doing this as a PR against the np.Tidy plugin which has a command: /List stubs: creates a note that lists all your notes that have wikilinks that lead nowhere. Or is there a reason why you think this needs to be a free-standing plugin? No strong preference as to where this lives. I was unsure if it should be in tidy, note helpers or its own plugin. I'm going to keep iterating on the code here, once I'm happy with it I'll move it to wherever makes the most sense. @aaronpoweruser I'm not yet sure what "unlinked references" are. But, yes, I agree that it sounds like a good candidate for NoteHelpers or Tidy. The former is more for things run on a particular note; the latter does things that run over all notes, or all recent notes. Note helpers it is I'm thinking about removing the large global search after switching to regex parsing the performance unacceptable it takes tens of minutes to scan the entire note database. So the plug-in will only work on note at a time. Once I finish up the code, I'll clean the documentation. The plug-in essentially scans are a given note and finds keywords that should link to other notes that are unlinked. It then creates a [[]] link to the found note. My use cases is using voice dictation and having to go back in manually create a [[]] link to another note. this just automated the process. I'll add on save trigger as well, so it should be fully automated I'm happy with this now. Feel free to take a look at it. I still need to add good public documentation and move to note helpers give me a few days @jgclark Can you review please so I can merge and release. @aaronpoweruser I'm confused. I've had a notification that you've done something new with this PR, but I can't see anything that's changed. Please advise. Ive just been reponding to comments. new PS incoming. Been using the new regex for the past week, with voice dictation and a on save tigger. On my notes my runtime has been sub 15ms consistently (Np writing the note to disk takes ~5ms). Its fast enough where I rarely tag notes by hand now. I still need to record a new demo gif, but i'm happy with the code. Old regex logic: New regex: Was a worthwhile effort, i'm finally comfortable with regexs. I'm releasing 0.19.2 with some bug fixes. We can release 0.20.0 once you have the documentation updates for this.
2025-04-01T06:37:19.854620
2018-02-27T19:52:29
300769663
{ "authors": [ "joelverhagen" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2149", "repo": "NuGet/NuGetGallery", "url": "https://github.com/NuGet/NuGetGallery/issues/5551" }
gharchive/issue
In build.ps1 scripts that fetch from ServerCommon, set SSL protocols A lot of machines default to SSL 3.0 and TLS 1.0, which is not compatible with GitHub. System.Net.WebException: The request was aborted: Could not create SSL/TLS secure channel. at Microsoft.PowerShell.Commands.WebRequestPSCmdlet.GetResponse(WebRequest request) at Microsoft.PowerShell.Commands.WebRequestPSCmdlet.ProcessRecord() Workaround is to run this on your machine: Set-ItemProperty -Path 'HKLM:\SOFTWARE\Wow6432Node\Microsoft\.NetFramework\v4.0.30319' -Name 'SchUseStrongCrypto' -Value '1' -Type DWord Set-ItemProperty -Path 'HKLM:\SOFTWARE\Microsoft\.NetFramework\v4.0.30319' -Name 'SchUseStrongCrypto' -Value '1' -Type DWord
2025-04-01T06:37:19.866694
2024-11-20T22:36:30
2677375508
{ "authors": [ "JaynouOliver", "NuayHL" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2150", "repo": "NuayHL/TimeStepGenerating", "url": "https://github.com/NuayHL/TimeStepGenerating/issues/1" }
gharchive/issue
Working with new datasets Does it support working with new Datasets? if I am trying to work with real_dataset = load_dataset("bitmind/AFHQ") fake_dataset = load_dataset("bitmind/AFHQ___RealVisXL_V4.0") How shall I setup the folder structure, a bit confused here Guidance for using custom datasets has been added to the README. You can find it in section "Custom Dataset" under "Download Dataset".
2025-04-01T06:37:19.880277
2023-04-13T01:43:56
1665534326
{ "authors": [ "kayhayen", "kennyhml" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2152", "repo": "Nuitka/Nuitka", "url": "https://github.com/Nuitka/Nuitka/issues/2166" }
gharchive/issue
FATAL: Error, the Windows SDK must be installed in Visual Studio. Hey, so ive seen this issue like twice here, unfortunately reading through neither of them fixed it for me. When I run py -m nuitka --version I get the following: 1.5.4 Commercial: None Python: 3.10.2 (tags/v3.10.2:a58ebcc, Jan 17 2022, 14:12:15) [MSC v.1929 64 bit (AMD64)] Flavor: CPython Official Executable: C:\Users\Admin\AppData\Local\Programs\Python\Python310\python.exe OS: Windows Arch: x86_64 WindowsRelease: 10 Version C compiler: C:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.16.27023\bin\HostX64\x64\cl.exe (cl 14.0). I installed nuitka via pip, python is standart cpython and installed via the official website. I do use a venv to execute the command I use a fair amount of packages in my project, but I dont think thta would be the problem as it compiles alright with the standart compiler, but I would like to use clang I execute py -m nuitka main.py --onefile --plugin-enable=pyside6 --nofollow-import-to=tkinter --windows-uac-admin --clang As you can see, ive tried basically every install: Update: I tried to update nuitka to the most recent version, exact same issue. So your version output shows you are using the community version, and your screenshot is for build tools. @kayhayen could you clarify? And I apologize for sending the build tools, the normal tab basically looks the same though, all SDKs installed So your version output shows you are using the community version. Indeed I am, but the projects description states: Visual Studio 2022 or higher on Windows, older versions will work but only supported for commercial users. (the community editions work just fine) So why is the community version not going to work? Thank you very much for your time! It is working fine. I notice a (2) in your screenshot title. Does that mean you have two of them installed? You might have previously bee using stuff that auto-installs things. I once tried to make Nuitka auto-install MSVC to its private parts with the correct components, but failed to do unfortunately. I think I saw these numbers in that case. You may want to configure the correct one. Oh yea, that (2) is weird! I'll see what I can find about that. Yes, you can waste my time.
2025-04-01T06:37:19.886113
2024-08-09T19:55:09
2458579837
{ "authors": [ "kayhayen", "pveikos" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2153", "repo": "Nuitka/Nuitka", "url": "https://github.com/Nuitka/Nuitka/issues/3058" }
gharchive/issue
Nuitka Module Importing External Numba .so file Nuitka Version 2.4.5, Python version 3.12.0 I am stuck on a question of having a Nuitka .so module import code outside of the module package that is compiled, specifically I want the compiled code inside a nuitka-ized module to reference code in a numba AOT-ized module. Thanks for the help in advance. I am trying to work around using both numba and nuitka for my project. See the project structure below: └── πŸ“nuitka_testing └── run_func.py └── πŸ“.venv └── πŸ“src └── πŸ“nuitka_funcs └── __init__.py └── test_nuitka_imports.py └── numba_modules.cpython-312-x86_64-linux-gnu.so └── __init__.py I have used numba's AOT capabilities to create numba_modules.cpython-312-x86_64-linux-gnu.so which houses all my numba code, which I then import in nuitka_funcs/test_nuitka_imports.py, the content of which is below: ## test_nuitka_imports.py ## from ..numba_modules import identity def run(x): y = identity(x) return y I can run my main function run_func.py (simply from src import run; run(5) ) when the nuitka_funcs folder is not compiled, but when i try to compile the nuitka_funcs folder into a .so using python -m nuitka --module nuitka_funcs --include-package=nuitka_funcs I get an assertion error AssertionError: <ModuleName ''>, crash report is attached: nuitka-crash-report.txt I gather it is because the code being compiled in the nuitka_funcs folder is importing code outside itself? If anyone could please help out on the right direction it'd be most appreciated. I just want to be able to reference outside, non-compiled code from inside a nuitka-ized .so module file, or any work-around for this use case. I have looked through other issues and they are either too vague to help me or don't seem to help me. Thank you! The example looks good to me, I will be trying to add Numba JIT support in the coming releases, so this will be useful as a starting point at least, before we see what larger packages also do.
2025-04-01T06:37:19.889472
2022-12-12T09:33:45
1491417215
{ "authors": [ "joouha", "kayhayen" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2154", "repo": "Nuitka/Nuitka", "url": "https://github.com/Nuitka/Nuitka/pull/1930" }
gharchive/pull-request
User Manual: Escape backslashes in Windows paths What does this PR do? This PR makes a minor correction to the User Manual - the back-slashes in the Windows file paths in the table on use-case 4 were not escaped, so were not visible when the page was rendered. Why was it initiated? Any relevant Issues? It was initiated because I was reading the User Manual and noticed missing back-slashes in the Windows file paths in the table on use-case 4. PR Checklist [x] Correct base branch selected? Should be develop branch. [x] Enabled commit hook or executed ./bin/autoformat-nuitka-source. [x] All tests still pass. Check the Developer Manual about Running the Tests. There are GitHub Actions tests that cover the most important things however, and you are welcome to rely on those, but they might not cover enough. [x] Ideally new features or fixed regressions ought to be covered via new tests. [x] Ideally new or changed features have documentation updates. Thanks for fixing this, very good.
2025-04-01T06:37:19.937594
2021-03-14T03:44:02
831056945
{ "authors": [ "abalmos", "aultac", "awlayton" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2155", "repo": "OADA/oada-srvc-docker", "url": "https://github.com/OADA/oada-srvc-docker/pull/83" }
gharchive/pull-request
Feature cli tokens Improved token admin script in auth and fixed bug in useradd script in users. Also added some readme info about using oadadeploy as an option for installation. It isn’t β€œhanging,” I’m guessing you didn’t pass any args and are probably running it with docker instead of docker-compose and probably forgot to pass β€œ-it” to docker. The script prompts you for anything that wasn’t passed: I.e. if you gave no username, it prompts you for one. If you use docker and forget the β€œ-it”, then you never see the prompt and can’t type the answer so it looks like it is hanging when it’s really just waiting for a username on stdin. Running it with docker-compose doesn’t require that you explicitly allocate a tty, so no extra flag necessary. I want it to operate that way because you don’t have to memorize the flags in order to just add a user. I kept testing it and couldn’t figure out why it wouldn’t prompt me for any of the missing items until I saw that condition. Aaron Sent from my iPhone On Mar 14, 2021, at 8:30 PM, Alex Layton @.***> wrote: ο»Ώ @awlayton commented on this pull request. In oada/services/users/useradd.js: @@ -28,7 +26,7 @@ async function findUserByUsername(username) { // The main event: async function run() { try { if (argv.h || argv.help || Object.keys(argv).length <= 1) { Why did you remove this condition? It was to fix it hanging if called with no arguments. With this change it hangs for me again. In oada/services/permissions-handler/server.js: @@ -88,6 +96,10 @@ responder.on('request', function handleReq(req) { responder.scopes = { read: true, write: true }; } else { // Check for read permission if (!req.scope || !req.scope.some) { Why not just use Array.isArray? In oada/services/auth/token.js: @@ -0,0 +1,139 @@ +const argv = require('minimist')(process.argv.slice(2)); +const cloneDeep = require('clone-deep'); +const chalk = require('chalk'); +const uuid = require('uuid'); +const { authorizations, users } = @.***/lib-arangodb'); + +if (!process.env.DEBUG) { process.env.DEBUG = 'token:info,token:error'; A CLI tool should just be directly printing anything for user interaction rather than forcing debug to display it, β€” You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub, or unsubscribe. Yes I intentionally passed no arguments. I agree you should not have to remember the flags, but for me nothing happened when I ran it with no args so I made it print the help and exit. What do you mean pass -it? Those are not flags for docker-compose run which is how I was running it. To be clear, when I ran docker-compose run users add it never exited or printed anything. So I added that check. You should use oadadeploy instead so you don’t have to remember docker details :). β€œoadadeploy admin useradd " If you don’t want to use the oadadeploy script, then: If users is running, simplest is to use exec: docker-compose exec users yarn run add Otherwise, if you try β€œrun” but you use it normally instead of by replacing the entrypoint w/ yarn, it will start a second users process listening and responding to Kafka. Aaron On Mar 15, 2021, at 9:56 PM, Alex Layton @.***> wrote: Merged #83 https://github.com/OADA/oada-srvc-docker/pull/83 into master. β€” You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/OADA/oada-srvc-docker/pull/83#event-4461892320, or unsubscribe https://github.com/notifications/unsubscribe-auth/AA2KEUH2A5MAEHSHI3GYYS3TD23ENANCNFSM4ZEQTZHA. No, docker-compose run users add should work. If it does not work the wait it is, it needs to be changed. With that condition in it works for me when I supply flags. Using run does not start another instance of the users service unless you are doing so in the useradd.js file. Wouldn’t your entrypoint have to be yarn in order for that to work? Aaron On Mar 16, 2021, at 12:46 AM, Alex Layton @.***> wrote: No, docker-compose run users add should work. If it does not work the wait it is, it needs to be changed. With that condition in it works for me when I supply flags. Using run does not start another instance of the users service unless you are doing so in the useradd.js file. β€” You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/OADA/oada-srvc-docker/pull/83#issuecomment-799944143, or unsubscribe https://github.com/notifications/unsubscribe-auth/AA2KEUFNDC6EJVBVHF2UYE3TD3PD5ANCNFSM4ZEQTZHA. It is already https://github.com/OADA/oada-srvc-docker/blob/c4688206138f841bdf081c827e3cbebeb7b88ba8/oada/utils/entrypoint.sh#L3
2025-04-01T06:37:19.957529
2022-02-08T19:38:04
1127678673
{ "authors": [ "bali182" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2156", "repo": "OAI/OpenAPI-Specification", "url": "https://github.com/OAI/OpenAPI-Specification/issues/2880" }
gharchive/issue
2022-02-08-Websockets As a follow up to #55, I'm adding a proposal for Websocket support. If this is not the right channel to do so, please let me know what's the official way of opening a proposal, as I don't see it in any of the docs, and I don't have the proposals folder when I forked the repo. Find the filled proposal template below: WebSockets Metadata Tag Value Proposal 2022-02-08-Websockets Authors Balazs Edes Review Manager TBD Status Proposal Implementations - Issues #55 Previous Revisions - Change Log Date Responsible Party Description 2022-02-08 Balazs Edes First draft created Introduction This proposal aims to make WebSocket api descriptions part of OpenAPI. Motivation Currently the OpenAPI standard is limited to describing what the http standard allows. More and more applications make use of WebSockets, which is a simple and standard way to build event driven systems, and runs in the most important platform end users see: the browser. Since there is demand for it, I propose to add the ability to describe WebSockets in OpenAPI schemas, as HTTP/WebSocket apis often complement each other. There are other initiatives to describe event driven messaging (eg.: AsyncAPI), but in terms of maturity I haven't seen anything close to OpenAPI. AsyncAPI is also trying to acomplish a wide array of other goals which delay it's maturity further. Looking at the traction that #55 gained over the years, I think it would be a reasonable goal to add the ability to describe WebSockets in OpenAPI. Proposed solution This proposal aims to add a new WebSocket Payload Object type to the spec, and 2 additional fields to the Operation Object, publish and subscribe (better name suggestions are wellcome), that would allow us describing how to connect to a ws server, and what messages can we expect to receive or send. Example openapi: '3.3.0' paths: /sample-socket: get: operationId: 'sampleSocket' publish: description: 'Short description about the messages you can push towards the server' content: application/json: schema: oneOf: - $ref: '#/components/schemas/CreateUserSchema' - $ref: '#/components/schemas/DeleteUserSchema' - $ref: '#/components/schemas/LikePostSchema' subscribe: description: 'Short description about the messages you can receive from the server' content: application/json: schema: oneOf: - $ref: '#/components/schemas/UserCreatedSchema' - $ref: '#/components/schemas/UserDeletedSchema' - $ref: '#/components/schemas/PostLikedSchema' Detailed design The WebSocket spec is currently very simple. From our point of view it allows: Connecting to an endpoint using a URL Sending messages in either binary or text format Receiving messages in either binary or text format This proposal aims to allow OpenAPI users to describe exactly the above mentioned points: What URL can I connect to, to talk to this WebSocket? What's the format and shape of data I'm allowed to send What's the format and shape of data I can expect to receive Important: This OpenAPI spec proposal doesn't concern itself with any frameworks, or any higher level abstractions people usually build on top of event driven messaging protocols, like topics/subjects/channels, implied message serialization, etc. It simply describes the above mentioned 3 points, and any further abstractions can be built on top of this, but not part of the core schema. Extension of the current OpenAPI spec Add a WebSocket Payload Object type (very similar to Request Body Object, except the required field - which makes no sense here). It has the following fields: description: string - A brief description of the payload. content: Map[string, Media Type Object] The content of the payload. Add an optional field in the Components Object type, so payloads can be reused: webSocketPayloads: Map[string, WebSocket Payload Object] - Reuseable payloads Add 2 optional fields on the Operation Object type: publish: Reference Object | WebSocket Payload Object - describes the message(s) the client can publish/push to the server. subscribe: Reference Object | WebSocket Payload Object - describes the message(s) the client can receive from the server. Simple chat example openapi: '3.3.0' components: schemas: ChatMessage: type: 'object' required: - 'from' - 'to' - 'message' properties: from: type: 'string' to: type: 'string' message: type: 'string' paths: /chat-using-socket: get: operationId: 'chatSocket' publish: description: 'You can publish chat messages to this server' content: application/json: schema: $ref: '#/components/schemas/ChatMessage' subscribe: description: 'You can expect chat messages from the server' content: application/json: schema: $ref: '#/components/schemas/UserCreatedSchema' Simple chat, but reusing payloads openapi: '3.3.0' components: schemas: ChatMessage: type: 'object' required: - 'from' - 'to' - 'message' properties: from: type: 'string' to: type: 'string' message: type: 'string' webSocketPayloads: ChatMessagePayload: description: 'A chat message payload' content: application/json: schema: $ref: '#/components/schemas/ChatMessage' paths: /chat-using-socket: get: operationId: 'chatSocket' publish: $ref: '#/components/webSocketPayloads/ChatMessagePayload' subscribe: $ref: '#/components/webSocketPayloads/ChatMessagePayload' Multiple message types Note since channels/topics/subjects cannot be expressed we can use a union type with a discriminator to mimic this. openapi: '3.3.0' components: schemas: CreateUserSchema: '...' DeleteUserSchema: '...' LikePostSchema: '...' PublishMessageSchema: discriminator: propertyName: 'action' mapping: CreateUser: '#/components/schemas/CreateUserSchema' CreateUser: '#/components/schemas/CreateUserSchema' LikePost: '#/components/schemas/LikePostSchema' oneOf: - $ref: '#/components/schemas/CreateUserSchema' - $ref: '#/components/schemas/DeleteUserSchema' - $ref: '#/components/schemas/LikePostSchema' UserCreatedSchema: '...' UserDeletedSchema: '...' PostLikedSchema: '...' SubscribeMessageSchema: discriminator: propertyName: 'type' mapping: UserCreated: '#/components/schemas/UserCreatedSchema' UserDeleted: '#/components/schemas/UserDeletedSchema' PostLiked: '#/components/schemas/PostLikedSchema' oneOf: - $ref: '#/components/schemas/UserCreatedSchema' - $ref: '#/components/schemas/UserDeletedSchema' - $ref: '#/components/schemas/PostLikedSchema' paths: /sample-socket: get: operationId: 'sampleSocket' publish: description: 'Short description about the messages you can push towards the server' content: application/json: schema: - $ref: '#/components/schemas/PublishMessageSchema' subscribe: description: 'Short description about the messages you can receive from the server' content: application/json: schema: - $ref: '#/components/schemas/SubscribeMessageSchema' Backwards compatibility As this change is additive (nothing is removed or modified in a non-backward compatible way) I don't see any backward compatibilty issues. Alternatives considered There can be many alternatives, my motivation was to add support for WebSocket api descriptions, while Adding as few new concepts as possible Allowing reuse the same way as with all other parts of the schema, through $refs Just describing what the standard allows (no frameworks) Not causing any breaking changes or backward compatibility issues I'm happy to take onboard any suggestions, alternatives, modifications to this! I'm sure there are many things I haven't thought of.
2025-04-01T06:37:19.968642
2017-03-01T13:47:21
211085719
{ "authors": [ "DavidBiesack", "MikeRalphson", "darrelmiller", "fehguy", "willdady" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2157", "repo": "OAI/OpenAPI-Specification", "url": "https://github.com/OAI/OpenAPI-Specification/issues/953" }
gharchive/issue
Clarify thinking behind .example object and .examples array ? Please could someone provide some background on the decision to have both .example and .examples[] properties on the requestBody, responses and schemas? It seems to me this might unnecessarily complicate documentation tooling, while the alternative is simply using a one-element array if only one example is provided. Extending an existing definition with a single example would involve more editing when adding a second one. An operation object does not have both a .tag and .tags[] properties. I also found the following wording confusing: SHOULD be in the correct format as specified [in the] parameter encoding I'm probably missing something. From my perspective there wasn't a strong motivation for keeping the "example" property. It was primarily for simplicity and backward compatibility. We have added a fair amount of complexity to some objects, there was some reluctance to force example arrays on everyone. We have made examples and example mutually exclusive, so for tooling it shouldn't be much more than an extra if statement to handle both properties. The examples/example have been added to the content object, so when tooling people implement it, it will automatically work for requestBody, responses, callbacks, links, complex parameters. The example has been left in for schemas. Maybe that isn't needed any more now that example objects are now reusable components. It is possible there could be some streamlining, but if we were to remove example I think it would be more for reducing conceptual complexity, rather than making tooling easier. @MikeRalphson does that answer your question? If so please close... @fehguy @darrelmiller Thanks for the response. I would however question the need for "backwards compatibility" when there has explicitly been a breaking change in the version of the specification. There seems to be no other examples of such mental backwards compatibility around servers, requestBodies etc. I'm not wholly convinced by the simplicity argument either I'm afraid, there is one spec to create, but many tools, each of which would have to take this wrinkle into account. If it looks like a wart, and sounds like a wart... If it is unlikely to change, or is still up for consideration outside the public sphere of GitHub issues, then yes, please close. Well this was actually requested and should hopefully be captured in the issues. Certainly nobody was creating work just for the sake of creating work. I'll have a hunt through the issues. Certainly nobody was creating work just for the sake of creating work. I hope I didn't come off as remotely suggesting that. No problem at all. I believe the examples arguments were around the requestBody changes Looks like Issue 800 is the most relevant. In which @DavidBiesack expressed a preference for the oneOf example / examples. I could find no other requests for this specific feature in the issues. I'll just chime in that my preference is strictly for ease of use of the API author. I think many cases one example is sufficient and using example: is very easy whereas using examples: and coding an array is just more tedious. I think tooling can pretty easily account for this and do the transformation for us; that's what computers are good at. Again, this is just my preference - -I tend to side on ease of use for people using OAS. Understood, but I still contend that it doesn't smell like any other part of the OpenAPI 3 spec. ;) It is a personal preference, I'm not strongly opposed to more uniform implementation (i.e. just examples: [ ]). (I do not think that preserving singular example: syntax just because it was singular in 2.0 is important here.) I agree with @MikeRalphson. To me this seems overly verbose when a single item examples array would suffice. If the argument for having both is for 'backwards compatibility' then I think this needs to be revisited since 3.0 is strictly not compatible with 2.0. I would argue the cognitive load of understanding the difference outweighs any tediousness of coding an array (literally adding a - in yaml, json slightly more work sure).
2025-04-01T06:37:20.106565
2023-10-11T21:29:06
1938772253
{ "authors": [ "turnerm" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2158", "repo": "OCHA-DAP/hapi-pipelines", "url": "https://github.com/OCHA-DAP/hapi-pipelines/pull/46" }
gharchive/pull-request
Hapi 131 bugfixes Lots of debugging here, had to: update the reqs to use the latest hapi-schema run the pg_dump command from the docker container add the HDX API key I guess the hapi-pipelines Dockerfile and publish workflow are redundant? @mcarans I think publish will not be needed as I don't think we'd need to publish to PyPI. The dockerfile could be kept until we see exactly what we need for V2. What do you think?
2025-04-01T06:37:20.357632
2022-08-22T13:20:22
1346415068
{ "authors": [ "catenacyber" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2159", "repo": "OISF/suricata-verify", "url": "https://github.com/OISF/suricata-verify/pull/913" }
gharchive/pull-request
dhcp: adds test about leasetime keyword Replaces #873 I want to build on it for other DHCP keywords ;-) cf https://redmine.openinfosecfoundation.org/issues/5506 Closing in favor of #919 now that recent Suricata PR has been merged
2025-04-01T06:37:20.411276
2020-11-14T06:38:49
742942972
{ "authors": [ "akuckartz", "skasberger" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2160", "repo": "OKFNat/offenewahlen_api", "url": "https://github.com/OKFNat/offenewahlen_api/issues/234" }
gharchive/issue
Still active ? The latest changes are three years old and recently lots of issues were closed without any explanation. Is this project still active? If not: can this please be stated at the top of the README ? The project is not active anymore, as Open Knowledge Austria does not exist anymore for more than 2 years. Started a new project from scratch 2 weeks ago, you can find here: https://github.com/skasberger/owat_api You could add that information here: https://github.com/OKFNat/offenewahlen_api
2025-04-01T06:37:20.414189
2024-10-14T18:32:14
2586760926
{ "authors": [ "angelaw7" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2161", "repo": "OKKM-insights/OKKM.insights", "url": "https://github.com/OKKM-insights/OKKM.insights/issues/123" }
gharchive/issue
Peer Review - SRS: Could add context diagram Artifact Under Review SRS Document Team Number for Team Doing the Review Team 10 Description of Issue Within section 6.2. The Context of the Work, you could potentially add a context diagram to outline how external entities interact with an internal software system. (I'm not sure if it's required for your case but it was required for the scientific computing template and I saw a previous group example add it as part of their revision 1, section 4.1.1.) It could be helpful to give the reader a more abstract understanding of the system before diving into the use case diagram. Team 10 TA: @Necried
2025-04-01T06:37:20.426762
2017-05-22T21:10:52
230520733
{ "authors": [ "JohnPerkin", "steampoweredlawngnome" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2162", "repo": "ONLYOFFICE/onlyoffice-owncloud", "url": "https://github.com/ONLYOFFICE/onlyoffice-owncloud/issues/71" }
gharchive/issue
Nextcloud 12's new editable public links should open docs with OnlyOffice Nextcloud 12 now allows public link shares to be writable, so that a user can share a document with a non-Nextcloud user and have them jump into the editor. This functionality is advertised for use with Collabora/CODE, but I imagine OnlyOffice could use it as well. Hello @steampoweredlawngnome The issue has already been discussed earlier https://github.com/ONLYOFFICE/onlyoffice-owncloud/issues/20
2025-04-01T06:37:20.435761
2020-08-20T09:46:19
682576780
{ "authors": [ "AdamHawtin", "ryangrundy7" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2163", "repo": "ONSdigital/census-rm-ops-ui", "url": "https://github.com/ONSdigital/census-rm-ops-ui/pull/3" }
gharchive/pull-request
1221 Search for case and show results Motivation and Context The first feature needed for ops-ui is to be able to search for cases via a postcode. This PR adds that functionality as well as IAP Audit logging and adding the CDN to ops-ui. What has changed Search for cases via postcode IAP audit logging Added flask/jinja templates to show postcode results. Added Flask blueprints for routing Now runs with Gunicorn How to test? Run with the other branches on the ticket. When you have cases in your database, you should be able to run the ops-ui and find cases with a postcode. Try running this in your GCP environment and see if you can connect to it. Anyone testing I can add to the IAP so you could try and connect to the domain set up in census-rm-ryangrundy Links Trello Missing columns, specified in the acceptance criteria (although I personally think it's far too many): org name case type estab type UPRN region @NickGrantONS Org name is part of the address summary that gets put into the table. Estab type and case type are included in the page. UPRN and Region aren't in the prototype that we were basing this on. Could be a question for Dan if they need to be included
2025-04-01T06:37:20.437141
2020-10-20T13:15:25
725575374
{ "authors": [ "rmccar" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2164", "repo": "ONSdigital/design-system", "url": "https://github.com/ONSdigital/design-system/pull/1110" }
gharchive/pull-request
Remove Back Link from Feedback thank you example What is the context of this PR? Remove Back Link from Feedback thank you example as it will not be on the page when implemented in runner How to review Check Feedback examples look and work as they should Not planning on removing back link from the pattern at the moment
2025-04-01T06:37:20.441041
2024-07-15T10:55:57
2408442312
{ "authors": [ "precious-onyenaucheya-ons", "rmccar" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2165", "repo": "ONSdigital/design-system", "url": "https://github.com/ONSdigital/design-system/pull/3268" }
gharchive/pull-request
update ons-grid--flex to ons-grid-flex What is the context of this PR? Fixes #3259. This PR renames utility class from ons-grid--flex to ons-grid-flex. Please note that this may be a breaking change. How to Resolve the Breaking Change Search your codebase for any occurrences of ons-grid--flex. Replace all instances with the new class name ons-grid-flex. How to review this PR Verify that all instances of ons-grid--flex have been renamed to ons-grid-flex. Ensure that all references in the documentation, examples, and tests have been updated accordingly. Confirm that all tests pass. Checklist This needs to be completed by the person raising the PR. [x] I have selected the correct Assignee [x] I have linked the correct Issue The PR description needs updating. Also as you have now started using the sass syntax in this file I was wondering if we want to restructure the rest of this file using sass syntax
2025-04-01T06:37:20.444313
2018-06-29T16:18:27
337058918
{ "authors": [ "CarlHembrough", "mattrout92" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2166", "repo": "ONSdigital/dp-api-tests", "url": "https://github.com/ONSdigital/dp-api-tests/pull/94" }
gharchive/pull-request
Feature/instance to code relation What Update end to end tests to import the CPIH codelist into neo4j. This allows the dimension importer to create relationships between the instance node and the codes that it relates to. How to review Review changes / test End to end test using these changes: https://github.com/ONSdigital/dp-api-tests/pull/94 https://github.com/ONSdigital/dp-dimension-importer/pull/37 https://github.com/ONSdigital/dp-dataset-exporter-xlsx/pull/68 https://github.com/ONSdigital/dp-dataset-exporter/pull/43 https://github.com/ONSdigital/dp-download-service/pull/29 Who can review Anyone Can you increase the default timeout to 30 or 45 seconds as I cannot get the instance to complete in time. Other than that, seems good πŸ‘
2025-04-01T06:37:20.484888
2019-06-04T22:15:31
452231721
{ "authors": [ "michael-adler" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2167", "repo": "OPAE/opae-sdk", "url": "https://github.com/OPAE/opae-sdk/pull/1300" }
gharchive/pull-request
Work around a glitch in the ASE DDR model The DDR simulation model's Avalon ready/enable protocol deliberately uses clock offsets, which wreak havoc even in the Qsys-supplied bridges. Adding a small delay to the simulated DDR clock passed to the AFU causes the AFU to sample signals when they are stable. This appears to have been the cause of random failures in hello_mem_afu simulation. Tests: http://sw-pert.altera.com/pert/pert.php?test_run_id=3410442
2025-04-01T06:37:20.520724
2021-06-09T01:48:46
915693799
{ "authors": [ "opstest2", "v-alji" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2168", "repo": "OPS-E2E-PPE/docs-archive-test-source", "url": "https://github.com/OPS-E2E-PPE/docs-archive-test-source/pull/2000" }
gharchive/pull-request
DocsArchive-<IP_ADDRESS>Auto archive content Please don't merge this PR until target archive repo PR https://github.com/OPS-E2E-PPE/docs-archive-test-target/pull/896 is merged into live branch. Auto archive content to https://github.com/opstest2/docs-archive-test-target.git Docs Build status updates of commit 18a80dc: :x: Validation status: errors Please follow instructions here which may help to resolve issue. File Status Preview URL Details :x:Error Details [Error-GitBranchDeletedOrForcePushed] Cannot sync git repo to specified commit because branch Release_Archive_master_2021-06-09-01-48-33 has been deleted or has been force pushed For more details, please refer to the build report. If you see build warnings/errors with permission issues, it might be due to single sign-on (SSO) enabled on Microsoft's GitHub organizations. Please follow instructions here to re-authorize your GitHub account to Docs Build. Note: Broken links written as relative paths are included in the above build report. For broken links written as absolute paths or external URLs, see the broken link report. Note: Your PR may contain errors or warnings unrelated to the files you changed. This happens when external dependencies like GitHub alias, Microsoft alias, cross repo links are updated. Please use these instructions to resolve them. For any questions, please:Try searching the docs.microsoft.com contributor guidesPost your question in the Docs support channel
2025-04-01T06:37:20.529790
2022-02-21T01:18:06
1145207025
{ "authors": [ "opstest2", "v-alji" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2169", "repo": "OPS-E2E-PPE/docs-archive-test-source", "url": "https://github.com/OPS-E2E-PPE/docs-archive-test-source/pull/2549" }
gharchive/pull-request
DocsArchive-<IP_ADDRESS>Auto archive content Please don't merge this PR until target archive repo PR https://github.com/OPS-E2E-PPE/docs-archive-test-target_82c94236-9731-4b30-ba51-24cc2fbf2b05/pull/1 is merged into live branch. Auto archive content to https://github.com/opstest2/docs-archive-test-target_82c94236-9731-4b30-ba51-24cc2fbf2b05.git Docs Build status updates of commit f8f79df: :x: Validation status: errors Please follow instructions here which may help to resolve issue. File Status Preview URL Details :x:Error Details [Error: GitBranchDeletedOrForcePushed] Cannot sync git repo to specified commit because branch Release_Archive_main_2022-02-21-01-17-20 has been deleted or has been force pushed For more details, please refer to the build report. If you see build warnings/errors with permission issues, it might be due to single sign-on (SSO) enabled on Microsoft's GitHub organizations. Please follow instructions here to re-authorize your GitHub account to Docs Build. Note: Broken links written as relative paths are included in the above build report. For broken links written as absolute paths or external URLs, see the broken link report. Note: Your PR may contain errors or warnings unrelated to the files you changed. This happens when external dependencies like GitHub alias, Microsoft alias, cross repo links are updated. Please use these instructions to resolve them. For any questions, please:Try searching the docs.microsoft.com contributor guidesPost your question in the Docs support channel
2025-04-01T06:37:20.537831
2021-05-14T05:23:34
891617495
{ "authors": [ "opstest2", "v-alji" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2170", "repo": "OPS-E2E-PPE/docs-archive-test-target", "url": "https://github.com/OPS-E2E-PPE/docs-archive-test-target/pull/811" }
gharchive/pull-request
DocsArchive-<IP_ADDRESS>Auto archive content Auto archive content from https://github.com/opstest2/docs-archive-test-source.git Docs Build status updates of commit fc4fbe9: :x: Validation status: errors Please follow instructions here which may help to resolve issue. File Status Preview URL Details :x:Error Details [Error-RuningBuildFailed] Some unexpected errors happened when running build, please open a ticket in https://aka.ms/SiteHelp and include the error report for our team to troubleshoot For more details, please refer to the build report. If you see build warnings/errors with permission issues, it might be due to single sign-on (SSO) enabled on Microsoft's GitHub organizations. Please follow instructions here to re-authorize your GitHub account to Docs Build. Note: Broken links written as relative paths are included in the above build report. For broken links written as absolute paths or external URLs, see the broken link report. Note: Your PR may contain errors or warnings unrelated to the files you changed. This happens when external dependencies like GitHub alias, Microsoft alias, cross repo links are updated. Please use these instructions to resolve them. For any questions, please:Try searching the docs.microsoft.com contributor guidesPost your question in the Docs support channel
2025-04-01T06:37:20.543727
2019-09-25T03:15:59
498023910
{ "authors": [ "VSC-Service-Account", "v-alji" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2171", "repo": "OPS-E2E-PPE/docs-archive-test", "url": "https://github.com/OPS-E2E-PPE/docs-archive-test/pull/124" }
gharchive/pull-request
DocsArchive-<IP_ADDRESS>Auto archive content Please don't merge this PR until target archive repo PR https://github.com/OPS-E2E-PPE/docs-archive-test-target/pull/57 is merged into live branch. Auto archive content to https://github.com/OPS-E2E-PPE/docs-archive-test-target.git Docs Build status updates of commit 816be5d: :x: Validation status: errors Please follow instructions here which may help to resolve issue. File Status Preview URL Details :x:Error Details [Error] Cannot sync git repo to specified commit because branch Archive_master_2019-09-25-11-15-07 has been deleted or has been force pushed: fatal: Couldn't find remote ref Archive_master_2019-09-25-11-15-07 For more details, please refer to the build report. Note: If you changed an existing file name or deleted a file, broken links in other files to the deleted or renamed file are listed only in the full build report.
2025-04-01T06:37:20.552391
2016-11-11T18:33:49
188821995
{ "authors": [ "aprokop", "dalg24", "sslattery" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2172", "repo": "ORNL-CEES/DataTransferKit", "url": "https://github.com/ORNL-CEES/DataTransferKit/issues/130" }
gharchive/issue
Add more warnings to Jenkins I think to we need to further clean up the code to get rid of warnings. Thus I suggest we add -Werror and additional warnings detection to Jenkins CI. At the moment I get things like this: [1059/1294] Building CXX object DataTransferKit/packages/Operators/test/CMakeFiles/dtk_hex_test_reference.dir/reference_implementation/DTK_ReferenceHexMesh.cpp.o In file included from ../../DataTransferKit/packages/Interface/src/Client/DTK_EntitySet.hpp:47:0, from ../../DataTransferKit/packages/Interface/src/OperatorVector/DTK_FunctionSpace.hpp:46, from ../../DataTransferKit/packages/Operators/test/reference_implementation/DTK_ReferenceHexMesh.hpp:45, from ../../DataTransferKit/packages/Operators/test/reference_implementation/DTK_ReferenceHexMesh.cpp:41: ../../DataTransferKit/packages/Interface/src/Client/DTK_EntityIterator.hpp:91:28: warning: β€˜virtual DataTransferKit::EntityIterator DataTransferKit::EntityIterator::operator++(int)’ was hidden [-Woverloaded-virtu al] virtual EntityIterator operator++( int ); ^~~~~~~~ In file included from ../../DataTransferKit/packages/Operators/test/reference_implementation/DTK_ReferenceHexMesh.cpp:50:0: ../../DataTransferKit/packages/Adapters/BasicGeometry/src/DTK_BasicEntitySet.hpp:83:21: warning: by β€˜virtual DataTransferKit::EntityIterator& DataTransferKit::BasicEntitySetIterator::operator++()’ [-Woverloaded -virtual] EntityIterator &operator++() override; ^~~~~~~~ One thing, though, is that some of Trilinos packages like STK are more warning-prone. I think the linear algebra portion, though, is mostly warnings free. Remember that we get warnings from TPLs when we build too that we have no control over. Would this flag cause the build to fail because a warning from a TPL? I usually have the following in my configure script: SYSTEM_HEADERS="" for header_dir in `echo $CPATH | sed 's/:/\ /g'`; do SYSTEM_HEADERS="$SYSTEM_HEADERS -isystem $header_dir" done ... -D CMAKE_CXX_FLAGS="-g -Wall -Wextra -Wno-unused-parameter $CXX_FLAGS_COMPILER_SPECIFIC $SYSTEM_HEADERS" This way the compiler would ignore the headers coming from the things in you $CPATH (or whatever you put there). So TPLs should be fine. Most issues I see coming from Trilinos packages. I think this is a good idea then as long as we can configure this to just fail on any DTK warnings. The problem is not so much the TPL but the other packages in Trilinos. Fixed in #134.
2025-04-01T06:37:20.554838
2024-08-29T20:52:30
2495555254
{ "authors": [ "gknapp1" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2173", "repo": "ORNL-MDF/Myna", "url": "https://github.com/ORNL-MDF/Myna/issues/24" }
gharchive/issue
Refactor bnpy data clustering apps Currently the data clustering implementation the bnpy-based apps is functional, but can be slow to converge when too much training data is added. It can also be awkward to re-cluster after the model is updated. In order to have automatic region of interest selection in Peregrine, the clustering implementation should be sped up. Some ideas for enhancements: use a representative sample for training instead of the entire dataset instead of training and clustering for each part sequentially, create training data based on all parts first and then cluster create a the Bnpy subclass of the base App class to streamline the apps themselves Removing this from the Peregrine demonstration milestone, because of the capability implemented in #25. Updating the bnpy app is still planned, but lower priority for the moment.
2025-04-01T06:37:20.622458
2024-05-12T17:23:40
2291425994
{ "authors": [ "PashaPal1974" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2174", "repo": "OX-Security-Demo/Multi-currency-management", "url": "https://github.com/OX-Security-Demo/Multi-currency-management/pull/1376" }
gharchive/pull-request
Upgrade dependency Upgrade<EMAIL_ADDRESS>to version 9.0.2 Please ignore this PR and close it. Generated by TestIM
2025-04-01T06:37:20.626505
2021-02-18T18:35:24
811357778
{ "authors": [ "ConversaBC", "Vohras2" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2175", "repo": "ObliqueNET/Server", "url": "https://github.com/ObliqueNET/Server/issues/1011" }
gharchive/issue
Sceptile and TRs I have been testing some TRs and TMs from gen 8 since the recent patch and discovered an issue. Sceptile is unable to learn Dragon Pulse through the Dragon Pulse TR. However, it learned Focus Blast through TRs, which means Dragon Pulse is glitched for Sceptile - likely other pokemon as well. I'm unsure if the screenshot will show, but I'll try anyway. This is working as intended. Sceptile didn't learn Dragon Pulse by TM in Gen 7, it was a tutor move. It still learns it by tutor as of now. Odd, the website PokemonDB shows that Sceptile can learn Dragon Pulse via TR in the Crown Tundra DLC. Are DLC mons not considered gen 8 mons? This will be fixed in the next day or two.
2025-04-01T06:37:20.627636
2018-01-21T11:18:32
290265006
{ "authors": [ "Hunter1220", "Rasgnarok" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2176", "repo": "ObliqueNET/Terra", "url": "https://github.com/ObliqueNET/Terra/issues/56" }
gharchive/issue
Incorrect Mega Medicham textures https://puu.sh/z6kWa/05f9da5705.png Sorry for the bad angle but it would appear that Mega Medicham is using the same textures as Sylveon, at least to some degree, as indicated by the out of place pinks and blues on the model. Fixed
2025-04-01T06:37:20.629907
2024-03-28T17:25:17
2213755239
{ "authors": [ "ajgosl", "ulrikpedersen" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2177", "repo": "Observatory-Sciences/aravis-detector", "url": "https://github.com/Observatory-Sciences/aravis-detector/pull/3" }
gharchive/pull-request
Configuration upgrades It was not possible to correctly set the rate. This has been resolved. The cache of config items was not being updated correctly, this has been resolved. Can't really claim much JS expertise, but it looks fine. Hope it works too πŸ˜„
2025-04-01T06:37:20.631251
2023-05-24T21:57:43
1724791563
{ "authors": [ "jagoosw", "navidcy" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2178", "repo": "OceanBioME/OceanBioME.jl", "url": "https://github.com/OceanBioME/OceanBioME.jl/issues/98" }
gharchive/issue
SLatissima parameters do not show This snipped https://github.com/OceanBioME/OceanBioME.jl/blob/1c5598ec482c97f92008250d6071c3d491a3e5c1/docs/src/appendix/params/SLatissima.md#L4-L5 doesn't seem to be doing anything. Oh yeah that doesn't work anymore. I'd quite like to document the default parameters somehow but haven't worked out how todo it yet.
2025-04-01T06:37:20.637180
2024-08-07T12:49:44
2453406892
{ "authors": [ "KOMPALALOKESH", "VeckoTheGecko", "erikvansebille" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2179", "repo": "OceanParcels/parcels", "url": "https://github.com/OceanParcels/parcels/issues/1642" }
gharchive/issue
Add FieldSet.from_a_grid_dataset() method Parcels has specific methods for loading B-grid and C-grid datasets, but not for A-grids. This is because A-grid (linear) interpolation is default, so FieldSet.from_netcdf() automatically assumes an A-grid. However, for consistency and to help new users, it might be nice to add an explicit FieldSet.from_a_grid_dataset() method to Parcels. This method would then simply call from_netcdf(). Note that while this is done, it might also be good to update a few mentions of from_bgrid() and from_cgrid() in the docstrings to the correct from_b_grid_dataset() and from_c_grid_dataset() Hi @erikvansebille , I'd like to work on this enhancement. Could you please assign this issue to me? Thank you! Thanks @KOMPALALOKESH, for wanting to pick this up! I've assigned this Issue to you. I think the key starting point is to make a new method in parcels/fieldset.py that returns a cls.from_netcdf(), a bit like how fieldset.from_b_grid() is structured but then without much (any?) adjustments. It would also be nice to add a unit test to tests/test_fieldset.py. Good luck, and let us know if you need any help! @KOMPALALOKESH if you're interested and want more than this issue, "Enable pyupgrade on ruff linting" from #1620 is well defined and might be interesting from a technical standpoint if you haven't worked with Ruff before. Its a nice Python tool, and modern Python is consolidating towards it for handling QAQC. @erikvansebille , This PR Link introduces the FieldSet.from_a_grid_dataset() method in "parcels/fieldset.py" to handle A-grid datasets. A corresponding unit test has been added in "tests/test_fieldset.py" to validate this functionality. Please review the changes and the test case. If everything is in order, kindly merge.
2025-04-01T06:37:20.721120
2022-03-18T04:27:25
1173176413
{ "authors": [ "infinite-persistence", "tzarebczan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2180", "repo": "OdyseeTeam/odysee-frontend", "url": "https://github.com/OdyseeTeam/odysee-frontend/pull/1141" }
gharchive/pull-request
Fix double email_user/confirm due to language render Add a persistent (per refreshed session) flag to skip any actions sparked by future renders after it has already been authenticated. Tested as much as I can, but do try at kp in case I broke something. We patched it up on the backend, but this should help anyway..will double check later tomorrow. Yeah, this will prevent it from jumping to SignUp flow page due to the re-render. (couldn't really replicate an actual problem -- had to force-render it to get the double confirm. Maybe only happens in certain locale).
2025-04-01T06:37:20.729237
2022-02-21T05:25:40
1145340854
{ "authors": [ "AnthonyPhan", "Prasad-MSFT", "v-chetsh" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2181", "repo": "OfficeDev/Microsoft-Teams-Samples", "url": "https://github.com/OfficeDev/Microsoft-Teams-Samples/issues/254" }
gharchive/issue
Pure client side Graph example Hi All, I have been studying the various examples provided here however Im unable to see any examples which provide a pure client side app which makes graph calls. Is there a reason for this? Im looking for some example which simply uses the authorization code flow with pkce to get a token and call the MS Graph. Is there som technical resson why this is not available? There is no such sample seems to be available. @AnthonyPhan - As per my understanding, the standard and recommended way to do the authentication at the server side and then call the API's at the client side. However you can implement the authorization code flow at client side and do the Graph API call's using generated token at client side. @AnthonyPhan - Please let us know if you need any further details or shall we close this issue? @v-chetsh it looks like the "simple-auth" service is being deprecated and as of a recent update of teamsfx the standard hello world uses a complete client side implementation of authorisation code flow with PKCE as suggested my this post. Here is a further reference: https://stackoverflow.com/q/70388145/3095420
2025-04-01T06:37:20.731487
2023-05-25T18:29:30
1726294584
{ "authors": [ "cristianoag", "gsv022" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2182", "repo": "OfficeDev/microsoft-teams-apps-company-communicator", "url": "https://github.com/OfficeDev/microsoft-teams-apps-company-communicator/issues/1054" }
gharchive/issue
Deployment form mentions that custom domains are recommended The documentation correctly states the advantages of having Azure Front Door for company communicator, however the deployment form we get when executing the default ARM deployment template states that custom domains are recommended. Could you please review that? Hi @cristianoag , Thanks for highlighting. We will check on this. Hi @cristianoag , In the new release, the above mentioned correction has been taken care. Please let us know if you have any concerns.
2025-04-01T06:37:20.735486
2023-12-01T07:14:12
2020258793
{ "authors": [ "gsv022", "reterfreeman" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2183", "repo": "OfficeDev/microsoft-teams-apps-company-communicator", "url": "https://github.com/OfficeDev/microsoft-teams-apps-company-communicator/issues/1245" }
gharchive/issue
Deploy failed with"There was a conflict. Conflict with existing ScmType: ExternalGit" Hi, I deploy CC 5.5 in another tenant, but it was failed, I found the closed issue#1078, follow the way (disconnect and reconnect, wait for Log show up: success(Active), and redeploy again, but it still fail with There was a conflict. Conflict with existing ScmType: ExternalGit I go back Deployment Center and External Git show up "Error fetching information", however I disconnect and connect again, redeploy always failed with "There was a conflict. Conflict with existing ScmType: ExternalGit" I am sure not any underscore (_) or space in any of the field values, include subscription & resource group. @gsv022 It's working, and deployed success. thanks for your support. Hi @reterfreeman , Thanks for the update. Please feel free to log a issue if anything comes up w.r.t standard version of CC
2025-04-01T06:37:20.748249
2024-02-27T11:26:21
2156345494
{ "authors": [ "Sheshank-TCS", "chrischowsos", "gsv022" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2184", "repo": "OfficeDev/microsoft-teams-apps-company-communicator", "url": "https://github.com/OfficeDev/microsoft-teams-apps-company-communicator/issues/1354" }
gharchive/issue
Author App is partially available to users if published. Hi Team, In the recent version (5.5.1) of the Company Communicator, I deployed both the Author and User apps via the Teams admin console. In earlier versions, if a non-authorized UPN user attempted to add the Author app, it would get added in Teams but would not open, displaying an authorization error. However, in version 5.5.1, users can add the app and also view the card design module, which was previously unavailable. Despite this, they still cannot send messages from the Author app. Can we enforce the same level of restriction as in the earlier version, where non-authorized users were completely restricted from accessing the app? Below are the screenshots of the old version author app added by random users after publishing it in Teams admin via Manage Apps. Below is the screenshot of the new version 5.5.1 author app added by random users after publishing it in Teams Admin via Manage Apps. Hi @Sheshank-TCS , Thanks for raising the query. Normally, in any version unauthorized users won't have access to CC author app unless their UPN is added to the UPN list. Coming to the error, check with that specific user on browser version of teams so that user will get relevant error when tries to access the application. Hi @gsv022 I have validated on teams web version also and its still available. Though the message cannot be sent but its not giving any access denied error. Please suggest how this can be fixed. Hi @Sheshank-TCS , Is it happening in New Teams as well? Yes, this is happening with New teams, Old teams and Web version. However old Company communicator app is restricted everywhere for the users who are not Authors. Hi @Sheshank-TCS , Please share your mail id so that we can discuss the issue internally. <EMAIL_ADDRESS>you can use this one. I have sent a mail. Please check and respond Hi Sai, Thanks for your response, I have reverted to it. I am facing the same issue, is there any update of this issue? Hey Chris, Still waiting to hear from the team. Hi @Sheshank-TCS , We have tried in different tenants and not able to replicate the issue. Please deploy one more instance and see the issue if persists. Keep us posted Working towards the implementation will keep you posted. Sure, please keep us posted Hi Sai, Just a query, have you validated setting up a second instance when existing instance exist. May be if you are setting up a single new instance it is not giving error. Hi @Sheshank-TCS , We will check out this scenario and let you know. @gsv022 We tried setting up the new instance and the Function app failed. Please suggest whether should I manually sync the resource from the deployment center again. Hi @Sheshank-TCS , To overcome the failed resources, Go to azure portal-->Go to resource group(Where the deployment has been done)-->Select prep-function(Any failed resource) from list of resources-->Go to Deployment Center-->Click on Settings. Below is the screenshot for reference. If the resource is not pointing to intended repository and showing empty. Please connect to the external Git and point to the official repository(like above) Once the status turns to Success(active) in the logs section then good to proceed with remaining steps of deployment. Please do the same for other failed resource as well. After performing the above steps, don't attempt for re-deploy again We have successfully deployed the latest App version 5.5.2 and the issue has been resolved it's throwing the disclaimer message as expected when nonauthorized members add the App. Thank you Sai for all the help and support, appreciate it.
2025-04-01T06:37:20.751002
2024-11-25T16:02:40
2691287759
{ "authors": [ "fajr365", "v-ajaysahu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2185", "repo": "OfficeDev/microsoft-teams-emergency-operations-center", "url": "https://github.com/OfficeDev/microsoft-teams-emergency-operations-center/issues/297" }
gharchive/issue
Issue with Adding Sharepoint Extension. When adding the sharepoint App, the documentation seems to be out of date for the new sharepoint. In the final step "do you trust eoc-extention..." New sharepoint has 2 options "Only enable this app" and "Enable this app and add it to all sites" What is the correct choice? @fajr365 , thank you for bringing this to our attention. We will update our documentation in the upcoming releases. The correct option will be "Enable this app and add it to all sites". I hope this clarifies your query. If you have any further questions or issues, please feel free to post here.
2025-04-01T06:37:20.791944
2022-03-24T10:34:25
1179317022
{ "authors": [ "AlexJerabek", "aletrovato" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2186", "repo": "OfficeDev/office-scripts-docs", "url": "https://github.com/OfficeDev/office-scripts-docs/issues/476" }
gharchive/issue
Problem with Page about Date I would like to report a problem with this page. Suggestion 1: As I am looking for information on working with dates in office script, the text has a problem with the link in this paragraph: "The next sample reads a date that's stored in Excel and translates it to a JavaScript Date object. It uses the date's numeric serial number as input for the JavaScript Date." The link is directing to the Now function page. Suggestion 2 The script shown can be simplified to the code below, because in this snippet of the original ( let excelDateValue = dateRange.getValue() as number;) returns an error. function main(workbook: ExcelScript.Workbook) { // Read a date at cell A1 from Excel. let dateRange = workbook.getActiveWorksheet().getRange("A1").getValue(); // Convert the Excel date to a JavaScript Date object. let javaScriptDate = new Date(Math.round((dateRange - 25569) * 86400 * 1000)); console.log(javaScriptDate); } Thank you! Document Details ⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking. ID: 98c58cbf-aaf7-3d3a-48f4-ce9cebcf331c Version Independent ID: 74821d34-eab6-ec0f-7a9f-55aca81975d9 Content: Basic scripts for Office Scripts in Excel on the web - Office Scripts Content Source: docs/resources/samples/excel-samples.md Product: excel Technology: scripts GitHub Login: @o365devx Microsoft Alias: o365devx Thanks for pointing out this issue @aletrovato. I'll get the page updated soon. How amazing the speed of the response. Thank you so much Alex!!! Alessandro Trovato Microsoft MVP / Instrutor LinkedIn Learning / Youtuber Blog, CatΓ‘logo Virtual das Aulas e cursos com certificado https://linktr.ee/alessandrotrovato Acesso ao Youtube http://www.youtube.com/aletrovato App gratuito na Play Store https://bit.ly/apptrovato com as aulas Compartilhando conhecimento, sempre! Em qui., 24 de mar. de 2022 Γ s 13:58, Alex Jerabek @.***> escreveu: Thanks for pointing out this issue @aletrovato https://github.com/aletrovato. I'll get the page updated soon. β€” Reply to this email directly, view it on GitHub https://github.com/OfficeDev/office-scripts-docs/issues/476#issuecomment-1077838127, or unsubscribe https://github.com/notifications/unsubscribe-auth/AC4HPXZNHH4EQYA6MVLQDI3VBSNLHANCNFSM5RQWXRRA . You are receiving this because you were mentioned.Message ID: @.***> Hi @aletrovato, First, thank you for raising this issue. I have reached out to the product team to see about adding convenience methods to Office Scripts to make working with dates easier and more intuitive. I wanted to follow up about your two suggestions. For the NOW function link, I had originally linked to that page because it contained the best description of the Excel date serial number. Is that link confusing? Should we just remove it? For the sample not working, I could not reproduce that issue. Could you please provide a screenshot of the error or a copy of the text that is produced? Thank you, Hi Alex! Your idea to develop a more practical way to work with dates is very welcome. When I tested today some scripts for date handling I ran into the problem of using the Date library (I didn't know it yet). The conversion of DataSerial types to standard JavaScript is confusing, perhaps it could also be an item in the Office Script help pages. I have written articles, produced a course and several classes for this on Youtube. I am trying to raise the flag and train users here to migrate to WEB and it would be of great help. Suggestion 1 NOW - The redirect got confusing, if I may suggest, it would be interesting to remove the link or include the information that the NOW page illustrates how the date will look after formatting. Suggestion 2: Solved. In the morning, when testing the code, getValue() was not recognized as a valid statement.Β  If I can help with anything, I am at your disposal. Thank you very much!
2025-04-01T06:37:20.812316
2020-04-27T15:21:20
607635831
{ "authors": [ "jasonjoh", "jmussman" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2187", "repo": "OfficeDev/outlook-dev-docs", "url": "https://github.com/OfficeDev/outlook-dev-docs/issues/862" }
gharchive/issue
Chrome Actionable message support fails for web-based Outlook in Chrome, but not in Firefox or Safari. It's well known that much Microsoft client-side code has problems with Chrome (not sure why, my own code works fine), but there should be an asterisk on this page for that. Document Details ⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking. ID: fc039d3b-b1e6-79ae-3b83-6e90d7d8af94 Version Independent ID: f72d624f-36b9-bab1-9c14-e9fc1e81082b Content: What are actionable messages in Office 365? - Outlook Developer Content Source: docs/actionable-messages/index.md Product: outlook Technology: o365-connectors GitHub Login: @jasonjoh Microsoft Alias: jasonjoh @jmussman Actionable messages are supported in Chrome. Do you have a repro of your failure you can share with us? You are correct. I rebooted since I posted that, and when I went back to take a screen shot for you it’s magically working. My bad then, sorry 😊 I guess welcome to the wonderful world of so many things impact so many others now days… I have no idea what could have caused that. From: Jason Johnston<EMAIL_ADDRESS>Reply-To: OfficeDev/outlook-dev-docs<EMAIL_ADDRESS>Date: Monday, April 27, 2020 at 6:57 PM To: OfficeDev/outlook-dev-docs<EMAIL_ADDRESS>Cc: Joel Mussman<EMAIL_ADDRESS>Mention<EMAIL_ADDRESS>Subject: Re: [OfficeDev/outlook-dev-docs] Chrome (#862) @jmussmanhttps://github.com/jmussman Actionable messages are supported in Chrome. Do you have a repro of your failure you can share with us? β€” You are receiving this because you were mentioned. Reply to this email directly, view it on GitHubhttps://github.com/OfficeDev/outlook-dev-docs/issues/862#issuecomment-620278257, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ABPCPO2CVWLJOOCUFHRV423ROYETVANCNFSM4MSAR7RA.
2025-04-01T06:37:20.829600
2024-02-21T15:07:50
2146988218
{ "authors": [ "OkuyanBoga" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2188", "repo": "OkuyanBoga/hc-qiskit-machine-learning", "url": "https://github.com/OkuyanBoga/hc-qiskit-machine-learning/issues/11" }
gharchive/issue
Update tutorials for deprecated circuit draw style #725 Technical discussion for #725 Update tutorials for deprecated circuit draw style #725
2025-04-01T06:37:20.863365
2017-04-25T07:53:57
224051909
{ "authors": [ "david-driscoll", "mickaelistria" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2191", "repo": "OmniSharp/csharp-language-server-protocol", "url": "https://github.com/OmniSharp/csharp-language-server-protocol/issues/1" }
gharchive/issue
Status of this LS compared to omnisharp-node-client? Hi @david-driscoll . Glad to see this moving forward and the node-client being dropped! I'd like to use it ASAP as a replacement for omnisharp-node-client in the Eclipse IDE integration. However, before doing it, I'd like to know whether it's already better than the omnisharp-node-client or if you think it's too early to switch (some features may be missing with the LS compared to previous one?). Thanks in advance @mickaelistria to early to switch for sure... this is the infrastructure for the work. I'm going to packagify this and use it in OmniSharp. The initial go forward plan is to add a new switch to OmniSharp --lsp that will kick it over into "lsp mode". I hope to get that done over the next month or so (gotta just find a day to crush out some code really). Then for third parties you'll just have to download the correct release from the https://github.com/OmniSharp/omnisharp-roslyn repository, and be able to use LSP natively. Work has started! https://github.com/OmniSharp/omnisharp-roslyn/pull/969
2025-04-01T06:37:20.910565
2020-12-04T18:56:59
757328697
{ "authors": [ "Elringus", "NTaylorMullen", "Skleni", "TylerLeonhardt", "anthony-c-martin", "david-driscoll", "rynowak", "s-KaiNet" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2192", "repo": "OmniSharp/csharp-language-server-protocol", "url": "https://github.com/OmniSharp/csharp-language-server-protocol/issues/456" }
gharchive/issue
[info request] Hosting the language server code in Blazor? We're using this library to provide LSP support for Bicep. We have an browser demo at https://aka.ms/bicepdemo which calls into the compiler code directly using Blazor/WASM, using Microsoft.JSInterop to compile, emit diagnostics, get semantic tokens, and pass the results back to the Monaco editor, but we're not using the language server for this - instead we've created a few custom functions for it. The monaco-languageclient library can be used to hook monaco up to a language server which would provide much of the functionality that VSCode offers in a browser. It would be extremely cool to be able to simply use the LSP in a browser without the need for a back-end server. I'm curious as to whether anyone has tried to run this server code via Blazor before. I've been experimenting with it and am able to get the initialize request/response negotiation to take place, but I don't see the client/registerCapability request come through from the server. I suspect there may be some sort of message pump that needs to run, but am not at all familiar with the Reactive library that's being used. Any pointers that you can give me would be awesome! Here's an example of the code changes I've been experimenting with to hook this up to monaco: main...antmarti/experiment/monaco_lsp I was thinking about this a week or two ago, in theory things should "just work" because the language server is just a .NET Standard library. This sounds fun, so I'll take a look at your code and see if I can get it running. I'm curious as to whether anyone has tried to run this server code via Blazor before. I've been experimenting with it and am able to get the initialize request/response negotiation to take place, but I don't see the client/registerCapability request come through from the server. I suspect there may be some sort of message pump that needs to run, but am not at all familiar with the Reactive library that's being used. Any pointers that you can give me would be awesome! If there is a message pump at play one of the challenges might be that pump using a dedicated thread. I don't think Blazor WASM supports threads yet. Certain BCL methods that interact with threading will "spin" and fry your CPU :) Thanks, I was thinking about this a week or two ago, in theory things should "just work" because the language server is just a .NET Standard library. This sounds fun, so I'll take a look at your code and see if I can get it running. Thanks! If you need any pointers in getting it running, let me know. It should just work if you clone the repo and run: cd src/playground npm i npm start I've stuck some haphazard logging in which should get written out to the browser console. For input the process scheduler runs on the thread pool, which I think should be fine. https://github.com/OmniSharp/csharp-language-server-protocol/blob/master/src/JsonRpc/InputHandler.cs#L86 For output however... I think it by default runs a dedicated thread. https://github.com/OmniSharp/csharp-language-server-protocol/blob/6371386ee90dde8a1c419f3dc8fc94e0c1bc35f0/src/JsonRpc/OutputHandler.cs#L58 Okay here's a possible quick fix. When setting up the server... try this. IScheduler should be System.Reactive.Scheduling.IScheduler (or something like that) options.Services.AddSingleton<IScheduler>(TaskPoolScheduler.Default); I theory that should kick the output handler to use the IScheduler provided in the container based on how I think DryIoc will pick constructors. @ryanbrandenburg @NTaylorMullen @TylerLeonhardt thoughts, should I just move to use the task pool scheduler for handing input/output? At the time a dedicated thread "made sense" but honestly it probably doesn't matter. Input is already on the task pool and working fine. Output isn't however it does ensure ordering, so we shouldn't there shouldn't really be any big problems. @ryanbrandenburg @NTaylorMullen @TylerLeonhardt thoughts, should I just move to use the task pool scheduler for handing input/output? At the time a dedicated thread "made sense" but honestly it probably doesn't matter. Having a dedicated thread has been risky because if something doesn't ConfigureAwait(false) and blocks you're doomed. We've actually encountered that issue once or twice in VS (as I'm sure you recall) so relying on the task pool scheduler doesn't sound awful. Are there any other drawbacks? All that said for extra background info, we run Razor's language server in-proc in VS today which I presume from the quick glance at this thread similar types of things are trying to be acheived. Here's where we create our own abstraction to start the spinup of the O# framework bits in VS: https://github.com/dotnet/aspnetcore-tooling/blob/feb060660bf14c9da3f284a72fe5f86390d3ab65/src/Razor/src/Microsoft.VisualStudio.LanguageServerClient.Razor/RazorLanguageServerClient.cs#L126-L129 And here's our actual abstraction that can rely on in-proc or out of proc streams: https://github.com/dotnet/aspnetcore-tooling/blob/feb060660bf14c9da3f284a72fe5f86390d3ab65/src/Razor/src/Microsoft.AspNetCore.Razor.LanguageServer/RazorLanguageServer.cs#L72-L75 Okay here's a possible quick fix. When setting up the server... try this. IScheduler should be System.Reactive.Scheduling.IScheduler (or something like that) options.Services.AddSingleton<IScheduler>(TaskPoolScheduler.Default); I gave this a go, but didn't see any observable difference in behavior. I did notice that the Reactive.Wasm library I'm trying to use to replace the schedule doesn't appear to be doing what it's meant to in .NET 5 - in particular these checks no longer seem to work: https://github.com/reactiveui/Reactive.Wasm/blob/a226dc0bb4f010c248568eb67b0b8e5b768358f5/src/System.Reactive.Wasm/Internal/WasmScheduler.cs#L136-L137 https://github.com/reactiveui/Reactive.Wasm/blob/a226dc0bb4f010c248568eb67b0b8e5b768358f5/src/System.Reactive.Wasm/Internal/WasmPlatformEnlightenmentProvider.cs#L27-L28 In theory if they were working, I should be able to do: options.Services.AddSingleton<IScheduler>(WasmScheduler.Default); I'll see if I can fix up the above checks locally and get that working. All that said for extra background info, we run Razor's language server in-proc in VS today which I presume from the quick glance at this thread similar types of things are trying to be acheived. Thanks for the pointers! We have the language server running as a standalone exe, which we use for VSCode integration, but as an experiment, I'm trying to see if we can also host the language server fully in a web browser, using Blazor/WASM without a backend - I think that's where the complexity is mostly coming from. Is that something your team has attempted by any chance? I think that's where the complexity is mostly coming from. Is that something your team has attempted by any chance? Ah, ya I can definitely imagine that being difficult πŸ˜„. No we haven't tried that but I can just imagine how the threading models may make things more difficult in addition to things like file watchers @david-driscoll I just got an end-to-end working with a very hacky change here: https://github.com/OmniSharp/csharp-language-server-protocol/blob/6371386ee90dde8a1c419f3dc8fc94e0c1bc35f0/src/JsonRpc/OutputHandler.cs#L67 Instead of adding the message to the queue, I just sent it directly with: ProcessOutputStream(value, CancellationToken.None).Wait(); So I think that definitely confirms that it's something to do with the scheduler. Interestingly, I noticed that in the version of the language server we're using (0.18.3) it is using TaskPoolScheduler.Default rather than EventLoopScheduler. I'm going to try and get a more solid PoC together by replacing IOutputHandler in the IoC container. interesting! options.Services.AddSingleton<IScheduler>(ImmediateScheduler.Instance); also appears to work. @anthony-c-martin are you on slack or msteams? So I'm running into an error Request client/registerCapability failed with message: i.languages.registerDocumentSemanticTokensProvider is not a function. Seems the monaco editor doesn't support semantic tokenization yet. However, disabling that things seem to work. Here's my branch for you reference from: https://github.com/Azure/bicep/compare/Azure:antmarti/experiment/monaco_lsp...david-driscoll:davidd/experiment/monaco_lsp Couple notes: I was building locally with the latest version of the library (0.19.0-beta.1) so the C# changes are the changes required based on the breaking changes I've documented. Also I was able to simplify the interop a little bit by using StreamMessageReader/StreamMessageWriter and a Duplex stream. This writes all the expected header information, so you don't have to serialize on the Blazor side, instead you just write the bytes directly into the pipe. Sending from the server to client also happens similarly. I've created this PR #458 so we can configure the schedulers specifically. So I'm running into an error Request client/registerCapability failed with message: i.languages.registerDocumentSemanticTokensProvider is not a function. Seems the monaco editor doesn't support semantic tokenization yet. However, disabling that things seem to work. Here's my branch for you reference from: Azure/bicep@Azure:antmarti/experiment/monaco_lsp...david-driscoll:davidd/experiment/monaco_lsp Couple notes: I was building locally with the latest version of the library (0.19.0-beta.1) so the C# changes are the changes required based on the breaking changes I've documented. Also I was able to simplify the interop a little bit by using StreamMessageReader/StreamMessageWriter and a Duplex stream. This writes all the expected header information, so you don't have to serialize on the Blazor side, instead you just write the bytes directly into the pipe. Sending from the server to client also happens similarly. This is AMAZING, thank you so much for your help! I ran into the same issue with the language client - looks like semantic support has only been added to a preview version, and that they haven't yet picked up the latest LSP spec. For now, since we've already implemented our own semantic token handler anyway, I've reverted back to using this for now until the language client has actual support for it. I've pushed a demo of this here: https://bicepdemo.z22.web.core.windows.net/experiment/lsp/index.html This has me thinking of making a Blazor Component that uses the monaco editor... but to try to make as much as possible of it actually live in C# and use the LanguageClient for interacting with it.... Other than the annoying part of converting the monaco api into C#... ugh. Right now I don't think I have the bandwidth to tie monaco and blazor together. I might spike something out next weekend. I looked at https://github.com/microsoft/monaco-editor/blob/master/monaco.d.ts and while I'm sure I could... that's a lot of code to keep in sync, so I would want to build out some sort of tool to integrate the two together. I found this project, and posted an issue there https://github.com/canhorn/EventHorizon.Blazor.TypeScript.Interop.Generator/issues/31 to see what might be needed to support generation interop with monaco.d.ts as I'm just not prepared for the maintenance that would entail. In the meantime there is recent activity on https://github.com/TypeFox/monaco-languageclient updating it to the latest version (that would include semantic tokens), you might be able to pin to the latest master branch and see if that works (I have not tried). @anthony-c-martin are you on slack or msteams? I'm on Teams -<EMAIL_ADDRESS> Shoot now I want to run the PowerShell language server in Blazor! I think you can, you'll just have to do something similar to the bicep solution using monaco + monaco-languageclient, it totally works, there might be some issues if you use the file system APIs but those can always be fixed. @TylerLeonhardt feel free to reach out if you'd like any pointers for the Bicep code! I'm skeptical the PowerShell API will "just work" in Blazor WASM but worth a shot. @anthony-c-martin how did you "start the language server" in Blazor WASM? I'd love to take a peak at how the language server is hooked up to Monaco Editor. For context, I've used the Monaco-languageclient before, but only their stdio option where the language server was running in a separate process on the machine. I'm skeptical the PowerShell API will "just work" in Blazor WASM but worth a shot. @anthony-c-martin how did you "start the language server" in Blazor WASM? I'd love to take a peak at how the language server is hooked up to Monaco Editor. [credit goes to @david-driscoll for a lot of this code] Here's where the server is being initialized: https://github.com/Azure/bicep/blob/16d7eb7fd5a92dadf6704f1c49e9246cf52b4da9/src/Bicep.Wasm/Interop.cs#L43-L58 The Server class is our own, but is really a thin wrapper around the Omnisharp Server class. The important pieces here are initializing the input/output pipes, and overriding the scheduler with ImmediateScheduler.Instance. Here's the C# method that the JS code invokes to send data from client to server: https://github.com/Azure/bicep/blob/16d7eb7fd5a92dadf6704f1c49e9246cf52b4da9/src/Bicep.Wasm/Interop.cs#L62 Here's where the C# code invokes the JS code to send data from server to client: https://github.com/Azure/bicep/blob/16d7eb7fd5a92dadf6704f1c49e9246cf52b4da9/src/Bicep.Wasm/Interop.cs#L78 Here's the JS code to setup the send/receive with the server: https://github.com/Azure/bicep/blob/16d7eb7fd5a92dadf6704f1c49e9246cf52b4da9/src/playground/src/helpers/lspInterop.ts#L24-L34 On startup I'm initializing the Blazor code from JS and setting the interop variable which can be used to invoke Blazor code with the following: https://github.com/Azure/bicep/blob/16d7eb7fd5a92dadf6704f1c49e9246cf52b4da9/src/playground/src/helpers/lspInterop.ts#L5-L14 If you follow through the TS code, you should be able to see how the above is hooked into monaco-languageclient. I'm probably going to try and refine this code at some point to see if I can clean up the use of globals, and also to see if I can use a webworker to run the Blazor code. Right now I don't think I have the bandwidth to tie monaco and blazor together. I might spike something out next weekend. I looked at https://github.com/microsoft/monaco-editor/blob/master/monaco.d.ts and while I'm sure I could... that's a lot of code to keep in sync, so I would want to build out some sort of tool to integrate the two together. Out of interest, what are the benefits of implementing the translation layer between LSP & monaco's "custom LSP" in C# vs relying on monaco-languageclient to do it? I quite like the clean separation of having the TS code handle the translation and communicating with the C# code via LSP. I just think it would be pretty cool to have a fully featured wrapper for monaco from the C# side. The added extra would make it it easier to consume using the client. Probably because then @david-driscoll could guarantee that the monaco language client was up-to-date on the LSP spec. Going to pin this issue for any passers by as it is truly a cool feature. FWIW I think we need one of https://github.com/dotnet/aspnetcore/issues/17730 or https://github.com/dotnet/aspnetcore/issues/5475 to really unlock the power of this, because at the moment synchronous dotnet code locks up the UI thread, which feels a little janky when typing. There's also this project which I haven't really investigated that might work as a stopgap: https://github.com/Tewr/BlazorWorker I think a web worker would be perfect. Your UI (TypeScript) starts the worker, and you interop with the worker using postmessage. The worker then just has to interop with the language server. Thank you guys, you helped me a lot to understand some ideas. I'm trying to build a small POC on blazor and monaco based C# code editor with code completion. However, I cannot get code completion to work. What I've done: in JS created monaco-editor and configured monaco-languageclient (in the same way as Bicep's playground) in wasm created an interop class: public class Interop { private LanguageServer languageServer; private readonly IJSRuntime jsRuntime; private readonly PipeWriter inputWriter; private readonly PipeReader outputReader; public Interop(IJSRuntime jsRuntime) { this.jsRuntime = jsRuntime; var inputPipe = new Pipe(); var outputPipe = new Pipe(); inputWriter = inputPipe.Writer; outputReader = outputPipe.Reader; languageServer = LanguageServer.PreInit(opts => { opts.WithInput(inputPipe.Reader); opts.WithOutput(outputPipe.Writer); opts.Services.AddSingleton<IScheduler>(ImmediateScheduler.Instance); }); Task.Run(() => RunAsync(CancellationToken.None)); Task.Run(() => ProcessInputStreamAsync()); } public async Task RunAsync(CancellationToken cancellationToken) { await languageServer.Initialize(cancellationToken); await languageServer.WaitForExit; } [JSInvokable] public async Task SendLspDataAsync(string jsonContent) { var cancelToken = CancellationToken.None; Console.WriteLine("jsonContent"); Console.WriteLine(jsonContent); await inputWriter.WriteAsync(Encoding.UTF8.GetBytes(jsonContent)).ConfigureAwait(false); } private async Task ProcessInputStreamAsync() { do { var result = await outputReader.ReadAsync(CancellationToken.None).ConfigureAwait(false); var buffer = result.Buffer; Console.WriteLine("ProcessInputStreamAsync"); await jsRuntime.InvokeVoidAsync("ReceiveLspData", Encoding.UTF8.GetString(buffer.Slice(buffer.Start, buffer.End))); outputReader.AdvanceTo(buffer.End, buffer.End); // Stop reading if there's no more data coming. if (result.IsCompleted && buffer.IsEmpty) { break; } // TODO: Add cancellation token } while (!CancellationToken.None.IsCancellationRequested); } } Code completion doesn't work, because I haven't registered CodeCompletionHandler. I don't understand which one to use, because in Bicep you use a custom completion handler, in my POC I would like to use O# completion handler. Do you have any hints on how to implement it? I've made a solution, that compiles C# project into single-file UMD library: https://github.com/Elringus/DotNetJS Tried to use the server with it, but not sure how to deal with input/output. Console.STD won't work, obviously. Can we somehow run the server via websocket? Tried to use the server with it, but not sure how to deal with input/output. Console.STD won't work, obviously. Can we somehow run the server via websocket? Nice, I'll check that library out! Here's how I've been doing things in my experimental branch - using a simple send/receive method to pass JSONRPC back and forth from JS <-> C#: C#: https://github.com/Azure/bicep/blob/4a4d193eff0492043ff350419c8c9693ad6d63b6/src/Bicep.Wasm/LspWorker.cs JS: https://github.com/Azure/bicep/blob/4a4d193eff0492043ff350419c8c9693ad6d63b6/src/playground/src/helpers/lspInterop.ts Not the most elegant/performant, but it works well enough for now. Being able to have client-side Blazor host a websocket would make this a lot nicer. @Elringus I'm currently trying to use your library to get our OmniSharp-based Language Server to run in an VS Code Web extension. This sounds very similar to what you want to achieve. May I ask if you already managed to get that working? My current status is that I can run the language server in a Blazor project (thanks to the information in this thread), but in the web extension the server never finishes initialization. @Skleni I've switched to Microsoft's reference LSP implementation in JS (https://github.com/microsoft/vscode-languageserver-node), while reusing the existing language-specific C# code via DotNetJS: β€” this way we can get up-to-date LSP implementation and native webworker transport layer out of the box, while keeping all the handlers logic in C#. Regarding VS Code, there were 2 issues with this workflow, but they're both solved in insiders stream now and should become available in the main stream in February: https://github.com/microsoft/vscode/issues/138413 https://github.com/microsoft/vscode/issues/138780
2025-04-01T06:37:20.913229
2015-04-24T17:28:19
70746878
{ "authors": [ "david-driscoll", "nosami" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2193", "repo": "OmniSharp/omnisharp-atom", "url": "https://github.com/OmniSharp/omnisharp-atom/pull/142" }
gharchive/pull-request
Fixed recursive require() between dock-view and omnisharp-atom, fixed a ... ...few (annoying) compiler errors. Fixes #139. Works a treat!
2025-04-01T06:37:20.916506
2021-12-21T03:08:38
1085389109
{ "authors": [ "dmgonch" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2194", "repo": "OmniSharp/omnisharp-roslyn", "url": "https://github.com/OmniSharp/omnisharp-roslyn/pull/2314" }
gharchive/pull-request
Add MSBuild project to solution and apply the change to Roslyn workspace as a unit I often hit 'Failed to add project to workspace: ' error in OmniSharp log when opening one of internal projects. I only see this issue on Linux (Ubuntu 20.04) and not on Windows. I always have omnisharp.enableMsBuildLoadProjectsOnDemand set to "true". The problem occurs often but not all the time. In both success and failure cases I see the following sequence of events (shortened to highlight only important parts): [info]: OmniSharp.MSBuild.ProjectManager Queue project update for '/home/<path_to_project>/ProjectName.csproj' [info]: OmniSharp.MSBuild.ProjectManager Loading project: /home/<path_to_project>/ProjectName.csproj [dbug]: OmniSharp.Stdio.Host ************ Request ************ { "Type": "request", "Seq": 7, "Command": "/filesChanged", "Arguments": [ { "FileName": "/home/<path_to_project>/obj/Debug/net6.0/linux-x64/ProjectName.AssemblyInfo.cs", "changeType": "Change" } ] } [dbug]: OmniSharp.Roslyn.BufferManager Adding transient file for /home/<path_to_project>/obj/Debug/net6.0/linux-x64/ProjectName.AssemblyInfo.cs [info]: OmniSharp.MSBuild.ProjectManager Successfully loaded project file '/home/<path_to_project>/ProjectName.csproj'. [info]: OmniSharp.MSBuild.ProjectManager Adding project '/home/<path_to_project>/ProjectName.csproj' In the case of the failure though the sequence is followed by: [fail]: OmniSharp.MSBuild.ProjectManager Failed to add project to workspace: '/home/<path_to_project>/ProjectName.csproj' ... [fail]: OmniSharp.MSBuild.ProjectManager Could not locate project in workspace: /home/<path_to_project>/ProjectName.csproj The documentation for "bool Workspace.TryApplyChanges(Solution newSolution)" states: β€œβ€¦ The specified solution must be one that originated from this workspace. If it is not, or the workspace has been updated since the solution was obtained from the workspace, then this method returns false …" ProjectManager.AddProject clearly passes the solution that is originated from the same workspace. Therefore, the workspace somehow gets changed b/w adding the project to the solution and applying the changes to the workspace. The theory is that the file change notification received for ProjectName.AssemblyInfo.cs causes the workspace change in ProjectManager.OnDirectoryFileChanged. Thus, the lock added around these two places. My testing shows that the change does address the issue I'm seeing. @filipw, @JoeRobich, appreciate if you could take a look.
2025-04-01T06:37:20.918708
2016-03-23T23:58:47
143109612
{ "authors": [ "DustinCampbell", "agocke" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2195", "repo": "OmniSharp/omnisharp-vscode", "url": "https://github.com/OmniSharp/omnisharp-vscode/issues/121" }
gharchive/issue
Intellisense appears after typing an open or close brace and autoselects "Action" This causes subsequent return presses on the keyboard to insert the text "Action" instead of moving to a new line. @agocke, are you still seeing this problem? I've been unable to reproduce it. @DustinCampbell I think this was a bad interaction between extensions. I'll try and repro and re-open if I find anything. thanks! No rush here. The idea that two extensions were interacting badly sounds like a good theory to me.
2025-04-01T06:37:20.956587
2023-02-16T15:25:25
1587885865
{ "authors": [ "JCrew0", "avirallariva", "nicolasiscoding", "patrick-rodgers" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2196", "repo": "OneDrive/samples", "url": "https://github.com/OneDrive/samples/issues/30" }
gharchive/issue
Document Sets are not visible when filtering on file extension in typesAndSources in configuration schema Issue: A user is required to pick a file of certain extension in the workflow. We pass the extension in the 'filters' parameter in 'typesAndSources' in the Picker Configuration Schema. The filtering seems to work fine. I receive the files with the required extension and the folders in the Sharepoint site. The issue is that the Document Sets are not visible. This hinders the user to pick files of similar extension in the Document Sets. The only possible options we can put in filters is string array with values like: - photo - video - audio - folder - file - extensions prefixed with '.' Possible Solutions: Treat the shared documents as folders Maybe add another parameter to filter based on the 2 content types, i.e. document sets and folder. Add Document Sets in the filters options Document Reference: https://learn.microsoft.com/en-us/onedrive/developer/controls/file-pickers/v8-schema?view=odsp-graph-online We will review this request with the engineering team. Hi @patrick-rodgers @avirallariva, what was the outcome of this? Hey @nicolasiscoding, did you ever get updates on this outside of this thread? If not, I can work with the team to provide some. Thanks!
2025-04-01T06:37:21.336007
2023-01-27T10:55:51
1559554478
{ "authors": [ "BlackMesa123", "ghost1372" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2197", "repo": "OneUIProject/oneui-design", "url": "https://github.com/OneUIProject/oneui-design/issues/21" }
gharchive/issue
How to update ProgressDialog? i want to update ProgressDialog Progress, this is my codes, where is wrong? progress dialog is not update and stay in 0 private var job: Job = Job() override val coroutineContext: CoroutineContext get() = IO + job fun getCirculars(){ swipRefresh.isRefreshing = true itemsData.clear() val dialog = ProgressDialog(mContext) dialog.setProgressStyle(ProgressDialog.STYLE_HORIZONTAL) dialog.setIndeterminate(true) dialog.setCancelable(false) dialog.setCanceledOnTouchOutside(false) dialog.max = 100 dialog.isIndeterminate = true dialog.show() launch { val operation = async { val doc: Document = Jsoup.connect(url) .timeout(0) .maxBodySize(0) .ignoreHttpErrors(true) .sslSocketFactory(CommonHelper.trustServer()) .get() val table: Elements = doc.select("table[class=\"table table-striped table-hover\"]") for (myTable in table) { val rows: Elements = myTable.select("tr") withContext(Dispatchers.Main){ dialog.isIndeterminate = false } for (i in 1 until rows.size) { withContext(Dispatchers.Main){ dialog.progress = (i / rows.size) * 100 } } } }.await() withContext(Dispatchers.Main) { // update UI dialog.dismiss() swipRefresh.isRefreshing = false } } } If you still have doubts about the lib's usage please join our official Telegram Group and ask us there. I'd prefer to keep the repository Issues page only for actual lib issues/bugs rather than for generic questions about the usage of the libs..
2025-04-01T06:37:21.354641
2024-04-10T14:28:16
2235741852
{ "authors": [ "cozvtieg9" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2198", "repo": "Onix-Systems/onix-flutter-project-generator", "url": "https://github.com/Onix-Systems/onix-flutter-project-generator/issues/58" }
gharchive/issue
Support Flavors for MacOs Is your feature request related to a problem? Please describe. When creating project with flavors, they configured for Android and OS only. There no MacOS cofiguration in flavorizr. Describe the solution you'd like Add MacOS to flavorizr configuration Planned in version 1.1.1(26)
2025-04-01T06:37:21.406404
2024-10-11T09:01:22
2580847102
{ "authors": [ "Shariq2003", "jeevan10017" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2199", "repo": "Open-Code-Crafters/FitFlex", "url": "https://github.com/Open-Code-Crafters/FitFlex/issues/165" }
gharchive/issue
Addition of Persistent Storage for UserData | ChatBot Enhancement | Optimisation of Recurrence Of Questioning I want to add the Persistent Storage for the optimisation of UserData Flow for ChatBot Flow. In this, I am going to implement the sessionStorage / localStorage for the userData storing that results in optimization of ChatBot Flow. Contributor Information Please check if you are a contributor from: [X] GSSoC-ext [X] Hacktoberfest @Shariq2003
2025-04-01T06:37:21.423581
2023-08-20T23:37:33
1858308682
{ "authors": [ "ReimarBauer", "joernu76" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2200", "repo": "Open-MSS/MSS", "url": "https://github.com/Open-MSS/MSS/issues/1915" }
gharchive/issue
TimeOut errors on bad internet connection/strange firewall We cannot connect to the MSColab Server on a guest WiFi with very long connection times (>2s), which may or may not be caused by a deep-inspecting firewall. The timeout should be configurable to provide a work around. Also, the current timeouts are defined as "magic number" in a wide range of places and the values should be collected to a single instance (as mentioned above, defined in the json configuration). If your ISP is throttling certain types of traffic or has poor peering with other networks, VPN will mask the traffic and the result is faster. https://www.cloudwards.net/vpn-internet-speed/#:~:text=Most of the time%2C you,your connection is capable of
2025-04-01T06:37:21.478494
2024-03-22T07:59:24
2201912507
{ "authors": [ "NanoCode012", "anttttti", "fozziethebeat" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2201", "repo": "OpenAccess-AI-Collective/axolotl", "url": "https://github.com/OpenAccess-AI-Collective/axolotl/issues/1430" }
gharchive/issue
Loading local Dataset saved via save_to_disk fails Please check that this issue hasn't been reported before. [X] I searched previous Bug Reports didn't find any similar reports. Expected Behavior I create some local dataset that I saved via calling ds.save_to_disk('some_path') If I then configure Axolotl to load a dataset from save_path such as datasets: - path: tmp_dataset ... I would expect it to load much like loading from a remote huggingface repo. Current behaviour The loaded dataset includes the contents of state.json rather than the actual dataset and everything fails. Steps to reproduce Save some dataset to disk via save_to_disk Configure Axolotl to read from that dataset See Axolotl crash Config yaml datasets: - path: local_dataset_saved_via_save_to_disk type: field_instruction: prompts field_output: responses field_system: system format: '[INST] {instruction} [/INST]' no_input_format: '[INST] {instruction} [/INST]' system_prompt: '' Possible solution In src/axolotl/utils/data.py I see that for local directory paths, datasets are loaded via load_dataset. load_from_disk seems to be the new preferred solution (this is also a comment). I manually hacked this in and everything worked without a problem. Would switching that to load_from_disk break anything? It seems like a reasonable quick fix (which I can do) Which Operating Systems are you using? [X] Linux [ ] macOS [ ] Windows Python Version 3.10 axolotl branch-commit main/4e69aa4 Acknowledgements [X] My issue title is concise, descriptive, and in title casing. [X] I have searched the existing issues to make sure this bug has not been reported yet. [X] I am using the latest version of axolotl. [X] I have provided enough information for the maintainers to reproduce and diagnose the issue. I've had to work around this for a while. Please use load_from_disk instead. The workaround is to manually delete "state.json" from the dataset directory, and axolotl will then read it. But other HF-compatible tools expect "state.json" so this is a hassle I do sometimes hate HuggingFace libraries. I assumed they'd solve cases like this but thems the beans. I think the right fix is to use the current loading behavior when data_files are specified but backoff to load_from_disk when a local directory is specified without data_files. That's what my PR does and it works at least for my case. load_from_disk is pretty limited tho it can't do much else but read from the one directory. Thanks for PR @fozziethebeat . Sorry I didn't get time to properly review. This got in faster than I expected! Thanks for the suggestions on testing, it made everything smoother~
2025-04-01T06:37:21.488610
2024-04-11T02:36:19
2236799445
{ "authors": [ "NanoCode012", "dsesclei", "winglian" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2202", "repo": "OpenAccess-AI-Collective/axolotl", "url": "https://github.com/OpenAccess-AI-Collective/axolotl/pull/1516" }
gharchive/pull-request
Switch to parallel FFD bin packing algorithm (closes #1492) Description Replace the existing sample packing algorithm with a parallel implementation of first-fit-decreasing. Motivation and Context I noticed recently that we could get denser sample packing with a different algorithm. Looking into it more, FFD performs just as well and is much faster than the heuristic I had πŸ˜…. We can run FFD in parallel without losing much performance by packing samples in groups rather than all at once. On an i9-14900k, it takes 2.2s to pack 1M samples with 99.7% efficiency (current multipack.py is 91.7% in 0.32s.) I removed the length estimates around packing in favor of just counting the batches, but let me know if I should add that back in. Two new config options are added: sample_packing_group_size controls the the number of samples packed by each process, and sample_packing_bin_size sets the number of samples that can be placed in one pack (may need to be increased for large context lengths.) How has this been tested? Tests have been updated to verify that packing is correct. Training appears to run the same, just with fewer steps. It seems reasonable that sorting the items in FFD would interfere with shuffling between epochs, but I haven't been able to find any evidence of that being the case. Testing against a few similarity metrics shows that even when we do the packing at once in one group, shuffling still generates a mostly new set of packs. Screenshots Some performance checks below for 1M items. I removed the length estimates around packing in favor of just counting the batches, but let me know if I should add that back in. I need to do some checking, but the estimates exist due to different processes getting different splits of data, so the actual count of packed samples can vary from process to process. When this happens, you get one process thinking it needs to run another step, but another process thinking it's done and they get out of sync. The estimate was the most sane way I could come up with having each process come up with a deterministic length. I'm open to other ideas to working around this. Could we generate all the packs, and then evenly split those up (like in the updated multipack.py)? I think each rank should then get an exact number of batches and stay in sync. Could we generate all the packs, and then evenly split those up (like in the updated multipack.py)? I think each rank should then get an exact number of batches and stay in sync. Perhaps we could do something like dispatch_batches=True to only run the packing on rank 0. I'm not 100% certain of the implications though Hey, this is very interesting. Should there be some full run comparisons to make sure that there is no loss in performance? Perhaps we could do something like dispatch_batches=True to only run the packing on rank 0. I'm not 100% certain of the implications though Gotcha, for now I'll keep this PR simple by leaving the packing estimates in. Ready for another look. Hey, this is very interesting. Should there be some full run comparisons to make sure that there is no loss in performance? Yeah definitely, once the code is greenlit/finalized I'll rent an instance to test it in a distributed setup. Hey @dsesclei we cherry picked and merged your fixes in #1619. Thanks! Would love to give you a shoutout if you're on twitter or discord and could share your handle. thanks! Thanks for getting this in Wing! No handles to give, but I appreciate it Thanks @dsesclei, I ended up having to revert the change b/c the loss was off by an order of magnitude. I need to dig into what the multipack sampler is outputting another time to see if there is something obvious that it is doing differently Oh gotcha, I'll look into it
2025-04-01T06:37:21.493254
2015-04-01T06:31:34
65616351
{ "authors": [ "Flavien", "hackable" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2203", "repo": "OpenAssets/colorcore", "url": "https://github.com/OpenAssets/colorcore/issues/8" }
gharchive/issue
List transactions It should show all transactions concerned with an address and balance being in open assets added or deducted for example https://www.coinprism.info/address/akV48Tav8nWZZgPWgHzXuQgMdM5jVJ53ros or get open assets details of a transaction based on get txid I may be wrong, but I don't think Bitcoind has an RPC call that does that. Though we could implement that for the chain.com provider. It would be very helpful if its implementable via chain.com api.
2025-04-01T06:37:21.497531
2024-05-20T13:11:28
2310107657
{ "authors": [ "SamuelHassine", "guillaumejparis" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2204", "repo": "OpenBAS-Platform/injectors-python", "url": "https://github.com/OpenBAS-Platform/injectors-python/issues/21" }
gharchive/issue
Cannot launch a HTTP inject HTTP injector do not work correctly critical : seen with product team Issue transferred to the proper repository.
2025-04-01T06:37:21.527857
2018-10-31T21:08:50
376161204
{ "authors": [ "cpacia", "rmisio" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2205", "repo": "OpenBazaar/openbazaar-go", "url": "https://github.com/OpenBazaar/openbazaar-go/issues/1272" }
gharchive/issue
Add the 'coin' to the 'payment' notification so that the coins icon can be displayed. Otherwise, it's just a generic icon, which just looks weird: done
2025-04-01T06:37:21.536346
2019-06-06T19:34:11
453202437
{ "authors": [ "cpacia", "placer14" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2206", "repo": "OpenBazaar/openbazaar-go", "url": "https://github.com/OpenBazaar/openbazaar-go/issues/1626" }
gharchive/issue
IPFS secio handshake patch likely not forward compatible In the code segment below when establishing a new outgoing connection s.remotePeer is set to the peerID of the node we're trying to connect to. Ultimately this will either be an old style (hashed) peerID or a new style inline key depending on the actual ID that is used when we try to make a connection. So far so good. Also peer.IDFromPublicKey is currently programmed to return an old style (hashed) peerID since we set AdvancedEnableInlining to false. OK . Now take a look at the secio handshake... this is their raw code without any modifications: // get peer id actualRemotePeer, err := peer.IDFromPublicKey(s.remote.permanentPubKey) if err != nil { return err } switch s.remotePeer { case actualRemotePeer: // All good. case "": // No peer set. We're accepting a remote connection. s.remotePeer = actualRemotePeer default: // Peer mismatch. Bail. s.insecure.Close() log.Debugf("expected peer %s, got peer %s", s.remotePeer, actualRemotePeer) return ErrWrongPeer } If we ever enable inline keys, old nodes which have not upgraded, when trying to connect to a new style peerID, will pass in a new style ID into s.remotePeer but actualRemotePeer will still be an old style key. Hence s.remotePeer != actualRemotePeer and the handshake will fail. If an upgraded node tries to connect to a non-upgraded peerID then peer.IDFromPublicKey will return a new style peerID while s.remotePeer is an old style key. Hence it will fail to connect. So to make our current release forward compatible with a future release using inline keys I've done: // get peer id actualRemotePeer, err := peer.IDFromPublicKey(s.remote.permanentPubKey) if err != nil { return err } switch s.remotePeer { case actualRemotePeer: // All good. case "": // No peer set. We're accepting a remote connection. s.remotePeer = actualRemotePeer default: pubkeyBytes, err := s.remote.permanentPubKey.Bytes() if err != nil { return err } oldMultihash, err := mh.Sum(pubkeyBytes, mh.SHA2_256, 32) if err != nil { return err } oldStylePeer, err := peer.IDB58Decode(oldMultihash.B58String()) if err != nil { return err } if s.remotePeer != oldStylePeer { // Peer mismatch. Bail. s.insecure.Close() log.Debugf("expected peer %s, got peer %s", s.remotePeer, actualRemotePeer) return ErrWrongPeer } } But looking at it, s.remotePeer would be an inline peerID and oldStylePeer would be an old style key, so I think this is screwed up and would prevent old nodes from connecting to new nodes. I think our current release should be a newStylePeer and compare it to s.remotePeer rather than an oldStylePeer. And then a subsequent release with inline keys should use the code snippet above. (edited) This will likely mean we need to fix this for next release and then push out the timeline for updating to the inline keys until at least enough people upgrade to this next release. Thinking through this problem with annotations around the code: primaryTest, err := peer.IDFromPublicKey(s.remote.permanentPubKey) // *snip* switch s.remotePeer { case primaryTest: // (Primary Case) All good. case "": // Ignore default: // (Backup Case) // preparation and checking *snipped* if s.remotePeer != oldStylePeer { // Failure Case *snip* } } Scenarios: OldNode (hashed) connecting to OldNode (hashed, inline off): remote (hashed) == primaryTest (hashed)... success in Primary Case OldNode (hashed) connecting to NewNode (inline, inline on): remote (hashed) != primaryTest (inline)... remote (hashed) == oldStyle (hashed) success in Backup NewNode (inline) connecting to OldNode (hashed, inline off): remote (inline) != primaryTest (hashed)... remote (inline) != oldStyle (hashed) ...always fails NewNode (inline) connecting to NewNode (inline, inline on) remote (inline) == primaryTest (inline)... success in Primary Case I see the problem you're describing... it seems the backup test is always stuck producing the hashed peerID when we seem to want the backupTest to use the other algo opposite from the primaryTest. I think the way forward for us will be a function ipfs.AlternativeIDFromPublicKey(crypto.PubKey) (peer.ID, error) which looks at the state of AdvancedEnableInlining and applies the opposite algorithm than what IDFromPublicKey applies. And then we can use that function inside of default until we've completed our inline key migration completely. Does that seem reasonable, @cpacia? That seems like a good approach. Better than when I was suggesting.
2025-04-01T06:37:21.539040
2016-08-09T17:46:17
170228125
{ "authors": [ "cpacia", "rmisio" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2207", "repo": "OpenBazaar/openbazaar-go", "url": "https://github.com/OpenBazaar/openbazaar-go/issues/73" }
gharchive/issue
Irregularly, but steadily getting IPNS pinned errors. I don't know what the exact steps to reproduce are, because it doesn't always happen. But, it has been happening enough to multiple developers that I think it requires attention. Anyhow, often times, after wiping your data and starting with a fresh server, POSTs on the profile are failing with a 'not pinned' error: Then, a subsequent fetch returns a 404, whereas a subsequent PUT fails with 'Profile already exists. Use PUT': Once again, i think the crux of the problem is that the Profile is erroneously 404ing, when the file does indeed exists. It shouldn't 404 in this case, otherwise the client has no reliable way of knowing whether to onboard or not and whether to save a profile via PUT or POST. The unpin error was because you were making successive puts before the previous publish completely. I moved the unpin operation to after the publish.
2025-04-01T06:37:21.622560
2020-08-11T09:25:31
676721914
{ "authors": [ "MKodde", "thijskh" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2208", "repo": "OpenConext/OpenConext-engineblock", "url": "https://github.com/OpenConext/OpenConext-engineblock/pull/877" }
gharchive/pull-request
Remove non conforming SHO oid from config ~The urn:oid:<IP_ADDRESS>.4.1.14<IP_ADDRESS>.15 attribute was previously released by default when the SHO was in the ARP. This oid is not valid and was added for historic reasons. It's not clear which SP's still rely on this attribute. So by default the attribute is not released by EB, but can be enabled using the 'eb.arp_remove_non_conforming_sho_attribute' feature flag.~ Simply removing the alias from the config was a more pragmatic solution. Federations relying on the alias can manually put it back if need be. See: https://www.pivotaltracker.com/story/show/164237578 Yes that should work too, the only drawback is that you'd have to keep track of an alternative configs/attributes.json file in you deploy scripts. If that works for you, I'd love to revert this change! PS the build 72 build break on a visual regression test that times out. It ran on the 74 test so this should not be a blocker. We already ship our own version of the attributes.json config file in https://github.com/OpenConext/OpenConext-deploy/blob/master/roles/engineblock/files/attributes.json At least it's then out of the EB product. Of course everyone using OpenConext-deploy will still have it. But we can consider to move the SURF-specific items in attributes.json to our ansible environments, we can do that separately.
2025-04-01T06:37:21.646781
2024-04-18T10:24:35
2250321437
{ "authors": [ "muhammad-anas087", "rbren" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2209", "repo": "OpenDevin/OpenDevin", "url": "https://github.com/OpenDevin/OpenDevin/issues/1205" }
gharchive/issue
ERROR:root:<class 'ImportError'>: cannot import name 'spawn' from 'pexpect' (C:\Users\muhammad_anas.virtualenvs\OpenDevin-59qZ_dP-\Lib\site-packages\pexpect_init_.py) Describe the bug File "C:\Users\muhammad_anas.virtualenvs\OpenDevin-59qZ_dP-\Lib\site-packages\pexpect\pxssh.py", line 23, in from pexpect import ExceptionPexpect, TIMEOUT, EOF, spawn ERROR:root:<class 'ImportError'>: cannot import name 'spawn' from 'pexpect' (C:\Users\muhammad_anas.virtualenvs\OpenDevin-59qZ_dP-\Lib\site-packages\pexpect_init_.py) Setup and configuration Current version: commit 426f3871235ce7e98ef60c0122f5de91c4974547 (HEAD -> main, origin/main, origin/HEAD) Author: sp.wack<EMAIL_ADDRESS>Date: Wed Apr 17 20:55:17 2024 +0300 setup env for controlled integration tests with redux (#1180) My operating system: Windows 10 My environment vars and other configuration (be sure to redact API keys): My model and agent (you can see these settings in the UI): Model: Agent: Commands I ran to install and run OpenDevin: uvicorn opendevin.server.listen:app --port 3000 Steps to Reproduce: 1. 2. 3. Logs, error messages, and screenshots: Additional Context Deduping with https://github.com/OpenDevin/OpenDevin/issues/1156 Setting SANDBOX_TYPE=exec may fix this
2025-04-01T06:37:21.650775
2024-07-19T21:45:34
2419977214
{ "authors": [ "tobitege" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2210", "repo": "OpenDevin/OpenDevin", "url": "https://github.com/OpenDevin/OpenDevin/pull/3040" }
gharchive/pull-request
(test) test_runtime_client.py to test _execute_bash() What is the problem that this fixes or functionality that this introduces? Does it fix any open issues? Enhancements to RuntimeClient's method _execute_bash. Comes with new unit test file. Belongs to #3031 Give a summary of what the PR does, explaining any non-trivial design decisions Improved parsing of pexpect output/prompt, with first iteration of interactive prompt detection. There's one test commented out at the end if someone else wants to try to get it working. Some notes: running the original "runtime_build.py" via command line produced an image of 23+ GB(!) in size version in this PR tries to "fix" that (down to < 4GB), but somethings still not right with the setup, I think added tweaks so both Ubuntu 22.04 and 24.04 can generated (2 packages have different setups: libgl1-mesa-glx and libasound2) Ugh, I messed up a line in the dockerfile generation, doh!
2025-04-01T06:37:21.654787
2023-08-05T13:57:28
1837774848
{ "authors": [ "ZhouYunsong-SJTU", "secret104278" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2211", "repo": "OpenDriveLab/OpenScene", "url": "https://github.com/OpenDriveLab/OpenScene/issues/2" }
gharchive/issue
The detail of data generation pipeline Dear OpenDriveLab Team, Thank you for sharing your outstanding work with the community. Will you public the details of your data generation pipeline? such as how you manage dynamic objects when merging multiple LiDAR frames in your data pipeline. Thanks in advance. Thanks for your interest. In order to ensure the iteration of the dataset, we currently do not have plans to release the code for data processing. However, for the process of occupancy generation, please refer to OccNet (https://github.com/OpenDriveLab/OccNet). We use box annotations to accumulate foreground objects and background point clouds separately.
2025-04-01T06:37:21.686029
2023-03-07T23:56:09
1614398420
{ "authors": [ "andrew-platt", "luwang00" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2212", "repo": "OpenFAST/openfast", "url": "https://github.com/OpenFAST/openfast/pull/1483" }
gharchive/pull-request
hd: add NBodyMod regression tests This is ready for merging. Feature or improvement description Flexible platform capability was added to OpenFAST v2.6.0 in HydroDyn. This included a new flag NBodyMod with three methods of handling multiple potential flow (WAMIT) bodies. No test cases were added at that time. This was a set of test cases developed by @mattEhall some time ago. Related issue, if one exists #1480 fixed a bug in NBodyMod = 1 Impacted areas of the software Testing of HydroDyn NBodyMod options only Test results, if applicable New test cases have been added. These are based on the OC4Semi test case, but treat the floating platform as 4 separate bodies (center column, and 3 corner columns): NBodyMod1 -- a single set of WAMIT files includes coupling terms between each body (PtfmRefxt/yt/zt/ztRot should match XBODY(1)/(2)/(3)/(4) in WAMIT and NBody should match NBODY in WAMIT) NBodyMod2 -- 4 separate WAMIT bodies neglecting couplings between each body and NBODY=1 with XBODY=0 in WAMIT (PtfmRefxt/yt/zt/ztRot may differ from XBODY(1)/(2)/(3)/(4) in WAMIT) NBodyMod3 -- 4 separate WAMIT bodies neglecting couplings between each body and NBODY=1 with XBODY=/0 in WAMIT (PtfmRefxt/yt/zt/ztRot should match XBODY(1)/(2)/(3)/(4) in WAMIT) @luwang00, could you review this? I've gone through the run files briefly. This is probably not critical for verification, but it makes more sense physically to set PropPot to TRUE for Member 1, the central column, which is already modeled as a potential-flow body. The surge, heave, and pitching moments all agree relatively well between the three models (the sway, roll, and yaw moments are very small and show differences). Lu also ran some quick tests that showed relatively close agreement between the models.
2025-04-01T06:37:21.702276
2015-06-09T13:01:51
86592613
{ "authors": [ "jodastephen", "yukiiwashita" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2213", "repo": "OpenGamma/Strata", "url": "https://github.com/OpenGamma/Strata/issues/309" }
gharchive/issue
FxPayment and NotionalExchange FxPayment and NotionalExchange have the same functionality. Review these objects and decide which one we use. FxPayment was replaced by Payment FX now uses the shared Payment pricer. Notional exchange does not.
2025-04-01T06:37:21.703101
2015-06-26T15:58:28
91286659
{ "authors": [ "jodastephen" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2214", "repo": "OpenGamma/Strata", "url": "https://github.com/OpenGamma/Strata/issues/354" }
gharchive/issue
Merge Analytics into Strata Take those parts of the Analytics repo that are being used and include them in Strata. Fixed by a06ec1c3e5d0ed48c64d42a38f3cb32087141479
2025-04-01T06:37:21.704204
2018-05-07T08:53:17
320725349
{ "authors": [ "sfchen", "shaoangwen" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2215", "repo": "OpenGene/fastp", "url": "https://github.com/OpenGene/fastp/issues/56" }
gharchive/issue
Feature Request: filter pair-end sequences match a pattern Hi, I have two paired fastq files in hand and want to filter out sequences match a given pattern from both files at the same time. for example, filter out sequences and their paired-end sequences which contains "AATGCTACGTGAC" You can specify "AATGCTACGTGAC" as adapter, and use -l to require minimum read length
2025-04-01T06:37:21.864741
2023-08-01T02:25:46
1830345323
{ "authors": [ "cubxxw" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2216", "repo": "OpenIMSDK/chat", "url": "https://github.com/OpenIMSDK/chat/pull/102" }
gharchive/pull-request
feat: add add labels πŸ” What type of PR is this? /kind documentation /kind feature πŸ‘€ What this PR does / why we need it: [x] My pull request adheres to the code style of this project [x] My code requires changes to the documentation [x] I have updated the documentation as required [x] All the tests have passed πŸ…° Which issue(s) this PR fixes: Fixes OpenIMSDK/Open-IM-Server#406 /create tag v1.1.1 "this is comment" /create tag v1.1.1 /create tag v1.1.1 comment /create tag v1.1.1 "comment"
2025-04-01T06:37:21.888405
2024-02-26T16:11:44
2154562734
{ "authors": [ "ddneilson" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2217", "repo": "OpenJobDescription/openjd-sessions-for-python", "url": "https://github.com/OpenJobDescription/openjd-sessions-for-python/pull/91" }
gharchive/pull-request
fix!: remove methods from public interface of WindowsSessionUser What was the problem/requirement? (What/Why) Two problems, really: The WindowsSessionUser class exposes a lot of functionality that are for internal use in the class constructor, and I don't think should be in the public interface. e.g. "validate_username_password" The regex-based validation of username & domain strikes me as not the way that this validation should be done. We should just be checking that the username/domain exists to the host. Two cases: a. We have a password; the logon check will fail if the use bad information. b. We don't have a password; we only allow this when the user is the process user. What was the solution? (How) Move methods from public to private by prefixing with an underscore. Remove the username & domain validation logic. It's handled by the logon & current-user checks that already exist. What is the impact of this change? Tidier code and interfaces. This is motivated by a change that I have upcoming. How was this change tested? The unit tests handle these cases; they've been updated as needed. I added some skipifs while in there to remove irrelevant platform-specific xfails. Was this change documented? N/A Is this a breaking change? BREAKING CHANGE BadUserNameException and BadDomainNameException have been removed. Many methods of WindowsSessionUser have been made private. By submitting this pull request, I confirm that you can use, modify, copy, and redistribute this contribution, under the terms of your choice. However, I do wonder about the potential computational cost of this, and whether it will be obvious to anyone creating a WindowsSesssionUser` that a logon will be attempted etc. Good observation in that it might be surprising. The computational cost seems to be essentially nil; it's pretty quick, and I don't anticipate the check being an issue at the scale that we're expecting -- logins spread out by seconds/minutes/hours/days, rather than microseconds. An option to make the check optional is a two-way door, so we can add it later if the need arises.
2025-04-01T06:37:21.899196
2022-10-18T12:29:34
1413140546
{ "authors": [ "jakub-pomykala" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2218", "repo": "OpenLiberty/cloud-hosted-guides", "url": "https://github.com/OpenLiberty/cloud-hosted-guides/issues/2241" }
gharchive/issue
Copy button on output codeblock in technical deep dive There is a copy button but as this is an example output it shouldn't be there. Another one: Another one in: "Augmenting the existing Jakarta RESTful Web Services annotations with OpenAPI annotations" Error message missing in: "Consuming the secured RESTful APIs by JWT" Another copyblock on output: "Deploying the microservice to Kubernetes"
2025-04-01T06:37:23.236678
2022-01-18T07:33:46
1106566875
{ "authors": [ "collado-mike", "mobuchowski", "wslulciuc" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2219", "repo": "OpenLineage/OpenLineage", "url": "https://github.com/OpenLineage/OpenLineage/pull/480" }
gharchive/pull-request
Add http client for Java Problem The openlineage-java lib. does not currently provide an HTTP client to emit run events. Closes: #430 Solution Add OpenLineageClient the implements the HTTP endpoint /api/v1/lineage to emit run events. Note, the HTTP client was heavily inspired by the marquez-java lib. import io.openlineage.client.OpenLineageClient; // (1) Create a new OpenLineageClient instance Note, by default, environments variables // OPENLINEAGE_URL and OPENLINEAGE_API_KEY are used To set the URL and API key // manually, use the constructor OpenLineageClient(BaseUrl, ApiKey) OpenLineageClient client = new OpenLineageClient(); // (2) Define a simple OpenLineage START or END event OpenLineage.Run startOrEndRun = ... // (3) Emit OpenLineage event client.emit(startOrEndRun) Other changes in this PR include: Add rule in .gitignore to ignore generated OL models Add spotless to format Java code Checklist [x] You've signed-off your work [x] Your pull request title follows our guidelines [ ] Your changes are accompanied by tests (if relevant) [ ] Your change contains a small diff and is self-contained [x] You've updated any relevant documentation (if relevant) [ ] You've updated the CHANGELOG.md with details about your change under the "Unreleased" section (if relevant, depending on the change, this may not be necessary) [ ] You've versioned the core OpenLineage model or facets according to SchemaVer (if relevant) Are we going to support arbitrary query params, as Spark HTTP client now does? https://github.com/OpenLineage/OpenLineage/pull/425 Are we going to support arbitrary query params, as Spark HTTP client now does? #425 Yeah, I think we should. Should the params be configured when creating the client, or provide it as an option when emitting an event? client.emit(event, queryParams) @wslulciuc I think client.emit(event) should be an overload of client.emit(event, queryParams) that fills in default query params from environment. Yeah, I think we should. Should the params be configured when creating the client, or provide it as an option when emitting an event? Personally, I think setting the query params at construction time is the way to go. I don't imagine the query params changing from one emit call to another. @mobuchowski / @collado-mike: I've documented the following the in the README.md, but here's the approach I recommend we use to configure the client with query params appended on each HTTP request: URI uri = new URIBuilder("http://localhost:5000") .addParameter("param0", "value0") .addParameter("param1", "value2") .build(); OpenLineageClient client = Clients.newClient(uri.toURL());
2025-04-01T06:37:23.271552
2023-06-27T09:39:01
1776498097
{ "authors": [ "cbielow", "greengypsy" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:2220", "repo": "OpenMS/OpenMS-docs", "url": "https://github.com/OpenMS/OpenMS-docs/pull/201" }
gharchive/pull-request
update installer win Describe the change Make user aware of Windows Defender when using our installer. (this PR is against develop -- should it be staging? (seems outdated, but the PR pretext says so...) PR checklist [ ] I have added description of the change I'm proposing in the OpenMS Documentation. [ ] I have read and followed OpenMS Documentation Contributing guidelines. [ ] I have attached a screenshot of the relevant area after this change. [ ] CHANGELOG.md is updated. [ ] I have added my name in CONTRIBUTING.md. Yes staging is obsolete. Develop is the main branch I believe.