id
int64
393k
2.82B
repo
stringclasses
68 values
title
stringlengths
1
936
body
stringlengths
0
256k
labels
stringlengths
2
508
priority
stringclasses
3 values
severity
stringclasses
3 values
2,555,113,843
neovim
multigrid UI: floating windows move when multigrid UI reconnects
### Problem When a multigrid UI reconnects, the floating windows move and the contents becomes empty. ### Steps to reproduce 1. Start a server `nvim --headless --listen /tmp/nvim.socket` 2. Connect with a multigrid UI like Neovide `neovide --server /tmp/nvim.socket` 3. Open a floating window like telescope 4. Disconnect from the server `:call chanclose(nvim_list_uis()[0].chan)` 5. Reconnect `neovide --server /tmp/nvim.socket` 6. Observe that the layout has changed like below Before: ![image](https://github.com/user-attachments/assets/d70122b3-c744-4beb-aaea-0296ad15b3b9) After: ![image](https://github.com/user-attachments/assets/85c230f4-247c-48b0-9026-1aa457a06577) ### Expected behavior The layout and the contents should not change ### Nvim version (nvim -v) 0.10.1 and latest master ### Vim (not Nvim) behaves the same? N/A ### Operating system/version Arch Linux ### Terminal name/version Neovide 0.13.3 ### $TERM environment variable N/A ### Installation pacman
bug,ui,ui-extensibility
low
Minor
2,555,115,608
neovim
multigrid UI: window contents disappear when multigrid UI reconnects
### Problem When a multigrid UI reconnects, the window content disappears and only the background is visible ### Steps to reproduce 1. Start a server `nvim --headless --listen /tmp/nvim.socket` 2. Connect with a multigrid UI like Neovide `neovide --server /tmp/nvim.socket` 3. Open a buffer 4. Disconnect from the server `:call chanclose(nvim_list_uis()[0].chan)` 5. Reconnect `neovide --server /tmp/nvim.socket` Observe that the window is empty: Before: ![image](https://github.com/user-attachments/assets/4c6d6bc8-d7df-45cb-8cd4-546ba1a027e5) After: ![image](https://github.com/user-attachments/assets/1b423cd0-21f4-4f3c-bc71-3aa6c0ff1f86) ### Expected behavior The window should keep it's contents. ### Nvim version (nvim -v) 0.10.1 and latest master ### Vim (not Nvim) behaves the same? N/A ### Operating system/version Arch Linux ### Terminal name/version Neovide 0.13.3 ### $TERM environment variable N/A ### Installation pacman
bug,ui,ui-extensibility
low
Minor
2,555,129,549
neovim
UI protocol: expose floatwin composition order (comp_index) to UIs
### Problem Neovim uses a hidden internal composition order, which determines the order the floating windows with the same z-index are drawn. Many UIs, including the very widespread telecope relies on this order to build the UIs. In particular, a there's no alpha blending between the windows of the same z-index, the last window in the index will always be used. A GUI like, Neovide, need to respect this order and do the same type of rendering, or things will look wrong. So that's in fact what we are doing right now, I reverse engineered the algorithm and duplicated it in Neovide [here](https://github.com/neovide/neovide/pull/2612). But it only works when the UI is connected from the start, and we receive absolutely all the events. It's also very fragile. ### Expected behavior The internal composition order is exposed in the UI protocol. NOTE: Another solution would be to require the whole plugin world to stop using overlapping z-indices. But this does still not work correctly in Neovide, because then we need a way to control the shadow casting and blur. Threating the whole z-index as one layer, is much easier to understand and does not require plugin authors to add additional Neovide specific annotations.
enhancement,ui-extensibility,floatwin
low
Major
2,555,151,120
svelte
Consider switching to a more mature parser
### Describe the problem There are many issues that have `acorn-typescript` as the upstream cause (#13179, #13125, #13409, #13188, [more](https://github.com/sveltejs/svelte/issues?q=acorn)), but `acorn-typescript` is [abandoned](https://github.com/TyrealHu/acorn-typescript/issues/61). It also has several open issues that would need to be addressed once forked or collaboration access is granted. ### Describe the proposed solution One of the following: - Fork `acorn-typescript`. The svelte team would become the primary maintainers but it wouldn't require any changes to the svelte source code. - Use an alternate parser, such as [`swc`](https://swc.rs/docs/usage/core#parse) or [`oxc`](https://github.com/oxc-project/oxc?tab=readme-ov-file#-ast-and-parser). SWC seems to use the same AST types as acorn, but oxc advertises faster speeds. These parsers are much more mature than `acorn-typescript` and are obviously actively maintained. (note: oxc would require changes to both esrap and the svelte source) My vote would be for the latter. The downsides (bigger dependencies and changes to svelte's source) are outweighed by the long-term advantage of using a more reliable library. ### Importance nice to have
compiler,needs discussion
medium
Major
2,555,163,651
vscode
`editor.fontLigatuers` doesn't wotk if the string is long
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes/No <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.93.1 - OS Version: windows 10 Steps to Reproduce: 1. Install `iosevka` font https://github.com/be5invis/Iosevka/releases/tag/v31.7.1 which has many variants 2. use the font in vscode and apply a long `fontLigatures` setting: ``` "editor.fontLigatures": "'dlig','ss20','cv06' 1,'cv08' 2,'cv09' 1,'cv10' 6,'cv17' 11,'cv19' 3,'cv20' 1,'cv26' 8,'cv27' 21,'cv28' 4,'cv36' 4,'cv37' 4,'cv39' 7,'cv42' 6,'cv44' 8,'cv45' 1,'cv47' 8,'cv48' 30,'cv49' 8,'cv50' 4,'cv51' 9,'cv52' 5,'cv54' 3,'cv55' 8,'VSAJ' 7,'VSAK' 2,'VSAL' 7,'VSAM' 7,'VLAG' 2", ``` The variant features do not work in this case, however, they will work if the string is sufficiently shortened: ``` "editor.fontLigatures": "'dlig','cv06' 1,'cv44' 8,'cv45' 1,'cv47' 8,'cv49' 8,'cv50' 4,'cv51' 9,'cv52' 5,'cv54' 3,'cv55' 8,'VSAJ' 7,'VSAK' 2,'VSAL' 7,'VSAM' 7,'VLAG' 2,'cv48' 30", ``` There seems to be a length limit on this string around 150 bytes, maybe increase it to 1024, etc.
bug,font-rendering,confirmation-pending
low
Critical
2,555,192,845
svelte
Slide transition of svelte 5 does not work correctly on iOS
### Describe the bug A slide transition on iOS/Safari does not „slide in“ an element step by step. Instead the content is show in one step and then the height is increased step by step. But as the content might overflow this results in an effect that does Not Look Like „slide in“. All other browsers work as expected. As a graphic is better than words look at this screencast: https://www.screencast.com/t/W7aQv2SZK5K ### Reproduction https://svelte-5-preview.vercel.app/#H4sIAAAAAAAACqVRTUvEMBD9K2MUdheUrh77BSILHsQVPHiwHrLNtIZNk5JMV5bQ_27TUPcHeJuZN_Pe441njVToWPrpmeYdspQ99j27ZXTuQ-NOqAin3pnB1mGSC3kqK12Rv5YNmB41ijH0FQUoVhWR5dpJkkanTkmBhfcgBsvnCTxst1sYx2XZ0VlhUbGO21ZO8D12GRx4fWytGbS4q40yNgUl229qjRKorRFnVMr8ZBWLLOVCNht8DiB8GKtEniyO_4deap_IZvK-DPLDQGQ0GF0rWR8Lv95AUcI6ZgMFXMVqM5b7t91r8vSyf9_lSbwKhDp3tZU9geK6nWIgV7FZSHa9sQQe5ghhhMaaDlbxKckl4lUWthUS_GneOOKE64Yrh5ssmI0aE-_0zc4I2UgULCU74Pg1_gJaN3c8CAIAAA== ### System Info ```shell System: OS: Windows 11 10.0.22631 CPU: (8) x64 11th Gen Intel(R) Core(TM) i7-1165G7 @ 2.80GHz Memory: 7.03 GB / 31.75 GB Binaries: Node: 20.17.0 - C:\Program Files\nodejs\node.EXE Yarn: 1.22.17 - ~\AppData\Roaming\npm\yarn.CMD npm: 10.8.2 - C:\Program Files\nodejs\npm.CMD pnpm: 9.11.0 - C:\Program Files\nodejs\pnpm.CMD Browsers: Edge: Chromium (127.0.2651.74) Internet Explorer: 11.0.22621.3527 ``` ### Severity blocking an upgrade
transition/animation
low
Critical
2,555,255,736
yt-dlp
Requesting Suppor for Chapman University Archive video downloads
### DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE - [X] I understand that I will be **blocked** if I *intentionally* remove or skip any mandatory\* field ### Checklist - [X] I'm reporting a new site support request - [X] I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels)) - [X] I've checked that all provided URLs are playable in a browser with the same IP and same login details - [X] I've checked that none of provided URLs [violate any copyrights](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#is-the-website-primarily-used-for-piracy) or contain any [DRM](https://en.wikipedia.org/wiki/Digital_rights_management) to the best of my knowledge - [X] I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates - [X] I've read the [guidelines for opening an issue](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#opening-an-issue) - [X] I've read about [sharing account credentials](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#are-you-willing-to-share-account-details-if-needed) and am willing to share it if required ### Region United States ### Example URLs - Video Listings: https://blogs.chapman.edu/huell-howser-archives/archives/ - Video Page: https://blogs.chapman.edu/huell-howser-archives/1991/01/10/preserving-the-past-californias-gold-207/ - Single Video Link: https://video.chapman.edu/V/Video?v=9842745&node=43207828&a=103486387&preload=false ### Provide a description that is worded well enough to be understood Requesting added support for the Chapman University video archives. I belive it's an HLS stream but I am not too sure. ### Provide verbose output that clearly demonstrates the problem - [X] Run **your** yt-dlp command with **-vU** flag added (`yt-dlp -vU <your command line>`) - [X] If using API, add `'verbose': True` to `YoutubeDL` params instead - [X] Copy the WHOLE output (starting with `[debug] Command-line config`) and insert it below ### Complete Verbose Output ```shell C:\Windows\System32>yt-dlp -vU --force-generic-extractor https://video.chapman.edu/V/Video?v=9842745&node=43207828&a=103486387&preload=false -vU [debug] Command-line config: ['-vU', '--force-generic-extractor', 'https://video.chapman.edu/V/Video?v=9842745'] [debug] Encodings: locale cp1252, fs utf-8, pref cp1252, out utf-8, error utf-8, screen utf-8 [debug] yt-dlp version stable@2024.08.06 from yt-dlp/yt-dlp [4d9231208] (pip) [debug] Python 3.12.3 (CPython AMD64 64bit) - Windows-11-10.0.22631-SP0 (OpenSSL 3.0.13 30 Jan 2024) [debug] exe versions: ffmpeg N-116328-gfa5a605542-20240718 (setts), ffprobe N-116328-gfa5a605542-20240718, phantomjs 2.1.1 [debug] Optional libraries: Cryptodome-3.20.0, brotli-1.1.0, certifi-2024.07.04, mutagen-1.47.0, requests-2.32.3, sqlite3-3.45.1, urllib3-2.2.2, websockets-12.0 [debug] Proxy map: {} [debug] Request Handlers: urllib, requests, websockets [debug] Loaded 1830 extractors [debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest [debug] Downloading _update_spec from https://github.com/yt-dlp/yt-dlp/releases/latest/download/_update_spec Current version: stable@2024.08.06 from yt-dlp/yt-dlp Latest version: stable@2024.09.27 from yt-dlp/yt-dlp ERROR: You installed yt-dlp with pip or using the wheel from PyPi; Use that to update [generic] Extracting URL: https://video.chapman.edu/V/Video?v=9842745 [generic] Video?v=9842745: Downloading webpage WARNING: [generic] Forcing generic information extractor [generic] Video?v=9842745: Extracting information [debug] Looking for embeds ERROR: Unsupported URL: https://video.chapman.edu/V/Video?v=9842745 Traceback (most recent call last): File "C:\Python312\Lib\site-packages\yt_dlp\YoutubeDL.py", line 1626, in wrapper return func(self, *args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Python312\Lib\site-packages\yt_dlp\YoutubeDL.py", line 1761, in __extract_info ie_result = ie.extract(url) ^^^^^^^^^^^^^^^ File "C:\Python312\Lib\site-packages\yt_dlp\extractor\common.py", line 740, in extract ie_result = self._real_extract(url) ^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Python312\Lib\site-packages\yt_dlp\extractor\generic.py", line 2526, in _real_extract raise UnsupportedError(url) yt_dlp.utils.UnsupportedError: Unsupported URL: https://video.chapman.edu/V/Video?v=9842745 node:internal/modules/cjs/loader:1148 throw err; ^ Error: Cannot find module 'C:\Windows\System32\=43207828' at Module._resolveFilename (node:internal/modules/cjs/loader:1145:15) at Module._load (node:internal/modules/cjs/loader:986:27) at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:174:12) at node:internal/main/run_main_module:28:49 { code: 'MODULE_NOT_FOUND', requireStack: [] } Node.js v20.13.1 'a' is not recognized as an internal or external command, operable program or batch file. 'preload' is not recognized as an internal or external command, operable program or batch file. ```
site-request,triage
low
Critical
2,555,287,891
PowerToys
Keyboard manager broke down after latest update
### Microsoft PowerToys version 0.84.1 ### Installation method PowerToys auto-update ### Running as admin Yes ### Area(s) with issue? Keyboard Manager ### Steps to reproduce [PowerToysReport_2024-09-29-21-27-26.zip](https://github.com/user-attachments/files/17182358/PowerToysReport_2024-09-29-21-27-26.zip) I'm not entirely sure if this is what caused it, but after installing the last update, I try to use a shortcut I had previously set up that worked just fine and it worked once before edge started using its default shortcut for that key combination that i didn't even know was a thing because I always used the shortcut I made. In the moment I had ctrl + NumPad1 and ctrl + NumPad3 assigned to send the text ">" and "<", respectively. I tried on other apps besides edge and it still didn't work at all. ### ✔️ Expected Behavior I was expecting for the shortcuts that I had already and used to work well to keep working. ### ❌ Actual Behavior None of the shortcuts I tried making nor the key rebindings worked after that, except for one or two, but for some reason the expected output of the latest shortcut was that of the prior shortcut, even when all prior shortcuts were deleted. ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Minor
2,555,304,960
yt-dlp
[youtube] 120-hour limit, imposed on live streams is wrong
### DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE - [X] I understand that I will be **blocked** if I *intentionally* remove or skip any mandatory\* field ### Checklist - [X] I'm requesting a feature unrelated to a specific site - [X] I've looked through the [README](https://github.com/yt-dlp/yt-dlp#readme) - [X] I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels)) - [X] I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates - [X] I've read the [guidelines for opening an issue](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#opening-an-issue) ### Provide a description that is worded well enough to be understood Please, make it at least 168 hours or do some actual probing for available timestamps. This hard limit is definitely not flexible and not everyone could patch their youtube.py, especially, Windows users. ### Provide verbose output that clearly demonstrates the problem - [X] Run **your** yt-dlp command with **-vU** flag added (`yt-dlp -vU <your command line>`) - [ ] If using API, add `'verbose': True` to `YoutubeDL` params instead - [X] Copy the WHOLE output (starting with `[debug] Command-line config`) and insert it below ### Complete Verbose Output ```shell [debug] Command-line config: ['-vU'] Deprecated Feature: Support for Windows 7/Server 2008 R2 has been deprecated. See https://github.com/yt-dlp/yt-dlp/issues/10086 for details. You may stop receiving updates on this version at any time! [debug] Encodings: locale cp1251, fs utf-8, pref cp1251, out utf-8 (No VT), error utf-8 (No VT), screen utf-8 (No VT) [debug] yt-dlp version nightly@2024.09.29.232819 from yt-dlp/yt-dlp-nightly-builds [6328e2e67] (win_exe) [debug] Python 3.8.10 (CPython AMD64 64bit) - Windows-7-6.1.7601-SP1 (OpenSSL 1.1.1k 25 Mar 2021) [debug] exe versions: ffmpeg git-2024-09-22-a577d31-ffmpeg-windows-build-helpers (fdk,setts), ffprobe git-2024-09-22-a577d31-ffm peg-windows-build-helpers [debug] Optional libraries: Cryptodome-3.20.0, brotli-1.1.0, certifi-2024.08.30, curl_cffi-0.5.10, mutagen-1.47.0, requests-2.32 .3, sqlite3-3.35.5, urllib3-2.2.3, websockets-13.1 [debug] Proxy map: {} [debug] Request Handlers: urllib, requests, websockets, curl_cffi [debug] Loaded 1838 extractors [debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp-nightly-builds/releases/latest Latest version: nightly@2024.09.29.232819 from yt-dlp/yt-dlp-nightly-builds yt-dlp is up to date (nightly@2024.09.29.232819 from yt-dlp/yt-dlp-nightly-builds) ```
site-enhancement,triage
low
Critical
2,555,332,481
go
proposal: slices: funcs `CollectKeys` and `CollectValues` to collect from an iter.Seq2
### Proposal Details The `slices` package currently has `Collect()` to perform `iter.Seq[T] -> []T`, but it doesn't have similar helper functions for collecting the keys or values from an `iter.Seq2[K, V]`. I propose to add the following two functions to `slices`: ```go func CollectKeys[K, V any](seq iter.Seq2[K, V]) []K { var s []K for k := range seq { s = append(s, k) } return s } func CollectValues[K, V any](seq iter.Seq2[K, V]) []V { var s []V for _, v := range seq { s = append(s, v) } return s } ```
Proposal
low
Minor
2,555,338,958
PowerToys
Workspaces error on streamdeck
### Microsoft PowerToys version 0.84.1 ### Installation method Microsoft Store ### Running as admin Yes ### Area(s) with issue? Workspaces ### Steps to reproduce Hi everyone, I created a workspace and created a shortcut icon on the desktop, when I try to add that file to my streamdeck and try to run that file it gives me this error: Incorrect command line arguments if I run the file with a double mouse click it works, why doesn't it work when I try to use the same file but via streamdeck? Is it possible to solve it? It's easier for me to manage the workspaces via streamdeck. thanks ### ✔️ Expected Behavior _No response_ ### ❌ Actual Behavior _No response_ ### Other Software _No response_
Issue-Bug,Needs-Triage,Product-Workspaces
low
Critical
2,555,361,878
ollama
Error: llama runner process has terminated: error loading modelvocabulary: cannot find tokenizer merges in model file
### What is the issue? time : 09/30/2024 script: ``` FROM "./model-quant.gguf" TEMPLATE """{{- if .System }} <|im_start|>system {{ .System }}<|im_end|> {{- end }} <|im_start|>user {{ .Prompt }}<|im_end|> <|im_start|>assistant """ SYSTEM """""" PARAMETER stop <|im_start|> PARAMETER stop <|im_end|> ``` The creation was successful, but the operation failed ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version ollama version is 0.3.12
bug
low
Critical
2,555,424,736
rust
access outside the bounds for given AllocRange
<!-- Thank you for finding an Internal Compiler Error! 🧊 If possible, try to provide a minimal verifiable example. You can read "Rust Bug Minimization Patterns" for how to create smaller examples. http://blog.pnkfx.org/blog/2019/11/18/rust-bug-minimization-patterns/ --> ### Code ```Rust #![feature(generic_const_exprs, generic_arg_infer, generic_const_items, associated_const_equality, adt_const_params)] #![allow(incomplete_features)] fn foo<const N: usize, const M: usize>(_: [(); N + 1 + M]) {} #![a] trait Owner { const C<const N: u32>: u32; } impl Owner for () { const C<const N: u32>: u32 = N; } fn take0<const N: usize>(_: impl Owner<C<N> = { N }>) {} fn main() { take0::<128>(()); } ``` ### Meta <!-- If you're using the stable version of the compiler, you should also check if the bug also exists in the beta or nightly versions. --> `rustc --version --verbose`: ``` rustc 1.83.0-nightly (7608018cb 2024-09-29) binary: rustc commit-hash: 7608018cbdac9e55d0d13529cf43adc33d53efcf commit-date: 2024-09-29 host: x86_64-apple-darwin release: 1.83.0-nightly LLVM version: 19.1.0 ``` ### Error output ``` error: an inner attribute is not permitted in this context --> ./9A756-1.rs:11:1 | 11 | #![this_just_works] | ^^^^^^^^^^^^^^^^^^^ 12 | 13 | / trait Owner { 14 | | const C<const N: u32>: u32; 15 | | } | |_- the inner attribute doesn't annotate this trait | = note: inner attributes, like `#![no_std]`, annotate the item enclosing them, and are usually found at the beginning of source files error: the constant `N` is not of type `u32` --> ./9A756-1.rs:21:40 | 21 | fn take0<const N: usize>(_: impl Owner<C<N> = { N }>) {} | ^^^^^^^^^^^^ expected `u32`, found `usize` | note: required by a const generic parameter in `Owner::C` --> ./9A756-1.rs:14:13 | 14 | const C<const N: u32>: u32; | ^^^^^^^^^^^^ required by this const generic parameter in `Owner::C` ``` <!-- Include a backtrace in the code block by setting `RUST_BACKTRACE=1` in your environment. E.g. `RUST_BACKTRACE=1 cargo build`. --> <details><summary><strong>Backtrace</strong></summary> <p> ``` thread 'rustc' panicked at /rustc/7608018cbdac9e55d0d13529cf43adc33d53efcf/compiler/rustc_middle/src/mir/interpret/allocation.rs:268:9: access outside the bounds for given AllocRange stack backtrace: 0: 0x10d3193e7 - std::backtrace::Backtrace::create::h697070a74b47c5a1 1: 0x10d319335 - std::backtrace::Backtrace::force_capture::hd4e20dc44637c239 2: 0x10b13a7ce - std[a533c313b02b6cfe]::panicking::update_hook::<alloc[3194a7a73d70d243]::boxed::Box<rustc_driver_impl[e0e0c21a20d76ec4]::install_ice_hook::{closure#0}>>::{closure#0} 3: 0x10d334af8 - std::panicking::rust_panic_with_hook::h9b093984c0b2ae3e 4: 0x10d334735 - std::panicking::begin_panic_handler::{{closure}}::h350248c9538498ad 5: 0x10d331f49 - std::sys::backtrace::__rust_end_short_backtrace::h51a3ed22bf72bc03 6: 0x10d3343ac - _rust_begin_unwind 7: 0x10ff4ff8f - core::panicking::panic_fmt::ha29ff0ee788aef57 8: 0x10b001b9f - <rustc_middle[3d266d78c9b198e4]::mir::interpret::allocation::AllocRange>::subrange 9: 0x10b0862dd - <rustc_const_eval[298ced3d0b3bd135]::interpret::memory::AllocRefMut<rustc_middle[3d266d78c9b198e4]::mir::interpret::pointer::CtfeProvenance, ()>>::write_scalar 10: 0x10b097b4a - <rustc_const_eval[298ced3d0b3bd135]::interpret::eval_context::InterpCx<rustc_const_eval[298ced3d0b3bd135]::const_eval::machine::CompileTimeMachine>>::write_immediate_to_mplace_no_validate 11: 0x10b097977 - <rustc_const_eval[298ced3d0b3bd135]::interpret::eval_context::InterpCx<rustc_const_eval[298ced3d0b3bd135]::const_eval::machine::CompileTimeMachine>>::write_immediate_no_validate::<rustc_const_eval[298ced3d0b3bd135]::interpret::place::MPlaceTy> 12: 0x10b0c993b - <rustc_const_eval[298ced3d0b3bd135]::interpret::eval_context::InterpCx<rustc_const_eval[298ced3d0b3bd135]::const_eval::machine::CompileTimeMachine>>::copy_op_no_validate::<rustc_const_eval[298ced3d0b3bd135]::interpret::operand::OpTy, rustc_const_eval[298ced3d0b3bd135]::interpret::place::MPlaceTy> 13: 0x10b0c278b - <rustc_const_eval[298ced3d0b3bd135]::interpret::eval_context::InterpCx<rustc_const_eval[298ced3d0b3bd135]::const_eval::machine::CompileTimeMachine>>::return_from_current_stack_frame 14: 0x10b0a1446 - <rustc_const_eval[298ced3d0b3bd135]::interpret::eval_context::InterpCx<rustc_const_eval[298ced3d0b3bd135]::const_eval::machine::CompileTimeMachine>>::eval_terminator 15: 0x10b0b9cd4 - rustc_const_eval[298ced3d0b3bd135]::const_eval::eval_queries::eval_to_allocation_raw_provider 16: 0x10c980cfc - rustc_query_impl[75f1704df0e38289]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_allocation_raw::dynamic_query::{closure#2}::{closure#0}, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>> 17: 0x10c93173e - <rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_allocation_raw::dynamic_query::{closure#2} as core[907a20686bf8bf7a]::ops::function::FnOnce<(rustc_middle[3d266d78c9b198e4]::ty::context::TyCtxt, rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>)>>::call_once 18: 0x10c748579 - rustc_query_system[13e780f7d1d88911]::query::plumbing::try_execute_query::<rustc_query_impl[75f1704df0e38289]::DynamicConfig<rustc_query_system[13e780f7d1d88911]::query::caches::DefaultCache<rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>>, false, false, false>, rustc_query_impl[75f1704df0e38289]::plumbing::QueryCtxt, false> 19: 0x10c9b823c - rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_allocation_raw::get_query_non_incr::__rust_end_short_backtrace 20: 0x10afc62a2 - rustc_middle[3d266d78c9b198e4]::query::plumbing::query_get_at::<rustc_query_system[13e780f7d1d88911]::query::caches::DefaultCache<rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>>> 21: 0x10b05e831 - rustc_const_eval[298ced3d0b3bd135]::const_eval::valtrees::eval_to_valtree 22: 0x10bae8d03 - <rustc_const_eval[298ced3d0b3bd135]::provide::{closure#0} as core[907a20686bf8bf7a]::ops::function::FnOnce<(rustc_middle[3d266d78c9b198e4]::ty::context::TyCtxt, rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>)>>::call_once 23: 0x10c97ee5c - rustc_query_impl[75f1704df0e38289]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_valtree::dynamic_query::{closure#2}::{closure#0}, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>> 24: 0x10c910a7e - <rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_valtree::dynamic_query::{closure#2} as core[907a20686bf8bf7a]::ops::function::FnOnce<(rustc_middle[3d266d78c9b198e4]::ty::context::TyCtxt, rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>)>>::call_once 25: 0x10c748579 - rustc_query_system[13e780f7d1d88911]::query::plumbing::try_execute_query::<rustc_query_impl[75f1704df0e38289]::DynamicConfig<rustc_query_system[13e780f7d1d88911]::query::caches::DefaultCache<rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>>, false, false, false>, rustc_query_impl[75f1704df0e38289]::plumbing::QueryCtxt, false> 26: 0x10c9b946c - rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_valtree::get_query_non_incr::__rust_end_short_backtrace 27: 0x10bf74b88 - rustc_middle[3d266d78c9b198e4]::query::plumbing::query_get_at::<rustc_query_system[13e780f7d1d88911]::query::caches::DefaultCache<rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>>> 28: 0x10bf796f3 - <rustc_middle[3d266d78c9b198e4]::ty::context::TyCtxt>::const_eval_global_id_for_typeck 29: 0x10bf79028 - <rustc_middle[3d266d78c9b198e4]::ty::context::TyCtxt>::const_eval_resolve_for_typeck 30: 0x10b9c469a - <rustc_infer[9b67ab4c7e1e7e36]::infer::InferCtxt>::const_eval_resolve 31: 0x10b9b3c9a - <rustc_infer[9b67ab4c7e1e7e36]::infer::InferCtxt>::try_const_eval_resolve 32: 0x10d0fd1cf - <rustc_trait_selection[1363494f2f4ec6d3]::traits::fulfill::FulfillProcessor as rustc_data_structures[3069d1c4d7b76162]::obligation_forest::ObligationProcessor>::process_obligation::{closure#0} 33: 0x10d0fc596 - <rustc_trait_selection[1363494f2f4ec6d3]::traits::fulfill::FulfillProcessor as rustc_data_structures[3069d1c4d7b76162]::obligation_forest::ObligationProcessor>::process_obligation 34: 0x10b602dfb - <rustc_data_structures[3069d1c4d7b76162]::obligation_forest::ObligationForest<rustc_trait_selection[1363494f2f4ec6d3]::traits::fulfill::PendingPredicateObligation>>::process_obligations::<rustc_trait_selection[1363494f2f4ec6d3]::traits::fulfill::FulfillProcessor> 35: 0x10b714556 - <rustc_trait_selection[1363494f2f4ec6d3]::traits::fulfill::FulfillmentContext<rustc_trait_selection[1363494f2f4ec6d3]::traits::FulfillmentError> as rustc_infer[9b67ab4c7e1e7e36]::traits::engine::TraitEngine<rustc_trait_selection[1363494f2f4ec6d3]::traits::FulfillmentError>>::select_where_possible 36: 0x10b7d18c2 - <rustc_hir_typeck[83d45668b9c71bf7]::fn_ctxt::FnCtxt>::check_argument_types 37: 0x10b76d5ce - <rustc_hir_typeck[83d45668b9c71bf7]::fn_ctxt::FnCtxt>::confirm_builtin_call 38: 0x10b8802b1 - <rustc_hir_typeck[83d45668b9c71bf7]::fn_ctxt::FnCtxt>::check_expr_kind 39: 0x10b79c073 - <rustc_hir_typeck[83d45668b9c71bf7]::fn_ctxt::FnCtxt>::check_expr_with_expectation_and_args 40: 0x10b7e087e - <rustc_hir_typeck[83d45668b9c71bf7]::fn_ctxt::FnCtxt>::check_block_with_expected 41: 0x10b79c073 - <rustc_hir_typeck[83d45668b9c71bf7]::fn_ctxt::FnCtxt>::check_expr_with_expectation_and_args 42: 0x10b79e237 - <rustc_hir_typeck[83d45668b9c71bf7]::fn_ctxt::FnCtxt>::check_return_expr 43: 0x10b8661c2 - rustc_hir_typeck[83d45668b9c71bf7]::check::check_fn 44: 0x10b85f9bc - rustc_hir_typeck[83d45668b9c71bf7]::typeck 45: 0x10c982a7a - rustc_query_impl[75f1704df0e38289]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[75f1704df0e38289]::query_impl::typeck::dynamic_query::{closure#2}::{closure#0}, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 8usize]>> 46: 0x10c7d8f4e - rustc_query_system[13e780f7d1d88911]::query::plumbing::try_execute_query::<rustc_query_impl[75f1704df0e38289]::DynamicConfig<rustc_query_system[13e780f7d1d88911]::query::caches::VecCache<rustc_span[51e7cf5b04f4660f]::def_id::LocalDefId, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 8usize]>>, false, false, false>, rustc_query_impl[75f1704df0e38289]::plumbing::QueryCtxt, false> 47: 0x10c9b1bca - rustc_query_impl[75f1704df0e38289]::query_impl::typeck::get_query_non_incr::__rust_end_short_backtrace 48: 0x10b3f209e - <rustc_middle[3d266d78c9b198e4]::hir::map::Map>::par_body_owners::<rustc_hir_analysis[e5ecc290d4e18640]::check_crate::{closure#4}>::{closure#0} 49: 0x10b57345c - rustc_hir_analysis[e5ecc290d4e18640]::check_crate 50: 0x10bb0a187 - rustc_interface[ee26b89eef7e70ed]::passes::run_required_analyses 51: 0x10bb0d170 - rustc_interface[ee26b89eef7e70ed]::passes::analysis 52: 0x10c982aca - rustc_query_impl[75f1704df0e38289]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[75f1704df0e38289]::query_impl::analysis::dynamic_query::{closure#2}::{closure#0}, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 1usize]>> 53: 0x10c7388ee - rustc_query_system[13e780f7d1d88911]::query::plumbing::try_execute_query::<rustc_query_impl[75f1704df0e38289]::DynamicConfig<rustc_query_system[13e780f7d1d88911]::query::caches::SingleCache<rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 1usize]>>, false, false, false>, rustc_query_impl[75f1704df0e38289]::plumbing::QueryCtxt, false> 54: 0x10c98f177 - rustc_query_impl[75f1704df0e38289]::query_impl::analysis::get_query_non_incr::__rust_end_short_backtrace 55: 0x10b0e3e77 - <rustc_interface[ee26b89eef7e70ed]::queries::QueryResult<&rustc_middle[3d266d78c9b198e4]::ty::context::GlobalCtxt>>::enter::<core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>, rustc_driver_impl[e0e0c21a20d76ec4]::run_compiler::{closure#0}::{closure#1}::{closure#5}> 56: 0x10b14195d - rustc_interface[ee26b89eef7e70ed]::interface::run_compiler::<core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>, rustc_driver_impl[e0e0c21a20d76ec4]::run_compiler::{closure#0}>::{closure#1} 57: 0x10b12d59c - std[a533c313b02b6cfe]::sys::backtrace::__rust_begin_short_backtrace::<rustc_interface[ee26b89eef7e70ed]::util::run_in_thread_with_globals<rustc_interface[ee26b89eef7e70ed]::util::run_in_thread_pool_with_globals<rustc_interface[ee26b89eef7e70ed]::interface::run_compiler<core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>, rustc_driver_impl[e0e0c21a20d76ec4]::run_compiler::{closure#0}>::{closure#1}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>>::{closure#0}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>>::{closure#0}::{closure#0}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>> 58: 0x10b143f1a - <<std[a533c313b02b6cfe]::thread::Builder>::spawn_unchecked_<rustc_interface[ee26b89eef7e70ed]::util::run_in_thread_with_globals<rustc_interface[ee26b89eef7e70ed]::util::run_in_thread_pool_with_globals<rustc_interface[ee26b89eef7e70ed]::interface::run_compiler<core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>, rustc_driver_impl[e0e0c21a20d76ec4]::run_compiler::{closure#0}>::{closure#1}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>>::{closure#0}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>>::{closure#0}::{closure#0}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>>::{closure#1} as core[907a20686bf8bf7a]::ops::function::FnOnce<()>>::call_once::{shim:vtable#0} 59: 0x10d33f95b - std::sys::pal::unix::thread::Thread::new::thread_start::hc5e64771570de417 60: 0x7ff801f5318b - __pthread_start rustc version: 1.83.0-nightly (7608018cb 2024-09-29) platform: x86_64-apple-darwin query stack during panic: #0 [eval_to_allocation_raw] const-evaluating + checking `<impl at ./9A756-1.rs:17:1: 17:18>::C` #1 [eval_to_valtree] evaluating type-level constant #2 [typeck] type-checking `main` #3 [analysis] running analysis passes on this crate end of query stack ``` </p> </details> ### Note <strong>Ice location:</strong> https://github.com/rust-lang/rust/blob/7608018cbdac9e55d0d13529cf43adc33d53efcf/compiler/rustc_middle/src/mir/interpret/allocation.rs#L257-L271
I-ICE,T-compiler,C-bug,S-bug-has-test,F-associated_const_equality,F-generic_const_items
low
Critical
2,555,431,963
rust
ICE: `invalid pointer unsizing` &str -> str
<!-- Thank you for finding an Internal Compiler Error! 🧊 If possible, try to provide a minimal verifiable example. You can read "Rust Bug Minimization Patterns" for how to create smaller examples. http://blog.pnkfx.org/blog/2019/11/18/rust-bug-minimization-patterns/ --> ### Code ```Rust #![feature(coerce_unsized)] use std::ops; trait Trait3<A> {} impl<A> ops::CoerceUnsized<A> for A where A: ?Sized {} fn main() { println!("Hello, world!"); } ``` Command: `rustc ./03FC8.rs` ### Meta <!-- If you're using the stable version of the compiler, you should also check if the bug also exists in the beta or nightly versions. --> `rustc --version --verbose`: ``` rustc 1.83.0-nightly (7608018cb 2024-09-29) binary: rustc commit-hash: 7608018cbdac9e55d0d13529cf43adc33d53efcf commit-date: 2024-09-29 host: x86_64-apple-darwin release: 1.83.0-nightly LLVM version: 19.1.0 ``` ### Error output ``` error[E0210]: type parameter `A` must be used as the type parameter for some local type (e.g., `MyStruct<A>`) --> ./03FC8.rs:9:6 | 9 | impl<A> ops::CoerceUnsized<A> for A where A: ?Sized {} | ^ type parameter `A` must be used as the type parameter for some local type | = note: implementing a foreign trait is only possible if at least one of the types for which it is implemented is local = note: only traits defined in the current crate can be implemented for a type parameter error[E0376]: the trait `CoerceUnsized` may only be implemented for a coercion between structures --> ./03FC8.rs:9:1 | 9 | impl<A> ops::CoerceUnsized<A> for A where A: ?Sized {} | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ error: internal compiler error: compiler/rustc_const_eval/src/interpret/cast.rs:465:17: invalid pointer unsizing &str -> str --> ./03FC8.rs:14:14 | 14 | println!("Hello, world!"); | ^^^^^^^^^^^^^^^ ``` <!-- Include a backtrace in the code block by setting `RUST_BACKTRACE=1` in your environment. E.g. `RUST_BACKTRACE=1 cargo build`. --> <details><summary><strong>Backtrace</strong></summary> <p> ``` thread 'rustc' panicked at compiler/rustc_const_eval/src/interpret/cast.rs:465:17: Box<dyn Any> stack backtrace: 0: 0x113f193e7 - std::backtrace::Backtrace::create::h697070a74b47c5a1 1: 0x113f19335 - std::backtrace::Backtrace::force_capture::hd4e20dc44637c239 2: 0x111d3a7ce - std[a533c313b02b6cfe]::panicking::update_hook::<alloc[3194a7a73d70d243]::boxed::Box<rustc_driver_impl[e0e0c21a20d76ec4]::install_ice_hook::{closure#0}>>::{closure#0} 3: 0x113f34af8 - std::panicking::rust_panic_with_hook::h9b093984c0b2ae3e 4: 0x111dbb617 - std[a533c313b02b6cfe]::panicking::begin_panic::<rustc_errors[e7c766416e916d12]::ExplicitBug>::{closure#0} 5: 0x111da7139 - std[a533c313b02b6cfe]::sys::backtrace::__rust_end_short_backtrace::<std[a533c313b02b6cfe]::panicking::begin_panic<rustc_errors[e7c766416e916d12]::ExplicitBug>::{closure#0}, !> 6: 0x116bf2f08 - std[a533c313b02b6cfe]::panicking::begin_panic::<rustc_errors[e7c766416e916d12]::ExplicitBug> 7: 0x111dd0806 - <rustc_errors[e7c766416e916d12]::diagnostic::BugAbort as rustc_errors[e7c766416e916d12]::diagnostic::EmissionGuarantee>::emit_producing_guarantee 8: 0x111baecb1 - <rustc_errors[e7c766416e916d12]::DiagCtxtHandle>::span_bug::<rustc_span[51e7cf5b04f4660f]::span_encoding::Span, alloc[3194a7a73d70d243]::string::String> 9: 0x111be2a98 - rustc_middle[3d266d78c9b198e4]::util::bug::opt_span_bug_fmt::<rustc_span[51e7cf5b04f4660f]::span_encoding::Span>::{closure#0} 10: 0x111be2ad7 - rustc_middle[3d266d78c9b198e4]::ty::context::tls::with_opt::<rustc_middle[3d266d78c9b198e4]::util::bug::opt_span_bug_fmt<rustc_span[51e7cf5b04f4660f]::span_encoding::Span>::{closure#0}, !>::{closure#0} 11: 0x111bc9975 - rustc_middle[3d266d78c9b198e4]::ty::context::tls::with_context_opt::<rustc_middle[3d266d78c9b198e4]::ty::context::tls::with_opt<rustc_middle[3d266d78c9b198e4]::util::bug::opt_span_bug_fmt<rustc_span[51e7cf5b04f4660f]::span_encoding::Span>::{closure#0}, !>::{closure#0}, !> 12: 0x116be1932 - rustc_middle[3d266d78c9b198e4]::util::bug::span_bug_fmt::<rustc_span[51e7cf5b04f4660f]::span_encoding::Span> 13: 0x111c6eba7 - <rustc_const_eval[298ced3d0b3bd135]::interpret::eval_context::InterpCx<rustc_const_eval[298ced3d0b3bd135]::const_eval::machine::CompileTimeMachine>>::unsize_into_ptr 14: 0x111c9d0bd - <rustc_const_eval[298ced3d0b3bd135]::interpret::eval_context::InterpCx<rustc_const_eval[298ced3d0b3bd135]::const_eval::machine::CompileTimeMachine>>::eval_statement 15: 0x111cb9c6c - rustc_const_eval[298ced3d0b3bd135]::const_eval::eval_queries::eval_to_allocation_raw_provider 16: 0x113580cfc - rustc_query_impl[75f1704df0e38289]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_allocation_raw::dynamic_query::{closure#2}::{closure#0}, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>> 17: 0x11353173e - <rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_allocation_raw::dynamic_query::{closure#2} as core[907a20686bf8bf7a]::ops::function::FnOnce<(rustc_middle[3d266d78c9b198e4]::ty::context::TyCtxt, rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>)>>::call_once 18: 0x113348579 - rustc_query_system[13e780f7d1d88911]::query::plumbing::try_execute_query::<rustc_query_impl[75f1704df0e38289]::DynamicConfig<rustc_query_system[13e780f7d1d88911]::query::caches::DefaultCache<rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>>, false, false, false>, rustc_query_impl[75f1704df0e38289]::plumbing::QueryCtxt, false> 19: 0x1135b823c - rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_allocation_raw::get_query_non_incr::__rust_end_short_backtrace 20: 0x111bc62a2 - rustc_middle[3d266d78c9b198e4]::query::plumbing::query_get_at::<rustc_query_system[13e780f7d1d88911]::query::caches::DefaultCache<rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>>> 21: 0x111cb6763 - rustc_const_eval[298ced3d0b3bd135]::const_eval::eval_queries::eval_to_const_value_raw_provider 22: 0x11358141c - rustc_query_impl[75f1704df0e38289]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_const_value_raw::dynamic_query::{closure#2}::{closure#0}, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>> 23: 0x11353704e - <rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_const_value_raw::dynamic_query::{closure#2} as core[907a20686bf8bf7a]::ops::function::FnOnce<(rustc_middle[3d266d78c9b198e4]::ty::context::TyCtxt, rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>)>>::call_once 24: 0x113348579 - rustc_query_system[13e780f7d1d88911]::query::plumbing::try_execute_query::<rustc_query_impl[75f1704df0e38289]::DynamicConfig<rustc_query_system[13e780f7d1d88911]::query::caches::DefaultCache<rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>>, false, false, false>, rustc_query_impl[75f1704df0e38289]::plumbing::QueryCtxt, false> 25: 0x1135b8e2c - rustc_query_impl[75f1704df0e38289]::query_impl::eval_to_const_value_raw::get_query_non_incr::__rust_end_short_backtrace 26: 0x112b74b88 - rustc_middle[3d266d78c9b198e4]::query::plumbing::query_get_at::<rustc_query_system[13e780f7d1d88911]::query::caches::DefaultCache<rustc_middle[3d266d78c9b198e4]::ty::ParamEnvAnd<rustc_middle[3d266d78c9b198e4]::mir::interpret::GlobalId>, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 24usize]>>> 27: 0x112b79563 - <rustc_middle[3d266d78c9b198e4]::ty::context::TyCtxt>::const_eval_global_id 28: 0x112b78e50 - <rustc_middle[3d266d78c9b198e4]::ty::context::TyCtxt>::const_eval_resolve 29: 0x112dbf811 - <rustc_const_eval[298ced3d0b3bd135]::interpret::eval_context::InterpCx<rustc_const_eval[298ced3d0b3bd135]::const_eval::dummy_machine::DummyMachine>>::eval_mir_constant::{closure#0} 30: 0x112ec3849 - <rustc_mir_transform[f683d3911b24b415]::known_panics_lint::ConstPropagator>::eval_constant 31: 0x112ec47e6 - <rustc_mir_transform[f683d3911b24b415]::known_panics_lint::ConstPropagator as rustc_middle[3d266d78c9b198e4]::mir::visit::Visitor>::visit_assign 32: 0x112ec9df2 - <rustc_mir_transform[f683d3911b24b415]::known_panics_lint::ConstPropagator as rustc_middle[3d266d78c9b198e4]::mir::visit::Visitor>::visit_basic_block_data 33: 0x112ec41bc - <rustc_mir_transform[f683d3911b24b415]::known_panics_lint::ConstPropagator as rustc_middle[3d266d78c9b198e4]::mir::visit::Visitor>::visit_body 34: 0x112ec0b70 - <rustc_mir_transform[f683d3911b24b415]::known_panics_lint::KnownPanicsLint as rustc_mir_transform[f683d3911b24b415]::pass_manager::MirLint>::run_lint 35: 0x112e652fa - rustc_mir_transform[f683d3911b24b415]::pass_manager::run_passes_inner 36: 0x112f155c9 - rustc_mir_transform[f683d3911b24b415]::run_analysis_to_runtime_passes 37: 0x112f1517a - rustc_mir_transform[f683d3911b24b415]::mir_drops_elaborated_and_const_checked 38: 0x11358265a - rustc_query_impl[75f1704df0e38289]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[75f1704df0e38289]::query_impl::mir_drops_elaborated_and_const_checked::dynamic_query::{closure#2}::{closure#0}, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 8usize]>> 39: 0x1133d8f4e - rustc_query_system[13e780f7d1d88911]::query::plumbing::try_execute_query::<rustc_query_impl[75f1704df0e38289]::DynamicConfig<rustc_query_system[13e780f7d1d88911]::query::caches::VecCache<rustc_span[51e7cf5b04f4660f]::def_id::LocalDefId, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 8usize]>>, false, false, false>, rustc_query_impl[75f1704df0e38289]::plumbing::QueryCtxt, false> 40: 0x1135982ea - rustc_query_impl[75f1704df0e38289]::query_impl::mir_drops_elaborated_and_const_checked::get_query_non_incr::__rust_end_short_backtrace 41: 0x11270ae48 - rustc_interface[ee26b89eef7e70ed]::passes::run_required_analyses 42: 0x11270d170 - rustc_interface[ee26b89eef7e70ed]::passes::analysis 43: 0x113582aca - rustc_query_impl[75f1704df0e38289]::plumbing::__rust_begin_short_backtrace::<rustc_query_impl[75f1704df0e38289]::query_impl::analysis::dynamic_query::{closure#2}::{closure#0}, rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 1usize]>> 44: 0x1133388ee - rustc_query_system[13e780f7d1d88911]::query::plumbing::try_execute_query::<rustc_query_impl[75f1704df0e38289]::DynamicConfig<rustc_query_system[13e780f7d1d88911]::query::caches::SingleCache<rustc_middle[3d266d78c9b198e4]::query::erase::Erased<[u8; 1usize]>>, false, false, false>, rustc_query_impl[75f1704df0e38289]::plumbing::QueryCtxt, false> 45: 0x11358f177 - rustc_query_impl[75f1704df0e38289]::query_impl::analysis::get_query_non_incr::__rust_end_short_backtrace 46: 0x111ce3e77 - <rustc_interface[ee26b89eef7e70ed]::queries::QueryResult<&rustc_middle[3d266d78c9b198e4]::ty::context::GlobalCtxt>>::enter::<core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>, rustc_driver_impl[e0e0c21a20d76ec4]::run_compiler::{closure#0}::{closure#1}::{closure#5}> 47: 0x111d4195d - rustc_interface[ee26b89eef7e70ed]::interface::run_compiler::<core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>, rustc_driver_impl[e0e0c21a20d76ec4]::run_compiler::{closure#0}>::{closure#1} 48: 0x111d2d59c - std[a533c313b02b6cfe]::sys::backtrace::__rust_begin_short_backtrace::<rustc_interface[ee26b89eef7e70ed]::util::run_in_thread_with_globals<rustc_interface[ee26b89eef7e70ed]::util::run_in_thread_pool_with_globals<rustc_interface[ee26b89eef7e70ed]::interface::run_compiler<core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>, rustc_driver_impl[e0e0c21a20d76ec4]::run_compiler::{closure#0}>::{closure#1}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>>::{closure#0}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>>::{closure#0}::{closure#0}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>> 49: 0x111d43f1a - <<std[a533c313b02b6cfe]::thread::Builder>::spawn_unchecked_<rustc_interface[ee26b89eef7e70ed]::util::run_in_thread_with_globals<rustc_interface[ee26b89eef7e70ed]::util::run_in_thread_pool_with_globals<rustc_interface[ee26b89eef7e70ed]::interface::run_compiler<core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>, rustc_driver_impl[e0e0c21a20d76ec4]::run_compiler::{closure#0}>::{closure#1}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>>::{closure#0}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>>::{closure#0}::{closure#0}, core[907a20686bf8bf7a]::result::Result<(), rustc_span[51e7cf5b04f4660f]::ErrorGuaranteed>>::{closure#1} as core[907a20686bf8bf7a]::ops::function::FnOnce<()>>::call_once::{shim:vtable#0} 50: 0x113f3f95b - std::sys::pal::unix::thread::Thread::new::thread_start::hc5e64771570de417 51: 0x7ff801f5318b - __pthread_start rustc version: 1.83.0-nightly (7608018cb 2024-09-29) platform: x86_64-apple-darwin query stack during panic: #0 [eval_to_allocation_raw] const-evaluating + checking `main::promoted[0]` #1 [eval_to_const_value_raw] simplifying constant for the type system `main::promoted[0]` #2 [mir_drops_elaborated_and_const_checked] elaborating drops for `main` #3 [analysis] running analysis passes on this crate end of query stack ``` </p> </details> ### Note <strong>Ice location: </strong> https://github.com/rust-lang/rust/blob/7608018cbdac9e55d0d13529cf43adc33d53efcf/compiler/rustc_const_eval/src/interpret/cast.rs#L460-L471 @rustbot label +F-coerce_unsized
I-ICE,T-compiler,C-bug,F-coerce_unsized,S-has-mcve,S-bug-has-test
low
Critical
2,555,436,830
godot
Android Editor [4.4.dev2]: Heavy Lag and Glitches during editing mash and mertarial.
### Tested versions 4.4.dev2 ### System information Android 10 - Godot 4.4.dev2 - Compatibility Renderer ### Issue description Editor is glitching and lagging too much when editing mash and mertarial in android editor. Need to optimize. https://github.com/user-attachments/assets/d1e0913f-c4ce-4e77-a4b2-1b932fad1427
performance
low
Major
2,555,443,085
flutter
EnsureVisible does not work going backwards on a CustomScrollView set to primary inheriting from NestedScrollView
### Steps to reproduce Steps: 1. Scroll past an item with the global key in the inner scroll view 2. Press the floating action button to navigate to the item with the key ('no') 3. Observe Notes: - works fine when primary is set to false on the CustomScrollView - works fine when there is no content inside of the outer scroll view (headersliverbuilder) - works fine when you have not scrolled past the item ('no') ### Expected results Should navigate to the item with the global key ('no') ### Actual results Navigated to the top of the screen, above the outer scroll view (green box) ### Code sample <details open><summary>Code sample</summary> ```dart import 'package:flutter/material.dart'; const String target = 'no'; void main() { runApp(const MainApp()); } class MainApp extends StatefulWidget { const MainApp({super.key}); @override State<MainApp> createState() => _MainAppState(); } final GlobalKey<NestedScrollViewState> nestedKey = GlobalKey(); final GlobalKey scrollKey = GlobalKey(); class _MainAppState extends State<MainApp> { List<String> stuff = [ "hello", "goodbye", "whats up", "howdy", "no", "yes", "maybe", "possibly", "likely", "could be" ]; @override Widget build(BuildContext context) { return MaterialApp( home: Scaffold( body: NestedScrollView( key: nestedKey, headerSliverBuilder: (context, innerBoxScrolled) => <Widget>[ SliverToBoxAdapter( child: Container( height: 300, color: Colors.green, ), ), ], body: SizedBox( height: 500, child: CustomScrollView(primary: true, slivers: [ SliverToBoxAdapter( child: Column( children: stuff .map((e) => GestureDetector( onTap: () {}, child: SizedBox( height: 200, child: ListTile( key: e == target ? scrollKey : null, leading: Text(e), )), )) .toList(), ), ), ]), ), ), floatingActionButton: FloatingActionButton(onPressed: () { if (nestedKey.currentState != null) { Scrollable.ensureVisible(scrollKey.currentContext!); } }), )); } } ``` </details> ### Screenshots or Video <details open> <summary>Screenshots / Video demonstration</summary> [Upload media here] https://github.com/user-attachments/assets/11db6a74-17fd-421a-9cdf-868918f425b6 </details> ### Logs Logs: empty <details open><summary>Logs</summary> ```console Showing iPhone 16 Pro Max logs: ``` </details> ### Flutter Doctor output <details open><summary>Doctor output</summary> ```console flutter doctor -v [✓] Flutter (Channel stable, 3.24.3, on macOS 14.5 23F79 darwin-arm64, locale en-AU) • Flutter version 3.24.3 on channel stable at /Users/**/Desktop/development/flutter • Upstream repository https://github.com/flutter/flutter.git • Framework revision 2663184aa7 (3 weeks ago), 2024-09-11 16:27:48 -0500 • Engine revision 36335019a8 • Dart version 3.5.3 • DevTools version 2.37.3 [✓] Android toolchain - develop for Android devices (Android SDK version 35.0.0) • Android SDK at /Users/**/Library/Android/sdk • Platform android-35, build-tools 35.0.0 • ANDROID_HOME = /Users/**/Library/Android/sdk • Java binary at: /Applications/Android Studio.app/Contents/jbr/Contents/Home/bin/java • Java version OpenJDK Runtime Environment (build 17.0.10+0-17.0.10b1087.21-11609105) • All Android licenses accepted. [✓] Xcode - develop for iOS and macOS (Xcode 16.0) • Xcode at /Applications/Xcode.app/Contents/Developer • Build 16A242d • CocoaPods version 1.15.2 [✓] Chrome - develop for the web • Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome [✓] Android Studio (version 2024.1) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/6351-dart • Java version OpenJDK Runtime Environment (build 17.0.10+0-17.0.10b1087.21-11609105) [✓] VS Code (version 1.93.1) • VS Code at /Applications/Visual Studio Code.app/Contents • Flutter extension version 3.96.0 [✓] Connected device (4 available) • iPhone 16 Pro Max (mobile) • 7EF3EBC3-5032-463D-9189-66B86C4FFF9E • ios • com.apple.CoreSimulator.SimRuntime.iOS-18-0 (simulator) • macOS (desktop) • macos • darwin-arm64 • macOS 14.5 23F79 darwin-arm64 • Mac Designed for iPad (desktop) • mac-designed-for-ipad • darwin • macOS 14.5 23F79 darwin-arm64 • Chrome (web) • chrome • web-javascript • Google Chrome 129.0.6668.70 [✓] Network resources • All expected network resources are available. • No issues found! ``` </details>
framework,f: scrolling,has reproducible steps,P2,team-framework,triaged-framework,found in release: 3.24,found in release: 3.26
low
Minor
2,555,457,404
pytorch
torch.set_float32_matmul_precision('medium') slower than regular BF16 matmul
### 🐛 Describe the bug According to the docs, setting the torch float32 matmul precision to 'medium' uses BF16 for the internal GEMM with a FP32 accumulator and returns the FP32 accumulator. My benchmark code (see below) says this is slower than rounding from FP32 to BF16 and then doing the matmul in BF16 and returning a BF16 accumulator. Why is this true? Shouldn't both be using the same BF16 tensorcores and the difference is in the accumulator precision? Is the overhead of using a higher precision accumulator *that much*? Benchmark script: ``` import torch import time sizes = [512, 1024, 2048, 4096, 8192, 16384] iters = 100 print('Regular FP32') for sz in sizes: a = torch.randn(sz, sz, device=0) b = torch.randn(sz, sz, device=0) total = 0 for i in range(iters): s = time.time() c = a @ b torch.cuda.synchronize() total += time.time() - s print(sz, total / iters) print() print('FP32 w/BF16 internal matmul') torch.set_float32_matmul_precision('medium') for sz in sizes: a = torch.randn(sz, sz, device=0) b = torch.randn(sz, sz, device=0) total = 0 for i in range(iters): s = time.time() c = a @ b torch.cuda.synchronize() total += time.time() - s print(sz, total / iters) print() print('regular BF16') for sz in sizes: a = torch.randn(sz, sz, device=0) b = torch.randn(sz, sz, device=0) total = 0 for i in range(iters): s = time.time() c = a.to(torch.bfloat16) @ b.to(torch.bfloat16) torch.cuda.synchronize() total += time.time() - s print(sz, total / iters) ``` Output on a H100 SXM: ``` Regular FP32 512 0.0016167497634887696 1024 7.171869277954102e-05 2048 0.0005328941345214843 4096 0.0026724767684936524 8192 0.021244921684265138 16384 0.16401963710784911 FP32 w/BF16 internal matmul 512 0.00036259174346923827 1024 0.00013562202453613282 2048 0.0001180720329284668 4096 0.0004148006439208984 8192 0.0031256651878356933 16384 0.02780592679977417 regular BF16 512 0.0008567142486572266 1024 0.00077117919921875 2048 7.407903671264648e-05 4096 0.00030952930450439453 8192 0.001969149112701416 16384 0.014865560531616211 ``` ### Versions ``` Collecting environment information... PyTorch version: 2.4.0+cu121 Is debug build: False CUDA used to build PyTorch: 12.1 ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.3 LTS (x86_64) GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version: Could not collect CMake version: version 3.22.1 Libc version: glibc-2.35 Python version: 3.12.4 | packaged by Anaconda, Inc. | (main, Jun 18 2024, 15:12:24) [GCC 11.2.0] (64-bit runtime) Python platform: Linux-5.15.0-94-generic-x86_64-with-glibc2.35 Is CUDA available: True CUDA runtime version: Could not collect CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: NVIDIA H100 80GB HBM3 GPU 1: NVIDIA H100 80GB HBM3 GPU 2: NVIDIA H100 80GB HBM3 GPU 3: NVIDIA H100 80GB HBM3 GPU 4: NVIDIA H100 80GB HBM3 GPU 5: NVIDIA H100 80GB HBM3 GPU 6: NVIDIA H100 80GB HBM3 GPU 7: NVIDIA H100 80GB HBM3 Nvidia driver version: 535.154.05 cuDNN version: Could not collect HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 48 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 176 On-line CPU(s) list: 0-175 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) Platinum 8468V CPU family: 6 Model: 143 Thread(s) per core: 2 Core(s) per socket: 44 Socket(s): 2 Stepping: 8 BogoMIPS: 4800.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves avx_vnni avx512_bf16 wbnoinvd arat avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b fsrm md_clear serialize tsxldtrk amx_bf16 avx512_fp16 amx_tile amx_int8 arch_capabilities Virtualization: VT-x Hypervisor vendor: KVM Virtualization type: full L1d cache: 5.5 MiB (176 instances) L1i cache: 5.5 MiB (176 instances) L2 cache: 352 MiB (88 instances) L3 cache: 32 MiB (2 instances) NUMA node(s): 2 NUMA node0 CPU(s): 0-87 NUMA node1 CPU(s): 88-175 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Unknown: No mitigations Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Enhanced IBRS, IBPB conditional, RSB filling, PBRSB-eIBRS SW sequence Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Mitigation; TSX disabled Versions of relevant libraries: [pip3] numpy==1.26.4 [pip3] torch==2.4.0 [pip3] triton==3.0.0 [conda] numpy 1.26.4 pypi_0 pypi [conda] torch 2.4.0 pypi_0 pypi [conda] triton 3.0.0 pypi_0 pypi ``` cc @msaroufim @ptrblck
module: performance,module: cuda,triaged
low
Critical
2,555,469,914
ui
[bug]: Unexpected token < in JSON at position 0
### Describe the bug # Bug Report: Error when adding v0.dev block to shadcn-ui ![GYsfOF4aoAAciwy](https://github.com/user-attachments/assets/ab2d9949-6679-47a2-bf94-d69476cf3b7a) ## Description When attempting to add a block from v0.dev to a shadcn-ui project, an error occurs during the execution of the `npx shadcn add` command. ## Steps to Reproduce 1. Start in a directory without a `package.json` file 2. Run the command: `npx shadcn add https://MYURL.v0.build/` 3. Follow the prompts to create a new Next.js project 4. The process fails at the "Checking registry" step ## Expected Behavior The command should successfully add the v0.dev block to the shadcn-ui project. ## Actual Behavior The process fails with the following error: ``` Something went wrong. Please check the error below for more details. If the problem persists, please open an issue on GitHub. Unexpected token < in JSON at position 0 ``` ## Environment - Operating System: macOS - npm version: 10.8.3 (after update) - Node.js version: v18.18.2 ## Additional Information - The issue persists after updating npm to the latest version (10.8.3) - The error occurs during the "Checking registry" step ## Possible Cause The error message "Unexpected token < in JSON at position 0" typically indicates that the response received is not valid JSON. This could be due to: 1. Network issues causing an HTML error page to be returned instead of JSON data 2. The registry endpoint returning an invalid response 3. A problem with the URL or the content at `https://MYURL.v0.build/` ### Affected component/components Published Blocks in v0 ### How to reproduce 1. Start in a directory without a `package.json` file 2. Run the command: `npx shadcn add https://MYURL.v0.build/` 3. Follow the prompts to create a new Next.js project 4. The process fails at the "Checking registry" step ### Codesandbox/StackBlitz link _No response_ ### Logs _No response_ ### System Info ```bash M2 MacBook Air, 14.6.1 (23G93) ``` ### Before submitting - [X] I've made research efforts and searched the documentation - [X] I've searched for existing issues
bug
low
Critical
2,555,472,676
TypeScript
Bug: null can be excluded by !== null in flow before typeof === 'object', but not excluded by predicate function.
### 🔎 Search Terms "null", "control flow", "narrowing", "narrow", "exclusion", "object", "typeof" ### 🕗 Version & Regression Information Potentially, we could expect the behaviors of `if (a === null)` and `if ((v => v !== null)(a))` (inferred as predicate since TS 5.5) should be the same, but it's actually not in specific cases. If it's combined with `typeof a === 'object'` check in the same flow, first one is correctly omits the null possibility, while second one doesn't. [<img width="540" alt="image" src="https://github.com/user-attachments/assets/3b3e2417-4912-456d-b2dc-b0dad8dec016">](https://www.typescriptlang.org/play/?ts=5.6.2#code/MYewdgzgLgBAlhAcgVwDapgXhgCgG4BcMyYA1mCAO5gCUWAfDHlptmGqgNwBQA9LzEEwAegH5uoSLABmWXAEMiJclVoMYAb25D4snAhToc8mnQBOAUyjIzYHjrh6oATwAOFkLPkwAhKxgA5CAARgBWFsBQAeZWNnbaQvL2-CLiAL483BLg0DAA5nLGSmQU1HSYjFoOet6sbBwx1rb2Qo64Lu6eMN5+2EFhEVGNcS2CSdopYtwZQA) - This changed between versions: - I confirmed that the exclusion in flow can be done since TS4.5. It's not regression, but the new feature seems to have been partially applied. - TS4.4 https://www.typescriptlang.org/play/?ts=4.4.4#code/MYewdgzgLgBAlhAcgVwDapgXhgCgG4BcMyYA1mCAO5gCURe8EMYaGmAfDA5j86wNwAoAPTCY4mAD0A-INCRYAMyy4AhkRLkqtLJwDegifGU4EKdDlU0aMAE4BTKMltghRuCagBPAA72QyqowAIS8AOQgAEYAVvbAUGE2Dk4ubhKqbqJSsgC+QoJy4NAwAOYqlhpkFNQ2HDAG7iZBPNgs6EmOzq6GEh643n4BMEGh2BExcQkdKd1GGYZZMoJ5QA - TS4.5 https://www.typescriptlang.org/play/?ts=4.5.5#code/MYewdgzgLgBAlhAcgVwDapgXhgCgG4BcMyYA1mCAO5gCURe8EMYaGmAfDA5j86wNwAoAPTCY4mAD0A-INCRYAMyy4AhkRLkqtLJwDegifGU4EKdDlU0aMAE4BTKMltghRuCagBPAA72QyqowAIS8AOQgAEYAVvbAUGE2Dk4ubhKqbqJSsgC+QoJy4NAwAOYqlhpkFNQ2HDAG7iZBPNgs6EmOzq6GEh643n4BMEGh2BExcQkdKd1GGYZZMoJ5QA - This changed in commit or PR (not sure for now, sorry) ### ⏯ Playground Link https://www.typescriptlang.org/play/?ts=5.6.2#code/MYewdgzgLgBAlhAcgVwDapgXhgCgG4BcMyYA1mCAO5gCUWAfDHlptmGqgNwBQA9LzEEwAegH5uoSLABmWXAEMiJclVoMYAb25D4snAhToc8mnQBOAUyjIzYHjrh6oATwAOFkLPkwAhKxgA5CAARgBWFsBQAeZWNnbaQvL2-CLiAL483BLg0DAA5nLGSmQU1HSYjFoOet6sbBwx1rb2Qo64Lu6eMN5+2EFhEVGNcS2CSdopYtwZQA ### 💻 Code ```ts const isNull = (v: unknown) => v === null; // ^? const f = (a: unknown) => { if (isNull(a)) return; if (typeof a !== 'object') return; a; // ^? }; const g = (a: unknown) => { if (a === null) return; if (typeof a !== 'object') return; a; // ^? }; ``` ### 🙁 Actual behavior Last `a` in `f` is inferred as `object | null`. ### 🙂 Expected behavior Last `a` in `f` would be inferred as `object`. ### Additional information about the issue Here is important to check `typeof === 'object'` after the check for null, because type narrowing step is formally, `unknown` → `object | null` → `object`. `=== null` alone cannot narrow anything on `unknown`. [If you swap the if-statements in above, you can see both is inferred as `object`.](https://www.typescriptlang.org/play/?ts=5.6.2#code/MYewdgzgLgBAlhAcgVwDapgXhgCgG4BcMyYA1mCAO5gCUWAfDHlptmGqgNwBQA9LzEEwAegH5uoSLABmWXAEMiJclVoMYAb25D4snFACeABwCmIWfJgBCVjADkIAEYArE8Ch26AJxNRkXsB4dOD0EFHQceRpvX39A7SF5IP4RcQBfHm4JcGgYAHM5SKUyCmo6TEYtYL1DU3MYSxtsBxc3Dxi-AKChEIUWNg4OuO7BJO0UsW4MoA)
Help Wanted,Possible Improvement
low
Critical
2,555,500,389
pytorch
pytorch 2.1.2+cu118, RTX 8000, backward show RuntimeError: Expected is_sm80 || is_sm90 to be true, but got false.
### 🐛 Describe the bug When I want to train qwen2.5-7B-instruct with using deepspeed, it shows the following erre: ``` Traceback (most recent call last): File "/home/work/ybs/deeplm/LLM/train.py", line 195, in <module> run_app() File "/home/work/ybs/deeplm/LLM/train.py", line 180, in run_app train( File "/home/work/ybs/deeplm/LLM/train.py", line 125, in train model.backward(loss) File "/root/miniconda3/envs/llm-agent/lib/python3.9/site-packages/deepspeed/utils/nvtx.py", line 15, in wrapped_fn ret_val = func(*args, **kwargs) File "/root/miniconda3/envs/llm-agent/lib/python3.9/site-packages/deepspeed/runtime/engine.py", line 1962, in backward self.optimizer.backward(loss, retain_graph=retain_graph) File "/root/miniconda3/envs/llm-agent/lib/python3.9/site-packages/deepspeed/runtime/fp16/fused_optimizer.py", line 335, in backward scaled_loss.backward(create_graph=create_graph, retain_graph=retain_graph) File "/root/miniconda3/envs/llm-agent/lib/python3.9/site-packages/torch/_tensor.py", line 492, in backward torch.autograd.backward( File "/root/miniconda3/envs/llm-agent/lib/python3.9/site-packages/torch/autograd/__init__.py", line 251, in backward Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass File "/root/miniconda3/envs/llm-agent/lib/python3.9/site-packages/torch/autograd/function.py", line 288, in apply return user_fn(self, *args) File "/root/miniconda3/envs/llm-agent/lib/python3.9/site-packages/torch/utils/checkpoint.py", line 288, in backward torch.autograd.backward(outputs_with_grad, args_with_grad) File "/root/miniconda3/envs/llm-agent/lib/python3.9/site-packages/torch/autograd/__init__.py", line 251, in backward Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass RuntimeError: Expected is_sm80 || is_sm90 to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.) ``` The env setting is listed below. I noticed that there was a same issue a year ago, and it was already solved in the last pytorch version. So I am confused why with pytorch 2.1.2+cu118, the error still shows again. Recall: when head_dim > 64, the error shows up, https://github.com/pytorch/pytorch/pull/99105 ### Versions Collecting environment information... PyTorch version: 2.1.2+cu118 Is debug build: False CUDA used to build PyTorch: 11.8 ROCM used to build PyTorch: N/A OS: CentOS Linux 7 (Core) (x86_64) GCC version: (GCC) 7.1.0 Clang version: Could not collect CMake version: Could not collect Libc version: glibc-2.17 Python version: 3.9.19 (main, May 6 2024, 19:43:03) [GCC 11.2.0] (64-bit runtime) Python platform: Linux-4.19.96-x86_64-with-glibc2.17 Is CUDA available: True CUDA runtime version: 11.8.89 CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: Quadro RTX 8000 GPU 1: Quadro RTX 8000 GPU 2: Quadro RTX 8000 GPU 3: Quadro RTX 8000 Nvidia driver version: 525.85.05 cuDNN version: Probably one of the following: /usr/lib64/libcudnn.so.8.9.7 /usr/lib64/libcudnn_adv_infer.so.8.9.7 /usr/lib64/libcudnn_adv_train.so.8.9.7 /usr/lib64/libcudnn_cnn_infer.so.8.9.7 /usr/lib64/libcudnn_cnn_train.so.8.9.7 /usr/lib64/libcudnn_ops_infer.so.8.9.7 /usr/lib64/libcudnn_ops_train.so.8.9.7 HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 80 On-line CPU(s) list: 0-79 Thread(s) per core: 2 Core(s) per socket: 20 Socket(s): 2 NUMA node(s): 2 Vendor ID: GenuineIntel CPU family: 6 Model: 85 Model name: Intel(R) Xeon(R) Gold 5218R CPU @ 2.10GHz Stepping: 7 CPU MHz: 1235.479 CPU max MHz: 4000.0000 CPU min MHz: 800.0000 BogoMIPS: 4200.00 Virtualization: VT-x L1d cache: 32K L1i cache: 32K L2 cache: 1024K L3 cache: 28160K NUMA node0 CPU(s): 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30,32,34,36,38,40,42,44,46,48,50,52,54,56,58,60,62,64,66,68,70,72,74,76,78 NUMA node1 CPU(s): 1,3,5,7,9,11,13,15,17,19,21,23,25,27,29,31,33,35,37,39,41,43,45,47,49,51,53,55,57,59,61,63,65,67,69,71,73,75,77,79 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cdp_l3 invpcid_single intel_ppin ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt clwb intel_pt avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts pku ospke avx512_vnni md_clear flush_l1d arch_capabilities Versions of relevant libraries: [pip3] flake8==6.1.0 [pip3] mypy-extensions==1.0.0 [pip3] numpy==1.22.2 [pip3] torch==2.1.2+cu118 [pip3] torchaudio==2.1.2+cu118 [pip3] torchvision==0.16.2+cu118 [pip3] triton==2.1.0 [conda] numpy 1.22.2 pypi_0 pypi [conda] torch 2.1.2+cu118 pypi_0 pypi [conda] torchaudio 2.1.2+cu118 pypi_0 pypi [conda] torchvision 0.16.2+cu118 pypi_0 pypi [conda] triton 2.1.0 pypi_0 pypi cc @ezyang @gchanan @zou3519 @kadeng @msaroufim @albanD @gqchen @pearu @nikitaved @soulitzer @Varal7 @xmfan @ptrblck @eqy @drisspg @mikaylagawarecki
needs reproduction,module: autograd,module: cuda,triaged,module: sdpa
low
Critical
2,555,589,469
yt-dlp
[youtube:tab] Shorts video. More than 100 from the playlist are not downloaded.
### DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE - [X] I understand that I will be **blocked** if I *intentionally* remove or skip any mandatory\* field ### Checklist - [X] I'm reporting that yt-dlp is broken on a **supported** site - [X] I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels)) - [X] I've checked that all provided URLs are playable in a browser with the same IP and same login details - [X] I've checked that all URLs and arguments with special characters are [properly quoted or escaped](https://github.com/yt-dlp/yt-dlp/wiki/FAQ#video-url-contains-an-ampersand--and-im-getting-some-strange-output-1-2839-or-v-is-not-recognized-as-an-internal-or-external-command) - [X] I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates - [X] I've read the [guidelines for opening an issue](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#opening-an-issue) - [X] I've read about [sharing account credentials](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#are-you-willing-to-share-account-details-if-needed) and I'm willing to share it if required ### Region _No response_ ### Provide a description that is worded well enough to be understood Shorts video. More than 100 from the playlist are not downloaded. If you download from the channel, then everything is ok. ### Provide verbose output that clearly demonstrates the problem - [X] Run **your** yt-dlp command with **-vU** flag added (`yt-dlp -vU <your command line>`) - [X] If using API, add `'verbose': True` to `YoutubeDL` params instead - [X] Copy the WHOLE output (starting with `[debug] Command-line config`) and insert it below ### Complete Verbose Output ```shell tube # /usr/local/bin/yt-dlp -Uv --flat-playlist "https://www.youtube.com/playlist?list=PL5Ib645yg8L0NO1Lj4Z3jaQJ9yeQOSa2r" [debug] Command-line config: ['-Uv', '--flat-playlist', 'https://www.youtube.com/playlist?list=PL5Ib645yg8L0NO1Lj4Z3jaQJ9yeQOSa2r'] [debug] Encodings: locale UTF-8, fs utf-8, pref UTF-8, out utf-8, error utf-8, screen utf-8 [debug] yt-dlp version nightly@2024.09.26.232938 from yt-dlp/yt-dlp-nightly-builds [eabb4680f] (pip) [debug] Python 3.8.17 (CPython x86_64 64bit) - Linux-5.4.17-2136.334.6.1.el8uek.x86_64-x86_64-with-glibc2.2.5 (OpenSSL 1.1.1k FIPS 25 Mar 2021, glibc 2.28) [debug] exe versions: ffmpeg 4.2.10, ffprobe 4.2.10 [debug] Optional libraries: Cryptodome-3.20.0, brotli-1.1.0, certifi-2024.08.30, mutagen-1.47.0, requests-2.32.3, sqlite3-3.26.0, urllib3-2.2.3, websockets-13.0.1 [debug] Proxy map: {} [debug] Request Handlers: urllib, requests, websockets [debug] Loaded 1838 extractors [debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp-nightly-builds/releases/latest [debug] Downloading _update_spec from https://github.com/yt-dlp/yt-dlp-nightly-builds/releases/latest/download/_update_spec Current version: nightly@2024.09.26.232938 from yt-dlp/yt-dlp-nightly-builds Latest version: nightly@2024.09.29.232819 from yt-dlp/yt-dlp-nightly-builds ERROR: You installed yt-dlp with pip or using the wheel from PyPi; Use that to update [youtube:tab] Extracting URL: https://www.youtube.com/playlist?list=PL5Ib645yg8L0NO1Lj4Z3jaQJ9yeQOSa2r [youtube:tab] PL5Ib645yg8L0NO1Lj4Z3jaQJ9yeQOSa2r: Downloading webpage [youtube:tab] PL5Ib645yg8L0NO1Lj4Z3jaQJ9yeQOSa2r: Redownloading playlist API JSON with unavailable videos [download] Downloading playlist: Shortss [youtube:tab] PL5Ib645yg8L0NO1Lj4Z3jaQJ9yeQOSa2r page 1: Downloading API JSON WARNING: [youtube:tab] Incomplete data received. Retrying (1/3)... [youtube:tab] PL5Ib645yg8L0NO1Lj4Z3jaQJ9yeQOSa2r page 1: Downloading API JSON WARNING: [youtube:tab] Incomplete data received. Retrying (2/3)... [youtube:tab] PL5Ib645yg8L0NO1Lj4Z3jaQJ9yeQOSa2r page 1: Downloading API JSON WARNING: [youtube:tab] Incomplete data received. Retrying (3/3)... [youtube:tab] PL5Ib645yg8L0NO1Lj4Z3jaQJ9yeQOSa2r page 1: Downloading API JSON WARNING: [youtube:tab] Incomplete data received. Giving up after 3 retries [youtube:tab] Playlist Shortss: Downloading 100 items of 301 [debug] The information of all playlist entries will be held in memory [download] Downloading item 1 of 100 [download] Downloading item 2 of 100 ... ```
site-bug,triage,site:youtube
low
Critical
2,555,605,005
PowerToys
Multi-Desktop Powertoy FancyZone settings would change from existing settings after rebooting computer
### Microsoft PowerToys version 0.84.1 ### Installation method GitHub ### Running as admin None ### Area(s) with issue? FancyZones ### Steps to reproduce Multi-Desktop Powertoy FancyZone settings would change from existing settings after rebooting computer For example My settings Multi-Desktop1: Monitor1 = A, Monitor2 = B Multi-Desktop2: Monitor1 = C, Monitor2 = D After rebooting Multi-Desktop1: Monitor1 = one of the default templates, Monitor2 = one of the default templates Multi-Desktop2: Monitor1 = A, Monitor2 = B It looks like this. Could you please fix this so that the settings are preserved after a reboot? ### ✔️ Expected Behavior Keep existing settings ### ❌ Actual Behavior Different settings ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Minor
2,555,669,370
node
Require of cached modules is not working after a while
### Version 18.20.4 ### Platform ```text Linux b3fb3508-30f0-4ff0-4220-0bc8 5.15.0-117-generic #127-Ubuntu SMP Fri Jul 5 20:13:28 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux ``` ### Subsystem - ### What steps will reproduce the bug? The issue we face happens as following: During the runtime of our backend, we happen to need some code extensions which we basically download and save on the disk ( on demand ofc ), and after we got them in place we require them in the code using the filePath that is being returned by the function which does the download & save on disk. [Yes we know thats not safe, but that works for us right now] Now, after some weeks or days even of the backend running, these demanded files that are can be found in the require.cache ( once we require the file the first time ) are getting somehow lost, or module.exports of that specific module gets lost. ### How often does it reproduce? Is there a required condition? Usually, we encounter this issue once every two weeks or similar. ( But from what our users report, its not always the same case ) ### What is the expected behavior? Why is that the expected behavior? We expect the require(filePath) to return the actual functions and other things exposed via module.exports. ### What do you see instead? Instead, we basically see the following: <img width="1333" alt="image" src="https://github.com/user-attachments/assets/1d2d3080-4160-4ce7-b460-b46f7f59c5b3"> we also tried to see somehow the require(file)'s keys <img width="650" alt="image" src="https://github.com/user-attachments/assets/091745a1-20d7-4e75-b676-769458217541"> ---- The strange thing is, the file is on the disk, and actually has content in it ( even if in the following picture we dont actually see any module.exports, trust me, those are there ) <img width="1685" alt="image" src="https://github.com/user-attachments/assets/06c7c4ac-07d4-4973-ae11-48e281965161"> While debugging, even if the file itself is on the disk, the debugger does not show it: <img width="900" alt="image" src="https://github.com/user-attachments/assets/d78ed093-0259-4d89-8e57-4e563ceccf28"> ---- as for the require.cache here it is: ![image](https://github.com/user-attachments/assets/d8bea7aa-b4a7-44a1-bb77-576465c51f63) ### Additional information I have to mention that usually, using delete require.cache(that require) and downloading the file again solves the issue, or just restarting the server itself.
loaders
low
Critical
2,555,671,814
kubernetes
add tests for scheduler-perf itself
/sig scheduling /kind cleanup [scheduler-perf](https://github.com/kubernetes/kubernetes/tree/master/test/integration/scheduler_perf) has few tests for itself at the moment; we roughly check the behavior with `-tags performance, short` or `-tags integration-test, short`, which could miss some potential issues with it. scheduler-perf right now serves a crucial role for the scheduler to keep our targets in various metrics, and we're adding more and more features into it. Unless we have proper testing for it, the more features we add into it, the more likely we could break some existing parts without noticing it since the reviewing would get more and more difficult. So, this issue aims to add tests for it. I guess it also requires some refactoring, especially around [`runWorkload`](https://github.com/kubernetes/kubernetes/blob/1bbe775d5ffb131636193fe0bc15a8fcc0cd6fd6/test/integration/scheduler_perf/scheduler_perf.go#L1270), which is currently huge. /cc @alculquicondor @macsko @dom4ha
kind/cleanup,sig/scheduling,needs-triage
low
Major
2,555,686,091
PowerToys
FancyZones: Set colour/opacity/blur for certain zones/inactive zones
### Description of the new feature / enhancement Set either certain zones, or the option for unused zones, to have a colour/opacity/blur level. E.g., on a wide monitor, you have a large zone in the middle, and one on each side. Right now, the two either side are transparent when not used. There could be options to:- 1. Not allow the zones on the left and right to be used, and then set their blur/colour/opacity levels. 2. When unused, the zones will blur/go a colour/change their opacity levels, until they are used. This will allow you to create a "focus" window in the middle, and remove distractions on the left and right. ### Scenario when this would be used? This will allow you to have focused zones that are used, as the disallowed/unused zones blur/change colour/go more opaque. ### Supporting information _No response_
Needs-Triage
low
Minor
2,555,872,026
material-ui
[docs] Icons aren't searchable anymore in Algolia
See https://github.com/mui/material-ui/pull/41330#issuecomment-2323517472 Icons used to be searchable in [v4](https://v4.mui.com/components/material-icons/#material-icons) ![image](https://github.com/user-attachments/assets/ed08080e-b29d-4cd6-9e5f-0ac76ad34b29) But even after [reverting](https://github.com/mui/material-ui/pull/43569) the icons virtualization, they still aren't searchable. <img width="951" alt="Screenshot 2024-09-30 at 09 54 56" src="https://github.com/user-attachments/assets/76b0b206-423e-4ecc-a757-0eb88845bc7c"> **Search keywords**:
bug 🐛,docs,package: icons,regression,scope: docs-infra
low
Minor
2,555,889,158
node
`require.resolve` cannot find module after calling `require.resolve` and installing dependencies
### Version v20.17.0 ### Platform ```text Darwin P03694NWF5 23.1.0 Darwin Kernel Version 23.1.0: Mon Oct 9 21:32:11 PDT 2023; root:xnu-10002.41.9~7/RELEASE_ARM64_T6030 arm64 ``` ### Subsystem require ### What steps will reproduce the bug? 1. Init an emty project ``` npm init ``` 2. Add a package. e.g. lodash ``` npm install lodash ``` 3. Execute this script, the `pkg` can be any dependencies that are not installed in the global environment and current project. e.g. rollup ```javascript const { exec } = require('child_process'); const pkg = 'rollup'; try { console.log(require.resolve(pkg)); } catch (e) { console.log(e); } const child = exec( `npm install ${pkg}`, () => { console.log(require.resolve(pkg)); } ); // sync log from child_process to main process child.stdout?.on('data', (data) => { process.stdout.write(data); }); child.stderr?.on('data', (data) => { process.stderr.write(data); }); ``` The output shows twice `Error: Cannot find module 'rollup'`, the first one is expected; But the second one seems to be a bug. ### How often does it reproduce? Is there a required condition? Emerge necessarily. The required condition is the first `require.resolve` call. Without it, the script works fine. ### What is the expected behavior? Why is that the expected behavior? Output the resolve path correctly after installing the package ### What do you see instead? ``` Error: Cannot find module 'rollup' Require stack: - /Users/Documents/test/index.js at Module._resolveFilename (node:internal/modules/cjs/loader:1225:15) at Function.resolve (node:internal/modules/helpers:190:19) at Object.<anonymous> (/Users/Documents/test/index.js:5:23) at Module._compile (node:internal/modules/cjs/loader:1469:14) at Module._extensions..js (node:internal/modules/cjs/loader:1548:10) at Module.load (node:internal/modules/cjs/loader:1288:32) at Module._load (node:internal/modules/cjs/loader:1104:12) at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:174:12) at node:internal/main/run_main_module:28:49 { code: 'MODULE_NOT_FOUND', requireStack: [ '/Users/Documents/test/index.js' ] } added 3 packages, and audited 5 packages in 1s found 0 vulnerabilities node:internal/modules/cjs/loader:1228 throw err; ^ Error: Cannot find module 'rollup' Require stack: - /Users/Documents/test/index.js at Module._resolveFilename (node:internal/modules/cjs/loader:1225:15) at Function.resolve (node:internal/modules/helpers:190:19) at /Users/Documents/test/index.js:13:25 at ChildProcess.exithandler (node:child_process:414:7) at ChildProcess.emit (node:events:519:28) at maybeClose (node:internal/child_process:1105:16) at ChildProcess._handle.onexit (node:internal/child_process:305:5) { code: 'MODULE_NOT_FOUND', requireStack: [ '/Users/Documents/test/index.js' ] } ``` ### Additional information _No response_
loaders,v20.x
low
Critical
2,555,921,544
rust
ICE: `expected int of size 8, but got size 1` in `rustc_middle/src/ty/consts/int.rs` with feature `adt_const_params`
<!-- Thank you for finding an Internal Compiler Error! 🧊 If possible, try to provide a minimal verifiable example. You can read "Rust Bug Minimization Patterns" for how to create smaller examples. http://blog.pnkfx.org/blog/2019/11/18/rust-bug-minimization-patterns/ --> ### Code (manually reduced) ```Rust #![feature(adt_const_params)] #![allow(incomplete_features)] struct ConstBytes<const T: &'static [*mut u8; 3]>; pub fn main() { let _: ConstBytes<b"AAA"> = ConstBytes::<b"BBB">; } ``` <details><summary><strong>Original Code</strong></summary> <p> ```Rust //@ revisions: full min #![cfg_attr(full, feature(adt_const_params))] #![cfg_attr(full, allow(incomplete_features))] struct ConstString<const T: &'static str>; //[min]~^ ERROR struct ConstBytes<const T: &'static [*mut u8; mem::size_of::<*mut T>() / mem::size_of::<*mut u8>()]>; //[min]~^ ERROR pub fn main() { let _: ConstString<"Hello"> = ConstString::<"Hello">; let _: ConstString<"Hello"> = None::<std::slice::Iter<()>>; //~ ERROR mismatched types let _: ConstString<"ℇ㇈↦"> = ConstString::<"ℇ㇈↦">; let _: ConstString<"ℇ㇈↦"> = ConstString::<"ℇ㇈↥">; //~ ERROR mismatched types let _: ConstBytes<b"AAA"> = ConstBytes::<{&[0x41, 0x41, 0x41]}>; let _: ConstBytes<b"AAA"> = ConstBytes::<b"BBB">; //~ ERROR mismatched types } // from tests/ui/const-generics/slice-const-param-mismatch.rs ``` </p> </details> ### Meta <!-- If you're using the stable version of the compiler, you should also check if the bug also exists in the beta or nightly versions. --> `rustc --version --verbose`: ``` rustc 1.83.0-nightly (7608018cb 2024-09-29) binary: rustc commit-hash: 7608018cbdac9e55d0d13529cf43adc33d53efcf commit-date: 2024-09-29 host: x86_64-apple-darwin release: 1.83.0-nightly LLVM version: 19.1.0 ``` ### Error output Command: `rustc` ``` error[E0741]: `&'static [*mut u8; 3]` can't be used as a const parameter type --> r_E364D18BCB8846EE07601029D3F8CBE898970C9D0A44FC6609A112B969C95B8D_2.rs:4:28 | 4 | struct ConstBytes<const T: &'static [*mut u8; 3]>; | ^^^^^^^^^^^^^^^^^^^^^ error: internal compiler error: compiler/rustc_middle/src/ty/consts/int.rs:263:13: expected int of size 8, but got size 1 ``` <!-- Include a backtrace in the code block by setting `RUST_BACKTRACE=1` in your environment. E.g. `RUST_BACKTRACE=1 cargo build`. --> <details><summary><strong>Backtrace</strong></summary> <p> ``` thread 'rustc' panicked at compiler/rustc_middle/src/ty/consts/int.rs:263:13: Box<dyn Any> stack backtrace: 0: std::panicking::begin_panic::<rustc_errors::ExplicitBug> 1: <rustc_errors::diagnostic::BugAbort as rustc_errors::diagnostic::EmissionGuarantee>::emit_producing_guarantee 2: rustc_middle::util::bug::opt_span_bug_fmt::<rustc_span::span_encoding::Span>::{closure#0} 3: rustc_middle::ty::context::tls::with_opt::<rustc_middle::util::bug::opt_span_bug_fmt<rustc_span::span_encoding::Span>::{closure#0}, !>::{closure#0} 4: rustc_middle::ty::context::tls::with_context_opt::<rustc_middle::ty::context::tls::with_opt<rustc_middle::util::bug::opt_span_bug_fmt<rustc_span::span_encoding::Span>::{closure#0}, !>::{closure#0}, !> 5: rustc_middle::util::bug::bug_fmt 6: <rustc_middle::ty::print::pretty::FmtPrinter as rustc_middle::ty::print::pretty::PrettyPrinter>::pretty_print_const_scalar_int 7: <rustc_middle::ty::print::pretty::FmtPrinter as rustc_middle::ty::print::pretty::PrettyPrinter>::pretty_print_const_valtree 8: <rustc_middle::ty::print::pretty::FmtPrinter as rustc_middle::ty::print::pretty::PrettyPrinter>::pretty_print_const 9: <rustc_middle::ty::print::pretty::FmtPrinter as rustc_middle::ty::print::pretty::PrettyPrinter>::pretty_print_const_valtree 10: <rustc_middle::ty::print::pretty::FmtPrinter as rustc_middle::ty::print::pretty::PrettyPrinter>::pretty_print_const_valtree 11: <rustc_middle::ty::print::pretty::FmtPrinter as rustc_middle::ty::print::pretty::PrettyPrinter>::pretty_print_const 12: <rustc_middle::ty::consts::Const as core::fmt::Display>::fmt 13: <rustc_trait_selection::error_reporting::TypeErrCtxt>::cmp 14: <rustc_trait_selection::error_reporting::TypeErrCtxt>::values_str 15: <rustc_trait_selection::error_reporting::TypeErrCtxt>::note_type_err 16: <rustc_trait_selection::error_reporting::TypeErrCtxt>::report_and_explain_type_error 17: <rustc_hir_typeck::fn_ctxt::FnCtxt>::demand_coerce_diag 18: <rustc_hir_typeck::fn_ctxt::FnCtxt>::check_decl 19: <rustc_hir_typeck::fn_ctxt::FnCtxt>::check_block_with_expected 20: <rustc_hir_typeck::fn_ctxt::FnCtxt>::check_expr_with_expectation_and_args 21: <rustc_hir_typeck::fn_ctxt::FnCtxt>::check_return_expr 22: rustc_hir_typeck::check::check_fn 23: rustc_hir_typeck::typeck [... omitted 1 frame ...] 24: <rustc_middle::hir::map::Map>::par_body_owners::<rustc_hir_analysis::check_crate::{closure#4}>::{closure#0} 25: rustc_hir_analysis::check_crate 26: rustc_interface::passes::run_required_analyses 27: rustc_interface::passes::analysis [... omitted 1 frame ...] 28: <rustc_interface::queries::QueryResult<&rustc_middle::ty::context::GlobalCtxt>>::enter::<core::result::Result<(), rustc_span::ErrorGuaranteed>, rustc_driver_impl::run_compiler::{closure#0}::{closure#1}::{closure#5}> 29: rustc_interface::interface::run_compiler::<core::result::Result<(), rustc_span::ErrorGuaranteed>, rustc_driver_impl::run_compiler::{closure#0}>::{closure#1} note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace. note: we would appreciate a bug report: https://github.com/rust-lang/rust/issues/new?labels=C-bug%2C+I-ICE%2C+T-compiler&template=ice.md note: please make sure that you have updated to the latest nightly note: please attach the file at `/Volumes/T7/workspace/blackbox icemaker 결과/240926_240401_blackbox_result/rustc-ice-2024-09-30T08_07_24-63216.txt` to your bug report query stack during panic: #0 [typeck] type-checking `main` #1 [analysis] running analysis passes on this crate end of query stack error: aborting due to 2 previous errors For more information about this error, try `rustc --explain E0741`. ``` </p> </details> ### Note - ICE location: `compiler/rustc_middle/src/ty/consts/int.rs Line-263` https://github.com/rust-lang/rust/blob/7608018cbdac9e55d0d13529cf43adc33d53efcf/compiler/rustc_middle/src/ty/consts/int.rs#L261-L265 - #126359 has identical ICE location. - but this report requires incomplete feature & regression in `nightly-2024-02-15` @rustbot label +F-adt_const_params
I-ICE,T-compiler,C-bug,F-adt_const_params,S-bug-has-test
low
Critical
2,555,946,509
flutter
Error where possible null is being asserted in rendering paragraph
### Steps to reproduce Spam asynchronous redraws of a Text element on macOS by resizing the window ### Expected results No Error ### Actual results ``` RenderParagraph.text (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/rendering/paragraph.dart:348) RenderParagraph.debugDescribeChildren (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/rendering/paragraph.dart:1278) DiagnosticableTreeNode.getChildren (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:2882) DiagnosticsNode.toJsonMap.<anonymous closure> (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:1611) DiagnosticsNode.toJsonMap (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:1652) DiagnosticsNode.toJsonList.<anonymous closure> (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:1677) MappedListIterable.elementAt (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/internal/iterable.dart:425) ListIterator.moveNext (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/internal/iterable.dart:354) new _GrowableList._ofEfficientLengthIterable (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/_internal/vm/lib/growable_array.dart:189) new _GrowableList.of (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/_internal/vm/lib/growable_array.dart:150) new List.of (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/_internal/vm/lib/array_patch.dart:39) ListIterable.toList (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/internal/iterable.dart:224) DiagnosticsNode.toJsonList (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:1678) DiagnosticsNode.toJsonMap.<anonymous closure> (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:1645) DiagnosticsNode.toJsonMap (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:1652) DiagnosticsNode.toJsonList.<anonymous closure> (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:1677) MappedListIterable.elementAt (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/internal/iterable.dart:425) ListIterator.moveNext (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/internal/iterable.dart:354) new _GrowableList._ofEfficientLengthIterable (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/_internal/vm/lib/growable_array.dart:189) new _GrowableList.of (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/_internal/vm/lib/growable_array.dart:150) new List.of (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/_internal/vm/lib/array_patch.dart:39) ListIterable.toList (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/bin/cache/pkg/sky_engine/lib/internal/iterable.dart:224) DiagnosticsNode.toJsonList (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:1678) DiagnosticsNode.toJsonMap.<anonymous closure> (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:1639) DiagnosticsNode.toJsonMap (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/foundation/diagnostics.dart:1652) WidgetInspectorService._nodeToJson (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/widgets/widget_inspector.dart:1735) WidgetInspectorService._reportStructuredError (/opt/homebrew/Caskroom/flutter/3.24.3/flutter/packages/flutter/lib/src/widgets/widget_inspector.dart:997) <asynchronous gap> (Unknown Source:0) ``` ### Code sample This is happening on a "Text" widget that is just `Text("string")` ### Screenshots or Video N/A ### Logs N/A ### Flutter Doctor output ``` Doctor summary (to see all details, run flutter doctor -v): [✓] Flutter (Channel stable, 3.24.3, on macOS 15.0 24A335 darwin-arm64, locale en-NZ) [!] Android toolchain - develop for Android devices (Android SDK version 33.0.1) ✗ cmdline-tools component is missing Run `path/to/sdkmanager --install "cmdline-tools;latest"` See https://developer.android.com/studio/command-line for more details. ✗ Android license status unknown. Run `flutter doctor --android-licenses` to accept the SDK licenses. See https://flutter.dev/to/macos-android-setup for more details. [✓] Xcode - develop for iOS and macOS (Xcode 16.0) [✓] Chrome - develop for the web [✓] Android Studio (version 2022.3) [✓] VS Code (version 1.93.1) [✓] Connected device (5 available) [✓] Network resources ! Doctor found issues in 1 category. ```
framework,a: error message,P2,team-text-input,triaged-text-input
low
Critical
2,555,964,875
transformers
Improve image processing time
### Feature request Optimize Transformers' image_processors to decrease image processing time, and reduce inference latency for vision models and vlms. ### Motivation The Transformers library relies on PIL (Pillow) for image preprocessing, which can become a major bottleneck during inference, especially with compiled models where the preprocessing time can dominate the overall inference time. ![image](https://github.com/user-attachments/assets/10e19b7a-5e48-4bef-84a1-430458e17586) ![image-1](https://github.com/user-attachments/assets/be4f06f3-befd-44b7-b743-10f09471fa58) In the examples above, the RT-DETR preprocessing necessitates only to resize the image, while the DETR one involves resize+normalize. In eager mode, image preprocessing takes a big part of the total inference time for RT-DETR, but is not the main bottleneck. However, with a compiled RT-DETR, image preprocessing takes up the majority of the inference time, underlining the necessity to optimize it. This is even clearer for DETR, where image preprocessing is already the main bottleneck in eager mode. However, alternative libraries exist that leverage available hardware more efficiently for faster image preprocessing. [OptimVision](https://github.com/yonigozlan/OptimVision) uses such libraries to get much better results compared to Transformers. Much more details on OptimVision and image processing methods comparison are available on this [Notion page](https://www.notion.so/huggingface2/OptimVision-Optimize-preprocessing-time-10f1384ebcac8091a12debb87fe5f591?pvs=4). ### Your contribution OptimVision is an experiment playground to optimize the different steps involved in inferring/training with vision models. The current fast image preprocessing in OptimVision is a proof of concept and is not yet ready to be merged into Transformers, but that this the ultimate goal :).
Feature request,Vision,optimization,Processing
low
Major
2,555,997,664
PowerToys
Seamless Night Light
### Description of the new feature / enhancement Instead of one setting for the night light which often seems too strong/soft at sunset/sunrise time it would be awesome to transition seamlessly and according to the actual daytime. ### Scenario when this would be used? Every-Day Use and especially for people who tend to work very early or late. ### Supporting information _No response_
Needs-Triage
low
Minor
2,556,016,964
ui
[bug]: useFormField can be used without being nested within FormField
### Describe the bug useFormField has a check that should prevent it from being used when it is not called within a FormField component. This error is not thrown because empty objects (the default of the context when it is not wrapped in a provider) are true when checking in the if condition. ### Affected component/components Form ### How to reproduce 1. Render any component that uses useFormField within a Form without being inside of a FormField component. 2. Don't see the error. ### Codesandbox/StackBlitz link https://stackblitz.com/edit/vitejs-vite-za3zz1?file=src%2FApp.tsx ### Logs _No response_ ### System Info ```bash MacOS, Chrome ``` ### Before submitting - [X] I've made research efforts and searched the documentation - [X] I've searched for existing issues
bug
low
Critical
2,556,030,128
vscode
Conflict resolver worked not expected when I click accpect incoming or current
Edit from @hediet: When accepting conflict 1 and 2 fast enough, the edit for conflict 2 does not take into account the (still ongoing) edit for conflict 1. --- <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.93.1 (system setup) - OS Version: Windows_NT x64 10.0.22631 Steps to Reproduce: 1. Some time I tried to resolve conflict in vscode. It worked ok when there was only on conflict, if two conflict came up, the first one could been resolved smooth by clicking accept incoming or current button. But the second will failed, after clicking one accept button, it didn't work. 2. Then I pressed ctrl+z to undo, and clicked the button again, it worked ok . ![ezgif-1-e8312ec3ca](https://github.com/user-attachments/assets/7b1ff00a-2609-42df-8680-0dd46f123d5a)
bug,merge-conflict
low
Critical
2,556,046,195
PowerToys
new feature
### Description of the new feature / enhancement You hold down some button While you hold the button keyboard layout switches to the selected one You release the button The keyboard layout switches to the original one ### Scenario when this would be used? write very quickly on the Russian keyboard # for example RightAlt+Shift+3 give us # on the Russian layout or the № on the English layout switching to Russian ### Supporting information _No response_
Idea-Enhancement,Product-Keyboard Shortcut Manager,Needs-Triage
low
Minor
2,556,094,545
vscode
Screen reader box incorrectly positioned when selecting multiple lines
The screen reader box incorrectly positioned when selecting multiple lines when using MacOS voice over This is because the edit context hidden element is incorrectly positioned on multi-line selections https://github.com/user-attachments/assets/3992fe5b-71ca-4a59-919e-10e27dc9f1ae
bug,editor-input
low
Minor
2,556,107,386
ollama
Warn users when input is longer than supported context
### What is the issue? Hi thanks for the great library! It would be great if there could be a warning when input is longer than supported context. Otherwise, if the text is silently truncated, the behavior can be confusing. ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
bug
low
Minor
2,556,196,884
langchain
ModuleNotFoundError: No module named 'rank_llm.rerank.zephyr_reranker' (update of rank_llm.rerank -> rank_llm.rerank.listwise.zephyr_reranker)
### Checked other resources - [X] I added a very descriptive title to this issue. - [X] I searched the LangChain documentation with the integrated search. - [X] I used the GitHub search to find a similar question and didn't find it. - [X] I am sure that this is a bug in LangChain rather than my code. - [X] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package). ### Example Code `from rank_llm.rerank.listwise.zephyr_reranker import ZephyrReranker` 🧨🧨 `from rank_llm.rerank.zephyr_reranker import ZephyrReranker` ### Error Message and Stack Trace (if applicable) ModuleNotFoundError: No module named 'rank_llm.rerank.zephyr_reranker' ### Description I think the `rank_llm` updated version changed the package locations: worked: `from rank_llm.rerank.listwise.zephyr_reranker import ZephyrReranker` not worked: `from rank_llm.rerank.zephyr_reranker import ZephyrReranker` ### System Info up2date langchain
🤖:bug,stale
low
Critical
2,556,198,362
PowerToys
Add a Highlight Border When Switching Windows
### Description of the new feature / enhancement When working on a large monitor with 3+ windows visible at a time, it can often be confusing to know which window is in focus, or which window you've instantly alt + tabbed to. My suggestion is to add an feature that adds a 1px blue border (or whatever) that fades out over 2 seconds or so, to whichever window becomes activated. This way, you will instantly know which one you've selected. Additionally - although a bit different/more technical, I think it could also be really useful to be able to proximity-switch to a window based on FancyZones zones. So basically you can "move" the active zone, which would function just like moving the active window around using hotkeys, except instead of moving the window, you're moving the "active zone", and whichever window was last accessed in the zone you land on, becomes active. ### Scenario when this would be used? This would be especially useful for large monitors where you quickly want to switch between windows. While the alt+tab windows overview is of course very useful, this requires you to "study" the windows for a second in order to choose the right one - meaning, if you simply quick swap with alt + tab, it's usually almost impossible to know which window you've now switched to, if it's one of the windows that are already visible and for instance, in another FancyZones zone. Additionally, especially with browsers, you'll often have multiple browsers open with very similar content on them, making it hard to distinguish which one you want. For instance, let's say you have your editor in one zone, and two github repos open in their own windows in other zones. Visually, on your monitor, you know which window you want to tab to, but with alt + tab, the window previews will look very similar, and I often find myself selecting the wrong window, and having to hit some other key to see which one I'm in (eg. up/down arrow key to see which one scrolls). I think the border glow would help with this, but switching active zones would be even better. ### Supporting information _No response_
Needs-Triage
low
Minor
2,556,216,269
godot
NodePath property targeting breaks when the targets script is overridden in an inherited scene.
### Tested versions Tested in 4.4-dev2_mono_win64 and 4.3-stable_mono_win64 ### System information Godot v4.4.dev2.mono - Windows 10.0.22631 - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 4080 (NVIDIA; 31.0.15.4633) - AMD Ryzen 7 5800X 8-Core Processor (16 Threads) ### Issue description When the target of an exported NodePath property has its script overridden in an inherited scene, the NodePath breaks and throws a ObjectDisposedException. ### Steps to reproduce To reproduce this bug, one needs a BaseScene, an InheritedScene (created via "New Inherited Scene" of the BaseScene), a BaseScript, a OtherScript and finally one ScriptWithExport with one export of the type Node. The BaseScene looks as follows: BaseScene (BaseScript) +ScriptWithExport (Exported variable Target targets BaseScene) The InheritedScene looks like this: InheritedScene (OtherScript) +ScriptWithExport (Exported variable Target targets InheritedScene) Inside the _Ready function of the ScriptWithExport function add one line which prints the Target. While the BaseScene works as expected, the InheritedScene will throw an ObjectDisposedException. Replacing OtherScript with BaseScript in the Inherited Scene, will result in a working condition again. The targeted node does not have to be the root node, the same error can be reproduced when any node has its script overridden and is the target of an exported NodePath property. ### Minimal reproduction project (MRP) [minimal-repro-inheritance.zip](https://github.com/user-attachments/files/17188270/minimal-repro-inheritance.zip)
bug,topic:editor
low
Critical
2,556,242,630
tensorflow
Multithreading is not working with teansorflow
### Issue type Bug ### Have you reproduced the bug with TensorFlow Nightly? Yes ### Source source ### TensorFlow version tensorflow==2.15.0.post1 ### Custom code Yes ### OS platform and distribution _No response_ ### Mobile device _No response_ ### Python version python:3.10.12 ### Bazel version _No response_ ### GCC/compiler version _No response_ ### CUDA/cuDNN version _No response_ ### GPU model and memory _No response_ ### Current behavior? I am using bert model for classification and serving the model with gunicorn worker_class=gthreads, tf.config.threading.set_intra_op_parallelism_threads(1) tf.config.threading.set_inter_op_parallelism_threads(1) when I using above two line of code the code is working fine as expected and if increase the number to more than 1, the code getting blocked at the below line of code # Make predictions outputs = model_obj(inputs) and also inorder to reduce the docker image size I am using RUN pip3 install torch==2.0.0+cpu -f https://download.pytorch.org/whl/torch_stable.html before installing all the dependencies Flask==2.2.5 g2p-en==2.1.0 gunicorn==21.2.0 jellyfish==1.0.3 kenlm==0.2.0 nltk==3.8.1 numpy==1.26.3 pandas==2.2.0 python-dotenv==1.0.1 requests==2.31.0 scikit-learn==1.4.0 semantic-router==0.0.17 semantic-router[fastembed] sentence-transformers==2.3.1 tensorflow==2.15.0.post1 tensorflow-hub==0.16.0 theano==1.0.5 transformers==4.37.2 Werkzeug==2.2.2 please tell me why is my code is getting blocked if I use more than 1 thread. ### Standalone code to reproduce the issue ```shell def intent_prediction(self, sentence, thresold_score): logger.info(f"Threshold Score: {thresold_score}") try: model_obj = intent_object_dict[self.model] except KeyError: logger.error("Model not found in intent_object_dict") model_obj = self.load_model() if INTENT_MODEL == "cohere": score, intent = self.cohere_intent_prediction(sentence, model_obj) else: score, intent = self.Bert_intent_prediction( sentence, model_obj, thresold_score ) return score, intent def Bert_intent_prediction(self, sentence, model_obj, thresold_score): inputs = self.load_BERT_tokenizer(sentence) # Make predictions outputs = model_obj(inputs) # Get predicted class probabilities = tf.nn.softmax(outputs.logits, axis=1) predicted_class = tf.argmax(probabilities, axis=1).numpy()[0] matching_score = probabilities[0][predicted_class].numpy() try: intent_data = intent_label_dict[self.model] except Exception as e: logger.error(f"error while getting label: {e}") intent_data = self.get_intent_labels() if matching_score >= thresold_score: logger.info( f"Matched Main Intent:\ {intent_data[predicted_class]},\ SCORE :{matching_score}" ) logger.info(f"Matched Sentence: {intent_data[predicted_class]}") else: logger.info(f"Intent not matched, score is {matching_score}") intent = intent_data[predicted_class] return str(matching_score), intent ``` ### Relevant log output ```shell 30-Sep-2024 15:50:42.761|INFO |__init__|I want my sofa get cleaned| __init__.py:171|Enter into PUNC for intent... 30-Sep-2024 15:50:42.761|INFO |phrase_sim|I want my sofa get cleaned| phrase_sim.py:72|Threshold Score: 0.7 after this the code is blocked ```
stat:awaiting tensorflower,type:bug,TF 2.15
low
Critical
2,556,259,842
PowerToys
Different Results in OpenAI Playground vs requests sent to API from Advanced Paste
### Microsoft PowerToys version 0.84.1 ### Installation method Microsoft Store ### Running as admin No ### Area(s) with issue? Advanced Paste ### Steps to reproduce Adding a custom action with very specific prompt gives a wrong result, where trying the same prompt in Playground gives the correct result. Using gpt-3.5-turbo in Playground so it is the same as Advanced paste. Example prompt: "0099-0ac8-0084" Strip dividers ":", "-", ".". Strip "0099". Divide to 4 octets with dots. Convert octets to dec. ### ✔️ Expected Behavior 10.200.0.132 ### ❌ Actual Behavior 172.200.0.132 ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Minor
2,556,347,611
react
Inability to prioritise hydration & react yields too willingly
# Summary React does not provide an API to set the priority on a component or suspense boundary for hydration. This means you cannot optimise your application to hydrate part of the app you know users will want to interact with first. Or in my case a part of the application I need to start rendering client side asap to replace a SSR skeleton with the correct personalised content that can only be rendered client side. The issue is compounded by the way React 18 yields the main thread during hydration to other higher priority events, which while a great idea in theory to improve FID and INP, in practice means 3rd party loaded (gtm) scripts delay the initial hydrating. It would be great if this logic could differentiate between user interactions and scripts added to the call stack? Detailed here https://github.com/reactwg/react-18/discussions/38#discussioncomment-837161 My real-world use case is an ecommerce application where marketing teams are loading via gtm vast amounts of 3rd party scripts, think tiktok, instagram, bing, and masses of gtm containers. While the ideal solution would be to either trim these down or move them to the worker threads via something like partytown (https://partytown.builder.io/), neither is realistic. As such we need a way to optimise around them. We can see, on slower windows machines and older mobiles, initial hydration of components take more than 5-8 seconds to start. This provides a negative experience to customers where skeletons are visible, and a worse UX than the legacy MVC sites. ##Work arounds As detailed here there are potential workarounds that already exist https://github.com/reactwg/react-18/discussions/130 Firstly the `unstable_scheduleHydration` API, however this has since been moved to the hydrateRoot function and is no longer accessible in Next.js (I believe?) (https://github.com/facebook/react/pull/22455/files) The second, and this is where you have to forgive me, is to creatively interpret the following ` Discrete events (eg. click/keypresses) trigger synchronous (selective) hydration in the capture phase if the code in its encapsulating Suspense boundary is ready. If the event can't be synchronously hydrated then we'll increase the priority of that boundary so it hydrates first when it's ready.` Which leads to this being an incredible performance optimisation ``` useEffect(() => { document.getElementById("body")?.click(); }, []); ``` Because React batches hydration together at Suspense layers, it does an initial render down to the Suspense boundary, then processes the useLayoutEffects, then useEffects, then finally starts the child boundary. This logic when applied in the parent and clicking an element in the first child boundary, forces React into a synchronous render, no longer yielding, and prioritising above all other suspense boundaries the one you clicked. ## Evidence Reproducible example here https://github.com/edqwerty1/hydrate Visit http://localhost:3000/slow for no click logic, and http://localhost:3000/fast for the improved version. In this scenario I really need Content Right to hydrate first as I must replace a SSR skeleton with my personalised content ### Before <img width="345" alt="image" src="https://github.com/user-attachments/assets/1410a856-e8c8-4bbc-974c-0d573666e0c2"> It is delayed by around 5 seconds, these timings are realistic to real world data. The gtm scripts don’t normally block for an entire second, but there are normally a lot more of them. ### After <img width="342" alt="image" src="https://github.com/user-attachments/assets/446cfd41-2977-4fe9-bc7e-4c08ca0959b0"> Content Right hydrated in 2.5s, half the time. And more importantly the time between App starting and my component rendering has reduced from 2.3s to 200ms! A very large saving and if I reorder the initial script tags I could in theory delay the 3rd party code until my entire app is ready. Now this of course isn't idea, by hacking the hydration logic I am potentially causing higher INP scores for real events, and there may be other issues.
Type: Discussion
medium
Major
2,556,369,243
svelte
Own TS types for types of input elements
### Describe the problem Now there is only a single type [HTMLInputAttributes](https://github.com/sveltejs/svelte/blob/svelte%405.0.0-next.260/packages/svelte/elements.d.ts#L1036) for all input elements and there is e.g. the library `flowbite-svelte` which uses this type to [extend](https://github.com/sveltejs/svelte/blob/svelte%405.0.0-next.260/packages/svelte/elements.d.ts#L1036) the props of a component with it. The advantage is that the maintainers don't have to implement each prop themselves but it comes with the disadvantage that the component has more props than it should have. ### Describe the proposed solution It would be nice to have types like for example `HTMLInputRadioAttributes` which are specific for one type of input element. Then libraries could extend the props of their components with these types. ### Importance nice to have
types / typescript
low
Minor
2,556,395,386
langchain
ChatDatabricks() unable to stream special characters
### Checked other resources - [X] I added a very descriptive title to this issue. - [X] I searched the LangChain documentation with the integrated search. - [X] I used the GitHub search to find a similar question and didn't find it. - [X] I am sure that this is a bug in LangChain rather than my code. - [X] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package). ### Example Code ## Config rag_chain_config.yaml ```yml databricks_resources: llm_endpoint_name: gpt-4o-mini vector_search_endpoint_name: VS_ENDPOINT input_example: messages: - content: Hvilken vinterrug sort havde mest udbytte i 2023? role: user - content: Det største udbytte i årets landsforsøg med sorter af vinterrug i 2023 var 112,1 hkg pr. ha, og det blev høstet i sorten KWS Cursor. role: assistant - content: Hvilken vinterhvede sort havde mest udbytte i 2023? role: user llm_config: llm_parameters: max_tokens: 1500 temperature: 0.01 llm_system_prompt_template: "\n You are a trusted assistant that helps answer\ \ questions based only on the provided information. If you do not know the answer\ \ to a question, you truthfully say you do not know. Here is some context which\ \ might or might not help you answer: {context}. Answer directly, do not repeat\ \ the question, do not start with something like: the answer to the question,\ \ do not add AI in front of your answer, do not say: here is the answer, do not\ \ mention the context or the question. You must provide your answer in the Danish\ \ language or in the detected language if the queried question from the user differs\ \ from Danish.\n If the answer contains multiple steps or points, provide the\ \ answer in a bullet format. \n Based on this context, answer this question:\ \ {question}\n" retriever_config: chunk_template: 'Passage: {chunk_text} ' data_pipeline_tag: poc parameters: k: 5 query_type: hybrid schema: chunk_text: chunked_text document_uri: path primary_key: chunk_id vector_search_index: VS_INDEX ``` ## Code example ```python from operator import itemgetter import mlflow import os from databricks.vector_search.client import VectorSearchClient from langchain_community.chat_models import ChatDatabricks from langchain_community.vectorstores import DatabricksVectorSearch from langchain_core.runnables import RunnableLambda from langchain_core.output_parsers import StrOutputParser from langchain_core.runnables import RunnablePassthrough from langchain_core.prompts import ( PromptTemplate, ChatPromptTemplate, ) ## Enable MLflow Tracing mlflow.langchain.autolog() ############ # Helper functions ############ # Return the string contents of the most recent message from the user def extract_user_query_string(chat_messages_array): return chat_messages_array[-1]["content"] # Return the chat history, which is is everything before the last question def extract_chat_history(chat_messages_array): return chat_messages_array[:-1] # FIT AND FINISH: We should not require a value here. model_config = mlflow.models.ModelConfig(development_config='rag_chain_config.yaml') databricks_resources = model_config.get("databricks_resources") retriever_config = model_config.get("retriever_config") llm_config = model_config.get("llm_config") ############ # Connect to the Vector Search Index ############ vs_client = VectorSearchClient(disable_notice=True) vs_index = vs_client.get_index( endpoint_name=databricks_resources.get("vector_search_endpoint_name"), index_name=retriever_config.get("vector_search_index"), ) vector_search_schema = retriever_config.get("schema") ############ # Turn the Vector Search index into a LangChain retriever ############ vector_search_as_retriever = DatabricksVectorSearch( vs_index, text_column=vector_search_schema.get("chunk_text"), columns=[ vector_search_schema.get("primary_key"), vector_search_schema.get("chunk_text"), vector_search_schema.get("document_uri"), ], ).as_retriever(search_kwargs=retriever_config.get("parameters")) ############ # Required to: # 1. Enable the RAG Studio Review App to properly display retrieved chunks # 2. Enable evaluation suite to measure the retriever ############ mlflow.models.set_retriever_schema( primary_key=vector_search_schema.get("primary_key"), text_column=vector_search_schema.get("chunk_text"), doc_uri=vector_search_schema.get( "document_uri" ), # Review App uses `doc_uri` to display chunks from the same document in a single view ) ############ # Method to format the docs returned by the retriever into the prompt ############ def format_context(docs): chunk_template = retriever_config.get("chunk_template") chunk_contents = [ chunk_template.format( chunk_text=d.page_content, document_uri=d.metadata[vector_search_schema.get("document_uri")], ) for d in docs ] return "".join(chunk_contents) ############ # Prompt Template for generation ############ prompt = ChatPromptTemplate.from_messages( [ ( # System prompt contains the instructions "system", llm_config.get("llm_system_prompt_template"), ), # User's question ("user", "{question}"), ] ) ############ # FM for generation ############ model = ChatDatabricks( endpoint=databricks_resources.get("llm_endpoint_name"), extra_params=llm_config.get("llm_parameters") ) ############ # RAG Chain ############ chain = ( { "question": itemgetter("messages") | RunnableLambda(extract_user_query_string), "context": itemgetter("messages") | RunnableLambda(extract_user_query_string) | vector_search_as_retriever | RunnableLambda(format_context), } | prompt | model | StrOutputParser() ) ## Tell MLflow logging where to find your chain. # `mlflow.models.set_model(model=...)` function specifies the LangChain chain to use for evaluation and deployment. This is required to log this chain to MLflow with `mlflow.langchain.log_model(...)`. mlflow.models.set_model(model=chain) # Streaming example output chain_input = { "messages": [ { "role": "user", "content": "Hvordan bekæmper man væselhale?", } ] } for chunk in chain.stream(chain_input): print(chunk, end="|", flush=True) ``` ### Error Message and Stack Trace (if applicable) ## Code example streams the following strange formatting of special Danish characters "æ, ø, å": ``` |-| Bek|æ|mp|else| skal| ske| sÃ¥| tid|ligt| som| muligt| efter| frem|spiring|. |-| An|vend|else| af| M|aten|o| Duo| |600| SC| er| effektiv|,| især| ved| tid|lig| behandling|. |-| Be|handling| med| |0|,|7| l| M|aten|o| Duo| pr|.| ha| den| |20|.| august| har| g|ivet| de| bedste| result|ater|. |-| T|ils|æt|ning| af| Boxer| kan| forbed|re| bek|æ|mp|elsen|,| men| kan| være| problem|at|isk| pÃ¥| grund| af| ford|amp|ning|. |-| Det| er| vigtigt| at| reduc|ere| væ|sel|hale| til| et| niveau|,| hvor| de| ikke| pÃ¥vir|ker| op|form|ering| og| sp|red|ning|.|| ``` However, when manually correcting the output: ```python for chunk in chain.stream(chain_input): # Decode using latin1 (or the incorrect encoding) and re-encode in UTF-8 corrected_chunk = chunk.encode('latin1').decode('utf-8') print(corrected_chunk, end="|", flush=True) ``` ``` |-| Bek|æ|mp|else| skal| ske| så| tid|ligt| som| muligt| efter| frem|spiring|. |-| An|vend|else| af| M|aten|o| Duo| |600| SC| er| effektiv|,| især| ved| tid|lig| behandling|. |-| Be|handling| med| |0|,|7| l| M|aten|o| Duo| pr|.| ha| den| |20|.| august| har| vist| sig| at| være| den| bedste| metode|. |-| T|ils|æt|ning| af| Boxer| kan| forbed|re| bek|æ|mp|elsen|,| men| kan| være| problem|at|isk| at| anv|ende| tid|ligt| på| grund| af| ford|amp|ning|. |-| Det| er| vigtigt| at| reduc|ere| væ|sel|hale| til| et| niveau|,| hvor| de| ikke| påvir|ker| op|form|ering| og| sp|red|ning|.|| ``` ## Error clearly happens in ChatDatabricks() ![image](https://github.com/user-attachments/assets/b70857e0-ecc5-4f15-9b69-3641acb5e3ef) ## Workaround is currently to turn off streaming ```python model = ChatDatabricks( endpoint=databricks_resources.get("llm_endpoint_name"), extra_params=llm_config.get("llm_parameters"), disable_streaming=True ) ``` ![image](https://github.com/user-attachments/assets/72be3796-134f-41d9-9724-9f2a34a55bdb) ### Description from langchain_community.chat_models import ChatDatabricks unable to stream special Danish characters. I expect to see "æ", "ø" and "å". Instead I see strange representations formatted like: "æt" and "Ã¥" ### System Info anyio==3.5.0 argon2-cffi==21.3.0 argon2-cffi-bindings==21.2.0 asttokens==2.0.5 attrs==22.1.0 backcall==0.2.0 beautifulsoup4==4.11.1 black==22.6.0 bleach==4.1.0 blinker==1.4 boto3==1.24.28 botocore==1.27.96 certifi==2022.12.7 cffi==1.15.1 chardet==4.0.0 charset-normalizer==2.0.4 click==8.0.4 comm==0.1.2 contourpy==1.0.5 cryptography==39.0.1 cycler==0.11.0 Cython==0.29.32 databricks-sdk==0.1.6 dbus-python==1.2.18 debugpy==1.6.7 decorator==5.1.1 defusedxml==0.7.1 distlib==0.3.7 distro==1.7.0 distro-info==1.1+ubuntu0.2 docstring-to-markdown==0.11 entrypoints==0.4 executing==0.8.3 facets-overview==1.1.1 fastjsonschema==2.19.1 filelock==3.13.1 fonttools==4.25.0 googleapis-common-protos==1.62.0 grpcio==1.48.2 grpcio-status==1.48.1 httplib2==0.20.2 idna==3.4 importlib-metadata==4.6.4 ipykernel==6.25.0 ipython==8.14.0 ipython-genutils==0.2.0 ipywidgets==7.7.2 jedi==0.18.1 jeepney==0.7.1 Jinja2==3.1.2 jmespath==0.10.0 joblib==1.2.0 jsonschema==4.17.3 jupyter-client==7.3.4 jupyter-server==1.23.4 jupyter_core==5.2.0 jupyterlab-pygments==0.1.2 jupyterlab-widgets==1.0.0 keyring==23.5.0 kiwisolver==1.4.4 launchpadlib==1.10.16 lazr.restfulclient==0.14.4 lazr.uri==1.0.6 lxml==4.9.1 MarkupSafe==2.1.1 matplotlib==3.7.0 matplotlib-inline==0.1.6 mccabe==0.7.0 mistune==0.8.4 more-itertools==8.10.0 mypy-extensions==0.4.3 nbclassic==0.5.2 nbclient==0.5.13 nbconvert==6.5.4 nbformat==5.7.0 nest-asyncio==1.5.6 nodeenv==1.8.0 notebook==6.5.2 notebook_shim==0.2.2 numpy==1.23.5 oauthlib==3.2.0 packaging==23.2 pandas==1.5.3 pandocfilters==1.5.0 parso==0.8.3 pathspec==0.10.3 patsy==0.5.3 pexpect==4.8.0 pickleshare==0.7.5 Pillow==9.4.0 platformdirs==2.5.2 plotly==5.9.0 pluggy==1.0.0 prometheus-client==0.14.1 prompt-toolkit==3.0.36 protobuf==4.24.0 psutil==5.9.0 psycopg2==2.9.3 ptyprocess==0.7.0 pure-eval==0.2.2 pyarrow==8.0.0 pyarrow-hotfix==0.5 pycparser==2.21 pydantic==1.10.6 pyflakes==3.1.0 Pygments==2.11.2 PyGObject==3.42.1 PyJWT==2.3.0 pyodbc==4.0.32 pyparsing==3.0.9 pyright==1.1.294 pyrsistent==0.18.0 python-apt==2.4.0+ubuntu4 python-dateutil==2.8.2 python-lsp-jsonrpc==1.1.1 python-lsp-server==1.8.0 pytoolconfig==1.2.5 pytz==2022.7 pyzmq==23.2.0 requests==2.28.1 rope==1.7.0 s3transfer==0.6.2 scikit-learn==1.1.1 scipy==1.10.0 seaborn==0.12.2 SecretStorage==3.3.1 Send2Trash==1.8.0 six==1.16.0 sniffio==1.2.0 soupsieve==2.3.2.post1 ssh-import-id==5.11 stack-data==0.2.0 statsmodels==0.13.5 tenacity==8.1.0 terminado==0.17.1 threadpoolctl==2.2.0 tinycss2==1.2.1 tokenize-rt==4.2.1 tomli==2.0.1 tornado==6.1 traitlets==5.7.1 typing_extensions==4.4.0 ujson==5.4.0 unattended-upgrades==0.1 urllib3==1.26.14 virtualenv==20.16.7 wadllib==1.3.6 wcwidth==0.2.5 webencodings==0.5.1 websocket-client==0.58.0 whatthepatch==1.0.2 widgetsnbextension==3.6.1 yapf==0.33.0 zipp==1.0.0
🤖:bug,investigate
low
Critical
2,556,415,156
flutter
[Semantics] Add support for setting the heading level on iOS and macOS
### Use case This is a follow-up of closed #97894. iOS and macOS platforms support [AccessibilityHeadingLevel](https://developer.apple.com/documentation/swiftui/accessibilityheadinglevel) which is an equivalent of [ARIA: heading role](https://developer.mozilla.org/en-US/docs/Web/Accessibility/ARIA/Roles/heading_role) on web. iOS support was mentioned by both the OP and in the comment: https://github.com/flutter/flutter/issues/97894#issuecomment-2150050231 but the related PR #41435 added the heading level support for the web only. Different heading levels can speed up the navigation in case of VoiceOver users. They can quickly navigate between subsequent headings (either any or at the same level only) using the keyboard shortcuts: https://support.apple.com/guide/voiceover/search-commands-cpvokys08/mac ### Proposal Required changes are analogous to those made in https://github.com/flutter/engine/pull/41435 for web. Here is how to set the heading level: https://developer.apple.com/documentation/swiftui/view/accessibilityheading(_:) iOS/macOS also have 6 heading levels (from 1 to 6) like web. The value of `0` can probably be mapped to [unspecified](https://developer.apple.com/documentation/swiftui/accessibilityheadinglevel/unspecified).
c: new feature,platform-ios,a: accessibility,platform-mac,c: proposal,P2,team-accessibility,triaged-accessibility
low
Minor
2,556,437,909
pytorch
Add Mcdropout layer
### 🚀 The feature, motivation and pitch I am working with MCDropout, and find the current workflow in pytorch suboptimal and misleading. The current way to enable Dropout at evaluation time to perform MCDropout (https://arxiv.org/pdf/1506.02142) is to set all the Dropout Layers to training mode, which is in my opinion misleading, since you are not training at that point. My implementation would be to set a second flag `mcdropout` that enables to use the Layer with dropout always turned on, so that one does not need to loop over the layers, maintaining a clearer and cleaner code structure. This is my suggestion: ``` class MCDropout(_DropoutNd): """ Implementation of MCDropout to the torch dropout layer. This adds a mcdropout flag, which turns dropout always on in training and evaluation mode. Args: p : probability of an element to be zeroed. Default: 0.5 mcdropout: if True, will always perform dropout inplace : if True, will perform dropout in-place """ def __init__( self, p: float = 0.5, mcdropout: bool = True, inplace: bool = False ) -> None: super().__init__(p, inplace) self.mcdroput = mcdropout def forward(self, input: torch.Tensor) -> torch.Tensor: return torch.nn.functional.dropout( input, self.p, self.training or self.mcdropout, self.inplace ) ``` The change adds a parameter `mcdropout` to the Dropout layer, and returns as a condition to use dropout or not `self.training or self.mcdropout`, thus avoiding the need to go over all the layers and turn the training mode on. This also enables to set only some layers to MCDropout layers in an easier way. ### Alternatives Currently to use mcdropout one has to go over all the modules in the model and set them to training mode. ```def enable_dropout(model): """ Function to enable the dropout layers during test-time """ for m in model.modules(): if m.__class__.__name__.startswith('Dropout'): m.train() ``` ### Additional context _No response_ cc @albanD @mruberry @jbschlosser @walterddr @mikaylagawarecki
module: nn,triaged
low
Minor
2,556,487,569
TypeScript
A new generic type can be used for file path completion(intellisense)
### 🔍 Search Terms "path intellisense","path completion","asset path completion" ### ✅ Viability Checklist - [X] This wouldn't be a breaking change in existing TypeScript/JavaScript code - [X] This wouldn't change the runtime behavior of existing JavaScript code - [X] This could be implemented without emitting different JS based on the types of the expressions - [X] This isn't a runtime feature (e.g. library functionality, non-ECMAScript syntax with JavaScript output, new syntax sugar for JS, etc.) - [X] This isn't a request to add a new utility type: https://github.com/microsoft/TypeScript/wiki/No-New-Utility-Types - [X] This feature would agree with the rest of our Design Goals: https://github.com/Microsoft/TypeScript/wiki/TypeScript-Design-Goals ### ⭐ Suggestion Add a tool generics so that some function parameters can enable file path type hints some assets in `../assets` ```ts function getImageUrl(name:T<"../assets">) { ... return ... } getImageUrl("// intellisense ``` ### 📃 Motivating Example Enable custom file path completion through generics ### 💻 Use Cases 1. What do you want to use this for? When I used [vite static resource introduction](https://vitejs.dev/guide/assets.html#new-url-url-import-meta-url), I used a function to wrap it, but the function parameters could not achieve type hints like `import`. the function like: ```ts function getImageUrl(name:string) { // lost file intellisense return new URL(`./dir/${name}.png`, import.meta.url).href } ``` 2. What shortcomings exist with current approaches? Only `import` statements support path completion and cannot be customized 3. What workarounds are you using in the meantime? i make a vscode extension
Suggestion,Awaiting More Feedback
low
Minor
2,556,533,310
PowerToys
Remapped key in client does BOTH actions on the remote
### Microsoft PowerToys version 0.84.1 ### Installation method Microsoft Store ### Running as admin Yes ### Area(s) with issue? Keyboard Manager ### Steps to reproduce Just got new Win 11 laptop and found out about Power Toys and it's keyboard remapping feature. I've enabled remapping Caps Lock -> Backspace, Scroll Lock -> Caps Lock and connected to the remote running the windows 10. Tried using remapped key but realised that BOTH original and remap are sent to/executed on remote. Most remote settings on both ends are defaults, I can't remember fiddling with them and the relevant client side settings are as follows: <img width="291" alt="image" src="https://github.com/user-attachments/assets/25dae5de-f2a5-4850-b98c-aaa94e73375d"> Remote host has the same remapping but done in the registry. ### ✔️ Expected Behavior Only remapped key to be send to the remote, or at the very least the original one. ### ❌ Actual Behavior BOTH original and remapped keys are sent ### Other Software Remote desktop connection
Issue-Bug,Needs-Triage
low
Minor
2,556,630,050
vscode
[Accessibility with screen reader] NVDA switch automatically the language in the accessible view of the embedded terminal
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes/No <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.93.1 - OS Version: Windows 11 professional - NVDA Version : 2024.3 with the OneCore synthesizer Steps to Reproduce: 1. Open VSCode with NVDA active 2. CTRL+ù to open the embedded terminal 3. Run a command, like for example in my case "quarto render" then enter to run the command 4. ALT+F2 to open the accessible view of the embedded terminal and listen to the compilation errors 5. i.e. My version of VSCode is in French, via the extension "French Language Pack for Visual Studio Code", because I'm a French native speaker, but the output of the terminal is in English, so I need to switch the language used by my screen reader to English to be able to understand what it is saying, otherwise it will read the content in the accessible view with a French accent, making it unintelligible 6. So I then try to change the language to hear the output of the accessible view, because by default NVDA is reading it with a French voice. To do that, I've tried to change the language on the fly, i.e. without going through the configuration profile, but only with the keys on my keyboard. 7. The selected reading voice become English on every other part of VSCode, except on the accessible view. Hello, As a new VSCode and blind user, I'm learning to use the Markdown language, and more specifically Quarto Markdown. I'm running Windows 11 professional, NVDA 2024.3 with the OneCore synthesizer and the “automatic language detection” option, and VSCode 1.93.1. I use the ALT+F2 shortcut to access the accessible view of the built-in terminal. Being French, I use VSCode in French thanks to the extension [French Language Pack for Visual Studio Code](https://marketplace.visualstudio.com/items?itemName=MS-CEINTL.vscode-language-pack-fr). When I open the accessible view of the embedded terminal, the textual content displayed by VSCode is in English, which is normal. But my problem is: how can I listen to this content with my screen reader using the right language voice? I've tried to change the language on the fly, i.e. without going through the configuration profile, but only with the keys on my keyboard. The selected voice is English everywhere, except in the accessible view. It's as if a language tag set the language of this window to French by default. How can I change this? I hope I've been as clear as possible in my request, this is the first time I've posted an issue here. Thank you very much, and have a nice day. Michaël Beuve
bug,accessibility
low
Critical
2,556,695,014
vscode
Terminal seems to flicker more than ever
I feel that the terminal flickers more when I resize it, not sure it ever did flicker so much actually: ![Image](https://github.com/user-attachments/assets/0f934c0c-e1da-4cd8-bf2b-9f31ed541919) Did something change in the past weeks? My terminal related settings: ```json "terminal.integrated.fontFamily": "Menlo", "terminal.integrated.fontSize": 13, "terminal.integrated.lineHeight": 1.4, "terminal.integrated.windowsEnableConpty": false, "terminal.integrated.showExitAlert": false, "terminal.integrated.tabs.enabled": true, "terminal.integrated.shellIntegration.enabled": true, "terminal.integrated.stickyScroll.enabled": false, "terminal.integrated.commandsToSkipShell": [ "workbench.action.openPreviousEditorFromHistory", "workbench.action.startTerminalSpeechToText", "workbench.action.stopTerminalSpeechToText" ], "terminal.integrated.initialHint": false, ```
bug,terminal-rendering
low
Minor
2,556,713,513
electron
[Bug]: Unicode console output on Windows is still problematic
### Preflight Checklist - [X] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project. - [X] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [X] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a bug report that matches the one I want to file, without success. ### Electron Version 32.1.2 ### What operating system(s) are you using? Windows ### Operating System Version Windows 10 Pro 22H2 ### What arch are you using? x64 ### Last Known Working Electron version _No response_ ### Expected Behavior When logging Unicode strings to the console in an Electron application on Windows, the output should correctly display the Unicode characters. ### Actual Behavior When running an Electron application on Windows, `console.log` statements containing Unicode strings produce gibberish or unreadable characters in the console output. To reproduce this issue, you could add the following lines to the beginning of your main process script: ```javascript console.log('Привет'); console.log('你好'); console.log('こんにちは'); console.log('مرحبا'); console.log('Γειά σου'); ``` Instead of displaying the correct Unicode characters, the console output appears as follows: ![image](https://github.com/user-attachments/assets/c4c09678-4e26-48c9-bf80-5ac59c76ea10) ### Testcase Gist URL _No response_ ### Additional Information This issue is related to the previously closed issue #5761 (https://github.com/electron/electron/issues/5761). Despite the original issue being closed, the problem still persists. The only known workaround is to run `chcp 65001` every time before running an Electron app.
platform/windows,bug :beetle:,has-repro-gist,32-x-y
low
Critical
2,556,721,240
excalidraw
Add support of Ukrainian (or Cyrillic in general) language for Lilita One and Comic Shanns fonts
Please, add Ukrainian language for following fonts: - Lilita One - Comic Shanns <img width="866" alt="image" src="https://github.com/user-attachments/assets/b017e70d-3c35-4565-bfc4-228fc07bf3b7"> <img width="848" alt="image" src="https://github.com/user-attachments/assets/9cc61510-c72c-48b2-b1a3-ccab78869858">
font
low
Minor
2,556,753,754
pytorch
BCE loss mps device
### 🐛 Describe the bug I have a problem training a model with the BCE loss using mps device. The model uses a linear layer to predict a the sign of a real value number (two classes). When I train it in cpu everything is fine, but when i pass to mps the prediction degenerates and predicts always the same "class" for all the training points. ### Versions Collecting environment information... PyTorch version: 2.3.0.dev20240229 Is debug build: False CUDA used to build PyTorch: None ROCM used to build PyTorch: N/A OS: macOS 14.6.1 (arm64) GCC version: Could not collect Clang version: Could not collect CMake version: version 3.29.2 Libc version: N/A Python version: 3.12.4 | packaged by conda-forge | (main, Jun 17 2024, 10:13:44) [Clang 16.0.6 ] (64-bit runtime) Python platform: macOS-14.6.1-arm64-arm-64bit Is CUDA available: False CUDA runtime version: No CUDA CUDA_MODULE_LOADING set to: N/A GPU models and configuration: No CUDA Nvidia driver version: No CUDA cuDNN version: No CUDA HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Apple M3 Max Versions of relevant libraries: [pip3] numpy==1.26.4 [pip3] torch==2.3.0.dev20240229 [pip3] torchaudio==2.2.0.dev20240229 [pip3] torchsummary==1.5.1 [pip3] torchvision==0.18.0.dev20240229 [conda] numpy 1.26.4 pypi_0 pypi [conda] torch 2.3.0.dev20240229 pypi_0 pypi [conda] torchaudio 2.2.0.dev20240229 pypi_0 pypi [conda] torchsummary 1.5.1 pypi_0 pypi [conda] torchvision 0.18.0.dev20240229 pypi_0 pypi cc @kulinseth @albanD @malfet @DenisVieriu97 @jhavukainen
triaged,module: correctness (silent),module: mps
low
Critical
2,556,757,282
neovim
`gq` fails on empty file with `E20: Mark not set`
### Problem Formatting with `gq` and `formatprg` fails on empty file (general new buffer) with error `E20: Mark not set`. ### Steps to reproduce 1. nvim --clean foo 2. `:set formatprg=fmt` 3. type `gqG` 4. command fails with `E20: Mark not set` tested with these `formatprg` values: - `fmt` - `jq` - `prettier --parser json` > `formatexpr` seems to be fine. tested with `xmlformat#Format()`, `v:lua.vim.lsp.formatexpr()` and `v:lua.require'conform'.formatexpr()`. > `:%!fmt` is also fine. ### Expected behavior `gq` command on empty file should not fail or print more meaningful error message. ### Nvim version (nvim -v) v0.11.0-dev-862+gdf915f3af ### Vim (not Nvim) behaves the same? yes, vim 8.2 ### Operating system/version ubuntu 22.04 (aarch64) ### Terminal name/version blink shell ### $TERM environment variable xterm-256color ### Installation build from repo
bug-vim,marks,normal-mode
low
Critical
2,556,771,318
vscode
Inconsistent button order between editor and panel
* open an editor (no tabs) * open chat view in secondary sidebad * the overflow (...) and close (X) commands are swapped IMO the editor does this rignt <img width="1069" alt="Screenshot 2024-09-30 at 15 56 31" src="https://github.com/user-attachments/assets/40817e6b-e6c5-45d8-b146-d09e2c6c8b7d">
workbench-tabs,under-discussion,layout
low
Minor
2,556,854,646
go
proposal: cmd/go: `-json` flag for `go version -m`
``` $ go version -m -json ~/go/bin/gopls flag provided but not defined: -json usage: go version [-m] [-v] [file ...] Run 'go help version' for details. ``` VS Code Go (TS/JS) uses `go version -m` when inspecting the tools versions. If `go version -m` can output JSON-encoded `runtime/debug.BuildSetting` format, it will help simplifying output parsing a lot.
Proposal,NeedsInvestigation,FeatureRequest,GoCommand,Proposal-FinalCommentPeriod
low
Critical
2,556,863,985
yt-dlp
[ixigua] Failed to parse JSON (caused by JSONDecodeError)
### DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE - [X] I understand that I will be **blocked** if I *intentionally* remove or skip any mandatory\* field ### Checklist - [X] I'm reporting that yt-dlp is broken on a **supported** site - [X] I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels)) - [X] I've checked that all provided URLs are playable in a browser with the same IP and same login details - [X] I've checked that all URLs and arguments with special characters are [properly quoted or escaped](https://github.com/yt-dlp/yt-dlp/wiki/FAQ#video-url-contains-an-ampersand--and-im-getting-some-strange-output-1-2839-or-v-is-not-recognized-as-an-internal-or-external-command) - [X] I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates - [X] I've read the [guidelines for opening an issue](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#opening-an-issue) - [X] I've read about [sharing account credentials](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#are-you-willing-to-share-account-details-if-needed) and I'm willing to share it if required ### Region china ### Provide a description that is worded well enough to be understood Error while try to download from ixigua.com. ### Provide verbose output that clearly demonstrates the problem - [X] Run **your** yt-dlp command with **-vU** flag added (`yt-dlp -vU <your command line>`) - [ ] If using API, add `'verbose': True` to `YoutubeDL` params instead - [X] Copy the WHOLE output (starting with `[debug] Command-line config`) and insert it below ### Complete Verbose Output ```shell [debug] Command-line config: ['-vU', 'https://www.ixigua.com/7401795461469700634', '--ignore-config', '--ignore-errors', '--cookies', 'e:\\cookies.txt', '--referer', 'https://www.ixigua.com/'] [debug] Encodings: locale cp936, fs utf-8, pref cp936, out utf-8, error utf-8, screen utf-8 [debug] yt-dlp version nightly@2024.09.29.232819 from yt-dlp/yt-dlp-nightly-builds [6328e2e67] (win_exe) [debug] Python 3.8.10 (CPython AMD64 64bit) - Windows-10-10.0.19045-SP0 (OpenSSL 1.1.1k 25 Mar 2021) [debug] exe versions: ffmpeg n7.0.2-19-g45ecf80f0e-20240928 (setts), ffprobe 2.7 [debug] Optional libraries: Cryptodome-3.20.0, brotli-1.1.0, certifi-2024.08.30, curl_cffi-0.5.10, mutagen-1.47.0, requests-2.32.3, sqlite3-3.35.5, urllib3-2.2.3, websockets-13.1 [debug] Proxy map: {} [debug] Request Handlers: urllib, requests, websockets, curl_cffi [debug] Loaded 1838 extractors [debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp-nightly-builds/releases/latest Latest version: nightly@2024.09.29.232819 from yt-dlp/yt-dlp-nightly-builds yt-dlp is up to date (nightly@2024.09.29.232819 from yt-dlp/yt-dlp-nightly-builds) [Ixigua] Extracting URL: https://www.ixigua.com/7401795461469700634 [Ixigua] 740179546146970063: Downloading webpage ERROR: [Ixigua] 740179546146970063: 740179546146970063: Failed to parse JSON (caused by JSONDecodeError('Extra data in \'atedData" = "functio\': line 2 column 37 (char 37)')); please report this issue on https://github.com/yt-dlp/yt-dlp/issues?q= , filling out the appropriate issue template. Confirm you are on the latest version using yt-dlp -U File "yt_dlp\extractor\common.py", line 741, in extract File "yt_dlp\extractor\ixigua.py", line 67, in _real_extract File "yt_dlp\extractor\ixigua.py", line 42, in _get_json_data File "yt_dlp\extractor\common.py", line 1093, in _parse_json File "yt_dlp\extractor\common.py", line 1076, in __print_error File "yt_dlp\utils\_utils.py", line 552, in decode File "json\decoder.py", line 340, in decode json.decoder.JSONDecodeError: Extra data: line 2 column 37 (char 37) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "yt_dlp\extractor\common.py", line 1090, in _parse_json File "json\__init__.py", line 370, in loads File "yt_dlp\utils\_utils.py", line 560, in decode json.decoder.JSONDecodeError: Extra data in 'atedData" = "functio': line 2 column 37 (char 37) ```
geo-blocked,site-bug,patch-available,needs-testing
low
Critical
2,556,932,774
vscode
how to disable Word Wrap globally
## Environment data Version: 1.93.1 (system setup) Commit: 38c31bc77e0dd6ae88a4e9cc93428cc27a56ba40 Date: 2024-09-11T17:20:05.685Z Electron: 30.4.0 ElectronBuildId: 10073054 Chromium: 124.0.6367.243 Node.js: 20.15.1 V8: 12.4.254.20-electron.0 OS: Windows_NT x64 10.0.22631 ## Expected behaviour Whenver we use the shortcut key "alt + z" it should disable wordwrap on all jupyter Notebook Cells ## Actual behaviour It is currently running on individual cell level and whenever we reopen the VSCode, it will be wordwrap again, so need to click each cell and click alt + z to remove word wraps ## Steps to reproduce: 1. create multiple notebook input cells with large string data in single line 2. inbetween create input cells with smaller strings 3. and then keep individual input cell unwrapped using alt + z 4. restart the VScode, and you will see all the cell input values are wordwrapped again. 5. even if we save all input cells with unwrapped data ( alt + z ), after vscode restart, it will wrapped again. Default I have made wordwrap Off everywhere. In below screenshot it shows "Modified elsewhere" popup, there also I have checked and made settings to Off. but still issue persists after restarting the VSCode. ![image](https://github.com/user-attachments/assets/29a3bf7a-668c-4bcc-b518-b1484e207a7e)
bug
low
Major
2,556,938,701
flutter
Create a migrator for android 35/16kb page size cmake flags for plugin_ffi
Follow on work from https://github.com/flutter/flutter/pull/155508 which updates the templates. Examples of updating the app cmake file here. Impl file: lib/src/migrations/cmake_native_assets_migration.dart Test dir: test/general.shard/migrations/cmake_project_migration_test.dart This bug will need a migrator for the plugin cmake file.
c: new feature,framework,fyi-ecosystem
low
Critical
2,556,961,325
ollama
Uneven split across GPUs
When loading a model across 2 GPUs, the layers are split evenly, but the GPU memory usage is quite a bit higher on the first GPU: ``` |=========================================+========================+======================| | 0 NVIDIA GeForce RTX 3090 On | 00000000:01:00.0 Off | N/A | | 55% 58C P0 188W / 275W | 23747MiB / 24576MiB | 27% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 1 NVIDIA GeForce RTX 3090 On | 00000000:02:00.0 Off | N/A | | 53% 49C P0 179W / 275W | 22519MiB / 24576MiB | 26% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ ``` This seems to be due do `CUDA0 compute buffer size` being ~1GB higher than the CUDA1 equivalent. When using text-generation-webui & llamacpp I'm able to specify a `50,51` split which results in the second GPU getting a layer or two more, thus balancing the memory usage, and allowing bigger models to run (or more layers to get offloaded). Does this exist? If not, is it possible? <details> <summary>Server log</summary> ``` Sep 30 15:00:25 dying-love 6f929011afab[1234]: time=2024-09-30T15:00:25.780Z level=INFO source=memory.go:326 msg="offload to cuda" layers.requested=78 layers.model=81 layers.offload=70 layers.split=35,35 memory.available="[23.3 GiB 23.3 GiB]" memory.gpu_overhead="0 B" memory.required.full="52.4 GiB" memory.required.partial="45.9 GiB" memory.required.kv="5.0 GiB" memory.required.allocations="[23.0 GiB 22.9 GiB]" memory.weights.total="44.3 GiB" memory.weights.repeating="43.3 GiB" memory.weights.nonrepeating="974.6 MiB" memory.graph.full="2.6 GiB" memory.graph.partial="2.6 GiB" Sep 30 15:00:25 dying-love 6f929011afab[1234]: time=2024-09-30T15:00:25.783Z level=INFO source=server.go:388 msg="starting llama server" cmd="/usr/lib/ollama/runners/cuda_v12/ollama_llama_server --model /root/.ollama/models/blobs/sha256-c9ff230988a3c90f5beec5da2ebbd8b77d953389b587cb7398c6abd671b7562f --ctx-size 16384 --batch-size 512 --embedding --log-disable --n-gpu-layers 78 --flash-attn --parallel 1 --tensor-split 35,35 --port 37883" Sep 30 15:00:25 dying-love 6f929011afab[1234]: time=2024-09-30T15:00:25.783Z level=INFO source=sched.go:449 msg="loaded runners" count=1 Sep 30 15:00:25 dying-love 6f929011afab[1234]: time=2024-09-30T15:00:25.783Z level=INFO source=server.go:587 msg="waiting for llama runner to start responding" Sep 30 15:00:25 dying-love 6f929011afab[1234]: time=2024-09-30T15:00:25.783Z level=INFO source=server.go:621 msg="waiting for server to become available" status="llm server error" Sep 30 15:00:25 dying-love 6f929011afab[1234]: INFO [main] build info | build=10 commit="3f6ec33" tid="139699938238464" timestamp=1727708425 Sep 30 15:00:25 dying-love 6f929011afab[1234]: INFO [main] system info | n_threads=16 n_threads_batch=16 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="139699938238464" timestamp=1727708425 total_threads=32 Sep 30 15:00:25 dying-love 6f929011afab[1234]: INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="31" port="37883" tid="139699938238464" timestamp=1727708425 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: loaded meta data with 35 key-value pairs and 963 tensors from /root/.ollama/models/blobs/sha256-c9ff230988a3c90f5beec5da2ebbd8b77d953389b587cb7398c6abd671b7562f (version GGUF V3 (latest)) Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 0: general.architecture str = qwen2 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 1: general.type str = model Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 2: general.name str = Qwen2.5 72B Instruct Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 3: general.finetune str = Instruct Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 4: general.basename str = Qwen2.5 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 5: general.size_label str = 72B Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 6: general.license str = other Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 7: general.license.name str = qwen Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 8: general.license.link str = https://huggingface.co/Qwen/Qwen2.5-7... Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 9: general.base_model.count u32 = 1 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 10: general.base_model.0.name str = Qwen2.5 72B Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 11: general.base_model.0.organization str = Qwen Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 12: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen2.5-72B Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 13: general.tags arr[str,2] = ["chat", "text-generation"] Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 14: general.languages arr[str,1] = ["en"] Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 15: qwen2.block_count u32 = 80 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 16: qwen2.context_length u32 = 32768 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 17: qwen2.embedding_length u32 = 8192 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 18: qwen2.feed_forward_length u32 = 29568 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 19: qwen2.attention.head_count u32 = 64 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 20: qwen2.attention.head_count_kv u32 = 8 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 21: qwen2.rope.freq_base f32 = 1000000.000000 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 22: qwen2.attention.layer_norm_rms_epsilon f32 = 0.000001 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 23: general.file_type u32 = 14 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 24: tokenizer.ggml.model str = gpt2 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 25: tokenizer.ggml.pre str = qwen2 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 26: tokenizer.ggml.tokens arr[str,152064] = ["!", "\"", "#", "$", "%", "&", "'", ... Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 27: tokenizer.ggml.token_type arr[i32,152064] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 28: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 29: tokenizer.ggml.eos_token_id u32 = 151645 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 30: tokenizer.ggml.padding_token_id u32 = 151643 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 31: tokenizer.ggml.bos_token_id u32 = 151643 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 32: tokenizer.ggml.add_bos_token bool = false Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 33: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - kv 34: general.quantization_version u32 = 2 Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - type f32: 401 tensors Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - type q5_0: 70 tensors Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - type q5_1: 10 tensors Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - type q4_K: 401 tensors Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - type q5_K: 80 tensors Sep 30 15:00:25 dying-love 6f929011afab[1234]: llama_model_loader: - type q6_K: 1 tensors Sep 30 15:00:26 dying-love 6f929011afab[1234]: time=2024-09-30T15:00:26.034Z level=INFO source=server.go:621 msg="waiting for server to become available" status="llm server loading model" Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_vocab: special tokens cache size = 22 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_vocab: token to piece cache size = 0.9310 MB Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: format = GGUF V3 (latest) Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: arch = qwen2 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: vocab type = BPE Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_vocab = 152064 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_merges = 151387 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: vocab_only = 0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_ctx_train = 32768 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_embd = 8192 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_layer = 80 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_head = 64 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_head_kv = 8 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_rot = 128 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_swa = 0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_embd_head_k = 128 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_embd_head_v = 128 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_gqa = 8 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_embd_k_gqa = 1024 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_embd_v_gqa = 1024 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: f_norm_eps = 0.0e+00 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: f_norm_rms_eps = 1.0e-06 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: f_clamp_kqv = 0.0e+00 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: f_max_alibi_bias = 0.0e+00 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: f_logit_scale = 0.0e+00 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_ff = 29568 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_expert = 0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_expert_used = 0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: causal attn = 1 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: pooling type = 0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: rope type = 2 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: rope scaling = linear Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: freq_base_train = 1000000.0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: freq_scale_train = 1 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: n_ctx_orig_yarn = 32768 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: rope_finetuned = unknown Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: ssm_d_conv = 0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: ssm_d_inner = 0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: ssm_d_state = 0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: ssm_dt_rank = 0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: ssm_dt_b_c_rms = 0 Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: model type = 70B Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: model ftype = Q4_K - Small Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: model params = 72.71 B Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: model size = 40.87 GiB (4.83 BPW) Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: general.name = Qwen2.5 72B Instruct Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: BOS token = 151643 '<|endoftext|>' Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: EOS token = 151645 '<|im_end|>' Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: PAD token = 151643 '<|endoftext|>' Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: LF token = 148848 'ÄĬ' Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: EOT token = 151645 '<|im_end|>' Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_print_meta: max token length = 256 Sep 30 15:00:26 dying-love 6f929011afab[1234]: ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no Sep 30 15:00:26 dying-love 6f929011afab[1234]: ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no Sep 30 15:00:26 dying-love 6f929011afab[1234]: ggml_cuda_init: found 2 CUDA devices: Sep 30 15:00:26 dying-love 6f929011afab[1234]: Device 0: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes Sep 30 15:00:26 dying-love 6f929011afab[1234]: Device 1: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes Sep 30 15:00:26 dying-love 6f929011afab[1234]: llm_load_tensors: ggml ctx size = 1.27 MiB Sep 30 15:00:27 dying-love 6f929011afab[1234]: time=2024-09-30T15:00:27.490Z level=INFO source=server.go:621 msg="waiting for server to become available" s> Sep 30 15:00:29 dying-love 6f929011afab[1234]: llm_load_tensors: offloading 78 repeating layers to GPU Sep 30 15:00:29 dying-love 6f929011afab[1234]: llm_load_tensors: offloaded 78/81 layers to GPU Sep 30 15:00:29 dying-love 6f929011afab[1234]: llm_load_tensors: CPU buffer size = 41850.31 MiB Sep 30 15:00:29 dying-love 6f929011afab[1234]: llm_load_tensors: CUDA0 buffer size = 19646.28 MiB Sep 30 15:00:29 dying-love 6f929011afab[1234]: llm_load_tensors: CUDA1 buffer size = 19530.78 MiB Sep 30 15:00:29 dying-love 6f929011afab[1234]: time=2024-09-30T15:00:29.545Z level=INFO source=server.go:621 msg="waiting for server to become available" s> Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: n_ctx = 16384 Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: n_batch = 512 Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: n_ubatch = 512 Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: flash_attn = 1 Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: freq_base = 1000000.0 Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: freq_scale = 1 Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_kv_cache_init: CUDA_Host KV buffer size = 128.00 MiB Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_kv_cache_init: CUDA0 KV buffer size = 2496.00 MiB Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_kv_cache_init: CUDA1 KV buffer size = 2496.00 MiB Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: KV self size = 5120.00 MiB, K (f16): 2560.00 MiB, V (f16): 2560.00 MiB Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: CUDA_Host output buffer size = 0.61 MiB Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: CUDA0 compute buffer size = 1287.53 MiB Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: CUDA1 compute buffer size = 163.50 MiB Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: CUDA_Host compute buffer size = 48.01 MiB Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: graph nodes = 2487 Sep 30 15:00:32 dying-love 6f929011afab[1234]: llama_new_context_with_model: graph splits = 33 ``` </details>
bug,memory
low
Critical
2,556,980,587
vscode
[json] Add support for extensions associating JSON schemas with documents dynamically (via API)
<!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> Add support for extensions associating JSON schemas with documents dynamically (via API). Right now, it is only possible to associate a JSON schema with a text document by: 1. using schemas directly built-in into VSCode, e.g., `package.json`, `.babelrc` , ... 2. explicitly [mapping it in the user settings](https://code.visualstudio.com/docs/languages/json#_mapping-in-the-user-settings) 3. explicitly [mapping it in the workspace](https://code.visualstudio.com/docs/languages/json#_mapping-to-a-schema-in-the-workspace) 4. explicitly [mapping it in an extension via a file-match pattern](https://code.visualstudio.com/docs/languages/json#_file-match-syntax) But all these associations are static and working only on a folder path and filename matching. I would like to be able to dynamically/programmatically associate a JSON schema with a document within an extension. E.g., the extension could first inspect the content of the file and check for specific well-known conditions in order to decide, which JSON schema (or if at all) should be associated with a given text document. E.g., within our organization we have a lot of different file formats with existing JSON schemas, but we don't have canonical/standardized folders and filenames for the files that would need validation against these JSON schemas, with most of them just ending on `*.json` somewhere in the folder tree. I could imagine a mechanism like this to associate a JSON schema with a JSON document `doc`: ``` registerJsonSchema(doc, schemaId, schemaSrc) ``` with a unique `schemaId` provided by the extension and `schemaSrc` either be `URL` or a local file reference. If the same `schemaId` is provided for a `doc` with no `schemaSrc` the JSON schema association is removed again from that document. Theoretically `schemaSrc` could also be a JSON schema object (similar to a [schema directly defined in the settings](https://code.visualstudio.com/docs/languages/json#_mapping-to-a-schema-defined-in-settings)), which would allow you to dynamically build up a schema by the extension and provide this for validation. And since a lot of configuration files are not just `JSON` files but also `YAML` files, some bonus points for allowing the association of JSON schemas also to `YAML` files, not just `JSON` files. Note, that there is also another related feature request: `Support for contentMatch in addition to fileMatch for extension contributes.jsonValidation` https://github.com/microsoft/vscode/issues/229696
feature-request,json
low
Minor
2,556,991,410
go
runtime:cpu1: TestGdbAutotmpTypes failures
``` #!watchflakes default <- pkg == "runtime:cpu1" && test == "TestGdbAutotmpTypes" ``` Issue created automatically to collect these failures. Example ([log](https://ci.chromium.org/b/8735486120996203153)): === RUN TestGdbAutotmpTypes === PAUSE TestGdbAutotmpTypes === CONT TestGdbAutotmpTypes runtime-gdb_test.go:78: gdb version 15.0 — [watchflakes](https://go.dev/wiki/Watchflakes)
NeedsInvestigation,arch-riscv,compiler/runtime
low
Critical
2,556,992,186
TypeScript
undefined is narrowed away incorrectly
### 🔎 Search Terms narrowing undefined, incorrect type narrowing, incorrect if statement narrowing ### 🕗 Version & Regression Information - This changed between versions 5.2 and 5.3 ### ⏯ Playground Link https://www.typescriptlang.org/play/?#code/CYUwxgNghgTiAEkoGdnwILwN4CgCQAksugBQCUAXPAC4AWAlmoxgNw4C+OOoSciA9gDtk1eAAco1aiBiCqmAD7wAroNAAzeoJDA261WGr0h8KGLEQAngAVJ02eWz566+CQlSZggPwA6IqRk8AC8ofDqUBDIIEG4eHge9oJs8QD0qfAAet74eHDUyrIpnAl2XinpWTnsQA ### 💻 Code ```ts declare class A { IsA(): this is A; } declare const pattern: A | undefined; function applyPattern() { if (pattern?.IsA() === false) { pattern; // ^? return; } pattern; // ^? } ``` ### 🙁 Actual behavior `pattern` is narrowed down to `A`, removing the undefined ### 🙂 Expected behavior `pattern` should only be narrowed to `A | undefined` ### Additional information about the issue _No response_
Bug,Help Wanted
low
Minor
2,557,016,328
pytorch
[RFC] Flight Precheck: ahead-of-hang collective mismatch detection
### 🚀 The feature, motivation and pitch Wondering if we would be interested in making a UX like this: ``` with flight_precheck: # user program ``` A couple motivations: (1) Would allow/encourage users to actively narrow a region for debugging, for example, 1 iteration. So that analysis is simpler. (2) Would leverage the fact that CPU runs ahead of GPU — when the context is exited, it can do a CPU-based all-gather (or similar) and pre-check, thus analysis made online / ahead of time. (3) The torch lib would contain the analysis algorithm, mitigating the need for releasing separate software (not easy to maintain if torch lib changes its recording format). (4) Can mostly reuse today’s FR tracing infra. Cc: @c-p-i-o ### Alternatives _No response_ ### Additional context _No response_ cc @XilunWu @H-Huang @awgu @wanchaol @fegin @fduwjj @wz337 @wconstab @d4l3k @c-p-i-o
oncall: distributed,triaged
low
Critical
2,557,027,641
flutter
sample code for `Custom image providers` throwing error on run
### Page URL https://api.flutter.dev/flutter/widgets/Image-class.html ### Page source _No response_ ### Describe the problem <img width="967" alt="image" src="https://github.com/user-attachments/assets/4335f8d8-0895-4805-8c92-013b8029f089"> Error message: ```shell unsupported library on the web: 'dart:io' Try removing the import and usage of the library. Unsupported operation: Platform._version ``` ### Expected fix Need to fix the given sample code to run with out error ### Additional context _No response_ ### I would like to fix this problem. - [ ] I will try and fix this problem on docs.flutter.dev.
framework,d: api docs,d: examples,platform-web,P2,team-framework,triaged-framework
low
Critical
2,557,052,974
pytorch
[RFC] ProcessGroupNCCL uses non-blocking API by default
### 🚀 The feature, motivation and pitch ## Motivation: - Gives better fault tolerance against hangs in comm init, comm destroy and P2P operations (involves dynamic connection performed by CPU). - Allow overlap between NCCL init and other stuff users want the main thread to do (e.g., model or data loader init) Here "non-blocking" refers to whether a NCCL API call would immediately return or block the host CPU (Traditionally, some NCCL APIs such as ncclCommInitRank may block for a little when rendezvous is performed.) If the user wants the main thread to do other stuff while NCCL is initializing, this mode would also help as it puts NCCL init to the background. ### Alternatives _No response_ ### Additional context This knob is control by PyTorch here: https://github.com/pytorch/pytorch/blob/26956980c6953b42e57ad889b87444c4abccec1c/torch/csrc/distributed/c10d/NCCLUtils.cpp#L90-L97 cc @XilunWu @H-Huang @awgu @wanchaol @fegin @fduwjj @wz337 @wconstab @d4l3k @c-p-i-o
oncall: distributed,module: nccl,module: c10d
low
Minor
2,557,072,866
pytorch
pre-dispatch tracing can't interpose on detach() calls from the autograd engine.
This boils down to the fact that our pre-dispatch mode stack is implemented in pure python, while our "normal" mode stack has C++ TLS associated with it. Say we want to trace out calls to `.backward()` using a "pre-dispatch" proxy mode. Autograd sometimes detaches saved variables while it executes the backward. It doesn't call `TensorBase::detach()` though - it calls `tensor.getIntrusivePtr()->shallow_copy_and_detach(...)` ([link](https://github.com/pytorch/pytorch/blob/main/torch/csrc/autograd/variable.h#L897)) `TensorImpl::shallow_copy_and_detach_core` has some special logic to figure out when it should dispatch to python, by consulting the C++ TLS about if there are any modes active ([link](https://github.com/pytorch/pytorch/blob/c13c7e11c553b2a848c158c8e2a5e20bbd5d7f1c/c10/core/TensorImpl.cpp#L499)). Our pre-dispatch mode stack doesn't have any C++ TLS, though, which means that `shallow_copy_and_detach()` won't know to dispatch to any pre-dispatch modes. cc @ezyang @chauhang @penguinwu @zou3519 @yf225
triaged,oncall: pt2,module: ProxyTensor,module: pt2-dispatcher
low
Minor
2,557,085,292
pytorch
[RFC] Investigate multi-thread support of c10d
### 🚀 The feature, motivation and pitch There are two drivers for this investigation: - People are starting to use torch.distributed APIs in multi-thread mode. We should investigate which usage is supported and which are not. For the mode not supported, we should document it so that it won't become a support of fait accompli. - Python 3.13t is coming (Oct 2024). This is the first official version with option to disable GIL. There may be interest in inference, data loading, etc, where torch.distributed should be compatible. ### Alternatives _No response_ ### Additional context _No response_ cc @XilunWu @H-Huang @awgu @wanchaol @fegin @fduwjj @wz337 @wconstab @d4l3k @c-p-i-o
oncall: distributed,module: c10d
low
Minor
2,557,120,862
pytorch
[RFC] Use NCCL topology detection for Inductor communication time estimation
### 🚀 The feature, motivation and pitch Today the bandwidth numbers in Inductor are more or less fixed, regardless of the system the job runs on. NCCL topology detection identifies whether the system is on NVLink, PCI-e, IB, or TCP sockets, and can provide a bandwidth matrix. A bonus is that the numbers have been sync'ed across all ranks, so that we can avoid different decisions made by compilers on different ranks. See https://github.com/NVIDIA/nccl/issues/830 for an example of the matrix. ### Alternatives _No response_ ### Additional context _No response_ cc @XilunWu @H-Huang @awgu @wanchaol @fegin @fduwjj @wz337 @wconstab @d4l3k @c-p-i-o @ezyang @chauhang @penguinwu @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @ColinPeppler @amjames @desertfire @xuanzhang816
oncall: distributed,triaged,module: nccl,oncall: pt2,module: inductor
low
Minor
2,557,134,462
ollama
ollama does not detect Quadro RTX 4000 - cuda driver library failed to get device context 801
### What is the issue? Hi All, I installed ollama both (on machine/docker) both with same behaviour of not detecting the GPU. Have LM Studio on the same machine which picks up GPU without any issues. ``` root@d50a3f8d8474:/# ollama run phi3.5:3.8b-mini-instruct-q2_K "" root@d50a3f8d8474:/# ollama ps NAME ID SIZE PROCESSOR UNTIL phi3.5:3.8b-mini-instruct-q2_K 45b8dc82a846 5.3 GB 100% CPU 4 minutes from now ``` **Installation** ``` [root@ai ~]# curl -fsSL https://ollama.com/install.sh | sh >>> Installing ollama to /usr/local >>> Downloading Linux amd64 bundle ######################################################################## 100.0%#=#=# ######################################################################## 100.0% >>> Creating ollama user... >>> Adding ollama user to render group... >>> Adding ollama user to video group... >>> Adding current user to ollama group... >>> Creating ollama systemd service... >>> Enabling and starting ollama service... Created symlink /etc/systemd/system/default.target.wants/ollama.service → /etc/systemd/system/ollama.service. >>> NVIDIA GPU installed. ``` *Logs from package installation* ``` [root@ai ~]# OLLAMA_DEBUG=1 ollama serve Error: listen tcp 127.0.0.1:11434: bind: address already in use [root@ai ~]# systemctl stop ollama [root@ai ~]# OLLAMA_DEBUG=1 ollama serve 2024/09/29 03:47:20 routes.go:1153: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:true OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://*] OLLAMA_SCHED_SPREAD:false OLLAMA_TMPDIR: ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2024-09-29T03:47:20.643-05:00 level=INFO source=images.go:753 msg="total blobs: 10" time=2024-09-29T03:47:20.672-05:00 level=INFO source=images.go:760 msg="total unused blobs removed: 0" time=2024-09-29T03:47:20.672-05:00 level=INFO source=routes.go:1200 msg="Listening on 127.0.0.1:11434 (version 0.3.12)" time=2024-09-29T03:47:20.673-05:00 level=INFO source=common.go:135 msg="extracting embedded files" dir=/tmp/ollama3184037398/runners time=2024-09-29T03:47:20.673-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cpu payload=linux/amd64/cpu/libggml.so.gz time=2024-09-29T03:47:20.673-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cpu payload=linux/amd64/cpu/libllama.so.gz time=2024-09-29T03:47:20.673-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cpu payload=linux/amd64/cpu/ollama_llama_server.gz time=2024-09-29T03:47:20.673-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cpu_avx payload=linux/amd64/cpu_avx/libggml.so.gz time=2024-09-29T03:47:20.674-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cpu_avx payload=linux/amd64/cpu_avx/libllama.so.gz time=2024-09-29T03:47:20.674-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cpu_avx payload=linux/amd64/cpu_avx/ollama_llama_server.gz time=2024-09-29T03:47:20.674-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cpu_avx2 payload=linux/amd64/cpu_avx2/libggml.so.gz time=2024-09-29T03:47:20.674-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cpu_avx2 payload=linux/amd64/cpu_avx2/libllama.so.gz time=2024-09-29T03:47:20.674-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cpu_avx2 payload=linux/amd64/cpu_avx2/ollama_llama_server.gz time=2024-09-29T03:47:20.674-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cuda_v11 payload=linux/amd64/cuda_v11/libggml.so.gz time=2024-09-29T03:47:20.674-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cuda_v11 payload=linux/amd64/cuda_v11/libllama.so.gz time=2024-09-29T03:47:20.674-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cuda_v11 payload=linux/amd64/cuda_v11/ollama_llama_server.gz time=2024-09-29T03:47:20.674-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cuda_v12 payload=linux/amd64/cuda_v12/libggml.so.gz time=2024-09-29T03:47:20.675-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cuda_v12 payload=linux/amd64/cuda_v12/libllama.so.gz time=2024-09-29T03:47:20.675-05:00 level=DEBUG source=common.go:168 msg=extracting runner=cuda_v12 payload=linux/amd64/cuda_v12/ollama_llama_server.gz time=2024-09-29T03:47:20.675-05:00 level=DEBUG source=common.go:168 msg=extracting runner=rocm_v60102 payload=linux/amd64/rocm_v60102/libggml.so.gz time=2024-09-29T03:47:20.675-05:00 level=DEBUG source=common.go:168 msg=extracting runner=rocm_v60102 payload=linux/amd64/rocm_v60102/libllama.so.gz time=2024-09-29T03:47:20.676-05:00 level=DEBUG source=common.go:168 msg=extracting runner=rocm_v60102 payload=linux/amd64/rocm_v60102/ollama_llama_server.gz time=2024-09-29T03:47:32.712-05:00 level=DEBUG source=common.go:294 msg="availableServers : found" file=/tmp/ollama3184037398/runners/cpu/ollama_llama_server time=2024-09-29T03:47:32.712-05:00 level=DEBUG source=common.go:294 msg="availableServers : found" file=/tmp/ollama3184037398/runners/cpu_avx/ollama_llama_server time=2024-09-29T03:47:32.713-05:00 level=DEBUG source=common.go:294 msg="availableServers : found" file=/tmp/ollama3184037398/runners/cpu_avx2/ollama_llama_server time=2024-09-29T03:47:32.713-05:00 level=DEBUG source=common.go:294 msg="availableServers : found" file=/tmp/ollama3184037398/runners/cuda_v11/ollama_llama_server time=2024-09-29T03:47:32.713-05:00 level=DEBUG source=common.go:294 msg="availableServers : found" file=/tmp/ollama3184037398/runners/cuda_v12/ollama_llama_server time=2024-09-29T03:47:32.713-05:00 level=DEBUG source=common.go:294 msg="availableServers : found" file=/tmp/ollama3184037398/runners/rocm_v60102/ollama_llama_server time=2024-09-29T03:47:32.713-05:00 level=INFO source=common.go:49 msg="Dynamic LLM libraries" runners="[rocm_v60102 cpu cpu_avx cpu_avx2 cuda_v11 cuda_v12]" time=2024-09-29T03:47:32.713-05:00 level=DEBUG source=common.go:50 msg="Override detection logic by setting OLLAMA_LLM_LIBRARY" time=2024-09-29T03:47:32.713-05:00 level=DEBUG source=sched.go:105 msg="starting llm scheduler" time=2024-09-29T03:47:32.713-05:00 level=INFO source=gpu.go:199 msg="looking for compatible GPUs" time=2024-09-29T03:47:32.713-05:00 level=DEBUG source=gpu.go:86 msg="searching for GPU discovery libraries for NVIDIA" time=2024-09-29T03:47:32.713-05:00 level=DEBUG source=gpu.go:468 msg="Searching for GPU library" name=libcuda.so* time=2024-09-29T03:47:32.713-05:00 level=DEBUG source=gpu.go:491 msg="gpu library search" globs="[/usr/local/lib/ollama/libcuda.so* /usr/local/cuda/lib64/libcuda.so* /root/libcuda.so* /usr/local/cuda*/targets/*/lib/libcuda.so* /usr/lib/*-linux-gnu/nvidia/current/libcuda.so* /usr/lib/*-linux-gnu/libcuda.so* /usr/lib/wsl/lib/libcuda.so* /usr/lib/wsl/drivers/*/libcuda.so* /opt/cuda/lib*/libcuda.so* /usr/local/cuda/lib*/libcuda.so* /usr/lib*/libcuda.so* /usr/local/lib*/libcuda.so*]" time=2024-09-29T03:47:32.715-05:00 level=DEBUG source=gpu.go:525 msg="discovered GPU libraries" paths=[/usr/lib64/libcuda.so.560.35.03] CUDA driver version: 12.6 time=2024-09-29T03:47:32.878-05:00 level=DEBUG source=gpu.go:118 msg="detected GPUs" count=1 library=/usr/lib64/libcuda.so.560.35.03 time=2024-09-29T03:47:32.907-05:00 level=INFO source=gpu.go:252 msg="error looking up nvidia GPU memory" error="cuda driver library failed to get device context 801" time=2024-09-29T03:47:32.907-05:00 level=DEBUG source=amd_linux.go:376 msg="amdgpu driver not detected /sys/module/amdgpu" time=2024-09-29T03:47:32.907-05:00 level=INFO source=gpu.go:347 msg="no compatible GPUs were discovered" releasing cuda driver library time=2024-09-29T03:47:32.907-05:00 level=INFO source=types.go:107 msg="inference compute" id=0 library=cpu variant=avx2 compute="" driver=0.0 name="" total="251.1 GiB" available="240.5 GiB" ``` **Logs from Docker installation** ``` [root@ai ~]# docker logs -f ollama 2024/09/30 15:58:28 routes.go:1153: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://*] OLLAMA_SCHED_SPREAD:false OLLAMA_TMPDIR: ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2024-09-30T15:58:28.508Z level=INFO source=images.go:753 msg="total blobs: 6" time=2024-09-30T15:58:28.509Z level=INFO source=images.go:760 msg="total unused blobs removed: 0" time=2024-09-30T15:58:28.509Z level=INFO source=routes.go:1200 msg="Listening on [::]:11434 (version 0.3.12)" time=2024-09-30T15:58:28.510Z level=INFO source=common.go:49 msg="Dynamic LLM libraries" runners="[cpu_avx cpu_avx2 cuda_v11 cuda_v12 cpu]" time=2024-09-30T15:58:28.510Z level=INFO source=gpu.go:199 msg="looking for compatible GPUs" time=2024-09-30T15:58:28.670Z level=INFO source=gpu.go:252 msg="error looking up nvidia GPU memory" error="cuda driver library failed to get device context 801" time=2024-09-30T15:58:28.670Z level=INFO source=gpu.go:347 msg="no compatible GPUs were discovered" time=2024-09-30T15:58:28.670Z level=INFO source=types.go:107 msg="inference compute" id=0 library=cpu variant=avx2 compute="" driver=0.0 name="" total="251.1 GiB" available="240.4 GiB" [GIN] 2024/09/30 - 15:59:07 | 200 | 94.19µs | 127.0.0.1 | HEAD "/" [GIN] 2024/09/30 - 15:59:07 | 200 | 10.954794ms | 127.0.0.1 | POST "/api/show" time=2024-09-30T15:59:07.334Z level=INFO source=server.go:103 msg="system memory" total="251.1 GiB" free="240.5 GiB" free_swap="4.0 GiB" time=2024-09-30T15:59:07.334Z level=INFO source=memory.go:326 msg="offload to cpu" layers.requested=-1 layers.model=33 layers.offload=0 layers.split="" memory.available="[240.5 GiB]" memory.gpu_overhead="0 B" memory.required.full="8.3 GiB" memory.required.partial="0 B" memory.required.kv="4.0 GiB" memory.required.allocations="[8.3 GiB]" memory.weights.total="7.4 GiB" memory.weights.repeating="7.3 GiB" memory.weights.nonrepeating="102.6 MiB" memory.graph.full="560.0 MiB" memory.graph.partial="681.0 MiB" time=2024-09-30T15:59:07.338Z level=INFO source=server.go:388 msg="starting llama server" cmd="/usr/lib/ollama/runners/cpu_avx2/ollama_llama_server --model /root/.ollama/models/blobs/sha256-8934d96d3f08982e95922b2b7a2c626a1fe873d7c3b06e8e56d7bc0a1fef9246 --ctx-size 8192 --batch-size 512 --embedding --log-disable --no-mmap --numa distribute --parallel 4 --port 39753" time=2024-09-30T15:59:07.339Z level=INFO source=sched.go:449 msg="loaded runners" count=1 time=2024-09-30T15:59:07.339Z level=INFO source=server.go:587 msg="waiting for llama runner to start responding" time=2024-09-30T15:59:07.339Z level=INFO source=server.go:621 msg="waiting for server to become available" status="llm server error" WARNING: /proc/sys/kernel/numa_balancing is enabled, this has been observed to impair performance INFO [main] build info | build=10 commit="070c75f" tid="140389372093376" timestamp=1727711947 INFO [main] system info | n_threads=20 n_threads_batch=20 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 1 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="140389372093376" timestamp=1727711947 total_threads=40 INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="39" port="39753" tid="140389372093376" timestamp=1727711947 llama_model_loader: loaded meta data with 23 key-value pairs and 291 tensors from /root/.ollama/models/blobs/sha256-8934d96d3f08982e95922b2b7a2c626a1fe873d7c3b06e8e56d7bc0a1fef9246 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. ``` I am able to get all the outputs ``` [root@ai ~]# nvidia-smi Sat Sep 28 01:07:12 2024 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 560.35.03 Driver Version: 560.35.03 CUDA Version: 12.6 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 Quadro RTX 4000 Off | 00000000:37:00.0 Off | N/A | | 30% 34C P8 9W / 125W | 1MiB / 8192MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | No running processes found | +-----------------------------------------------------------------------------------------+ ``` ``` [root@ai ~]# nvcc --version nvcc: NVIDIA (R) Cuda compiler driver Copyright (c) 2005-2024 NVIDIA Corporation Built on Wed_Aug_14_10:10:22_PDT_2024 Cuda compilation tools, release 12.6, V12.6.68 Build cuda_12.6.r12.6/compiler.34714021_0 ``` **OS** Linux Rocky 9.4 ``` [root@ai ~]# uname -r 5.14.0-427.37.1.el9_4.x86_64 ``` **logs** ``` [root@ai ~]# sudo dmesg | grep -i nvidia [ 1.704573] Loaded X.509 cert 'Rocky Enterprise Software Foundation: Nvidia GPU OOT Signing 101: 816ba9c770e6960cefe378020865d4ebbc352a7d' [ 6.270595] input: HDA NVidia HDMI/DP,pcm=3 as /devices/pci0000:36/0000:36:00.0/0000:37:00.1/sound/card0/input6 [ 6.270694] input: HDA NVidia HDMI/DP,pcm=7 as /devices/pci0000:36/0000:36:00.0/0000:37:00.1/sound/card0/input7 [ 6.270796] input: HDA NVidia HDMI/DP,pcm=8 as /devices/pci0000:36/0000:36:00.0/0000:37:00.1/sound/card0/input8 [ 6.270843] input: HDA NVidia HDMI/DP,pcm=9 as /devices/pci0000:36/0000:36:00.0/0000:37:00.1/sound/card0/input9 [ 7.812685] nvidia: loading out-of-tree module taints kernel. [ 7.812696] nvidia: module license 'NVIDIA' taints kernel. [ 7.836076] nvidia: module verification failed: signature and/or required key missing - tainting kernel [ 7.950180] nvidia-nvlink: Nvlink Core is being initialized, major device number 510 [ 7.951760] nvidia 0000:37:00.0: enabling device (0140 -> 0143) [ 7.951842] nvidia 0000:37:00.0: vgaarb: VGA decodes changed: olddecodes=io+mem,decodes=none:owns=none [ 8.001592] NVRM: loading NVIDIA UNIX x86_64 Kernel Module 560.35.03 Fri Aug 16 21:39:15 UTC 2024 [ 8.115320] nvidia_uvm: module uses symbols from proprietary module nvidia, inheriting taint. [ 8.252789] nvidia-uvm: Loaded the UVM driver, major device number 508. [ 8.307886] nvidia-modeset: Loading NVIDIA Kernel Mode Setting Driver for UNIX platforms 560.35.03 Fri Aug 16 21:21:48 UTC 2024 [ 8.323807] [drm] [nvidia-drm] [GPU ID 0x00003700] Loading driver [ 9.814993] [drm] Initialized nvidia-drm 0.0.0 20160202 for 0000:37:00.0 on minor 1 [ 9.815921] nvidia 0000:37:00.0: [drm] Cannot find any crtc or sizes ``` **additional logs** ``` [root@ai ~]# sudo dmesg | grep -i nvrm [ 8.001592] NVRM: loading NVIDIA UNIX x86_64 Kernel Module 560.35.03 Fri Aug 16 21:39:15 UTC 2024 ``` ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.12
bug,linux,nvidia,needs more info
low
Critical
2,557,152,512
PowerToys
*Multiply NumPad missing as an option to rekey from or to
### Microsoft PowerToys version v0.84.1 ### Installation method Microsoft Store ### Running as admin No ### Area(s) with issue? Keyboard Manager ### Steps to reproduce Remapping -subtract key to *multiply key and *multiply key to -subtract key on NumPad ### ✔️ Expected Behavior I expect to be able to choose the * (multiply) key from the drop down under Select. It does not exist in the drop downs for either Select or To Send. ### ❌ Actual Behavior When selecting - (subtract) in the Select column, I can switch to Send Text in the To Send section and enter an * Asterisk. The * (multiply) NumPad key then needs to be remapped, but I can't remap it because the option doesn't appear in the dropdown. ### Other Software _No response_
Issue-Bug,Needs-Triage
low
Minor
2,557,187,369
godot
SkeletonModification2DCCDIK causes errors when a second instance of itself it instantiated
### Tested versions Godot version v4.3.stable.official [77dcf97d8] System information Windows 10, ryzen 5, radeon 3080 ### System information v4.3.stable.official [77dcf97d8], Windows 10, ryzen 5, radeon 3080 ### Issue description When using SkeletonModification2DCCDIK within SkeletonModificationStack2D editor displays huge amount of error messages (before execution). However this happens only when SkeletonModification2DCCDIK is present (not deleted) **AND** a second instance of the character is loaded. A single instance of the character works perfectly fine. ![image](https://github.com/user-attachments/assets/02519731-1588-4e4f-aee3-16533b85837b) ### Steps to reproduce -Create Skeleton2D (based on Poylgon2D) -Assign SkeletonModificationStack2D and SkeletonModification2DCCDIK -Load Character scene into a new project -Duplicate character scene ![image](https://github.com/user-attachments/assets/2b20371c-6f05-45f8-97cd-6771ac9ba2ea) ### Minimal reproduction project (MRP) [minimal test.zip](https://github.com/user-attachments/files/17193664/minimal.test.zip)
bug,topic:animation
low
Critical
2,557,193,410
pytorch
[RFC] Allow user to abort group (followed by custom recovery)
### 🚀 The feature, motivation and pitch Advanced users want to control ProcessGroup abort and recovery, instead of being controlled by watchdog (such as SIGABRT). See https://github.com/pytorch/pytorch/issues/136312 for such ask. In this situation, an abort group API is necessary because that's the only way to stop a spinning NCCL kernel. There is a draft `_abort_process_group` implementation in https://github.com/pytorch/pytorch/pull/132291. Next step is to formalize it, test it and expose as user API. ### Alternatives _No response_ ### Additional context _No response_ cc @XilunWu @H-Huang @awgu @wanchaol @fegin @fduwjj @wz337 @wconstab @d4l3k @c-p-i-o
oncall: distributed,module: c10d
low
Minor
2,557,223,454
pytorch
[RFC] Accelerate intra-node communication via NCCL buffer registration
### 🚀 The feature, motivation and pitch Registering user buffer with NCCL allows for zero copy. While `ncclCommRegister` API allowed this for network operations previously, NCCL 2.23 also allows this for intra-node operations. This is an investigation task wrt: - how to integrate this feature in ProcessGroupNCCL; - how to expose this capability to user; - how to integrate into Inductor memory planning. Related to Symmetric Memory, async TP. Cc: @yifuwang ### Alternatives _No response_ ### Additional context _No response_ cc @XilunWu @H-Huang @awgu @wanchaol @fegin @fduwjj @wz337 @wconstab @d4l3k @c-p-i-o
oncall: distributed,module: nccl
low
Minor
2,557,261,310
TypeScript
[Control flow] Array filter with typeguard works differently following array initial value
### 🔎 Search Terms filter + typeguard + control flow + strictNullChecks ### 🕗 Version & Regression Information - This is the behavior in every version I tried, and I reviewed the FAQ for entries about filter + typeguards + control flow + strictNullChecks ### ⏯ Playground Link https://www.typescriptlang.org/play/?strict=false&noImplicitAny=false&strictFunctionTypes=false&strictPropertyInitialization=false&strictBindCallApply=false&noImplicitThis=false&noImplicitReturns=false&alwaysStrict=false&esModuleInterop=false&declaration=false&target=0&jsx=0&module=0&ts=5.6.2&experimentalDecorators=false&emitDecoratorMetadata=false#code/MYewdgzgLgBATgQwO4C4bTgSzAcwNoC6MAPjAIJyICeAPBtjiTGAK4A2bAfDALwyEBuAFBDQkWADNMbKAFM4sgCZp6uQr3jIAdFJnyAFPuiyADiqhZcTVhwCU50zEwR0Fhr25QqJ2SAmvHHiCYAHJVHBDbGAB6aJgAaQB5ETFoV0scACVkcwz1UgpqOjcrUhsuDTwAIhM2EBMqgmFRcDTwgDFpOQVldIZ1PnDspB0ugyM5Mz7S5nY2ewCTJxdwjxgvHz9F3mCwkoio2ITklvFZjmHc-qICygRaVbK57j48cqaU1thyzr0eq7URD45WGoz+hmMU0e53mDiWzmmjB4nm8vn8kJ2fD2GUiMTiiXinzO5QACmwWBAAMr7S6I-LkO4PfbWZ6VGp1BoAGhhH1OaVJ5Kp+1+3SUAPwQJhZIp1IyoN03Qhk3FLLscOWiLWGzR2yCWPCuKOBKEQA ### 💻 Code ```ts const raw: string[] | (string | null)[] = []; const filtered: string[] = raw.filter((step: string | null): step is string => typeof step === 'string') // KO // Type '(string | null)[]' is not assignable to type 'string[]'. ``` ```ts const raw: string[] |(string | null)[] = ["plop", null]; const filtered: string[] = raw.filter((step: string | null): step is string => typeof step === 'string') // OK no error ``` ### 🙁 Actual behavior There is a strange difference in control flow when you filter nullish values in an array regarding the declaration of that array. This is definitely a edge case. The type `string[] | (string | null)[]` comes from an union type in my case. ### 🙂 Expected behavior No difference and typeguard should work (regardless the fact it is valid). ### Additional information about the issue I first thought it was a regression due to #57465 but I tested it on several version. The issue has always been here. I looked into a lot of issues about typeguards and filters but unfortunately, could not find one similar to my issue :cry: Thanks anyway for the great job
Help Wanted,Possible Improvement
low
Critical
2,557,286,643
godot
(Funny) Godot icon looking extended in Task Manager
### Tested versions - Reproducible in: `master` ### System information Windows 11 24H2 and 23H2 - `master` ### Issue description ![image](https://github.com/user-attachments/assets/966f264e-4153-4b93-acea-9ffd5458702d) ### Steps to reproduce - Launch Godot - Launch Task Manager ### Minimal reproduction project (MRP) N/A
bug,platform:windows,topic:porting,topic:thirdparty
low
Minor
2,557,293,752
flutter
DD for missing Semantics role
### Document Link https://flutter.dev/go/semantics-roles ### What problem are you solving? Audit the missing semantics roles in flutter
a: accessibility,P3,design doc,:scroll:,team-accessibility,triaged-accessibility
low
Minor
2,557,303,538
go
proposal: syscall/js: method or function to await a js.Value
### Proposal Details Let's say I have an async JavaScript function: ```js globalThis.myAsyncFunction = async () => 42; ``` Ideally I would be able to do one of these: ```go var n int n = js.Global().Call("myAsyncFunction").Wait().Int() n = js.Global().Call("myAsyncFunction").Await().Int() n = js.Await(js.Global().Call("myAsyncFunction")).Int() n = (<-js.Global().Call("myAsyncFunction").Chan()).Int() ``` Creating a promise in Go code for use by JavaScript code is _ok_-ish (it's not great): just use the `new Promise(callback)` constructor with a Go-defined `js.Func` callback. ```go var jsDoThingCallback = js.FuncOf(func(this Value, args []Value) interface{} { r1 := <-myChannelFromSomewhere r2, err := someFuncThatUsesChans(r1) if err != nil { args[1].Invoke(errorConstructor.New(err.Error())) return nil } r3, err := someFuncThatSleeps(r2) if err != nil { args[1].Invoke(errorConstructor.New(err.Error())) return nil } args[0].Invoke(r3) return nil }) func DoThingAsync() js.Value { return promiseConstructor.New(jsDoThingCallback) } ``` The other direction -- unwrapping a JavaScript `Promise` instance on the Go-side by waiting for it -- is worse. ```go // Ugh. I have to manage this intermediate channel myself and remember to handle some edge cases. var n int p := promiseConstructor.Call("resolve", js.Global().Call("myAsyncFunction")) type result struct { value js.Value err error } ch := make(chan result, 1) onFulfilled := js.FuncOf(func(this js.Value, args []js.Value) interface{} { ch <- result{args[0], nil} close(ch) return nil }) defer onFulfilled.Release() onRejected := js.FuncOf(func(this js.Value, args []js.Value) interface{} { ch <- result{js.Value{}, js.Error{args[0]}} close(ch) return nil }) defer onRejected.Release() p.Call("then", onFulfilled, onRejected) r := <-ch if r.err != nil { panic(r.err) } n = r.value.Int() // Would have to do *all that again* for another `await nextFunctionThatUsesResult(n)` 😭 ``` Here's the algorithm for the ECMAScript 2025 `Await( value )` abstract operation: https://tc39.es/ecma262/multipage/control-abstraction-objects.html#await It seems like the `.Wait()` method convention is already in the standard library with `sync.WaitGroup` `wg.Wait()`. Here's an idea for how that might look in Go code. I'm not a Go channels wizard so this might be the completely wrong way to do this. ```go // Wait waits for the thenable v to fulfill or reject and returns the resulting value or error. // This is equivalent to the await operator in JavaScript. func (v js.Value) Wait() (js.Value, error) { p := promiseConstructor.Call("resolve", v) type result struct { value js.Value err error } ch := make(chan result, 1) onFulfilled := js.FuncOf(func(this js.Value, args []js.Value) interface{} { ch <- result{args[0], nil} close(ch) return nil }) defer onFulfilled.Release() onRejected := js.FuncOf(func(this js.Value, args []js.Value) interface{} { ch <- result{js.Value{}, js.Error{args[0]}} close(ch) return nil }) defer onRejected.Release() p.Call("then", onFulfilled, onRejected) r := <-ch // Unsure if want to panic on error, or return the error. // If a synchronous function throws an error it panics. Don't know whether to stick // to that convention or to return a (js.Value, error) multivalue. return r.value, r.err // OR if r.err == nil { return r.value } else { panic(r.err) } } ``` There's already some of this promise stuff in the Go standard library https://github.com/golang/go/blob/1d0f5c478ac176fa99d0f3d6bd540e5fb422187a/src/net/http/roundtrip_js.go#L129-L245 so it seems like this is a _thing that people need to do_. I think that providing a `.Wait()` or `Await(v)` or something would be a good way to "one good way to do it"-ify this.
Proposal
low
Critical
2,557,366,732
godot
Global illumination doesn't update when force drawing a `SubViewport` if main viewport is disabled
### Tested versions - 4.4.dev - 4.3.1.rc ### System information Godot v4.4.dev (e3213aaef) - Windows 11 - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 3070 (NVIDIA; 31.0.15.4633) ### Issue description SDFGI takes some frames to converge the global illumination (setting `rendering/global_illumination/sdfgi/frames_to_converge`). However, when force drawing a `SubViewport` like the below example, global illumination doesn't update if the main viewport is disabled. ```gdscript func _render_subviewport(subviewport: SubViewport, iterations: int = 30, disable_main = true) -> Image: var scene_tree = Engine.get_main_loop() as SceneTree var root_viewport = scene_tree.root.get_viewport_rid() if disable_main: # Disable main viewport so it doesn't redrawn RenderingServer.viewport_set_active(root_viewport, false) # Render SubViewport for i in iterations: await RenderingServer.frame_pre_draw RenderingServer.viewport_set_update_mode(render.get_viewport_rid(), RenderingServer.VIEWPORT_UPDATE_ONCE) RenderingServer.force_draw(true, 1.0 / iterations) await RenderingServer.frame_post_draw if disable_main: # Enable main viewport again RenderingServer.viewport_set_active(root_viewport, true) await RenderingServer.frame_post_draw # image data doesn't updates correctly without this.. return subviewport.get_texture().get_image() ``` Left: expected. Right: result of multiple `force_draw()`, lighting didn't iterate and it always looks like the first iteration. ![image](https://github.com/user-attachments/assets/56f4923a-4a75-4fa6-a608-b73dae1a6101) It only updates correclty when the main viewport is enabled during `force_draw()`. ![image](https://github.com/user-attachments/assets/73c3bf23-9bf0-42b3-a3ac-8eb01a3b417d) But it slows down rendering since `force_draw()` draws all active viewports although I don't need the main viewport. ### Steps to reproduce * Open the MRP. It contains the test project of the images above. * Run the project. The left is a realtime `SubViewport`. The right creates a `SubViewport` dynamically, force draw multiple times and shows a `TextureRect` with the resulting image. * Toggle "Disable main viewport". * Rendering only looks correct when the main viewport is enabled. ### Minimal reproduction project (MRP) [sdfgi-update.zip](https://github.com/user-attachments/files/17194488/sdfgi-update.zip)
bug,topic:rendering,topic:3d
low
Major
2,557,367,012
flutter
[Scribe] Hover icon
### Use case Android's Scribe feature supports showing an icon when a stylus hovers over a field that supports Scribe handwriting input. Flutter should support this feature too. ![Screenshot_20240930-111720](https://github.com/user-attachments/assets/71dc95cd-eefe-4d57-8292-2884a77840d5) ### Proposal The hover cursor can be set in `onResolvePointerIcon` in InputConnection in FlutterView.java. To know where the fields are that support Scribe, we'd likely have to send the rects over from the framework. ### See also Scribe issue: https://github.com/flutter/flutter/issues/115607 Scribe support added in https://github.com/flutter/flutter/pull/148784 and https://github.com/flutter/engine/pull/52943.
a: text input,c: new feature,platform-android,engine,c: proposal,P1,team-text-input,triaged-text-input
medium
Minor
2,557,376,889
flutter
Document dirtying the source tree on CI (including `.gitignore`'d) has caching implications
_Forked from https://github.com/flutter/engine/pull/55475#issuecomment-2383597893:_ > CI does filesystem level caching, so even files that are .gitignored in the source tree may persist between CI runs It is surprising and, with the exception of the [raw code in another repo](https://flutter.googlesource.com/recipes/+/refs/heads/main/recipe_modules/cache/api.py), undocumented behavior. We should document what the implications of this caching are, what types of errors or undefined behavior it can lead to, and what we expect tools and test authors to do (for example, using cleanup hooks to delete files, but perhaps more than that?)
engine,P2,c: tech-debt,team-engine,triaged-engine
low
Critical
2,557,380,138
PowerToys
App Hibernation Module
### Description of the new feature / enhancement Allows the user to "hibernate" an application. It will suspend the process + any subprocesses and hide it from the taskbar. There would then be some sort of UI which shows all of the hibernated apps and allows you to bring them back out of hibernation. ### Scenario when this would be used? When you have a game open and you want to switch to doing something else without the game using up a load of resources, while still being able to return to it without needing to wait for it to boot up again. ### Supporting information _No response_
Needs-Triage
low
Minor
2,557,380,524
pytorch
[user triton] error out with export/AOTI if dynamic shape is specialized at tl.constexpr
### 🚀 The feature, motivation and pitch If you're using PT2 with a custom triton kernel, and the custom triton kernel takes a tl.constexpr parameter, then we specialize the dynamic shape (https://github.com/pytorch/pytorch/pull/136512). But for export/AOTI, that's not going to work: if the input really is dynamic, then you're going to run into issues at runtime. So we should make sure this is handled well, and error out. ### Alternatives _No response_ ### Additional context _No response_ cc @ezyang @chauhang @penguinwu @bobrenjc93 @avikchaudhuri @gmagogsfm @zhxchen17 @tugsbayasgalan @angelayi @suo @ydwu4 @desertfire @chenyang78 @oulgen @aakhundov
triaged,oncall: pt2,module: dynamic shapes,export-triaged,oncall: export,module: aotinductor,module: user triton
low
Critical
2,557,395,571
flutter
Verify that `dart test` has acceptable semantics for engine CI
_Forked from https://github.com/flutter/engine/pull/55475_ some questions that shouldn't be buried on that PR. I'm not sure what to do with this information (maybe there is an ask for the engine team to improve the infra or file downstream bugs?), but filing this for posterity to divorce it from merging https://github.com/flutter/engine/pull/55475 (which is unrelated to adding `dart test`, `dart test` already runs on engine CI). --- > Does `dart test` do an implicit `pub get`? Yes. > Do we need a flag to skip implicit `pub get`s for `dart test`? Unclear (based on the [discussion so far](https://github.com/flutter/engine/pull/55475)), leaning no. > Does `dart test` write to the source tree at all? Yes, since we consider `.gitignore`'d directories - `.dart_tool/test`. > Do we need to isolate `dart test` in a way where it doesn't write to the source tree? Unclear (based on the [discussion so far](https://github.com/flutter/engine/pull/55475)), leaning no. > Does dart test send telemetry? Should we be explicitly disabling that on CI, somehow? `package:unified_analytics` detect if we're running on CI and disable telemetry automatically > Is it possible to ensure that no other process connects to a compilation server? Yes, with [`--resident-compiler-info-file`](https://github.com/flutter/engine/pull/55475#issuecomment-2383903545)
engine,P2,c: tech-debt,team-engine,triaged-engine
low
Critical
2,557,401,866
ollama
Tool call support in Qwen 2.5 hallucinates with Maybe pattern
According to https://python.useinstructor.com/concepts/maybe/. There is an issue with tool calling in a case like this: ```json { "messages": [ { "role": "system", "content": "Today's date is 2024-09-30. Please consider this when processing the availability information.\nIf you cannot extract the start date, use today.\nThis is the list of employees, with the initials, employee ID, full name, and skills:\n...\n\nDO NOT invent data. DO NOT hallucinate!" }, { "role": "user", "content": "When does our colleague XYZ have two days available for a 2 days appointment?" } ], "model": "qwen2.5:7b-instruct-fp16", "tool_choice": { "type": "function", "function": { "name": "MaybeAvailabilityRequest" } }, "tools": [ { "type": "function", "function": { "name": "MaybeAvailabilityRequest", "description": "Correctly extracted `MaybeAvailabilityRequest` with all the required parameters with correct types", "parameters": { "$defs": { "AvailabilityRequest": { "properties": { "personIds": { "description": "List of person IDs to check availability for", "items": { "type": "integer" }, "title": "Personids", "type": "array" }, "startDate": { "description": "Start date for the availability check", "title": "Startdate", "type": "string" }, "endDate": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "description": "End date for the availability check", "title": "Enddate" }, "numberOfConsecutiveDays": { "description": "Number of consecutive days required", "title": "Numberofconsecutivedays", "type": "integer" } }, "required": [ "personIds", "startDate", "endDate", "numberOfConsecutiveDays" ], "title": "AvailabilityRequest", "type": "object" } }, "properties": { "result": { "anyOf": [ { "$ref": "#/$defs/AvailabilityRequest" }, { "type": "null" } ], "default": null }, "error": { "default": false, "title": "Error", "type": "boolean" }, "message": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "title": "Message" } }, "type": "object", "required": [] } } } ] } ``` it answers with this: ``` json { "id": "chatcmpl-485", "object": "chat.completion", "created": 1727721983, "model": "qwen2.5:7b-instruct-fp16", "system_fingerprint": "fp_ollama", "choices": [ { "index": 0, "message": { "role": "assistant", "content": "", "tool_calls": [ { "id": "call_oe8h5as1", "type": "function", "function": { "name": "MaybeAvailabilityRequest", "arguments": "{\"error\":false,\"message\":null,\"result\":{\"availableDateRange\":[{\"end_date\":\"2024-10-03\",\"start_date\":\"2024-10-01\"}]}}" } } ] }, "finish_reason": "tool_calls" } ], "usage": { "prompt_tokens": 477, "completion_tokens": 158, "total_tokens": 635 } } ``` Which is obviously wrong and not following the JSON schema from the tool call. When I use non function calling and craft the prompt manually, it always gets the answer right. .cc @JianxinMa Thanks!
feature request
low
Critical
2,557,413,238
pytorch
DISABLED test_open_device_registration (__main__.ExtensionBackendTests)
Platforms: linux, rocm, slow This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_open_device_registration&suite=ExtensionBackendTests&limit=100) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/30857493223). Over the past 3 hours, it has been determined flaky in 8 workflow(s) with 16 failures and 8 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_open_device_registration` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. <details><summary>Sample error message</summary> ``` Traceback (most recent call last): File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 2147, in _run_ninja_build subprocess.run( File "/opt/conda/envs/py_3.9/lib/python3.9/subprocess.py", line 528, in run raise CalledProcessError(retcode, process.args, subprocess.CalledProcessError: Command '['ninja', '-v', '-j', '6']' returned non-zero exit status 1. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 1764, in _jit_compile _write_ninja_file_and_build_library( File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 1876, in _write_ninja_file_and_build_library _run_ninja_build( File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 2163, in _run_ninja_build raise RuntimeError(message) from e RuntimeError: Error building extension 'extension_device' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/var/lib/jenkins/workspace/test/inductor/test_extension_backend.py", line 75, in setUpClass cls.module = torch.utils.cpp_extension.load( File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 1357, in load return _jit_compile( File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 1779, in _jit_compile baton.release() File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/utils/file_baton.py", line 50, in release os.remove(self.lock_file_path) FileNotFoundError: [Errno 2] No such file or directory: '/var/lib/jenkins/.cache/torch_extensions/py39_cpu/extension_device/lock' ``` </details> Test file path: `inductor/test_extension_backend.py` cc @clee2000 @ezyang @chauhang @penguinwu @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @ColinPeppler @amjames @desertfire
triaged,module: flaky-tests,skipped,oncall: pt2,module: inductor
low
Critical
2,557,413,242
pytorch
DISABLED test_byte_tensor_assignment (__main__.TestAdvancedIndexing)
Platforms: mac, macos This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_byte_tensor_assignment&suite=TestAdvancedIndexing&limit=100) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/30861081049). Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_byte_tensor_assignment` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. <details><summary>Sample error message</summary> ``` Traceback (most recent call last): File "/Users/ec2-user/runner/_work/pytorch/pytorch/test/test_mps.py", line 11657, in test_byte_tensor_assignment self.assertEqual(len(w), 1) File "/Users/ec2-user/runner/_work/_temp/conda_environment_11107079890/lib/python3.9/site-packages/torch/testing/_internal/common_utils.py", line 3889, in assertEqual raise error_metas.pop()[0].to_error( AssertionError: Scalars are not equal! Expected 1 but got 2. Absolute difference: 1 Relative difference: 1.0 To execute this test, run the following from the base repo dir: python test/test_mps.py TestAdvancedIndexing.test_byte_tensor_assignment This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 ``` </details> Test file path: `test_mps.py` cc @clee2000 @malfet @albanD @kulinseth @DenisVieriu97 @jhavukainen
triaged,module: flaky-tests,module: macos,skipped,module: mps
low
Critical
2,557,416,541
flutter
Flavor API should be derived at the `flutter assemble` level so it is set correctly when run from native tooling, like Android Studio or Xcode
Currently, the app flavor is passed into the `FLUTTER_APP_FLAVOR` based on the `--flavor` flag passed into the Flutter tool, as implemented in https://github.com/flutter/flutter/pull/134179 https://github.com/flutter/flutter/blob/9441f9d48fce1d0b425628731dd6ecab8c8b0826/packages/flutter_tools/lib/src/runner/flutter_command.dart#L1341-L1352 However, that solution doesn't work when run from native tooling, like Android Studio of Xcode, which calls only into `flutter assemble`, so `FLUTTER_APP_FLAVOR` is unset. The proposed community solution at https://github.com/flutter/website/issues/11178 for iOS/macOS requires `DART_DEFINES` to be individually set per Xcode scheme/config, which would be very easy to get out-of-sync, and essentially unnecessary given that `assemble` already has all the information it needs to parse which flavor should be used. It would be better if the flavor parsing logic were pushed down into `assemble`.
platform-ios,tool,platform-mac,P2,team-tool,triaged-tool
low
Major
2,557,492,864
vscode
[macOS] Custom terminal not found on first launch
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.93.1 - OS Version: macOS Sonoma 14.5 We're using a custom terminal (provided by our extension) set as the default terminal in a .code-workspace. When launching VS Code, the wrong terminal is being used for the first terminal instance created. Every subsequent terminal is created using our custom terminal as expected. This can lead to issues when using our extension because we are providing environment variables to the terminal that are missing when the wrong terminal is used. I've tried adjusting the extension activation settings and confirmed that the extension is indeed activating. I am able to reproduce the same issue using the terminal-sample in the VS Code extension samples repo, so it doesn't appear to be specific to our extension. I can only reproduce this issue on macOS - I could not reproduce on Windows 10. Steps to Reproduce: 1. Clone the `terminal-sample` (https://github.com/microsoft/vscode-extension-samples/tree/main/terminal-sample). 2. Build and debug the sample. 3. Close the Terminal view, if open. 4. Create a workspace (`File > Save Workspace As...`) and save it anywhere. 5. In the workspace settings, set the custom terminal as the default: ``` "settings": { "terminal.integrated.defaultProfile.osx": "Terminal API Profile", } ``` 6. Reload the debugged VS Code Window using the Command Palette (`Developer: Reload Window`) 7. Open the workspace settings and note the warning under the "Terminal API Profile": <img width="1512" alt="Screenshot 2024-09-30 at 3 20 42 PM" src="https://github.com/user-attachments/assets/83c69b4e-3b29-45a9-b345-193b7ead1678"> 8. Create a new terminal. This should use the "Terminal API Profile" by default, but it opens `zsh` instead. 9. Create another new terminal. The "Terminal API Profile" is correctly used.
bug,terminal-profiles
low
Critical
2,557,506,046
langchain
KeyError: 'Field "metadata" does not exist in schema'
### Checked other resources - [X] I added a very descriptive title to this issue. - [X] I searched the LangChain documentation with the integrated search. - [X] I used the GitHub search to find a similar question and didn't find it. - [X] I am sure that this is a bug in LangChain rather than my code. - [X] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package). ### Example Code The DataFrame is in the following format. ![data_frame](https://github.com/user-attachments/assets/6b9b9411-ee77-4b4d-b588-15819c9774d3) ```python import lancedb uri = "data/sample-lancedb" db = lancedb.connect(uri,) try: table = db.create_table("movies",data=md_final) except: table = db.open_table("movies") ``` ```python from langchain.vectorstores import LanceDB from langchain.chains import RetrievalQA import os from langchain_huggingface import HuggingFaceEmbeddings embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2") docsearch = LanceDB(connection = db, embedding = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2"), table_name="movies", vector_key="vector", id_key='title', text_key='text', uri=uri, ) ``` The error occurs in the following code. ```python query = "I'm looking for an animated action movie. What could you suggest to me?" docs = docsearch.similarity_search(query) ``` ### Error Message and Stack Trace (if applicable) File c:\Users\jefer\Documents\Livros\LLMs\BUILDING_LLM_POWERED_APPLICATIONS\Building-LLM-Powered-Applications\.conda\Lib\site-packages\langchain_community\vectorstores\lancedb.py:524, in LanceDB.similarity_search(self, query, k, name, filter, fts, **kwargs) [500](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:500) def similarity_search( [501](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:501) self, [502](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:502) query: str, (...) [507](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:507) **kwargs: Any, [508](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:508) ) -> List[Document]: [509](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:509) """Return documents most similar to the query [510](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:510) [511](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:511) Args: (...) [522](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:522) List of documents most similar to the query. [523](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:523) """ --> [524](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:524) res = self.similarity_search_with_score( [525](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:525) query=query, k=k, name=name, filter=filter, fts=fts, score=False, **kwargs [526](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:526) ) [527](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:527) return res File c:\Users\jefer\Documents\Livros\LLMs\BUILDING_LLM_POWERED_APPLICATIONS\Building-LLM-Powered-Applications\.conda\Lib\site-packages\langchain_community\vectorstores\lancedb.py:498, in LanceDB.similarity_search_with_score(self, query, k, filter, **kwargs) [496](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:496) embedding = self._embedding.embed_query(query) [497](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:497) res = self._query(embedding, k, filter=filter, **kwargs) --> [498](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:498) return self.results_to_docs(res, score=score) File c:\Users\jefer\Documents\Livros\LLMs\BUILDING_LLM_POWERED_APPLICATIONS\Building-LLM-Powered-Applications\.conda\Lib\site-packages\langchain_community\vectorstores\lancedb.py:155, in LanceDB.results_to_docs(self, results, score) [152](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:152) score_col = None [154](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:154) if score_col is None or not score: --> [155](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:155) return [ [156](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:156) Document( [157](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:157) page_content=results[self._text_key][idx].as_py(), [158](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:158) metadata=results["metadata"][idx].as_py(), [159](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:159) ) [160](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:160) for idx in range(len(results)) [161](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:161) ] [162](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:162) elif score_col and score: [163](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:163) return [ [164](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:164) ( [165](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:165) Document( (...) [171](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:171) for idx in range(len(results)) [172](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:172) ] File c:\Users\jefer\Documents\Livros\LLMs\BUILDING_LLM_POWERED_APPLICATIONS\Building-LLM-Powered-Applications\.conda\Lib\site-packages\langchain_community\vectorstores\lancedb.py:158, in <listcomp>(.0) [152](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:152) score_col = None [154](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:154) if score_col is None or not score: [155](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:155) return [ [156](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:156) Document( [157](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:157) page_content=results[self._text_key][idx].as_py(), --> [158](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:158) metadata=results["metadata"][idx].as_py(), [159](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:159) ) [160](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:160) for idx in range(len(results)) [161](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:161) ] [162](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:162) elif score_col and score: [163](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:163) return [ [164](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:164) ( [165](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:165) Document( (...) [171](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:171) for idx in range(len(results)) [172](file:///C:/Users/jefer/Documents/Livros/LLMs/BUILDING_LLM_POWERED_APPLICATIONS/Building-LLM-Powered-Applications/.conda/Lib/site-packages/langchain_community/vectorstores/lancedb.py:172) ] File c:\Users\jefer\Documents\Livros\LLMs\BUILDING_LLM_POWERED_APPLICATIONS\Building-LLM-Powered-Applications\.conda\Lib\site-packages\pyarrow\table.pxi:1539, in pyarrow.lib._Tabular.__getitem__() File c:\Users\jefer\Documents\Livros\LLMs\BUILDING_LLM_POWERED_APPLICATIONS\Building-LLM-Powered-Applications\.conda\Lib\site-packages\pyarrow\table.pxi:1625, in pyarrow.lib._Tabular.column() File c:\Users\jefer\Documents\Livros\LLMs\BUILDING_LLM_POWERED_APPLICATIONS\Building-LLM-Powered-Applications\.conda\Lib\site-packages\pyarrow\table.pxi:1561, in pyarrow.lib._Tabular._ensure_integer_index() KeyError: 'Field "metadata" does not exist in schema' ### Description **I connected to the LanceDB database and tried to create a table called "movies" using the data from the md_final DataFrame. If the table already existed, I opened it. Then, I initialized an instance of LanceDB with the "sentence-transformers/all-MiniLM-L6-v2" embeddings model, configuring the vector key, ID key, and text key options. I defined a text query and called the similarity_search method to obtain the relevant documents.** ### System Info ```bash System Information ------------------ > OS: Windows > OS Version: 10.0.22631 > Python Version: 3.11.9 | packaged by Anaconda, Inc. | (main, Apr 19 2024, 16:40:41) [MSC v.1916 64 bit (AMD64)] Package Information ------------------- > langchain_core: 0.3.6 > langchain: 0.3.1 > langchain_community: 0.2.17 > langsmith: 0.1.129 > langchain_chroma: 0.1.4 > langchain_groq: 0.2.0 > langchain_huggingface: 0.1.0 > langchain_mistralai: 0.2.0 > langchain_openai: 0.2.1 > langchain_text_splitters: 0.3.0 Optional packages not installed ------------------------------- > langgraph > langserve Other Dependencies ------------------ > aiohttp: 3.10.5 > async-timeout: Installed. No version info available. > chromadb: 0.5.3 > dataclasses-json: 0.6.7 > fastapi: 0.115.0 > groq: 0.11.0 > httpx: 0.27.2 > httpx-sse: 0.4.0 > huggingface-hub: 0.24.7 > jsonpatch: 1.33 > numpy: 1.26.4 > openai: 0.28.0 > orjson: 3.10.7 > packaging: 23.2 > pydantic: 2.9.1 > PyYAML: 6.0.2 > requests: 2.32.3 > sentence-transformers: 3.1.1 > SQLAlchemy: 2.0.35 > tenacity: 8.5.0 > tiktoken: 0.7.0 > tokenizers: 0.19.1 > transformers: 4.44.2 > typing-extensions: 4.12.2 ```
Ɑ: vector store,investigate
low
Critical
2,557,554,098
vscode
API to access the shell type for a given terminal
Currently there is no API available on the terminal that allows extensions to detect the shell type. This is needed for Python where there are shell specific activation scripts.
feature-request,api,terminal,on-testplan,api-proposal
low
Major
2,557,561,078
pytorch
[inductor] requires_stride_order doesn't work with unbacked symints
```py import torch from torch import Tensor torch._dynamo.config.capture_dynamic_output_shape_ops = True @torch.library.custom_op("mylib::foo", mutates_args={}) def foo(x: Tensor) -> Tensor: return x.clone() @foo.register_fake def _(x): u0 = torch.library.get_ctx().new_dynamic_size() u1 = torch.library.get_ctx().new_dynamic_size() u2 = torch.library.get_ctx().new_dynamic_size() return x.new_empty(u0, u1, u2) @torch.library.custom_op("mylib::bar", mutates_args={}) def bar(x: Tensor) -> Tensor: return x.clone() @bar.register_fake def _(x): return torch.empty_like(x) x = torch.randn(2, 3, 4) @torch.compile(fullgraph=True) def f(x): y = foo(x) z = bar(y) return z f(x) print("done") ``` In this case we should be able to infer the order of the sym strides (they are `u1 * u2, u2, 1`). Maybe we just need a sorting algorithm that understand unbacked symints cc @ezyang @chauhang @penguinwu @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @ColinPeppler @amjames @desertfire @bdhirsh
triaged,module: custom-operators,oncall: pt2,module: dynamic shapes,module: inductor,module: pt2-dispatcher
low
Minor
2,557,584,776
deno
Deno test runs tests in workspace twice if test is not in a member
Title isn't quite accurate, it also only happens if you pass a path to deno test. Repro: ```jsonc // deno.jsonc { "workspace": [ "member" ] } ``` ``` ❯ mkdir member; touch member/deno.json ``` ```ts // main_test.ts (in root of workspace) Deno.test(function myCoolTest() { console.log("running test..."); }) ``` ``` ❯ deno test . running 1 test from ./main_test.ts myCoolTest ... ------- output ------- running test... ----- output end ----- myCoolTest ... ok (0ms) running 1 test from ./main_test.ts myCoolTest ... ------- post-test output ------- running test... ----- post-test output end ----- myCoolTest ... ok (0ms) ok | 2 passed | 0 failed (11ms) ``` If you omit the `.` or the test is in `member`, this doesn't occur
bug,testing,workspaces
low
Critical
2,557,591,356
node
`assert.deepEqual()` report difference between promises during `test`
### Version v22.9.0 ### Platform ```text Linux regseblaptop 6.8.0-45-generic #45-Ubuntu SMP PREEMPT_DYNAMIC Fri Aug 30 12:02:04 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux ``` ### Subsystem Ubuntu 24.04.1 LTS ### What steps will reproduce the bug? 1. Create `test.mjs` ```javascript import assert from "node:assert/strict"; import { mock, test } from "node:test"; test("promise", () => { assert.deepEqual(Promise.resolve("foo"), Promise.resolve("foo")); }); test("mock, promise and await", async () => { const fn = mock.fn(() => "foo"); await fn(Promise.resolve("bar")); assert.deepEqual(fn.mock.calls[0].arguments[0], Promise.resolve("bar")); }); ``` 2. `node --test test.mjs` ### How often does it reproduce? Is there a required condition? Always. ### What is the expected behavior? Why is that the expected behavior? The `assert.deepEqual()` doesn't report difference between the two identical promises. ``` ✔ promise (1.830551ms) ✔ mock, promise and await (0.533274ms) ℹ tests 2 ℹ suites 0 ℹ pass 2 ℹ fail 0 ℹ cancelled 0 ℹ skipped 0 ℹ todo 0 ℹ duration_ms 75.905452 ``` ### What do you see instead? The `assert.deepEqual()` find difference between the two identical promises. ``` ✖ promise (5.761508ms) AssertionError [ERR_ASSERTION]: Expected values to be strictly deep-equal: + actual - expected Promise { 'foo', + [Symbol(async_id_symbol)]: 39, - [Symbol(async_id_symbol)]: 40, [Symbol(trigger_async_id_symbol)]: 17 } at TestContext.<anonymous> (file:///home/regseb/dev/tc/mock/test.mjs:5:12) at Test.runInAsyncScope (node:async_hooks:211:14) at Test.run (node:internal/test_runner/test:930:25) at Test.start (node:internal/test_runner/test:829:17) at startSubtestAfterBootstrap (node:internal/test_runner/harness:289:17) { generatedMessage: true, code: 'ERR_ASSERTION', actual: [Promise], expected: [Promise], operator: 'deepStrictEqual' } ✖ mock, promise and await (0.791194ms) AssertionError [ERR_ASSERTION]: Expected values to be strictly deep-equal: + actual - expected Promise { 'bar', + [Symbol(async_id_symbol)]: 52, + [Symbol(trigger_async_id_symbol)]: 20 - [Symbol(async_id_symbol)]: 66, - [Symbol(trigger_async_id_symbol)]: 54 } at TestContext.<anonymous> (file:///home/regseb/dev/tc/mock/test.mjs:11:12) at async Test.run (node:internal/test_runner/test:931:9) at async Test.processPendingSubtests (node:internal/test_runner/test:629:7) { generatedMessage: true, code: 'ERR_ASSERTION', actual: [Promise], expected: [Promise], operator: 'deepStrictEqual' } ℹ tests 2 ℹ suites 0 ℹ pass 0 ℹ fail 2 ℹ cancelled 0 ℹ skipped 0 ℹ todo 0 ℹ duration_ms 82.541815 ✖ failing tests: test at test.mjs:4:1 ✖ promise (5.761508ms) AssertionError [ERR_ASSERTION]: Expected values to be strictly deep-equal: + actual - expected Promise { 'foo', + [Symbol(async_id_symbol)]: 39, - [Symbol(async_id_symbol)]: 40, [Symbol(trigger_async_id_symbol)]: 17 } at TestContext.<anonymous> (file:///home/regseb/dev/tc/mock/test.mjs:5:12) at Test.runInAsyncScope (node:async_hooks:211:14) at Test.run (node:internal/test_runner/test:930:25) at Test.start (node:internal/test_runner/test:829:17) at startSubtestAfterBootstrap (node:internal/test_runner/harness:289:17) { generatedMessage: true, code: 'ERR_ASSERTION', actual: [Promise], expected: [Promise], operator: 'deepStrictEqual' } test at test.mjs:8:1 ✖ mock, promise and await (0.791194ms) AssertionError [ERR_ASSERTION]: Expected values to be strictly deep-equal: + actual - expected Promise { 'bar', + [Symbol(async_id_symbol)]: 52, + [Symbol(trigger_async_id_symbol)]: 20 - [Symbol(async_id_symbol)]: 66, - [Symbol(trigger_async_id_symbol)]: 54 } at TestContext.<anonymous> (file:///home/regseb/dev/tc/mock/test.mjs:11:12) at async Test.run (node:internal/test_runner/test:931:9) at async Test.processPendingSubtests (node:internal/test_runner/test:629:7) { generatedMessage: true, code: 'ERR_ASSERTION', actual: [Promise], expected: [Promise], operator: 'deepStrictEqual' } ``` ### Additional information ~~The problem only occurs in `test`.~~ https://github.com/nodejs/node/issues/55198#issuecomment-2384180276 1. Create `index.mjs` ```javascript import assert from "node:assert/strict"; import { mock } from "node:test"; assert.deepEqual(Promise.resolve("foo"), Promise.resolve("foo")); const fn = mock.fn(() => "foo"); await fn(Promise.resolve("bar")); assert.deepEqual(fn.mock.calls[0].arguments[0], Promise.resolve("bar")); ``` 2. `node index.mjs` 3. The `assert.deepEqual()` doesn't report difference between the two identical promises. :+1:
assert,promises,test_runner
medium
Critical
2,557,600,127
pytorch
Namespace error when using torch::unravel_index to implement a custom function in cuda.
### 🚀 The feature, motivation and pitch Thank you all for the great library. I wanted to please ask wlease where is the unravel_index in located in the cpp source code of torch? since it seems it is not in the namespace ? God bless ### Alternatives _No response_ ### Additional context _No response_ cc @mruberry @rgommers
triaged,module: numpy
low
Critical
2,557,611,784
ui
[bug]: Tooltip have a problem with the index
### Describe the bug ![Screenshot_1](https://github.com/user-attachments/assets/d7e0a0ac-eb47-4a75-9aed-11d58453d2e9) ``` <Tooltip> <TooltipTrigger className="w-full"> <Button variant="ghost" className="flex w-full justify-center gap-2"> <IconComponent className="h-4 w-4" /> </Button> </TooltipTrigger> <TooltipContent side="right"> <p>{text}</p> </TooltipContent> </Tooltip> ``` The problem is that it always stays below, even though I put a higher z-index ### Affected component/components Tooltip ### How to reproduce 1. Clone component 2. Usage component ### Codesandbox/StackBlitz link _No response_ ### Logs _No response_ ### System Info ```bash Windows, Arc and Brave ``` ### Before submitting - [X] I've made research efforts and searched the documentation - [X] I've searched for existing issues
bug
low
Critical
2,557,651,202
PowerToys
Sort Pattern and Replace-With lists by most recently used
### Description of the new feature / enhancement Sort the "Pattern" and "Replace With" drop-down history lists by most recently used, rather than the current sort which seems random. ### Scenario when this would be used? At times I am going back and forth between a small number of patterns/filenames. It'd be nice to have my more recently-used items near the top of the history lists instead of randomly distributed (or even disappearing) from the lists. ### Supporting information N/A
Needs-Triage
low
Minor
2,557,678,038
vscode
`inlineChat.viewInChat` command from notebook will cause pylance to crash
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes/No <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.93.1 - OS Version: Windows 11 Enterprise Version 24H2 (OS Build 26100.1742) Steps to Reproduce: 1. Install Python and Pylance extensions 2. Open a notebook or create a notebook file with the following content, and press `ctrl + i` to open the copilot inline chat ```python # notebook1.ipynb def foo(): print("foo") ``` 3. Ask for a code change, then run inlineChat.viewInChat on the response (ctrl+down)![Image](https://github.com/user-attachments/assets/b966f00a-773a-4307-a262-c964c2babba6) 4. Pylance crashes with bad request ``` 2024-09-30 14:25:21.458 [info] [Error - 2:25:21 PM] Request textDocument/documentSymbol failed. 2024-09-30 14:25:21.458 [info] Message: Request textDocument/documentSymbol failed with message: Debug Failure. Notebook file should not be passed to getWorkspaceForFile: vscode-chat-code-block:/c%3A/Users/stellahuang/source/projects/issue6427/notebook1.ipynb?6ff16a3c-1351-4aff-837f-d29e7cb883a7 Code: -32603 ``` When inspecting the [textDocument/documentSymbol](https://microsoft.github.io/language-server-protocol/specifications/lsp/3.17/specification/#textDocument_documentSymbol) request from the Pylance side, I noticed that the `DocumentSymbolParams.textDocument.uri = 'vscode-chat-code-block:/c%3A/Users/stellahuang/source/projects/issue6427/notebook1.ipynb?3574a4e6-6692-4217-86b1-0ca8750d25f6'`. It looks like the query is for the notebook file itself, not for a specific cell, which seems like a bug. Related bug opened in Pylance: https://github.com/microsoft/pylance-release/issues/6454
under-discussion,notebook,inline-chat
low
Critical
2,557,678,204
flutter
[A11y] Prelaunch report and internal android a11y testing framework can work with Flutter
### Use case Currently the PDC prelaunch report skips Flutter app in all android version. Pre API 34, all warning are muted Post API 34, PDC isn't able to read into Flutter ANI and produce a useful report. there isn't anything we can do about pre 34 because Android doesn't provide an API to look into virtual view, but we should make sure post API 34 the PDC can correctly identify issues in Flutter apps. ### Proposal make sure internal a11y testing framework can run on flutter post API 34
P2,fyi-android,team-accessibility,triaged-accessibility
low
Minor
2,557,734,894
flutter
Support iOS targets in `et run`
`et run` (which builds the engine, and passes `--local-engine` flags to `flutter run`) does not support iOS. Besides finding a champion that regularly builds and debugs iOS engines to help develop (and dogfood) this workflow, we also might have assumptions in `et` that don't hold for iOS builds. For example, [`<RunTarget>.buildConfigFor`](https://github.com/flutter/engine/blob/bd36636b9a83dfdbb7ef4a0c8fa9ad66f04c7625/tools/engine_tool/lib/src/run_utils.dart#L44-L55) makes an assumption about a 1:1 mapping from a target platform to an engine build: ```dart String buildConfigFor(String mode) { switch (targetPlatform) { case 'android-arm64': return 'android_${mode}_arm64'; case 'darwin': return 'host_$mode'; case 'web-javascript': return 'chrome_$mode'; default: throw UnimplementedError('No mapping for $targetPlatform'); } } ``` However, for iOS, there isn't such a 1:1 mapping, as there are separate builds for an iOS simulator or device. `RunTarget` would need to be expanded to receive both the `TargetPlatform` _and_ [`emulator`](https://github.com/flutter/flutter/blob/9441f9d48fce1d0b425628731dd6ecab8c8b0826/packages/flutter_tools/lib/src/device.dart#L899) key.
platform-ios,P3,team-engine,triaged-engine,e: engine-tool
low
Critical