id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1732847625
definitionAndBoundSpan empty on destructured parameter ES6 has added a new syntax called object destructuring, which can be used in variable assignment and function argument list. My request is to add support for auto-jump and finding references for parameters written in destructuring format. Consider this: function test({a, b}) { console.log(a); console.log(b); console.log({a, b}); } test({ a: 1, b: 2, }); Currently if I hold Ctrl and click on the variables in the function body, I can jump back to the variable names in the parameter list. But the reverse is not yet supported, i.e. holding Ctrl and clicking on a or b should show the references to where they are used in the function body. Hope that this functionality can be added so that I can better see where and how many times each destructured variable is used in the function. Our behavior here is consistent with the desugaring form: If you want to change it you can, but I don't consider this a bug.
gharchive/issue
2023-05-24T03:22:39
2025-04-01T06:44:57.693104
{ "authors": [ "RyanCavanaugh", "ytxmobile98" ], "repo": "microsoft/TypeScript", "url": "https://github.com/microsoft/TypeScript/issues/54453", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1797519816
not recognizing path mappings for files with a .jsw extension, even though having set "files.associations": {"*.jsw": "typescript"} Type: Bug opened with Jetbrains IDE, and all types and modules works properly. create tsconfig file that has "references" and "path" property. something like "references": [ { "path": "sub/tsconfig.json" } ] add "sub/tsconfig.json" to import types "compilerOptions": { "typeRoots": ["../"], "lib": ["ES2020"], "types": ["dist/types/backend", "dist/types/node"] } add vscode setting "files.associations": {"*.jsw": "javascript"} create test.js and 'test.jsw" import module and types compare test.js and test.jsw VS Code version: Code 1.79.2 (Universal) (695af097c7bd098fbf017ce3ac85e09bbc5dda06, 2023-06-14T08:58:52.392Z) OS version: Darwin arm64 22.5.0 Modes: System Info Item Value CPUs Apple M2 (8 x 24) GPU Status 2d_canvas: enabledcanvas_oop_rasterization: disabled_offdirect_rendering_display_compositor: disabled_off_okgpu_compositing: enabledmetal: disabled_offmultiple_raster_threads: enabled_onopengl: enabled_onrasterization: enabledraw_draw: disabled_off_okvideo_decode: enabledvideo_encode: enabledvulkan: disabled_offwebgl: enabledwebgl2: enabledwebgpu: enabled Load (avg) 3, 2, 2 Memory (System) 16.00GB (0.11GB free) Process Argv . --crash-reporter-id 9f0ac6eb-66f7-4287-8ae9-03c0d2e7316c Screen Reader no VM 0% Extensions (19) Extension Author (truncated) Version aws-toolkit-vscode ama 1.78.0 dscodegpt Dan 2.1.13 dart-code Dar 3.66.0 vscode-eslint dba 2.4.2 githistory don 0.6.20 vscode-html-css ecm 1.13.1 prettier-vscode esb 9.16.0 copilot Git 1.93.189 copilot-labs Git 0.14.884 elixir-ls Jak 0.15.1 git-graph mhu 1.30.0 vscode-dotnet-runtime ms- 1.6.0 live-server ms- 0.4.8 remote-repositories ms- 0.36.0 LiveServer rit 5.7.9 vscode-icons vsc 12.4.0 volar Vue 1.8.2 vscode-typescript-vue-plugin Vue 1.8.2 html-css-class-completion Zig 1.20.0 TypeScript doesn't read VS Code's file extension settings
gharchive/issue
2023-06-28T01:06:49
2025-04-01T06:44:57.710070
{ "authors": [ "RyanCavanaugh", "jhlee111" ], "repo": "microsoft/TypeScript", "url": "https://github.com/microsoft/TypeScript/issues/54957", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
119019213
Visual Studio: Hyperlink @see to referenced type Hi, In the following // file A module foo { export interface Bar {} } // file B module foofoo { /** Defines options for creating an instance of @see foo.Bar */ export interface BarOptions { } } it will be useful if the text referenced by @see (i.e. foo.Bar) is hyperlinked to interface Bar in file A. +1 @sandersn @Kingwl We've just upgraded to version 4.1 RC and this issue is not fixed for namespaces as outlined above. The issue used the old syntax module for namespace but that shouldn't change anything. Please can you reopen? Thanks Doube check and It seems work. @Kingwl The @see only works when it is on a separate line. For example, this does not work /** Creates an instance of @see foo.bar */ interface foo { } But this works /** * Creates an instance of * @see foo.bar */ interface foo { } I'm not sure whether you see this as a bug, but it's certainly surprising behaviour. I'm using Visual Studio Community 2019 Version 16.7.7 Typescript 4.1 Also # 2 you seem to have something that looks like Code Lens (the 0 references | 0 implementations. How do you enable that? Also # 3 I need to hold down the control key for the hyperlink to appear. Thanks @NoelAbrahams I'm pretty sure that JSDoc tags are only allowed at the beginning of the line. There are also inline tags like foo {@link Bar bar} baz but they require braces. A regular tag in the middle of sentence should just be interpreted as regular text. @phaux thanks for looking that up. I looked up what appears to be the official documentation and there is no clear definition of whether @see should appear on a newline or not — although in the examples they've provided the tag does always appear on a newline. The fault is in the specification, for lack of clarity, and implementers appear to have decided on not supporting inline @see. @NoelAbrahams As you noted, jsdoc.app really only gives examples of one nested tag -- @link. The rest of them are always the first thing on a line. TS mostly supports inline tags, but the implementation is quirky since it's best-effort, not really a committed feature. Specifically, an inline tag works if the line starts with a non-inline tag, or is a line following such a line: /** @param foo @see foo.bar */ or /** @param foo - a long description * also @see foo.bar */ This is all down to the state machine we use for parsing, so it could be improved to also support /** for more info @see foo.bar */ @sandersn Thanks for your PR, but it seems that currently the ts-sever in VS Code doesn't support references unless a comment is at the begin of a block / declaration sentence. I tried such usages: // this works if (foo) { /** satisfy {@link ConditionsIfFoo} */ str = getBar() } // this doesn't work if (foo) { /** satisfy {@link ConditionsIfFoo} */ str = getBar() } // this doesn't work if (foo) { } /** all the below satisfy {@link ConditionsIfFoo} */ else if (1) { } else if (2) { } else { } While highlighting works well for all the 3 cases. @gdh1995 can you open a new issue? That's a missing feature that needs a detailed proposal. Currently jsdoc only works when attached to a declaration of some kind. The exceptions are @typedef and @callback, so @link could work like those. However, all 3 tags have complex scope rules, and we'd have to think about how those rules would need to change for @link. Sorry I didn't know a hint would require so many details. I'm unable to give such a proposal, so let me forget it. I'll update my code to try to make tsserver happy.
gharchive/issue
2015-11-26T10:52:21
2025-04-01T06:44:57.721393
{ "authors": [ "Kingwl", "NoelAbrahams", "gdh1995", "holdfenytolvaj", "phaux", "sandersn" ], "repo": "microsoft/TypeScript", "url": "https://github.com/microsoft/TypeScript/issues/5802", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2274318194
[isolatedDeclarations][5.5] Adding satisfies to a const expression makes it require explicit type annotation 🔎 Search Terms isolated declarations, satisfies, const 🕗 Version & Regression Information This changed in commit or PR https://github.com/microsoft/TypeScript/pull/58201 ⏯ Playground Link https://tsplay.dev/wRbGEw 💻 Code export const foo = { a: 42 } as const satisfies Record<string, number>; 🙁 Actual behavior Variable must have an explicit type annotation with --isolatedDeclarations.ts(9010) 🙂 Expected behavior No errors Additional information about the issue No response My impression was that satisfies may impact the contextual type of an expression and cause its type to change, such that its type may not be syntactically inferable. (Need to dig up that example again, though.) My impression was that satisfies may impact the contextual type of an expression and cause its type to change, such that its type may not be syntactically inferable. (Need to dig up that example again, though.) Yep I've definitely seen cases where satisfies influences the contextual type of an expression, but I wonder if the as const + direct assignment to a variable is a sufficiently well-defined scenario that we can always trivially extract the type. It is not unfortunately: const x = [function() { return "A"}] as const satisfies Array<() => "A"> const x = [function() { return "A"}] as const Playground Link I wonder if the problem boils down to it not being possible to determine whether the const inference of an expression or the satisfies inference or combination of the two leads to the more specific type, then can we defer that decision down the line, so that export const x = [function() { return "A"}] as const satisfies Array<() => "A"> would be emitted as export declare const x: Resolve<readonly [() => string], [() => "A"]> Are there cases where knowing only the const inference of an expression + the satisfies type isn't sufficient to determine the contextually inferred type? Forgive me for pushing on this, as I'm sure folks are eager to treat this as an open and shut case. We have hundreds of instances of this pattern in our code base, the majority in which are cases where satisfies wouldn't affect the contextually inferred type. We adopted the operator early on under a different a different assumption The new satisfies operator lets us validate that the type of an expression matches some type, without changing the resulting type of that expression[0] So while we could migrate back to our old pattern (since duplicating the structure of the expression with an explicit type is the less desirable) of doing a type test like ``tsx function upcast(value: T): void { return value } export const foo = { a: 42 } as const; upcast<Record<string, number>>(foo); asking developers to write unidiomatic code with an abstruse explanation is an outcome I want to push against. [0] https://devblogs.microsoft.com/typescript/announcing-typescript-4-9/#satisfies satisfies invoking contextual typing is a necessary evil imo; you want e.g. [42] satisfies [number] to succeed, but for that you need to contextually type [42] by [number] so that it doesn't widen to an array type before it can be checked. But that in turn means that [42] ultimately gets inferred as a tuple type instead of an array type. There's no mechanism to "reverse" that after-the-fact because it actually changes which type is initially inferred for the expression. There's a general class of "operators that affect the actual type", including but not limited to: satisfies ... ? ... : ... Some forms of destructuring + defualting basically anything with widening / freshness (including Symbol, unfortunately) Isolated declarations is designed to be syntax-only, meaning an external implementation (or TS internally) must always be able to detect when there's a problem or produce an equivalent annotation / initializer. If we do want to allow these things, then dts emit is going to have to gain some sort of syntax or builin type to describe these behaviors, but right now, they don't really exist. If the satisfies operator doesn't change the type of the inferred expression, then it's always safe to rewrite this into two lines: export const foo = { a: 42 } as const; foo satisfies Record<string, number>; If the satisfies operator doesn't change the type of the inferred expression, then it's always safe to rewrite this into two lines: export const foo = { a: 42 } as const; foo satisfies Record<string, number>; While visually will take some getting used to, I suppose that's a good enough solution (thank you for providing one and not just mansplaining how things work).
gharchive/issue
2024-05-01T22:58:14
2025-04-01T06:44:57.735097
{ "authors": [ "MichaelMitchell-at", "RyanCavanaugh", "dragomirtitian", "fatcerberus", "jakebailey" ], "repo": "microsoft/TypeScript", "url": "https://github.com/microsoft/TypeScript/issues/58397", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
548059005
Fix parsing nested parameter types of @callback JSDoc tag Fixes #36101 There was a conditional statement which prohibited parsing nested parameter types when the parent jsdoc tag is a @callback in parser.ts. After thorough consideration, there seems no reason why this should be handled differently than a function() JDSoc. Attached is a test and updated baseline. I was unable to find a way to confirm the transformed declaration AST is correct after applying this fix. Existing JSDoc conformance tests do not test the resulting declaration files eighter. Please suggest more tests and additional testing strategies if possible. Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.:x: yin sign nowYou have signed the CLA already but the status is still pending? Let us recheck it. I notice, I didn't run all the tests properly before submitting this PR. This bugfix broke the test case callbackTagNamespaced. I am on it. All problems fixed. Now it seems that the CI checks fail for no well-explained reason. This is the case for multiple new PRs. I'll put some more research into this problem later in the next week. @yin can you regenerate and commit the baselines after your change? This PR hasn't seen any activity for quite a while, so I'm going to close it to keep the number of open PRs manageable. Feel free to open a fresh PR or continue the discussion here. I am working with my employer to obtain permission to work on this. Please, be patient, we have more serious problems on hands ATM. Keep this open. I tested this change while creating declarations for Apify SDK. That was the reason, why w decided to push this contribution. We had to create @typedef's to workaround, but that clutters our autogenerated Docs with single-purpose classes. Regards / S pozdravom Matej Gagyi On Thu, Apr 2, 2020 at 6:32 PM Andrew Branch notifications@github.com wrote: @andrewbranch commented on this pull request. In tests/baselines/reference/callbackTagNestedParameter.js https://github.com/microsoft/TypeScript/pull/36131#discussion_r402448948 : +type WorksWithPeopleCallback = (person: @param {String} person.name @param {Number} [person.age] +) => void; I think you weren’t seeing this because you hadn’t yet generated declaration files via @declaration: true. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/microsoft/TypeScript/pull/36131#discussion_r402448948, or unsubscribe https://github.com/notifications/unsubscribe-auth/AACRYNLANFHTRVTPVCACXBDRKS4YBANCNFSM4KFHHWGQ . @yin were you able to get permission to work on this? Either way, do you want to keep working on it? Yes, I got permission. Just got tired of debugging a few months ago, so I postponed it. Do you want me to look into this? If you're up for it, we always appreciate contributions. =) But declaration emit is a tricky part of the code to fix. @yin do you still want to work on this right now? If not, I will close it and you can tell me later if you want to restart. just close it for now. It turned to be much more difficult than I anticipated. I'll come back to it maybe next quarter. On Wed, Feb 17, 2021 at 12:07 AM Nathan Shively-Sanders < notifications@github.com> wrote: @yin https://github.com/yin do you still want to work on this right now? If not, I will close it and you can tell me later if you want to restart. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/microsoft/TypeScript/pull/36131#issuecomment-780175882, or unsubscribe https://github.com/notifications/unsubscribe-auth/AACRYNPAOBALOITLNXODXHLS7L3ELANCNFSM4KFHHWGQ .
gharchive/pull-request
2020-01-10T12:42:19
2025-04-01T06:44:57.751573
{ "authors": [ "andrewbranch", "msftclas", "sandersn", "yin" ], "repo": "microsoft/TypeScript", "url": "https://github.com/microsoft/TypeScript/pull/36131", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
685957379
Partially disable inference recursion tracking changes Experiment to see if partially disabling inference recursion tracking changes from recursive conditional types PR resolves the OOM we're seeing in the RWC test suites. @typescript-bot test this @typescript-bot test this @weswigham Hmm, the RWC suites are failing even when I back out the changes that could potentially consume more memory by generating more work. I'm pretty sure there's nothing left in the recursive conditional types PR that could be the cause. Wonder if something else could be? @weswigham Also, here I ran the RWC tests before merging the PR and I'm pretty certain they passed (sadly that build has now gone away). The fact that none of this repros locally of course makes it extra hard to track down. RWC should be fixed now~ No idea how a change merged on the timeline after your change was merged somehow affected your change's RWC run, though.
gharchive/pull-request
2020-08-26T02:33:14
2025-04-01T06:44:57.755103
{ "authors": [ "ahejlsberg", "weswigham" ], "repo": "microsoft/TypeScript", "url": "https://github.com/microsoft/TypeScript/pull/40256", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
798857937
Error message improvements for unions with identical discriminants This PR would fix #40934, but we will be changing the core logic used in #42556. However, I'm opening this because I wanted to point out the sorts of ~improvements~ changes we can see though if we're willing do a little extra work to track multiple object types with identical discriminants though. That logic can possibly be incorporated into #42556. @typescript-bot pack this This experiment is pretty old, so I'm going to close it to reduce the number of open PRs.
gharchive/pull-request
2021-02-02T02:04:38
2025-04-01T06:44:57.756994
{ "authors": [ "DanielRosenwasser", "andrewbranch", "sandersn" ], "repo": "microsoft/TypeScript", "url": "https://github.com/microsoft/TypeScript/pull/42598", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
918305134
Avoid unexpected any Fixes nothing Fixes nothing 🤣 This isn't an any, it's an evolving let. By the time it's read at the bottom of the function, it's been assigned on all CFA branches, and has the expected type. If we were to read it before definite initialization, it'd be an implicit any, and if an incorrect type were written to it, we'd get a type error at the call to createVariableDeclarationList. it's been assigned on all CFA branches. and has the expected type. Okay. It's seems a expected any here.
gharchive/pull-request
2021-06-11T06:52:36
2025-04-01T06:44:57.759262
{ "authors": [ "Kingwl", "RyanCavanaugh", "fatcerberus" ], "repo": "microsoft/TypeScript", "url": "https://github.com/microsoft/TypeScript/pull/44547", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1086059385
feat(27615): fixAddMissingMember should work for type literal Fixes #27615 The original issue (#27615) said "this should work with types defined in jsdoc syntax", but from looking at the code in the codefix, I'm not sure if that could go in the same PR or would be better left separate. @a-tarasyuk @sandersn what do you think? @gabritto Thanks for the review. You are right, all actions in fixAddMissingMember don't support JSDoc. I would divide adding JSDoc support into new issue/PR.
gharchive/pull-request
2021-12-21T17:28:52
2025-04-01T06:44:57.761102
{ "authors": [ "a-tarasyuk", "gabritto" ], "repo": "microsoft/TypeScript", "url": "https://github.com/microsoft/TypeScript/pull/47212", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1345026794
Replace eslint rulesdir with eslint-plugin-local-rules, convert eslint rules to JS --rulesdir is deprecated in favor of plugins. Although there's no official support for a local plugin, eslint-plugin-local-rules can do what we need. This, combined with converting our rules to JS, means that we don't need any configuration in VS Code or to run the CLI, nor do we need a build step to run after clone or branch changes (leaving the diagnostics as the only remaining build step on clone/branch change). The only gotcha here is that people working on branches between these two configs may get an error in their editor, depending on which config they have in .vscode/settings.json, since we don't check that in. Should should be ready for review; I don't think I have any other changes.
gharchive/pull-request
2022-08-20T02:06:47
2025-04-01T06:44:57.762891
{ "authors": [ "jakebailey" ], "repo": "microsoft/TypeScript", "url": "https://github.com/microsoft/TypeScript/pull/50380", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
691556225
Persist /mnt/wsl mountpoints Is your feature request related to a problem? Please describe. In WSL2, I'd like to be able to share my home directory between distros. So when Ubuntu (my main distro) comes up, it adds a symlink from /mnt/wsl/homeuser/ to /home/user, and when OpenSuse comes up, it also mounts from /mnt/wsl/homeuser/ to /home/user. Describe the solution you'd like I understand /mnt/wsl enables the sharing between distros but unfortunately /mnt/wsl uses tmpfs, so when wsl shutsdown you lose everything that was there. Describe alternatives you've considered I could host the files on my main distro (Ubuntu) and share from Ubuntu to /mnt/wsl/homeuser/, and when OpenSuse comes up, it does the opposite. Unfortunately, for that to occur, Ubuntu must come up first, and must stay up, consuming memory and cpu cycles. Additional context Initial request started at #5177. Also related to #689 and to a possible wsl --mount described in https://github.com/MicrosoftDocs/WSL/pull/824. Is there a workaround for this? I need to be able to persist a folder in /mnt/wsl for kubernetes/docker-for-win. Actual folder can be elsewhere, but I at least need a way to persistently mount bind in /mnt/wsl. Maybe you have an idea for a workaround, @therealkenc ? Use .bashrc and /etc/suders tricks, same as always for what-would-be normally a systemd thing. Analogous firing up (say) sshd et cetera at first launch. /mnt/wsl should really only be used for mounts, both bind-mounts across WSL2 distros (the original use-case per https://github.com/microsoft/WSL/issues/4577#issuecomment-545192865), and wsl --mount, not for actually storing data. Persistent data storage outside a distro falls under #689 (per https://github.com/microsoft/WSL/issues/5177#issuecomment-685211096), but the mount points also need to be recreated on startup, for use-cases like Docker Desktop kubernetes hostPath mounts that reference other distros, see https://github.com/docker/for-win/issues/7023#issuecomment-774891325. This introduces a minor ordering issue, as all the mounts need to be up before any distribution starts (the k8s hostPath PV use-case from https://github.com/microsoft/WSL/issues/5177), but of course the mount has no real way to know which distribution is the source of the mount (it's just a bind-mount between two mnt namespaces); it can't rely on running distribution startup scripts, as it wouldn't know that it needs to start docker-desktop last so that its mounts into other distros function correctly. Another issue here is where do we create said data/home partition? Seems to be we would have to deal with that inside wsl. So maybe like using docker we can create a persistent drive with the wsl command and then copy HOME to it, then tell wsl to mount it to HOME? This is a use-case for #689, per https://github.com/microsoft/WSL/issues/5177#issuecomment-685211096: A way/place to store data inside WSL but outside any particular distro's storage, so that it can be shared with all distros without needing to start a particular distro first. Then you would just need what was originally described in this ticket: When a distro starts, before you access $HOME, you replace with with a symlink/bind-mount to wherever that block device is mounted. (Which could be /mnt/wsl, but that will depend on the way #689 is addressed, and if necessary any follow-on work to auto-mount such block devices). if you just want to share small files such as .bashrc and other config files just store them on /mnt/c/. I use /mnt/c/devtools (c:\devtools) for my bash profile special configs and my entire collection of bash scripts. My distros just source /mnt/c/devtools...../.bashrc and this is also git repository I sync across multiple machines all using multiple wsl distros. All my custom scripts reside in NTFS and are accessible by all distros and git bash and PowerShell simply because I added the bin folder to the windows env path. I have found no performance impact at all, my distros still start and run my scripts instantly. PS: another trick I do is make all distros use git credential helper from windows and windows composer.phar so they never need to ask for git authentication once windows has triggered a browser sign in.
gharchive/issue
2020-09-03T01:36:58
2025-04-01T06:44:57.773869
{ "authors": [ "TBBle", "b-hayes", "giggio", "stefanloerwald", "therealkenc" ], "repo": "microsoft/WSL", "url": "https://github.com/microsoft/WSL/issues/5851", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
980398173
Emojis made of multiple code points break tmux status bar rendering Windows Build Number Microsoft Windows [Version 10.0.19043.1165] WSL Version [X] WSL 2 [ ] WSL 1 Kernel Version 5.10.16.3-microsoft-standard-WSL2 Distro Version Ubuntu 20.04 Other Software tmux 2.6 or 3.0a Windows Terminal 1.9.1942.0 or Ubuntu Windows Store app 2004.2021.222.0 Repro Steps In Ubuntu command line: tmux new-session -n 🕊️ Expected Behavior As on native Ubuntu: tmux status bar has only one line; clicking a window's name in the status bar switches to it Actual Behavior tmux status bar wraps and spills onto a second line; clicking on a window's name in the status bar does nothing Diagnostic Logs No response @duhowett this looks like a console issue. How can we diagnose it ? FYI I also posted this in the Terminal GH repo and apparently this is a known issue https://github.com/microsoft/terminal/issues/11053 Is there a workaround for this in the meantime (other than just not using those characters)?
gharchive/issue
2021-08-26T15:49:09
2025-04-01T06:44:57.779037
{ "authors": [ "OneBlue", "armanschwarz", "carlpaten", "lilred" ], "repo": "microsoft/WSL", "url": "https://github.com/microsoft/WSL/issues/7355", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1171191520
Consider backporting PMU support for Alder Lake Is your feature request related to a problem? Please describe. There is no PMU driver for the Intel Alder Lake platform. Hardware event sampling does not work with VTune or perf. vtune -collect hotspots -knob sampling-mode=hw -knob sampling-interval=0.5 /home/trym/source/stud/tdt4186/practical2/build/release/webserver /tmp 8889 12 24 vtune: Error: Unable to perform driverless collection on this platform. vtune: Error: Cannot enable event-based sampling collection: Architectural Performance Monitoring version is 0. Make sure the vPMU feature is enabled in your hypervisor. root@DESKTOP-CMKEO60:~# dmesg | grep -i pmu [ 0.177428] Performance Events: unsupported p6 CPU model 151 no PMU driver, software events only. Describe the solution you'd like Backport the driver or provide an alternate solution. Additional context AFAIK, the only workarounds are to use Hyper-V or dual boot. https://www.phoronix.com/scan.php?page=news_item&px=Linux-5.13-Perf-Alder-Lake https://lore.kernel.org/lkml/20210311114814.GA5829@zn.tnic/T/ cpuid.txt Reportedly a newer kernel alone is not enough: https://github.com/microsoft/WSL/issues/4678#issuecomment-1138625875 Your cpuid report does not list the features either: Architecture Performance Monitoring Features (0xa): version ID = 0x0 (0) number of counters per logical processor = 0x0 (0) bit width of counter = 0x0 (0) length of EBX bit vector = 0x0 (0) core cycle event not available = false instruction retired event not available = false reference cycles event not available = false last-level cache ref event not available = false last-level cache miss event not avail = false branch inst retired event not available = false branch mispred retired event not avail = false fixed counter 0 supported = false ... fixed counter 31 supported = false number of fixed counters = 0x0 (0) bit width of fixed counters = 0x0 (0) anythread deprecation = false @benhillis are you aware of this issue and is there maybe even a fix coming? It's very frustrating not to have PMU support for Alderlake… (Especially since Hyper-V supports it, as @trympet pointed out) I've hit this today again, since I wanted to profile something under Linux. Why is this not being fixed?!? "Plain" Hyper-V already supports PMU's with Alderlake! I hope you can understand that this is really frustrating as a user 😞... This is also an issue for me as well. This is also an issue for me as well +1. I've also asked on Twitter: https://twitter.com/clemenswasser/status/1669265762991714304 Seems like we're just being ghosted 💀, which is extremely disappointing since many require performance counters support and they already work when using Hyper-V... @benhillis @craigloewen-msft ping, are you working on this? I've once again looked into this and this still hasn't been fixed. In the old issue, I noticed this comment: https://github.com/microsoft/WSL/issues/4678#issuecomment-1142331647 Which seems to have documented the root of the issue pretty well. The problem is that the WSL VM hasn't activated Perfmon (the arch_perfmon feature is missing), which seems to be a hard requirement for newer Intel CPUs for performance counters to work on Linux. Instructions for enabling Perfmon are in the Hyper-V documentation. I could validate this by running the following command: $ cpuid | grep 'performance monitor' performance monitor support available = false performance monitor support available = false [...] Sadly, we can't just call Set-VMProcessor MyVMName -Perfmon @("ipt", "pmu", "lbr", "pebs") on the WSL VM, as it seems to be hidden. I only managed to list the vm by running hcsdiag list, but it seems to absent for all hyper-v commands. @benhillis @craigloewen-msft Since we now know what is missing, could you please activate all Perfmon features for the WSL VMs so that perf and other software which use performance counters now finally work on newer CPUs?
gharchive/issue
2022-03-16T15:33:26
2025-04-01T06:44:57.788555
{ "authors": [ "Trass3r", "clemenswasser", "samlihaha", "trympet", "tyler274" ], "repo": "microsoft/WSL", "url": "https://github.com/microsoft/WSL/issues/8155", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1560452756
ping works in wsl2 ubuntu, but wget, apt doesn't. DNS does work. Version 10.0.22621.1194 WSL Version [X] WSL 2 [ ] WSL 1 Kernel Version 5.15.79.1 Distro Version Ubunto 22.04 Other Software No response Repro Steps wget https://www.microsoft.com no response, but does translate to ip address Expected Behavior a result Actual Behavior hang Diagnostic Logs No response traceroute msn.com works , wget http://msn.com just hangs. so what can effect web access but not traceroute ? I have a problem that seems similar to this. Connectivity starts out fine but after a while things like apt, curl, get stop working, just hang. Problem is cleared for a while by shutting down WSL, logging out and back into Windows, then relaunching WSL. I do not know what triggers the internet connectivity to stop working. Like OP, I see name resolution and pings working. Iam facing the same problem.Connectivity starts out fine but after a while things like apt, curl, get stop working. Same here. Pings works but all conections to any repo were dead. Same here as well. ping, traceroute work just fine but wget, curl, apt don't. Could you please follow the steps below and attach the diagnostic logs? https://github.com/microsoft/WSL/blob/master/CONTRIBUTING.md#collect-wsl-logs-for-networking-issues Doing these solved my problem: Open Hyper-V Manager as administrator Select your pc, open Virtual Switch Manager Select WSL Set to external network Select the network card the traffic runs through Then login to wsl2 terminal and configure an IP address. E.g. sudo ip addr flush dev eth0 sudo dhclient eth0 This is where i found it: https://stackoverflow.com/a/62438375/10853017
gharchive/issue
2023-01-27T21:28:20
2025-04-01T06:44:57.796210
{ "authors": [ "AnQueth", "RustTurakulov", "abdullah-bin-hasan", "chanpreetdhanjal", "daueee", "kamorrissey" ], "repo": "microsoft/WSL", "url": "https://github.com/microsoft/WSL/issues/9550", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1562033979
WSL2: Cisco anyconnect not connecting after running wsl --update Hi, I have been using wsl2 for several months now on windows 11 and my guest OS is Ubuntu (Linux 5.15.79.1-microsoft-standard-WSL2) I recently got a prompt when starting Ubuntu console about WSL now being available on Microsoft Store and the prompt indicated I can simply run wsl --update to update to the new WSL. So all I did was to run this command. After performing this update, I can no longer connect to VPN using Cisco Anyconnect. I did not change anything else (like updating Ubuntu or Cisco Anyconnect client) Now I am getting this error: VPN establishment capability for a remote user is disabled. A VPN connection will not be established I have verified that I can connect to same VPN portal using both Windows and other stand alone Linux laptop, and even another Ubuntu running using Virtual box, So the VPN server is working fine for other Linux and windows clients. WSL version: 1.0.3.0 Kernel version: 5.15.79.1 WSLg version: 1.0.47 MSRDC version: 1.2.3575 Direct3D version: 1.606.4 anyconnect-linux64-4.10.06079 Windows version: 10.0.22621.1105 I have a similar issue. After I did the wsl --update prompt to use the Store version, my VS Code no longer see any distro dans doesn't work in WSL mode anymore. And today after I did a Docker Desktop update to the latest version, it refused to start so I had to competely reset it; it works on Windows, but on the WSL Integration page, it also says that I don't have any WSL 2 distro installed. (which is not true, I have Ubuntu 22.04 installed and it's working fine). So after the wsl --update, there are two different programs which say that I don't have any WSL 2 distro installed. @frivard-coveo, What version of wsl are you running wsl --version? What version of wsl are you running wsl --version? ❯ wsl --version WSL version: 1.0.3.0 Kernel version: 5.15.79.1 WSLg version: 1.0.47 MSRDC version: 1.2.3575 Direct3D version: 1.606.4 DXCore version: 10.0.25131.1002-220531-1700.rs-onecore-base2-hyp Windows version: 10.0.22621.1105 Same issue here on WSL 1.1.3.0 Just experienced the exact same issue because I set up a new laptop where I installed WSL via the store. However, this only occurs when I enable systemd in the wsl.conf file. Whenever I disable it and start a few services manually it connects fine again. @MirChamb3r I tried this too by setting the systemd config in /etc/wsl.conf to false. Unfortunately, this did not fix it for me. Is there anything else you changed?
gharchive/issue
2023-01-30T09:15:46
2025-04-01T06:44:57.803054
{ "authors": [ "DarkVen0m", "davidroth", "frivard-coveo", "kodergeek", "pmartincic" ], "repo": "microsoft/WSL", "url": "https://github.com/microsoft/WSL/issues/9561", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
845299700
The desktop application has an authentication step which opens up new browser and has a pop up attached to it which has a checkbox. The desktop application has an authentication step which opens up new chrome browser and has a pop up attached to it which has a checkbox. I need to select the checkbox and click on the button to move ahead. How can this be achieved? can you please elaborate? The elements are not inspect able as the right click is disabled Something like this. https://intellitect.com/selenium-chrome-csharp/ We add a static ChromeDriverService in ou assemblyInitalization clean up the chrome driver instances. Also, remembering to do chromedriver.Quit() in test cleanup prevents weird errors in downstream tests. Hi @sahaiswat I can think of three ways to handle this case but all involve using desktop session Attach opened browser using desktop session and then find root element by '//*' and calculator 'OK' button offsets by using paint application. Use move_to_element_with_offset method to click on detected element with coordinates. Create a desktop session and perform the steps in #1 without attaching it to the application. (It will be slow) Create a desktop session and find element using an image. You might have to create your own implementation as winappdriver natively doesn't support this feature.
gharchive/issue
2021-03-30T21:28:44
2025-04-01T06:44:57.811521
{ "authors": [ "liljohnak", "sahaiswat", "shoaibmansoor" ], "repo": "microsoft/WinAppDriver", "url": "https://github.com/microsoft/WinAppDriver/issues/1488", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1456498012
SystemEvents.UserPreferenceChanged cannot be triggered Describe the bug I tried to attach an event handler to SystemEvents.UserPreferenceChanged. However, it seems that the event is not triggered for the WinUI 3 desktop app using the Windows App SDK. Steps to reproduce the bug You may just create a simple project with a main window and a main page on it. For the page_OnLoaded event, add: using Microsoft.Win32; # OnLoaded SystemEvents.UserPreferenceChanged += SystemEventsOnUserPreferenceChanged; private void SystemEventsOnUserPreferenceChanged(object sender, UserPreferenceChangedEventArgs e) { throw new NotImplementedException(); } # OnUnloaded SystemEvents.UserPreferenceChanged -= SystemEventsOnUserPreferenceChanged; Expected behavior No response Screenshots No response NuGet package version Windows App SDK 1.2.1: 1.2.221116.1 Packaging type Packaged (MSIX) Windows version Windows 11 version 22H2 (22621, 2022 Update) IDE Visual Studio 2022 Additional context No response A workaround is to handle WM_SETTINGCHANGE (tested on Windows 10 21H1) A workaround is to handle WM_SETTINGCHANGE (tested on Windows 10 21H1) Could you show me your code? With SetWindowSubclass, like in some samples I posted (MainWindow.xaml.cs for declarations) on main window handle hWnd : hWnd = WinRT.Interop.WindowNative.GetWindowHandle(this); SubClassDelegate = new SUBCLASSPROC(WindowSubClass); bool bReturn = SetWindowSubclass(hWnd, SubClassDelegate, 0, 0); with (I tested by changing System colors and I get : "Settings = ImmersiveColorSet") : public const int WM_WININICHANGE = 0x001A; public const int WM_SETTINGCHANGE = WM_WININICHANGE; private int WindowSubClass(IntPtr hWnd, uint uMsg, IntPtr wParam, IntPtr lParam, IntPtr uIdSubclass, uint dwRefData) { switch (uMsg) { case WM_SETTINGCHANGE: { // Console.Beep(6000, 10); string sText = Marshal.PtrToStringUni(lParam); System.Diagnostics.Debug.WriteLine(string.Format("Settings = {0}", sText)); } break; } return DefSubclassProc(hWnd, uMsg, wParam, lParam); } The events of SystemEvents class cannot be raised because when the SystemEventInvokeInfo object captures the SynchronizationContext of the current thread, the SynchronizationContext.Send is called when RaiseEvent, and DispatcherQueueSynchronizationContext has not implemented the Send method. Solution: SystemEvents.InvokeOnEventsThread(() => { // No need to set, the default is null //SynchronizationContext.SetSynchronizationContext(null); SystemEvents.DisplaySettingsChanged += (s, a) => { Debug.WriteLine("DisplaySettingsChanged"); }; });
gharchive/issue
2022-11-19T14:10:38
2025-04-01T06:44:57.819477
{ "authors": [ "ArvinZJC", "castorix", "cnbluefire" ], "repo": "microsoft/WindowsAppSDK", "url": "https://github.com/microsoft/WindowsAppSDK/issues/3158", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
978446229
remove MRTCore .net sdk version Remove the specific .net sdk version requirement for MRTCore. 5.0.302+ is probably installed for most of the people (if not all). It would be annoying you have to keep a version of 5.0.302 to build MRTCore locally, and I don't have enough space on C drive :). Build pipeline will do a version check based on .\build\versions.props. /azp run
gharchive/pull-request
2021-08-24T20:11:56
2025-04-01T06:44:57.821288
{ "authors": [ "huichen123" ], "repo": "microsoft/WindowsAppSDK", "url": "https://github.com/microsoft/WindowsAppSDK/pull/1307", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
852973066
Questions about model evaluation I have two questions regarding model evaluation: Why is the frame to be evaluated (x) multiplied by 100 during call to modelwork function? isn't it affecting data normalization? Is it supposed the data must be normalized before the call to sr_cnn_eval? thank you in advance. I realized that gen_set class makes the same magnitude scaling to the data during training.
gharchive/issue
2021-04-08T01:39:24
2025-04-01T06:44:57.832999
{ "authors": [ "jdariasl" ], "repo": "microsoft/anomalydetector", "url": "https://github.com/microsoft/anomalydetector/issues/34", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
615770596
Azure Devops Pipeline error - 'Unknown argument --async' Adding the command --async works fine locally allowing me to add --async true to stop listening for results once all the files have been passed to appcenter. Adding the same command to the Azure pipeline i get this error - Unknown argument --async - task: AppCenterTest@1 inputs: appFile: '$(Build.ArtifactStagingDirectory)\AndroidBuild\App.Android.apk' artifactsDirectory: '$(System.ArtifactsDirectory)\AppCenterTest' frameworkOption: 'uitest' uiTestBuildDirectory: '$(Build.ArtifactStagingDirectory)\TestStuff' uiTestToolsDirectory: '$(Build.ArtifactStagingDirectory)\TestStuff' prepareOptions: '--include-category $(TestInclude) --exclude-category $(TestExclude) --async $(TestAsync)' credentialsOption: 'serviceEndpoint' serverEndpoint: '*****' appSlug: '*******' devices: '$(Devices)' series: '$(TestSeries)' localeOption: 'en_GB' This is needed as when all the UI tests run the duration is longer than 60m and the hosted pipeline timesout after 60m which shows it as a failure, it also means and entire pipeline is busy for 60m for no reason. Wouldn't it be possible to use Do not wait for test result instead? Closing as stale.
gharchive/issue
2020-05-11T10:23:07
2025-04-01T06:44:57.839483
{ "authors": [ "DmitriyKirakosyan", "IeuanWalker", "Oddj0b" ], "repo": "microsoft/appcenter", "url": "https://github.com/microsoft/appcenter/issues/1852", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
479091881
JUnit 5 for Android Describe the solution you'd like Please add support for JUnit 5 for Android. Describe alternatives you've considered None. I have JUnit 5 tests, and they cannot be run on AppCenter. Thanks @humblehacker for the request! @Oddj0b would you mind taking a look please?
gharchive/issue
2019-08-09T17:28:40
2025-04-01T06:44:57.841056
{ "authors": [ "amchew", "humblehacker" ], "repo": "microsoft/appcenter", "url": "https://github.com/microsoft/appcenter/issues/859", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
2565109036
Add support for context caching What feature would you like to be added? Certain APIs have support for context/prompt/prefix caching, notably Gemini and Claude as well as any local LLM. OpenAI also started supporting this not long ago. This enables calls that condition on the same prefix (chat history) to be faster as they cache the context. However, I am not sure if we can have caching infrastructure be independent of the API provider or rather support caching for each provider. Why is this needed? Speed up inference time @husseinmozannar Moved your issue here.
gharchive/issue
2024-10-03T22:38:20
2025-04-01T06:44:57.842995
{ "authors": [ "ekzhu" ], "repo": "microsoft/autogen", "url": "https://github.com/microsoft/autogen/issues/3636", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
519072735
Update ubuntu-latest to Zulu 13 Azul Zulu OpenJDK 13 has been available since September, but tonight I learned that it still isn't pre-installed in the Microsoft-hosted ubuntu-latest pool (at least not with a path that follows the pattern of previous versions): /home/vsts/work/_temp/bf6bce63-30e5-48f2-abfa-2e76f8006e04.sh: line 1: /usr/lib/jvm/zulu-13-azure-amd64/bin/java: No such file or directory https://github.com/Pr0methean/BetterRandom/blob/master/azure.yml partially works around this issue, since it uses AdoptOpenJDK's JDK13 build with OpenJ9 rather than another JDK12 build in the matrix. (OpenJ9 isn't available through Zulu anyway, so the jobs download it rather than using pre-installed copies.) But this would probably be much harder in a build system that didn't respect the JAVA_HOME environment variable consistently as Maven does. @Pr0methean - This issue is with the image of the host running the agent, not with the agent code itself. Can you open this issue on this project https://github.com/microsoft/azure-pipelines-image-generation ? Unfortunately, I do not have a way of transferring it for you.
gharchive/issue
2019-11-07T06:40:21
2025-04-01T06:44:57.848725
{ "authors": [ "Pr0methean", "jtpetty" ], "repo": "microsoft/azure-pipelines-agent", "url": "https://github.com/microsoft/azure-pipelines-agent/issues/2582", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
521822522
Cache persist doesn't work on server 2012 R2 or 2016 (I assume) because it doesn't inclued tar Having issue with Tasks? Since this is the embedded cache task I believe this belongs here. Agent Version and Platform Version of your agent? 2.160 on prem OS of the machine running the agent? Windows 2012 R2 Azure DevOps Type and Version dev.azure.com https://dev.azure.com/gmollc What's not working? CacheBeta@1 isn't saving the cache because tar isn't included in 2012 R2. Agent and Worker's Diagnostic Logs 2019-11-12T21:42:35.5033236Z ##[debug]Starting 'tar' with arguments '-h -v -cf "4092ddf38c57400cbae3ca85f9ae5178_archive.tar" -C "D:\SharedBuild2\1\s\.pipeline-cache\nuget" .'... 2019-11-12T21:42:36.0435164Z Information, ApplicationInsightsTelemetrySender correlated 1 events with X-TFS-Session 8a46dbca-e561-4dd7-ab21-92c8c86e4de1 2019-11-12T21:42:36.0512117Z ##[error]The system cannot find the file specified 2019-11-12T21:42:36.0520742Z ##[debug]Processed: ##vso[task.logissue type=error;]The system cannot find the file specified 2019-11-12T21:42:36.0521458Z ##[debug]Processed: ##vso[task.complete result=Failed;] 2019-11-12T21:42:36.1115284Z ##[debug] at System.Diagnostics.Process.StartWithCreateProcess(ProcessStartInfo startInfo) at System.Diagnostics.Process.Start() at Agent.Plugins.PipelineCache.TarUtils.RunProcessAsync(AgentTaskPluginExecutionContext context, ProcessStartInfo processStartInfo, Func`3 additionalTaskToExecuteWhilstRunningProcess, Action actionOnFailure, CancellationToken cancellationToken) at Agent.Plugins.PipelineCache.TarUtils.RunProcessAsync(AgentTaskPluginExecutionContext context, ProcessStartInfo processStartInfo, Func`3 additionalTaskToExecuteWhilstRunningProcess, Action actionOnFailure, CancellationToken cancellationToken) at Agent.Plugins.PipelineCache.TarUtils.ArchiveFilesToTarAsync(AgentTaskPluginExecutionContext context, String inputPath, CancellationToken cancellationToken) at Agent.Plugins.PipelineCache.PipelineCacheServer.GetUploadPathAsync(ContentFormat contentFormat, AgentTaskPluginExecutionContext context, String path, CancellationToken cancellationToken) at Agent.Plugins.PipelineCache.PipelineCacheServer.UploadAsync(AgentTaskPluginExecutionContext context, Fingerprint fingerprint, String path, CancellationToken cancellationToken, ContentFormat contentFormat) at Agent.Plugins.PipelineCache.SavePipelineCacheV0.ProcessCommandInternalAsync(AgentTaskPluginExecutionContext context, Fingerprint fingerprint, Func`1 restoreKeysGenerator, String path, CancellationToken token) at Agent.Plugins.PipelineCache.PipelineCacheTaskPluginBase.RunAsync(AgentTaskPluginExecutionContext context, CancellationToken token) at Agent.Plugins.PipelineCache.SavePipelineCacheV0.RunAsync(AgentTaskPluginExecutionContext context, CancellationToken token) at Agent.PluginHost.Program.Main(String[] args) 2019-11-12T21:42:36.1219290Z ##[section]Finishing: Cache NuGet packages You people are going to think I'm crazy, but I'm not... Here a screenshot of the procmon trace looking for tar.exe: Here is a screenshot of it NEVER looking for 7z.exe from the same trace: The ONLY way I can see that happen is if somehow isWindows is false here: https://github.com/microsoft/azure-pipelines-agent/blob/7fc04a145e5158368e8e4fea7f2d996e8a053389/src/Agent.Plugins/PipelineCache/TarUtils.cs#L190 but I don't see how that is possible!?!?!?! @fadnavistanmay Can you take a look? Just to add more confusion to the mix. I provisioned a new 2019 server and everthing works (same pipeline). I'm going to try a 2012 R2 new. Then I'm going to put a new agent on my non-working 2012 r2. Will update this post as I investigate. Hi @jabbera - Glad you were able to make it work on the 2019 server. Just a quick note, the error you are getting is in the SaveCache task, which always use tar; the code snippet you have shared is during the Restore Cache task, where we look for 7z and use that if present. No I’m having issue persisting the cache. It’s a miss so there isn’t a download. This fails on a new 2012 R2 server as well. @jabbera - did you explicitly installed tar on those machines, and it still fails?? You are using self-hosted agents, yes? For now, you can fall back to old behavior by setting the variable AZP_CACHING_CONTENT_FORMAT to FileSs. @fadnavistanmay where from? Isn't that a 2019 only feature? (Unless you are talking about the mingw tar?) Additionally why would you have the 7 zip fallback for downloading caches but not uploading them? It doesn't make much sense to me. For whatever reason that doesn't work either. Here are snippits from my log: [2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'ALLUSERSPROFILE': 'C:\ProgramData' [2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'APPDATA': 'C:\Users\<SNIP>\AppData\Roaming' [2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'AZP_CACHING_CONTENT_FORMAT': 'Files' [2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'CommonProgramFiles': 'C:\Program Files\Common Files' [2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'CommonProgramFiles(x86)': 'C:\Program Files (x86)\Common Files' [2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'CommonProgramW6432': 'C:\Program Files\Common Files' 2019-11-13T20:38:27.5300788Z ##[debug]Processed: ##vso[telemetry.publish area=AzurePipelinesAgent;feature=PipelineCache]{"FileCount":"0","PlanId":"b8fde65d-718b-4f2c-8a48-b3ff33b7809f","JobId":"0ab14b9f-e499-56d5-97b1-fd98b70ea339","TaskInstanceId":"aa203359-c3b8-5e0f-0fa6-2cb7b8dcce2a","CacheResult":"Miss","ActionDurationMs":"892","ActionName":"PipelineCache.RestoreCache","ActionResult":"Success","AttemptNumber":"1","ItemCount":"0","Level":"ThirdParty","CreatedUtcNow":"2019-11-13T20:38:26.5686059Z","SentUtcNow":"2019-11-13T20:38:27.4717776Z","BaseAddress":"https://vsblobprodcus3.vsblob.visualstudio.com/Ad31b77fc-dfbf-4068-8419-a4cc92bbfac6/","X_TFS_Session":"2ebdd9b2-d83c-46a7-836f-1080713ffccc","DeploymentEnvironment":"PRODUCTION","DeploymentEnvironmentIsProduction":"True","VSOAccount":"vsblobprodcus3","OSName":"Microsoft Windows","OSVersion":"6.3.9600","FrameworkDescription":".NET Core ","ProcessName":"Agent.PluginHost","DotNetReleaseDword":"-1","Version":"18.159.29324.0 built by: master (a2f0ba0f2a)","ExceptionCount":"0"} 2019-11-13T20:38:27.5302449Z ##[debug]Starting 'tar' with arguments '-h -v -cf "15fad05c549e42999f7259172fa315b6_archive.tar" -C "D:\SharedBuild1\2\s\.pipeline-cache\nuget" .'... 2019-11-13T20:38:28.1735655Z Information, ApplicationInsightsTelemetrySender correlated 1 events with X-TFS-Session 2ebdd9b2-d83c-46a7-836f-1080713ffccc 2019-11-13T20:38:28.1814906Z ##[error]The system cannot find the file specified 2019-11-13T20:38:28.1824931Z ##[debug]Processed: ##vso[task.logissue type=error;]The system cannot find the file specified 2019-11-13T20:38:28.1825655Z ##[debug]Processed: ##vso[task.complete result=Failed;] 2019-11-13T20:38:28.1988197Z ##[debug] at System.Diagnostics.Process.StartWithCreateProcess(ProcessStartInfo startInfo) at System.Diagnostics.Process.Start() at Agent.Plugins.PipelineCache.TarUtils.RunProcessAsync(AgentTaskPluginExecutionContext context, ProcessStartInfo processStartInfo, Func`3 additionalTaskToExecuteWhilstRunningProcess, Action actionOnFailure, CancellationToken cancellationToken) at Agent.Plugins.PipelineCache.TarUtils.RunProcessAsync(AgentTaskPluginExecutionContext context, ProcessStartInfo processStartInfo, Func`3 additionalTaskToExecuteWhilstRunningProcess, Action actionOnFailure, CancellationToken cancellationToken) at Agent.Plugins.PipelineCache.TarUtils.ArchiveFilesToTarAsync(AgentTaskPluginExecutionContext context, String inputPath, CancellationToken cancellationToken) at Agent.Plugins.PipelineCache.PipelineCacheServer.GetUploadPathAsync(ContentFormat contentFormat, AgentTaskPluginExecutionContext context, String path, CancellationToken cancellationToken) at Agent.Plugins.PipelineCache.PipelineCacheServer.UploadAsync(AgentTaskPluginExecutionContext context, Fingerprint fingerprint, String path, CancellationToken cancellationToken, ContentFormat contentFormat) at Agent.Plugins.PipelineCache.SavePipelineCacheV0.ProcessCommandInternalAsync(AgentTaskPluginExecutionContext context, Fingerprint fingerprint, Func`1 restoreKeysGenerator, String path, CancellationToken token) at Agent.Plugins.PipelineCache.PipelineCacheTaskPluginBase.RunAsync(AgentTaskPluginExecutionContext context, CancellationToken token) at Agent.Plugins.PipelineCache.SavePipelineCacheV0.RunAsync(AgentTaskPluginExecutionContext context, CancellationToken token) at Agent.PluginHost.Program.Main(String[] args) 2019-11-13T20:38:28.2093865Z ##[section]Finishing: Cache NuGet packages Hi @jabbera - The reason we went for tarring for upload was, the performance was almost the same for creating a "tar" - for both the tar and 7z process, but while downloading (untarring) - 7z gives much better performance. Could you please tell us how/where are you setting the environment variable. And if you could give us the redacted logs of the entire build, that would be helpful. Will it be possible for you. to give @johnterickson and me , permissions for https://dev.azure.com/gmollc , to investigate? Thanks. I’m setting the variable as a system variable. I’ll attach the logs in the AM. Feel free to access the org. PS: the reason the tar thing doesn’t really work well is that tar is not an option on 2016 or 2012. IMO: You should fall back on 7 zip, and if that isn’t there fall back on files all automatically unless you don’t plan on supporting those platforms as build platforms. Hi @jabbera - I still can't access the org. Could you please check. Thanks. I thought you meant on the backend. We don’t allow guest accounts in our tenant. I can open a ticket if that gives you more ability to get in. Otherwise I can do a screen share with you. My bad. Let's check the logs first, if nothing significant comes up, we can do a screen share. @fadnavistanmay I was able to reproduce this on my public azure devops. It really is as simple as installing the agent on a 2012 r2 server: https://dev.azure.com/mike-barry/Demo2012Issue/_build/results?buildId=420&view=results @fadnavistanmay any chance to look at this in my public repo? Let's make this error clearer that TAR needs to be on the path I had this issue on Windows Server 2016. I had to install 7zip and gnuwin32 tar manually and add to the PATH. I'm running self hosted agents as a service, so I restarted the service after adding to the PATH - this fixed the problem for me (saving and restoring from the cache). @johnterickson / @jtpetty Will the documentation be updated to include the hard dependency on these applications because at the moment there isn't anything: https://docs.microsoft.com/en-us/azure/devops/pipelines/caching/index?view=azure-devops I agree with @jabbera though, should this task really have these hard dependencies at all or should it install these applications within the tools cache within $(Agent.ToolsDirectory) Hi @garfbradaz , sorry for the delayed in documentation update. We have a PR out, it looks like it haven't merged yet. I'll follow up to get the docs updated. Docs are live: https://docs.microsoft.com/en-us/azure/devops/pipelines/caching/index?view=azure-devops#required-software-on-self-hosted-agent Thanks @fadnavistanmay! @fadnavistanmay Can we add C:\Program Files\Git\usr\bin to the search path for tar on Windows? Any suggestions of how to cache the below mongodb installation would be helpfull script: | wget -qO - https://www.mongodb.org/static/pgp/server-3.6.asc | sudo apt-key add - echo "deb [ arch=amd64,arm64 ] https://repo.mongodb.org/apt/ubuntu xenial/mongodb-org/3.6 multiverse" | sudo tee /etc/apt/sources.list.d/mongodb-org-3.6.list sudo apt-get update sudo apt-get install -y mongodb-org=3.6.16 mongodb-org-server=3.6.16 mongodb-org-shell=3.6.16 mongodb-org-mongos=3.6.16 mongodb-org-tools=3.6.16 sudo service mongod start Any suggestions of how to cache the below mongodb installation would be helpfull script: | wget -qO - https://www.mongodb.org/static/pgp/server-3.6.asc | sudo apt-key add - echo "deb [ arch=amd64,arm64 ] https://repo.mongodb.org/apt/ubuntu xenial/mongodb-org/3.6 multiverse" | sudo tee /etc/apt/sources.list.d/mongodb-org-3.6.list sudo apt-get update sudo apt-get install -y mongodb-org=3.6.16 mongodb-org-server=3.6.16 mongodb-org-shell=3.6.16 mongodb-org-mongos=3.6.16 mongodb-org-tools=3.6.16 sudo service mongod start Hi @johnterickson , @b-barthel already has a PR out for "Can we add C:\Program Files\Git\usr\bin to the search path for tar on Windows?" PR: https://github.com/b-barthel/azure-pipelines-agent/commit/6e0d9b7a64cf48c9e558b3e81f86850916ec3e2a This has been rolled out, closing.
gharchive/issue
2019-11-12T21:47:36
2025-04-01T06:44:57.872355
{ "authors": [ "fadnavistanmay", "garfbradaz", "jabbera", "johnterickson", "shrutyraos", "stephenmichaelf" ], "repo": "microsoft/azure-pipelines-agent", "url": "https://github.com/microsoft/azure-pipelines-agent/issues/2595", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
618039522
Feature: Share cache across pipelines Required Information Type: Feature Enter Task Name: Cache (https://docs.microsoft.com/en-us/azure/devops/pipelines/tasks/utility/cache?view=azure-devops / https://github.com/microsoft/azure-pipelines-tasks/tree/master/Tasks/CacheV2) Environment Azure Pipelines (yml) Can provide account / build / pipeline information privately Agent - Hosted vs2017-win2016 Issue Description I have a number of pipelines that run on the same hosted agent (around 50), for the same repository. All the pipelines use the same templates with lots of pipeline specific variables. I use the following cache task in a build template: - task: Cache@2 condition: and(succeeded(), eq('${{ parameters.enableNugetCaching }}', true)) displayName: NuGet Cache inputs: key: 'nuget | "${{ parameters.solution }}" | "$(Agent.OS)" | ${{ parameters.nugetConfig }},**/packages.config,!**/bin/**,!**/obj/**' restoreKeys: | nuget | "${{ parameters.solution }}" | "$(Agent.OS)" nuget | "${{ parameters.solution }}" path: '${{ parameters.nugetPackagesDirectorySource }}' Caching works perfectly when I run a single pipeline twice - it restores the expected cache, and creates a new cache if one of the matched files changes, as you would expect. However, when I run another pipeline on the same repo / set of code, the cache generates exactly the same key, but does not "match" against the existing cache - it generates a new one. As this cache is around 600mb in size, the "pipeline cache" for the Azure DevOps organisation is now around 30Gb when it should be 600mb. I would expect extra caches to only generate when the source input files change, and the old caches to expire after 30 days (would be handy for this to be customised too, but that's not as important). It also means that all 50 pipelines take around 2 mins extra each, eating into the pipeline minutes. I've attached a screenshot of a compare between the two cache job runs in two separate pipelines. Everything except the X-TFS-Session identifier are exactly the same. Ideally, the cache key should be shared between pipelines that run on the same repo. In addition, if I was able to share this cache between pipelines I would be able to reduce the build by another 2 or 3 mins as I could cache the main solution binaries too - which don't change that often. It's an odd scenario, but one which suits this particular client's build requirements perfectly. I've also posted here: https://developercommunity.visualstudio.com/idea/1030422/share-cache-across-pipelines.html Task logs Cache log comparisons I am considering taking a look into producing a PR for this, let me know if its something you're interested in. Hi @Bidthedog - Thanks for offering to help! Unfortunately, the code changes required here are server-side. This is trickier than it might seem at first because this is an insidious attack vector. Even if I don't have write access to a repo or it's CI build, I could go create a new pipeline that reads from that repo, but puts something "evil" in the cache. The CI build (that I don't have access to) would then read that "evil" cache entry and the build would carry forward my injected "evil" bits. Knowing the above, one way to share artifacts between pipelines/projects is through Packages. In fact, there is a (non-official Azure DevOps but written by MSFT employees) task that acts similarly to Pipeline Caching but uses Universal Packages: https://github.com/Microsoft/azure-pipelines-artifact-caching-tasks If you use it, you'll just have to be very careful about the permissions you have set. OK, thank you for your response. It's not the end of the world, it just means that MUCH more cache space is taken up, and less-frequently-run builds do not take advantage of the cache; some of the pipelines - as you might imagine - are not executed regularly, so it would be handy if they used the cache when they do run. Others run multiple times per day. Tbh, I'd much rather do this with a single pipeline, but my client's software architecture just doesn't make it feasible at present. This pipeline/branch scoping makes cache task extremely inefficient. In my environment, CI builds unable to use cache produced by PR build (different pipelines). Literally, pipelines produce and store all this cached data for nothing. This pipeline/branch scoping makes cache task extremely inefficient. In my environment, CI builds unable to use cache produced by PR build (different pipelines). Literally, pipelines produce and store all this cached data for nothing. I have to agree. I've actually turned package caching off now I've moved to a self-hosted agent, because it's quicker to use the local server's cache than manage a cache per pipeline. One of my clients has 24 main pipelines (and counting) that should mostly use the same cache. Another scenario (although, somehow similar to @gaikovoi's) in which this feature would come handy: We build Python conda environment as part of our PR and CI builds. The operation takes around 5 minutes, but it rarely needs to be redone, as the environment stays unchanged for long time. We can use current cache mechanism efficiently for CI builds, but not for PR build: since the cache is per build & per pipeline, and PR branches are short-lived, we end up building environment most of the times instead of using cached value. Ideally, we would like to use the environment created by CI builds for our PR build We can't really use https://github.com/Microsoft/azure-pipelines-artifact-caching-tasks, as its documentation advices against using it for artifacts produced outside of the repo's directory (this is the case for conda environements) Another scenario (although, somehow similar to @gaikovoi's) in which this feature would come handy: We build Python conda environment as part of our PR and CI builds. The operation takes around 5 minutes, but it rarely needs to be redone, as the environment stays unchanged for long time. We can use current cache mechanism efficiently for CI builds, but not for PR build: since the cache is per build & per pipeline, and PR branches are short-lived, we end up building environment most of the times instead of using cached value. Ideally, we would like to use the environment created by CI builds for our PR build We can't really use https://github.com/Microsoft/azure-pipelines-artifact-caching-tasks, as its documentation advices against using it for artifacts produced outside of the repo's directory (this is the case for conda environements) You can use pipeline artefacts to do this. I have one pipeline that downloads the caches every night and uploads them to Azure Pipeline Artifacts which is free (not Azure Artifacts which costs per GB), then my other pipelines download the pipeline artefact at the beginning of each run. Don't forget if you say "Cache per Pipeline" that is wrong for me, because actually it's "cache per job per pipeline. I can repeatable prove that running one a sequence of 3 jobs where one i and 3 is maven caching under the same key, i don't see items restored on 3 which has been produced (and cached) in 1 even although job 2 in between is taking at least 5 minutes. So what is really the reliable definition for a minimum hit rate if its even on "same" pipeline for a dependent sequence of jobs. This is a very useful feature if can get it in. this is actually not stale. In a monorepo scenario with loads of pipelines using the same dependencies, this leads to dumb amounts of caching thats not shareable across pipelines.. Any solution to this would be awesome @Squixx Azure devops is stale :/ Shouldnt be to hard to allow us to run github tasks on azure devops ;) @Squixx You have a typo , you might have meant gitLAB - or do you want to run into same issue again in x years ;) Just kidding Any news on this? It would be a great feature.
gharchive/issue
2020-05-14T08:40:05
2025-04-01T06:44:57.889140
{ "authors": [ "Bidthedog", "Jon889", "Squixx", "bharathns", "cforce", "gaikovoi", "gsarapura", "johnterickson", "makukl" ], "repo": "microsoft/azure-pipelines-tasks", "url": "https://github.com/microsoft/azure-pipelines-tasks/issues/12901", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1497394093
AzureKeyVault fails with 'endpoints_resolution_error' Question, Bug, or Feature? Type: Bug Enter Task Name: AzureKeyVault Environment Agent - Private, running under Azure DevOps OS: RHEL 8.6 Version: 2.214.2 Issue Description Secrets downloading takes 30+ sec to complete with the following error: "Could not fetch access token for Azure. Status code: endpoints_resolution_error, status message: Error: could not resolve endpoints. Please check network and try again. Detail: ClientConfigurationError: untrusted_authority: The provided authority is not a trusted authority. Please include this authority in the knownAuthorities config parameter." Update - 18 Dec 2022 I've managed to mitigate the error by removing the HTTP proxy and going directly to the internet to download the secrets. So whatever change has been done to the AzureKeyVault to download secrets, it doesn't take into consideration that the secrets download can be done from behind a proxy. And yeah, we set all the proxy environment variables; they're just being ignored. I am seeing this issue as well. When trying to do a recent deployment we received an error during Download Secrets step which uses AzureKeyVaultV2 task. We noticed there was recent version downloaded 1.214.2 which seems to introduce the error. I am running azure devops agent on windows server 2019 with proxy configured. Initialize Job Step: Error during Download Secret step that occurs about 30 seconds later: @jtterry2856 - Are you running on Self Hosted Agent ? I got this error only when i am running on Self Hosted Agent with proxy enabled @RaviChandraMadipadiga yes I am running self hosted agent We're experiencing the same issue since the 14th. Self hosted agent (ver 2.193.0) using a proxy. Could not fetch access token for Azure. Status code: endpoints_resolution_error, status message: Error: could not resolve endpoints. Please check network and try again. Detail: ClientConfigurationError: untrusted_authority: The provided authority is not a trusted authority. Please include this authority in the knownAuthorities config parameter. It seems to be related to the AzureKeyVault task 2.214.2. When we use the previous version (2.211.1), deployments succeed again. However, I am not aware of an elegant way of forcing the agent to use a specific version for this task, since it is automatically selected because of a linked variable group, linked to the keyvault. Anyone? @edohussl - Yes , i tried now changing the version to 1* in Azure Key Vault task and it succeeded but whereas with the Azure key vault variable group it is auto picking the version to latest as 1.214.3 as so the issue is occurring. Do we have any feasibility to change the version of the task to pick for key vault variable group ? @RaviChandraMadipadiga Nope, you can't select the minor plugin version in the release pipeline. You can do that in the YAML pipeline, but YAML pipelines don't support the Deployment Groups. You can use Environments for YAML pipelines, but those don't support sharing of the VMs - as you can't have the same VM in multiple Enviornments... 'Consistency' wasn't apparently a key concern on mind of whoever design this entire thing. Maybe Jenkins isn't that bad after all... If we're blocked for too long I'll be making the case for CloudBees.... We have the same problem and we are not able to deactivate the proxy. Is there a possibility to set max version of a task for the agent setup? We are not able to enable direct internet access on our VMs. However, we did implement a workaround (it's more like hack though) to ensure the agent uses an older version of the AzureKeyVault task. We simply copied over the files from the old version (2.211.1) to the new version (2.214.2) in the 'work_tasks\AzureKeyVault...' folder. I do think this is an issue that has to be fixed in the task or agent. How do we get this under the attention of Microsoft? For me it works again! It seems that the task has been updated to 2.214.3 and now it runs smoothly. @kwasiak - Seems like they have updated the task version now to 214.3 and it's working fine now. Thank you for raising this issue in GitHub by seeing my question in community forum.it really helped all of us . Works for me as well when I tested pipeline today. New version 1.214.3 was downloaded and "download secrets" step completed successfully. Same here, works with 2.214.3. Although, it is strange that I see version 2.214.3 reported and other see 1.214.3 reported. Thanks @ all, the new AzureKeyVault task 1.214.3 works for us too. the new AzureKeyVault task 1.215.0 have the same error again... It is happening again with the newly installed AzureKeyVault task 2.215.0. correct, just got an update from our DevOps team... I really need to start thinking about Jenkins.... I was troubleshooting a network/proxy related issue (407) regarding the AzureKeyVault task, but during the troubleshooting yesterday, I ran into the "endpoints_resolution_error". I came across this issue here and indeed what happened is that they pushed a new version of this task (2.215.0) yesterday, in which it is broken again ("endpoints_resolution_error"). However, a few hours later they pushed 2.215.1 and that one seems to be working fine again. I faced the same problem with delete item in adf task. it may relate... In my case solved by add prerequisite task to set HTTP_PROXY env to the session. steps: powershell: 'Write-Host "##vso[task.setvariable variable=HTTP_PROXY;]http://{proxy_host}:{port}"'
gharchive/issue
2022-12-14T21:06:44
2025-04-01T06:44:57.907240
{ "authors": [ "ChristianFehlinger", "RaviChandraMadipadiga", "RobRybber", "StuhlfauthKlaus", "aomerkii", "armin-pfaeffle", "edohussl", "enginhorzum", "jtterry2856", "kwasiak", "prfj", "vireshov" ], "repo": "microsoft/azure-pipelines-tasks", "url": "https://github.com/microsoft/azure-pipelines-tasks/issues/17485", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2074578917
Azure Arc-enabled data services- update secure string variable Azure Arc-enabled data services- update secure string variable LGTM as well!
gharchive/pull-request
2024-01-10T15:09:16
2025-04-01T06:44:57.917753
{ "authors": [ "alsanch", "lanicolas" ], "repo": "microsoft/azure_arc", "url": "https://github.com/microsoft/azure_arc/pull/2350", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
793880559
Gke postgres scenario Closes #369 Fixed 19.png. I dont see what you mean about 34-38. Looks like same IP to me? Fixed 19.png. I dont see what you mean about 34-38. Looks like same IP to me?
gharchive/pull-request
2021-01-26T02:44:45
2025-04-01T06:44:57.918960
{ "authors": [ "dkirby-ms" ], "repo": "microsoft/azure_arc", "url": "https://github.com/microsoft/azure_arc/pull/373", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
572060909
A11y_SQLAzureDataStudio_Notebook_Create new notebook_AI4D: The control type of editor is set as "Custom" Check out Accessibility Insights! - Identify accessibility bugs before check-in and make bug fixing faster and easier.” GitHubTags:#A11y_SQLAzureDataStudioBenchmark;#A11yMAS;#A11ySev2;#A11yTCS;#SQL Azure Data Studio;#Benchmark;#DesktopApp;#Win32;#AI4D;#MAS1.3.1;#MAS4.1.2;#MAS4.2.1; Environment Details: Application Name: Azure Data Studio Version: 1.16.0-insider OS: Windows_NT x64 10.0.18363 Additional Details:AMAS References:MAS4.1.2, MAS4.2.1, MAS1.3.1 Tools: accessibility insights for desktop Repro Steps: Launch Azure Data Studio. Click File then New Notebook. Launch accessibility insights for desktop Hover over the editors. Actual The control set for editors is set as custom both code and text editor. Expected: The control should be set as editor and announced by a screen reader. User Impact: Screen reader users will not know that have reached the editor. Attachment link for Reference @fsteffi this control encompasses both an editor and the editor's outputs. I'm worried that us saying that the control type as an editor would be confusing, since it's not strictly true. Any thoughts? @chlafreniere FYI I'm going to bring this one to the UX office hours to see what they say - since yeah there isn't a clear indication of what role this should actually be (editor isn't even an actual aria role) Thanks, @Charles-Gagnon !!! #closed; GitHubTags:#SQLADS-Win32-Feb2020;
gharchive/issue
2020-02-27T12:47:38
2025-04-01T06:44:57.925564
{ "authors": [ "Charles-Gagnon", "chlafreniere", "fsteffi", "mstechie", "v-jagansai" ], "repo": "microsoft/azuredatastudio", "url": "https://github.com/microsoft/azuredatastudio/issues/9367", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2158224087
gobject-introspection: upgrade 1.71.0->1.79.1; meson: upgrade 1.2.1->1.3.1; glib: upgrade 2.78.1->2.79.1 Merge Checklist All boxes should be checked before merging the PR (just tick any boxes which don't apply to this PR) [X] The toolchain has been rebuilt successfully (or no changes were made to it) [X] The toolchain/worker package manifests are up-to-date [X] Any updated packages successfully build (or no packages were changed) [X] Packages depending on static components modified in this PR (Golang, *-static subpackages, etc.) have had their Release tag incremented. [X] Package tests (%check section) have been verified with RUN_CHECK=y for existing SPEC files, or added to new SPEC files [X] All package sources are available [X] cgmanifest files are up-to-date and sorted (./cgmanifest.json, ./toolkit/scripts/toolchain/cgmanifest.json, .github/workflows/cgmanifest.json) [X] LICENSE-MAP files are up-to-date (./SPECS/LICENSES-AND-NOTICES/data/licenses.json, ./SPECS/LICENSES-AND-NOTICES/LICENSES-MAP.md, ./SPECS/LICENSES-AND-NOTICES/LICENSE-EXCEPTIONS.PHOTON) [X] All source files have up-to-date hashes in the *.signatures.json files [X] sudo make go-tidy-all and sudo make go-test-coverage pass [X] Documentation has been updated to match any changes to the build system [X] Ready to merge Summary What does the PR accomplish, why was it needed? Change Log gobject-introspection: upgrade 1.71.0->1.79.1; meson: upgrade 1.2.1->1.3.1; glib: upgrade 2.78.1->2.79.1 Skip some meson tests because of their flakiness or we don't support them Add python-packaging to build_official_toolchain_rpms as required for new glib version Does this affect the toolchain? NO Test Methodology Buddy Build Id Full Build Id Looks like toolchain build is failing for me because of python3-packaging-23.2-2.azl3.noarch.rpm https://dev.azure.com/mariner-org/mariner/_build/results?buildId=515772&view=logs&j=db98f19e-da46-5e4d-a4ba-372cf3771a92&t=8831c5ce-9476-5d69-9bf5-bfd7485723f1&l=464717
gharchive/pull-request
2024-02-28T06:35:42
2025-04-01T06:44:57.935588
{ "authors": [ "BettyRain" ], "repo": "microsoft/azurelinux", "url": "https://github.com/microsoft/azurelinux/pull/8140", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2296586900
python-werkzeug: Patch CVE-2024-34069 Merge Checklist All boxes should be checked before merging the PR (just tick any boxes which don't apply to this PR) [ ] The toolchain has been rebuilt successfully (or no changes were made to it) [ ] The toolchain/worker package manifests are up-to-date [ ] Any updated packages successfully build (or no packages were changed) [ ] Packages depending on static components modified in this PR (Golang, *-static subpackages, etc.) have had their Release tag incremented. [ ] Package tests (%check section) have been verified with RUN_CHECK=y for existing SPEC files, or added to new SPEC files [ ] All package sources are available [ ] cgmanifest files are up-to-date and sorted (./cgmanifest.json, ./toolkit/scripts/toolchain/cgmanifest.json, .github/workflows/cgmanifest.json) [ ] LICENSE-MAP files are up-to-date (./SPECS/LICENSES-AND-NOTICES/data/licenses.json, ./SPECS/LICENSES-AND-NOTICES/LICENSES-MAP.md, ./SPECS/LICENSES-AND-NOTICES/LICENSE-EXCEPTIONS.PHOTON) [ ] All source files have up-to-date hashes in the *.signatures.json files [ ] sudo make go-tidy-all and sudo make go-test-coverage pass [ ] Documentation has been updated to match any changes to the build system [ ] If you are adding/removing a .spec file that has multiple-versions supported, please add @microsoft/cbl-mariner-multi-package-reviewers team as reviewer (Eg. golang has 2 versions 1.18, 1.21+) [ ] Ready to merge Summary Patch CVE-2024-34069 Change Log CVE-2024-34069 Does this affect the toolchain? NO Links to CVEs https://nvd.nist.gov/vuln/detail/CVE-2024-34069 Test Methodology Pipeline build id: 570115 The one failing ptests from the PR check wasn't triggered by this change - merging. Auto cherry-pick results: main :white_check_mark: -> https://github.com/microsoft/azurelinux/pull/9118 Auto cherry-pick pipeline run -> https://dev.azure.com/mariner-org/mariner/_build/results?buildId=570548&view=results
gharchive/pull-request
2024-05-14T23:39:05
2025-04-01T06:44:57.945999
{ "authors": [ "CBL-Mariner-Bot", "PawelWMS", "fintelia" ], "repo": "microsoft/azurelinux", "url": "https://github.com/microsoft/azurelinux/pull/9104", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1173208301
Cadl Syntax Review Hold a review of cadl syntax with arch board Gather feedbacks from early adopters on cadl syntax related to #318 Syntax to review: [x] Allow using above the global namespace from https://github.com/Azure/cadl-azure/issues/1320 Closing as dupolicate of https://github.com/Azure/cadl-azure/issues/1518
gharchive/issue
2022-03-18T05:40:40
2025-04-01T06:44:57.958158
{ "authors": [ "markcowl" ], "repo": "microsoft/cadl", "url": "https://github.com/microsoft/cadl/issues/337", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1174647723
[BUG] Incorrect casing of "GitHub" Description GitHub is incorrectly written as "Github". Expected "GitHub" Need to add a permanent linter check for this one and fail the build if the word GitHub case is incorrect and not a part of a link.
gharchive/issue
2022-03-20T18:52:22
2025-04-01T06:44:57.959449
{ "authors": [ "NenoLoje", "shiranr" ], "repo": "microsoft/code-with-engineering-playbook", "url": "https://github.com/microsoft/code-with-engineering-playbook/issues/813", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
2333894276
Investigate --no-compile with PipReport detector A potential improvement for the PipReport detector is to see if --no-compile produces any benefit to performance when used in conjunction with the other existing parameters. Validate that the dependency graph is still the same and that there is a notable improvement in detection time. From limited local testing, this seems to have no effect on report generation times. I'll try and investigate further later on.
gharchive/issue
2024-06-04T16:05:23
2025-04-01T06:44:58.000907
{ "authors": [ "cobya" ], "repo": "microsoft/component-detection", "url": "https://github.com/microsoft/component-detection/issues/1149", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1129481299
clientgen elements can't read their outputs back in without additional parsing I have a clientgen code object for the query query Farms { farms { id } } For a schema with type Farm { id: String! } The output for this query when run by hand includes the data and possibly error fields in the root of the returned payload - however the generated parseResponse(response::Value response) begins parsing for farms in the root document - not data Combined with the JSON parsing of the returned document - it would make serialising the final response on the server and deserialising it back into C++ on the client simpler. ie: client::query::Farms::parseResponse(response::parseJSON(str)) Otherwise you need to do your own parse for data and errors beforehand. Not gonna call this a bug but I would like to know if that is intentional behaviour or not. Doing a quick parse for data or errors is pretty easy - but could be standardised to save some boilerplate. There's a function called parseServiceResponse in GraphQLClient.h which should do the trick. It'll split the response into data and error members, and then you can parse the data independently with the generated function from clientgen. This has come up before, though. Maybe clientgen should output a comment with parseResponse referencing parseServiceResponse to make this clearer.
gharchive/issue
2022-02-10T05:40:44
2025-04-01T06:44:58.004937
{ "authors": [ "ALTinners", "wravery" ], "repo": "microsoft/cppgraphqlgen", "url": "https://github.com/microsoft/cppgraphqlgen/issues/216", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
530932986
Safety of manipulating member variables in async callback? In my code, I fire off a task at startup, which when completed assigns the result to a member variable: // Error checking ignored using namespace winrt::Windows::Foundation; using namespace winrt::Windows::ApplicationModel; using Collections::IVectorView; StartupTask::GetForCurrentPackageAsync().Completed([this](const IAsyncOperation<IVectorView<StartupTask>> &op, AsyncStatus) { m_StartupTask = op.GetResults().GetAt(0); }); There doesn't seem to be actually much details on how safe that is (except if this gets destroyed, which is already greatly covered) because as far as I understand, the result of the operation might very well run on another thread (documentation on apartments isn't very clear about this one) and therefore face concurrency issues. This class currently does not have any thread safety included, and I am not sure if I should add some around m_StartupTask to avoid concurrency issues (or if WinRT consumer types already are thread-safe) In general, implementations should be agile but that's up to the implementation. An agile object may be accessed from any thread or apartment. If the API is not agile then you can use apartment_context or resume_foreground depending on the API. Regardless of apartment context, any concurrency you introduce (such as the example above) needs appropriate locking if you think there may be a race. WinRT does not provide any such thread-safety automatically, except what may be provided implicitly by using something like resume_foreground or thread_pool to queue work onto a single thread.
gharchive/issue
2019-12-02T07:05:25
2025-04-01T06:44:58.007789
{ "authors": [ "kennykerr", "sylveon" ], "repo": "microsoft/cppwinrt", "url": "https://github.com/microsoft/cppwinrt/issues/437", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
740145161
Support execution on a RAM drive See: https://github.com/microsoft/cppwinrt/pull/770 Feel free to reopen if there's interest in exploring this further.
gharchive/issue
2020-11-10T18:31:01
2025-04-01T06:44:58.009049
{ "authors": [ "kennykerr" ], "repo": "microsoft/cppwinrt", "url": "https://github.com/microsoft/cppwinrt/issues/786", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2493821113
Environments sometime don't load and clicking 'Create new environment' causes a crash Dev Home version 0.1701.597.0 Windows build number 10.0.22631.0 Other software N / A Steps to reproduce the bug I haven't got solid steps to repro this every time, but it's happening quite frequently with me visiting the environments page. Don't really know what triggers it. What happens is that sometimes Environments don't load properly, and it shows "No environments found". When I click to "Create new environment", devhome crashes. See the video for better understanding. First time the environments load correctly, but second time they don't. https://github.com/user-attachments/assets/225f190d-91ce-4089-a5c5-d5828b19a718 Expected result Environments screen to always load correctly Actual result Environments not always loaded, and cause a crash when you click "Create new environment" Included System Information Windows: Windows.Desktop v10.0.22631.4037 System Architecture: X64 Included Extensions Information DevHome GitHub Extension version 0.1700.597.0 /logs Hi @mdanish-kh, would you be able to file a feedback hub bug for this issue and provide us with the link? That will help us diagnose this issue faster. You can launch the Feedback hub app by entering the Windows Key + the F key on your keyboard at the same time. During your feedback, there will be an option to recreate the issue in your video. Feedback hub will capture the data needed for further debugging on our side. Thank you! @bbonaby Feedback Hub was not cooperating yesterday, so it took me some time to get back. This should be the issue: https://aka.ms/AAs318i I also used the "Recreate the issue" and was able to successfully repro it. Let me know if you need anything else from me Fixed as a side effect of #3907
gharchive/issue
2024-08-29T08:44:51
2025-04-01T06:44:58.018104
{ "authors": [ "bbonaby", "huzaifa-d", "mdanish-kh" ], "repo": "microsoft/devhome", "url": "https://github.com/microsoft/devhome/issues/3726", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1924728150
bpf_map_get_next_key should return the first key if key not found As per information from the Isovalent folks, the bpf_map_get_next_key should start the iteration of the BPF map at the beginning of the map if the key is not found. Adding the specs as per the official documentation here: https://man7.org/linux/man-pages/man2/bpf.2.html BPF_MAP_GET_NEXT_KEY The BPF_MAP_GET_NEXT_KEY command looks up an element by key in the map referred to by the file descriptor fd and sets the next_key pointer to the key of the next element. int bpf_get_next_key(int fd, const void *key, void *next_key) { union bpf_attr attr = { .map_fd = fd, .key = ptr_to_u64(key), .next_key = ptr_to_u64(next_key), }; return bpf(BPF_MAP_GET_NEXT_KEY, &attr, sizeof(attr)); } If key is found, the operation returns zero and sets the next_key pointer to the key of the next element. If key is not found, the operation returns zero and sets the next_key pointer to the key of the first element. If key is the last element, -1 is returned and [errno](https://man7.org/linux/man-pages/man3/errno.3.html) is set to ENOENT. Other possible [errno](https://man7.org/linux/man-pages/man3/errno.3.html) values are ENOMEM, EFAULT, EPERM, and EINVAL. This method can be used to iterate over all elements in the map.
gharchive/issue
2023-10-03T18:48:41
2025-04-01T06:44:58.020222
{ "authors": [ "Alan-Jowett", "gtrevi" ], "repo": "microsoft/ebpf-for-windows", "url": "https://github.com/microsoft/ebpf-for-windows/issues/2942", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2096463017
Reindex job stuck at "Queued" Describe the bug After submitting a reindex job, the job remains queued indefinitely with no progress. The issue persists after deleting and resubmitting the job. Similar to #2200 ? FHIR Version? R4 Data provider? CosmosDB To Reproduce Steps to reproduce the behavior: Create search parameter, test on a single resource (success) Submit reindex job for new parameter (success) Query reindex job ID - status is stuck at "Queued" Expected behavior Reindex job should run Actual behavior Reindex job remains "Queued" Thanks for reporting the issue @elgib - Are you observing this issue on OSS FHIR service or are you using managed Azure FHIR service (Azure API for FHIR / Azure Health Data Services)? @EXPEkesheth we're using the OSS FHIR service @elgib , have added in our queue for investigation. Can you please share with us search parameter json used for creating custom search parameter. Will inform once we have more details/ questions. #114359 @EXPEkesheth here's an example of one of our custom search parameters. { "resourceType": "SearchParameter", "id": "e46bd3c4-f278-4039-841b-892e931596fe", "meta": { "versionId": "1", "lastUpdated": "2023-11-22T12:54:08.333+00:00" }, "url": "http://1beat.care/fhir/search-parameters#patient-care-unit", "name": "patient-care-unit", "status": "draft", "description": "Reference to Organization resource that represents the care unit currently responsible for the patient.", "code": "care-unit", "base": [ "Patient" ], "type": "reference", "expression": "Patient.extension.where(url = 'http://1beat.care/fhir/extensions#patient-care-unit').value", "target": [ "Organization" ] } Thanks for looking into this. We would appreciate any updates as this is blocking key areas of work for our team. Any short-term recommendations would also be helpful -- for example, should we try rolling back to an earlier version? @EXPEkesheth are there any updates on this issue, or timeframes for a fix? @elgib - As you are using the OSS FHIR service , you would need to explicitly enable Reindex in deployment template (https://github.com/microsoft/fhir-server/blob/main/samples/templates/default-azuredeploy-docker.json#L272). Please ensure this setting is enabled in your instance. Have you used reindex capability in FHIR server OSS before? @EXPEkesheth we have enabled reindex operations and run several reindex jobs successfully in the past. The last successful reindex was late November 2023. We have not made any changes to our setup since then. @elgib Thanks for the information. We will look into the issue and get back incase of any questions. @elgib - we recently made improvements in reindex operation. We hope it helps address your issue, can you please execute reindex and let us know the outcome?
gharchive/issue
2024-01-23T16:22:39
2025-04-01T06:44:58.028623
{ "authors": [ "EXPEkesheth", "elgib" ], "repo": "microsoft/fhir-server", "url": "https://github.com/microsoft/fhir-server/issues/3684", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
802423482
Upgrade dotnet sdk to 5.0.100 Description Upgrade dotnet sdk to 5.0.100 Related issues Addresses [issue #]. Couple of latest PRs are failing due to this (screenshot below) error Testing Describe how this change was tested. FHIR Team Checklist [x] Update the title of the PR to be succinct and less than 50 characters [ ] Add a milestone to the PR for the sprint that it is merged (i.e. add S47) [x] Tag the PR with the type of update: Bug, Dependencies, Enhancement, or New-Feature [ ] Tag the PR with Azure API for FHIR if this will release to the managed service Review squash-merge requirements Semver Change (docs) Patch|Skip|Feature|Breaking (reason) 👏 I think these tests will be fixed by #1650 I think +semver: feature 🚀 Thoughts? Good point!
gharchive/pull-request
2021-02-05T19:37:22
2025-04-01T06:44:58.034672
{ "authors": [ "Ivanidzo4ka", "brendankowitz", "rbans96" ], "repo": "microsoft/fhir-server", "url": "https://github.com/microsoft/fhir-server/pull/1645", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1865148454
Fix FHIR Search query with _count=0 and SqlCustomQueryTest issue Description Fix FHIR Serach query to return result same as _summary=count when passing _count=0. Previously, 400 response was there for _count=0. Fix SqlCustomQueryTest issue. Updated following logic to fix this issue. Adds logic to remove the _count param and replaces it with _summary param in SearchOptionsFactory . Updates GetSummaryTypeOrDefault method in HttpContextExtensions class to return summary type. Related issues Addresses [issue #3230 , #105699]. Testing Adds unit test under SearchOperationsFactoryTests Adds unit test under HttpContextExtensionsTests FHIR Team Checklist Update the title of the PR to be succinct and less than 65 characters Add a milestone to the PR for the sprint that it is merged (i.e. add S47) Tag the PR with the type of update: Bug, Build, Dependencies, Enhancement, New-Feature or Documentation Tag the PR with Open source, Azure API for FHIR (CosmosDB or common code) or Azure Healthcare APIs (SQL or common code) to specify where this change is intended to be released. [ ] CI is green before merge Review squash-merge requirements Semver Change (docs) Patch|Skip|Feature|Breaking (reason) What happens if a user specifies both _count=0 and _summary? This looks like it will take whichever is later in the query parameter array. I'm going to run a test on this. What happens if a user specifies both _count=0 and _summary? This looks like it will take whichever is later in the query parameter array. I'm going to run a test on this. Yeah, these queries give different results: https://localhost:44348/Patient?_summary=text&_count=0 https://localhost:44348/Patient?_count=0&_summary=text The first one gives the number of Patients and nothing else. The second one gives the text summary. The FHIR spec doesn't say what should be done when multiple summary parameters are given. I feel this should be an error. Also, we have a bug. When returning just the count of resources we don't return a self link. But section 3.2.1.7.5 of the FHIR search spec says we should. I know we have an existing issue about the order of search parameters affecting results. @brendankowitz thoughts? Nevermind, I found the section in the spec. It says this is left up to implementations, but they recommend returning an error. Note that with the exception of _include and _revinclude, search result parameters SHOULD only appear once in a search. If such a parameter appears more than once, the behavior is undefined and a server MAY treat the situation as an error @mahajan-xor I don't think this needs to be part of this PR as it is a wider issue in our service. It is a bug though, and we should track it. I'll make an item for it.
gharchive/pull-request
2023-08-24T13:21:25
2025-04-01T06:44:58.045319
{ "authors": [ "LTA-Thinking", "mahajan-xor" ], "repo": "microsoft/fhir-server", "url": "https://github.com/microsoft/fhir-server/pull/3491", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
466663954
Auto-update for packages related to 'Microsoft.CodeAnalysis' Updates package 'Microsoft.CodeAnalysis.FxCopAnalyzers' to version '2.9.3' Updates package 'Microsoft.CodeAnalysis.CSharp.Scripting' to version '3.1.0' 2.9.3 Seems to have a bug that flags CA1062 (Check for null parameter) in a place where it is actually being checked (ServiceDescriptorExtensions : WithMetadata). 2.9.4-beta1.final seems to have it fixed, but it is a preview version. We should revisit this when 2.9.4 is released. Closing as 2.9.4 is released
gharchive/pull-request
2019-07-11T05:01:28
2025-04-01T06:44:58.047688
{ "authors": [ "MicrosoftHealthService", "YazanMSFT", "brendankowitz" ], "repo": "microsoft/fhir-server", "url": "https://github.com/microsoft/fhir-server/pull/573", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2623355936
fix: [Local storage for theme not being set] in [FluentDesignTheme] 🐛 Bug Report Theme is not saved in localstorage when using StorageName 💻 Repro or Code Sample Add to App.razor or main layout: <FluentDesignTheme Mode="DesignThemeModes.Default" OfficeColor="OfficeColor.Teams" StorageName="theme" /> 🤔 Expected Behavior Theme should be saved in localstorage without need to switch it manually for the first time 😯 Current Behavior When you set a theme like this: <FluentDesignTheme Mode="DesignThemeModes.Default" OfficeColor="OfficeColor.Teams" StorageName="theme" /> it's not being added to localstorage untill you switch the theme. Right now, the only way to save the theme to localstorage is to bind values and switch the theme for example using some button, only then the theme is added to localstorage. <FluentDesignTheme @bind-Mode="@Mode" @bind-OfficeColor="@ThemeColor" StorageName="theme" https://github.com/user-attachments/assets/4de13519-002d-4109-a6a5-eba8bc2151e4 💁 Possible Solution Not sure 🔦 Context I want to save the theme in localstorage with values mode: null and primaryColor: Teams, so it will switch according to user settings. Right now sometimes the theme is being set wrong and having the values in localstorage could help. 🌍 Your Environment Windows 11, Edge, FluentUI v4.10.3, .NET 8.0.4 This is how the component works: nothing is saved in the LocalStorage until it is needed and the default values are used. We can't change this behaviour. However, you can use the OnLoaded event to get the theme settings when your application starts. This is how the component works: nothing is saved in the LocalStorage until it is needed and the default values are used. We can't change this behaviour. However, you can use the OnLoaded event to get the theme settings when your application starts. Alright I understand. Thank you 😊
gharchive/issue
2024-10-30T09:08:15
2025-04-01T06:44:58.053741
{ "authors": [ "Nikkoro", "dvoituron" ], "repo": "microsoft/fluentui-blazor", "url": "https://github.com/microsoft/fluentui-blazor/issues/2885", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2224123852
Update from upstream and fix missing set -eux Update upstream submodule and resolve patch conflicts. Upstream is now using a builder stage, so some things are reordered. They have some infrastructure set up to actually share more work as a result (https://github.com/docker-library/golang/commit/a6fd6eceb0cb26da2fceefb4353768c472f84420) but for us, it doesn't have any meaningful impact. Fix a missing set -eux that could have caused errors to be improperly ignored. Add a command to upgrade all distro packages for Azure Linux and Debian. https://github.com/microsoft/go-images/pull/290 Regenerate the Dockerfiles with dockerupdate -f. A few more small fixes, would appreciate a set of eyes taking another look. 😄
gharchive/pull-request
2024-04-03T23:39:15
2025-04-01T06:44:58.441042
{ "authors": [ "dagood" ], "repo": "microsoft/go-images", "url": "https://github.com/microsoft/go-images/pull/294", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2597474974
Grand Overhaul: Refactored Core Structure, Introduced New Features, E… Refactoring of graphrag/query Module Description This pull request introduces a significant refactoring of the graphrag/query module within the GraphRAG project. The primary objectives of this refactoring are: Decoupling the Query Module: Transform the query component into an independent package, fully decoupled from other modules. Enhancing Code Reusability and Modularity: Implement a modular design for the entire lifecycle of the GraphRAG query pipeline, promoting loose coupling and facilitating future maintenance and extension. Improving the Python API: Provide a more user-friendly and convenient Python API, simplifying the creation and management of GraphRAG clients. Eliminating Redundancies: Remove redundant modules and parameters (e.g., the question_gen module), streamlining the codebase. Comprehensive Documentation: Add detailed docstrings and extensive type annotations throughout the codebase, ensuring code reliability and passing mypy checks. Enhanced CLI and GUI Tools: Introduce a more powerful CLI tool with rich parameter combinations and an optional GUI built with PyQt6. Unified Streaming Implementation: Employ a more elegant approach to handle both streaming and non-streaming outputs within a single method. Related Issues N/A Proposed Changes 1. Project Layout The codebase has been reorganized, promoting separation of concerns and ease of navigation. The new structure is as follows: query/ ├── __init__.py # Package initialization ├── __main__.py # CLI entry point ├── _base_client.py # Base client templates ├── _cli/ # CLI layer │ ├── __init__.py │ ├── _api.py # CLI API │ ├── _cli.py # CLI main program │ ├── _qt/ # GUI layer │ │ ├── __init__.py │ │ └── _app.py # GUI main program │ └── _utils.py # CLI utilities ├── _client.py # GraphRAG clients ├── _config.py # Configuration classes ├── _defaults.py # Default constants ├── _search/ # Search layer │ ├── __init__.py │ ├── _context/ # Context module │ │ ├── __init__.py │ │ ├── _builders/ # Context builders │ │ ├── _loaders/ # Context loaders │ │ └── _types.py # Type hints │ ├── _defaults.py # Search layer defaults │ ├── _engine/ # Engine module │ │ ├── __init__.py │ │ ├── _base_engine.py # Base engine template │ │ ├── _global.py # Global search engine │ │ └── _local.py # Local search engine │ ├── _input/ # Input module │ │ ├── __init__.py │ │ ├── _loaders/ # Input loaders │ │ └── _retrieval/ # Input retrieval │ ├── _llm/ # LLM module │ │ ├── __init__.py │ │ ├── _base_llm.py # Base LLM template │ │ ├── _chat.py # Chat LLM │ │ ├── _embedding.py # Text Embedding │ │ └── _types.py # Type hints │ ├── _model/ # Data models │ └── _types/ # Type hints │ ├── __init__.py │ ├── _search.py │ ├── _search_chunk.py │ ├── _search_verbose.py │ └── _search_chunk_verbose.py ├── _utils/ # Utilities │ ├── __init__.py │ ├── _text.py # Text utilities │ └── _utils.py # General utilities ├── _vector_stores/ # Vector storage layer │ ├── __init__.py │ ├── _base_vector_store.py │ └── _lancedb.py ├── _version.py # Version information ├── errors.py # Error types └── types.py # Type hints The query module is now fully decoupled from other modules, making it usable as a standalone package. The code is reorganized to promote modularity, facilitating easier maintenance and potential future extensions. 2. Enhanced Python API 2.1 Initialize Users can easily create a GraphRAGClient instance using configuration file, dictionary, environment variables or configuration object. a) From Configuration File e.g., from graphrag.query import GraphRAGClient config_file = "config.yaml" client = GraphRAGClient.from_config_file(config_file) The configuration file can be in YAML, JSON, or TOML format. Refer to the graphrag.example.yaml file for an example. b) From Configuration Dictionary e.g., from graphrag.query import AsyncGraphRAGClient config = { "chat": { "api_key": "API_KEY", "base_url": "BASE_URL", "model": "MODEL" }, "embedding": { "api_key": "API_KEY", "base_url": "BASE_URL", "model": "MODEL" } } client = AsyncGraphRAGClient.from_config_dict(config) c) From Configuration Object If you prefer to use a configuration object and an optional logger, you can pass them directly to the constructor: import logging from graphrag.query import ( ChatLLMConfig, EmbeddingConfig, GraphRAGClient, GraphRAGConfig, ) logger = logging.getLogger(__name__) config = GraphRAGConfig( chat=ChatLLMConfig(api_key="API_KEY", base_url="BASE_URL", model="MODEL"), embedding=EmbeddingConfig(api_key="API_KEY", base_url="BASE_URL", model="MODEL") ) client = GraphRAGClient(config=config, logger=logger) d) From Environment Variables You can also initialize a client using environment variables: export GRAPHRAG_QUERY__CHAT_LLM__API_KEY=API_KEY export GRAPHRAG_QUERY__CHAT_LLM__MODEL=MODEL export GRAPHRAG_QUERY__EMBEDDING__API_KEY=API_KEY export GRAPHRAG_QUERY__EMBEDDING__MODEL=MODEL Or create .env file in the project root directory: GRAPHRAG_QUERY__CHAT_LLM__API_KEY=API_KEY GRAPHRAG_QUERY__CHAT_LLM__MODEL=MODEL GRAPHRAG_QUERY__EMBEDDING__API_KEY=API_KEY GRAPHRAG_QUERY__EMBEDDING__MODEL=MODEL Then initialize the client: from graphrag.query import GraphRAGClient, GraphRAGConfig config = GraphRAGConfig() client = GraphRAGClient(config=config) 2.2 Chatting with GraphRAG a) Simple Chat You can chat with GraphRAG using the chat method: from graphrag.query import GraphRAGClient client: GraphRAGClient = ... response = client.chat( engine="local", message=[ {"role": "user", "content": "What is the purpose of life?"}, {"role": "assistant", "content": "The purpose of life is to be happy."}, {"role": "user", "content": "What is the meaning of happiness?"} ], ) print(response.choice.message.content) Or, in streaming mode: from graphrag.query import GraphRAGClient client: GraphRAGClient = ... response = client.chat( engine="local", message=[ {"role": "user", "content": "What is the purpose of life?"}, {"role": "assistant", "content": "The purpose of life is to be happy."}, {"role": "user", "content": "What is the meaning of happiness?"} ], stream=True ) for chunk in response: print(chunk.choice.delta.content, end="") client.close() # Close the client c) Using with Statement You can also use the with statement to manage the client's lifecycle: from graphrag.query import GraphRAGClient, GraphRAGConfig config: GraphRAGConfig = ... with GraphRAGClient(config=config) as client: response = client.chat( engine="local", message=[ {"role": "user", "content": "What is the purpose of life?"}, {"role": "assistant", "content": "The purpose of life is to be happy."}, {"role": "user", "content": "What is the meaning of happiness?"} ], stream=True ) for chunk in response: print(chunk.choice.delta.content, end="") d) Verbose Search Results If you want to collect verbose search results, you can set the verbose parameter to True: from graphrag.query import GraphRAGClient client: GraphRAGClient = ... response = client.chat( engine="local", message=[ {"role": "user", "content": "What is the purpose of life?"}, {"role": "assistant", "content": "The purpose of life is to be happy."}, {"role": "user", "content": "What is the meaning of happiness?"} ], verbose=True ) print(response.model_dump()) Or, in streaming mode: from graphrag.query import GraphRAGClient client: GraphRAGClient = ... response = client.chat( engine="local", message=[ {"role": "user", "content": "What is the purpose of life?"}, {"role": "assistant", "content": "The purpose of life is to be happy."}, {"role": "user", "content": "What is the meaning of happiness?"} ], streaming=True, verbose=True ) for chunk in response: print(chunk.model_dump()) e) Async Client AsyncGraphRAGClient provides an asynchronous version of the GraphRAGClient: import asyncio from graphrag.query import AsyncGraphRAGClient, GraphRAGConfig config: GraphRAGConfig = ... async def main(): client = AsyncGraphRAGClient(config=config) response = await client.chat( engine="local", message=[ {"role": "user", "content": "What is the purpose of life?"}, {"role": "assistant", "content": "The purpose of life is to be happy."}, {"role": "user", "content": "What is the meaning of happiness?"} ], streaming=True ) async for chunk in response: print(chunk.choice.delta.content, end="") await client.close() # Or you can use the async context manager asyncio.run(main()) 3. Streamlined CLI and GUI Tools 3.1 CLI Parameters Execute the following command: python -m graphrag.query --help To see the available options: usage: python -m query [-h] [--verbose] [--engine {local,global}] [--stream] --chat-api-key CHAT_API_KEY [--chat-base-url CHAT_BASE_URL] --chat-model CHAT_MODEL --embedding-api-key EMBEDDING_API_KEY [--embedding-base-url EMBEDDING_BASE_URL] --embedding-model EMBEDDING_MODEL --context-dir CONTEXT_DIR [--mode {console,gui}] [--sys-prompt SYS_PROMPT] [-V] GraphRAG Query CLI options: -h, --help show this help message and exit --verbose, -v enable verbose logging (default: False) --engine {local,global}, -e {local,global} engine to use for the query (default: local) --stream, -s enable streaming output (default: False) --chat-api-key CHAT_API_KEY, -k CHAT_API_KEY API key for the Chat API (default: None) --chat-base-url CHAT_BASE_URL, -b CHAT_BASE_URL base URL for the chat API (default: None) --chat-model CHAT_MODEL, -m CHAT_MODEL model to use for the chat API (default: None) --embedding-api-key EMBEDDING_API_KEY, -K EMBEDDING_API_KEY API key for the embedding API (default: None) --embedding-base-url EMBEDDING_BASE_URL, -B EMBEDDING_BASE_URL base URL for the embedding API (default: None) --embedding-model EMBEDDING_MODEL, -M EMBEDDING_MODEL model to use for the embedding API (default: None) --context-dir CONTEXT_DIR, -c CONTEXT_DIR directory containing the context data (default: None) --mode {console,gui}, -o {console,gui} mode to execute the GraphRAG engine (default: console) --sys-prompt SYS_PROMPT, -p SYS_PROMPT system prompt file in TXT format to use for the local engine (default: None) -V, --version show program's version number and exit 3.2 Usage Examples We can get started with the CLI from the corpus used in the GraphRAG official tutorial: curl https://www.gutenberg.org/cache/epub/24022/pg24022.txt -o ./input/pg24022.txt Then running the indexing pipeline. Ommited for brevity. a) Console Mode python -m graphrag.query --engine local \ --chat-api-key API_KEY \ --chat-model MODEL \ --embedding-api-key API_KEY \ --embedding-model MODEL \ --context-dir ./output \ --mode console \ --stream Or, more concisely: python -m graphrag.query -e local \ -k API_KEY \ -m MODEL \ -K API_KEY \ -M MODEL \ -c ./output \ -o console \ -s Here is an example screenshot: b) GUI Mode python -m graphrag.query --engine local \ --chat-api-key API_KEY \ --chat-model MODEL \ --embedding-api-key API_KEY \ --embedding-model MODEL \ --context-dir ./output \ --mode gui Here is an example screenshot: 4. Web API Applied the refactored query module to a web service in the graphrag-server repository, providing an OpenAI-compatible Chat API interface. git clone https://github.com/6ixGODD/graphrag-server.git cd graphrag-server Modify the .env file with the appropriate API keys and models. cp .env.example .env Write a simple Python script to execute the web service: from server import create_app app = create_app() if __name__ == '__main__': import uvicorn uvicorn.run(app, host='127.0.0.1', port=8000) Then you can use the OpenAI SDK to interact with the web service: import openai client = openai.OpenAI( api_key="API_KEY", base_url="http://127.0.0.1:8000/api", ) Detailed documentation and deployment instructions (e.g., using Gunicorn and Docker) will be provided in future updates. Currently, there is no detailed docstring documentation for the web service; this will be added subsequently. Checklist [x] I have tested these changes locally. [x] I have reviewed the code changes. [x] I have updated the documentation (if necessary). [ ] I have added appropriate unit tests (if applicable). Additional Notes As mentioned, this PR involves significant code changes, but I believe it is a positive step forward. With thorough testing, it will provide developers a more stable and modular version of GraphRAG for integration into their applications, leading to greater overall benefits. However, for this PR to be merged, some additional documentation work and test case development may require collaboration with the official team. PLEASE REVIEW THIS! PEOPLE ARE WAITING!!!! PLEASE REVIEW THIS! PEOPLE ARE WAITING!!!! You need to rebase to main.
gharchive/pull-request
2024-10-18T13:05:54
2025-04-01T06:44:58.465674
{ "authors": [ "6ixGODD", "JoedNgangmeni", "knguyen1" ], "repo": "microsoft/graphrag", "url": "https://github.com/microsoft/graphrag/pull/1295", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2124836272
shorthands.flex(1) should produce flex: 1 1 0% instead of flex: 1 1 0px When you set a css property to flex: 1 in any of the major browsers the result isflex: 1 1 0%. But the griffel shorthand shorthands.flex(1) produces flex: 1 1 0px. What's maybe even weirder is that the standard actually defines that flex: <positive-number> is the equivalent of flex: <positive-number> 1 0;. But somehow all browser vendors decided that this is just wrong and produce flex: 1 1 0% as mentioned above. Either way following reality or following the standard griffel produces an incorrect output. Ideally it should follow how the major browser vendors do it to have the same developer experience as specifying flex: 1, because otherwise this can be a real pitfall. https://github.com/w3c/csswg-drafts/issues/5742 @stefan-schweiger shorthands.flex() have been deprecated as CSS shorthands are supported natively, check #531 😉
gharchive/issue
2024-02-08T10:37:03
2025-04-01T06:44:58.470431
{ "authors": [ "layershifter", "stefan-schweiger" ], "repo": "microsoft/griffel", "url": "https://github.com/microsoft/griffel/issues/502", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1025784953
Update the Type field name to PolicyType for SetPolicy The "Type" fieldname was interfering with a internal "Type" field. Changing it to "PolicyType" Signed-off-by: netal netalgupta17@gmail.com @netal You'll need to run go mod vendor and go mod tidy in the /test directory in this repo and re-push to fix the CI. Also, I'm having trouble finding uses/examples of SetPolicySetting. I don't see any in this repo or Kubernetes so not sure how it's used. As of now, there are no customers using Setpolicy. Azure NPM is the first customer using this policy. @netal or @dcantah or @kevpar any ETA on when this PR will be merged and released ? @vakalapa @netal Sorry for the delay, this looks fine to us. Going to squash the commits and check in shortly @dcantah when can we have a release with this version ?
gharchive/pull-request
2021-10-13T23:38:02
2025-04-01T06:44:58.473421
{ "authors": [ "dcantah", "netal", "vakalapa" ], "repo": "microsoft/hcsshim", "url": "https://github.com/microsoft/hcsshim/pull/1194", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
848819012
Incorrect(?) SDK Invalidation: SDK assumes all JWKs must conform to SHC rules Problem SHC IG specifies requirements for public JWK to specify signature validation against (use: sig; alg: ES256, etc). However, the SHC IG does not specify that all public keys hosted a domain.com/.well-known/jwks.json must be a SHC-conforming JWK: it's possible that orgs might want to host other various JWKs there and only the JWK with the kid matching that in the VC must conform. Solution Update to only validate against the JWK matching the kid, not all keys in the key-set https://github.com/microsoft/health-cards-validation-SDK/blob/fb6a527401e7e49d9bd0433c465ab4bac6498c31/src/shcKeyValidator.ts#L174 cc @jmandel to make sure I'm interpreting the SHC IG correctly @christianpaquin assuming we are aligned that this is a problem worth solving, I'm happy to help with a patch here. Yes, @dleve123, JWK sets certainly can have more than one key. The tool currently gives warnings for keys that cannot be used for Health Card issuance. The same key validation code runs when you validate a key set directly or when one is downloaded while validating a health card. We chose to be overly verbose, to allow developers to be warned about the contents of their key set (before putting them online), to make sure they don't include something in there by mistake. What you propose is certainly what a real-life verifier should do. We could be more lenient when validating a health card, not displaying the warnings if one key in the set is ok and can be used with a card, but being a developer tool, it feels like we should report as much as we can. What you propose is certainly what a real-life verifier should do. We could be more lenient when validating a health card, not displaying the warnings if one key in the set is ok and can be used with a card, but being a developer tool, it feels like we should report as much as we can. Gotcha - I appreciate the nuanced thinking between warning and errors! I would learn towards not including warnings due to unrelated keys, but am fine with either decision here. Feel free to close at your discretion. It might be worth adding a way for developers to explicitly opt out of certain checks (e.g., a suppression file, or a suppression CLI flag). The first time you get a warning, it might come with advice like: If you don't want to see this warning in the future you can [...] While I'm exercising my imagination...the suppression could be scoped to types of errors, or even types of errors combined with data (like, "don't warn me about extra keys" vs "don't warn me about this extra key). (The common version of this is for Node testing frameworks, where you can specify --include or --exclude type flags to filter out the set of tests applied.0 It might be worth adding a way for developers to explicitly opt out of certain checks (e.g., a suppression file, or a suppression CLI flag). The first time you get a warning, it might come with advice like: Yes, and it should be fairly easy to do since our logger takes in specific error codes which should be easy to filter out. Ok, so I'll close this, and keep @jmandel's suggestion alive on issue #35.
gharchive/issue
2021-04-01T22:06:13
2025-04-01T06:44:58.480372
{ "authors": [ "christianpaquin", "dleve123", "jmandel" ], "repo": "microsoft/health-cards-validation-SDK", "url": "https://github.com/microsoft/health-cards-validation-SDK/issues/33", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1361700697
Microbit Blutooth doesn't work after adding Jacdac extension Describe the bug I'm able to pair microbit with an android smartphone and display the data sent via bluetooth to microbit led display. bluetooth.onUartDataReceived(serial.delimiters(Delimiters.NewLine), function () { basic.showString(bluetooth.uartReadUntil(serial.delimiters(Delimiters.NewLine))) }) bluetooth.startUartService() basic.forever(function () { }) Once I add the jacdac extension, can't pair with the microbit or connect already paired. Seems like adding the jacdac extension disables the BLE of the microbit. Jacdac extension version for makecode "microsoft/pxt-jacdac 1.8.22" Desktop (please complete the following information): OS: Windows 10 Browser Edge Smartphone (please complete the following information): Device: Android Unfortunately, Jacdac and Bluetooth are incompatible. It seems that we are missing the annotation so that MakeCode refuses to use the 2 togher: [ ] add flag in pxt.json to prevent using with bluetooth [ ] update makecode codes From what I can see here it did work: https://forum.makecode.com/t/video-use-outdated-iphone-to-code-and-download-jacdac-to-microbit-v2/15152 For debugging it is also often handy to use a wireless connection (as well as for classrooms that have only iPads etc.). Shouldn't be the option available to use it together? It would be good, if after Jacdac has been used, the device could still communicate via bluetooth to a device. Right now the pairing pattern appears, but it can't be connected/flashed. More can be found here: https://github.com/microsoft/pxt-calliope/issues/257 though it must be changed in Jacdac unfortunately. Does it not work even with the pairing pattern described in https://support.microbit.org/support/solutions/articles/19000051025-pairing-and-flashing-code-via-bluetooth ? Does it not work even with the pairing pattern described in https://support.microbit.org/support/solutions/articles/19000051025-pairing-and-flashing-code-via-bluetooth ? Unfortunately not. Though it is interesting to see that triple reset also doesn't work, just A+B+Reset – but that has no effect! The app doesn't recognize an advertised device. Does it not work even with the pairing pattern described in https://support.microbit.org/support/solutions/articles/19000051025-pairing-and-flashing-code-via-bluetooth ? I also made some tests with the Jacdac extension 1.9.25, the micro:bit 2.20 and the micro:bit iOS app 3.10. I was able to reproduce the behavior as mentioned by arijitx and joernalraun. If the micro:bit runs a firmware with the jacdac extension enabled, it is still possible to go into pairing mode by A+B+reset, but also 3x reset button works. But the app does not allow any pairing or program transfer to an already paired micro:bit and shows a timeout error: I had a look at the advertised Bluetooth services using the nRF connect app and noticed that a micro:bit with the Jacdac extension in pairing mode is still advertising as a device but does not include in the 'Buttonless DFU with bonds' service, that a micro:bit without the jacdac extension advertises Without Jacdac Extension With Jacdac Extension @pelikhan You mentioned that a firmware update via app should still be possible as long as the pairing mode (pattern) is used, right? Is there a thing that I can still try out? Do the project settings in makecode need to be adjusted? I suggest reviewing the preprocessor flags for BLE in microbit and flag if we haven't missed one.
gharchive/issue
2022-09-05T10:04:23
2025-04-01T06:44:58.491438
{ "authors": [ "arijitx", "fabianhugo", "joernalraun", "pelikhan" ], "repo": "microsoft/jacdac", "url": "https://github.com/microsoft/jacdac/issues/1209", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
2413275745
Fix transaction problem in SQL Server Memory DeleteAsync method Motivation and Context (Why the change? What's the scenario?) This PR fixes an issue when trying to delete a document from SQL Server Memory. High level description (Approach, Design) The DeleteDocumentHandler class calls the memory GetListAsync method and then deletes each returned records, using an async enumeration: https://github.com/microsoft/kernel-memory/blob/2ff894c99d30531fe83a9134c35ec4608583ef59/service/Core/Handlers/DeleteDocumentHandler.cs#L45-L54 The issue is related to how records are returned in SqlServerMemory.GetListAsync: https://github.com/microsoft/kernel-memory/blob/2ff894c99d30531fe83a9134c35ec4608583ef59/extensions/SQLServer/SQLServer/SqlServerMemory.cs#L291-L294 In this code, records are yield inside a DataReader loop, so the connection is kept open until the end of the list. As shown above, DeleteDocumentHandler calls DeleteAsync while iterating on records. In case of SQL Server: https://github.com/microsoft/kernel-memory/blob/2ff894c99d30531fe83a9134c35ec4608583ef59/extensions/SQLServer/SQLServer/SqlServerMemory.cs#L137-L168 This method tries to open a new connection and execute commands within a transaction. Because it is called while executing GetListAsync, there already is an opened connection that is reading data, so it causes a deadlock that ends with a Timeout exception. This PR fixes the issue modifying the GetListAsync method so that it reads all the data, closes the connection and then returns the list. @marcominerva I can't reproduce the problem you're describing here. I added a new test here https://github.com/microsoft/kernel-memory/pull/718/files and the test is working fine. Is the test missing something? I'd like to reproduce the issue before making any change, to be sure the fix works and we won't have future regressions. @dluc You need to import at least two documents, then try to delete the first one. I have created a small repro here: https://github.com/marcominerva/KernelMemorySqlServerIssue Use the first API to upload two different documents, then try to delete the first and you'll obtain a Timeout exception from SQL Server. I merged in the ItDeletesRecords test I added in #718. The test is failing on my local machine, using MSSQL docker image. Could you take a look? @dluc I have just tried the ItDeletesRecords test on this sqlserver-disposefix branch and on my machine it passes correctly (I'm using a LocalDB instance). What kind of failure do you encounter? Getting a timeout exception on batch upserts. Could you try the run-mssql.sh script which runs the docker image? Tried with the container created by run-mssql.sh, but still works correctly. It sounds strange that you get a timeout exception on batch upserts (so, you stop before the delete), because upserts don't call GetListAsync, that it the only method I have modified. I have also tried to set the MemoryDbUpsertBatchSize explicitly in DefaultTests.cs: this._memory = builder .With(new KernelMemoryConfig { DefaultIndexName = "default4tests", DataIngestion = new() { MemoryDbUpsertBatchSize = 64 } }) Do other SQL Server tests work on your local machine? Do other SQL Server tests work on your local machine? some other tests are failing, I'm starting to think if it's a problem with docker and macOS. On main branch: On this PR branch: Execution 1 Execution 2, after restarting the docker image: Execution 1 log Microsoft.Data.SqlClient.SqlException: Execution Timeout Expired. The timeout period elapsed prior to completion of the operation or the server is no... Microsoft.Data.SqlClient.SqlException Execution Timeout Expired. The timeout period elapsed prior to completion of the operation or the server is not responding. at Microsoft.Data.SqlClient.SqlCommand.EndExecuteNonQueryAsync(IAsyncResult asyncResult) at System.Threading.Tasks.TaskFactory`1.FromAsyncCoreLogic(IAsyncResult iar, Func`2 endFunction, Action`1 endAction, Task`1 promise, Boolean requiresSynchronization) --- End of stack trace from previous location --- at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+MoveNext() in km/extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 533 at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+MoveNext() in km/extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 543 at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+System.Threading.Tasks.Sources.IValueTaskSource<System.Boolean>.GetResult() at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertAsync(String index, MemoryRecord record, CancellationToken cancellationToken) in km/extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 435 at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertAsync(String index, MemoryRecord record, CancellationToken cancellationToken) in km/extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 435 at Microsoft.KernelMemory.Handlers.SaveRecordsHandler.SaveRecordAsync(DataPipeline pipeline, IMemoryDb db, MemoryRecord record, HashSet`1 createdIndexes, CancellationToken cancellationToken) in km/service/Core/Handlers/SaveRecordsHandler.cs:line 266 at Microsoft.KernelMemory.Handlers.SaveRecordsHandler.InvokeAsync(DataPipeline pipeline, CancellationToken cancellationToken) in km/service/Core/Handlers/SaveRecordsHandler.cs:line 214 at Microsoft.KernelMemory.Pipeline.InProcessPipelineOrchestrator.RunPipelineAsync(DataPipeline pipeline, CancellationToken cancellationToken) in km/service/Core/Pipeline/InProcessPipelineOrchestrator.cs:line 174 at Microsoft.KernelMemory.Pipeline.BaseOrchestrator.ImportDocumentAsync(String index, DocumentUploadRequest uploadRequest, IContext context, CancellationToken cancellationToken) in km/service/Core/Pipeline/BaseOrchestrator.cs:line 121 at Microsoft.KM.Core.FunctionalTests.DefaultTestCases.RecordDeletionTest.ItDeletesRecords(IKernelMemory memory, IMemoryDb db, Action`1 log) in km/service/tests/Core.FunctionalTests/DefaultTestCases/RecordDeletionTest.cs:line 30 at Microsoft.SQLServer.FunctionalTests.DefaultTests.ItDeletesRecords() in km/extensions/SQLServer/SQLServer.FunctionalTests/DefaultTests.cs:line 119 at Xunit.DependencyInjection.DependencyInjectionTestInvoker.AsyncStack(Task task, Activity activity) System.ComponentModel.Win32Exception Unknown error: 258 Exception doesn't have a stacktrace Execution 2 log Microsoft.Data.SqlClient.SqlException: Transaction (Process ID 73) was deadlocked on communication buffer resources with another process and has been ... Microsoft.Data.SqlClient.SqlException Transaction (Process ID 73) was deadlocked on communication buffer resources with another process and has been chosen as the deadlock victim. Rerun the transaction. at Microsoft.Data.SqlClient.SqlConnection.OnError(SqlException exception, Boolean breakConnection, Action`1 wrapCloseInAction) at Microsoft.Data.SqlClient.SqlInternalConnection.OnError(SqlException exception, Boolean breakConnection, Action`1 wrapCloseInAction) at Microsoft.Data.SqlClient.TdsParser.ThrowExceptionAndWarning(TdsParserStateObject stateObj, SqlCommand command, Boolean callerHasConnectionLock, Boolean asyncClose) at Microsoft.Data.SqlClient.TdsParser.TryRun(RunBehavior runBehavior, SqlCommand cmdHandler, SqlDataReader dataStream, BulkCopySimpleResultSet bulkCopyHandler, TdsParserStateObject stateObj, Boolean& dataReady) at Microsoft.Data.SqlClient.SqlCommand.FinishExecuteReader(SqlDataReader ds, RunBehavior runBehavior, String resetOptionsString, Boolean isInternal, Boolean forDescribeParameterEncryption, Boolean shouldCacheForAlwaysEncrypted) at Microsoft.Data.SqlClient.SqlCommand.CompleteAsyncExecuteReader(Boolean isInternal, Boolean forDescribeParameterEncryption) at Microsoft.Data.SqlClient.SqlCommand.InternalEndExecuteNonQuery(IAsyncResult asyncResult, Boolean isInternal, String endMethod) at Microsoft.Data.SqlClient.SqlCommand.EndExecuteNonQueryInternal(IAsyncResult asyncResult) at Microsoft.Data.SqlClient.SqlCommand.EndExecuteNonQueryAsync(IAsyncResult asyncResult) at System.Threading.Tasks.TaskFactory`1.FromAsyncCoreLogic(IAsyncResult iar, Func`2 endFunction, Action`1 endAction, Task`1 promise, Boolean requiresSynchronization) --- End of stack trace from previous location --- at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+MoveNext() in km//extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 533 at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+MoveNext() in km//extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 543 at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+System.Threading.Tasks.Sources.IValueTaskSource<System.Boolean>.GetResult() at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertAsync(String index, MemoryRecord record, CancellationToken cancellationToken) in km//extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 435 at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertAsync(String index, MemoryRecord record, CancellationToken cancellationToken) in km//extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 435 at Microsoft.KernelMemory.Handlers.SaveRecordsHandler.SaveRecordAsync(DataPipeline pipeline, IMemoryDb db, MemoryRecord record, HashSet`1 createdIndexes, CancellationToken cancellationToken) in km//service/Core/Handlers/SaveRecordsHandler.cs:line 266 at Microsoft.KernelMemory.Handlers.SaveRecordsHandler.InvokeAsync(DataPipeline pipeline, CancellationToken cancellationToken) in km//service/Core/Handlers/SaveRecordsHandler.cs:line 214 at Microsoft.KernelMemory.Pipeline.InProcessPipelineOrchestrator.RunPipelineAsync(DataPipeline pipeline, CancellationToken cancellationToken) in km//service/Core/Pipeline/InProcessPipelineOrchestrator.cs:line 174 at Microsoft.KernelMemory.Pipeline.BaseOrchestrator.ImportDocumentAsync(String index, DocumentUploadRequest uploadRequest, IContext context, CancellationToken cancellationToken) in km//service/Core/Pipeline/BaseOrchestrator.cs:line 121 at Microsoft.KM.Core.FunctionalTests.DefaultTestCases.RecordDeletionTest.ItDeletesRecords(IKernelMemory memory, IMemoryDb db, Action`1 log) in km//service/tests/Core.FunctionalTests/DefaultTestCases/RecordDeletionTest.cs:line 30 at Microsoft.SQLServer.FunctionalTests.DefaultTests.ItDeletesRecords() in km//extensions/SQLServer/SQLServer.FunctionalTests/DefaultTests.cs:line 119 at Xunit.DependencyInjection.DependencyInjectionTestInvoker.AsyncStack(Task task, Activity activity)
gharchive/pull-request
2024-07-17T10:47:01
2025-04-01T06:44:58.508094
{ "authors": [ "dluc", "marcominerva" ], "repo": "microsoft/kernel-memory", "url": "https://github.com/microsoft/kernel-memory/pull/712", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1049045442
Partial Revert of #608, Updating Min Target to Ensure 8.1 Backcompat #608 Changed the WindowsTargetPlatform to hard target Windows 8.1, so that Psychonauts 2 would be able to run on Windows 7. However, this change breaks XAL build pipelines, as our Windows2019 images don't support this scenario. We can do a partial revert and add Windows 8.1 backwards compatibility by specifying a MinTargetVersion Since I didn't create a new branch for this fix, it's re-adding the commits for updating the Android NDK. That was already merged in #631
gharchive/pull-request
2021-11-09T19:52:10
2025-04-01T06:44:58.510063
{ "authors": [ "SahilAshar" ], "repo": "microsoft/libHttpClient", "url": "https://github.com/microsoft/libHttpClient/pull/632", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
449810864
export the blocks to DXF (BLOCKS section of the DXF file) How to output blocks with primitives when exporting to DXF? Hi @PodgPow , I may also be needing to have blocks exported, when #133 is fully implemented. Can you add details about your use case? I want to use blocks for exporting to the DXF path, consisting of several circuits (in SVG everything is fine with this). Can you say what advantage it is to have things in blocks? One example - is the transfer to a DXF shape with a hole, so that you can work in AutoCAD (ArtCAM) with the shape as a solid (not fragmented by contour) object I would like to use makerjs to produce parametric dxf AAMA files for CNC textile cutters. (Lectra Gerber etc) Some of the software used by these machines requires dxf files with blocks to separate the patterns from one another. There are other options where each pattern could be exported as it's own dxf file but I'd prefer to keep it all together if possible so that I can use the file to reference the parts together. It looks like all the magic is happening in the packages/maker.js/src/core/dxf.ts file. I've downloaded the Autodesk DXF reference but honestly, it's pretty overwhelming. It looks like there needs to be a blocks section added to the exported file and then the block references can be included as an entity to inside the drawing with it's own origin? Is there a simpler reference that I'm missing somewhere? Thanks HI @cgowen , you're not missing anything - the DXF format is not super easy. I've done a small amount of digging through Audodesk references and files I exported with the exdxf library and have come up with a short breakdown of how to add blocks to a dxf file. I still don't know enough javascript to make this happen. I'm hoping that having the file structure may be helpful to someone who might be a better coder than I am. https://anonfiles.com/w5DaHbm2uc/DXF_Format_Cheatsheet_pdf https://anonfiles.com/u8D7H4mfu7/handmade_dxf
gharchive/issue
2019-05-29T13:32:20
2025-04-01T06:44:58.524001
{ "authors": [ "PodgPow", "PowPodg", "cgowen", "danmarshall" ], "repo": "microsoft/maker.js", "url": "https://github.com/microsoft/maker.js/issues/411", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
554476418
TabView should not allow Narrator users to deselect current tab Describe the bug Steps to reproduce the bug Steps to reproduce the behavior: Turn on Narrator. Launch XAML Controls Gallery. Activate Tabview from All Controls Page. Navigate to any tab and press caps+enter to select it now again caps+enter and observe. Expected behavior Nothing should happen, the de-selection should be ignored. Screenshots Version Info Xaml Controls Gallery version 1.2.12.0. NuGet package version: Additional context Copied from internal bug I think this issue was fixed with the recent changes we made to the TabView automationpeer. This behavior is not reproducable in the MUXControlsTestApp anymore, so I think it's fine to close it now. @StephenLPeters @ranjeshj FYI
gharchive/issue
2020-01-24T00:01:39
2025-04-01T06:44:58.529552
{ "authors": [ "YuliKl", "chingucoding" ], "repo": "microsoft/microsoft-ui-xaml", "url": "https://github.com/microsoft/microsoft-ui-xaml/issues/1880", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
559210864
URI encoded references are not resolved in JSON schema VSCode version: 1.41.1 Browser: Electron, Chrome OS: MacOS This schema encoded special characters in the reference URI. The standard says that references must be valid URIs: https://json-schema.org/draft/2019-09/json-schema-core.html#ref and can, therefore, be encoded. { "$schema": "http://json-schema.org/draft-07/schema#", "definitions": { "Foo<number>": { "type": "object", "properties": {"q1": {"enum": ["x1", "x2"]}} } }, "type": "object", "properties": { "p1": {"enum": ["v1", "v2"]}, "p2": {"$ref": "#/definitions/Foo%3Cnumber%3E"} } } Monaco doesn't correctly resolve the references. { "$schema": "https://gist.githubusercontent.com/domoritz/3cd0ddb8cad62ad611f301b2a8743ebf/raw/3bc4ff9350a8e9a7acb54d9683ce3a6d3c179377/schema.json", } duplicate of https://github.com/microsoft/vscode-json-languageservice/issues/49 Ahh, that’s where I filed the issue. Thanks.
gharchive/issue
2020-02-03T16:59:21
2025-04-01T06:44:58.541139
{ "authors": [ "aeschli", "domoritz" ], "repo": "microsoft/monaco-editor", "url": "https://github.com/microsoft/monaco-editor/issues/1804", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
446833176
Fix C# sample code String literal does not need to be @-escaped. Alternatively, one could remove one of the backslashes in the line. The @ is used to test/show off the colorizer.
gharchive/pull-request
2019-05-21T21:26:52
2025-04-01T06:44:58.542346
{ "authors": [ "DaMightyZombie", "alexandrudima" ], "repo": "microsoft/monaco-editor", "url": "https://github.com/microsoft/monaco-editor/pull/1450", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2140347358
Netperf version 1.5 TODO: PGO/SPGO Support We need to add support for PGO runs and then the automation to push back the PGO files to in a GitHub PR. Mentioning SPGO here; collecting and pushing to SPGO is another option, we can do one or the other or both.
gharchive/issue
2024-02-17T17:29:51
2025-04-01T06:44:58.547228
{ "authors": [ "mtfriesen", "nibanks" ], "repo": "microsoft/netperf", "url": "https://github.com/microsoft/netperf/issues/76", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2122979608
Is Memory supported as return type in dotnet? I tried to use Memory<byte> as return type of a method on dotnet side and call the method from Javascript. However, although the generated .d.ts file contains the method signature, the Javascript code is not able to call that method. The error message says the object does not have such method/property/function. I tried to change the return type to string or byte[] then the method could be called from Javascript. Is Memory<byte> supported? BTW, I suspected marshaller code needs to be generated and utlised, however, I couldn't find any generated file. I've created another issue for that: #207 A .NET Memory<byte> should get converted to JS UInt8Array, in both typedefs and runtime. But there might be some other reason you're unable to call the method. Can you check what properties are defined on the JS object? Object.keys(obj) Do you see other expected methods/properties on the object? How are you getting the JS instance of the class? Can you check what properties are defined on the JS object? Object.keys(obj) If you mean the JS object returned from the .NET method I want to call, because the method can't be called (the JS class instance does not have such function on it) there is no way to get the returned value. Do you see other expected methods/properties on the object? Other methods exist on the class instance as long as their return types are not Memory<T>. I tried changing the return type from Memory<byte> to string or byte[], both made that method available to Node.js. How are you getting the JS instance of the class? The type definition was generated through MSBuild. I believe the class instance was created through reflection becaue I suspect the code generate didn't work (as described in #207 ) I hope there is a way to tell whether refection is used or generated code is used. I suspect the missing of that method is caused by the missing of generated code. BTW, currently I use base64 encoded string to pass binary data from .NET to Node.js as a workaround. I confirmed in #220 with more testing that Memory<byte> is working as expected. So there is likely some other problem here. the JS class instance does not have such function on it This sounds like there was some other problem with the method that is unrelated to the fact that it returns Memory<byte>. If you want to troubleshoot further, it might help to share relevant snippets of the code. I'm closing this for now since the originally reported problem is not reproducible. But feel free to re-open with more information if you're still having this problem.
gharchive/issue
2024-02-07T12:57:12
2025-04-01T06:44:58.555289
{ "authors": [ "james-hu", "jasongin" ], "repo": "microsoft/node-api-dotnet", "url": "https://github.com/microsoft/node-api-dotnet/issues/208", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1444570283
Release 5.20.0 Summary of the Pull Request What is this about? Release 5.20.0 Added Service: Added endpoint to download agent binaries to support the unmanaged node scenario. #2600 Service: Added additional error handling when updating VMSS nodes. #2607 Changed Service: Added additional logging when using the decommission node policy. #2605 Agent/Supervisor/Proxy: Updated third-party Rust dependencies.#2608 Service: Added optional retry_limit when connecting to the repro machine. #2609 Fixed Service: Fixed status top in C# implementation. #2604 Codecov Report Merging #2614 (5fecc23) into main (ff85b80) will decrease coverage by 4.97%. The diff coverage is 17.94%. @@ Coverage Diff @@ ## main #2614 +/- ## ========================================== - Coverage 29.88% 24.91% -4.98% ========================================== Files 290 121 -169 Lines 35846 12533 -23313 ========================================== - Hits 10714 3122 -7592 + Misses 25132 9411 -15721 Impacted Files Coverage Δ src/ApiService/ApiService/Functions/ReproVmss.cs 0.00% <0.00%> (ø) src/ApiService/ApiService/Functions/Tasks.cs 22.22% <0.00%> (ø) src/ApiService/ApiService/UserCredentials.cs 8.82% <8.33%> (+2.26%) :arrow_up: ...ice/ApiService/onefuzzlib/EndpointAuthorization.cs 22.13% <11.11%> (-0.76%) :arrow_down: src/ApiService/ApiService/Functions/Jobs.cs 76.59% <100.00%> (ø) src/ApiService/ApiService/OneFuzzTypes/Model.cs 72.05% <100.00%> (+0.07%) :arrow_up: src/agent/coverage/src/lib.rs src/agent/onefuzz-telemetry/src/lib.rs src/agent/onefuzz-agent/src/config.rs src/agent/coverage/src/cobertura.rs ... and 165 more Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
gharchive/pull-request
2022-11-10T21:14:07
2025-04-01T06:44:58.572774
{ "authors": [ "codecov-commenter", "mgreisen" ], "repo": "microsoft/onefuzz", "url": "https://github.com/microsoft/onefuzz/pull/2614", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2123977263
No information when tokenizer can't load config files To repro, run the following code, without the Llama tokenizer.model available. import onnxruntime_genai as og import time print("Loading model...") model=og.Model("model", og.DeviceType.CPU) print("Model loaded") tokenizer=model.create_tokenizer() print("Tokenizer created") # Keep asking for input prompts in an loop while True: text = input("Input:") input_tokens = tokenizer.encode(text) params=og.search_params(model) params.max_length = 64 params.input_ids = input_tokens start_time=time.time() output_tokens=model.generate(params) run_time=time.time()-start_time; print(f"Tokens: {len(output_tokens)} Time: {run_time:.2f} Tokens per second: {len(output_tokens)/run_time:.2f}") print("Output:") print(tokenizer.decode(output_tokens)) print() print() tokenizer=model.create_tokenizer() RuntimeError Fixed 2/26/24
gharchive/issue
2024-02-07T22:04:10
2025-04-01T06:44:58.574678
{ "authors": [ "natke" ], "repo": "microsoft/onnxruntime-genai", "url": "https://github.com/microsoft/onnxruntime-genai/issues/65", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2753136655
Figure out what cr8 should do in AccessVpState AccessVpState stubs out cr8 read/writes. Figure out if/how we should wire this up. Supposedly it's only used for the instruction emulator, which doesn't need cr8. These error tracing calls will be downgraded to trace. There's also updates to rflags in this path For CR8 @chris-oo thinks when we fix the instruction emulator to only grab state it needs, this goes away, but we need to confirm
gharchive/issue
2024-12-20T17:12:33
2025-04-01T06:44:58.668390
{ "authors": [ "cperezvargas" ], "repo": "microsoft/openvmm", "url": "https://github.com/microsoft/openvmm/issues/564", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1169023939
Non-blocking EventPipeEventSource Ctor This PR makes the constructor for EventPipeEventSource non-blocking. Previously, the constructor would block until data was sent on the stream. This brings EventPipeEventSource closer in behavior to other similar APIs in the TraceEvent library. One issue that isn't solved by this PR (yet) is that we need to handle the case where users try to get metadata about the source before data arrives. Before this PR, users were guaranteed to have that information when the ctor returned. Now, this is only true for opening a completed file, hence the special casing for files. Currently, this PR makes no attempt to stop users from reading metadata properties. I am hoping for some feedback on whether this should throw an exception or now. Until this is decided, I'll leave this in draft mode. Currently, this PR makes no attempt to stop users from reading metadata properties I'd say throwing is desirable if the values aren't initialized yet, but if it was a significant hassle to implement I don't think it would be awful to return some default 0/null/empty string values instead. What do you think @brianrob? I'd say throwing is desirable if the values aren't initialized yet Agreed. There are Debug.Asserts scattered around that check whether this metadata is set before certain operations. These are mostly buried in the TraceLog APIs from what I saw. These caused the file-based tests to fail until I special cased the file-based ctor. Depending on @brianrob's thoughts, I'll look at making these properties throw if Process hasn't been called yet. I would tend to agree that if there is metadata that can't be filled before returning from the constructor, then throwing is a good idea. If you look at how we handle realtime ETW sessions, the session is created and setup before you get the source from it, which means that the source can make calls to the session to get metadata. Would such a pattern help here? If not, I would say throwing is the next best thing. Would such a pattern help here? I think yes, however, the API for configuring, starting, and stopping an EventPipe session is in the diagnostics client library and not in TraceEvent. That makes it a little more complicated to make this pattern happen. Right now, the flow for "trace on startup" is this: [Diag Client Lib] Create configuration (list of providers, buffersize, etc.) [Diag Client Lib] Start EventPipe session (get Stream object) [TraceEvent lib] Create EventPipeEventSource using EventPipeEventSource(Stream stream) ctor [TraceEvent lib] Call EventPipeEventSource.Process() [Diag Client lib] Resume target [Diag Client lib] Stop EventPipe session In the "trace on startup" scenario, we need to wait for the target process to resume before the session starts. Without this change, the ctor call in step 3 would block until we resume the target. After this change, the ctor would complete, but the metadata wouldn't be populated yet. Any usage of that EventPipeEventSource before resuming the target could cause errors in other parts of TraceEvent regardless of whether EventPipeEventSource.Process has been called. The metadata should only be present after step 5. Gotcha. So this becomes a bit more complicated. I think it is OK to simplify and just throw in APIs that don't have data yet. Just wanted to make sure that there wasn't another pattern that might help. @josalem, I see that you just pushed another commit here. Please let me know when this is ready for another review, or is ready for merge. Thanks. I think this is ready for final review and merge at your discretion.
gharchive/pull-request
2022-03-14T23:38:49
2025-04-01T06:44:58.676646
{ "authors": [ "brianrob", "josalem", "noahfalk" ], "repo": "microsoft/perfview", "url": "https://github.com/microsoft/perfview/pull/1588", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1084939105
[Feature]: Modification of network response Feature request Follow-up on #1816 This is supported and documented in the Node.js version of playwright. Node.js section: https://playwright.dev/docs/network#modify-responses It does not seem exist for the .NET version. Merging into https://github.com/microsoft/playwright-dotnet/issues/1905
gharchive/issue
2021-12-20T15:47:56
2025-04-01T06:44:58.679584
{ "authors": [ "krokofant", "mxschmitt" ], "repo": "microsoft/playwright-dotnet", "url": "https://github.com/microsoft/playwright-dotnet/issues/1904", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2344983315
FIle Upload box doesn't appear. Version 1.44.0 Steps to reproduce In a app that I am using, file upload box doesn't appear when clicking the choose file element as below: page.waitForFileChooser(() -> locator1.click(); Tried with set force true but same issues page.waitForFileChooser(() -> locator1.click(new Locator.ClickOptions().setForce(true)); When I click the choose file with a simple click with the above waitforfilechooser wrapper method, it works just fine and opens the file upload dialog. locator1.click(new Locator.ClickOptions().setForce(true)) Any idea what the issue could be. Expected behavior FIle upload box should appear Actual behavior File upload box not appearing Additional context No response Environment Mac OS Playwright 1.44.0 Java 17 Chromium Please follow our bug template and provide a minimal self-contained project that we could run locally to reproduce the problem.
gharchive/issue
2024-06-10T22:49:59
2025-04-01T06:44:58.683596
{ "authors": [ "maheshwg", "yury-s" ], "repo": "microsoft/playwright-java", "url": "https://github.com/microsoft/playwright-java/issues/1596", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1186807691
Hello, boss. I think this code will overflow memory after running for a few hours? Excuse me, where is the problem? Please check it for me. I have been checking this code for a long time and can't find the cause of memory overflow package com.artisan.spider; import com.artisan.spider.consts.Consts; import com.artisan.spider.domain.SpError; import com.artisan.spider.domain.SpLawChl; import com.artisan.spider.domain.SpLawLar; import com.artisan.spider.service.ISpErrorService; import com.artisan.spider.service.impl.SpLawChlServiceImpl; import com.artisan.spider.service.impl.SpLawLarServiceImpl; import com.artisan.spider.util.DateUtil; import com.artisan.spider.util.SpiderUtil; import com.artisan.spider.util.SpringContextUtil; import com.microsoft.playwright.Browser; import com.microsoft.playwright.BrowserContext; import com.microsoft.playwright.Page; import com.microsoft.playwright.Playwright; import lombok.extern.slf4j.Slf4j; import org.junit.jupiter.api.Test; import org.springframework.beans.factory.annotation.Autowired; import org.springframework.boot.test.context.SpringBootTest; import java.util.ArrayList; import java.util.List; @SpringBootTest @Slf4j public class DffggzDfgfxwjTaskTest01 { @Autowired ISpErrorService iSpErrorService; @Test public void collectionInfo(){ try (Playwright playwright = Playwright.create()) { String pageUrl = Consts.dfgfxwj; Browser browser = playwright.chromium().launch(); BrowserContext context = browser.newContext(); Page page = context.newPage(); page.navigate(String.format(pageUrl, 1)); int pageNum = Integer.valueOf(page.querySelector(".page").innerText().split("/")[1].replace("页", "").trim()); for (int i = 1; i <= 3; i++) { SpiderUtil.pageChuli("dfflgz",context,page, pageUrl, i, pageNum); } } } } ` `public static void pageChuli(String mark,BrowserContext browserContext, Page page, String pageUrl, int currentNum, int pageNum) { Locator rows = page.locator("td a"); int count = rows.count(); log.info("每页的大小--》{},当前页数{}", count,currentNum); for (int k = 0; k < count; k++) { // 判断是不是曾经抓取过 QueryWrapper<SpLawChl> queryWrapper=new QueryWrapper(); queryWrapper.eq("page_url",rows.nth(k).getAttribute("href")); int lawNumber=SpringContextUtil.getBean(ISpLawChlService.class).count(queryWrapper); if(lawNumber==0){ Page detail = null; try{ detail= browserContext.newPage(); String herf= Consts.rootUrl+"/"+rows.nth(k).getAttribute("href"); detail.navigate(herf); log.info("第{}页-第{}篇文章的标题----->{},url为{},", currentNum, k + 1, rows.nth(k).textContent(),herf); String title=detail.locator(".qw-bt").textContent(); // 发布部门也可能存在多个 Locator fbbmLocator = detail.locator("td:right-of(:text(\"发布部门\"))"); String fbbm = ""; String fbbmCode = ""; if (fbbmLocator.count() != 0) { Locator lb=fbbmLocator.nth(0).locator("a"); if (lb.count() != 0) { if (lb.count() > 1) { for (int i = 0; i < lb.count(); i++) { Locator locator=lb.nth(i); fbbmCode += getType(locator,"fdep_id","&menuname"); fbbm +=locator.textContent(); if(i!= lb.count()-1){ fbbmCode+=","; fbbm+=","; } } } else { Locator locator=lb.nth(0); fbbm = locator.textContent(); fbbmCode=getType(locator,"fdep_id","&menuname"); } } } String fwzh = ""; Locator fbzhLocator = detail.locator("td:right-of(:text(\"发文字号\"))"); if (fbzhLocator.count() != 0) { fwzh = fbzhLocator.nth(0).textContent(); } String fbrq = ""; Locator fbrqLocator = detail.locator("td:right-of(:text(\"发布日期\"))"); if (fbrqLocator.count() != 0) { fbrq = fbrqLocator.nth(0).textContent(); } String ssrq = ""; Locator ssrqLocator = detail.locator("td:right-of(:text(\"实施日期\"))"); if (ssrqLocator.count() != 0) { ssrq = ssrqLocator.nth(0).textContent(); } String ppbm = ""; String ppbmCode = ""; Locator ppbmLocator = detail.locator("td:right-of(:text(\"批准部门\"))"); if (ppbmLocator.count() != 0) { Locator lb=ppbmLocator.nth(0).locator("a"); if (lb.count() != 0) { if (lb.count() > 1) { for (int i = 0; i < lb.count(); i++) { Locator locator=lb.nth(i); ppbmCode += getType(locator,"pdep_id","&menuname"); ppbm +=locator.textContent(); if(i!= lb.count()-1){ ppbmCode+=","; ppbm+=","; } } } else { Locator locator=lb.nth(0); ppbm = locator.textContent(); ppbmCode=getType(locator,"pdep_id","&menuname"); } } } String sxx = ""; String sxxCode=""; Locator sxxLocator = detail.locator("td:right-of(:text(\"时效性\"))"); if (sxxLocator.count() != 0) { Locator locator=sxxLocator.nth(0).locator("a"); sxx = locator.textContent(); sxxCode=getType(locator,"shixiao_id","&menuname"); } String xljb = ""; String xljbCode=""; Locator xljbLocator = detail.locator("td:right-of(:text(\"效力级别\"))"); if (xljbLocator.count() != 0) { Locator locator=xljbLocator.nth(0).locator("a"); xljb = locator.textContent(); xljbCode=SpiderUtil.getType(locator,"xiaoli_id","&menuname"); } // 法规类别有可能是多个 String fglb = ""; String fglbCode=""; Locator fglbLocator = detail.locator("td:right-of(:text(\"法规类别\"))"); if (fglbLocator.count() != 0) { // 如果法规类别有多个 Locator lb = fglbLocator.nth(0).locator("a"); if (lb.count() != 0) { if (lb.count() > 0) { for (int i = 0; i < lb.count(); i++) { Locator locator=lb.nth(i); fglb += locator.textContent(); fglbCode += getType(locator,"sort_id","&menuname"); if(i!= lb.count()-1){ fglb+=","; fglbCode+=","; } } } else { Locator locator=lb.nth(0); fglb = locator.textContent(); fglbCode=getType(locator,"sort_id","&menuname"); } } } String content = detail.locator("#div_content").innerHTML(); log.info("发布部门名称,编码--》{},{}", fbbm,fbbmCode); log.info("发文字号--》{}", fwzh); log.info("发布日期--》{}", fbrq); log.info("实施日期--》{}", ssrq); log.info("时效性名称,编码--》{},{}", sxx,sxxCode); log.info("效力级别名称,编码--》{},{}", xljb,xljbCode); log.info("法规类别名称,编码--》{}", fglb,fglbCode); log.info("*******************************第{}页-第{}篇文章处理完成", currentNum, k + 1); SpLawChl spLawChl=new SpLawChl(); spLawChl.setId(UuidUtils.getUUid()); spLawChl.setFbDeptName(fbbm); spLawChl.setFbDeptCode(fbbmCode); spLawChl.setPzDeptName(ppbm); spLawChl.setPzDeptCode(ppbmCode); spLawChl.setLssuedNumber(fwzh); spLawChl.setReleaseDate(fbrq); spLawChl.setTimeName(sxx); spLawChl.setTimeCode(sxxCode); spLawChl.setXiaoliName(xljb); spLawChl.setXiaoliCode(xljbCode); spLawChl.setFaguiName(fglb); spLawChl.setFaguiCode(fglbCode); spLawChl.setImplDate(ssrq); spLawChl.setTitle(title); spLawChl.setFullText(content); spLawChl.setStatus("1"); spLawChl.setCreateTime(DateUtil.getCurrentTime()); spLawChl.setPageUrl(rows.nth(k).getAttribute("href")); spLawChl.setFgfl(mark); spLawChl.setIsCreate("0"); SpringContextUtil.getBean(ISpLawChlService.class).save(spLawChl); // 把数据发布到相应的消息对列中 SpringContextUtil.getBean(KafkaTemplate.class).send("messageQueen", spLawChl.getId()); detail.close(); }catch (Exception e){ e.printStackTrace(); detail.close(); log.info("*******************************第{}页-第{}篇文章处理失败", currentNum, k + 1); log.info("错误链接:"+Consts.rootUrl+"/"+rows.nth(k).getAttribute("href")); SpError spError=new SpError(); spError.setCreateTime(DateUtil.getCurrentTime()); spError.setPageUrl(rows.nth(k).getAttribute("href")); SpringContextUtil.getBean(ISpErrorService.class).save(spError); continue; } }else{ continue; } } // 跳转列表页 if (pageNum != currentNum) { page.navigate(String.format(pageUrl, currentNum + 1)); } } In the inner loop you reuse one and the same BrowserContext for all pages( detail= browserContext.newPage()), it may hold some of the network resources until it's closed. We usually recommend working with each page in a new context for better isolation and so that you don't unnecessarily consume memory. The example is quite generic and impossible to run locally so there is not much we can help with except for general suggestions based on our understanding of the code. In the inner loop you reuse one and the same BrowserContext for all pages( detail= browserContext.newPage()), it may hold some of the network resources until it's closed. We usually recommend working with each page in a new context for better isolation and so that you don't unnecessarily consume memory. The example is quite generic and impossible to run locally so there is not much we can help with except for general suggestions based on our understanding of the code. Thank you. I'm trying to modify the experiment Hello, boss, because I'm multi-layer for loop nesting, can I reuse one browsercontext for 40 pages? The previous list page was 10000 pages, with 40 articles per page, so it may be the problem of reusing browsercontext you said, resulting in memory overflow Hello, after testing, the memory will still rise.. Thank you. It seems that the problem has been solved. After putting it on the server, the memory is now normal after a few hours
gharchive/issue
2022-03-30T18:05:23
2025-04-01T06:44:58.692175
{ "authors": [ "wasd345", "yury-s" ], "repo": "microsoft/playwright-java", "url": "https://github.com/microsoft/playwright-java/issues/875", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
858066873
[BUG]: Not Getting Capacity Alerts if Approved Capacity is Null Describe the bug If the Approved Capacity is null, then you do not get any Capacity Alerts Component (please tell us which flow or app you are experiencing issues with): Solution Core Version 2.0 App or Flow Get Capacity Alerts To Reproduce Steps to reproduce the behavior: set the Approved Capacity to null Expected behavior If the actual capacity is > 0 and is less than the approved capacity is null, then I would have expected there to have been a Capacity Alert. There is no capacity alert if the approved capacity is null. This is because the "List Environment Capacity Information" has a filter in it which filters out record where the approved capacity is null (See screenshot below) Screenshots Hello. This is by design. We do not want to send alerts unless the admin has explicitly decided to add capacity limits for the environments. You just need to enter these to be some default value if you want to have alerts for all your environments. Thanks for using CoE How do you add these default values ? From: Jenefer Monroe @.> Sent: 14 April 2021 18:18 To: microsoft/powerapps-tools @.> Cc: Nigel Price @.>; Author @.> Subject: Re: [microsoft/powerapps-tools] [BUG]: Not Getting Capacity Alerts if Approved Capacity is Null (#932) Hello. This is by design. We do not want to send alerts unless the admin has explicitly decided to add capacity limits for the environments. You just need to enter these to be some default value if you want to have alerts for all your environments. Thanks for using CoE — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHubhttps://github.com/microsoft/powerapps-tools/issues/932#issuecomment-819684766, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ACAXJUEOYVLMD3LWGCUIA5LTIXFADANCNFSM425XVHBQ. If you're looking for a daily capacity report, you may find it easier to set up a separate flow following these steps: https://docs.microsoft.com/en-us/power-platform/admin/programmability-tutorial-create-daily-capacity-report
gharchive/issue
2021-04-14T16:32:58
2025-04-01T06:44:58.757063
{ "authors": [ "JeneferM-MSFT", "NPrice99", "manuelap-msft" ], "repo": "microsoft/powerapps-tools", "url": "https://github.com/microsoft/powerapps-tools/issues/932", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2332237275
Invalid instrumentation key Describe the bug I am following this tutorial: https://learn.microsoft.com/en-us/azure/machine-learning/prompt-flow/how-to-trace-local-sdk?view=azureml-api-2&tabs=python I logged in using azure cli and then entered the command to set up my trace destination: pf config set trace.destination=azureml://subscriptions/<your_subscription_id>/resourcegroups/<your_resourcegroup_name>/providers/Microsoft.MachineLearningServices/workspaces//<your_azureml_workspace_name> This created a CosmosDB. I promptflow does not do this every time I run it on different machines. I tried to send traces using the trace-autogen-groupchat.ipynb, and I got this error: ERROR:opencensus.ext.azure.common.transport:Non-retryable server side error 400: {"itemsReceived":2,"itemsAccepted":0,"appId":null,"errors":[{"index":0,"statusCode":400,"message":"Invalid instrumentation key"},{"index":1,"statusCode":400,"message":"Invalid instrumentation key"}]}. ERROR:opencensus.ext.azure.common.transport:Non-retryable server side error 400: {"itemsReceived":1,"itemsAccepted":0,"appId":null,"errors":[{"index":0,"statusCode":400,"message":"Invalid instrumentation key"}]}. How To Reproduce the bug Steps to reproduce the behavior, how frequent can you experience the bug: Download the trace-autogen-groupchat.ipynb notebook from this repo Follow the instructions for setting the trace destination https://learn.microsoft.com/en-us/azure/machine-learning/prompt-flow/how-to-trace-local-sdk?view=azureml-api-2&tabs=python Try to run the code in the autogen section, where agents generate messages and send them using tracing. Expected behavior I expected the information to be sent without any error. Using Windows, Python 11, and the latest version of PromptFlow. Hi @tyler-suard-parker , thank you for reaching out. I think there are several questions in this issue: pf config set trace.destination does create a Cosmos DB, but one workspace will only have one Cosmos resource, so I understand when you run same command on different machines, only the first time you will see the Cosmos setup process (and wait for that). For the error, it's wired to me because we did use opencensus before, but in #2175 , which is merged 3 months ago, we already migrate to Open Telemetry. Could you please share some screenshots for your error? and provide your prompt flow version with pf -v? so that maybe we can re-produce your error and better investigate on that.
gharchive/issue
2024-06-03T23:32:11
2025-04-01T06:44:58.768681
{ "authors": [ "tyler-suard-parker", "zhengfeiwang" ], "repo": "microsoft/promptflow", "url": "https://github.com/microsoft/promptflow/issues/3373", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1822982439
Type Mismatch Errors with @types/react and TypeScript in PXT-Microbit Project Describe the bug When trying to build the PXT-Microbit project, a series of TypeScript errors occur related to the @types/react package. The errors indicate that subsequent property declarations must have the same type, which seems to be a mismatch between the types defined in the @types/react package. To Reproduce Steps to reproduce the behavior: Clone the PXT-Microbit repository (https://github.com/microsoft/pxt-microbit). Run npm install to install the necessary packages. Run pxt serve to build and serve the project. See the TypeScript errors related to @types/react. Expected behavior The project should build without any TypeScript errors and serve locally for development. Desktop (please complete the following information): OS: MacOS Node.js Version: LTS TypeScript Version: 4.4.3 @types/react Version: 17.0.5 Additional context I've tried several solutions to resolve these errors, including updating the @types/react package to the latest version, deleting the node_modules directory and reinstalling the packages, and adding "skipLibCheck": true to the tsconfig.json file. Unfortunately, none of these solutions have resolved the issue @emoltz could you run node -v and get the version of node.js you are running? 18.17.0, but I tried it with 20.5.0 and v14 via nvm This is also affecting the build of the pxt repo, i.e. when I run npm run build I get an error related to a type mismatch with the Provider component. This issue seems to extend to all pxt-related repos.
gharchive/issue
2023-07-26T18:57:42
2025-04-01T06:44:58.783869
{ "authors": [ "emoltz", "jwunderl" ], "repo": "microsoft/pxt", "url": "https://github.com/microsoft/pxt/issues/9619", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1292898377
VSCode option to turn off syntax reporting of future Python versions Hello, I have the following workflow. I remotely connect to a cluster access machine (via VSCode over SSH), where I make changes to my repository. The access machine runs Python 3.4, thus I assume Pylance operates with it to analyze my repo. The issue is that the code that I develop runs on cluster nodes with Python 3.8, and my code extensively uses some of newer Python features like f-strings and type annotations. Pylance annoingly complains about "Format string literals (f-strings) require Python 3.6 or newer" at hundreds and hundreads of places of my repository. I haven't found any VSCode option to turn these errors off. It would be great if there was an option like "python.analysis.diagnosticSeverityOverrides": { "reportFuturePythonSyntax": "none", }, hellow, sorry for the errors... well one way to solve the problem is to install a new version of python this can somehow help and the good thing about this is that you can install multiple python versions so you can choose which version you need to run your code On Mon, 4 Jul 2022 at 12:36, Alex Larionov @.***> wrote: Hello, I have the following workflow. I remotely connect to a cluster access machine (via VSCode over SSH), where I make changes to my repository. The access machine runs Python 3.4, thus I assume Pylance operates with it to analyze my repo. The issue is that the code that I develop runs on cluster nodes with Python 3.8, and my code extensively uses some of newer Python features like f-strings. Pylance annoingly complains about "Format string literals (f-strings) require Python 3.6 or newer" at hundreds and hundreads of places of my repository. I haven't found any VSCode option to turn these errors off. It would be great if there was an option like "python.analysis.diagnosticSeverityOverrides": { "reportFuturePythonSyntax": "none", }, — Reply to this email directly, view it on GitHub https://github.com/microsoft/pylance-release/issues/2994, or unsubscribe https://github.com/notifications/unsubscribe-auth/AYGBJN7AHYYZMUOCXUK4WRDVSKV77ANCNFSM52SRSIFA . You are receiving this because you are subscribed to this thread.Message ID: @.***> @BerylXavier Hi, as I said, I work on a cluster, and managing its software is beyond my control By default, pyright (the type checker upon which pylance is built) assumes that you are running a version of python that matches the selected python interpreter. You can override this by creating a pyrightconfig.json file in the root directory of your project and including { "pythonVersion": "3.8" }. For additional configuration options, refer to this documentation. Works like a charm! Though having unrelated files in the project root annoys a bit If you would prefer to use a pyproject.toml file, pyright also supports that.
gharchive/issue
2022-07-04T09:35:46
2025-04-01T06:44:58.791804
{ "authors": [ "BerylXavier", "erictraut", "laralex" ], "repo": "microsoft/pylance-release", "url": "https://github.com/microsoft/pylance-release/issues/2994", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
607920860
Package Microsoft.Quantum.ProjectTemplates not found Hello, I'm trying to install QKD for developing in C# and command line. So I installed the .NET Core SDK version 3.1.100 x64, then I tried to install templates with command line : dotnet new -i Microsoft.Quantum.ProjectTemplates But this command line give me an error : C:\Users\Julien.templateengine\dotnetcli\v3.1.100\scratch\restore.csproj : error NU1101: Package Microsoft.Quantum.ProjectTemplates not found. no package associated to this ID exists in ressources : C:\Program Files\dotnet\sdk\NuGetFallbackFolder Restoration failed 195,48 ms for C:\Users\Julien.templateengine\dotnetcli\v3.1.100\scratch\restore.csproj. (I have translated it from french). My computer is Windows 10 x64. I made exactly same steps on an other computer (Windows 10 x64 also with same .NET SDK version) and it works well and I'm able to run the "hello world" Q# sample. I suspect the reason is because I have Visual Studio 2013 installed on the computer, do you think it could be the reason ? Any help to solve my problem ? Thank you very much ! Best regards. Julien. Thanks for your patience. Since this concerns the project templates, moved over to the qsharp-compiler repo where those are hosted. It looks like your current NuGet.Config should work, such that I'm a bit confused by the errors that you've listed. That said, that's also the default configuration for NuGet, such that I'd suggest temporarily moving your config file somewhere else and seeing if it works that way.
gharchive/issue
2020-01-16T20:55:40
2025-04-01T06:44:58.807251
{ "authors": [ "JulesMhz", "cgranade" ], "repo": "microsoft/qsharp-compiler", "url": "https://github.com/microsoft/qsharp-compiler/issues/420", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1337743198
How to execute my own queries in BERT When we use the BERT model to preprocess our own queries (already transformed by the spider). We found that the preprocessing cannot crate a new dev.jsonl file. What should we do? Hey @robinzixuan where you able to figure out the reason I am facing the same issue but for some databases not for all.
gharchive/issue
2022-08-12T22:09:13
2025-04-01T06:44:58.808466
{ "authors": [ "pmane-uptycs", "robinzixuan" ], "repo": "microsoft/rat-sql", "url": "https://github.com/microsoft/rat-sql/issues/69", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1281505931
Update CODEOWNERS Description Removing @riedgar-ms from CODEOWNERS file. Adding e2e package with @vinuthakaranth, @tongyu-microsoft and myself as owners. Checklist [ ] I have added screenshots above for all UI changes. [ ] I have added e2e tests for all UI changes. [ ] Documentation was updated if it was needed. Codecov Report Merging #1513 (68cc074) into main (46d1ab7) will decrease coverage by 11.69%. The diff coverage is n/a. @@ Coverage Diff @@ ## main #1513 +/- ## =========================================== - Coverage 87.27% 75.58% -11.70% =========================================== Files 108 10 -98 Lines 5108 258 -4850 =========================================== - Hits 4458 195 -4263 + Misses 650 63 -587 Flag Coverage Δ unittests 75.58% <ø> (-11.70%) :arrow_down: Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage Δ responsibleai/responsibleai/_managers/__init__.py responsibleai/responsibleai/_input_processing.py ...ponsibleai/responsibleai/_managers/base_manager.py raiwidgets/raiwidgets/cohort.py raiwidgets/raiwidgets/__init__.py raiwidgets/raiwidgets/fairness_dashboard.py raiutils/raiutils/common/__init__.py ...iwidgets/raiwidgets/fairness_metric_calculation.py ...iwidgets/raiwidgets/explanation_dashboard_input.py responsibleai/responsibleai/__init__.py ... and 88 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 46d1ab7...68cc074. Read the comment docs.
gharchive/pull-request
2022-06-23T01:58:59
2025-04-01T06:44:58.833279
{ "authors": [ "codecov-commenter", "romanlutz" ], "repo": "microsoft/responsible-ai-toolbox", "url": "https://github.com/microsoft/responsible-ai-toolbox/pull/1513", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1521884248
[Big data] Counterfactuals - part 1 Signed-off-by: vinutha karanth vinutha.karanth@gmail.com Description This PR contains all the other changes needed for counterfactuals other than changes in libs/counterfactuals folder. Next part will contain all changes in libs/counterfactuals folder. Checklist [ ] I have added screenshots above for all UI changes. [ ] I have added e2e tests for all UI changes. [ ] Documentation was updated if it was needed. Codecov Report Merging #1888 (83132f9) into main (a4920bf) will decrease coverage by 7.48%. The diff coverage is n/a. @@ Coverage Diff @@ ## main #1888 +/- ## ========================================== - Coverage 93.33% 85.85% -7.49% ========================================== Files 93 29 -64 Lines 4559 523 -4036 ========================================== - Hits 4255 449 -3806 + Misses 304 74 -230 Flag Coverage Δ unittests 85.85% <ø> (-7.49%) :arrow_down: Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage Δ ...onsibleai/responsibleai/serialization_utilities.py ...ibleai/_tools/shared/state_directory_management.py ...s/erroranalysis/_internal/error_report/__init__.py erroranalysis/erroranalysis/report/error_report.py ...ponsibleai/responsibleai/_tools/shared/__init__.py ...sponsibleai/responsibleai/rai_insights/__init__.py ...leai/responsibleai/modelanalysis/model_analysis.py ...bleai/responsibleai/_tools/causal/causal_config.py ...leai/responsibleai/databalanceanalysis/__init__.py .../responsibleai/modelanalysis/constants/__init__.py ... and 54 more Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
gharchive/pull-request
2023-01-06T03:58:05
2025-04-01T06:44:58.850735
{ "authors": [ "codecov-commenter", "vinuthakaranth" ], "repo": "microsoft/responsible-ai-toolbox", "url": "https://github.com/microsoft/responsible-ai-toolbox/pull/1888", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1829055453
Running the same container in AKS instead of WebApp @sarah-widder How about creating the same deployment in AKS, should a separate solution be created for that or can this repo be extended to include an AKS deployment as well? This should be not difficult to implement, given there's already a dockerfile ready with all the dependencies. I'm not completely sure about the authentication layer though, need to look closer. For now I've disabled the authentication, just to test whether it works here: https://github.com/microsoft/sample-app-aoai-chatGPT/blob/main/frontend/src/pages/chat/Chat.tsx amended lines 35-44 like that: const getUserInfoList = async () => { const userInfoList = await getUserInfo(); // if (userInfoList.length === 0 && window.location.hostname !== "127.0.0.1") { // setShowAuthMessage(true); // } // else { // setShowAuthMessage(false); // } setShowAuthMessage(false); } Now the authentication prompt doesn't appear, however the chat isn't working still due to name resolution error. It is the same error in both WebApp (which I deploy from the azure portal by clicking "Deploy to WebApp") and when I'm running it in kubernetes. The error: Error communicating with OpenAI: HTTPSConnectionPool(host='my-cognitive-account-chatgpt4.openai.azure.com', port=443): Max retries exceeded with url: //openai/deployments/gpt-4_0314/chat/completions?api-version=2023-03-15-preview (Caused by NameResolutionError("<urllib3.connection.HTTPSConnection object at 0x7f28c71e7950>: Failed to resolve 'my-cognitive-account-chatgpt4.openai.azure.com' ([Errno -2] Name does not resolve)")) Can you suggest why am I getting that? My OpenAI account is located in FranceCentral region (because this is where GPT-4 is available), the edpoint is: https://francecentral.api.cognitive.microsoft.com AZURE_OPENAI_RESOURCE: "my-cognitive-account-chatgpt4" -- the name of OpenAI (cognitive services account of type OpenAI) AZURE_OPENAI_MODEL: "gpt-4_0314" -- deployment name AZURE_OPENAI_MODEL_NAME: "gpt-4" -- model name But how does it know where this resource is deployed, which subscription and which resource group? aahh the problem is that I was missing a custom domain, thus my endpoint didn't look like customdomain.openai.azure.com Now just need to be able to setup the same authentication mechanism (azure AD), as was done in WebApp, for my application running in AKS. @pamelafox can you help with that perhaps? Ok, i've solved the authentication problem, by following these steps (a great article by the way): https://kristhecodingunicorn.com/post/k8s_nginx_oauth/#configure-nginx-ingress-controller and yeah, i really enjoy talking to myself here... :) @ealasgarov Sorry, we're playing whack-a-mole on OpenAI repository issue trackers right now. Here's a write-up of how I enabled AAD for this repo, if it helps: For sample-app-aoai-chatGPT, I automated the process of creating an app registration and protecting the app service with that app with a combination of hooks and Bicep. The hooks are declared here: https://github.com/microsoft/sample-app-aoai-chatGPT/blob/main/azure.yaml For the pre provision hook, auth_init.sh calls auth_init.py: https://github.com/microsoft/sample-app-aoai-chatGPT/blob/main/scripts/auth_init.py That script makes REST API calls to https://graph.microsoft.com/v1.0/applications in order to create a new app registration. It then sets AUTH_APP_ID, AUTH_CLIENT_ID, and AUTH_CLIENT_SECRET. For the provisioning step, AUTH_CLIENT_ID and AUTH_CLIENT_SECRET are passed in main.parameters.json: https://github.com/microsoft/sample-app-aoai-chatGPT/blob/main/infra/main.parameters.json Those parameters get passed into the appservice module here: https://github.com/microsoft/sample-app-aoai-chatGPT/blob/5b311a9f74797b771dad2b515126f9ec91a3dabe/infra/main.bicep#L96 That appservice.bicep module adds the identity provider here: https://github.com/microsoft/sample-app-aoai-chatGPT/blob/5b311a9f74797b771dad2b515126f9ec91a3dabe/infra/core/host/appservice.bicep#L103 For the post provision hook, auth_update.sh calls auth_update.py: https://github.com/microsoft/sample-app-aoai-chatGPT/blob/main/scripts/auth_update.py That code makes a REST API to update the redirect URIs for the registered application to include the deployed URL endpoint. This all works great locally! However, it doesn't work on CI/CD as the pipeline principal doesn't have the permission needed to create an application registration. Thanks for the reply Pamela, much appreciated! I guess I've now sorted everything out and things seem to work, including azure authentication, except for one problem - each time after the first successful answer, I'm getting this error when asking a 2nd question: Error Requests to the Creates a completion for the chat message Operation under Azure OpenAI API version 2023-03-15-preview have exceeded token rate limit of your current OpenAI S0 pricing tier. Please retry after 7 seconds. Please go here: https://aka.ms/oai/quotaincrease if you would like to further increase the default rate limit. I cannot replicate this on OpenAI studio Playground with the same deployment/model, there I can ask 10 questions one after another and everything works fine. Not sure why is that... But I will the open a separate issue for that one. imo it would be great to just pass the following environment variables if possible: AUTH_TENANT_ID AUTH_CLIENT_ID AUTH_CLIENT_SECRET what do you guys think? @ealasgarov I am having issue still on this, can you share your ingress files and deployment files please if possible
gharchive/issue
2023-07-31T12:12:29
2025-04-01T06:44:58.871588
{ "authors": [ "Breee", "breddy-lgamerica", "ealasgarov", "pamelafox" ], "repo": "microsoft/sample-app-aoai-chatGPT", "url": "https://github.com/microsoft/sample-app-aoai-chatGPT/issues/121", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1634164401
init and attempting to upload new package from workflow. On the step: - name: CLI init run: msstore init . -n ${{ secrets.PUBLISHER_DISPLAY_NAME }} --package --publish The workflow fails with the error message Unhandled exception: System.NotSupportedException: Cannot show selection prompt since the current terminal isn't interactive. My goal is to build and package the application and upload a submission to the dev center, and it appeared that init is the closest command to do so, though I am not sure if it selects any particular app/submission to upload the built package to. Hello @yaakovschectman. Just to make sure I have full context, for what type of application is this for? (Flutter, React Native, Electron, WinUI, etc). The init command is probably not the one you need here. You can call the init only once in your development computer. The CLI will connect with Partner Center, list all the applications in your account, and let you interactively select which application you want to configure your project. Then, it will detect the project type and properly configure your project. That command is not supposed to be called from CI/CD environments. What you want is probably a package command with the --publish option. Does that make sense? 3 steps: Init (once, in the dev box). Package (in CI/CD), it will build and generate the MSIX. Publish (in CI/CD), it will upload the MSIX to the store, to the right application that was selected in the init command. I think so. I have run msstore init locally within the repository and pushed the changes it caused to the remote, then run package and publish in the workflow. Is there something that must be done in order to create/select a submission? msstore publish produces the below error: Creating new Submission 💥 Error while creating submission. Please try again. Could not retrieve submission. Please try again. Is this the very first submission of the application, or is this an update? If it is an update, what is the ApplicationCategory of this app? Important to notice that if this is the first submission of an application, it needs to happen from an interactive console, as there are many missing information for the application, such as no category selected, for example. It is an update, and it is listed as "Type: Game" Hello, just want to check with you if you've any thoughts on the error message. I'm really not sure how to debug this one. Thanks. It took me a long time to find this issue, but I believe we've fixed it. Will ship a new version shortly. Ok, new release is out on GitHub Actions. Just run your build again, as it should pick up the latest version now. Thank you for the update. I am currently running into the below error upon the publish command: │ InvalidParameterValue │ You don’t have packages that support all of the │ │ │ device families you’ve selected in Device family │ │ │ availability. Upload a package that supports each │ │ │ device family you’ve selected, or uncheck the box │ │ │ for any device families that this submission doesn’t │ │ │ support. Device families: Holographic | If this is creating a new submission, where is it getting the selected available families from? On the MS dev center, it looks like one can only select device families on a submission that's already started. Is there a CLI argument to specify the device families? This is a separate issue that we are aware of, and we are already fixing it. The CLI right now uses the same families for every type of project, but this should not be the case. It should use what previous submissions are using and have default per project type (UWP vs Win32 implies some of these, so different defaults). If the submission is failed, you should be able to just change it and re-submit, without the CLI. @yaakovschectman v0.1.23 should have fixed this issue. Let me know if you are still having problems.
gharchive/issue
2023-03-21T15:27:59
2025-04-01T06:44:58.881350
{ "authors": [ "azchohfi", "yaakovschectman" ], "repo": "microsoft/setup-msstore-cli", "url": "https://github.com/microsoft/setup-msstore-cli/issues/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1067679350
ANSI Escape Sequence Handling Issue Windows Terminal version 1.12.211020001-release1.12 Windows build number 10.0.19041.0 Other Software julia version 1.6 and above Steps to reproduce Launch julia inside of Windows Terminal Expected Behavior Launching julia outside of Windows Terminal with Legacy Console Mode enabled: Windows Terminal should render the ANSI escapes as above. Actual Behavior When julia is launched in Windows Terminal, ANSI escapes are printed directly to the screen. In Windows Terminal, launched from pwsh: In Windows Terminal, launched from pwsh (julia --color=no): In Windows Terminal, launched as a new tab: In Windows Terminal, launched as a new tab (julia --color=no): Launched directly, Legacy Console Mode disabled: (glad you made it from stackoverflow ☺️) HUH. I wonder if they're setting the VT reg key. Can you export the contents of HKCU\Console to a .reg file, rename it to .txt, and share it here? I'm guessing that there's a VirtualTerminalLevel that's set to something other than 0, and they're using that to manually enable VT support, rather than using SetConsoleMode Here's the output with a few non-relevant applications removed from the list. console.txt You know what, I'm smarter than that. When julia is working correctly, are you launching it from a shortcut? Something like julia.lnk, either on the desktop, or the Taskbar, or the Start Menu? There are two places that a console app can have VirtualTerminalLevel set - one is in the registry, the other is in the lnk. Though, I didn't.... (thought clipped in favor of another theory) Presumably, using the run dialog (win+r) to launch v:\PortableApps\julia\bin\julia.exe would also not work. https://github.com/microsoft/terminal/blob/284257a38392c85a2f7d7ac7e77e20a05242ec64/src/host/srvinit.cpp#L194-L201 We should be enabling virtual terminal processing for Terminal client apps, always. That implies that julia is actually _manually disabling it?! https://github.com/JuliaLang/julia/blob/f9bb6f8fa8c66ca00f215cbbc4e4bc5addc43156/stdlib/REPL/src/LineEdit.jl#L1375 That's the only SetConsoleMode reference in julia. Maybe there's something else I'm missing. x-ref: https://github.com/JuliaLang/julia/issues/43273 I don't launch it through a link. In Windows Terminal, the settings are done this way: Otherwise, I use start v:\PortableApps\julia\bin\julia.exe to launch it. After the problem magically resolving itself when i moved my ~/.julia directory, I did some more digging. It looks like the issue is tied to a macro that I created to clear the screen which was called in a different macro in my startup script. replacing the macro with its content removes the ANSI character issue. Well, that's good to know. Thanks for following up!
gharchive/issue
2021-11-30T21:11:14
2025-04-01T06:44:58.893859
{ "authors": [ "andrewraddatz", "zadjii-msft" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/11848", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1068029037
Why not add session management and file transfer functions? When I connect ssh server in Windows, the best tools to do it is Mobaxterm, but It's too complex and ugly. I really like Microsoft Terminal, Why not add session management and file transfer features to it. It will be more popular. File transfer is probably best left for something like scp, a tool that's actually dedicated to it. Could you elaborate more on what you mean by "session management"? I suspect by "session management" they mean something along the lines of #1280. In apps like MobaXterm and PuTTY you can setup a session (like Windows Terminal profile) with your SSH connection parameters preconfigured, and the same concept can be used for file protocols like FTP and SFTP. I think it was decided that this would best be implemented as a plugin though.
gharchive/issue
2021-12-01T06:51:33
2025-04-01T06:44:58.896182
{ "authors": [ "b1tkeeper", "j4james", "zadjii-msft" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/11855", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1358180737
Display problems, color Settings and some interactive buttons Windows Terminal version 1.15.2282.0 Windows build number 10.0.19044.1889 Other Software No response Steps to reproduce That's how it started. There's something wrong with the interface Expected Behavior No response Actual Behavior When I configure the new theme color, the interface looks like this, and the system interaction button is not visible When I hit Close, the confirm button does the same thing Thanks for the suggestion! This is actually already being tracked by another issue on our repo - please refer to #13382 for more discussion. /dup #13382
gharchive/issue
2022-09-01T03:06:04
2025-04-01T06:44:58.899657
{ "authors": [ "nianjiuhuiyi", "zadjii-msft" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/13895", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1663193468
Add support for LNM (Line Feed/New Line Mode) Description of the new feature/enhancement When LNM is set, the linefeed controls (LF, FF, and VT) will also trigger a carriage return. When reset they, they won't. This is essentially the inverse of our existing console mode, DISABLE_NEWLINE_AUTO_RETURN. However, LNM has an additional effect on input. When set, the Return key generates both a carriage return and a line feed. When reset it only generates the carriage return. It's quite widely implemented, but I don't think VTE supports it, which suggests it's probably not that widely used. However, it is a requirement for meeting VT level 1 conformance, which is why I would like for us to support it. Proposed technical implementation details (optional) I had originally thought we could map it directly to the DISABLE_NEWLINE_AUTO_RETURN mode, since we're already using that to determine how linefeed controls are interpreted. However, when a Windows console app has DISABLE_NEWLINE_AUTO_RETURN reset (which is the equivalent of LNM being set), we don't typically want the Return key to behave differently. So my idea was this: We add a new input mode that specifically handles the LNM behavior for the Return key, which by default is disabled. And we treat the DISABLE_NEWLINE_AUTO_RETURN mode as an inverse alias for the LNM output behavior (as we already do). When those two match, we're in a valid VT state, and we can use the LNM mode to toggle them both at the same time. But when they're out of sync (which is the default state for a Windows console app), we just act as if the LNM mode is not supported, i.e. we don't respond to any attempts to change it, and DECRQM reports the mode as unknown. Does that seem like a reasonable approach to take? The current default behavior for Windows console apps (i.e. when DISABLE_NEWLINE_AUTO_RETURN is reset) is to execute both a carriage and a line feed when LF is output, and to generate just CR when the Return key is pressed. This behavior does not match either of the LNM states. When LNM is set, a Return key will generate both CR and LF, which a lot of console apps will register as a double key press. And when LNM is reset, you won't automatically get carriage returns when outputting *nix-style text content with \n line endings (I'm guessing that was the main reason the console added the auto-return functionality in the first place).
gharchive/issue
2023-04-11T20:48:06
2025-04-01T06:44:58.905550
{ "authors": [ "j4james" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/15167", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1773546809
Saving settings moves/looses focus to Startup item Windows Terminal version 1.17.11461.0 Windows build number 10.0.19045.2965 Other Software No response Steps to reproduce Click Save button in Settings dialog (for any profile, nested item) Expected Behavior Focus won't be lost/moved; current settings pane/context won't be changed/lost. Actual Behavior Focus switches to Startup item, quiting some (distant) profile settings pane, unexpectedly changing context. This is a known bug, but I couldn't find any existing issue that mentioned this. It happens because clicking save reloads the entire settings page.
gharchive/issue
2023-06-25T22:59:43
2025-04-01T06:44:58.908355
{ "authors": [ "lhecker", "lv-gh" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/15600", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
479398133
Unable to overwrite custom tabTitle from within shell Environment Windows build number: 10.0.18956.1000 Windows Terminal version (if applicable): 0.3.2171.0 Any other software? No other software relevant I believe Steps to reproduce Configure a custom tabTitle for cmd or PowerShell shells in the profiles.json file, e.g.: [...] "tabTitle" : "Windows PowerShell", [...] Open a cmd or PowerShell tab, verify the customized tabTitle is appearing. Try to set a custom title programmatically from within the shell, e.g. in PowerShell: $Host.UI.RawUI.WindowTitle = 'Custom Title :)' # OR [Console]::Title = 'Custom Title 2 :)' or in command prompt: TITLE Custom Title :) Expected behavior We should be able to programmatically overwrite the customized tabTitle set in the profiles.json file - programmatically setting the tab title DOES work when no custom tabTitle is specified in profiles.json. I believe the order of precedence should be, from lowest to highest: Default title / executable path Gets overwritten by tabTitle property in profiles.json Gets overwritten by scripts or programs setting their own titles Actual behavior When a customized tabTitle is specified, it cannot be overwriten or changed from within the shell. tabTitle is an override. If you don’t want the title to be overridden, don’t use it. If you’re already planning on using the shell to set the title, you do not need tabTitle. :smile: There is an open pull request that offers another better option for titling. I disagree. I would say tabTitle is a customization preference that's used to have something nicer than a filesystem path by default. However, the applications running within the shell have better context awareness than the hosting Terminal application - and command-line applications commonly set their own "window" titles, so this should take precedence. I must re-state that I feel this is a bug. Cosnider this: When I have not customized my tabTitle and I start vim from a command prompt, my tab title is changed to [Unnamed] - VIM, as the application author intended. This information is also VERY helpful to me the user, who can not tell my VIM tab apart from my 4 other command prompt tabs. With a custom tabTitle set, this feature is completely taken away. You say tabTitle is an override, but it shouldn't be. It should be a default value, modifyable at any point by any application. Please consider re-opening the issue, or introducing a defaultTabTitle property that is not a hard override. You should follow #2373. There are a few issues littering this repository explaining why things are the way they are, and why 2373 is the correct fix. That should ameliorate most of your concerns here. Thanks, I had since found that issue after you illuded to an open PR regarding this. I will follow it, and it is completely fine to close this as "duplicate / close enough duplicate", it was just your original statement that tabTitle should be a set-in-stone, static, non-context-aware override that I took issue with. And as I saw now, I'm not alone. cheers
gharchive/issue
2019-08-11T18:40:32
2025-04-01T06:44:58.917173
{ "authors": [ "DHowett-MSFT", "jantari" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/2393", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
355484211
Font that was once rendering correctly, now shows empty boxes I am on Windows 10 Pro 1803 (17134.228) on a Dell XPS 15 9570 with the 4K screen (not sure if this matters). I have been using "Ubuntu mono derivative Powerline" without issues on WSL. Yesterday I noticed that my zsh prompt no longer renders the font correctly; it now shows empty boxes instead of >>> (three arrows). Here is what it used to look like: I'm really not sure what triggered this; here is my Windows Update log: Hey @burhan - in the erroneous case, what font is your Console using? The same font, I haven't changed the console settings. On Thu, Aug 30, 2018, 7:07 PM Rich Turner notifications@github.com wrote: Hey @burhan https://github.com/burhan - in the erroneous case, what font is your Console using? — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/Microsoft/console/issues/243#issuecomment-417374775, or mute the thread https://github.com/notifications/unsubscribe-auth/AAkz6NeU-xqShgtZjZOl1GmeVvQlVKwNks5uWA3BgaJpZM4WTCRS . -- -- Burhan Khalid Sent from a mobile device @Burhan: And do you have a link to the font you're using? The font is installed as per the properties window. The font installed is the same one as in your link @bitcrazed I am not sure if this is related, but in VS code, the font renders correctly Could you share your zsh configuration declaring the color chevrons? ${SSH_TTY:+"%F{9}%n%f%F{7}@%f%F{3}%m%f "}%F{4}${_prompt_sorin_pwd}%(!. %B%F{1}#%f%b.)${editor_info[keymap]} And what does your .~/zshrc look like? @burhan, The fonts in provided VS Code and cmd.exe screenshots look different to the naked eye. While the one in cmd.exe definitely looks like Ubuntu Mono, the VS Code one most certainly doesn't. I wonder if this is the same problem as #3257? There's a lot more discussion there and a potential workaround, selecting "Install for all users" on the font may get it working properly. I wonder if this is the same problem as #3257? There's a lot more discussion there and a potential workaround, selecting "Install for all users" on the font may get it working properly. Hey so this was filed like, four years ago, and obviously there's been a lot of work done since then. Is this still happening? I think we added some better support for unsupported glyphs in the console in the latest Win11 bits. Then there's also the Terminal, which may have never had this issue at all. We never did nail down a specific repro for this one, so maybe it's gone/?
gharchive/issue
2018-08-30T08:45:01
2025-04-01T06:44:58.927703
{ "authors": [ "bitcrazed", "burhan", "mikedld", "qidydl", "zadjii-msft" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/243", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
586118816
Weird behavior after changing cursor shape in vim Environment Windows build number: 10.0.18363.0 Windows Terminal version (if applicable): 0.10.761.0 Steps to reproduce Launch vim from either WSL or SSH of a Linux host Set cursor shape options: blinking block in normal mode, blinking vertical bar in insert mode let &t_SI .= "\<Esc>[5 q" let &t_EI .= "\<Esc>[0 q" The shape does change as what we expect, but I observed weird behaviors after that. Expected behavior When exiting insert mode, i.e. press i then <ESC>, the cursor stays at where it was if I didn't move it. When exiting insert mode, the cursor shape should change as soon as <ESC> pressed. Restore the default shape after exiting vim Actual behavior Press i then <ESC>, the cursor moves ahead one character. The shape didn't change until the next blink. The cursor stays like a block after exiting. You can find an explanation of why the cursor moves backwards on the VIM stack exchange site. This is expected behaviour, and I see the same behaviour in other terminals emulators too. The delayed shape changing is because an ESC character can also be the start of a character sequence representing another key, so VIM needs to wait and see if there are more characters following before it can be certain you've actually pressed Esc. Again, this is expected behaviour. DECSCUSR 0 is officially defined as the block cursor so that's why you get a block on exit. Some terminals do interpret that as the user's preferred cursor, though, and it's possible we may support that one day too (see issue #1604) Thanks for the comprehensive writeup, James. :smile:
gharchive/issue
2020-03-23T10:57:42
2025-04-01T06:44:58.934710
{ "authors": [ "DHowett-MSFT", "dianlujitao", "j4james" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/5084", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
707882172
startup commands from command line for wt Ability to set startup commands from command line startup. What I want is to be able to set startup command of the terminal tab/pane from a command line option for wt. Something like "--startup" option. There is an example of the usage down below. Proposed technical implementation details Example is for settings an environmental variable for git bash. Each pane in each tab gets its own prefix group name preceeded by the path. Open wt and 2 split panes with git bash and just execute TITLEPREFIX=G1 in each pane and see for yourself. :) But basically it is for executing a command initially for each tab/pane. example usage: (imagine as single line, switched to multiline for readability) wt -d {PATH1} --startup "TITLEPREFIX={Group1}" ; split-pane -V -d {PATH2} --startup "TITLEPREFIX={Group1}" ; split-pane -V -d {PATH3} --startup "TITLEPREFIX={Group1}"; new-tab -d {PATH4} --startup "TITLEPREFIX={Group2}"; split-pane -V -d {PATH5} --startup "TITLEPREFIX={Group2}" ; new-tab --title web -d {PATH6} --startup "TITLEPREFIX={Group3}"; split-pane -V -d {PATH7} --startup "TITLEPREFIX={Group3}"; new-tab --title mobile -d {PATH8} --startup "TITLEPREFIX={Group4}"; focus-tab -t 0 Maybe #6776 ? yes, seems like #5528
gharchive/issue
2020-09-24T06:04:54
2025-04-01T06:44:58.939474
{ "authors": [ "alper-batioglu", "skyline75489" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/7720", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
818009122
Opening terminal and opening a second window then closing the first window with ctrl+shift+w freezes terminal Environment Windows build number: Microsoft Windows [Version 10.0.19042.844] Windows Terminal version: 1.5.10411.0 Steps to reproduce Open up a new fresh terminal instance Open up a new tab (you now have a total of two tabs open) Switch to the previous tab (the one you started out with when you opened up the terminal) Close the tab by using the default keybinding to close a pane which is ctrl+shift+w (see this for reference https://docs.microsoft.com/en-us/windows/terminal/customize-settings/actions#close-pane) Expected behavior The previous tab (the first tab) closes and focus goes to the recently second opened tab. Actual behavior The whole terminal freezes and I have to either close the terminal by using alt+F4 or terminate the process in task manager. Notes This issue seems closely related to this one https://github.com/microsoft/terminal/issues/5799 but it still doesn't work for me. Well. That's certainly weird. You're not using anything like focus mode, you running the Terminal full-screen? Could you share your settings? Are you using any sort of accessibility software (Narrator, NVDA)? I'm having a hard time reproing this, but I also only have 1.6+ builds installed 😕 Ah, you right! This issue only applies to when the terminal is in focus mode (not full-screen nor default mode). I've updated the issue with the new information. Here is my settings: { "$schema": "https://aka.ms/terminal-profiles-schema", "defaultProfile": "{2c4de342-38b7-51cf-b940-2309a097f518}", "copyOnSelect": false, "copyFormatting": false, "launchMode": "focus", "profiles": { "defaults": { "useAcrylic": true, "acrylicOpacity": 0.8, "fontSize": 9, "cursorShape": "filledBox" }, "list": [ { "guid": "{2c4de342-38b7-51cf-b940-2309a097f518}", "hidden": false, "name": "Ubuntu", "source": "Windows.Terminal.Wsl", "startingDirectory" : "//wsl$/Ubuntu/home/anders" }, { "guid": "{0caa0dad-35be-5f56-a8ff-afceeeaa6101}", "name": "Command Prompt", "commandline": "cmd.exe", "hidden": false }, { "guid": "{61c54bbd-c2c6-5271-96e7-009a87ff44bf}", "name": "Windows PowerShell", "commandline": "powershell.exe", "hidden": false }, { "guid": "{b453ae62-4e3d-5e58-b989-0a998ec441b8}", "hidden": false, "name": "Azure Cloud Shell", "source": "Windows.Terminal.Azure" } ] }, "schemes": [], "actions": [ { "command": "find", "keys": "ctrl+shift+f" }, { "command": { "action": "splitPane", "split": "auto", "splitMode": "duplicate" }, "keys": "alt+shift+d" }, { "command": "toggleFocusMode", "keys": "shift+f11" } ] } @zadjii-msft - wasn't it solved in 1.6 by https://github.com/microsoft/terminal/pull/8549? I've just tested it in 1.6 and it works! Should've updated!
gharchive/issue
2021-02-27T20:13:21
2025-04-01T06:44:58.946618
{ "authors": [ "AndysonDK", "Don-Vito", "zadjii-msft" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/9306", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
867077001
The WT can't create a new tab based on the parent root when i onpen from the right click menu Description of the new feature/enhancement The WT can't create a new tab based on the parent root when i onpen from the right click menu. After update the new version, a content is created on my right click menu,which named "Open in windows terminal ".When I click it from "D:/myfile:" ,although the first page is base on my folder root which is "D:/myfile", the root of new tab is "c:/windows/system32" again. It is unconvenient!! WISH TO BE SOLVED!! Proposed technical implementation details (optional) Thanks for the suggestion! This is actually already being tracked by another issue on our repo - please refer to #8933 for more discussion. /dup #8933
gharchive/issue
2021-04-25T18:15:14
2025-04-01T06:44:58.949316
{ "authors": [ "Michaelzhouisnotwhite", "zadjii-msft" ], "repo": "microsoft/terminal", "url": "https://github.com/microsoft/terminal/issues/9947", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
391455941
TFS specific TestContext variables are not present in Azure DevOps when executing TestCase from TestPlan Description I used to use variables populated by TestAdapter in TextContext (during TestInitialize, i believe) when executing tests from Test Plan. These are: Tfs_TestPlanId Tfs_TestCaseId Tfs_TestPointId Tfs_TestConfigurationId Tfs_TestConfigurationName Tfs_TfsServerCollectionUrl Tfs_TeamProject Recently when I came back to my solution I noticed that these variables are missing and not populated when executing on Azure DevOps. I am sure that when Azure DevOps was called VSTS the variables were still populated. Executed on: Azure DevOps by: VSTest v2 Written locally using: MSTest v2 Steps to reproduce Define any TestCase using MsTest v2 Created TestPlan in Azure DevOps Connect Defined MsTest TestCase to TestCase in TestPlan (RMB + Associate To Test Case) Add some logging code to the testCase and set it to be executed after Test Case initailization to check TestContext Run TestCase from TestPlan Verify if TestContext data has been populated with: Tfs_TestPlanId Tfs_TestCaseId Tfs_TestPointId Tfs_TestConfigurationId Tfs_TestConfigurationName Tfs_TfsServerCollectionUrl Tfs_TeamProject Expected behavior Variables: Tfs_TestPlanId Tfs_TestCaseId Tfs_TestPointId Tfs_TestConfigurationId Tfs_TestConfigurationName Tfs_TfsServerCollectionUrl Tfs_TeamProject Should be present in TestContext after testCase initialization Actual behavior Tfs_TestPlanId Tfs_TestCaseId Tfs_TestPointId Tfs_TestConfigurationId Tfs_TestConfigurationName Tfs_TfsServerCollectionUrl Tfs_TeamProject Are missing in TestContext Environment Executed on: Azure DevOps (Azure DevOps TestPlan) by: VSTest v2 Written locally using: MSTest v2 @borsooq : Which version of MSTest nuget packages are you using? This functionality was added in 1.4.0 version of MSTest.TestAdapter and MSTest.TestFramework. Please try with latest packages and see it that works for you! Hello, it is MSTest V2 1.3.2. Seems like we have the reason :) Will check that out. I can see that property TCMTestPropertiesJSONFile has been added to TestContext.Properties. The property contains JSON string that can be deserialized. I used that and it does the job :) @borsooq I am facing the same issue. Could you please explain which property did you use to get the TCM Test data. I don't see TCMTestPropertiesJSONFile in the TextContext.Properties. @jayaranigarg Could you please suggest how we can use the TCMTestPropertiesJSONFile to get TCM data?
gharchive/issue
2018-12-16T10:54:20
2025-04-01T06:44:58.991163
{ "authors": [ "borsooq", "jayaranigarg", "tarenbd" ], "repo": "microsoft/testfx", "url": "https://github.com/microsoft/testfx/issues/541", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2760546251
MSTEST0037 (proper assert analyzer): Don't report for user defined equality operators Fixes #4414 I think it should. It should be the compiler responsibility to give us the correct OperatorMethod with the correct MethodKind.
gharchive/pull-request
2024-12-27T07:56:42
2025-04-01T06:44:58.992355
{ "authors": [ "Youssef1313" ], "repo": "microsoft/testfx", "url": "https://github.com/microsoft/testfx/pull/4456", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
637547176
About how to submit SROIE competition Sorry for disturbing the problem not related to this repo. About how to submit SROIE competition. I have finished my model for SROIE Task3, however, I do not know how to submit my results. here is the snapshot about my results. My results pass the auto examination and reports 'The method is not calculated yet'. So, if there some thing wrong with my reuslts? Can you help me. VERY THANKS. @persistforever Sometimes the website will return this status. Please wait or resubmit the results @persistforever hey can you please help me how you preprocess the dataset and how you map the key-value. @persistforever, I also give same issue. Did you solve this problem ? @persistforever, I also give same issue. Did you solve this problem ?
gharchive/issue
2020-06-12T07:30:08
2025-04-01T06:44:58.995310
{ "authors": [ "buiquangmanhhp1999", "kbrajwani", "persistforever", "wolfshow" ], "repo": "microsoft/unilm", "url": "https://github.com/microsoft/unilm/issues/178", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2356331995
Raise errors for malformed overlay port manifests For https://github.com/microsoft/vcpkg/issues/39276. This PR will make vcpkg no longer tolerate broken overlay port manifests. Initially meant for vcpkg ci, the change also affects commands like install and search. This is acceptable: As long as broken ports are ignored, installation plans and results of "search" must be considered invalid. Errors are raised for: malformed JSON (broken-manifests/malformed) missing name (broken-manifests/broken-no-name) missing version (broken-manifests/broken-no-version) Note that errors were already raised before this change when --overlay-ports pointed directly to a single broken port. No errors are raised for the remaining cases in ci/* (unchanged), including the test for missing dependencies of inactive features (new: ci/feature-missing-dep). This PR will make vcpkg no longer tolerate broken overlay port manifests. What exactly is the behavior here? It shouldn't break on vcpkg install overlayA if overlayB is not used but broken. This PR will make vcpkg no longer tolerate broken overlay port manifests. What exactly is the behavior here? It shouldn't break on vcpkg install overlayA if overlayB is not used but broken. vcpkg ci must fail if overlayB is broken. overlayB is vcpkg-ci-openimageio. The effect on vcpkg install overlayA is debatable. The effect on vcpkg install overlayA is debatable. In particular, you cannot determine correct dependencies of overlayA if overlayB would overlay one of the dependencies. vcpkg ci must fail I fully agree with that. In particular, you cannot determine correct dependencies of overlayA if overlayB would overlay one of the dependencies. that would mean overlayB is used. I am only interested in the case that overlayB is not used by anything and just exists. Overlays are for me working directories and erroring on incomplete and yet unused work is just annoying. A warning would be ok. Just don't straight out error for stuff which isn't required. Overlays are for me working directories and erroring on incomplete and yet unused work is just annoying. A warning would be ok. Just don't straight out error for stuff which isn't required. I understand the WIP nature of overlay port directories. However, silently ignoring (or just issueing a warning for) a potential overlay port just because @Neumann-A forgot to add a required comma isn't acceptable even for vcpkg install foo IMO. vcpkg must know the available ports to determine the installation plan. But the only hard requirement is to be able to parse the manifest for the port name and version. In a more strict sense, only the port name is needed, and it is defined by the directory name [if the directory has a manifest]. Hm, I don't think the problems exposed by the vcpkg-ci-openimageio fiasko are sufficiently understood. The ci command effectively names everything reachable, so the intended effect of exploding in the registry repo CI runs would still be achieved with this framing. vcpkg-ci-openimageio isn't used as a dependency, so its breakage went unnoticed. Unnoticed even in presence of =pass in the baseline. Do we need to see this as a depedency? It is not unreasonable, but it wouldn't make this much easier. vcpkg install foo --overlay-ports=broken\bar shouldn't explode because bar is malformed, unless it's named as a dependency of foo. The point is: ATM it will never explode "because bar is malformed". It will only explode when bar is missing. Imagine it wasn't vcpkg-ci-openimageio but an openssl hotfix. vcpkg would silently use the main repo's vulnerable openssl because the author forgot one comma in the manifest. Is this acceptable? vcpkg would silently use the main repo's vulnerable openssl because the author forgot one comma in the manifest. AS far as I know it is not silent since vcpkg prints the location where something is taken from. If it is from an external registry you would have to run x-add-version which would check the format. So the problem probably only really applies to overlays. The problem is that changing the behavior to always break independent of actual requirements/dependencies breaks potentially existing users unnecessary. So the fix needs to only break if it is asked for, meaning you have a broken openssl in the overlay and you a requesting openssl to be build, then it should stop and yell at the user that openssl is broken. Implementing this probably requires keeping a list of broken ports around and compare that against the requested stuff. The interesting cases then will be how to handle multiple overlays with different broken stuff, e.g.: --overlay-ports=broken\foo --overlay-ports=broken\bar --overlay-ports=broken\abc AS far as I know it is not silent since vcpkg prints the location where something is taken from. IMO any overlay port modification which introduces a syntax error must lead to an immediate full error. Anything else, such as using a regular registry port, counts as "silent". Yes, this change can cause interruption. That's why I tried to explicitly mention the effects in the top post. But the only "bad" consequence is that users must fix (or remove) broken manifests. It will not cause picking overlay ports which weren't pick before the change. It might even turn some users' attentition to the fact that desired overlay ports where not used. I still don't see the real benefit of broken manifest being silent ignored which justifies keeping that behavior. That's like keeping bugs unfixed because the it was always broken. Hm, I don't think the problems exposed by the vcpkg-ci-openimageio fiasko are sufficiently understood. The ci command effectively names everything reachable, so the intended effect of exploding in the registry repo CI runs would still be achieved with this framing. vcpkg-ci-openimageio isn't used as a dependency, so its breakage went unnoticed. To ci, everything is referenced. Or, at least, should be. Unnoticed even in presence of =pass in the baseline. Do we need to see this as a depedency? It is not unreasonable, but it wouldn't make the fix much easier. In particular if vcpkg eventually wants to move to "no baseline entry means =pass". IMO this is a different bug but should also be fixed. If there's =pass but, for instance, the --overlay-ports part was missing in the command line, that should still fail, and this change wouldn't fix that. vcpkg install foo --overlay-ports=broken\bar shouldn't explode because bar is malformed, unless it's named as a dependency of foo. The point is: ATM it will never explode "because bar is malformed". It will explode only when bar is missing. Do I misunderstand what this PR does then? It looks like this PR will make that explode even if the contents of that overlay-ports directory never participate in the plan whatsoever. Imagine it wasn't vcpkg-ci-openimageio but an openssl hotfix. vcpkg would silently use the main repo's vulnerable openssl because the author forgot one comma in the manifest. Is this acceptable? In that case, the name openssl participated in the plan, so a malformed openssl in an overlay-ports directory should explode. My point is that names which never participate in the plan at all should not result in failures. For ci, all names in all overlay-ports directories participate in the plan, so they should indeed all result in failures. In the interest of making sure that this horse is dead, what I mean is that, given the following: broken-overlays/openssl overlays/openssl vcpkg install --overlay-ports=overlays curl[openssl] # should succeed # should fail because a broken openssl overlay-port exists; I believe your PR fixes this vcpkg install --overlay-ports=overlays --overlay-ports=broken-overlays curl[openssl] vcpkg install --overlay-ports=broken-overlays curl[openssl] # should succeed, there's no reason to ever have loaded the openssl overlay-port # I believe your PR breaks this vcpkg install --overlay-ports=broken-overlays vcpkg-cmake # should fail because a broken openssl overlay-port was referenced vcpkg ci --overlay-ports=broken-overlays I still don't see the real benefit of broken manifest being silent ignored which justifies keeping that behavior. That's like keeping bugs unfixed because the it was always broken. I think it's reasonable to have directories in a directory passed to overlay-ports which aren't intended to be ports in the first place I think it's reasonable to have directories in a directory passed to overlay-ports which aren't intended to be ports in the first place That's still possible. They are just not allowed to have malformed vcpkg.json or CONTROL. My point is that names which never participate in the plan at all should not result in failures. My point is that vcpkg doesn't know if it participates in the plan at the time it loads the manifests. My point is that vcpkg doesn't know if it participates in the plan at the time it loads the manifests. I believe that it does know that; the dependency planner doesn't read and parse every vcpkg.json or CONTROL in the repo before doing an install. Maybe this change is fine and that tests needed to change suggests some other bug. I'm taking a look... Indeed I tested this and load_overlay_ports never gets called while running install; this change might actually already do the thing Neumann-A and I are asking for. I tested and indeed this already works. I added an explicit test that this didn't break.
gharchive/pull-request
2024-06-17T04:25:50
2025-04-01T06:44:59.019096
{ "authors": [ "BillyONeal", "Neumann-A", "dg0yt" ], "repo": "microsoft/vcpkg-tool", "url": "https://github.com/microsoft/vcpkg-tool/pull/1435", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
840506828
Error Installing CMake in Windows64 Describe the bug https://github.com/YDLIDAR/YDLidar-SDK/blob/master/doc/howto/how_to_build_and_install.md We are following the instructions above and kept getting the error: error while loading cmake, the port directory does not exist, error: failed to load port from C:\src\win64\vcpkg\portd\cmake note: uploading vcpkg by rerunning bootstrap-vcpkg may resolve this failure We made a directory called cmake inside of ports and now we are getting this error: Environment OS: Window Compiler: revision To Reproduce Steps to reproduce the behavior: ./vcpkg install cmake See error Repro code when Expected behavior successfully install and move to next step (integrate install, cc build && \ cmake .."-DCMAKE_TOOLCHAIN_FILE.....) Failure logs Additional context https://github.com/YDLIDAR/YDLidar-SDK/blob/master/doc/howto/how_to_build_and_install.md We are trying to follow these instructions to use python to interface with the X2 YDLiDar for a project. We are trying to extract the raw data from the sensor Hi @madeline-m Thanks for posting this issue. Since cmake is listed in scripts/test_port/cmake, which is different from these ports in ports/ directory. So It cannot be installed like this. You can copy cmake folder in https://github.com/microsoft/vcpkg/tree/master/scripts/test_ports to ports folder, then try to rebuild. Just to make it clear, the port cmake provided by vcpkg in test_ports/ folder is used to test these ports in vcpkg. I'm not clear about this document https://github.com/YDLIDAR/YDLidar-SDK/blob/master/doc/howto/how_to_build_and_install.md. But I think this is not the proper way to install cmake. Closing this issue since it should not a problem in vcpkg.
gharchive/issue
2021-03-25T02:50:50
2025-04-01T06:44:59.027907
{ "authors": [ "NancyLi1013", "madeline-m" ], "repo": "microsoft/vcpkg", "url": "https://github.com/microsoft/vcpkg/issues/16869", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1062108752
make libzip for mingw have error make libzip for win is OK,bug make libzip for mingw is Error! D:\vcpkg-master>vcpkg install libzip[bzip2]:x64-mingw-static Computing installation plan... The following packages will be built and installed: libzip[bzip2,core,default-aes,wincrypto]:x64-mingw-static -> 1.8.0 Detecting compiler hash for triplet x64-mingw-static... Restored 0 packages from C:\Users\Administrator\AppData\Local\vcpkg\archives in 246 us. Use --debug to see more details. Starting package 1/1: libzip:x64-mingw-static Building package libzip[bzip2,core,default-aes,wincrypto]:x64-mingw-static... -- Using community triplet x64-mingw-static. This triplet configuration is not guaranteed to succeed. -- [COMMUNITY] Loading triplet configuration from: D:\vcpkg-master\triplets\community\x64-mingw-static.cmake -- Using cached nih-at-libzip-v1.8.0.tar.gz. -- Cleaning sources at D:/vcpkg-master/buildtrees/libzip/src/v1.8.0-57e7193039.clean. Use --editable to skip cleaning for the packages you specify. -- Extracting source D:/vcpkg-master/downloads/nih-at-libzip-v1.8.0.tar.gz -- Applying patch fix-dependency.patch -- Using source at D:/vcpkg-master/buildtrees/libzip/src/v1.8.0-57e7193039.clean -- Configuring x64-mingw-static -- Building x64-mingw-static-dbg -- Building x64-mingw-static-rel -- Fixing pkgconfig file: D:/vcpkg-master/packages/libzip_x64-mingw-static/lib/pkgconfig/libzip.pc -- Using cached msys-mingw-w64-i686-pkg-config-0.29.2-2-any.pkg.tar.zst. -- Using cached msys-mingw-w64-i686-libwinpthread-git-8.0.0.5906.c9a21571-1-any.pkg.tar.zst. -- Using msys root at D:/vcpkg-master/downloads/tools/msys2/aa5af7b2aa7e90e8 -- Fixing pkgconfig file: D:/vcpkg-master/packages/libzip_x64-mingw-static/debug/lib/pkgconfig/libzip.pc -- Installing: D:/vcpkg-master/packages/libzip_x64-mingw-static/share/libzip/copyright -- Performing post-build validation -- Performing post-build validation done Stored binary cache: C:\Users\Administrator\AppData\Local\vcpkg\archives\e7\e7367733ed889d42cc52d811dee7f40c1cb877ba7a817972ba201a41ad55e553.zip Installing package libzip[bzip2,core,default-aes,wincrypto]:x64-mingw-static... Elapsed time for package libzip:x64-mingw-static: 32.74 s Total elapsed time: 35.13 s The package libzip provides CMake targets: find_package(libzip CONFIG REQUIRED) target_link_libraries(main PRIVATE libzip::zip) make libzip for mingw-x64-static file size is 277kb error code win64/libzip.a(zip_crypto_win.c.obj):zip_crypto_win.c:(.text+0xeb5): undefined reference to `BCryptCloseAlgorithmProvider' This issue hasn’t been updated in 3 month, if it is still an issue, please reopen this issue.
gharchive/issue
2021-11-24T07:56:41
2025-04-01T06:44:59.036247
{ "authors": [ "JackBoosY", "Moodsky" ], "repo": "microsoft/vcpkg", "url": "https://github.com/microsoft/vcpkg/issues/21643", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1168348477
vcpkg.exe install boost:x64-mingw-static Host Environment OS: [e.g. Windows/Linux etc...] Compiler: revision To Reproduce Steps to reproduce the behavior: ./vcpkg install xxxx Failure logs -Cut and paste the appropriate build messages from the console output. -Please attach any additional failure logs mentioned in the console output. Additional context Add any other context about the problem here, such as what you have already tried to resolve the issue. We don't have enough information to solve this issue, unfortunately; if you have any more information to help us solve this, please reopen!
gharchive/issue
2022-03-14T13:05:05
2025-04-01T06:44:59.039316
{ "authors": [ "Adela0814", "codezhy" ], "repo": "microsoft/vcpkg", "url": "https://github.com/microsoft/vcpkg/issues/23545", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1416074694
[icu] build failure: Detected whitespace in root directory. Computing installation plan... The following packages will be built and installed: * icu[core]:x64-windows -> 71.1 * jasper[core,default-features,opengl]:x64-windows -> 2.0.33#6 * jsoncpp[core]:x64-windows -> 1.9.5 * libharu[core]:x64-windows -> 2.4.2 * libiconv[core]:x64-windows -> 1.17 * libjpeg-turbo[core]:x64-windows -> 2.1.4 * libogg[core]:x64-windows -> 1.3.5 * libtheora[core]:x64-windows -> 1.2.0alpha1-20170719#4 * libwebp[core,libwebpmux,nearlossless,simd,unicode]:x64-windows -> 1.2.4 * libxml2[core]:x64-windows -> 2.9.14#1 * netcdf-c[core,dap,hdf5,nczarr,netcdf-4,platform-default-features]:x64-windows -> 4.8.1#2 * nlohmann-json[core]:x64-windows -> 3.11.2 * openssl[core]:x64-windows -> 3.0.5#5 pcl[core,qt,visualization,vtk]:x64-windows -> 1.12.0#6 * pcre2[core]:x64-windows -> 10.40 * pegtl-2[core]:x64-windows -> 2.8.3#1 * pkgconf[core]:x64-windows -> 1.8.0#3 * proj[core,net,tiff]:x64-windows -> 9.1.0 * pugixml[core]:x64-windows -> 1.12.1#1 * qhull[core]:x64-windows -> 8.0.2#3 * qt5-activeqt[core]:x64-windows -> 5.15.6 * qt5-base[core]:x64-windows -> 5.15.6#1 * qt5-declarative[core]:x64-windows -> 5.15.6 * qt5-imageformats[core]:x64-windows -> 5.15.6 * qt5-svg[core]:x64-windows -> 5.15.6 * qt5-tools[core]:x64-windows -> 5.15.6 * qt5-xmlpatterns[core]:x64-windows -> 5.15.6 * sqlite3[core,tool]:x64-windows -> 3.39.2 * tiff[core,jpeg,lzma,zip]:x64-windows -> 4.4.0#1 * utfcpp[core]:x64-windows -> 3.2.1#1 * vcpkg-pkgconfig-get-modules[core]:x64-windows -> 2022-02-10#1 * vtk[core,qt]:x64-windows -> 9.0.3-pv5.9.1#12 Additional packages (*) will be modified to complete this operation. Detecting compiler hash for triplet x64-windows... Restored 0 package(s) from C:\Users\DELL\AppData\Local\vcpkg\archives in 765 us. Use --debug to see more details. Installing 1/32 icu:x64-windows... Building icu[core]:x64-windows... -- Using cached icu4c-71_1-src.tgz. -- Cleaning sources at D:/Program Files (x86)/vcpkg/vcpkg/buildtrees/icu/src/c-71_1-src-08c83f7de1.clean. Use --editable to skip cleaning for the packages you specify. -- Extracting source D:/Program Files (x86)/vcpkg/vcpkg/downloads/icu4c-71_1-src.tgz -- Applying patch disable-escapestr-tool.patch -- Applying patch remove-MD-from-configure.patch -- Applying patch fix_parallel_build_on_windows.patch -- Applying patch fix-extra.patch -- Applying patch mingw-dll-install.patch -- Applying patch disable-static-prefix.patch -- Applying patch fix-win-build.patch -- Applying patch check-autoconf-archive.patch -- Using source at D:/Program Files (x86)/vcpkg/vcpkg/buildtrees/icu/src/c-71_1-src-08c83f7de1.clean -- Found external ninja('1.10.2'). -- Getting CMake variables for x64-windows CMake Warning at scripts/cmake/vcpkg_configure_make.cmake:203 (message): Detected whitespace in root directory. Please move the path to one without whitespaces! The required tools do not handle whitespaces correctly and the build will most likely fail Call Stack (most recent call first): ports/icu/portfile.cmake:53 (vcpkg_configure_make) scripts/ports.cmake:147 (include) -- Using cached msys-gzip-1.11-1-x86_64.pkg.tar.zst. -- Using cached msys-bash-5.1.008-1-x86_64.pkg.tar.zst. -- Using cached msys-autoconf-2.71-3-any.pkg.tar.zst. -- Using cached msys-autoconf-archive-2019.01.06-1-any.pkg.tar.xz. -- Using cached msys-diffutils-3.8-2-x86_64.pkg.tar.zst. -- Using cached msys-binutils-2.37-5-x86_64.pkg.tar.zst. -- Using cached msys-libtool-2.4.6-9-x86_64.pkg.tar.xz. -- Using cached msys-file-5.41-2-x86_64.pkg.tar.zst. -- Using cached msys-zlib-1.2.11-1-x86_64.pkg.tar.xz. -- Using cached msys-libbz2-1.0.8-3-x86_64.pkg.tar.zst. -- Using cached msys-coreutils-8.32-2-x86_64.pkg.tar.zst. -- Using cached msys-grep-3.0-2-x86_64.pkg.tar.xz. -- Using cached msys-sed-4.8-2-x86_64.pkg.tar.zst. -- Using cached msys-libpcre-8.45-1-x86_64.pkg.tar.zst. -- Using cached msys-m4-1.4.19-2-x86_64.pkg.tar.zst. -- Using cached msys-automake-wrapper-11-4-any.pkg.tar.zst. -- Using cached msys-gawk-5.1.0-2-x86_64.pkg.tar.zst. -- Using cached msys-mpfr-4.1.0-1-x86_64.pkg.tar.zst. -- Using cached msys-gmp-6.2.1-1-x86_64.pkg.tar.zst. -- Using cached msys-libreadline-8.1.001-1-x86_64.pkg.tar.zst. -- Using cached msys-ncurses-6.2-2-x86_64.pkg.tar.zst. -- Using cached msys-automake1.16-1.16.3-3-any.pkg.tar.zst. -- Using cached msys-perl-5.32.1-2-x86_64.pkg.tar.zst. -- Using cached msys-libcrypt-2.1-3-x86_64.pkg.tar.zst. -- Using cached msys-pkg-config-0.29.2-4-x86_64.pkg.tar.zst. -- Using cached msys-make-4.3-3-x86_64.pkg.tar.zst. -- Using cached msys-findutils-4.8.0-1-x86_64.pkg.tar.zst. -- Using cached msys-libintl-0.21-1-x86_64.pkg.tar.zst. -- Using cached msys-libiconv-1.16-2-x86_64.pkg.tar.zst. -- Using cached msys-gcc-libs-11.2.0-3-x86_64.pkg.tar.zst. -- Using cached msys-msys2-runtime-3.2.0-8-x86_64.pkg.tar.zst. -- Using cached msys-which-2.21-4-x86_64.pkg.tar.zst. -- Using msys root at D:/Program Files (x86)/vcpkg/vcpkg/downloads/tools/msys2/16df26917335bb2d -- Generating configure for x64-windows -- Finished generating configure for x64-windows -- Using cached msys-mingw-w64-i686-pkg-config-0.29.2-3-any.pkg.tar.zst. -- Using cached msys-mingw-w64-i686-libwinpthread-git-9.0.0.6373.5be8fcd83-1-any.pkg.tar.zst. -- Using msys root at D:/Program Files (x86)/vcpkg/vcpkg/downloads/tools/msys2/9a1ec3f33446b195 -- Configuring x64-windows-dbg CMake Error at scripts/cmake/vcpkg_execute_required_process.cmake:96 (message): Command failed: "D:/Program Files (x86)/vcpkg/vcpkg/downloads/tools/msys2/16df26917335bb2d/usr/bin/bash.exe" --noprofile --norc --debug -c "V=1 CPP='compile cl.exe -E' CC='compile cl.exe' CC_FOR_BUILD='compile cl.exe' CPP_FOR_BUILD='compile cl.exe -E' CXX_FOR_BUILD='compile cl.exe' CXX='compile cl.exe' RC='windres-rc rc.exe' WINDRES='windres-rc rc.exe' AR='ar-lib lib.exe' LD='link.exe -verbose' RANLIB=':' STRIP=':' NM='dumpbin.exe -symbols -headers' DLLTOOL='link.exe -verbose -dll' CCAS=':' AS=':' ./../src/c-71_1-src-08c83f7de1.clean/source/configure --build=x86_64-pc-mingw32 \"--enable-icu-build-win\" \"--disable-samples\" \"--disable-tests\" \"--disable-layoutex\" \"ac_cv_prog_ac_ct_STRIP=:\" \"gl_cv_double_slash_root=yes\" \"ac_cv_func_memmove=yes\" \"--disable-silent-rules\" \"--verbose\" \"--enable-shared\" \"--disable-static\" \"--enable-debug\" \"--disable-release\" \"--prefix=/D/Program Files (x86)/vcpkg/vcpkg/installed/x64-windows/debug\" \"--bindir=\\${prefix}/../tools/icu/debug/bin\" \"--sbindir=\\${prefix}/../tools/icu/debug/sbin\" \"--libdir=\\${prefix}/lib\" \"--includedir=\\${prefix}/../include\" \"--datarootdir=\\${prefix}/share/icu\"" Working Directory: D:/Program Files (x86)/vcpkg/vcpkg/buildtrees/icu/x64-windows-dbg Error code: 1 See logs for more information: D:\Program Files (x86)\vcpkg\vcpkg\buildtrees\icu\config-x64-windows-dbg-config.log D:\Program Files (x86)\vcpkg\vcpkg\buildtrees\icu\config-x64-windows-dbg-out.log D:\Program Files (x86)\vcpkg\vcpkg\buildtrees\icu\config-x64-windows-dbg-err.log Call Stack (most recent call first): scripts/cmake/vcpkg_configure_make.cmake:808 (vcpkg_execute_required_process) ports/icu/portfile.cmake:53 (vcpkg_configure_make) scripts/ports.cmake:147 (include) error: building icu:x64-windows failed with: BUILD_FAILED error: Please ensure you're using the latest port files with `git pull` and `vcpkg update`. Then check for known issues at: https://github.com/microsoft/vcpkg/issues?q=is%3Aissue+is%3Aopen+in%3Atitle+icu You can submit a new issue at: https://github.com/microsoft/vcpkg/issues/new?template=report-package-build-failure.md&title=[icu]+Build+error Include '[icu] Build error' in your bug report title, the following version information in your bug description, and attach any relevant failure logs from above. vcpkg-tool version: 2022-10-12-b586c2752f75bcc3f6a243749e9a0d94d0d93ccd vcpkg-scripts version: 94ce0dab5 2022-10-19 (5 hours ago) Please use the prefilled template from D:\Program Files (x86)\vcpkg\vcpkg\installed\vcpkg\issue_body.md when reporting your issue. Please move vcpkg to a directory without spaces and try again. :) does the issue still occur? We hope your question was answered to your satisfaction; if it wasn't, you can reopen with more info.
gharchive/issue
2022-10-20T06:56:50
2025-04-01T06:44:59.043285
{ "authors": [ "FrankXie05", "munitioner" ], "repo": "microsoft/vcpkg", "url": "https://github.com/microsoft/vcpkg/issues/27344", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2426322228
[New Port Request] libhsl Library name libhsl Library description HSL. A collection of Fortran codes for large scale scientific computation. http://www.hsl.rl.ac.uk/ Source repository URL https://www.hsl.rl.ac.uk/catalogue/index.html Project homepage (if different from the source repository) https://www.hsl.rl.ac.uk/index.html Anything else that is useful to know when adding (such as optional features the library may have that should be included) No response why i request after build and install ipopt, i copied the sample cpp_example.cpp MyNLP.hpp MyNLP.cpp from vcpkg\buildtrees\coin-or-ipopt\src\30c9ada089-5859d8f5b5.clean\examples\Cpp_example. then i changed the #include: from #include "IpIpoptApplication.hpp" to #include "coin-or/IpIpoptApplication.hpp" but the program genarated runtime exception: Exception of type: DYNAMIC_LIBRARY_FAILURE in file ".././../src/30c9ada089-5859d8f5b5.clean/src/Common/IpLibraryLoader.cpp" at line 59: Exception message: Error 126 while loading DLL libhsl.dll: cannot find specified module. EXIT: Library loading failure. then i tried to type in cmd prompt: C:\Users\Huang>vcpkg search hsl The result may be outdated. Run `git pull` to get the latest results. If your port is not listed, please open an issue at and/or consider making a pull request. - https://github.com/Microsoft/vcpkg/issues the result is empty. 。 我遇到相同的问题并解决了:关键在于Ipopt包bin\ipopt-3.dll文件的大小。 我最初使用vcpkg安装了Ipopt,在VS和Clion上配置都出现该问题。 最后在Ipopt的git项目下载了Ipopt-3.14.16-win64-msvs2019-md.zip 包并配置后能正常运行,对比发现:vcpkg包中的bin\ipopt-3.dll文件大小仅几MB,而官网包下的对应文件100多MB。这就是问题原因。 I encountered the same issue and resolved it: the key lies in the size of the bin\ipopt-3.dll file in the Ipopt package. Initially, I installed Ipopt using vcpkg and encountered the issue when configuring both VS and CLion. Eventually, I downloaded the Ipopt-3.14.16-win64-msvs2019-md.zip package from the Ipopt Git project, configured it, and everything worked correctly. Upon comparison, I found that the bin\ipopt-3.dll file in the vcpkg package is only a few megabytes, whereas the corresponding file in the official package is over 100 MB. This discrepancy was the root cause of the problem. Got it ! Thank you @DuskandDawn 我遇到相同的问题并解决了:关键在于Ipopt包bin\ipopt-3.dll文件的大小。 我最初使用vcpkg安装了Ipopt,在VS和Clion上配置都出现该问题。 最后在Ipopt的git项目下载了Ipopt-3.14.16-win64-msvs2019-md.zip 包并配置后能正常运行,对比发现:vcpkg包中的bin\ipopt-3.dll文件大小仅几MB,而官网包下的对应文件100多MB。这就是问题原因。 I encountered the same issue and resolved it: the key lies in the size of the bin\ipopt-3.dll file in the Ipopt package. Initially, I installed Ipopt using vcpkg and encountered the issue when configuring both VS and CLion. Eventually, I downloaded the Ipopt-3.14.16-win64-msvs2019-md.zip package from the Ipopt Git project, configured it, and everything worked correctly. Upon comparison, I found that the bin\ipopt-3.dll file in the vcpkg package is only a few megabytes, whereas the corresponding file in the official package is over 100 MB. This discrepancy was the root cause of the problem. @DuskandDawn 我猜是官方仓库默认静态链接了 libhsl,而 vcpkg 编译时改用动态链接。我当时没试着找 vcpkg 文件夹里有没有 libhsl.dll。现在我换了机器,还没试。 I guess the official repository statically links libhsl by default, while vcpkg uses dynamic linking during compilation. I didn't try to find libhsl.dll in the vcpkg folder at that time. Now I've switched machines and haven't tried it yet.
gharchive/issue
2024-07-24T00:01:38
2025-04-01T06:44:59.053595
{ "authors": [ "DuskandDawn", "HuangDuoYan", "SKNo" ], "repo": "microsoft/vcpkg", "url": "https://github.com/microsoft/vcpkg/issues/40060", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
369799681
Obscure vcpkg/CMake error when "Visual C++ tools for CMake" are not installed I've had a bunch of libraries installed no problem, but then I tried installing openblas, and it failed on the project generation phase. vcpkg output mentioned a log file, which contained the following CMake error: Visual Studio 15 2017 could not find any instance of Visual Studio. Some googling lead me to a solution: install "Visual C++ tools for CMake" through the VS 2017 installer. Couldn't vcpkg parse the logs to detect this specific error in order to suggest a solution, to save the users some time and effort? @VioletGiraffe Yes, it could, vcpkg is open source and any PR, deemed useful, will be merged after CLA signing. I'm watching this space. Now, vcpkg will install cmake automaticlly when cmake not found.
gharchive/issue
2018-10-13T12:54:01
2025-04-01T06:44:59.055749
{ "authors": [ "JackBoosY", "VioletGiraffe", "degski" ], "repo": "microsoft/vcpkg", "url": "https://github.com/microsoft/vcpkg/issues/4470", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
538329973
How do I ... When asking a question please also include where you looked for an answer (so we can update the documentation if needed). @Tsunoda1048 , thanks for your suggestion! For the answer to the question, some documented, all docs located in https://github.com/microsoft/vcpkg/tree/master/docs. Some not, they will be tagged to 'documentation' if needed. We'd like to provide users the link to docs if there has.
gharchive/issue
2019-12-16T10:43:13
2025-04-01T06:44:59.057347
{ "authors": [ "PhoebeHui", "Tsunoda1048" ], "repo": "microsoft/vcpkg", "url": "https://github.com/microsoft/vcpkg/issues/9337", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
667124914
[cgal] Upgrade CGAL to 5.1 This PR upgrades the CGAL pkg to the version 5.1. For now it is only a test, as we are only at the beta stage. @maxGimeno: I have updated the package to v5.1. I think you can un-draft the PR. @LilyWangL: I am surprised that the CI test were fine, even if I have not changed the SHA256. Thanks for your contribution!
gharchive/pull-request
2020-07-28T14:34:28
2025-04-01T06:44:59.059209
{ "authors": [ "BillyONeal", "lrineau", "maxGimeno" ], "repo": "microsoft/vcpkg", "url": "https://github.com/microsoft/vcpkg/pull/12614", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
802768771
[V8] Update to Stable version 8.8.278.14 This change also includes a change to use the Windows SDK as reported by the WinSDK ENV variable. Describe the pull request Update to V8 8.8.278.14 What does your PR fix? Fixes # May fix #15351 Which triplets are supported/not supported? Have you updated the CI baseline? Same as before x64-windows, x64-windows-static, x86-windows, x86-windows-static, x64-linux Does your PR follow the maintainer guide? Yes. Thanks! Thanks, that's what I wanted to know :) Hi @JackBoosY , anything left to do for this merge? let me know. @ras0219-msft ping for merge this PR. @ras0219-msft Friendly poke here 😁 Depends on https://github.com/microsoft/vcpkg/pull/17341.
gharchive/pull-request
2021-02-06T19:11:34
2025-04-01T06:44:59.064073
{ "authors": [ "JackBoosY", "Kwizatz", "ras0219-msft" ], "repo": "microsoft/vcpkg", "url": "https://github.com/microsoft/vcpkg/pull/16077", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1441811576
[folly] Fix feature's find_dependency Describe the pull request What does your PR fix? Fixes #27422, add feature option for some find_dependency, this is regression from PR #25335. cc @jiayuehua You must actually control CMAKE_REQUIRE_FIND_PACKAGE_ZLIB from the portfile. I cannot see this was fixed. Sigh, sorry, I dropped the ball there. I saw the comment in the files view but not the comment back here in 'conversation' view. I super hate GitHub's code review tools sometimes :(
gharchive/pull-request
2022-11-09T10:14:19
2025-04-01T06:44:59.066452
{ "authors": [ "BillyONeal", "LilyWangLL" ], "repo": "microsoft/vcpkg", "url": "https://github.com/microsoft/vcpkg/pull/27728", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }