id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1732847625 | definitionAndBoundSpan empty on destructured parameter
ES6 has added a new syntax called object destructuring, which can be used in variable assignment and function argument list.
My request is to add support for auto-jump and finding references for parameters written in destructuring format.
Consider this:
function test({a, b}) {
console.log(a);
console.log(b);
console.log({a, b});
}
test({
a: 1,
b: 2,
});
Currently if I hold Ctrl and click on the variables in the function body, I can jump back to the variable names in the parameter list. But the reverse is not yet supported, i.e. holding Ctrl and clicking on a or b should show the references to where they are used in the function body.
Hope that this functionality can be added so that I can better see where and how many times each destructured variable is used in the function.
Our behavior here is consistent with the desugaring form:
If you want to change it you can, but I don't consider this a bug.
| gharchive/issue | 2023-05-24T03:22:39 | 2025-04-01T06:44:57.693104 | {
"authors": [
"RyanCavanaugh",
"ytxmobile98"
],
"repo": "microsoft/TypeScript",
"url": "https://github.com/microsoft/TypeScript/issues/54453",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1797519816 | not recognizing path mappings for files with a .jsw extension, even though having set "files.associations": {"*.jsw": "typescript"}
Type: Bug
opened with Jetbrains IDE, and all types and modules works properly.
create tsconfig file that has "references" and "path" property. something like
"references": [
{
"path": "sub/tsconfig.json"
}
]
add "sub/tsconfig.json" to import types
"compilerOptions": {
"typeRoots": ["../"],
"lib": ["ES2020"],
"types": ["dist/types/backend", "dist/types/node"]
}
add vscode setting
"files.associations": {"*.jsw": "javascript"}
create test.js and 'test.jsw"
import module and types
compare test.js and test.jsw
VS Code version: Code 1.79.2 (Universal) (695af097c7bd098fbf017ce3ac85e09bbc5dda06, 2023-06-14T08:58:52.392Z)
OS version: Darwin arm64 22.5.0
Modes:
System Info
Item
Value
CPUs
Apple M2 (8 x 24)
GPU Status
2d_canvas: enabledcanvas_oop_rasterization: disabled_offdirect_rendering_display_compositor: disabled_off_okgpu_compositing: enabledmetal: disabled_offmultiple_raster_threads: enabled_onopengl: enabled_onrasterization: enabledraw_draw: disabled_off_okvideo_decode: enabledvideo_encode: enabledvulkan: disabled_offwebgl: enabledwebgl2: enabledwebgpu: enabled
Load (avg)
3, 2, 2
Memory (System)
16.00GB (0.11GB free)
Process Argv
. --crash-reporter-id 9f0ac6eb-66f7-4287-8ae9-03c0d2e7316c
Screen Reader
no
VM
0%
Extensions (19)
Extension
Author (truncated)
Version
aws-toolkit-vscode
ama
1.78.0
dscodegpt
Dan
2.1.13
dart-code
Dar
3.66.0
vscode-eslint
dba
2.4.2
githistory
don
0.6.20
vscode-html-css
ecm
1.13.1
prettier-vscode
esb
9.16.0
copilot
Git
1.93.189
copilot-labs
Git
0.14.884
elixir-ls
Jak
0.15.1
git-graph
mhu
1.30.0
vscode-dotnet-runtime
ms-
1.6.0
live-server
ms-
0.4.8
remote-repositories
ms-
0.36.0
LiveServer
rit
5.7.9
vscode-icons
vsc
12.4.0
volar
Vue
1.8.2
vscode-typescript-vue-plugin
Vue
1.8.2
html-css-class-completion
Zig
1.20.0
TypeScript doesn't read VS Code's file extension settings
| gharchive/issue | 2023-06-28T01:06:49 | 2025-04-01T06:44:57.710070 | {
"authors": [
"RyanCavanaugh",
"jhlee111"
],
"repo": "microsoft/TypeScript",
"url": "https://github.com/microsoft/TypeScript/issues/54957",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
119019213 | Visual Studio: Hyperlink @see to referenced type
Hi,
In the following
// file A
module foo {
export interface Bar {}
}
// file B
module foofoo {
/** Defines options for creating an instance of @see foo.Bar */
export interface BarOptions {
}
}
it will be useful if the text referenced by @see (i.e. foo.Bar) is hyperlinked to interface Bar in file A.
+1
@sandersn @Kingwl
We've just upgraded to version 4.1 RC and this issue is not fixed for namespaces as outlined above. The issue used the old syntax module for namespace but that shouldn't change anything.
Please can you reopen? Thanks
Doube check and It seems work.
@Kingwl
The @see only works when it is on a separate line. For example,
this does not work
/** Creates an instance of @see foo.bar */
interface foo {
}
But this works
/**
* Creates an instance of
* @see foo.bar
*/
interface foo {
}
I'm not sure whether you see this as a bug, but it's certainly surprising behaviour.
I'm using
Visual Studio Community 2019 Version 16.7.7
Typescript 4.1
Also # 2 you seem to have something that looks like Code Lens (the 0 references | 0 implementations. How do you enable that?
Also # 3 I need to hold down the control key for the hyperlink to appear.
Thanks
@NoelAbrahams I'm pretty sure that JSDoc tags are only allowed at the beginning of the line. There are also inline tags like foo {@link Bar bar} baz but they require braces. A regular tag in the middle of sentence should just be interpreted as regular text.
@phaux thanks for looking that up. I looked up what appears to be the official documentation and there is no clear definition of whether @see should appear on a newline or not — although in the examples they've provided the tag does always appear on a newline.
The fault is in the specification, for lack of clarity, and implementers appear to have decided on not supporting inline @see.
@NoelAbrahams As you noted, jsdoc.app really only gives examples of one nested tag -- @link. The rest of them are always the first thing on a line.
TS mostly supports inline tags, but the implementation is quirky since it's best-effort, not really a committed feature. Specifically, an inline tag works if the line starts with a non-inline tag, or is a line following such a line:
/** @param foo @see foo.bar */
or
/** @param foo - a long description
* also @see foo.bar */
This is all down to the state machine we use for parsing, so it could be improved to also support
/** for more info @see foo.bar */
@sandersn Thanks for your PR, but it seems that currently the ts-sever in VS Code doesn't support references unless a comment is at the begin of a block / declaration sentence.
I tried such usages:
// this works
if (foo) {
/** satisfy {@link ConditionsIfFoo} */
str = getBar()
}
// this doesn't work
if (foo) { /** satisfy {@link ConditionsIfFoo} */
str = getBar()
}
// this doesn't work
if (foo)
{
}
/** all the below satisfy {@link ConditionsIfFoo} */
else if (1) {
} else if (2) {
} else {
}
While highlighting works well for all the 3 cases.
@gdh1995 can you open a new issue? That's a missing feature that needs a detailed proposal. Currently jsdoc only works when attached to a declaration of some kind. The exceptions are @typedef and @callback, so @link could work like those. However, all 3 tags have complex scope rules, and we'd have to think about how those rules would need to change for @link.
Sorry I didn't know a hint would require so many details. I'm unable to give such a proposal, so let me forget it. I'll update my code to try to make tsserver happy.
| gharchive/issue | 2015-11-26T10:52:21 | 2025-04-01T06:44:57.721393 | {
"authors": [
"Kingwl",
"NoelAbrahams",
"gdh1995",
"holdfenytolvaj",
"phaux",
"sandersn"
],
"repo": "microsoft/TypeScript",
"url": "https://github.com/microsoft/TypeScript/issues/5802",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2274318194 | [isolatedDeclarations][5.5] Adding satisfies to a const expression makes it require explicit type annotation
🔎 Search Terms
isolated declarations, satisfies, const
🕗 Version & Regression Information
This changed in commit or PR https://github.com/microsoft/TypeScript/pull/58201
⏯ Playground Link
https://tsplay.dev/wRbGEw
💻 Code
export const foo = {
a: 42
} as const satisfies Record<string, number>;
🙁 Actual behavior
Variable must have an explicit type annotation with --isolatedDeclarations.ts(9010)
🙂 Expected behavior
No errors
Additional information about the issue
No response
My impression was that satisfies may impact the contextual type of an expression and cause its type to change, such that its type may not be syntactically inferable. (Need to dig up that example again, though.)
My impression was that satisfies may impact the contextual type of an expression and cause its type to change, such that its type may not be syntactically inferable. (Need to dig up that example again, though.)
Yep I've definitely seen cases where satisfies influences the contextual type of an expression, but I wonder if the as const + direct assignment to a variable is a sufficiently well-defined scenario that we can always trivially extract the type.
It is not unfortunately:
const x = [function() { return "A"}] as const satisfies Array<() => "A">
const x = [function() { return "A"}] as const
Playground Link
I wonder if the problem boils down to it not being possible to determine whether the const inference of an expression or the satisfies inference or combination of the two leads to the more specific type, then can we defer that decision down the line, so that
export const x = [function() { return "A"}] as const satisfies Array<() => "A">
would be emitted as
export declare const x: Resolve<readonly [() => string], [() => "A"]>
Are there cases where knowing only the const inference of an expression + the satisfies type isn't sufficient to determine the contextually inferred type?
Forgive me for pushing on this, as I'm sure folks are eager to treat this as an open and shut case. We have hundreds of instances of this pattern in our code base, the majority in which are cases where satisfies wouldn't affect the contextually inferred type. We adopted the operator early on under a different a different assumption
The new satisfies operator lets us validate that the type of an expression matches some type, without changing the resulting type of that expression[0]
So while we could migrate back to our old pattern (since duplicating the structure of the expression with an explicit type is the less desirable) of doing a type test like
``tsx
function upcast(value: T): void { return value }
export const foo = {
a: 42
} as const;
upcast<Record<string, number>>(foo);
asking developers to write unidiomatic code with an abstruse explanation is an outcome I want to push against.
[0] https://devblogs.microsoft.com/typescript/announcing-typescript-4-9/#satisfies
satisfies invoking contextual typing is a necessary evil imo; you want e.g. [42] satisfies [number] to succeed, but for that you need to contextually type [42] by [number] so that it doesn't widen to an array type before it can be checked. But that in turn means that [42] ultimately gets inferred as a tuple type instead of an array type. There's no mechanism to "reverse" that after-the-fact because it actually changes which type is initially inferred for the expression.
There's a general class of "operators that affect the actual type", including but not limited to:
satisfies
... ? ... : ...
Some forms of destructuring + defualting
basically anything with widening / freshness (including Symbol, unfortunately)
Isolated declarations is designed to be syntax-only, meaning an external implementation (or TS internally) must always be able to detect when there's a problem or produce an equivalent annotation / initializer.
If we do want to allow these things, then dts emit is going to have to gain some sort of syntax or builin type to describe these behaviors, but right now, they don't really exist.
If the satisfies operator doesn't change the type of the inferred expression, then it's always safe to rewrite this into two lines:
export const foo = {
a: 42
} as const;
foo satisfies Record<string, number>;
If the satisfies operator doesn't change the type of the inferred expression, then it's always safe to rewrite this into two lines:
export const foo = {
a: 42
} as const;
foo satisfies Record<string, number>;
While visually will take some getting used to, I suppose that's a good enough solution (thank you for providing one and not just mansplaining how things work).
| gharchive/issue | 2024-05-01T22:58:14 | 2025-04-01T06:44:57.735097 | {
"authors": [
"MichaelMitchell-at",
"RyanCavanaugh",
"dragomirtitian",
"fatcerberus",
"jakebailey"
],
"repo": "microsoft/TypeScript",
"url": "https://github.com/microsoft/TypeScript/issues/58397",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
548059005 | Fix parsing nested parameter types of @callback JSDoc tag
Fixes #36101
There was a conditional statement which prohibited parsing nested parameter types when the parent jsdoc tag is a @callback in parser.ts. After thorough consideration, there seems no reason why this should be handled differently than a function() JDSoc.
Attached is a test and updated baseline. I was unable to find a way to confirm the transformed declaration AST is correct after applying this fix. Existing JSDoc conformance tests do not test the resulting declaration files eighter.
Please suggest more tests and additional testing strategies if possible.
Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.:x: yin sign nowYou have signed the CLA already but the status is still pending? Let us recheck it.
I notice, I didn't run all the tests properly before submitting this PR. This bugfix broke the test case callbackTagNamespaced. I am on it.
All problems fixed. Now it seems that the CI checks fail for no well-explained reason. This is the case for multiple new PRs. I'll put some more research into this problem later in the next week.
@yin can you regenerate and commit the baselines after your change?
This PR hasn't seen any activity for quite a while, so I'm going to close it to keep the number of open PRs manageable. Feel free to open a fresh PR or continue the discussion here.
I am working with my employer to obtain permission to work on this. Please, be patient, we have more serious problems on hands ATM.
Keep this open.
I tested this change while creating declarations for Apify SDK. That was
the reason, why w decided to push this contribution. We had to create
@typedef's to workaround, but that clutters our autogenerated Docs with
single-purpose classes.
Regards / S pozdravom
Matej Gagyi
On Thu, Apr 2, 2020 at 6:32 PM Andrew Branch notifications@github.com
wrote:
@andrewbranch commented on this pull request.
In tests/baselines/reference/callbackTagNestedParameter.js
https://github.com/microsoft/TypeScript/pull/36131#discussion_r402448948
:
+type WorksWithPeopleCallback = (person:
@param {String} person.name
@param {Number} [person.age]
+) => void;
I think you weren’t seeing this because you hadn’t yet generated
declaration files via @declaration: true.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/microsoft/TypeScript/pull/36131#discussion_r402448948,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AACRYNLANFHTRVTPVCACXBDRKS4YBANCNFSM4KFHHWGQ
.
@yin were you able to get permission to work on this? Either way, do you want to keep working on it?
Yes, I got permission. Just got tired of debugging a few months ago, so I postponed it. Do you want me to look into this?
If you're up for it, we always appreciate contributions. =) But declaration emit is a tricky part of the code to fix.
@yin do you still want to work on this right now? If not, I will close it and you can tell me later if you want to restart.
just close it for now. It turned to be much more difficult than I
anticipated. I'll come back to it maybe next quarter.
On Wed, Feb 17, 2021 at 12:07 AM Nathan Shively-Sanders <
notifications@github.com> wrote:
@yin https://github.com/yin do you still want to work on this right
now? If not, I will close it and you can tell me later if you want to
restart.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/microsoft/TypeScript/pull/36131#issuecomment-780175882,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AACRYNPAOBALOITLNXODXHLS7L3ELANCNFSM4KFHHWGQ
.
| gharchive/pull-request | 2020-01-10T12:42:19 | 2025-04-01T06:44:57.751573 | {
"authors": [
"andrewbranch",
"msftclas",
"sandersn",
"yin"
],
"repo": "microsoft/TypeScript",
"url": "https://github.com/microsoft/TypeScript/pull/36131",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
685957379 | Partially disable inference recursion tracking changes
Experiment to see if partially disabling inference recursion tracking changes from recursive conditional types PR resolves the OOM we're seeing in the RWC test suites.
@typescript-bot test this
@typescript-bot test this
@weswigham Hmm, the RWC suites are failing even when I back out the changes that could potentially consume more memory by generating more work. I'm pretty sure there's nothing left in the recursive conditional types PR that could be the cause. Wonder if something else could be?
@weswigham Also, here I ran the RWC tests before merging the PR and I'm pretty certain they passed (sadly that build has now gone away). The fact that none of this repros locally of course makes it extra hard to track down.
RWC should be fixed now~
No idea how a change merged on the timeline after your change was merged somehow affected your change's RWC run, though.
| gharchive/pull-request | 2020-08-26T02:33:14 | 2025-04-01T06:44:57.755103 | {
"authors": [
"ahejlsberg",
"weswigham"
],
"repo": "microsoft/TypeScript",
"url": "https://github.com/microsoft/TypeScript/pull/40256",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
798857937 | Error message improvements for unions with identical discriminants
This PR would fix #40934, but we will be changing the core logic used in #42556.
However, I'm opening this because I wanted to point out the sorts of ~improvements~ changes we can see though if we're willing do a little extra work to track multiple object types with identical discriminants though. That logic can possibly be incorporated into #42556.
@typescript-bot pack this
This experiment is pretty old, so I'm going to close it to reduce the number of open PRs.
| gharchive/pull-request | 2021-02-02T02:04:38 | 2025-04-01T06:44:57.756994 | {
"authors": [
"DanielRosenwasser",
"andrewbranch",
"sandersn"
],
"repo": "microsoft/TypeScript",
"url": "https://github.com/microsoft/TypeScript/pull/42598",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
918305134 | Avoid unexpected any
Fixes nothing
Fixes nothing
🤣
This isn't an any, it's an evolving let. By the time it's read at the bottom of the function, it's been assigned on all CFA branches, and has the expected type. If we were to read it before definite initialization, it'd be an implicit any, and if an incorrect type were written to it, we'd get a type error at the call to createVariableDeclarationList.
it's been assigned on all CFA branches. and has the expected type.
Okay. It's seems a expected any here.
| gharchive/pull-request | 2021-06-11T06:52:36 | 2025-04-01T06:44:57.759262 | {
"authors": [
"Kingwl",
"RyanCavanaugh",
"fatcerberus"
],
"repo": "microsoft/TypeScript",
"url": "https://github.com/microsoft/TypeScript/pull/44547",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1086059385 | feat(27615): fixAddMissingMember should work for type literal
Fixes #27615
The original issue (#27615) said "this should work with types defined in jsdoc syntax", but from looking at the code in the codefix, I'm not sure if that could go in the same PR or would be better left separate. @a-tarasyuk @sandersn what do you think?
@gabritto Thanks for the review. You are right, all actions in fixAddMissingMember don't support JSDoc. I would divide adding JSDoc support into new issue/PR.
| gharchive/pull-request | 2021-12-21T17:28:52 | 2025-04-01T06:44:57.761102 | {
"authors": [
"a-tarasyuk",
"gabritto"
],
"repo": "microsoft/TypeScript",
"url": "https://github.com/microsoft/TypeScript/pull/47212",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1345026794 | Replace eslint rulesdir with eslint-plugin-local-rules, convert eslint rules to JS
--rulesdir is deprecated in favor of plugins. Although there's no official support for a local plugin, eslint-plugin-local-rules can do what we need. This, combined with converting our rules to JS, means that we don't need any configuration in VS Code or to run the CLI, nor do we need a build step to run after clone or branch changes (leaving the diagnostics as the only remaining build step on clone/branch change).
The only gotcha here is that people working on branches between these two configs may get an error in their editor, depending on which config they have in .vscode/settings.json, since we don't check that in.
Should should be ready for review; I don't think I have any other changes.
| gharchive/pull-request | 2022-08-20T02:06:47 | 2025-04-01T06:44:57.762891 | {
"authors": [
"jakebailey"
],
"repo": "microsoft/TypeScript",
"url": "https://github.com/microsoft/TypeScript/pull/50380",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
691556225 | Persist /mnt/wsl mountpoints
Is your feature request related to a problem? Please describe.
In WSL2, I'd like to be able to share my home directory between distros. So when Ubuntu (my main distro) comes up, it adds a symlink from /mnt/wsl/homeuser/ to /home/user, and when OpenSuse comes up, it also mounts from /mnt/wsl/homeuser/ to /home/user.
Describe the solution you'd like
I understand /mnt/wsl enables the sharing between distros but unfortunately /mnt/wsl uses tmpfs, so when wsl shutsdown you lose everything that was there.
Describe alternatives you've considered
I could host the files on my main distro (Ubuntu) and share from Ubuntu to /mnt/wsl/homeuser/, and when OpenSuse comes up, it does the opposite. Unfortunately, for that to occur, Ubuntu must come up first, and must stay up, consuming memory and cpu cycles.
Additional context
Initial request started at #5177. Also related to #689 and to a possible wsl --mount described in https://github.com/MicrosoftDocs/WSL/pull/824.
Is there a workaround for this? I need to be able to persist a folder in /mnt/wsl for kubernetes/docker-for-win. Actual folder can be elsewhere, but I at least need a way to persistently mount bind in /mnt/wsl.
Maybe you have an idea for a workaround, @therealkenc ?
Use .bashrc and /etc/suders tricks, same as always for what-would-be normally a systemd thing. Analogous firing up (say) sshd et cetera at first launch.
/mnt/wsl should really only be used for mounts, both bind-mounts across WSL2 distros (the original use-case per https://github.com/microsoft/WSL/issues/4577#issuecomment-545192865), and wsl --mount, not for actually storing data. Persistent data storage outside a distro falls under #689 (per https://github.com/microsoft/WSL/issues/5177#issuecomment-685211096), but the mount points also need to be recreated on startup, for use-cases like Docker Desktop kubernetes hostPath mounts that reference other distros, see https://github.com/docker/for-win/issues/7023#issuecomment-774891325.
This introduces a minor ordering issue, as all the mounts need to be up before any distribution starts (the k8s hostPath PV use-case from https://github.com/microsoft/WSL/issues/5177), but of course the mount has no real way to know which distribution is the source of the mount (it's just a bind-mount between two mnt namespaces); it can't rely on running distribution startup scripts, as it wouldn't know that it needs to start docker-desktop last so that its mounts into other distros function correctly.
Another issue here is where do we create said data/home partition? Seems to be we would have to deal with that inside wsl. So maybe like using docker we can create a persistent drive with the wsl command and then copy HOME to it, then tell wsl to mount it to HOME?
This is a use-case for #689, per https://github.com/microsoft/WSL/issues/5177#issuecomment-685211096: A way/place to store data inside WSL but outside any particular distro's storage, so that it can be shared with all distros without needing to start a particular distro first.
Then you would just need what was originally described in this ticket: When a distro starts, before you access $HOME, you replace with with a symlink/bind-mount to wherever that block device is mounted. (Which could be /mnt/wsl, but that will depend on the way #689 is addressed, and if necessary any follow-on work to auto-mount such block devices).
if you just want to share small files such as .bashrc and other config files just store them on /mnt/c/.
I use /mnt/c/devtools (c:\devtools) for my bash profile special configs and my entire collection of bash scripts.
My distros just source /mnt/c/devtools...../.bashrc and this is also git repository I sync across multiple machines all using multiple wsl distros.
All my custom scripts reside in NTFS and are accessible by all distros and git bash and PowerShell simply because I added the bin folder to the windows env path.
I have found no performance impact at all, my distros still start and run my scripts instantly.
PS: another trick I do is make all distros use git credential helper from windows and windows composer.phar so they never need to ask for git authentication once windows has triggered a browser sign in.
| gharchive/issue | 2020-09-03T01:36:58 | 2025-04-01T06:44:57.773869 | {
"authors": [
"TBBle",
"b-hayes",
"giggio",
"stefanloerwald",
"therealkenc"
],
"repo": "microsoft/WSL",
"url": "https://github.com/microsoft/WSL/issues/5851",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
980398173 | Emojis made of multiple code points break tmux status bar rendering
Windows Build Number
Microsoft Windows [Version 10.0.19043.1165]
WSL Version
[X] WSL 2
[ ] WSL 1
Kernel Version
5.10.16.3-microsoft-standard-WSL2
Distro Version
Ubuntu 20.04
Other Software
tmux 2.6 or 3.0a
Windows Terminal 1.9.1942.0 or Ubuntu Windows Store app 2004.2021.222.0
Repro Steps
In Ubuntu command line:
tmux new-session -n 🕊️
Expected Behavior
As on native Ubuntu: tmux status bar has only one line; clicking a window's name in the status bar switches to it
Actual Behavior
tmux status bar wraps and spills onto a second line; clicking on a window's name in the status bar does nothing
Diagnostic Logs
No response
@duhowett this looks like a console issue. How can we diagnose it ?
FYI I also posted this in the Terminal GH repo and apparently this is a known issue https://github.com/microsoft/terminal/issues/11053
Is there a workaround for this in the meantime (other than just not using those characters)?
| gharchive/issue | 2021-08-26T15:49:09 | 2025-04-01T06:44:57.779037 | {
"authors": [
"OneBlue",
"armanschwarz",
"carlpaten",
"lilred"
],
"repo": "microsoft/WSL",
"url": "https://github.com/microsoft/WSL/issues/7355",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1171191520 | Consider backporting PMU support for Alder Lake
Is your feature request related to a problem? Please describe.
There is no PMU driver for the Intel Alder Lake platform. Hardware event sampling does not work with VTune or perf.
vtune -collect hotspots -knob sampling-mode=hw -knob sampling-interval=0.5 /home/trym/source/stud/tdt4186/practical2/build/release/webserver /tmp 8889 12 24
vtune: Error: Unable to perform driverless collection on this platform.
vtune: Error: Cannot enable event-based sampling collection: Architectural Performance Monitoring version is 0. Make sure the vPMU feature is enabled in your hypervisor.
root@DESKTOP-CMKEO60:~# dmesg | grep -i pmu
[ 0.177428] Performance Events: unsupported p6 CPU model 151 no PMU driver, software events only.
Describe the solution you'd like
Backport the driver or provide an alternate solution.
Additional context
AFAIK, the only workarounds are to use Hyper-V or dual boot.
https://www.phoronix.com/scan.php?page=news_item&px=Linux-5.13-Perf-Alder-Lake
https://lore.kernel.org/lkml/20210311114814.GA5829@zn.tnic/T/
cpuid.txt
Reportedly a newer kernel alone is not enough: https://github.com/microsoft/WSL/issues/4678#issuecomment-1138625875
Your cpuid report does not list the features either:
Architecture Performance Monitoring Features (0xa):
version ID = 0x0 (0)
number of counters per logical processor = 0x0 (0)
bit width of counter = 0x0 (0)
length of EBX bit vector = 0x0 (0)
core cycle event not available = false
instruction retired event not available = false
reference cycles event not available = false
last-level cache ref event not available = false
last-level cache miss event not avail = false
branch inst retired event not available = false
branch mispred retired event not avail = false
fixed counter 0 supported = false
...
fixed counter 31 supported = false
number of fixed counters = 0x0 (0)
bit width of fixed counters = 0x0 (0)
anythread deprecation = false
@benhillis
are you aware of this issue and is there maybe even a fix coming?
It's very frustrating not to have PMU support for Alderlake…
(Especially since Hyper-V supports it, as @trympet pointed out)
I've hit this today again, since I wanted to profile something under Linux.
Why is this not being fixed?!? "Plain" Hyper-V already supports PMU's with Alderlake!
I hope you can understand that this is really frustrating as a user 😞...
This is also an issue for me as well.
This is also an issue for me as well +1.
I've also asked on Twitter: https://twitter.com/clemenswasser/status/1669265762991714304
Seems like we're just being ghosted 💀, which is extremely disappointing since many require performance counters support and they already work when using Hyper-V...
@benhillis @craigloewen-msft ping, are you working on this?
I've once again looked into this and this still hasn't been fixed. In the old issue, I noticed this comment: https://github.com/microsoft/WSL/issues/4678#issuecomment-1142331647
Which seems to have documented the root of the issue pretty well. The problem is that the WSL VM hasn't activated Perfmon (the arch_perfmon feature is missing), which seems to be a hard requirement for newer Intel CPUs for performance counters to work on Linux. Instructions for enabling Perfmon are in the Hyper-V documentation. I could validate this by running the following command:
$ cpuid | grep 'performance monitor'
performance monitor support available = false
performance monitor support available = false
[...]
Sadly, we can't just call Set-VMProcessor MyVMName -Perfmon @("ipt", "pmu", "lbr", "pebs") on the WSL VM, as it seems to be hidden. I only managed to list the vm by running hcsdiag list, but it seems to absent for all hyper-v commands.
@benhillis @craigloewen-msft
Since we now know what is missing, could you please activate all Perfmon features for the WSL VMs so that perf and other software which use performance counters now finally work on newer CPUs?
| gharchive/issue | 2022-03-16T15:33:26 | 2025-04-01T06:44:57.788555 | {
"authors": [
"Trass3r",
"clemenswasser",
"samlihaha",
"trympet",
"tyler274"
],
"repo": "microsoft/WSL",
"url": "https://github.com/microsoft/WSL/issues/8155",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1560452756 | ping works in wsl2 ubuntu, but wget, apt doesn't. DNS does work.
Version
10.0.22621.1194
WSL Version
[X] WSL 2
[ ] WSL 1
Kernel Version
5.15.79.1
Distro Version
Ubunto 22.04
Other Software
No response
Repro Steps
wget https://www.microsoft.com
no response, but does translate to ip address
Expected Behavior
a result
Actual Behavior
hang
Diagnostic Logs
No response
traceroute msn.com works , wget http://msn.com just hangs. so what can effect web access but not traceroute ?
I have a problem that seems similar to this. Connectivity starts out fine but after a while things like apt, curl, get stop working, just hang. Problem is cleared for a while by shutting down WSL, logging out and back into Windows, then relaunching WSL. I do not know what triggers the internet connectivity to stop working. Like OP, I see name resolution and pings working.
Iam facing the same problem.Connectivity starts out fine but after a while things like apt, curl, get stop working.
Same here. Pings works but all conections to any repo were dead.
Same here as well. ping, traceroute work just fine but wget, curl, apt don't.
Could you please follow the steps below and attach the diagnostic logs? https://github.com/microsoft/WSL/blob/master/CONTRIBUTING.md#collect-wsl-logs-for-networking-issues
Doing these solved my problem:
Open Hyper-V Manager as administrator
Select your pc, open Virtual Switch Manager
Select WSL
Set to external network
Select the network card the traffic runs through
Then login to wsl2 terminal and configure an IP address. E.g.
sudo ip addr flush dev eth0
sudo dhclient eth0
This is where i found it: https://stackoverflow.com/a/62438375/10853017
| gharchive/issue | 2023-01-27T21:28:20 | 2025-04-01T06:44:57.796210 | {
"authors": [
"AnQueth",
"RustTurakulov",
"abdullah-bin-hasan",
"chanpreetdhanjal",
"daueee",
"kamorrissey"
],
"repo": "microsoft/WSL",
"url": "https://github.com/microsoft/WSL/issues/9550",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1562033979 | WSL2: Cisco anyconnect not connecting after running wsl --update
Hi,
I have been using wsl2 for several months now on windows 11 and my guest OS is Ubuntu (Linux 5.15.79.1-microsoft-standard-WSL2)
I recently got a prompt when starting Ubuntu console about WSL now being available on Microsoft Store and the prompt indicated I can simply run wsl --update to update to the new WSL. So all I did was to run this command. After performing this update, I can no longer connect to VPN using Cisco Anyconnect. I did not change anything else (like updating Ubuntu or Cisco Anyconnect client) Now I am getting this error:
VPN establishment capability for a remote user is disabled. A VPN connection will not be established
I have verified that I can connect to same VPN portal using both Windows and other stand alone Linux laptop, and even another Ubuntu running using Virtual box, So the VPN server is working fine for other Linux and windows clients.
WSL version: 1.0.3.0
Kernel version: 5.15.79.1
WSLg version: 1.0.47
MSRDC version: 1.2.3575
Direct3D version: 1.606.4
anyconnect-linux64-4.10.06079
Windows version: 10.0.22621.1105
I have a similar issue.
After I did the wsl --update prompt to use the Store version, my VS Code no longer see any distro dans doesn't work in WSL mode anymore.
And today after I did a Docker Desktop update to the latest version, it refused to start so I had to competely reset it; it works on Windows, but on the WSL Integration page, it also says that I don't have any WSL 2 distro installed. (which is not true, I have Ubuntu 22.04 installed and it's working fine).
So after the wsl --update, there are two different programs which say that I don't have any WSL 2 distro installed.
@frivard-coveo, What version of wsl are you running wsl --version?
What version of wsl are you running wsl --version?
❯ wsl --version
WSL version: 1.0.3.0
Kernel version: 5.15.79.1
WSLg version: 1.0.47
MSRDC version: 1.2.3575
Direct3D version: 1.606.4
DXCore version: 10.0.25131.1002-220531-1700.rs-onecore-base2-hyp
Windows version: 10.0.22621.1105
Same issue here on WSL 1.1.3.0
Just experienced the exact same issue because I set up a new laptop where I installed WSL via the store.
However, this only occurs when I enable systemd in the wsl.conf file. Whenever I disable it and start a few services manually it connects fine again.
@MirChamb3r
I tried this too by setting the systemd config in /etc/wsl.conf to false.
Unfortunately, this did not fix it for me.
Is there anything else you changed?
| gharchive/issue | 2023-01-30T09:15:46 | 2025-04-01T06:44:57.803054 | {
"authors": [
"DarkVen0m",
"davidroth",
"frivard-coveo",
"kodergeek",
"pmartincic"
],
"repo": "microsoft/WSL",
"url": "https://github.com/microsoft/WSL/issues/9561",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
845299700 | The desktop application has an authentication step which opens up new browser and has a pop up attached to it which has a checkbox.
The desktop application has an authentication step which opens up new chrome browser and has a pop up attached to it which has a checkbox.
I need to select the checkbox and click on the button to move ahead. How can this be achieved?
can you please elaborate?
The elements are not inspect able as the right click is disabled
Something like this. https://intellitect.com/selenium-chrome-csharp/
We add a static ChromeDriverService in ou assemblyInitalization clean up the chrome driver instances. Also, remembering to do chromedriver.Quit() in test cleanup prevents weird errors in downstream tests.
Hi @sahaiswat I can think of three ways to handle this case but all involve using desktop session
Attach opened browser using desktop session and then find root element by '//*' and calculator 'OK' button offsets by using paint application. Use move_to_element_with_offset method to click on detected element with coordinates.
Create a desktop session and perform the steps in #1 without attaching it to the application. (It will be slow)
Create a desktop session and find element using an image. You might have to create your own implementation as winappdriver natively doesn't support this feature.
| gharchive/issue | 2021-03-30T21:28:44 | 2025-04-01T06:44:57.811521 | {
"authors": [
"liljohnak",
"sahaiswat",
"shoaibmansoor"
],
"repo": "microsoft/WinAppDriver",
"url": "https://github.com/microsoft/WinAppDriver/issues/1488",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1456498012 | SystemEvents.UserPreferenceChanged cannot be triggered
Describe the bug
I tried to attach an event handler to SystemEvents.UserPreferenceChanged. However, it seems that the event is not triggered for the WinUI 3 desktop app using the Windows App SDK.
Steps to reproduce the bug
You may just create a simple project with a main window and a main page on it. For the page_OnLoaded event, add:
using Microsoft.Win32;
# OnLoaded
SystemEvents.UserPreferenceChanged += SystemEventsOnUserPreferenceChanged;
private void SystemEventsOnUserPreferenceChanged(object sender, UserPreferenceChangedEventArgs e)
{
throw new NotImplementedException();
}
# OnUnloaded
SystemEvents.UserPreferenceChanged -= SystemEventsOnUserPreferenceChanged;
Expected behavior
No response
Screenshots
No response
NuGet package version
Windows App SDK 1.2.1: 1.2.221116.1
Packaging type
Packaged (MSIX)
Windows version
Windows 11 version 22H2 (22621, 2022 Update)
IDE
Visual Studio 2022
Additional context
No response
A workaround is to handle WM_SETTINGCHANGE
(tested on Windows 10 21H1)
A workaround is to handle WM_SETTINGCHANGE (tested on Windows 10 21H1)
Could you show me your code?
With SetWindowSubclass, like in some samples I posted (MainWindow.xaml.cs for declarations)
on main window handle hWnd :
hWnd = WinRT.Interop.WindowNative.GetWindowHandle(this);
SubClassDelegate = new SUBCLASSPROC(WindowSubClass);
bool bReturn = SetWindowSubclass(hWnd, SubClassDelegate, 0, 0);
with (I tested by changing System colors and I get : "Settings = ImmersiveColorSet") :
public const int WM_WININICHANGE = 0x001A;
public const int WM_SETTINGCHANGE = WM_WININICHANGE;
private int WindowSubClass(IntPtr hWnd, uint uMsg, IntPtr wParam, IntPtr lParam, IntPtr uIdSubclass, uint dwRefData)
{
switch (uMsg)
{
case WM_SETTINGCHANGE:
{
// Console.Beep(6000, 10);
string sText = Marshal.PtrToStringUni(lParam);
System.Diagnostics.Debug.WriteLine(string.Format("Settings = {0}", sText));
}
break;
}
return DefSubclassProc(hWnd, uMsg, wParam, lParam);
}
The events of SystemEvents class cannot be raised because when the SystemEventInvokeInfo object captures the SynchronizationContext of the current thread, the SynchronizationContext.Send is called when RaiseEvent, and DispatcherQueueSynchronizationContext has not implemented the Send method.
Solution:
SystemEvents.InvokeOnEventsThread(() =>
{
// No need to set, the default is null
//SynchronizationContext.SetSynchronizationContext(null);
SystemEvents.DisplaySettingsChanged += (s, a) =>
{
Debug.WriteLine("DisplaySettingsChanged");
};
});
| gharchive/issue | 2022-11-19T14:10:38 | 2025-04-01T06:44:57.819477 | {
"authors": [
"ArvinZJC",
"castorix",
"cnbluefire"
],
"repo": "microsoft/WindowsAppSDK",
"url": "https://github.com/microsoft/WindowsAppSDK/issues/3158",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
978446229 | remove MRTCore .net sdk version
Remove the specific .net sdk version requirement for MRTCore. 5.0.302+ is probably installed for most of the people (if not all). It would be annoying you have to keep a version of 5.0.302 to build MRTCore locally, and I don't have enough space on C drive :). Build pipeline will do a version check based on .\build\versions.props.
/azp run
| gharchive/pull-request | 2021-08-24T20:11:56 | 2025-04-01T06:44:57.821288 | {
"authors": [
"huichen123"
],
"repo": "microsoft/WindowsAppSDK",
"url": "https://github.com/microsoft/WindowsAppSDK/pull/1307",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
852973066 | Questions about model evaluation
I have two questions regarding model evaluation:
Why is the frame to be evaluated (x) multiplied by 100 during call to modelwork function? isn't it affecting data normalization?
Is it supposed the data must be normalized before the call to sr_cnn_eval?
thank you in advance.
I realized that gen_set class makes the same magnitude scaling to the data during training.
| gharchive/issue | 2021-04-08T01:39:24 | 2025-04-01T06:44:57.832999 | {
"authors": [
"jdariasl"
],
"repo": "microsoft/anomalydetector",
"url": "https://github.com/microsoft/anomalydetector/issues/34",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
615770596 | Azure Devops Pipeline error - 'Unknown argument --async'
Adding the command --async works fine locally allowing me to add --async true to stop listening for results once all the files have been passed to appcenter.
Adding the same command to the Azure pipeline i get this error - Unknown argument --async
- task: AppCenterTest@1
inputs:
appFile: '$(Build.ArtifactStagingDirectory)\AndroidBuild\App.Android.apk'
artifactsDirectory: '$(System.ArtifactsDirectory)\AppCenterTest'
frameworkOption: 'uitest'
uiTestBuildDirectory: '$(Build.ArtifactStagingDirectory)\TestStuff'
uiTestToolsDirectory: '$(Build.ArtifactStagingDirectory)\TestStuff'
prepareOptions: '--include-category $(TestInclude) --exclude-category $(TestExclude) --async $(TestAsync)'
credentialsOption: 'serviceEndpoint'
serverEndpoint: '*****'
appSlug: '*******'
devices: '$(Devices)'
series: '$(TestSeries)'
localeOption: 'en_GB'
This is needed as when all the UI tests run the duration is longer than 60m and the hosted pipeline timesout after 60m which shows it as a failure, it also means and entire pipeline is busy for 60m for no reason.
Wouldn't it be possible to use Do not wait for test result instead?
Closing as stale.
| gharchive/issue | 2020-05-11T10:23:07 | 2025-04-01T06:44:57.839483 | {
"authors": [
"DmitriyKirakosyan",
"IeuanWalker",
"Oddj0b"
],
"repo": "microsoft/appcenter",
"url": "https://github.com/microsoft/appcenter/issues/1852",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
479091881 | JUnit 5 for Android
Describe the solution you'd like
Please add support for JUnit 5 for Android.
Describe alternatives you've considered
None. I have JUnit 5 tests, and they cannot be run on AppCenter.
Thanks @humblehacker for the request! @Oddj0b would you mind taking a look please?
| gharchive/issue | 2019-08-09T17:28:40 | 2025-04-01T06:44:57.841056 | {
"authors": [
"amchew",
"humblehacker"
],
"repo": "microsoft/appcenter",
"url": "https://github.com/microsoft/appcenter/issues/859",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2565109036 | Add support for context caching
What feature would you like to be added?
Certain APIs have support for context/prompt/prefix caching, notably Gemini and Claude as well as any local LLM. OpenAI also started supporting this not long ago. This enables calls that condition on the same prefix (chat history) to be faster as they cache the context. However, I am not sure if we can have caching infrastructure be independent of the API provider or rather support caching for each provider.
Why is this needed?
Speed up inference time
@husseinmozannar Moved your issue here.
| gharchive/issue | 2024-10-03T22:38:20 | 2025-04-01T06:44:57.842995 | {
"authors": [
"ekzhu"
],
"repo": "microsoft/autogen",
"url": "https://github.com/microsoft/autogen/issues/3636",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
519072735 | Update ubuntu-latest to Zulu 13
Azul Zulu OpenJDK 13 has been available since September, but tonight I learned that it still isn't pre-installed in the Microsoft-hosted ubuntu-latest pool (at least not with a path that follows the pattern of previous versions):
/home/vsts/work/_temp/bf6bce63-30e5-48f2-abfa-2e76f8006e04.sh: line 1: /usr/lib/jvm/zulu-13-azure-amd64/bin/java: No such file or directory
https://github.com/Pr0methean/BetterRandom/blob/master/azure.yml partially works around this issue, since it uses AdoptOpenJDK's JDK13 build with OpenJ9 rather than another JDK12 build in the matrix. (OpenJ9 isn't available through Zulu anyway, so the jobs download it rather than using pre-installed copies.) But this would probably be much harder in a build system that didn't respect the JAVA_HOME environment variable consistently as Maven does.
@Pr0methean - This issue is with the image of the host running the agent, not with the agent code itself. Can you open this issue on this project https://github.com/microsoft/azure-pipelines-image-generation ? Unfortunately, I do not have a way of transferring it for you.
| gharchive/issue | 2019-11-07T06:40:21 | 2025-04-01T06:44:57.848725 | {
"authors": [
"Pr0methean",
"jtpetty"
],
"repo": "microsoft/azure-pipelines-agent",
"url": "https://github.com/microsoft/azure-pipelines-agent/issues/2582",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
521822522 | Cache persist doesn't work on server 2012 R2 or 2016 (I assume) because it doesn't inclued tar
Having issue with Tasks?
Since this is the embedded cache task I believe this belongs here.
Agent Version and Platform
Version of your agent? 2.160 on prem
OS of the machine running the agent? Windows 2012 R2
Azure DevOps Type and Version
dev.azure.com
https://dev.azure.com/gmollc
What's not working?
CacheBeta@1 isn't saving the cache because tar isn't included in 2012 R2.
Agent and Worker's Diagnostic Logs
2019-11-12T21:42:35.5033236Z ##[debug]Starting 'tar' with arguments '-h -v -cf "4092ddf38c57400cbae3ca85f9ae5178_archive.tar" -C "D:\SharedBuild2\1\s\.pipeline-cache\nuget" .'...
2019-11-12T21:42:36.0435164Z Information, ApplicationInsightsTelemetrySender correlated 1 events with X-TFS-Session 8a46dbca-e561-4dd7-ab21-92c8c86e4de1
2019-11-12T21:42:36.0512117Z ##[error]The system cannot find the file specified
2019-11-12T21:42:36.0520742Z ##[debug]Processed: ##vso[task.logissue type=error;]The system cannot find the file specified
2019-11-12T21:42:36.0521458Z ##[debug]Processed: ##vso[task.complete result=Failed;]
2019-11-12T21:42:36.1115284Z ##[debug] at System.Diagnostics.Process.StartWithCreateProcess(ProcessStartInfo startInfo)
at System.Diagnostics.Process.Start()
at Agent.Plugins.PipelineCache.TarUtils.RunProcessAsync(AgentTaskPluginExecutionContext context, ProcessStartInfo processStartInfo, Func`3 additionalTaskToExecuteWhilstRunningProcess, Action actionOnFailure, CancellationToken cancellationToken)
at Agent.Plugins.PipelineCache.TarUtils.RunProcessAsync(AgentTaskPluginExecutionContext context, ProcessStartInfo processStartInfo, Func`3 additionalTaskToExecuteWhilstRunningProcess, Action actionOnFailure, CancellationToken cancellationToken)
at Agent.Plugins.PipelineCache.TarUtils.ArchiveFilesToTarAsync(AgentTaskPluginExecutionContext context, String inputPath, CancellationToken cancellationToken)
at Agent.Plugins.PipelineCache.PipelineCacheServer.GetUploadPathAsync(ContentFormat contentFormat, AgentTaskPluginExecutionContext context, String path, CancellationToken cancellationToken)
at Agent.Plugins.PipelineCache.PipelineCacheServer.UploadAsync(AgentTaskPluginExecutionContext context, Fingerprint fingerprint, String path, CancellationToken cancellationToken, ContentFormat contentFormat)
at Agent.Plugins.PipelineCache.SavePipelineCacheV0.ProcessCommandInternalAsync(AgentTaskPluginExecutionContext context, Fingerprint fingerprint, Func`1 restoreKeysGenerator, String path, CancellationToken token)
at Agent.Plugins.PipelineCache.PipelineCacheTaskPluginBase.RunAsync(AgentTaskPluginExecutionContext context, CancellationToken token)
at Agent.Plugins.PipelineCache.SavePipelineCacheV0.RunAsync(AgentTaskPluginExecutionContext context, CancellationToken token)
at Agent.PluginHost.Program.Main(String[] args)
2019-11-12T21:42:36.1219290Z ##[section]Finishing: Cache NuGet packages
You people are going to think I'm crazy, but I'm not... Here a screenshot of the procmon trace looking for tar.exe:
Here is a screenshot of it NEVER looking for 7z.exe from the same trace:
The ONLY way I can see that happen is if somehow isWindows is false here:
https://github.com/microsoft/azure-pipelines-agent/blob/7fc04a145e5158368e8e4fea7f2d996e8a053389/src/Agent.Plugins/PipelineCache/TarUtils.cs#L190
but I don't see how that is possible!?!?!?!
@fadnavistanmay Can you take a look?
Just to add more confusion to the mix. I provisioned a new 2019 server and everthing works (same pipeline). I'm going to try a 2012 R2 new. Then I'm going to put a new agent on my non-working 2012 r2. Will update this post as I investigate.
Hi @jabbera - Glad you were able to make it work on the 2019 server. Just a quick note, the error you are getting is in the SaveCache task, which always use tar; the code snippet you have shared is during the Restore Cache task, where we look for 7z and use that if present.
No I’m having issue persisting the cache. It’s a miss so there isn’t a download.
This fails on a new 2012 R2 server as well.
@jabbera - did you explicitly installed tar on those machines, and it still fails?? You are using self-hosted agents, yes?
For now, you can fall back to old behavior by setting the variable AZP_CACHING_CONTENT_FORMAT to FileSs.
@fadnavistanmay where from? Isn't that a 2019 only feature? (Unless you are talking about the mingw tar?) Additionally why would you have the 7 zip fallback for downloading caches but not uploading them? It doesn't make much sense to me.
For whatever reason that doesn't work either. Here are snippits from my log:
[2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'ALLUSERSPROFILE': 'C:\ProgramData'
[2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'APPDATA': 'C:\Users\<SNIP>\AppData\Roaming'
[2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'AZP_CACHING_CONTENT_FORMAT': 'Files'
[2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'CommonProgramFiles': 'C:\Program Files\Common Files'
[2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'CommonProgramFiles(x86)': 'C:\Program Files (x86)\Common Files'
[2019-11-13 20:27:16Z INFO EnvironmentCapabilitiesProvider] Adding 'CommonProgramW6432': 'C:\Program Files\Common Files'
2019-11-13T20:38:27.5300788Z ##[debug]Processed: ##vso[telemetry.publish area=AzurePipelinesAgent;feature=PipelineCache]{"FileCount":"0","PlanId":"b8fde65d-718b-4f2c-8a48-b3ff33b7809f","JobId":"0ab14b9f-e499-56d5-97b1-fd98b70ea339","TaskInstanceId":"aa203359-c3b8-5e0f-0fa6-2cb7b8dcce2a","CacheResult":"Miss","ActionDurationMs":"892","ActionName":"PipelineCache.RestoreCache","ActionResult":"Success","AttemptNumber":"1","ItemCount":"0","Level":"ThirdParty","CreatedUtcNow":"2019-11-13T20:38:26.5686059Z","SentUtcNow":"2019-11-13T20:38:27.4717776Z","BaseAddress":"https://vsblobprodcus3.vsblob.visualstudio.com/Ad31b77fc-dfbf-4068-8419-a4cc92bbfac6/","X_TFS_Session":"2ebdd9b2-d83c-46a7-836f-1080713ffccc","DeploymentEnvironment":"PRODUCTION","DeploymentEnvironmentIsProduction":"True","VSOAccount":"vsblobprodcus3","OSName":"Microsoft Windows","OSVersion":"6.3.9600","FrameworkDescription":".NET Core ","ProcessName":"Agent.PluginHost","DotNetReleaseDword":"-1","Version":"18.159.29324.0 built by: master (a2f0ba0f2a)","ExceptionCount":"0"}
2019-11-13T20:38:27.5302449Z ##[debug]Starting 'tar' with arguments '-h -v -cf "15fad05c549e42999f7259172fa315b6_archive.tar" -C "D:\SharedBuild1\2\s\.pipeline-cache\nuget" .'...
2019-11-13T20:38:28.1735655Z Information, ApplicationInsightsTelemetrySender correlated 1 events with X-TFS-Session 2ebdd9b2-d83c-46a7-836f-1080713ffccc
2019-11-13T20:38:28.1814906Z ##[error]The system cannot find the file specified
2019-11-13T20:38:28.1824931Z ##[debug]Processed: ##vso[task.logissue type=error;]The system cannot find the file specified
2019-11-13T20:38:28.1825655Z ##[debug]Processed: ##vso[task.complete result=Failed;]
2019-11-13T20:38:28.1988197Z ##[debug] at System.Diagnostics.Process.StartWithCreateProcess(ProcessStartInfo startInfo)
at System.Diagnostics.Process.Start()
at Agent.Plugins.PipelineCache.TarUtils.RunProcessAsync(AgentTaskPluginExecutionContext context, ProcessStartInfo processStartInfo, Func`3 additionalTaskToExecuteWhilstRunningProcess, Action actionOnFailure, CancellationToken cancellationToken)
at Agent.Plugins.PipelineCache.TarUtils.RunProcessAsync(AgentTaskPluginExecutionContext context, ProcessStartInfo processStartInfo, Func`3 additionalTaskToExecuteWhilstRunningProcess, Action actionOnFailure, CancellationToken cancellationToken)
at Agent.Plugins.PipelineCache.TarUtils.ArchiveFilesToTarAsync(AgentTaskPluginExecutionContext context, String inputPath, CancellationToken cancellationToken)
at Agent.Plugins.PipelineCache.PipelineCacheServer.GetUploadPathAsync(ContentFormat contentFormat, AgentTaskPluginExecutionContext context, String path, CancellationToken cancellationToken)
at Agent.Plugins.PipelineCache.PipelineCacheServer.UploadAsync(AgentTaskPluginExecutionContext context, Fingerprint fingerprint, String path, CancellationToken cancellationToken, ContentFormat contentFormat)
at Agent.Plugins.PipelineCache.SavePipelineCacheV0.ProcessCommandInternalAsync(AgentTaskPluginExecutionContext context, Fingerprint fingerprint, Func`1 restoreKeysGenerator, String path, CancellationToken token)
at Agent.Plugins.PipelineCache.PipelineCacheTaskPluginBase.RunAsync(AgentTaskPluginExecutionContext context, CancellationToken token)
at Agent.Plugins.PipelineCache.SavePipelineCacheV0.RunAsync(AgentTaskPluginExecutionContext context, CancellationToken token)
at Agent.PluginHost.Program.Main(String[] args)
2019-11-13T20:38:28.2093865Z ##[section]Finishing: Cache NuGet packages
Hi @jabbera - The reason we went for tarring for upload was, the performance was almost the same for creating a "tar" - for both the tar and 7z process, but while downloading (untarring) - 7z gives much better performance.
Could you please tell us how/where are you setting the environment variable. And if you could give us the redacted logs of the entire build, that would be helpful.
Will it be possible for you. to give @johnterickson and me , permissions for https://dev.azure.com/gmollc , to investigate?
Thanks.
I’m setting the variable as a system variable.
I’ll attach the logs in the AM.
Feel free to access the org.
PS: the reason the tar thing doesn’t really work well is that tar is not an option on 2016 or 2012. IMO: You should fall back on 7 zip, and if that isn’t there fall back on files all automatically unless you don’t plan on supporting those platforms as build platforms.
Hi @jabbera - I still can't access the org. Could you please check.
Thanks.
I thought you meant on the backend. We don’t allow guest accounts in our tenant. I can open a ticket if that gives you more ability to get in. Otherwise I can do a screen share with you.
My bad. Let's check the logs first, if nothing significant comes up, we can do a screen share.
@fadnavistanmay I was able to reproduce this on my public azure devops. It really is as simple as installing the agent on a 2012 r2 server: https://dev.azure.com/mike-barry/Demo2012Issue/_build/results?buildId=420&view=results
@fadnavistanmay any chance to look at this in my public repo?
Let's make this error clearer that TAR needs to be on the path
I had this issue on Windows Server 2016. I had to install 7zip and gnuwin32 tar manually and add to the PATH. I'm running self hosted agents as a service, so I restarted the service after adding to the PATH - this fixed the problem for me (saving and restoring from the cache).
@johnterickson / @jtpetty Will the documentation be updated to include the hard dependency on these applications because at the moment there isn't anything:
https://docs.microsoft.com/en-us/azure/devops/pipelines/caching/index?view=azure-devops
I agree with @jabbera though, should this task really have these hard dependencies at all or should it install these applications within the tools cache within $(Agent.ToolsDirectory)
Hi @garfbradaz , sorry for the delayed in documentation update. We have a PR out, it looks like it haven't merged yet. I'll follow up to get the docs updated.
Docs are live: https://docs.microsoft.com/en-us/azure/devops/pipelines/caching/index?view=azure-devops#required-software-on-self-hosted-agent
Thanks @fadnavistanmay!
@fadnavistanmay Can we add C:\Program Files\Git\usr\bin to the search path for tar on Windows?
Any suggestions of how to cache the below mongodb installation would be helpfull
script: |
wget -qO - https://www.mongodb.org/static/pgp/server-3.6.asc | sudo apt-key add -
echo "deb [ arch=amd64,arm64 ] https://repo.mongodb.org/apt/ubuntu xenial/mongodb-org/3.6 multiverse" | sudo tee /etc/apt/sources.list.d/mongodb-org-3.6.list
sudo apt-get update
sudo apt-get install -y mongodb-org=3.6.16 mongodb-org-server=3.6.16 mongodb-org-shell=3.6.16 mongodb-org-mongos=3.6.16 mongodb-org-tools=3.6.16
sudo service mongod start
Any suggestions of how to cache the below mongodb installation would be helpfull
script: |
wget -qO - https://www.mongodb.org/static/pgp/server-3.6.asc | sudo apt-key add -
echo "deb [ arch=amd64,arm64 ] https://repo.mongodb.org/apt/ubuntu xenial/mongodb-org/3.6 multiverse" | sudo tee /etc/apt/sources.list.d/mongodb-org-3.6.list
sudo apt-get update
sudo apt-get install -y mongodb-org=3.6.16 mongodb-org-server=3.6.16 mongodb-org-shell=3.6.16 mongodb-org-mongos=3.6.16 mongodb-org-tools=3.6.16
sudo service mongod start
Hi @johnterickson , @b-barthel already has a PR out for "Can we add C:\Program Files\Git\usr\bin to the search path for tar on Windows?"
PR: https://github.com/b-barthel/azure-pipelines-agent/commit/6e0d9b7a64cf48c9e558b3e81f86850916ec3e2a
This has been rolled out, closing.
| gharchive/issue | 2019-11-12T21:47:36 | 2025-04-01T06:44:57.872355 | {
"authors": [
"fadnavistanmay",
"garfbradaz",
"jabbera",
"johnterickson",
"shrutyraos",
"stephenmichaelf"
],
"repo": "microsoft/azure-pipelines-agent",
"url": "https://github.com/microsoft/azure-pipelines-agent/issues/2595",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
618039522 | Feature: Share cache across pipelines
Required Information
Type: Feature
Enter Task Name: Cache (https://docs.microsoft.com/en-us/azure/devops/pipelines/tasks/utility/cache?view=azure-devops / https://github.com/microsoft/azure-pipelines-tasks/tree/master/Tasks/CacheV2)
Environment
Azure Pipelines (yml)
Can provide account / build / pipeline information privately
Agent - Hosted vs2017-win2016
Issue Description
I have a number of pipelines that run on the same hosted agent (around 50), for the same repository. All the pipelines use the same templates with lots of pipeline specific variables. I use the following cache task in a build template:
- task: Cache@2
condition: and(succeeded(), eq('${{ parameters.enableNugetCaching }}', true))
displayName: NuGet Cache
inputs:
key: 'nuget | "${{ parameters.solution }}" | "$(Agent.OS)" | ${{ parameters.nugetConfig }},**/packages.config,!**/bin/**,!**/obj/**'
restoreKeys: |
nuget | "${{ parameters.solution }}" | "$(Agent.OS)"
nuget | "${{ parameters.solution }}"
path: '${{ parameters.nugetPackagesDirectorySource }}'
Caching works perfectly when I run a single pipeline twice - it restores the expected cache, and creates a new cache if one of the matched files changes, as you would expect.
However, when I run another pipeline on the same repo / set of code, the cache generates exactly the same key, but does not "match" against the existing cache - it generates a new one. As this cache is around 600mb in size, the "pipeline cache" for the Azure DevOps organisation is now around 30Gb when it should be 600mb. I would expect extra caches to only generate when the source input files change, and the old caches to expire after 30 days (would be handy for this to be customised too, but that's not as important). It also means that all 50 pipelines take around 2 mins extra each, eating into the pipeline minutes.
I've attached a screenshot of a compare between the two cache job runs in two separate pipelines. Everything except the X-TFS-Session identifier are exactly the same. Ideally, the cache key should be shared between pipelines that run on the same repo.
In addition, if I was able to share this cache between pipelines I would be able to reduce the build by another 2 or 3 mins as I could cache the main solution binaries too - which don't change that often. It's an odd scenario, but one which suits this particular client's build requirements perfectly.
I've also posted here:
https://developercommunity.visualstudio.com/idea/1030422/share-cache-across-pipelines.html
Task logs
Cache log comparisons
I am considering taking a look into producing a PR for this, let me know if its something you're interested in.
Hi @Bidthedog - Thanks for offering to help! Unfortunately, the code changes required here are server-side. This is trickier than it might seem at first because this is an insidious attack vector.
Even if I don't have write access to a repo or it's CI build, I could go create a new pipeline that reads from that repo, but puts something "evil" in the cache. The CI build (that I don't have access to) would then read that "evil" cache entry and the build would carry forward my injected "evil" bits.
Knowing the above, one way to share artifacts between pipelines/projects is through Packages. In fact, there is a (non-official Azure DevOps but written by MSFT employees) task that acts similarly to Pipeline Caching but uses Universal Packages: https://github.com/Microsoft/azure-pipelines-artifact-caching-tasks
If you use it, you'll just have to be very careful about the permissions you have set.
OK, thank you for your response. It's not the end of the world, it just means that MUCH more cache space is taken up, and less-frequently-run builds do not take advantage of the cache; some of the pipelines - as you might imagine - are not executed regularly, so it would be handy if they used the cache when they do run. Others run multiple times per day.
Tbh, I'd much rather do this with a single pipeline, but my client's software architecture just doesn't make it feasible at present.
This pipeline/branch scoping makes cache task extremely inefficient. In my environment, CI builds unable to use cache produced by PR build (different pipelines). Literally, pipelines produce and store all this cached data for nothing.
This pipeline/branch scoping makes cache task extremely inefficient. In my environment, CI builds unable to use cache produced by PR build (different pipelines). Literally, pipelines produce and store all this cached data for nothing.
I have to agree. I've actually turned package caching off now I've moved to a self-hosted agent, because it's quicker to use the local server's cache than manage a cache per pipeline. One of my clients has 24 main pipelines (and counting) that should mostly use the same cache.
Another scenario (although, somehow similar to @gaikovoi's) in which this feature would come handy:
We build Python conda environment as part of our PR and CI builds. The operation takes around 5 minutes, but it rarely needs to be redone, as the environment stays unchanged for long time.
We can use current cache mechanism efficiently for CI builds, but not for PR build: since the cache is per build & per pipeline, and PR branches are short-lived, we end up building environment most of the times instead of using cached value. Ideally, we would like to use the environment created by CI builds for our PR build
We can't really use https://github.com/Microsoft/azure-pipelines-artifact-caching-tasks, as its documentation advices against using it for artifacts produced outside of the repo's directory (this is the case for conda environements)
Another scenario (although, somehow similar to @gaikovoi's) in which this feature would come handy:
We build Python conda environment as part of our PR and CI builds. The operation takes around 5 minutes, but it rarely needs to be redone, as the environment stays unchanged for long time.
We can use current cache mechanism efficiently for CI builds, but not for PR build: since the cache is per build & per pipeline, and PR branches are short-lived, we end up building environment most of the times instead of using cached value. Ideally, we would like to use the environment created by CI builds for our PR build
We can't really use https://github.com/Microsoft/azure-pipelines-artifact-caching-tasks, as its documentation advices against using it for artifacts produced outside of the repo's directory (this is the case for conda environements)
You can use pipeline artefacts to do this. I have one pipeline that downloads the caches every night and uploads them to Azure Pipeline Artifacts which is free (not Azure Artifacts which costs per GB), then my other pipelines download the pipeline artefact at the beginning of each run.
Don't forget if you say "Cache per Pipeline" that is wrong for me, because actually it's "cache per job per pipeline.
I can repeatable prove that running one a sequence of 3 jobs where one i and 3 is maven caching under the same key, i don't see items restored on 3 which has been produced (and cached) in 1 even although job 2 in between is taking at least 5 minutes.
So what is really the reliable definition for a minimum hit rate if its even on "same" pipeline for a dependent sequence of jobs.
This is a very useful feature if can get it in.
this is actually not stale. In a monorepo scenario with loads of pipelines using the same dependencies, this leads to dumb amounts of caching thats not shareable across pipelines..
Any solution to this would be awesome
@Squixx Azure devops is stale :/
Shouldnt be to hard to allow us to run github tasks on azure devops ;)
@Squixx You have a typo , you might have meant gitLAB - or do you want to run into same issue again in x years ;)
Just kidding
Any news on this? It would be a great feature.
| gharchive/issue | 2020-05-14T08:40:05 | 2025-04-01T06:44:57.889140 | {
"authors": [
"Bidthedog",
"Jon889",
"Squixx",
"bharathns",
"cforce",
"gaikovoi",
"gsarapura",
"johnterickson",
"makukl"
],
"repo": "microsoft/azure-pipelines-tasks",
"url": "https://github.com/microsoft/azure-pipelines-tasks/issues/12901",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1497394093 | AzureKeyVault fails with 'endpoints_resolution_error'
Question, Bug, or Feature?
Type: Bug
Enter Task Name: AzureKeyVault
Environment
Agent - Private, running under Azure DevOps
OS: RHEL 8.6
Version: 2.214.2
Issue Description
Secrets downloading takes 30+ sec to complete with the following error:
"Could not fetch access token for Azure. Status code: endpoints_resolution_error, status message: Error: could not resolve endpoints. Please check network and try again. Detail: ClientConfigurationError: untrusted_authority: The provided authority is not a trusted authority. Please include this authority in the knownAuthorities config parameter."
Update - 18 Dec 2022
I've managed to mitigate the error by removing the HTTP proxy and going directly to the internet to download the secrets. So whatever change has been done to the AzureKeyVault to download secrets, it doesn't take into consideration that the secrets download can be done from behind a proxy. And yeah, we set all the proxy environment variables; they're just being ignored.
I am seeing this issue as well. When trying to do a recent deployment we received an error during Download Secrets step which uses AzureKeyVaultV2 task. We noticed there was recent version downloaded 1.214.2 which seems to introduce the error. I am running azure devops agent on windows server 2019 with proxy configured.
Initialize Job Step:
Error during Download Secret step that occurs about 30 seconds later:
@jtterry2856 - Are you running on Self Hosted Agent ?
I got this error only when i am running on Self Hosted Agent with proxy enabled
@RaviChandraMadipadiga yes I am running self hosted agent
We're experiencing the same issue since the 14th. Self hosted agent (ver 2.193.0) using a proxy.
Could not fetch access token for Azure. Status code: endpoints_resolution_error, status message: Error: could not resolve endpoints. Please check network and try again. Detail: ClientConfigurationError: untrusted_authority: The provided authority is not a trusted authority. Please include this authority in the knownAuthorities config parameter.
It seems to be related to the AzureKeyVault task 2.214.2. When we use the previous version (2.211.1), deployments succeed again.
However, I am not aware of an elegant way of forcing the agent to use a specific version for this task, since it is automatically selected because of a linked variable group, linked to the keyvault. Anyone?
@edohussl - Yes , i tried now changing the version to 1* in Azure Key Vault task and it succeeded but whereas with the Azure key vault variable group it is auto picking the version to latest as 1.214.3 as so the issue is occurring. Do we have any feasibility to change the version of the task to pick for key vault variable group ?
@RaviChandraMadipadiga Nope, you can't select the minor plugin version in the release pipeline. You can do that in the YAML pipeline, but YAML pipelines don't support the Deployment Groups. You can use Environments for YAML pipelines, but those don't support sharing of the VMs - as you can't have the same VM in multiple Enviornments... 'Consistency' wasn't apparently a key concern on mind of whoever design this entire thing. Maybe Jenkins isn't that bad after all... If we're blocked for too long I'll be making the case for CloudBees....
We have the same problem and we are not able to deactivate the proxy.
Is there a possibility to set max version of a task for the agent setup?
We are not able to enable direct internet access on our VMs. However, we did implement a workaround (it's more like hack though) to ensure the agent uses an older version of the AzureKeyVault task. We simply copied over the files from the old version (2.211.1) to the new version (2.214.2) in the 'work_tasks\AzureKeyVault...' folder.
I do think this is an issue that has to be fixed in the task or agent. How do we get this under the attention of Microsoft?
For me it works again! It seems that the task has been updated to 2.214.3 and now it runs smoothly.
@kwasiak - Seems like they have updated the task version now to 214.3 and it's working fine now.
Thank you for raising this issue in GitHub by seeing my question in community forum.it really helped all of us .
Works for me as well when I tested pipeline today. New version 1.214.3 was downloaded and "download secrets" step completed successfully.
Same here, works with 2.214.3. Although, it is strange that I see version 2.214.3 reported and other see 1.214.3 reported.
Thanks @ all,
the new AzureKeyVault task 1.214.3 works for us too.
the new AzureKeyVault task 1.215.0 have the same error again...
It is happening again with the newly installed AzureKeyVault task 2.215.0.
correct, just got an update from our DevOps team... I really need to start thinking about Jenkins....
I was troubleshooting a network/proxy related issue (407) regarding the AzureKeyVault task, but during the troubleshooting yesterday, I ran into the "endpoints_resolution_error". I came across this issue here and indeed what happened is that they pushed a new version of this task (2.215.0) yesterday, in which it is broken again ("endpoints_resolution_error"). However, a few hours later they pushed 2.215.1 and that one seems to be working fine again.
I faced the same problem with delete item in adf task. it may relate...
In my case solved by add prerequisite task to set HTTP_PROXY env to the session.
steps:
powershell: 'Write-Host "##vso[task.setvariable variable=HTTP_PROXY;]http://{proxy_host}:{port}"'
| gharchive/issue | 2022-12-14T21:06:44 | 2025-04-01T06:44:57.907240 | {
"authors": [
"ChristianFehlinger",
"RaviChandraMadipadiga",
"RobRybber",
"StuhlfauthKlaus",
"aomerkii",
"armin-pfaeffle",
"edohussl",
"enginhorzum",
"jtterry2856",
"kwasiak",
"prfj",
"vireshov"
],
"repo": "microsoft/azure-pipelines-tasks",
"url": "https://github.com/microsoft/azure-pipelines-tasks/issues/17485",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2074578917 | Azure Arc-enabled data services- update secure string variable
Azure Arc-enabled data services- update secure string variable
LGTM as well!
| gharchive/pull-request | 2024-01-10T15:09:16 | 2025-04-01T06:44:57.917753 | {
"authors": [
"alsanch",
"lanicolas"
],
"repo": "microsoft/azure_arc",
"url": "https://github.com/microsoft/azure_arc/pull/2350",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
793880559 | Gke postgres scenario
Closes #369
Fixed 19.png.
I dont see what you mean about 34-38. Looks like same IP to me?
Fixed 19.png.
I dont see what you mean about 34-38. Looks like same IP to me?
| gharchive/pull-request | 2021-01-26T02:44:45 | 2025-04-01T06:44:57.918960 | {
"authors": [
"dkirby-ms"
],
"repo": "microsoft/azure_arc",
"url": "https://github.com/microsoft/azure_arc/pull/373",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
572060909 | A11y_SQLAzureDataStudio_Notebook_Create new notebook_AI4D: The control type of editor is set as "Custom"
Check out Accessibility Insights! - Identify accessibility bugs before check-in and make bug fixing faster and easier.”
GitHubTags:#A11y_SQLAzureDataStudioBenchmark;#A11yMAS;#A11ySev2;#A11yTCS;#SQL Azure Data Studio;#Benchmark;#DesktopApp;#Win32;#AI4D;#MAS1.3.1;#MAS4.1.2;#MAS4.2.1;
Environment Details:
Application Name: Azure Data Studio
Version: 1.16.0-insider
OS: Windows_NT x64 10.0.18363
Additional Details:AMAS References:MAS4.1.2, MAS4.2.1, MAS1.3.1
Tools: accessibility insights for desktop
Repro Steps:
Launch Azure Data Studio.
Click File then New Notebook.
Launch accessibility insights for desktop
Hover over the editors.
Actual
The control set for editors is set as custom both code and text editor.
Expected:
The control should be set as editor and announced by a screen reader.
User Impact:
Screen reader users will not know that have reached the editor.
Attachment link for Reference
@fsteffi this control encompasses both an editor and the editor's outputs. I'm worried that us saying that the control type as an editor would be confusing, since it's not strictly true. Any thoughts?
@chlafreniere FYI I'm going to bring this one to the UX office hours to see what they say - since yeah there isn't a clear indication of what role this should actually be (editor isn't even an actual aria role)
Thanks, @Charles-Gagnon !!!
#closed;
GitHubTags:#SQLADS-Win32-Feb2020;
| gharchive/issue | 2020-02-27T12:47:38 | 2025-04-01T06:44:57.925564 | {
"authors": [
"Charles-Gagnon",
"chlafreniere",
"fsteffi",
"mstechie",
"v-jagansai"
],
"repo": "microsoft/azuredatastudio",
"url": "https://github.com/microsoft/azuredatastudio/issues/9367",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2158224087 | gobject-introspection: upgrade 1.71.0->1.79.1; meson: upgrade 1.2.1->1.3.1; glib: upgrade 2.78.1->2.79.1
Merge Checklist
All boxes should be checked before merging the PR (just tick any boxes which don't apply to this PR)
[X] The toolchain has been rebuilt successfully (or no changes were made to it)
[X] The toolchain/worker package manifests are up-to-date
[X] Any updated packages successfully build (or no packages were changed)
[X] Packages depending on static components modified in this PR (Golang, *-static subpackages, etc.) have had their Release tag incremented.
[X] Package tests (%check section) have been verified with RUN_CHECK=y for existing SPEC files, or added to new SPEC files
[X] All package sources are available
[X] cgmanifest files are up-to-date and sorted (./cgmanifest.json, ./toolkit/scripts/toolchain/cgmanifest.json, .github/workflows/cgmanifest.json)
[X] LICENSE-MAP files are up-to-date (./SPECS/LICENSES-AND-NOTICES/data/licenses.json, ./SPECS/LICENSES-AND-NOTICES/LICENSES-MAP.md, ./SPECS/LICENSES-AND-NOTICES/LICENSE-EXCEPTIONS.PHOTON)
[X] All source files have up-to-date hashes in the *.signatures.json files
[X] sudo make go-tidy-all and sudo make go-test-coverage pass
[X] Documentation has been updated to match any changes to the build system
[X] Ready to merge
Summary
What does the PR accomplish, why was it needed?
Change Log
gobject-introspection: upgrade 1.71.0->1.79.1; meson: upgrade 1.2.1->1.3.1; glib: upgrade 2.78.1->2.79.1
Skip some meson tests because of their flakiness or we don't support them
Add python-packaging to build_official_toolchain_rpms as required for new glib version
Does this affect the toolchain?
NO
Test Methodology
Buddy Build Id
Full Build Id
Looks like toolchain build is failing for me because of python3-packaging-23.2-2.azl3.noarch.rpm
https://dev.azure.com/mariner-org/mariner/_build/results?buildId=515772&view=logs&j=db98f19e-da46-5e4d-a4ba-372cf3771a92&t=8831c5ce-9476-5d69-9bf5-bfd7485723f1&l=464717
| gharchive/pull-request | 2024-02-28T06:35:42 | 2025-04-01T06:44:57.935588 | {
"authors": [
"BettyRain"
],
"repo": "microsoft/azurelinux",
"url": "https://github.com/microsoft/azurelinux/pull/8140",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2296586900 | python-werkzeug: Patch CVE-2024-34069
Merge Checklist
All boxes should be checked before merging the PR (just tick any boxes which don't apply to this PR)
[ ] The toolchain has been rebuilt successfully (or no changes were made to it)
[ ] The toolchain/worker package manifests are up-to-date
[ ] Any updated packages successfully build (or no packages were changed)
[ ] Packages depending on static components modified in this PR (Golang, *-static subpackages, etc.) have had their Release tag incremented.
[ ] Package tests (%check section) have been verified with RUN_CHECK=y for existing SPEC files, or added to new SPEC files
[ ] All package sources are available
[ ] cgmanifest files are up-to-date and sorted (./cgmanifest.json, ./toolkit/scripts/toolchain/cgmanifest.json, .github/workflows/cgmanifest.json)
[ ] LICENSE-MAP files are up-to-date (./SPECS/LICENSES-AND-NOTICES/data/licenses.json, ./SPECS/LICENSES-AND-NOTICES/LICENSES-MAP.md, ./SPECS/LICENSES-AND-NOTICES/LICENSE-EXCEPTIONS.PHOTON)
[ ] All source files have up-to-date hashes in the *.signatures.json files
[ ] sudo make go-tidy-all and sudo make go-test-coverage pass
[ ] Documentation has been updated to match any changes to the build system
[ ] If you are adding/removing a .spec file that has multiple-versions supported, please add @microsoft/cbl-mariner-multi-package-reviewers team as reviewer (Eg. golang has 2 versions 1.18, 1.21+)
[ ] Ready to merge
Summary
Patch CVE-2024-34069
Change Log
CVE-2024-34069
Does this affect the toolchain?
NO
Links to CVEs
https://nvd.nist.gov/vuln/detail/CVE-2024-34069
Test Methodology
Pipeline build id: 570115
The one failing ptests from the PR check wasn't triggered by this change - merging.
Auto cherry-pick results:
main :white_check_mark: -> https://github.com/microsoft/azurelinux/pull/9118
Auto cherry-pick pipeline run -> https://dev.azure.com/mariner-org/mariner/_build/results?buildId=570548&view=results
| gharchive/pull-request | 2024-05-14T23:39:05 | 2025-04-01T06:44:57.945999 | {
"authors": [
"CBL-Mariner-Bot",
"PawelWMS",
"fintelia"
],
"repo": "microsoft/azurelinux",
"url": "https://github.com/microsoft/azurelinux/pull/9104",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1173208301 | Cadl Syntax Review
Hold a review of cadl syntax with arch board
Gather feedbacks from early adopters on cadl syntax
related to #318
Syntax to review:
[x] Allow using above the global namespace from https://github.com/Azure/cadl-azure/issues/1320
Closing as dupolicate of https://github.com/Azure/cadl-azure/issues/1518
| gharchive/issue | 2022-03-18T05:40:40 | 2025-04-01T06:44:57.958158 | {
"authors": [
"markcowl"
],
"repo": "microsoft/cadl",
"url": "https://github.com/microsoft/cadl/issues/337",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1174647723 | [BUG] Incorrect casing of "GitHub"
Description
GitHub is incorrectly written as "Github".
Expected
"GitHub"
Need to add a permanent linter check for this one and fail the build if the word GitHub case is incorrect and not a part of a link.
| gharchive/issue | 2022-03-20T18:52:22 | 2025-04-01T06:44:57.959449 | {
"authors": [
"NenoLoje",
"shiranr"
],
"repo": "microsoft/code-with-engineering-playbook",
"url": "https://github.com/microsoft/code-with-engineering-playbook/issues/813",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2333894276 | Investigate --no-compile with PipReport detector
A potential improvement for the PipReport detector is to see if --no-compile produces any benefit to performance when used in conjunction with the other existing parameters. Validate that the dependency graph is still the same and that there is a notable improvement in detection time.
From limited local testing, this seems to have no effect on report generation times. I'll try and investigate further later on.
| gharchive/issue | 2024-06-04T16:05:23 | 2025-04-01T06:44:58.000907 | {
"authors": [
"cobya"
],
"repo": "microsoft/component-detection",
"url": "https://github.com/microsoft/component-detection/issues/1149",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1129481299 | clientgen elements can't read their outputs back in without additional parsing
I have a clientgen code object for the query
query Farms { farms { id } }
For a schema with
type Farm {
id: String!
}
The output for this query when run by hand includes the data and possibly error fields in the root of the returned payload - however the generated parseResponse(response::Value response) begins parsing for farms in the root document - not data
Combined with the JSON parsing of the returned document - it would make serialising the final response on the server and deserialising it back into C++ on the client simpler.
ie:
client::query::Farms::parseResponse(response::parseJSON(str))
Otherwise you need to do your own parse for data and errors beforehand.
Not gonna call this a bug but I would like to know if that is intentional behaviour or not. Doing a quick parse for data or errors is pretty easy - but could be standardised to save some boilerplate.
There's a function called parseServiceResponse in GraphQLClient.h which should do the trick. It'll split the response into data and error members, and then you can parse the data independently with the generated function from clientgen.
This has come up before, though. Maybe clientgen should output a comment with parseResponse referencing parseServiceResponse to make this clearer.
| gharchive/issue | 2022-02-10T05:40:44 | 2025-04-01T06:44:58.004937 | {
"authors": [
"ALTinners",
"wravery"
],
"repo": "microsoft/cppgraphqlgen",
"url": "https://github.com/microsoft/cppgraphqlgen/issues/216",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
530932986 | Safety of manipulating member variables in async callback?
In my code, I fire off a task at startup, which when completed assigns the result to a member variable:
// Error checking ignored
using namespace winrt::Windows::Foundation;
using namespace winrt::Windows::ApplicationModel;
using Collections::IVectorView;
StartupTask::GetForCurrentPackageAsync().Completed([this](const IAsyncOperation<IVectorView<StartupTask>> &op, AsyncStatus)
{
m_StartupTask = op.GetResults().GetAt(0);
});
There doesn't seem to be actually much details on how safe that is (except if this gets destroyed, which is already greatly covered) because as far as I understand, the result of the operation might very well run on another thread (documentation on apartments isn't very clear about this one) and therefore face concurrency issues. This class currently does not have any thread safety included, and I am not sure if I should add some around m_StartupTask to avoid concurrency issues (or if WinRT consumer types already are thread-safe)
In general, implementations should be agile but that's up to the implementation. An agile object may be accessed from any thread or apartment. If the API is not agile then you can use apartment_context or resume_foreground depending on the API. Regardless of apartment context, any concurrency you introduce (such as the example above) needs appropriate locking if you think there may be a race. WinRT does not provide any such thread-safety automatically, except what may be provided implicitly by using something like resume_foreground or thread_pool to queue work onto a single thread.
| gharchive/issue | 2019-12-02T07:05:25 | 2025-04-01T06:44:58.007789 | {
"authors": [
"kennykerr",
"sylveon"
],
"repo": "microsoft/cppwinrt",
"url": "https://github.com/microsoft/cppwinrt/issues/437",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
740145161 | Support execution on a RAM drive
See: https://github.com/microsoft/cppwinrt/pull/770
Feel free to reopen if there's interest in exploring this further.
| gharchive/issue | 2020-11-10T18:31:01 | 2025-04-01T06:44:58.009049 | {
"authors": [
"kennykerr"
],
"repo": "microsoft/cppwinrt",
"url": "https://github.com/microsoft/cppwinrt/issues/786",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2493821113 | Environments sometime don't load and clicking 'Create new environment' causes a crash
Dev Home version
0.1701.597.0
Windows build number
10.0.22631.0
Other software
N / A
Steps to reproduce the bug
I haven't got solid steps to repro this every time, but it's happening quite frequently with me visiting the environments page. Don't really know what triggers it. What happens is that sometimes Environments don't load properly, and it shows "No environments found". When I click to "Create new environment", devhome crashes. See the video for better understanding. First time the environments load correctly, but second time they don't.
https://github.com/user-attachments/assets/225f190d-91ce-4089-a5c5-d5828b19a718
Expected result
Environments screen to always load correctly
Actual result
Environments not always loaded, and cause a crash when you click "Create new environment"
Included System Information
Windows: Windows.Desktop v10.0.22631.4037
System Architecture: X64
Included Extensions Information
DevHome GitHub Extension version 0.1700.597.0
/logs
Hi @mdanish-kh, would you be able to file a feedback hub bug for this issue and provide us with the link? That will help us diagnose this issue faster.
You can launch the Feedback hub app by entering the Windows Key + the F key on your keyboard at the same time. During your feedback, there will be an option to recreate the issue in your video. Feedback hub will capture the data needed for further debugging on our side.
Thank you!
@bbonaby Feedback Hub was not cooperating yesterday, so it took me some time to get back. This should be the issue: https://aka.ms/AAs318i
I also used the "Recreate the issue" and was able to successfully repro it. Let me know if you need anything else from me
Fixed as a side effect of #3907
| gharchive/issue | 2024-08-29T08:44:51 | 2025-04-01T06:44:58.018104 | {
"authors": [
"bbonaby",
"huzaifa-d",
"mdanish-kh"
],
"repo": "microsoft/devhome",
"url": "https://github.com/microsoft/devhome/issues/3726",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1924728150 | bpf_map_get_next_key should return the first key if key not found
As per information from the Isovalent folks, the bpf_map_get_next_key should start the iteration of the BPF map at the beginning of the map if the key is not found.
Adding the specs as per the official documentation here: https://man7.org/linux/man-pages/man2/bpf.2.html
BPF_MAP_GET_NEXT_KEY
The BPF_MAP_GET_NEXT_KEY command looks up an element by
key in the map referred to by the file descriptor fd and
sets the next_key pointer to the key of the next element.
int
bpf_get_next_key(int fd, const void *key, void *next_key)
{
union bpf_attr attr = {
.map_fd = fd,
.key = ptr_to_u64(key),
.next_key = ptr_to_u64(next_key),
};
return bpf(BPF_MAP_GET_NEXT_KEY, &attr, sizeof(attr));
}
If key is found, the operation returns zero and sets the
next_key pointer to the key of the next element. If key
is not found, the operation returns zero and sets the
next_key pointer to the key of the first element. If key
is the last element, -1 is returned and [errno](https://man7.org/linux/man-pages/man3/errno.3.html) is set to
ENOENT. Other possible [errno](https://man7.org/linux/man-pages/man3/errno.3.html) values are ENOMEM, EFAULT,
EPERM, and EINVAL. This method can be used to iterate
over all elements in the map.
| gharchive/issue | 2023-10-03T18:48:41 | 2025-04-01T06:44:58.020222 | {
"authors": [
"Alan-Jowett",
"gtrevi"
],
"repo": "microsoft/ebpf-for-windows",
"url": "https://github.com/microsoft/ebpf-for-windows/issues/2942",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2096463017 | Reindex job stuck at "Queued"
Describe the bug
After submitting a reindex job, the job remains queued indefinitely with no progress. The issue persists after deleting and resubmitting the job. Similar to #2200 ?
FHIR Version?
R4
Data provider?
CosmosDB
To Reproduce
Steps to reproduce the behavior:
Create search parameter, test on a single resource (success)
Submit reindex job for new parameter (success)
Query reindex job ID - status is stuck at "Queued"
Expected behavior
Reindex job should run
Actual behavior
Reindex job remains "Queued"
Thanks for reporting the issue @elgib - Are you observing this issue on OSS FHIR service or are you using managed Azure FHIR service (Azure API for FHIR / Azure Health Data Services)?
@EXPEkesheth we're using the OSS FHIR service
@elgib , have added in our queue for investigation. Can you please share with us search parameter json used for creating custom search parameter. Will inform once we have more details/ questions.
#114359
@EXPEkesheth here's an example of one of our custom search parameters.
{ "resourceType": "SearchParameter", "id": "e46bd3c4-f278-4039-841b-892e931596fe", "meta": { "versionId": "1", "lastUpdated": "2023-11-22T12:54:08.333+00:00" }, "url": "http://1beat.care/fhir/search-parameters#patient-care-unit", "name": "patient-care-unit", "status": "draft", "description": "Reference to Organization resource that represents the care unit currently responsible for the patient.", "code": "care-unit", "base": [ "Patient" ], "type": "reference", "expression": "Patient.extension.where(url = 'http://1beat.care/fhir/extensions#patient-care-unit').value", "target": [ "Organization" ] }
Thanks for looking into this. We would appreciate any updates as this is blocking key areas of work for our team. Any short-term recommendations would also be helpful -- for example, should we try rolling back to an earlier version?
@EXPEkesheth are there any updates on this issue, or timeframes for a fix?
@elgib - As you are using the OSS FHIR service , you would need to explicitly enable Reindex in deployment template (https://github.com/microsoft/fhir-server/blob/main/samples/templates/default-azuredeploy-docker.json#L272). Please ensure this setting is enabled in your instance.
Have you used reindex capability in FHIR server OSS before?
@EXPEkesheth we have enabled reindex operations and run several reindex jobs successfully in the past. The last successful reindex was late November 2023. We have not made any changes to our setup since then.
@elgib Thanks for the information. We will look into the issue and get back incase of any questions.
@elgib - we recently made improvements in reindex operation. We hope it helps address your issue, can you please execute reindex and let us know the outcome?
| gharchive/issue | 2024-01-23T16:22:39 | 2025-04-01T06:44:58.028623 | {
"authors": [
"EXPEkesheth",
"elgib"
],
"repo": "microsoft/fhir-server",
"url": "https://github.com/microsoft/fhir-server/issues/3684",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
802423482 | Upgrade dotnet sdk to 5.0.100
Description
Upgrade dotnet sdk to 5.0.100
Related issues
Addresses [issue #].
Couple of latest PRs are failing due to this (screenshot below) error
Testing
Describe how this change was tested.
FHIR Team Checklist
[x] Update the title of the PR to be succinct and less than 50 characters
[ ] Add a milestone to the PR for the sprint that it is merged (i.e. add S47)
[x] Tag the PR with the type of update: Bug, Dependencies, Enhancement, or New-Feature
[ ] Tag the PR with Azure API for FHIR if this will release to the managed service
Review squash-merge requirements
Semver Change (docs)
Patch|Skip|Feature|Breaking (reason)
👏
I think these tests will be fixed by #1650
I think +semver: feature 🚀
Thoughts?
Good point!
| gharchive/pull-request | 2021-02-05T19:37:22 | 2025-04-01T06:44:58.034672 | {
"authors": [
"Ivanidzo4ka",
"brendankowitz",
"rbans96"
],
"repo": "microsoft/fhir-server",
"url": "https://github.com/microsoft/fhir-server/pull/1645",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1865148454 | Fix FHIR Search query with _count=0 and SqlCustomQueryTest issue
Description
Fix FHIR Serach query to return result same as _summary=count when passing _count=0.
Previously, 400 response was there for _count=0.
Fix SqlCustomQueryTest issue.
Updated following logic to fix this issue.
Adds logic to remove the _count param and replaces it with _summary param in SearchOptionsFactory .
Updates GetSummaryTypeOrDefault method in HttpContextExtensions class to return summary type.
Related issues
Addresses [issue #3230 , #105699].
Testing
Adds unit test under SearchOperationsFactoryTests
Adds unit test under HttpContextExtensionsTests
FHIR Team Checklist
Update the title of the PR to be succinct and less than 65 characters
Add a milestone to the PR for the sprint that it is merged (i.e. add S47)
Tag the PR with the type of update: Bug, Build, Dependencies, Enhancement, New-Feature or Documentation
Tag the PR with Open source, Azure API for FHIR (CosmosDB or common code) or Azure Healthcare APIs (SQL or common code) to specify where this change is intended to be released.
[ ] CI is green before merge
Review squash-merge requirements
Semver Change (docs)
Patch|Skip|Feature|Breaking (reason)
What happens if a user specifies both _count=0 and _summary? This looks like it will take whichever is later in the query parameter array. I'm going to run a test on this.
What happens if a user specifies both _count=0 and _summary? This looks like it will take whichever is later in the query parameter array. I'm going to run a test on this.
Yeah, these queries give different results:
https://localhost:44348/Patient?_summary=text&_count=0
https://localhost:44348/Patient?_count=0&_summary=text
The first one gives the number of Patients and nothing else.
The second one gives the text summary.
The FHIR spec doesn't say what should be done when multiple summary parameters are given. I feel this should be an error.
Also, we have a bug. When returning just the count of resources we don't return a self link. But section 3.2.1.7.5 of the FHIR search spec says we should.
I know we have an existing issue about the order of search parameters affecting results. @brendankowitz thoughts?
Nevermind, I found the section in the spec. It says this is left up to implementations, but they recommend returning an error.
Note that with the exception of _include and _revinclude, search result parameters SHOULD only appear once in a search. If such a parameter appears more than once, the behavior is undefined and a server MAY treat the situation as an error
@mahajan-xor I don't think this needs to be part of this PR as it is a wider issue in our service. It is a bug though, and we should track it. I'll make an item for it.
| gharchive/pull-request | 2023-08-24T13:21:25 | 2025-04-01T06:44:58.045319 | {
"authors": [
"LTA-Thinking",
"mahajan-xor"
],
"repo": "microsoft/fhir-server",
"url": "https://github.com/microsoft/fhir-server/pull/3491",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
466663954 | Auto-update for packages related to 'Microsoft.CodeAnalysis'
Updates package 'Microsoft.CodeAnalysis.FxCopAnalyzers' to version '2.9.3'
Updates package 'Microsoft.CodeAnalysis.CSharp.Scripting' to version '3.1.0'
2.9.3 Seems to have a bug that flags CA1062 (Check for null parameter) in a place where it is actually being checked (ServiceDescriptorExtensions : WithMetadata). 2.9.4-beta1.final seems to have it fixed, but it is a preview version. We should revisit this when 2.9.4 is released.
Closing as 2.9.4 is released
| gharchive/pull-request | 2019-07-11T05:01:28 | 2025-04-01T06:44:58.047688 | {
"authors": [
"MicrosoftHealthService",
"YazanMSFT",
"brendankowitz"
],
"repo": "microsoft/fhir-server",
"url": "https://github.com/microsoft/fhir-server/pull/573",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2623355936 | fix: [Local storage for theme not being set] in [FluentDesignTheme]
🐛 Bug Report
Theme is not saved in localstorage when using StorageName
💻 Repro or Code Sample
Add to App.razor or main layout:
<FluentDesignTheme Mode="DesignThemeModes.Default" OfficeColor="OfficeColor.Teams" StorageName="theme" />
🤔 Expected Behavior
Theme should be saved in localstorage without need to switch it manually for the first time
😯 Current Behavior
When you set a theme like this:
<FluentDesignTheme Mode="DesignThemeModes.Default" OfficeColor="OfficeColor.Teams" StorageName="theme" />
it's not being added to localstorage untill you switch the theme.
Right now, the only way to save the theme to localstorage is to bind values and switch the theme for example using some button, only then the theme is added to localstorage.
<FluentDesignTheme @bind-Mode="@Mode" @bind-OfficeColor="@ThemeColor" StorageName="theme"
https://github.com/user-attachments/assets/4de13519-002d-4109-a6a5-eba8bc2151e4
💁 Possible Solution
Not sure
🔦 Context
I want to save the theme in localstorage with values mode: null and primaryColor: Teams, so it will switch according to user settings. Right now sometimes the theme is being set wrong and having the values in localstorage could help.
🌍 Your Environment
Windows 11, Edge, FluentUI v4.10.3, .NET 8.0.4
This is how the component works: nothing is saved in the LocalStorage until it is needed and the default values are used.
We can't change this behaviour.
However, you can use the OnLoaded event to get the theme settings when your application starts.
This is how the component works: nothing is saved in the LocalStorage until it is needed and the default values are used.
We can't change this behaviour.
However, you can use the OnLoaded event to get the theme settings when your application starts.
Alright I understand. Thank you 😊
| gharchive/issue | 2024-10-30T09:08:15 | 2025-04-01T06:44:58.053741 | {
"authors": [
"Nikkoro",
"dvoituron"
],
"repo": "microsoft/fluentui-blazor",
"url": "https://github.com/microsoft/fluentui-blazor/issues/2885",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2224123852 | Update from upstream and fix missing set -eux
Update upstream submodule and resolve patch conflicts.
Upstream is now using a builder stage, so some things are reordered. They have some infrastructure set up to actually share more work as a result (https://github.com/docker-library/golang/commit/a6fd6eceb0cb26da2fceefb4353768c472f84420) but for us, it doesn't have any meaningful impact.
Fix a missing set -eux that could have caused errors to be improperly ignored. Add a command to upgrade all distro packages for Azure Linux and Debian.
https://github.com/microsoft/go-images/pull/290
Regenerate the Dockerfiles with dockerupdate -f.
A few more small fixes, would appreciate a set of eyes taking another look. 😄
| gharchive/pull-request | 2024-04-03T23:39:15 | 2025-04-01T06:44:58.441042 | {
"authors": [
"dagood"
],
"repo": "microsoft/go-images",
"url": "https://github.com/microsoft/go-images/pull/294",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2597474974 | Grand Overhaul: Refactored Core Structure, Introduced New Features, E…
Refactoring of graphrag/query Module
Description
This pull request introduces a significant refactoring of the graphrag/query module within the GraphRAG project. The
primary objectives of this refactoring are:
Decoupling the Query Module: Transform the query component into an independent package, fully decoupled from
other modules.
Enhancing Code Reusability and Modularity: Implement a modular design for the entire lifecycle of the GraphRAG
query pipeline, promoting loose coupling and facilitating future maintenance and extension.
Improving the Python API: Provide a more user-friendly and convenient Python API, simplifying the creation and
management of GraphRAG clients.
Eliminating Redundancies: Remove redundant modules and parameters (e.g., the question_gen module), streamlining
the codebase.
Comprehensive Documentation: Add detailed docstrings and extensive type annotations throughout the codebase,
ensuring code reliability and passing mypy checks.
Enhanced CLI and GUI Tools: Introduce a more powerful CLI tool with rich parameter combinations and an optional
GUI built with PyQt6.
Unified Streaming Implementation: Employ a more elegant approach to handle both streaming and non-streaming
outputs within a single method.
Related Issues
N/A
Proposed Changes
1. Project Layout
The codebase has been reorganized, promoting separation of concerns and ease of navigation. The new structure is as
follows:
query/
├── __init__.py # Package initialization
├── __main__.py # CLI entry point
├── _base_client.py # Base client templates
├── _cli/ # CLI layer
│ ├── __init__.py
│ ├── _api.py # CLI API
│ ├── _cli.py # CLI main program
│ ├── _qt/ # GUI layer
│ │ ├── __init__.py
│ │ └── _app.py # GUI main program
│ └── _utils.py # CLI utilities
├── _client.py # GraphRAG clients
├── _config.py # Configuration classes
├── _defaults.py # Default constants
├── _search/ # Search layer
│ ├── __init__.py
│ ├── _context/ # Context module
│ │ ├── __init__.py
│ │ ├── _builders/ # Context builders
│ │ ├── _loaders/ # Context loaders
│ │ └── _types.py # Type hints
│ ├── _defaults.py # Search layer defaults
│ ├── _engine/ # Engine module
│ │ ├── __init__.py
│ │ ├── _base_engine.py # Base engine template
│ │ ├── _global.py # Global search engine
│ │ └── _local.py # Local search engine
│ ├── _input/ # Input module
│ │ ├── __init__.py
│ │ ├── _loaders/ # Input loaders
│ │ └── _retrieval/ # Input retrieval
│ ├── _llm/ # LLM module
│ │ ├── __init__.py
│ │ ├── _base_llm.py # Base LLM template
│ │ ├── _chat.py # Chat LLM
│ │ ├── _embedding.py # Text Embedding
│ │ └── _types.py # Type hints
│ ├── _model/ # Data models
│ └── _types/ # Type hints
│ ├── __init__.py
│ ├── _search.py
│ ├── _search_chunk.py
│ ├── _search_verbose.py
│ └── _search_chunk_verbose.py
├── _utils/ # Utilities
│ ├── __init__.py
│ ├── _text.py # Text utilities
│ └── _utils.py # General utilities
├── _vector_stores/ # Vector storage layer
│ ├── __init__.py
│ ├── _base_vector_store.py
│ └── _lancedb.py
├── _version.py # Version information
├── errors.py # Error types
└── types.py # Type hints
The query module is now fully decoupled from other modules, making it usable as a standalone package.
The code is reorganized to promote modularity, facilitating easier maintenance and potential future extensions.
2. Enhanced Python API
2.1 Initialize
Users can easily create a GraphRAGClient instance using configuration file, dictionary, environment variables or
configuration object.
a) From Configuration File
e.g.,
from graphrag.query import GraphRAGClient
config_file = "config.yaml"
client = GraphRAGClient.from_config_file(config_file)
The configuration file can be in YAML, JSON, or TOML format. Refer to the graphrag.example.yaml file for an example.
b) From Configuration Dictionary
e.g.,
from graphrag.query import AsyncGraphRAGClient
config = {
"chat": {
"api_key": "API_KEY",
"base_url": "BASE_URL",
"model": "MODEL"
},
"embedding": {
"api_key": "API_KEY",
"base_url": "BASE_URL",
"model": "MODEL"
}
}
client = AsyncGraphRAGClient.from_config_dict(config)
c) From Configuration Object
If you prefer to use a configuration object and an optional logger, you can pass them directly to the constructor:
import logging
from graphrag.query import (
ChatLLMConfig,
EmbeddingConfig,
GraphRAGClient,
GraphRAGConfig,
)
logger = logging.getLogger(__name__)
config = GraphRAGConfig(
chat=ChatLLMConfig(api_key="API_KEY", base_url="BASE_URL", model="MODEL"),
embedding=EmbeddingConfig(api_key="API_KEY", base_url="BASE_URL", model="MODEL")
)
client = GraphRAGClient(config=config, logger=logger)
d) From Environment Variables
You can also initialize a client using environment variables:
export GRAPHRAG_QUERY__CHAT_LLM__API_KEY=API_KEY
export GRAPHRAG_QUERY__CHAT_LLM__MODEL=MODEL
export GRAPHRAG_QUERY__EMBEDDING__API_KEY=API_KEY
export GRAPHRAG_QUERY__EMBEDDING__MODEL=MODEL
Or create .env file in the project root directory:
GRAPHRAG_QUERY__CHAT_LLM__API_KEY=API_KEY
GRAPHRAG_QUERY__CHAT_LLM__MODEL=MODEL
GRAPHRAG_QUERY__EMBEDDING__API_KEY=API_KEY
GRAPHRAG_QUERY__EMBEDDING__MODEL=MODEL
Then initialize the client:
from graphrag.query import GraphRAGClient, GraphRAGConfig
config = GraphRAGConfig()
client = GraphRAGClient(config=config)
2.2 Chatting with GraphRAG
a) Simple Chat
You can chat with GraphRAG using the chat method:
from graphrag.query import GraphRAGClient
client: GraphRAGClient = ...
response = client.chat(
engine="local",
message=[
{"role": "user", "content": "What is the purpose of life?"},
{"role": "assistant", "content": "The purpose of life is to be happy."},
{"role": "user", "content": "What is the meaning of happiness?"}
],
)
print(response.choice.message.content)
Or, in streaming mode:
from graphrag.query import GraphRAGClient
client: GraphRAGClient = ...
response = client.chat(
engine="local",
message=[
{"role": "user", "content": "What is the purpose of life?"},
{"role": "assistant", "content": "The purpose of life is to be happy."},
{"role": "user", "content": "What is the meaning of happiness?"}
],
stream=True
)
for chunk in response:
print(chunk.choice.delta.content, end="")
client.close() # Close the client
c) Using with Statement
You can also use the with statement to manage the client's lifecycle:
from graphrag.query import GraphRAGClient, GraphRAGConfig
config: GraphRAGConfig = ...
with GraphRAGClient(config=config) as client:
response = client.chat(
engine="local",
message=[
{"role": "user", "content": "What is the purpose of life?"},
{"role": "assistant", "content": "The purpose of life is to be happy."},
{"role": "user", "content": "What is the meaning of happiness?"}
],
stream=True
)
for chunk in response:
print(chunk.choice.delta.content, end="")
d) Verbose Search Results
If you want to collect verbose search results, you can set the verbose parameter to True:
from graphrag.query import GraphRAGClient
client: GraphRAGClient = ...
response = client.chat(
engine="local",
message=[
{"role": "user", "content": "What is the purpose of life?"},
{"role": "assistant", "content": "The purpose of life is to be happy."},
{"role": "user", "content": "What is the meaning of happiness?"}
],
verbose=True
)
print(response.model_dump())
Or, in streaming mode:
from graphrag.query import GraphRAGClient
client: GraphRAGClient = ...
response = client.chat(
engine="local",
message=[
{"role": "user", "content": "What is the purpose of life?"},
{"role": "assistant", "content": "The purpose of life is to be happy."},
{"role": "user", "content": "What is the meaning of happiness?"}
],
streaming=True,
verbose=True
)
for chunk in response:
print(chunk.model_dump())
e) Async Client
AsyncGraphRAGClient provides an asynchronous version of the GraphRAGClient:
import asyncio
from graphrag.query import AsyncGraphRAGClient, GraphRAGConfig
config: GraphRAGConfig = ...
async def main():
client = AsyncGraphRAGClient(config=config)
response = await client.chat(
engine="local",
message=[
{"role": "user", "content": "What is the purpose of life?"},
{"role": "assistant", "content": "The purpose of life is to be happy."},
{"role": "user", "content": "What is the meaning of happiness?"}
],
streaming=True
)
async for chunk in response:
print(chunk.choice.delta.content, end="")
await client.close() # Or you can use the async context manager
asyncio.run(main())
3. Streamlined CLI and GUI Tools
3.1 CLI Parameters
Execute the following command:
python -m graphrag.query --help
To see the available options:
usage: python -m query [-h] [--verbose] [--engine {local,global}] [--stream] --chat-api-key CHAT_API_KEY [--chat-base-url CHAT_BASE_URL] --chat-model CHAT_MODEL
--embedding-api-key EMBEDDING_API_KEY [--embedding-base-url EMBEDDING_BASE_URL] --embedding-model EMBEDDING_MODEL --context-dir CONTEXT_DIR
[--mode {console,gui}] [--sys-prompt SYS_PROMPT] [-V]
GraphRAG Query CLI
options:
-h, --help show this help message and exit
--verbose, -v enable verbose logging (default: False)
--engine {local,global}, -e {local,global}
engine to use for the query (default: local)
--stream, -s enable streaming output (default: False)
--chat-api-key CHAT_API_KEY, -k CHAT_API_KEY
API key for the Chat API (default: None)
--chat-base-url CHAT_BASE_URL, -b CHAT_BASE_URL
base URL for the chat API (default: None)
--chat-model CHAT_MODEL, -m CHAT_MODEL
model to use for the chat API (default: None)
--embedding-api-key EMBEDDING_API_KEY, -K EMBEDDING_API_KEY
API key for the embedding API (default: None)
--embedding-base-url EMBEDDING_BASE_URL, -B EMBEDDING_BASE_URL
base URL for the embedding API (default: None)
--embedding-model EMBEDDING_MODEL, -M EMBEDDING_MODEL
model to use for the embedding API (default: None)
--context-dir CONTEXT_DIR, -c CONTEXT_DIR
directory containing the context data (default: None)
--mode {console,gui}, -o {console,gui}
mode to execute the GraphRAG engine (default: console)
--sys-prompt SYS_PROMPT, -p SYS_PROMPT
system prompt file in TXT format to use for the local engine (default: None)
-V, --version show program's version number and exit
3.2 Usage Examples
We can get started with the CLI from the corpus used in the GraphRAG official tutorial:
curl https://www.gutenberg.org/cache/epub/24022/pg24022.txt -o ./input/pg24022.txt
Then running the indexing pipeline. Ommited for brevity.
a) Console Mode
python -m graphrag.query --engine local \
--chat-api-key API_KEY \
--chat-model MODEL \
--embedding-api-key API_KEY \
--embedding-model MODEL \
--context-dir ./output \
--mode console \
--stream
Or, more concisely:
python -m graphrag.query -e local \
-k API_KEY \
-m MODEL \
-K API_KEY \
-M MODEL \
-c ./output \
-o console \
-s
Here is an example screenshot:
b) GUI Mode
python -m graphrag.query --engine local \
--chat-api-key API_KEY \
--chat-model MODEL \
--embedding-api-key API_KEY \
--embedding-model MODEL \
--context-dir ./output \
--mode gui
Here is an example screenshot:
4. Web API
Applied the refactored query module to a web service in
the graphrag-server repository, providing an OpenAI-compatible
Chat API interface.
git clone https://github.com/6ixGODD/graphrag-server.git
cd graphrag-server
Modify the .env file with the appropriate API keys and models.
cp .env.example .env
Write a simple Python script to execute the web service:
from server import create_app
app = create_app()
if __name__ == '__main__':
import uvicorn
uvicorn.run(app, host='127.0.0.1', port=8000)
Then you can use the OpenAI SDK to interact with the web service:
import openai
client = openai.OpenAI(
api_key="API_KEY",
base_url="http://127.0.0.1:8000/api",
)
Detailed documentation and deployment instructions (e.g., using Gunicorn and Docker) will be provided in future
updates.
Currently, there is no detailed docstring documentation for the web service; this will be added subsequently.
Checklist
[x] I have tested these changes locally.
[x] I have reviewed the code changes.
[x] I have updated the documentation (if necessary).
[ ] I have added appropriate unit tests (if applicable).
Additional Notes
As mentioned, this PR involves significant code changes, but I believe it is a positive step forward. With thorough
testing, it will provide developers a more stable and modular version of GraphRAG for integration into their
applications, leading to greater overall benefits.
However, for this PR to be merged, some additional documentation work and test case development may require
collaboration with the official team.
PLEASE REVIEW THIS! PEOPLE ARE WAITING!!!!
PLEASE REVIEW THIS! PEOPLE ARE WAITING!!!!
You need to rebase to main.
| gharchive/pull-request | 2024-10-18T13:05:54 | 2025-04-01T06:44:58.465674 | {
"authors": [
"6ixGODD",
"JoedNgangmeni",
"knguyen1"
],
"repo": "microsoft/graphrag",
"url": "https://github.com/microsoft/graphrag/pull/1295",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2124836272 | shorthands.flex(1) should produce flex: 1 1 0% instead of flex: 1 1 0px
When you set a css property to flex: 1 in any of the major browsers the result isflex: 1 1 0%. But the griffel shorthand shorthands.flex(1) produces flex: 1 1 0px.
What's maybe even weirder is that the standard actually defines that flex: <positive-number> is the equivalent of flex: <positive-number> 1 0;. But somehow all browser vendors decided that this is just wrong and produce flex: 1 1 0% as mentioned above.
Either way following reality or following the standard griffel produces an incorrect output. Ideally it should follow how the major browser vendors do it to have the same developer experience as specifying flex: 1, because otherwise this can be a real pitfall.
https://github.com/w3c/csswg-drafts/issues/5742
@stefan-schweiger shorthands.flex() have been deprecated as CSS shorthands are supported natively, check #531 😉
| gharchive/issue | 2024-02-08T10:37:03 | 2025-04-01T06:44:58.470431 | {
"authors": [
"layershifter",
"stefan-schweiger"
],
"repo": "microsoft/griffel",
"url": "https://github.com/microsoft/griffel/issues/502",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1025784953 | Update the Type field name to PolicyType for SetPolicy
The "Type" fieldname was interfering with a internal "Type" field. Changing it to "PolicyType"
Signed-off-by: netal netalgupta17@gmail.com
@netal You'll need to run go mod vendor and go mod tidy in the /test directory in this repo and re-push to fix the CI. Also, I'm having trouble finding uses/examples of SetPolicySetting. I don't see any in this repo or Kubernetes so not sure how it's used.
As of now, there are no customers using Setpolicy. Azure NPM is the first customer using this policy.
@netal or @dcantah or @kevpar any ETA on when this PR will be merged and released ?
@vakalapa @netal Sorry for the delay, this looks fine to us. Going to squash the commits and check in shortly
@dcantah when can we have a release with this version ?
| gharchive/pull-request | 2021-10-13T23:38:02 | 2025-04-01T06:44:58.473421 | {
"authors": [
"dcantah",
"netal",
"vakalapa"
],
"repo": "microsoft/hcsshim",
"url": "https://github.com/microsoft/hcsshim/pull/1194",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
848819012 | Incorrect(?) SDK Invalidation: SDK assumes all JWKs must conform to SHC rules
Problem
SHC IG specifies requirements for public JWK to specify signature validation against (use: sig; alg: ES256, etc). However, the SHC IG does not specify that all public keys hosted a domain.com/.well-known/jwks.json must be a SHC-conforming JWK: it's possible that orgs might want to host other various JWKs there and only the JWK with the kid matching that in the VC must conform.
Solution
Update to only validate against the JWK matching the kid, not all keys in the key-set
https://github.com/microsoft/health-cards-validation-SDK/blob/fb6a527401e7e49d9bd0433c465ab4bac6498c31/src/shcKeyValidator.ts#L174
cc @jmandel to make sure I'm interpreting the SHC IG correctly
@christianpaquin assuming we are aligned that this is a problem worth solving, I'm happy to help with a patch here.
Yes, @dleve123, JWK sets certainly can have more than one key. The tool currently gives warnings for keys that cannot be used for Health Card issuance. The same key validation code runs when you validate a key set directly or when one is downloaded while validating a health card. We chose to be overly verbose, to allow developers to be warned about the contents of their key set (before putting them online), to make sure they don't include something in there by mistake.
What you propose is certainly what a real-life verifier should do. We could be more lenient when validating a health card, not displaying the warnings if one key in the set is ok and can be used with a card, but being a developer tool, it feels like we should report as much as we can.
What you propose is certainly what a real-life verifier should do. We could be more lenient when validating a health card, not displaying the warnings if one key in the set is ok and can be used with a card, but being a developer tool, it feels like we should report as much as we can.
Gotcha - I appreciate the nuanced thinking between warning and errors! I would learn towards not including warnings due to unrelated keys, but am fine with either decision here. Feel free to close at your discretion.
It might be worth adding a way for developers to explicitly opt out of certain checks (e.g., a suppression file, or a suppression CLI flag). The first time you get a warning, it might come with advice like:
If you don't want to see this warning in the future you can [...]
While I'm exercising my imagination...the suppression could be scoped to types of errors, or even types of errors combined with data (like, "don't warn me about extra keys" vs "don't warn me about this extra key).
(The common version of this is for Node testing frameworks, where you can specify --include or --exclude type flags to filter out the set of tests applied.0
It might be worth adding a way for developers to explicitly opt out of certain checks (e.g., a suppression file, or a suppression CLI flag). The first time you get a warning, it might come with advice like:
Yes, and it should be fairly easy to do since our logger takes in specific error codes which should be easy to filter out.
Ok, so I'll close this, and keep @jmandel's suggestion alive on issue #35.
| gharchive/issue | 2021-04-01T22:06:13 | 2025-04-01T06:44:58.480372 | {
"authors": [
"christianpaquin",
"dleve123",
"jmandel"
],
"repo": "microsoft/health-cards-validation-SDK",
"url": "https://github.com/microsoft/health-cards-validation-SDK/issues/33",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1361700697 | Microbit Blutooth doesn't work after adding Jacdac extension
Describe the bug
I'm able to pair microbit with an android smartphone and display the data sent via bluetooth to microbit led display.
bluetooth.onUartDataReceived(serial.delimiters(Delimiters.NewLine), function () {
basic.showString(bluetooth.uartReadUntil(serial.delimiters(Delimiters.NewLine)))
})
bluetooth.startUartService()
basic.forever(function () {
})
Once I add the jacdac extension, can't pair with the microbit or connect already paired. Seems like adding the jacdac extension disables the BLE of the microbit. Jacdac extension version for makecode "microsoft/pxt-jacdac 1.8.22"
Desktop (please complete the following information):
OS: Windows 10
Browser Edge
Smartphone (please complete the following information):
Device: Android
Unfortunately, Jacdac and Bluetooth are incompatible. It seems that we are missing the annotation so that MakeCode refuses to use the 2 togher:
[ ] add flag in pxt.json to prevent using with bluetooth
[ ] update makecode codes
From what I can see here it did work: https://forum.makecode.com/t/video-use-outdated-iphone-to-code-and-download-jacdac-to-microbit-v2/15152
For debugging it is also often handy to use a wireless connection (as well as for classrooms that have only iPads etc.). Shouldn't be the option available to use it together?
It would be good, if after Jacdac has been used, the device could still communicate via bluetooth to a device. Right now the pairing pattern appears, but it can't be connected/flashed. More can be found here: https://github.com/microsoft/pxt-calliope/issues/257 though it must be changed in Jacdac unfortunately.
Does it not work even with the pairing pattern described in https://support.microbit.org/support/solutions/articles/19000051025-pairing-and-flashing-code-via-bluetooth ?
Does it not work even with the pairing pattern described in https://support.microbit.org/support/solutions/articles/19000051025-pairing-and-flashing-code-via-bluetooth ?
Unfortunately not. Though it is interesting to see that triple reset also doesn't work, just A+B+Reset – but that has no effect! The app doesn't recognize an advertised device.
Does it not work even with the pairing pattern described in https://support.microbit.org/support/solutions/articles/19000051025-pairing-and-flashing-code-via-bluetooth ?
I also made some tests with the Jacdac extension 1.9.25, the micro:bit 2.20 and the micro:bit iOS app 3.10. I was able to reproduce the behavior as mentioned by arijitx and joernalraun. If the micro:bit runs a firmware with the jacdac extension enabled, it is still possible to go into pairing mode by A+B+reset, but also 3x reset button works.
But the app does not allow any pairing or program transfer to an already paired micro:bit and shows a timeout error:
I had a look at the advertised Bluetooth services using the nRF connect app and noticed that a micro:bit with the Jacdac extension in pairing mode is still advertising as a device but does not include in the 'Buttonless DFU with bonds' service, that a micro:bit without the jacdac extension advertises
Without Jacdac Extension
With Jacdac Extension
@pelikhan You mentioned that a firmware update via app should still be possible as long as the pairing mode (pattern) is used, right? Is there a thing that I can still try out? Do the project settings in makecode need to be adjusted?
I suggest reviewing the preprocessor flags for BLE in microbit and flag if we haven't missed one.
| gharchive/issue | 2022-09-05T10:04:23 | 2025-04-01T06:44:58.491438 | {
"authors": [
"arijitx",
"fabianhugo",
"joernalraun",
"pelikhan"
],
"repo": "microsoft/jacdac",
"url": "https://github.com/microsoft/jacdac/issues/1209",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2413275745 | Fix transaction problem in SQL Server Memory DeleteAsync method
Motivation and Context (Why the change? What's the scenario?)
This PR fixes an issue when trying to delete a document from SQL Server Memory.
High level description (Approach, Design)
The DeleteDocumentHandler class calls the memory GetListAsync method and then deletes each returned records, using an async enumeration:
https://github.com/microsoft/kernel-memory/blob/2ff894c99d30531fe83a9134c35ec4608583ef59/service/Core/Handlers/DeleteDocumentHandler.cs#L45-L54
The issue is related to how records are returned in SqlServerMemory.GetListAsync:
https://github.com/microsoft/kernel-memory/blob/2ff894c99d30531fe83a9134c35ec4608583ef59/extensions/SQLServer/SQLServer/SqlServerMemory.cs#L291-L294
In this code, records are yield inside a DataReader loop, so the connection is kept open until the end of the list. As shown above, DeleteDocumentHandler calls DeleteAsync while iterating on records. In case of SQL Server:
https://github.com/microsoft/kernel-memory/blob/2ff894c99d30531fe83a9134c35ec4608583ef59/extensions/SQLServer/SQLServer/SqlServerMemory.cs#L137-L168
This method tries to open a new connection and execute commands within a transaction. Because it is called while executing GetListAsync, there already is an opened connection that is reading data, so it causes a deadlock that ends with a Timeout exception.
This PR fixes the issue modifying the GetListAsync method so that it reads all the data, closes the connection and then returns the list.
@marcominerva I can't reproduce the problem you're describing here. I added a new test here https://github.com/microsoft/kernel-memory/pull/718/files and the test is working fine. Is the test missing something?
I'd like to reproduce the issue before making any change, to be sure the fix works and we won't have future regressions.
@dluc You need to import at least two documents, then try to delete the first one.
I have created a small repro here: https://github.com/marcominerva/KernelMemorySqlServerIssue
Use the first API to upload two different documents, then try to delete the first and you'll obtain a Timeout exception from SQL Server.
I merged in the ItDeletesRecords test I added in #718.
The test is failing on my local machine, using MSSQL docker image. Could you take a look?
@dluc I have just tried the ItDeletesRecords test on this sqlserver-disposefix branch and on my machine it passes correctly (I'm using a LocalDB instance). What kind of failure do you encounter?
Getting a timeout exception on batch upserts. Could you try the run-mssql.sh script which runs the docker image?
Tried with the container created by run-mssql.sh, but still works correctly. It sounds strange that you get a timeout exception on batch upserts (so, you stop before the delete), because upserts don't call GetListAsync, that it the only method I have modified. I have also tried to set the MemoryDbUpsertBatchSize explicitly in DefaultTests.cs:
this._memory = builder
.With(new KernelMemoryConfig
{
DefaultIndexName = "default4tests",
DataIngestion = new()
{
MemoryDbUpsertBatchSize = 64
}
})
Do other SQL Server tests work on your local machine?
Do other SQL Server tests work on your local machine?
some other tests are failing, I'm starting to think if it's a problem with docker and macOS.
On main branch:
On this PR branch:
Execution 1
Execution 2, after restarting the docker image:
Execution 1 log
Microsoft.Data.SqlClient.SqlException: Execution Timeout Expired. The timeout period elapsed prior to completion of the operation or the server is no...
Microsoft.Data.SqlClient.SqlException
Execution Timeout Expired. The timeout period elapsed prior to completion of the operation or the server is not responding.
at Microsoft.Data.SqlClient.SqlCommand.EndExecuteNonQueryAsync(IAsyncResult asyncResult)
at System.Threading.Tasks.TaskFactory`1.FromAsyncCoreLogic(IAsyncResult iar, Func`2 endFunction, Action`1 endAction, Task`1 promise, Boolean requiresSynchronization)
--- End of stack trace from previous location ---
at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+MoveNext() in km/extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 533
at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+MoveNext() in km/extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 543
at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+System.Threading.Tasks.Sources.IValueTaskSource<System.Boolean>.GetResult()
at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertAsync(String index, MemoryRecord record, CancellationToken cancellationToken) in km/extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 435
at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertAsync(String index, MemoryRecord record, CancellationToken cancellationToken) in km/extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 435
at Microsoft.KernelMemory.Handlers.SaveRecordsHandler.SaveRecordAsync(DataPipeline pipeline, IMemoryDb db, MemoryRecord record, HashSet`1 createdIndexes, CancellationToken cancellationToken) in km/service/Core/Handlers/SaveRecordsHandler.cs:line 266
at Microsoft.KernelMemory.Handlers.SaveRecordsHandler.InvokeAsync(DataPipeline pipeline, CancellationToken cancellationToken) in km/service/Core/Handlers/SaveRecordsHandler.cs:line 214
at Microsoft.KernelMemory.Pipeline.InProcessPipelineOrchestrator.RunPipelineAsync(DataPipeline pipeline, CancellationToken cancellationToken) in km/service/Core/Pipeline/InProcessPipelineOrchestrator.cs:line 174
at Microsoft.KernelMemory.Pipeline.BaseOrchestrator.ImportDocumentAsync(String index, DocumentUploadRequest uploadRequest, IContext context, CancellationToken cancellationToken) in km/service/Core/Pipeline/BaseOrchestrator.cs:line 121
at Microsoft.KM.Core.FunctionalTests.DefaultTestCases.RecordDeletionTest.ItDeletesRecords(IKernelMemory memory, IMemoryDb db, Action`1 log) in km/service/tests/Core.FunctionalTests/DefaultTestCases/RecordDeletionTest.cs:line 30
at Microsoft.SQLServer.FunctionalTests.DefaultTests.ItDeletesRecords() in km/extensions/SQLServer/SQLServer.FunctionalTests/DefaultTests.cs:line 119
at Xunit.DependencyInjection.DependencyInjectionTestInvoker.AsyncStack(Task task, Activity activity)
System.ComponentModel.Win32Exception
Unknown error: 258
Exception doesn't have a stacktrace
Execution 2 log
Microsoft.Data.SqlClient.SqlException: Transaction (Process ID 73) was deadlocked on communication buffer resources with another process and has been ...
Microsoft.Data.SqlClient.SqlException
Transaction (Process ID 73) was deadlocked on communication buffer resources with another process and has been chosen as the deadlock victim. Rerun the transaction.
at Microsoft.Data.SqlClient.SqlConnection.OnError(SqlException exception, Boolean breakConnection, Action`1 wrapCloseInAction)
at Microsoft.Data.SqlClient.SqlInternalConnection.OnError(SqlException exception, Boolean breakConnection, Action`1 wrapCloseInAction)
at Microsoft.Data.SqlClient.TdsParser.ThrowExceptionAndWarning(TdsParserStateObject stateObj, SqlCommand command, Boolean callerHasConnectionLock, Boolean asyncClose)
at Microsoft.Data.SqlClient.TdsParser.TryRun(RunBehavior runBehavior, SqlCommand cmdHandler, SqlDataReader dataStream, BulkCopySimpleResultSet bulkCopyHandler, TdsParserStateObject stateObj, Boolean& dataReady)
at Microsoft.Data.SqlClient.SqlCommand.FinishExecuteReader(SqlDataReader ds, RunBehavior runBehavior, String resetOptionsString, Boolean isInternal, Boolean forDescribeParameterEncryption, Boolean shouldCacheForAlwaysEncrypted)
at Microsoft.Data.SqlClient.SqlCommand.CompleteAsyncExecuteReader(Boolean isInternal, Boolean forDescribeParameterEncryption)
at Microsoft.Data.SqlClient.SqlCommand.InternalEndExecuteNonQuery(IAsyncResult asyncResult, Boolean isInternal, String endMethod)
at Microsoft.Data.SqlClient.SqlCommand.EndExecuteNonQueryInternal(IAsyncResult asyncResult)
at Microsoft.Data.SqlClient.SqlCommand.EndExecuteNonQueryAsync(IAsyncResult asyncResult)
at System.Threading.Tasks.TaskFactory`1.FromAsyncCoreLogic(IAsyncResult iar, Func`2 endFunction, Action`1 endAction, Task`1 promise, Boolean requiresSynchronization)
--- End of stack trace from previous location ---
at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+MoveNext() in km//extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 533
at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+MoveNext() in km//extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 543
at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertBatchAsync(String index, IEnumerable`1 records, CancellationToken cancellationToken)+System.Threading.Tasks.Sources.IValueTaskSource<System.Boolean>.GetResult()
at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertAsync(String index, MemoryRecord record, CancellationToken cancellationToken) in km//extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 435
at Microsoft.KernelMemory.MemoryDb.SQLServer.SqlServerMemory.UpsertAsync(String index, MemoryRecord record, CancellationToken cancellationToken) in km//extensions/SQLServer/SQLServer/SqlServerMemory.cs:line 435
at Microsoft.KernelMemory.Handlers.SaveRecordsHandler.SaveRecordAsync(DataPipeline pipeline, IMemoryDb db, MemoryRecord record, HashSet`1 createdIndexes, CancellationToken cancellationToken) in km//service/Core/Handlers/SaveRecordsHandler.cs:line 266
at Microsoft.KernelMemory.Handlers.SaveRecordsHandler.InvokeAsync(DataPipeline pipeline, CancellationToken cancellationToken) in km//service/Core/Handlers/SaveRecordsHandler.cs:line 214
at Microsoft.KernelMemory.Pipeline.InProcessPipelineOrchestrator.RunPipelineAsync(DataPipeline pipeline, CancellationToken cancellationToken) in km//service/Core/Pipeline/InProcessPipelineOrchestrator.cs:line 174
at Microsoft.KernelMemory.Pipeline.BaseOrchestrator.ImportDocumentAsync(String index, DocumentUploadRequest uploadRequest, IContext context, CancellationToken cancellationToken) in km//service/Core/Pipeline/BaseOrchestrator.cs:line 121
at Microsoft.KM.Core.FunctionalTests.DefaultTestCases.RecordDeletionTest.ItDeletesRecords(IKernelMemory memory, IMemoryDb db, Action`1 log) in km//service/tests/Core.FunctionalTests/DefaultTestCases/RecordDeletionTest.cs:line 30
at Microsoft.SQLServer.FunctionalTests.DefaultTests.ItDeletesRecords() in km//extensions/SQLServer/SQLServer.FunctionalTests/DefaultTests.cs:line 119
at Xunit.DependencyInjection.DependencyInjectionTestInvoker.AsyncStack(Task task, Activity activity)
| gharchive/pull-request | 2024-07-17T10:47:01 | 2025-04-01T06:44:58.508094 | {
"authors": [
"dluc",
"marcominerva"
],
"repo": "microsoft/kernel-memory",
"url": "https://github.com/microsoft/kernel-memory/pull/712",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1049045442 | Partial Revert of #608, Updating Min Target to Ensure 8.1 Backcompat
#608 Changed the WindowsTargetPlatform to hard target Windows 8.1, so that Psychonauts 2 would be able to run on Windows 7. However, this change breaks XAL build pipelines, as our Windows2019 images don't support this scenario. We can do a partial revert and add Windows 8.1 backwards compatibility by specifying a MinTargetVersion
Since I didn't create a new branch for this fix, it's re-adding the commits for updating the Android NDK. That was already merged in #631
| gharchive/pull-request | 2021-11-09T19:52:10 | 2025-04-01T06:44:58.510063 | {
"authors": [
"SahilAshar"
],
"repo": "microsoft/libHttpClient",
"url": "https://github.com/microsoft/libHttpClient/pull/632",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
449810864 | export the blocks to DXF (BLOCKS section of the DXF file)
How to output blocks with primitives when exporting to DXF?
Hi @PodgPow , I may also be needing to have blocks exported, when #133 is fully implemented. Can you add details about your use case?
I want to use blocks for exporting to the DXF path, consisting of several circuits (in SVG everything is fine with this).
Can you say what advantage it is to have things in blocks?
One example - is the transfer to a DXF shape with a hole, so that you can work in AutoCAD (ArtCAM) with the shape as a solid (not fragmented by contour) object
I would like to use makerjs to produce parametric dxf AAMA files for CNC textile cutters. (Lectra Gerber etc)
Some of the software used by these machines requires dxf files with blocks to separate the patterns from one another.
There are other options where each pattern could be exported as it's own dxf file but I'd prefer to keep it all together if possible so that I can use the file to reference the parts together.
It looks like all the magic is happening in the packages/maker.js/src/core/dxf.ts file. I've downloaded the Autodesk DXF reference but honestly, it's pretty overwhelming.
It looks like there needs to be a blocks section added to the exported file and then the block references can be included as an entity to inside the drawing with it's own origin?
Is there a simpler reference that I'm missing somewhere?
Thanks
HI @cgowen , you're not missing anything - the DXF format is not super easy.
I've done a small amount of digging through Audodesk references and files I exported with the exdxf library and have come up with a short breakdown of how to add blocks to a dxf file. I still don't know enough javascript to make this happen.
I'm hoping that having the file structure may be helpful to someone who might be a better coder than I am.
https://anonfiles.com/w5DaHbm2uc/DXF_Format_Cheatsheet_pdf
https://anonfiles.com/u8D7H4mfu7/handmade_dxf
| gharchive/issue | 2019-05-29T13:32:20 | 2025-04-01T06:44:58.524001 | {
"authors": [
"PodgPow",
"PowPodg",
"cgowen",
"danmarshall"
],
"repo": "microsoft/maker.js",
"url": "https://github.com/microsoft/maker.js/issues/411",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
554476418 | TabView should not allow Narrator users to deselect current tab
Describe the bug
Steps to reproduce the bug
Steps to reproduce the behavior:
Turn on Narrator.
Launch XAML Controls Gallery.
Activate Tabview from All Controls Page.
Navigate to any tab and press caps+enter to select it now again caps+enter and observe.
Expected behavior
Nothing should happen, the de-selection should be ignored.
Screenshots
Version Info
Xaml Controls Gallery version 1.2.12.0.
NuGet package version:
Additional context
Copied from internal bug
I think this issue was fixed with the recent changes we made to the TabView automationpeer. This behavior is not reproducable in the MUXControlsTestApp anymore, so I think it's fine to close it now.
@StephenLPeters @ranjeshj FYI
| gharchive/issue | 2020-01-24T00:01:39 | 2025-04-01T06:44:58.529552 | {
"authors": [
"YuliKl",
"chingucoding"
],
"repo": "microsoft/microsoft-ui-xaml",
"url": "https://github.com/microsoft/microsoft-ui-xaml/issues/1880",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
559210864 | URI encoded references are not resolved in JSON schema
VSCode version: 1.41.1
Browser: Electron, Chrome
OS: MacOS
This schema encoded special characters in the reference URI. The standard says that references must be valid URIs: https://json-schema.org/draft/2019-09/json-schema-core.html#ref and can, therefore, be encoded.
{
"$schema": "http://json-schema.org/draft-07/schema#",
"definitions": {
"Foo<number>": {
"type": "object",
"properties": {"q1": {"enum": ["x1", "x2"]}}
}
},
"type": "object",
"properties": {
"p1": {"enum": ["v1", "v2"]},
"p2": {"$ref": "#/definitions/Foo%3Cnumber%3E"}
}
}
Monaco doesn't correctly resolve the references.
{
"$schema": "https://gist.githubusercontent.com/domoritz/3cd0ddb8cad62ad611f301b2a8743ebf/raw/3bc4ff9350a8e9a7acb54d9683ce3a6d3c179377/schema.json",
}
duplicate of https://github.com/microsoft/vscode-json-languageservice/issues/49
Ahh, that’s where I filed the issue. Thanks.
| gharchive/issue | 2020-02-03T16:59:21 | 2025-04-01T06:44:58.541139 | {
"authors": [
"aeschli",
"domoritz"
],
"repo": "microsoft/monaco-editor",
"url": "https://github.com/microsoft/monaco-editor/issues/1804",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
446833176 | Fix C# sample code
String literal does not need to be @-escaped.
Alternatively, one could remove one of the backslashes in the line.
The @ is used to test/show off the colorizer.
| gharchive/pull-request | 2019-05-21T21:26:52 | 2025-04-01T06:44:58.542346 | {
"authors": [
"DaMightyZombie",
"alexandrudima"
],
"repo": "microsoft/monaco-editor",
"url": "https://github.com/microsoft/monaco-editor/pull/1450",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2140347358 | Netperf version 1.5 TODO: PGO/SPGO Support
We need to add support for PGO runs and then the automation to push back the PGO files to in a GitHub PR.
Mentioning SPGO here; collecting and pushing to SPGO is another option, we can do one or the other or both.
| gharchive/issue | 2024-02-17T17:29:51 | 2025-04-01T06:44:58.547228 | {
"authors": [
"mtfriesen",
"nibanks"
],
"repo": "microsoft/netperf",
"url": "https://github.com/microsoft/netperf/issues/76",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2122979608 | Is Memory supported as return type in dotnet?
I tried to use Memory<byte> as return type of a method on dotnet side and call the method from Javascript.
However, although the generated .d.ts file contains the method signature, the Javascript code is not able to call that method. The error message says the object does not have such method/property/function. I tried to change the return type to string or byte[] then the method could be called from Javascript.
Is Memory<byte> supported?
BTW, I suspected marshaller code needs to be generated and utlised, however, I couldn't find any generated file. I've created another issue for that: #207
A .NET Memory<byte> should get converted to JS UInt8Array, in both typedefs and runtime.
But there might be some other reason you're unable to call the method.
Can you check what properties are defined on the JS object? Object.keys(obj)
Do you see other expected methods/properties on the object?
How are you getting the JS instance of the class?
Can you check what properties are defined on the JS object? Object.keys(obj)
If you mean the JS object returned from the .NET method I want to call, because the method can't be called (the JS class instance does not have such function on it) there is no way to get the returned value.
Do you see other expected methods/properties on the object?
Other methods exist on the class instance as long as their return types are not Memory<T>. I tried changing the return type from Memory<byte> to string or byte[], both made that method available to Node.js.
How are you getting the JS instance of the class?
The type definition was generated through MSBuild.
I believe the class instance was created through reflection becaue I suspect the code generate didn't work (as described in #207 )
I hope there is a way to tell whether refection is used or generated code is used.
I suspect the missing of that method is caused by the missing of generated code.
BTW, currently I use base64 encoded string to pass binary data from .NET to Node.js as a workaround.
I confirmed in #220 with more testing that Memory<byte> is working as expected. So there is likely some other problem here.
the JS class instance does not have such function on it
This sounds like there was some other problem with the method that is unrelated to the fact that it returns Memory<byte>. If you want to troubleshoot further, it might help to share relevant snippets of the code.
I'm closing this for now since the originally reported problem is not reproducible. But feel free to re-open with more information if you're still having this problem.
| gharchive/issue | 2024-02-07T12:57:12 | 2025-04-01T06:44:58.555289 | {
"authors": [
"james-hu",
"jasongin"
],
"repo": "microsoft/node-api-dotnet",
"url": "https://github.com/microsoft/node-api-dotnet/issues/208",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1444570283 | Release 5.20.0
Summary of the Pull Request
What is this about?
Release 5.20.0
Added
Service: Added endpoint to download agent binaries to support the unmanaged node scenario. #2600
Service: Added additional error handling when updating VMSS nodes. #2607
Changed
Service: Added additional logging when using the decommission node policy. #2605
Agent/Supervisor/Proxy: Updated third-party Rust dependencies.#2608
Service: Added optional retry_limit when connecting to the repro machine. #2609
Fixed
Service: Fixed status top in C# implementation. #2604
Codecov Report
Merging #2614 (5fecc23) into main (ff85b80) will decrease coverage by 4.97%.
The diff coverage is 17.94%.
@@ Coverage Diff @@
## main #2614 +/- ##
==========================================
- Coverage 29.88% 24.91% -4.98%
==========================================
Files 290 121 -169
Lines 35846 12533 -23313
==========================================
- Hits 10714 3122 -7592
+ Misses 25132 9411 -15721
Impacted Files
Coverage Δ
src/ApiService/ApiService/Functions/ReproVmss.cs
0.00% <0.00%> (ø)
src/ApiService/ApiService/Functions/Tasks.cs
22.22% <0.00%> (ø)
src/ApiService/ApiService/UserCredentials.cs
8.82% <8.33%> (+2.26%)
:arrow_up:
...ice/ApiService/onefuzzlib/EndpointAuthorization.cs
22.13% <11.11%> (-0.76%)
:arrow_down:
src/ApiService/ApiService/Functions/Jobs.cs
76.59% <100.00%> (ø)
src/ApiService/ApiService/OneFuzzTypes/Model.cs
72.05% <100.00%> (+0.07%)
:arrow_up:
src/agent/coverage/src/lib.rs
src/agent/onefuzz-telemetry/src/lib.rs
src/agent/onefuzz-agent/src/config.rs
src/agent/coverage/src/cobertura.rs
... and 165 more
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
| gharchive/pull-request | 2022-11-10T21:14:07 | 2025-04-01T06:44:58.572774 | {
"authors": [
"codecov-commenter",
"mgreisen"
],
"repo": "microsoft/onefuzz",
"url": "https://github.com/microsoft/onefuzz/pull/2614",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2123977263 | No information when tokenizer can't load config files
To repro, run the following code, without the Llama tokenizer.model available.
import onnxruntime_genai as og
import time
print("Loading model...")
model=og.Model("model", og.DeviceType.CPU)
print("Model loaded")
tokenizer=model.create_tokenizer()
print("Tokenizer created")
# Keep asking for input prompts in an loop
while True:
text = input("Input:")
input_tokens = tokenizer.encode(text)
params=og.search_params(model)
params.max_length = 64
params.input_ids = input_tokens
start_time=time.time()
output_tokens=model.generate(params)
run_time=time.time()-start_time;
print(f"Tokens: {len(output_tokens)} Time: {run_time:.2f} Tokens per second: {len(output_tokens)/run_time:.2f}")
print("Output:")
print(tokenizer.decode(output_tokens))
print()
print()
tokenizer=model.create_tokenizer()
RuntimeError
Fixed 2/26/24
| gharchive/issue | 2024-02-07T22:04:10 | 2025-04-01T06:44:58.574678 | {
"authors": [
"natke"
],
"repo": "microsoft/onnxruntime-genai",
"url": "https://github.com/microsoft/onnxruntime-genai/issues/65",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2753136655 | Figure out what cr8 should do in AccessVpState
AccessVpState stubs out cr8 read/writes. Figure out if/how we should wire this up. Supposedly it's only used for the instruction emulator, which doesn't need cr8.
These error tracing calls will be downgraded to trace. There's also updates to rflags in this path
For CR8 @chris-oo thinks when we fix the instruction emulator to only grab state it needs, this goes away, but we need to confirm
| gharchive/issue | 2024-12-20T17:12:33 | 2025-04-01T06:44:58.668390 | {
"authors": [
"cperezvargas"
],
"repo": "microsoft/openvmm",
"url": "https://github.com/microsoft/openvmm/issues/564",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1169023939 | Non-blocking EventPipeEventSource Ctor
This PR makes the constructor for EventPipeEventSource non-blocking. Previously, the constructor would block until data was sent on the stream. This brings EventPipeEventSource closer in behavior to other similar APIs in the TraceEvent library.
One issue that isn't solved by this PR (yet) is that we need to handle the case where users try to get metadata about the source before data arrives. Before this PR, users were guaranteed to have that information when the ctor returned. Now, this is only true for opening a completed file, hence the special casing for files. Currently, this PR makes no attempt to stop users from reading metadata properties. I am hoping for some feedback on whether this should throw an exception or now. Until this is decided, I'll leave this in draft mode.
Currently, this PR makes no attempt to stop users from reading metadata properties
I'd say throwing is desirable if the values aren't initialized yet, but if it was a significant hassle to implement I don't think it would be awful to return some default 0/null/empty string values instead. What do you think @brianrob?
I'd say throwing is desirable if the values aren't initialized yet
Agreed.
There are Debug.Asserts scattered around that check whether this metadata is set before certain operations. These are mostly buried in the TraceLog APIs from what I saw. These caused the file-based tests to fail until I special cased the file-based ctor.
Depending on @brianrob's thoughts, I'll look at making these properties throw if Process hasn't been called yet.
I would tend to agree that if there is metadata that can't be filled before returning from the constructor, then throwing is a good idea. If you look at how we handle realtime ETW sessions, the session is created and setup before you get the source from it, which means that the source can make calls to the session to get metadata. Would such a pattern help here? If not, I would say throwing is the next best thing.
Would such a pattern help here?
I think yes, however, the API for configuring, starting, and stopping an EventPipe session is in the diagnostics client library and not in TraceEvent. That makes it a little more complicated to make this pattern happen.
Right now, the flow for "trace on startup" is this:
[Diag Client Lib] Create configuration (list of providers, buffersize, etc.)
[Diag Client Lib] Start EventPipe session (get Stream object)
[TraceEvent lib] Create EventPipeEventSource using EventPipeEventSource(Stream stream) ctor
[TraceEvent lib] Call EventPipeEventSource.Process()
[Diag Client lib] Resume target
[Diag Client lib] Stop EventPipe session
In the "trace on startup" scenario, we need to wait for the target process to resume before the session starts. Without this change, the ctor call in step 3 would block until we resume the target. After this change, the ctor would complete, but the metadata wouldn't be populated yet. Any usage of that EventPipeEventSource before resuming the target could cause errors in other parts of TraceEvent regardless of whether EventPipeEventSource.Process has been called. The metadata should only be present after step 5.
Gotcha. So this becomes a bit more complicated.
I think it is OK to simplify and just throw in APIs that don't have data yet. Just wanted to make sure that there wasn't another pattern that might help.
@josalem, I see that you just pushed another commit here. Please let me know when this is ready for another review, or is ready for merge. Thanks.
I think this is ready for final review and merge at your discretion.
| gharchive/pull-request | 2022-03-14T23:38:49 | 2025-04-01T06:44:58.676646 | {
"authors": [
"brianrob",
"josalem",
"noahfalk"
],
"repo": "microsoft/perfview",
"url": "https://github.com/microsoft/perfview/pull/1588",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1084939105 | [Feature]: Modification of network response
Feature request
Follow-up on #1816
This is supported and documented in the Node.js version of playwright.
Node.js section: https://playwright.dev/docs/network#modify-responses
It does not seem exist for the .NET version.
Merging into https://github.com/microsoft/playwright-dotnet/issues/1905
| gharchive/issue | 2021-12-20T15:47:56 | 2025-04-01T06:44:58.679584 | {
"authors": [
"krokofant",
"mxschmitt"
],
"repo": "microsoft/playwright-dotnet",
"url": "https://github.com/microsoft/playwright-dotnet/issues/1904",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2344983315 | FIle Upload box doesn't appear.
Version
1.44.0
Steps to reproduce
In a app that I am using, file upload box doesn't appear when clicking the choose file element as below:
page.waitForFileChooser(() -> locator1.click();
Tried with set force true but same issues
page.waitForFileChooser(() -> locator1.click(new Locator.ClickOptions().setForce(true));
When I click the choose file with a simple click with the above waitforfilechooser wrapper method, it works just fine and opens the file upload dialog.
locator1.click(new Locator.ClickOptions().setForce(true))
Any idea what the issue could be.
Expected behavior
FIle upload box should appear
Actual behavior
File upload box not appearing
Additional context
No response
Environment
Mac OS
Playwright 1.44.0
Java 17
Chromium
Please follow our bug template and provide a minimal self-contained project that we could run locally to reproduce the problem.
| gharchive/issue | 2024-06-10T22:49:59 | 2025-04-01T06:44:58.683596 | {
"authors": [
"maheshwg",
"yury-s"
],
"repo": "microsoft/playwright-java",
"url": "https://github.com/microsoft/playwright-java/issues/1596",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1186807691 | Hello, boss. I think this code will overflow memory after running for a few hours? Excuse me, where is the problem? Please check it for me. I have been checking this code for a long time and can't find the cause of memory overflow
package com.artisan.spider;
import com.artisan.spider.consts.Consts;
import com.artisan.spider.domain.SpError;
import com.artisan.spider.domain.SpLawChl;
import com.artisan.spider.domain.SpLawLar;
import com.artisan.spider.service.ISpErrorService;
import com.artisan.spider.service.impl.SpLawChlServiceImpl;
import com.artisan.spider.service.impl.SpLawLarServiceImpl;
import com.artisan.spider.util.DateUtil;
import com.artisan.spider.util.SpiderUtil;
import com.artisan.spider.util.SpringContextUtil;
import com.microsoft.playwright.Browser;
import com.microsoft.playwright.BrowserContext;
import com.microsoft.playwright.Page;
import com.microsoft.playwright.Playwright;
import lombok.extern.slf4j.Slf4j;
import org.junit.jupiter.api.Test;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.boot.test.context.SpringBootTest;
import java.util.ArrayList;
import java.util.List;
@SpringBootTest
@Slf4j
public class DffggzDfgfxwjTaskTest01 {
@Autowired
ISpErrorService iSpErrorService;
@Test
public void collectionInfo(){
try (Playwright playwright = Playwright.create()) {
String pageUrl = Consts.dfgfxwj;
Browser browser = playwright.chromium().launch();
BrowserContext context = browser.newContext();
Page page = context.newPage();
page.navigate(String.format(pageUrl, 1));
int pageNum = Integer.valueOf(page.querySelector(".page").innerText().split("/")[1].replace("页", "").trim());
for (int i = 1; i <= 3; i++) {
SpiderUtil.pageChuli("dfflgz",context,page, pageUrl, i, pageNum);
}
}
}
}
`
`public static void pageChuli(String mark,BrowserContext browserContext, Page page, String pageUrl, int currentNum, int pageNum) {
Locator rows = page.locator("td a");
int count = rows.count();
log.info("每页的大小--》{},当前页数{}", count,currentNum);
for (int k = 0; k < count; k++) {
// 判断是不是曾经抓取过
QueryWrapper<SpLawChl> queryWrapper=new QueryWrapper();
queryWrapper.eq("page_url",rows.nth(k).getAttribute("href"));
int lawNumber=SpringContextUtil.getBean(ISpLawChlService.class).count(queryWrapper);
if(lawNumber==0){
Page detail = null;
try{
detail= browserContext.newPage();
String herf= Consts.rootUrl+"/"+rows.nth(k).getAttribute("href");
detail.navigate(herf);
log.info("第{}页-第{}篇文章的标题----->{},url为{},", currentNum, k + 1, rows.nth(k).textContent(),herf);
String title=detail.locator(".qw-bt").textContent();
// 发布部门也可能存在多个
Locator fbbmLocator = detail.locator("td:right-of(:text(\"发布部门\"))");
String fbbm = "";
String fbbmCode = "";
if (fbbmLocator.count() != 0) {
Locator lb=fbbmLocator.nth(0).locator("a");
if (lb.count() != 0) {
if (lb.count() > 1) {
for (int i = 0; i < lb.count(); i++) {
Locator locator=lb.nth(i);
fbbmCode += getType(locator,"fdep_id","&menuname");
fbbm +=locator.textContent();
if(i!= lb.count()-1){
fbbmCode+=",";
fbbm+=",";
}
}
} else {
Locator locator=lb.nth(0);
fbbm = locator.textContent();
fbbmCode=getType(locator,"fdep_id","&menuname");
}
}
}
String fwzh = "";
Locator fbzhLocator = detail.locator("td:right-of(:text(\"发文字号\"))");
if (fbzhLocator.count() != 0) {
fwzh = fbzhLocator.nth(0).textContent();
}
String fbrq = "";
Locator fbrqLocator = detail.locator("td:right-of(:text(\"发布日期\"))");
if (fbrqLocator.count() != 0) {
fbrq = fbrqLocator.nth(0).textContent();
}
String ssrq = "";
Locator ssrqLocator = detail.locator("td:right-of(:text(\"实施日期\"))");
if (ssrqLocator.count() != 0) {
ssrq = ssrqLocator.nth(0).textContent();
}
String ppbm = "";
String ppbmCode = "";
Locator ppbmLocator = detail.locator("td:right-of(:text(\"批准部门\"))");
if (ppbmLocator.count() != 0) {
Locator lb=ppbmLocator.nth(0).locator("a");
if (lb.count() != 0) {
if (lb.count() > 1) {
for (int i = 0; i < lb.count(); i++) {
Locator locator=lb.nth(i);
ppbmCode += getType(locator,"pdep_id","&menuname");
ppbm +=locator.textContent();
if(i!= lb.count()-1){
ppbmCode+=",";
ppbm+=",";
}
}
} else {
Locator locator=lb.nth(0);
ppbm = locator.textContent();
ppbmCode=getType(locator,"pdep_id","&menuname");
}
}
}
String sxx = "";
String sxxCode="";
Locator sxxLocator = detail.locator("td:right-of(:text(\"时效性\"))");
if (sxxLocator.count() != 0) {
Locator locator=sxxLocator.nth(0).locator("a");
sxx = locator.textContent();
sxxCode=getType(locator,"shixiao_id","&menuname");
}
String xljb = "";
String xljbCode="";
Locator xljbLocator = detail.locator("td:right-of(:text(\"效力级别\"))");
if (xljbLocator.count() != 0) {
Locator locator=xljbLocator.nth(0).locator("a");
xljb = locator.textContent();
xljbCode=SpiderUtil.getType(locator,"xiaoli_id","&menuname");
}
// 法规类别有可能是多个
String fglb = "";
String fglbCode="";
Locator fglbLocator = detail.locator("td:right-of(:text(\"法规类别\"))");
if (fglbLocator.count() != 0) {
// 如果法规类别有多个
Locator lb = fglbLocator.nth(0).locator("a");
if (lb.count() != 0) {
if (lb.count() > 0) {
for (int i = 0; i < lb.count(); i++) {
Locator locator=lb.nth(i);
fglb += locator.textContent();
fglbCode += getType(locator,"sort_id","&menuname");
if(i!= lb.count()-1){
fglb+=",";
fglbCode+=",";
}
}
} else {
Locator locator=lb.nth(0);
fglb = locator.textContent();
fglbCode=getType(locator,"sort_id","&menuname");
}
}
}
String content = detail.locator("#div_content").innerHTML();
log.info("发布部门名称,编码--》{},{}", fbbm,fbbmCode);
log.info("发文字号--》{}", fwzh);
log.info("发布日期--》{}", fbrq);
log.info("实施日期--》{}", ssrq);
log.info("时效性名称,编码--》{},{}", sxx,sxxCode);
log.info("效力级别名称,编码--》{},{}", xljb,xljbCode);
log.info("法规类别名称,编码--》{}", fglb,fglbCode);
log.info("*******************************第{}页-第{}篇文章处理完成", currentNum, k + 1);
SpLawChl spLawChl=new SpLawChl();
spLawChl.setId(UuidUtils.getUUid());
spLawChl.setFbDeptName(fbbm);
spLawChl.setFbDeptCode(fbbmCode);
spLawChl.setPzDeptName(ppbm);
spLawChl.setPzDeptCode(ppbmCode);
spLawChl.setLssuedNumber(fwzh);
spLawChl.setReleaseDate(fbrq);
spLawChl.setTimeName(sxx);
spLawChl.setTimeCode(sxxCode);
spLawChl.setXiaoliName(xljb);
spLawChl.setXiaoliCode(xljbCode);
spLawChl.setFaguiName(fglb);
spLawChl.setFaguiCode(fglbCode);
spLawChl.setImplDate(ssrq);
spLawChl.setTitle(title);
spLawChl.setFullText(content);
spLawChl.setStatus("1");
spLawChl.setCreateTime(DateUtil.getCurrentTime());
spLawChl.setPageUrl(rows.nth(k).getAttribute("href"));
spLawChl.setFgfl(mark);
spLawChl.setIsCreate("0");
SpringContextUtil.getBean(ISpLawChlService.class).save(spLawChl);
// 把数据发布到相应的消息对列中
SpringContextUtil.getBean(KafkaTemplate.class).send("messageQueen", spLawChl.getId());
detail.close();
}catch (Exception e){
e.printStackTrace();
detail.close();
log.info("*******************************第{}页-第{}篇文章处理失败", currentNum, k + 1);
log.info("错误链接:"+Consts.rootUrl+"/"+rows.nth(k).getAttribute("href"));
SpError spError=new SpError();
spError.setCreateTime(DateUtil.getCurrentTime());
spError.setPageUrl(rows.nth(k).getAttribute("href"));
SpringContextUtil.getBean(ISpErrorService.class).save(spError);
continue;
}
}else{
continue;
}
}
// 跳转列表页
if (pageNum != currentNum) {
page.navigate(String.format(pageUrl, currentNum + 1));
}
}
In the inner loop you reuse one and the same BrowserContext for all pages( detail= browserContext.newPage()), it may hold some of the network resources until it's closed. We usually recommend working with each page in a new context for better isolation and so that you don't unnecessarily consume memory. The example is quite generic and impossible to run locally so there is not much we can help with except for general suggestions based on our understanding of the code.
In the inner loop you reuse one and the same BrowserContext for all pages( detail= browserContext.newPage()), it may hold some of the network resources until it's closed. We usually recommend working with each page in a new context for better isolation and so that you don't unnecessarily consume memory. The example is quite generic and impossible to run locally so there is not much we can help with except for general suggestions based on our understanding of the code.
Thank you. I'm trying to modify the experiment
Hello, boss, because I'm multi-layer for loop nesting, can I reuse one browsercontext for 40 pages? The previous list page was 10000 pages, with 40 articles per page, so it may be the problem of reusing browsercontext you said, resulting in memory overflow
Hello, after testing, the memory will still rise..
Thank you. It seems that the problem has been solved. After putting it on the server, the memory is now normal after a few hours
| gharchive/issue | 2022-03-30T18:05:23 | 2025-04-01T06:44:58.692175 | {
"authors": [
"wasd345",
"yury-s"
],
"repo": "microsoft/playwright-java",
"url": "https://github.com/microsoft/playwright-java/issues/875",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
858066873 | [BUG]: Not Getting Capacity Alerts if Approved Capacity is Null
Describe the bug
If the Approved Capacity is null, then you do not get any Capacity Alerts
Component (please tell us which flow or app you are experiencing issues with):
Solution Core
Version 2.0
App or Flow Get Capacity Alerts
To Reproduce
Steps to reproduce the behavior:
set the Approved Capacity to null
Expected behavior
If the actual capacity is > 0 and is less than the approved capacity is null, then I would have expected there to have been a Capacity Alert. There is no capacity alert if the approved capacity is null. This is because the "List Environment Capacity Information" has a filter in it which filters out record where the approved capacity is null (See screenshot below)
Screenshots
Hello. This is by design. We do not want to send alerts unless the admin has explicitly decided to add capacity limits for the environments. You just need to enter these to be some default value if you want to have alerts for all your environments.
Thanks for using CoE
How do you add these default values ?
From: Jenefer Monroe @.>
Sent: 14 April 2021 18:18
To: microsoft/powerapps-tools @.>
Cc: Nigel Price @.>; Author @.>
Subject: Re: [microsoft/powerapps-tools] [BUG]: Not Getting Capacity Alerts if Approved Capacity is Null (#932)
Hello. This is by design. We do not want to send alerts unless the admin has explicitly decided to add capacity limits for the environments. You just need to enter these to be some default value if you want to have alerts for all your environments.
Thanks for using CoE
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHubhttps://github.com/microsoft/powerapps-tools/issues/932#issuecomment-819684766, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ACAXJUEOYVLMD3LWGCUIA5LTIXFADANCNFSM425XVHBQ.
If you're looking for a daily capacity report, you may find it easier to set up a separate flow following these steps: https://docs.microsoft.com/en-us/power-platform/admin/programmability-tutorial-create-daily-capacity-report
| gharchive/issue | 2021-04-14T16:32:58 | 2025-04-01T06:44:58.757063 | {
"authors": [
"JeneferM-MSFT",
"NPrice99",
"manuelap-msft"
],
"repo": "microsoft/powerapps-tools",
"url": "https://github.com/microsoft/powerapps-tools/issues/932",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2332237275 | Invalid instrumentation key
Describe the bug
I am following this tutorial: https://learn.microsoft.com/en-us/azure/machine-learning/prompt-flow/how-to-trace-local-sdk?view=azureml-api-2&tabs=python
I logged in using azure cli and then entered the command to set up my trace destination:
pf config set trace.destination=azureml://subscriptions/<your_subscription_id>/resourcegroups/<your_resourcegroup_name>/providers/Microsoft.MachineLearningServices/workspaces//<your_azureml_workspace_name>
This created a CosmosDB. I promptflow does not do this every time I run it on different machines.
I tried to send traces using the trace-autogen-groupchat.ipynb, and I got this error:
ERROR:opencensus.ext.azure.common.transport:Non-retryable server side error 400: {"itemsReceived":2,"itemsAccepted":0,"appId":null,"errors":[{"index":0,"statusCode":400,"message":"Invalid instrumentation key"},{"index":1,"statusCode":400,"message":"Invalid instrumentation key"}]}.
ERROR:opencensus.ext.azure.common.transport:Non-retryable server side error 400: {"itemsReceived":1,"itemsAccepted":0,"appId":null,"errors":[{"index":0,"statusCode":400,"message":"Invalid instrumentation key"}]}.
How To Reproduce the bug
Steps to reproduce the behavior, how frequent can you experience the bug:
Download the trace-autogen-groupchat.ipynb notebook from this repo
Follow the instructions for setting the trace destination https://learn.microsoft.com/en-us/azure/machine-learning/prompt-flow/how-to-trace-local-sdk?view=azureml-api-2&tabs=python
Try to run the code in the autogen section, where agents generate messages and send them using tracing.
Expected behavior
I expected the information to be sent without any error.
Using Windows, Python 11, and the latest version of PromptFlow.
Hi @tyler-suard-parker , thank you for reaching out. I think there are several questions in this issue:
pf config set trace.destination does create a Cosmos DB, but one workspace will only have one Cosmos resource, so I understand when you run same command on different machines, only the first time you will see the Cosmos setup process (and wait for that).
For the error, it's wired to me because we did use opencensus before, but in #2175 , which is merged 3 months ago, we already migrate to Open Telemetry. Could you please share some screenshots for your error? and provide your prompt flow version with pf -v? so that maybe we can re-produce your error and better investigate on that.
| gharchive/issue | 2024-06-03T23:32:11 | 2025-04-01T06:44:58.768681 | {
"authors": [
"tyler-suard-parker",
"zhengfeiwang"
],
"repo": "microsoft/promptflow",
"url": "https://github.com/microsoft/promptflow/issues/3373",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1822982439 | Type Mismatch Errors with @types/react and TypeScript in PXT-Microbit Project
Describe the bug
When trying to build the PXT-Microbit project, a series of TypeScript errors occur related to the @types/react package. The errors indicate that subsequent property declarations must have the same type, which seems to be a mismatch between the types defined in the @types/react package.
To Reproduce
Steps to reproduce the behavior:
Clone the PXT-Microbit repository (https://github.com/microsoft/pxt-microbit).
Run npm install to install the necessary packages.
Run pxt serve to build and serve the project.
See the TypeScript errors related to @types/react.
Expected behavior
The project should build without any TypeScript errors and serve locally for development.
Desktop (please complete the following information):
OS: MacOS
Node.js Version: LTS
TypeScript Version: 4.4.3
@types/react Version: 17.0.5
Additional context
I've tried several solutions to resolve these errors, including updating the @types/react package to the latest version, deleting the node_modules directory and reinstalling the packages, and adding "skipLibCheck": true to the tsconfig.json file. Unfortunately, none of these solutions have resolved the issue
@emoltz could you run node -v and get the version of node.js you are running?
18.17.0, but I tried it with 20.5.0 and v14 via nvm
This is also affecting the build of the pxt repo, i.e. when I run npm run build I get an error related to a type mismatch with the Provider component. This issue seems to extend to all pxt-related repos.
| gharchive/issue | 2023-07-26T18:57:42 | 2025-04-01T06:44:58.783869 | {
"authors": [
"emoltz",
"jwunderl"
],
"repo": "microsoft/pxt",
"url": "https://github.com/microsoft/pxt/issues/9619",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1292898377 | VSCode option to turn off syntax reporting of future Python versions
Hello, I have the following workflow. I remotely connect to a cluster access machine (via VSCode over SSH), where I make changes to my repository. The access machine runs Python 3.4, thus I assume Pylance operates with it to analyze my repo. The issue is that the code that I develop runs on cluster nodes with Python 3.8, and my code extensively uses some of newer Python features like f-strings and type annotations.
Pylance annoingly complains about "Format string literals (f-strings) require Python 3.6 or newer" at hundreds and hundreads of places of my repository. I haven't found any VSCode option to turn these errors off. It would be great if there was an option like
"python.analysis.diagnosticSeverityOverrides": {
"reportFuturePythonSyntax": "none",
},
hellow, sorry for the errors... well one way to solve the problem is to
install a new version of python this can somehow help and the good thing
about this is that you can install multiple python versions so you can
choose which version you need to run your code
On Mon, 4 Jul 2022 at 12:36, Alex Larionov @.***> wrote:
Hello, I have the following workflow. I remotely connect to a cluster
access machine (via VSCode over SSH), where I make changes to my
repository. The access machine runs Python 3.4, thus I assume Pylance
operates with it to analyze my repo. The issue is that the code that I
develop runs on cluster nodes with Python 3.8, and my code extensively uses
some of newer Python features like f-strings.
Pylance annoingly complains about "Format string literals (f-strings)
require Python 3.6 or newer" at hundreds and hundreads of places of my
repository. I haven't found any VSCode option to turn these errors off. It
would be great if there was an option like
"python.analysis.diagnosticSeverityOverrides": {
"reportFuturePythonSyntax": "none",
},
—
Reply to this email directly, view it on GitHub
https://github.com/microsoft/pylance-release/issues/2994, or unsubscribe
https://github.com/notifications/unsubscribe-auth/AYGBJN7AHYYZMUOCXUK4WRDVSKV77ANCNFSM52SRSIFA
.
You are receiving this because you are subscribed to this thread.Message
ID: @.***>
@BerylXavier Hi, as I said, I work on a cluster, and managing its software is beyond my control
By default, pyright (the type checker upon which pylance is built) assumes that you are running a version of python that matches the selected python interpreter. You can override this by creating a pyrightconfig.json file in the root directory of your project and including { "pythonVersion": "3.8" }. For additional configuration options, refer to this documentation.
Works like a charm! Though having unrelated files in the project root annoys a bit
If you would prefer to use a pyproject.toml file, pyright also supports that.
| gharchive/issue | 2022-07-04T09:35:46 | 2025-04-01T06:44:58.791804 | {
"authors": [
"BerylXavier",
"erictraut",
"laralex"
],
"repo": "microsoft/pylance-release",
"url": "https://github.com/microsoft/pylance-release/issues/2994",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
607920860 | Package Microsoft.Quantum.ProjectTemplates not found
Hello,
I'm trying to install QKD for developing in C# and command line. So I installed the .NET Core SDK version 3.1.100 x64, then I tried to install templates with command line :
dotnet new -i Microsoft.Quantum.ProjectTemplates
But this command line give me an error :
C:\Users\Julien.templateengine\dotnetcli\v3.1.100\scratch\restore.csproj : error NU1101: Package Microsoft.Quantum.ProjectTemplates not found. no package associated to this ID exists in ressources : C:\Program Files\dotnet\sdk\NuGetFallbackFolder
Restoration failed 195,48 ms for C:\Users\Julien.templateengine\dotnetcli\v3.1.100\scratch\restore.csproj.
(I have translated it from french).
My computer is Windows 10 x64.
I made exactly same steps on an other computer (Windows 10 x64 also with same .NET SDK version) and it works well and I'm able to run the "hello world" Q# sample.
I suspect the reason is because I have Visual Studio 2013 installed on the computer, do you think it could be the reason ? Any help to solve my problem ?
Thank you very much !
Best regards.
Julien.
Thanks for your patience. Since this concerns the project templates, moved over to the qsharp-compiler repo where those are hosted. It looks like your current NuGet.Config should work, such that I'm a bit confused by the errors that you've listed. That said, that's also the default configuration for NuGet, such that I'd suggest temporarily moving your config file somewhere else and seeing if it works that way.
| gharchive/issue | 2020-01-16T20:55:40 | 2025-04-01T06:44:58.807251 | {
"authors": [
"JulesMhz",
"cgranade"
],
"repo": "microsoft/qsharp-compiler",
"url": "https://github.com/microsoft/qsharp-compiler/issues/420",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1337743198 | How to execute my own queries in BERT
When we use the BERT model to preprocess our own queries (already transformed by the spider). We found that the preprocessing cannot crate a new dev.jsonl file. What should we do?
Hey @robinzixuan where you able to figure out the reason I am facing the same issue but for some databases not for all.
| gharchive/issue | 2022-08-12T22:09:13 | 2025-04-01T06:44:58.808466 | {
"authors": [
"pmane-uptycs",
"robinzixuan"
],
"repo": "microsoft/rat-sql",
"url": "https://github.com/microsoft/rat-sql/issues/69",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1281505931 | Update CODEOWNERS
Description
Removing @riedgar-ms from CODEOWNERS file. Adding e2e package with @vinuthakaranth, @tongyu-microsoft and myself as owners.
Checklist
[ ] I have added screenshots above for all UI changes.
[ ] I have added e2e tests for all UI changes.
[ ] Documentation was updated if it was needed.
Codecov Report
Merging #1513 (68cc074) into main (46d1ab7) will decrease coverage by 11.69%.
The diff coverage is n/a.
@@ Coverage Diff @@
## main #1513 +/- ##
===========================================
- Coverage 87.27% 75.58% -11.70%
===========================================
Files 108 10 -98
Lines 5108 258 -4850
===========================================
- Hits 4458 195 -4263
+ Misses 650 63 -587
Flag
Coverage Δ
unittests
75.58% <ø> (-11.70%)
:arrow_down:
Flags with carried forward coverage won't be shown. Click here to find out more.
Impacted Files
Coverage Δ
responsibleai/responsibleai/_managers/__init__.py
responsibleai/responsibleai/_input_processing.py
...ponsibleai/responsibleai/_managers/base_manager.py
raiwidgets/raiwidgets/cohort.py
raiwidgets/raiwidgets/__init__.py
raiwidgets/raiwidgets/fairness_dashboard.py
raiutils/raiutils/common/__init__.py
...iwidgets/raiwidgets/fairness_metric_calculation.py
...iwidgets/raiwidgets/explanation_dashboard_input.py
responsibleai/responsibleai/__init__.py
... and 88 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 46d1ab7...68cc074. Read the comment docs.
| gharchive/pull-request | 2022-06-23T01:58:59 | 2025-04-01T06:44:58.833279 | {
"authors": [
"codecov-commenter",
"romanlutz"
],
"repo": "microsoft/responsible-ai-toolbox",
"url": "https://github.com/microsoft/responsible-ai-toolbox/pull/1513",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1521884248 | [Big data] Counterfactuals - part 1
Signed-off-by: vinutha karanth vinutha.karanth@gmail.com
Description
This PR contains all the other changes needed for counterfactuals other than changes in libs/counterfactuals folder.
Next part will contain all changes in libs/counterfactuals folder.
Checklist
[ ] I have added screenshots above for all UI changes.
[ ] I have added e2e tests for all UI changes.
[ ] Documentation was updated if it was needed.
Codecov Report
Merging #1888 (83132f9) into main (a4920bf) will decrease coverage by 7.48%.
The diff coverage is n/a.
@@ Coverage Diff @@
## main #1888 +/- ##
==========================================
- Coverage 93.33% 85.85% -7.49%
==========================================
Files 93 29 -64
Lines 4559 523 -4036
==========================================
- Hits 4255 449 -3806
+ Misses 304 74 -230
Flag
Coverage Δ
unittests
85.85% <ø> (-7.49%)
:arrow_down:
Flags with carried forward coverage won't be shown. Click here to find out more.
Impacted Files
Coverage Δ
...onsibleai/responsibleai/serialization_utilities.py
...ibleai/_tools/shared/state_directory_management.py
...s/erroranalysis/_internal/error_report/__init__.py
erroranalysis/erroranalysis/report/error_report.py
...ponsibleai/responsibleai/_tools/shared/__init__.py
...sponsibleai/responsibleai/rai_insights/__init__.py
...leai/responsibleai/modelanalysis/model_analysis.py
...bleai/responsibleai/_tools/causal/causal_config.py
...leai/responsibleai/databalanceanalysis/__init__.py
.../responsibleai/modelanalysis/constants/__init__.py
... and 54 more
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
| gharchive/pull-request | 2023-01-06T03:58:05 | 2025-04-01T06:44:58.850735 | {
"authors": [
"codecov-commenter",
"vinuthakaranth"
],
"repo": "microsoft/responsible-ai-toolbox",
"url": "https://github.com/microsoft/responsible-ai-toolbox/pull/1888",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1829055453 | Running the same container in AKS instead of WebApp
@sarah-widder
How about creating the same deployment in AKS, should a separate solution be created for that or can this repo be extended to include an AKS deployment as well? This should be not difficult to implement, given there's already a dockerfile ready with all the dependencies.
I'm not completely sure about the authentication layer though, need to look closer.
For now I've disabled the authentication, just to test whether it works
here:
https://github.com/microsoft/sample-app-aoai-chatGPT/blob/main/frontend/src/pages/chat/Chat.tsx
amended lines 35-44 like that:
const getUserInfoList = async () => {
const userInfoList = await getUserInfo();
// if (userInfoList.length === 0 && window.location.hostname !== "127.0.0.1") {
// setShowAuthMessage(true);
// }
// else {
// setShowAuthMessage(false);
// }
setShowAuthMessage(false);
}
Now the authentication prompt doesn't appear, however the chat isn't working still due to name resolution error. It is the same error in both WebApp (which I deploy from the azure portal by clicking "Deploy to WebApp") and when I'm running it in kubernetes.
The error:
Error communicating with OpenAI: HTTPSConnectionPool(host='my-cognitive-account-chatgpt4.openai.azure.com', port=443): Max retries exceeded with url: //openai/deployments/gpt-4_0314/chat/completions?api-version=2023-03-15-preview (Caused by NameResolutionError("<urllib3.connection.HTTPSConnection object at 0x7f28c71e7950>: Failed to resolve 'my-cognitive-account-chatgpt4.openai.azure.com' ([Errno -2] Name does not resolve)"))
Can you suggest why am I getting that? My OpenAI account is located in FranceCentral region (because this is where GPT-4 is available), the edpoint is: https://francecentral.api.cognitive.microsoft.com
AZURE_OPENAI_RESOURCE: "my-cognitive-account-chatgpt4" -- the name of OpenAI (cognitive services account of type OpenAI)
AZURE_OPENAI_MODEL: "gpt-4_0314" -- deployment name
AZURE_OPENAI_MODEL_NAME: "gpt-4" -- model name
But how does it know where this resource is deployed, which subscription and which resource group?
aahh the problem is that I was missing a custom domain, thus my endpoint didn't look like customdomain.openai.azure.com
Now just need to be able to setup the same authentication mechanism (azure AD), as was done in WebApp, for my application running in AKS.
@pamelafox can you help with that perhaps?
Ok, i've solved the authentication problem, by following these steps (a great article by the way): https://kristhecodingunicorn.com/post/k8s_nginx_oauth/#configure-nginx-ingress-controller
and yeah, i really enjoy talking to myself here... :)
@ealasgarov Sorry, we're playing whack-a-mole on OpenAI repository issue trackers right now. Here's a write-up of how I enabled AAD for this repo, if it helps:
For sample-app-aoai-chatGPT, I automated the process of creating an app registration and protecting the app service with that app with a combination of hooks and Bicep.
The hooks are declared here:
https://github.com/microsoft/sample-app-aoai-chatGPT/blob/main/azure.yaml
For the pre provision hook, auth_init.sh calls auth_init.py:
https://github.com/microsoft/sample-app-aoai-chatGPT/blob/main/scripts/auth_init.py
That script makes REST API calls to https://graph.microsoft.com/v1.0/applications in order to create a new app registration. It then sets AUTH_APP_ID, AUTH_CLIENT_ID, and AUTH_CLIENT_SECRET.
For the provisioning step, AUTH_CLIENT_ID and AUTH_CLIENT_SECRET are passed in main.parameters.json:
https://github.com/microsoft/sample-app-aoai-chatGPT/blob/main/infra/main.parameters.json
Those parameters get passed into the appservice module here:
https://github.com/microsoft/sample-app-aoai-chatGPT/blob/5b311a9f74797b771dad2b515126f9ec91a3dabe/infra/main.bicep#L96
That appservice.bicep module adds the identity provider here:
https://github.com/microsoft/sample-app-aoai-chatGPT/blob/5b311a9f74797b771dad2b515126f9ec91a3dabe/infra/core/host/appservice.bicep#L103
For the post provision hook, auth_update.sh calls auth_update.py:
https://github.com/microsoft/sample-app-aoai-chatGPT/blob/main/scripts/auth_update.py
That code makes a REST API to update the redirect URIs for the registered application to include the deployed URL endpoint.
This all works great locally! However, it doesn't work on CI/CD as the pipeline principal doesn't have the permission needed to create an application registration.
Thanks for the reply Pamela, much appreciated!
I guess I've now sorted everything out and things seem to work, including azure authentication, except for one problem - each time after the first successful answer, I'm getting this error when asking a 2nd question:
Error
Requests to the Creates a completion for the chat message Operation under Azure OpenAI API version 2023-03-15-preview have exceeded token rate limit of your current OpenAI S0 pricing tier. Please retry after 7 seconds. Please go here: https://aka.ms/oai/quotaincrease if you would like to further increase the default rate limit.
I cannot replicate this on OpenAI studio Playground with the same deployment/model, there I can ask 10 questions one after another and everything works fine. Not sure why is that... But I will the open a separate issue for that one.
imo it would be great to just pass the following environment variables if possible:
AUTH_TENANT_ID
AUTH_CLIENT_ID
AUTH_CLIENT_SECRET
what do you guys think?
@ealasgarov I am having issue still on this, can you share your ingress files and deployment files please if possible
| gharchive/issue | 2023-07-31T12:12:29 | 2025-04-01T06:44:58.871588 | {
"authors": [
"Breee",
"breddy-lgamerica",
"ealasgarov",
"pamelafox"
],
"repo": "microsoft/sample-app-aoai-chatGPT",
"url": "https://github.com/microsoft/sample-app-aoai-chatGPT/issues/121",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1634164401 | init and attempting to upload new package from workflow.
On the step:
- name: CLI init
run: msstore init . -n ${{ secrets.PUBLISHER_DISPLAY_NAME }} --package --publish
The workflow fails with the error message Unhandled exception: System.NotSupportedException: Cannot show selection prompt since the current terminal isn't interactive.
My goal is to build and package the application and upload a submission to the dev center, and it appeared that init is the closest command to do so, though I am not sure if it selects any particular app/submission to upload the built package to.
Hello @yaakovschectman.
Just to make sure I have full context, for what type of application is this for? (Flutter, React Native, Electron, WinUI, etc).
The init command is probably not the one you need here. You can call the init only once in your development computer. The CLI will connect with Partner Center, list all the applications in your account, and let you interactively select which application you want to configure your project. Then, it will detect the project type and properly configure your project. That command is not supposed to be called from CI/CD environments. What you want is probably a package command with the --publish option.
Does that make sense?
3 steps:
Init (once, in the dev box).
Package (in CI/CD), it will build and generate the MSIX.
Publish (in CI/CD), it will upload the MSIX to the store, to the right application that was selected in the init command.
I think so. I have run msstore init locally within the repository and pushed the changes it caused to the remote, then run package and publish in the workflow. Is there something that must be done in order to create/select a submission? msstore publish produces the below error:
Creating new Submission
💥 Error while creating submission. Please try again.
Could not retrieve submission. Please try again.
Is this the very first submission of the application, or is this an update?
If it is an update, what is the ApplicationCategory of this app?
Important to notice that if this is the first submission of an application, it needs to happen from an interactive console, as there are many missing information for the application, such as no category selected, for example.
It is an update, and it is listed as "Type: Game"
Hello, just want to check with you if you've any thoughts on the error message. I'm really not sure how to debug this one. Thanks.
It took me a long time to find this issue, but I believe we've fixed it.
Will ship a new version shortly.
Ok, new release is out on GitHub Actions.
Just run your build again, as it should pick up the latest version now.
Thank you for the update. I am currently running into the below error upon the publish command:
│ InvalidParameterValue │ You don’t have packages that support all of the │
│ │ device families you’ve selected in Device family │
│ │ availability. Upload a package that supports each │
│ │ device family you’ve selected, or uncheck the box │
│ │ for any device families that this submission doesn’t │
│ │ support. Device families: Holographic |
If this is creating a new submission, where is it getting the selected available families from? On the MS dev center, it looks like one can only select device families on a submission that's already started. Is there a CLI argument to specify the device families?
This is a separate issue that we are aware of, and we are already fixing it.
The CLI right now uses the same families for every type of project, but this should not be the case. It should use what previous submissions are using and have default per project type (UWP vs Win32 implies some of these, so different defaults).
If the submission is failed, you should be able to just change it and re-submit, without the CLI.
@yaakovschectman v0.1.23 should have fixed this issue. Let me know if you are still having problems.
| gharchive/issue | 2023-03-21T15:27:59 | 2025-04-01T06:44:58.881350 | {
"authors": [
"azchohfi",
"yaakovschectman"
],
"repo": "microsoft/setup-msstore-cli",
"url": "https://github.com/microsoft/setup-msstore-cli/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1067679350 | ANSI Escape Sequence Handling Issue
Windows Terminal version
1.12.211020001-release1.12
Windows build number
10.0.19041.0
Other Software
julia version 1.6 and above
Steps to reproduce
Launch julia inside of Windows Terminal
Expected Behavior
Launching julia outside of Windows Terminal with Legacy Console Mode enabled:
Windows Terminal should render the ANSI escapes as above.
Actual Behavior
When julia is launched in Windows Terminal, ANSI escapes are printed directly to the screen.
In Windows Terminal, launched from pwsh:
In Windows Terminal, launched from pwsh (julia --color=no):
In Windows Terminal, launched as a new tab:
In Windows Terminal, launched as a new tab (julia --color=no):
Launched directly, Legacy Console Mode disabled:
(glad you made it from stackoverflow ☺️)
HUH. I wonder if they're setting the VT reg key.
Can you export the contents of HKCU\Console to a .reg file, rename it to .txt, and share it here?
I'm guessing that there's a VirtualTerminalLevel that's set to something other than 0, and they're using that to manually enable VT support, rather than using SetConsoleMode
Here's the output with a few non-relevant applications removed from the list.
console.txt
You know what, I'm smarter than that. When julia is working correctly, are you launching it from a shortcut? Something like julia.lnk, either on the desktop, or the Taskbar, or the Start Menu?
There are two places that a console app can have VirtualTerminalLevel set - one is in the registry, the other is in the lnk. Though, I didn't.... (thought clipped in favor of another theory)
Presumably, using the run dialog (win+r) to launch v:\PortableApps\julia\bin\julia.exe would also not work.
https://github.com/microsoft/terminal/blob/284257a38392c85a2f7d7ac7e77e20a05242ec64/src/host/srvinit.cpp#L194-L201
We should be enabling virtual terminal processing for Terminal client apps, always. That implies that julia is actually _manually disabling it?!
https://github.com/JuliaLang/julia/blob/f9bb6f8fa8c66ca00f215cbbc4e4bc5addc43156/stdlib/REPL/src/LineEdit.jl#L1375
That's the only SetConsoleMode reference in julia. Maybe there's something else I'm missing.
x-ref: https://github.com/JuliaLang/julia/issues/43273
I don't launch it through a link. In Windows Terminal, the settings are done this way:
Otherwise, I use start v:\PortableApps\julia\bin\julia.exe to launch it.
After the problem magically resolving itself when i moved my ~/.julia directory, I did some more digging. It looks like the issue is tied to a macro that I created to clear the screen which was called in a different macro in my startup script.
replacing the macro with its content removes the ANSI character issue.
Well, that's good to know. Thanks for following up!
| gharchive/issue | 2021-11-30T21:11:14 | 2025-04-01T06:44:58.893859 | {
"authors": [
"andrewraddatz",
"zadjii-msft"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/11848",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1068029037 | Why not add session management and file transfer functions?
When I connect ssh server in Windows, the best tools to do it is Mobaxterm, but It's too complex and ugly. I really like Microsoft Terminal, Why not add session management and file transfer features to it. It will be more popular.
File transfer is probably best left for something like scp, a tool that's actually dedicated to it.
Could you elaborate more on what you mean by "session management"?
I suspect by "session management" they mean something along the lines of #1280. In apps like MobaXterm and PuTTY you can setup a session (like Windows Terminal profile) with your SSH connection parameters preconfigured, and the same concept can be used for file protocols like FTP and SFTP. I think it was decided that this would best be implemented as a plugin though.
| gharchive/issue | 2021-12-01T06:51:33 | 2025-04-01T06:44:58.896182 | {
"authors": [
"b1tkeeper",
"j4james",
"zadjii-msft"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/11855",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1358180737 | Display problems, color Settings and some interactive buttons
Windows Terminal version
1.15.2282.0
Windows build number
10.0.19044.1889
Other Software
No response
Steps to reproduce
That's how it started. There's something wrong with the interface
Expected Behavior
No response
Actual Behavior
When I configure the new theme color, the interface looks like this, and the system interaction button is not visible
When I hit Close, the confirm button does the same thing
Thanks for the suggestion! This is actually already being tracked by another issue on our repo - please refer to #13382 for more discussion.
/dup #13382
| gharchive/issue | 2022-09-01T03:06:04 | 2025-04-01T06:44:58.899657 | {
"authors": [
"nianjiuhuiyi",
"zadjii-msft"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/13895",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1663193468 | Add support for LNM (Line Feed/New Line Mode)
Description of the new feature/enhancement
When LNM is set, the linefeed controls (LF, FF, and VT) will also trigger a carriage return. When reset they, they won't. This is essentially the inverse of our existing console mode, DISABLE_NEWLINE_AUTO_RETURN.
However, LNM has an additional effect on input. When set, the Return key generates both a carriage return and a line feed. When reset it only generates the carriage return.
It's quite widely implemented, but I don't think VTE supports it, which suggests it's probably not that widely used. However, it is a requirement for meeting VT level 1 conformance, which is why I would like for us to support it.
Proposed technical implementation details (optional)
I had originally thought we could map it directly to the DISABLE_NEWLINE_AUTO_RETURN mode, since we're already using that to determine how linefeed controls are interpreted. However, when a Windows console app has DISABLE_NEWLINE_AUTO_RETURN reset (which is the equivalent of LNM being set), we don't typically want the Return key to behave differently.
So my idea was this: We add a new input mode that specifically handles the LNM behavior for the Return key, which by default is disabled. And we treat the DISABLE_NEWLINE_AUTO_RETURN mode as an inverse alias for the LNM output behavior (as we already do).
When those two match, we're in a valid VT state, and we can use the LNM mode to toggle them both at the same time. But when they're out of sync (which is the default state for a Windows console app), we just act as if the LNM mode is not supported, i.e. we don't respond to any attempts to change it, and DECRQM reports the mode as unknown.
Does that seem like a reasonable approach to take?
The current default behavior for Windows console apps (i.e. when DISABLE_NEWLINE_AUTO_RETURN is reset) is to execute both a carriage and a line feed when LF is output, and to generate just CR when the Return key is pressed. This behavior does not match either of the LNM states.
When LNM is set, a Return key will generate both CR and LF, which a lot of console apps will register as a double key press. And when LNM is reset, you won't automatically get carriage returns when outputting *nix-style text content with \n line endings (I'm guessing that was the main reason the console added the auto-return functionality in the first place).
| gharchive/issue | 2023-04-11T20:48:06 | 2025-04-01T06:44:58.905550 | {
"authors": [
"j4james"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/15167",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1773546809 | Saving settings moves/looses focus to Startup item
Windows Terminal version
1.17.11461.0
Windows build number
10.0.19045.2965
Other Software
No response
Steps to reproduce
Click Save button in Settings dialog (for any profile, nested item)
Expected Behavior
Focus won't be lost/moved; current settings pane/context won't be changed/lost.
Actual Behavior
Focus switches to Startup item, quiting some (distant) profile settings pane, unexpectedly changing context.
This is a known bug, but I couldn't find any existing issue that mentioned this. It happens because clicking save reloads the entire settings page.
| gharchive/issue | 2023-06-25T22:59:43 | 2025-04-01T06:44:58.908355 | {
"authors": [
"lhecker",
"lv-gh"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/15600",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
479398133 | Unable to overwrite custom tabTitle from within shell
Environment
Windows build number: 10.0.18956.1000
Windows Terminal version (if applicable): 0.3.2171.0
Any other software? No other software relevant I believe
Steps to reproduce
Configure a custom tabTitle for cmd or PowerShell shells in the profiles.json file, e.g.:
[...]
"tabTitle" : "Windows PowerShell",
[...]
Open a cmd or PowerShell tab, verify the customized tabTitle is appearing.
Try to set a custom title programmatically from within the shell, e.g. in PowerShell:
$Host.UI.RawUI.WindowTitle = 'Custom Title :)'
# OR
[Console]::Title = 'Custom Title 2 :)'
or in command prompt:
TITLE Custom Title :)
Expected behavior
We should be able to programmatically overwrite the customized tabTitle set in the profiles.json file - programmatically setting the tab title DOES work when no custom tabTitle is specified in profiles.json.
I believe the order of precedence should be, from lowest to highest:
Default title / executable path
Gets overwritten by tabTitle property in profiles.json
Gets overwritten by scripts or programs setting their own titles
Actual behavior
When a customized tabTitle is specified, it cannot be overwriten or changed from within the shell.
tabTitle is an override. If you don’t want the title to be overridden, don’t use it. If you’re already planning on using the shell to set the title, you do not need tabTitle. :smile:
There is an open pull request that offers another better option for titling.
I disagree.
I would say tabTitle is a customization preference that's used to have something nicer than a filesystem path by default.
However, the applications running within the shell have better context awareness than the hosting Terminal application - and command-line applications commonly set their own "window" titles,
so this should take precedence. I must re-state that I feel this is a bug.
Cosnider this:
When I have not customized my tabTitle and I start vim from a command prompt, my tab title is changed to [Unnamed] - VIM, as the application author intended. This information is also VERY helpful to me the user, who can not tell my VIM tab apart from my 4 other command prompt tabs.
With a custom tabTitle set, this feature is completely taken away.
You say tabTitle is an override, but it shouldn't be. It should be a default value, modifyable at any point by any application.
Please consider re-opening the issue, or introducing a defaultTabTitle property that is not a hard override.
You should follow #2373. There are a few issues littering this repository explaining why things are the way they are, and why 2373 is the correct fix. That should ameliorate most of your concerns here.
Thanks, I had since found that issue after you illuded to an open PR regarding this.
I will follow it, and it is completely fine to close this as "duplicate / close enough duplicate",
it was just your original statement that tabTitle should be a set-in-stone, static, non-context-aware override that I took issue with. And as I saw now, I'm not alone.
cheers
| gharchive/issue | 2019-08-11T18:40:32 | 2025-04-01T06:44:58.917173 | {
"authors": [
"DHowett-MSFT",
"jantari"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/2393",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
355484211 | Font that was once rendering correctly, now shows empty boxes
I am on Windows 10 Pro 1803 (17134.228) on a Dell XPS 15 9570 with the 4K screen (not sure if this matters).
I have been using "Ubuntu mono derivative Powerline" without issues on WSL. Yesterday I noticed that my zsh prompt no longer renders the font correctly; it now shows empty boxes instead of >>> (three arrows).
Here is what it used to look like:
I'm really not sure what triggered this; here is my Windows Update log:
Hey @burhan - in the erroneous case, what font is your Console using?
The same font, I haven't changed the console settings.
On Thu, Aug 30, 2018, 7:07 PM Rich Turner notifications@github.com wrote:
Hey @burhan https://github.com/burhan - in the erroneous case, what
font is your Console using?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/Microsoft/console/issues/243#issuecomment-417374775,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AAkz6NeU-xqShgtZjZOl1GmeVvQlVKwNks5uWA3BgaJpZM4WTCRS
.
--
--
Burhan Khalid
Sent from a mobile device
@Burhan: And do you have a link to the font you're using?
The font is installed as per the properties window.
The font installed is the same one as in your link @bitcrazed
I am not sure if this is related, but in VS code, the font renders correctly
Could you share your zsh configuration declaring the color chevrons?
${SSH_TTY:+"%F{9}%n%f%F{7}@%f%F{3}%m%f "}%F{4}${_prompt_sorin_pwd}%(!. %B%F{1}#%f%b.)${editor_info[keymap]}
And what does your .~/zshrc look like?
@burhan, The fonts in provided VS Code and cmd.exe screenshots look different to the naked eye. While the one in cmd.exe definitely looks like Ubuntu Mono, the VS Code one most certainly doesn't.
I wonder if this is the same problem as #3257? There's a lot more discussion there and a potential workaround, selecting "Install for all users" on the font may get it working properly.
I wonder if this is the same problem as #3257? There's a lot more discussion there and a potential workaround, selecting "Install for all users" on the font may get it working properly.
Hey so this was filed like, four years ago, and obviously there's been a lot of work done since then. Is this still happening? I think we added some better support for unsupported glyphs in the console in the latest Win11 bits. Then there's also the Terminal, which may have never had this issue at all. We never did nail down a specific repro for this one, so maybe it's gone/?
| gharchive/issue | 2018-08-30T08:45:01 | 2025-04-01T06:44:58.927703 | {
"authors": [
"bitcrazed",
"burhan",
"mikedld",
"qidydl",
"zadjii-msft"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/243",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
586118816 | Weird behavior after changing cursor shape in vim
Environment
Windows build number: 10.0.18363.0
Windows Terminal version (if applicable): 0.10.761.0
Steps to reproduce
Launch vim from either WSL or SSH of a Linux host
Set cursor shape options: blinking block in normal mode, blinking vertical bar in insert mode
let &t_SI .= "\<Esc>[5 q"
let &t_EI .= "\<Esc>[0 q"
The shape does change as what we expect, but I observed weird behaviors after that.
Expected behavior
When exiting insert mode, i.e. press i then <ESC>, the cursor stays at where it was if I didn't move it.
When exiting insert mode, the cursor shape should change as soon as <ESC> pressed.
Restore the default shape after exiting vim
Actual behavior
Press i then <ESC>, the cursor moves ahead one character.
The shape didn't change until the next blink.
The cursor stays like a block after exiting.
You can find an explanation of why the cursor moves backwards on the VIM stack exchange site. This is expected behaviour, and I see the same behaviour in other terminals emulators too.
The delayed shape changing is because an ESC character can also be the start of a character sequence representing another key, so VIM needs to wait and see if there are more characters following before it can be certain you've actually pressed Esc. Again, this is expected behaviour.
DECSCUSR 0 is officially defined as the block cursor so that's why you get a block on exit. Some terminals do interpret that as the user's preferred cursor, though, and it's possible we may support that one day too (see issue #1604)
Thanks for the comprehensive writeup, James. :smile:
| gharchive/issue | 2020-03-23T10:57:42 | 2025-04-01T06:44:58.934710 | {
"authors": [
"DHowett-MSFT",
"dianlujitao",
"j4james"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/5084",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
707882172 | startup commands from command line for wt
Ability to set startup commands from command line startup.
What I want is to be able to set startup command of the terminal tab/pane from a command line option for wt.
Something like "--startup" option. There is an example of the usage down below.
Proposed technical implementation details
Example is for settings an environmental variable for git bash. Each pane in each tab gets its own prefix group name preceeded by the path. Open wt and 2 split panes with git bash and just execute TITLEPREFIX=G1 in each pane and see for yourself. :)
But basically it is for executing a command initially for each tab/pane.
example usage: (imagine as single line, switched to multiline for readability)
wt -d {PATH1} --startup "TITLEPREFIX={Group1}" ;
split-pane -V -d {PATH2} --startup "TITLEPREFIX={Group1}" ;
split-pane -V -d {PATH3} --startup "TITLEPREFIX={Group1}";
new-tab -d {PATH4} --startup "TITLEPREFIX={Group2}";
split-pane -V -d {PATH5} --startup "TITLEPREFIX={Group2}" ;
new-tab --title web -d {PATH6} --startup "TITLEPREFIX={Group3}";
split-pane -V -d {PATH7} --startup "TITLEPREFIX={Group3}";
new-tab --title mobile -d {PATH8} --startup "TITLEPREFIX={Group4}";
focus-tab -t 0
Maybe #6776 ?
yes, seems like #5528
| gharchive/issue | 2020-09-24T06:04:54 | 2025-04-01T06:44:58.939474 | {
"authors": [
"alper-batioglu",
"skyline75489"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/7720",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
818009122 | Opening terminal and opening a second window then closing the first window with ctrl+shift+w freezes terminal
Environment
Windows build number: Microsoft Windows [Version 10.0.19042.844]
Windows Terminal version: 1.5.10411.0
Steps to reproduce
Open up a new fresh terminal instance
Open up a new tab (you now have a total of two tabs open)
Switch to the previous tab (the one you started out with when you opened up the terminal)
Close the tab by using the default keybinding to close a pane which is ctrl+shift+w (see this for reference https://docs.microsoft.com/en-us/windows/terminal/customize-settings/actions#close-pane)
Expected behavior
The previous tab (the first tab) closes and focus goes to the recently second opened tab.
Actual behavior
The whole terminal freezes and I have to either close the terminal by using alt+F4 or terminate the process in task manager.
Notes
This issue seems closely related to this one https://github.com/microsoft/terminal/issues/5799 but it still doesn't work for me.
Well. That's certainly weird. You're not using anything like focus mode, you running the Terminal full-screen? Could you share your settings? Are you using any sort of accessibility software (Narrator, NVDA)?
I'm having a hard time reproing this, but I also only have 1.6+ builds installed 😕
Ah, you right! This issue only applies to when the terminal is in focus mode (not full-screen nor default mode).
I've updated the issue with the new information.
Here is my settings:
{
"$schema": "https://aka.ms/terminal-profiles-schema",
"defaultProfile": "{2c4de342-38b7-51cf-b940-2309a097f518}",
"copyOnSelect": false,
"copyFormatting": false,
"launchMode": "focus",
"profiles":
{
"defaults":
{
"useAcrylic": true,
"acrylicOpacity": 0.8,
"fontSize": 9,
"cursorShape": "filledBox"
},
"list":
[
{
"guid": "{2c4de342-38b7-51cf-b940-2309a097f518}",
"hidden": false,
"name": "Ubuntu",
"source": "Windows.Terminal.Wsl",
"startingDirectory" : "//wsl$/Ubuntu/home/anders"
},
{
"guid": "{0caa0dad-35be-5f56-a8ff-afceeeaa6101}",
"name": "Command Prompt",
"commandline": "cmd.exe",
"hidden": false
},
{
"guid": "{61c54bbd-c2c6-5271-96e7-009a87ff44bf}",
"name": "Windows PowerShell",
"commandline": "powershell.exe",
"hidden": false
},
{
"guid": "{b453ae62-4e3d-5e58-b989-0a998ec441b8}",
"hidden": false,
"name": "Azure Cloud Shell",
"source": "Windows.Terminal.Azure"
}
]
},
"schemes": [],
"actions":
[
{ "command": "find", "keys": "ctrl+shift+f" },
{ "command": { "action": "splitPane", "split": "auto", "splitMode": "duplicate" }, "keys": "alt+shift+d" },
{ "command": "toggleFocusMode", "keys": "shift+f11" }
]
}
@zadjii-msft - wasn't it solved in 1.6 by https://github.com/microsoft/terminal/pull/8549?
I've just tested it in 1.6 and it works! Should've updated!
| gharchive/issue | 2021-02-27T20:13:21 | 2025-04-01T06:44:58.946618 | {
"authors": [
"AndysonDK",
"Don-Vito",
"zadjii-msft"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/9306",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
867077001 | The WT can't create a new tab based on the parent root when i onpen from the right click menu
Description of the new feature/enhancement
The WT can't create a new tab based on the parent root when i onpen from the right click menu.
After update the new version, a content is created on my right click menu,which named "Open in windows terminal ".When I click it from "D:/myfile:" ,although the first page is base on my folder root which is "D:/myfile", the root of new tab is "c:/windows/system32" again. It is unconvenient!! WISH TO BE SOLVED!!
Proposed technical implementation details (optional)
Thanks for the suggestion! This is actually already being tracked by another issue on our repo - please refer to #8933 for more discussion.
/dup #8933
| gharchive/issue | 2021-04-25T18:15:14 | 2025-04-01T06:44:58.949316 | {
"authors": [
"Michaelzhouisnotwhite",
"zadjii-msft"
],
"repo": "microsoft/terminal",
"url": "https://github.com/microsoft/terminal/issues/9947",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
391455941 | TFS specific TestContext variables are not present in Azure DevOps when executing TestCase from TestPlan
Description
I used to use variables populated by TestAdapter in TextContext (during TestInitialize, i believe) when executing tests from Test Plan. These are:
Tfs_TestPlanId
Tfs_TestCaseId
Tfs_TestPointId
Tfs_TestConfigurationId
Tfs_TestConfigurationName
Tfs_TfsServerCollectionUrl
Tfs_TeamProject
Recently when I came back to my solution I noticed that these variables are missing and not populated when executing on Azure DevOps. I am sure that when Azure DevOps was called VSTS the variables were still populated.
Executed on:
Azure DevOps
by:
VSTest v2
Written locally using:
MSTest v2
Steps to reproduce
Define any TestCase using MsTest v2
Created TestPlan in Azure DevOps
Connect Defined MsTest TestCase to TestCase in TestPlan (RMB + Associate To Test Case)
Add some logging code to the testCase and set it to be executed after Test Case initailization to check TestContext
Run TestCase from TestPlan
Verify if TestContext data has been populated with:
Tfs_TestPlanId
Tfs_TestCaseId
Tfs_TestPointId
Tfs_TestConfigurationId
Tfs_TestConfigurationName
Tfs_TfsServerCollectionUrl
Tfs_TeamProject
Expected behavior
Variables:
Tfs_TestPlanId
Tfs_TestCaseId
Tfs_TestPointId
Tfs_TestConfigurationId
Tfs_TestConfigurationName
Tfs_TfsServerCollectionUrl
Tfs_TeamProject
Should be present in TestContext after testCase initialization
Actual behavior
Tfs_TestPlanId
Tfs_TestCaseId
Tfs_TestPointId
Tfs_TestConfigurationId
Tfs_TestConfigurationName
Tfs_TfsServerCollectionUrl
Tfs_TeamProject
Are missing in TestContext
Environment
Executed on:
Azure DevOps (Azure DevOps TestPlan)
by:
VSTest v2
Written locally using:
MSTest v2
@borsooq : Which version of MSTest nuget packages are you using? This functionality was added in 1.4.0 version of MSTest.TestAdapter and MSTest.TestFramework. Please try with latest packages and see it that works for you!
Hello, it is MSTest V2 1.3.2. Seems like we have the reason :)
Will check that out.
I can see that property TCMTestPropertiesJSONFile has been added to TestContext.Properties. The property contains JSON string that can be deserialized. I used that and it does the job :)
@borsooq I am facing the same issue. Could you please explain which property did you use to get the TCM Test data. I don't see TCMTestPropertiesJSONFile in the TextContext.Properties.
@jayaranigarg Could you please suggest how we can use the TCMTestPropertiesJSONFile to get TCM data?
| gharchive/issue | 2018-12-16T10:54:20 | 2025-04-01T06:44:58.991163 | {
"authors": [
"borsooq",
"jayaranigarg",
"tarenbd"
],
"repo": "microsoft/testfx",
"url": "https://github.com/microsoft/testfx/issues/541",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2760546251 | MSTEST0037 (proper assert analyzer): Don't report for user defined equality operators
Fixes #4414
I think it should. It should be the compiler responsibility to give us the correct OperatorMethod with the correct MethodKind.
| gharchive/pull-request | 2024-12-27T07:56:42 | 2025-04-01T06:44:58.992355 | {
"authors": [
"Youssef1313"
],
"repo": "microsoft/testfx",
"url": "https://github.com/microsoft/testfx/pull/4456",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
637547176 | About how to submit SROIE competition
Sorry for disturbing the problem not related to this repo.
About how to submit SROIE competition.
I have finished my model for SROIE Task3, however, I do not know how to submit my results.
here is the snapshot about my results. My results pass the auto examination and reports 'The method is not calculated yet'. So, if there some thing wrong with my reuslts?
Can you help me. VERY THANKS.
@persistforever Sometimes the website will return this status. Please wait or resubmit the results
@persistforever hey can you please help me how you preprocess the dataset and how you map the key-value.
@persistforever, I also give same issue. Did you solve this problem ?
@persistforever, I also give same issue. Did you solve this problem ?
| gharchive/issue | 2020-06-12T07:30:08 | 2025-04-01T06:44:58.995310 | {
"authors": [
"buiquangmanhhp1999",
"kbrajwani",
"persistforever",
"wolfshow"
],
"repo": "microsoft/unilm",
"url": "https://github.com/microsoft/unilm/issues/178",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2356331995 | Raise errors for malformed overlay port manifests
For https://github.com/microsoft/vcpkg/issues/39276.
This PR will make vcpkg no longer tolerate broken overlay port manifests. Initially meant for vcpkg ci, the change also affects commands like install and search. This is acceptable: As long as broken ports are ignored, installation plans and results of "search" must be considered invalid.
Errors are raised for:
malformed JSON (broken-manifests/malformed)
missing name (broken-manifests/broken-no-name)
missing version (broken-manifests/broken-no-version)
Note that errors were already raised before this change when --overlay-ports pointed directly to a single broken port.
No errors are raised for the remaining cases in ci/* (unchanged), including the test for missing dependencies of inactive features (new: ci/feature-missing-dep).
This PR will make vcpkg no longer tolerate broken overlay port manifests.
What exactly is the behavior here? It shouldn't break on vcpkg install overlayA if overlayB is not used but broken.
This PR will make vcpkg no longer tolerate broken overlay port manifests.
What exactly is the behavior here? It shouldn't break on vcpkg install overlayA if overlayB is not used but broken.
vcpkg ci must fail if overlayB is broken. overlayB is vcpkg-ci-openimageio.
The effect on vcpkg install overlayA is debatable.
The effect on vcpkg install overlayA is debatable.
In particular, you cannot determine correct dependencies of overlayA if overlayB would overlay one of the dependencies.
vcpkg ci must fail
I fully agree with that.
In particular, you cannot determine correct dependencies of overlayA if overlayB would overlay one of the dependencies.
that would mean overlayB is used. I am only interested in the case that overlayB is not used by anything and just exists. Overlays are for me working directories and erroring on incomplete and yet unused work is just annoying. A warning would be ok. Just don't straight out error for stuff which isn't required.
Overlays are for me working directories and erroring on incomplete and yet unused work is just annoying. A warning would be ok. Just don't straight out error for stuff which isn't required.
I understand the WIP nature of overlay port directories. However, silently ignoring (or just issueing a warning for) a potential overlay port just because @Neumann-A forgot to add a required comma isn't acceptable even for vcpkg install foo IMO. vcpkg must know the available ports to determine the installation plan.
But the only hard requirement is to be able to parse the manifest for the port name and version. In a more strict sense, only the port name is needed, and it is defined by the directory name [if the directory has a manifest].
Hm, I don't think the problems exposed by the vcpkg-ci-openimageio fiasko are sufficiently understood.
The ci command effectively names everything reachable, so the intended effect of exploding in the registry repo CI runs would still be achieved with this framing.
vcpkg-ci-openimageio isn't used as a dependency, so its breakage went unnoticed.
Unnoticed even in presence of =pass in the baseline. Do we need to see this as a depedency? It is not unreasonable, but it wouldn't make this much easier.
vcpkg install foo --overlay-ports=broken\bar shouldn't explode because bar is malformed, unless it's named as a dependency of foo.
The point is: ATM it will never explode "because bar is malformed". It will only explode when bar is missing.
Imagine it wasn't vcpkg-ci-openimageio but an openssl hotfix. vcpkg would silently use the main repo's vulnerable openssl because the author forgot one comma in the manifest. Is this acceptable?
vcpkg would silently use the main repo's vulnerable openssl because the author forgot one comma in the manifest.
AS far as I know it is not silent since vcpkg prints the location where something is taken from. If it is from an external registry you would have to run x-add-version which would check the format. So the problem probably only really applies to overlays.
The problem is that changing the behavior to always break independent of actual requirements/dependencies breaks potentially existing users unnecessary. So the fix needs to only break if it is asked for, meaning you have a broken openssl in the overlay and you a requesting openssl to be build, then it should stop and yell at the user that openssl is broken.
Implementing this probably requires keeping a list of broken ports around and compare that against the requested stuff. The interesting cases then will be how to handle multiple overlays with different broken stuff, e.g.:
--overlay-ports=broken\foo --overlay-ports=broken\bar --overlay-ports=broken\abc
AS far as I know it is not silent since vcpkg prints the location where something is taken from.
IMO any overlay port modification which introduces a syntax error must lead to an immediate full error. Anything else, such as using a regular registry port, counts as "silent".
Yes, this change can cause interruption. That's why I tried to explicitly mention the effects in the top post.
But the only "bad" consequence is that users must fix (or remove) broken manifests.
It will not cause picking overlay ports which weren't pick before the change.
It might even turn some users' attentition to the fact that desired overlay ports where not used.
I still don't see the real benefit of broken manifest being silent ignored which justifies keeping that behavior. That's like keeping bugs unfixed because the it was always broken.
Hm, I don't think the problems exposed by the vcpkg-ci-openimageio fiasko are sufficiently understood.
The ci command effectively names everything reachable, so the intended effect of exploding in the registry repo CI runs would still be achieved with this framing.
vcpkg-ci-openimageio isn't used as a dependency, so its breakage went unnoticed.
To ci, everything is referenced. Or, at least, should be.
Unnoticed even in presence of =pass in the baseline. Do we need to see this as a depedency? It is not unreasonable, but it wouldn't make the fix much easier. In particular if vcpkg eventually wants to move to "no baseline entry means =pass".
IMO this is a different bug but should also be fixed. If there's =pass but, for instance, the --overlay-ports part was missing in the command line, that should still fail, and this change wouldn't fix that.
vcpkg install foo --overlay-ports=broken\bar shouldn't explode because bar is malformed, unless it's named as a dependency of foo.
The point is: ATM it will never explode "because bar is malformed". It will explode only when bar is missing.
Do I misunderstand what this PR does then? It looks like this PR will make that explode even if the contents of that overlay-ports directory never participate in the plan whatsoever.
Imagine it wasn't vcpkg-ci-openimageio but an openssl hotfix. vcpkg would silently use the main repo's vulnerable openssl because the author forgot one comma in the manifest. Is this acceptable?
In that case, the name openssl participated in the plan, so a malformed openssl in an overlay-ports directory should explode.
My point is that names which never participate in the plan at all should not result in failures. For ci, all names in all overlay-ports directories participate in the plan, so they should indeed all result in failures.
In the interest of making sure that this horse is dead, what I mean is that, given the following:
broken-overlays/openssl
overlays/openssl
vcpkg install --overlay-ports=overlays curl[openssl] # should succeed
# should fail because a broken openssl overlay-port exists; I believe your PR fixes this
vcpkg install --overlay-ports=overlays --overlay-ports=broken-overlays curl[openssl]
vcpkg install --overlay-ports=broken-overlays curl[openssl]
# should succeed, there's no reason to ever have loaded the openssl overlay-port
# I believe your PR breaks this
vcpkg install --overlay-ports=broken-overlays vcpkg-cmake
# should fail because a broken openssl overlay-port was referenced
vcpkg ci --overlay-ports=broken-overlays
I still don't see the real benefit of broken manifest being silent ignored which justifies keeping that behavior. That's like keeping bugs unfixed because the it was always broken.
I think it's reasonable to have directories in a directory passed to overlay-ports which aren't intended to be ports in the first place
I think it's reasonable to have directories in a directory passed to overlay-ports which aren't intended to be ports in the first place
That's still possible. They are just not allowed to have malformed vcpkg.json or CONTROL.
My point is that names which never participate in the plan at all should not result in failures.
My point is that vcpkg doesn't know if it participates in the plan at the time it loads the manifests.
My point is that vcpkg doesn't know if it participates in the plan at the time it loads the manifests.
I believe that it does know that; the dependency planner doesn't read and parse every vcpkg.json or CONTROL in the repo before doing an install. Maybe this change is fine and that tests needed to change suggests some other bug. I'm taking a look...
Indeed I tested this and load_overlay_ports never gets called while running install; this change might actually already do the thing Neumann-A and I are asking for.
I tested and indeed this already works. I added an explicit test that this didn't break.
| gharchive/pull-request | 2024-06-17T04:25:50 | 2025-04-01T06:44:59.019096 | {
"authors": [
"BillyONeal",
"Neumann-A",
"dg0yt"
],
"repo": "microsoft/vcpkg-tool",
"url": "https://github.com/microsoft/vcpkg-tool/pull/1435",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
840506828 | Error Installing CMake in Windows64
Describe the bug
https://github.com/YDLIDAR/YDLidar-SDK/blob/master/doc/howto/how_to_build_and_install.md
We are following the instructions above and kept getting the error:
error while loading cmake, the port directory does not exist, error: failed to load port from C:\src\win64\vcpkg\portd\cmake
note: uploading vcpkg by rerunning bootstrap-vcpkg may resolve this failure
We made a directory called cmake inside of ports and now we are getting this error:
Environment
OS: Window
Compiler: revision
To Reproduce
Steps to reproduce the behavior:
./vcpkg install cmake
See error
Repro code when
Expected behavior
successfully install and move to next step (integrate install, cc build && \ cmake .."-DCMAKE_TOOLCHAIN_FILE.....)
Failure logs
Additional context
https://github.com/YDLIDAR/YDLidar-SDK/blob/master/doc/howto/how_to_build_and_install.md
We are trying to follow these instructions to use python to interface with the X2 YDLiDar for a project. We are trying to extract the raw data from the sensor
Hi @madeline-m
Thanks for posting this issue.
Since cmake is listed in scripts/test_port/cmake, which is different from these ports in ports/ directory.
So It cannot be installed like this.
You can copy cmake folder in https://github.com/microsoft/vcpkg/tree/master/scripts/test_ports to ports folder, then try to rebuild.
Just to make it clear, the port cmake provided by vcpkg in test_ports/ folder is used to test these ports in vcpkg.
I'm not clear about this document https://github.com/YDLIDAR/YDLidar-SDK/blob/master/doc/howto/how_to_build_and_install.md. But I think this is not the proper way to install cmake.
Closing this issue since it should not a problem in vcpkg.
| gharchive/issue | 2021-03-25T02:50:50 | 2025-04-01T06:44:59.027907 | {
"authors": [
"NancyLi1013",
"madeline-m"
],
"repo": "microsoft/vcpkg",
"url": "https://github.com/microsoft/vcpkg/issues/16869",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1062108752 | make libzip for mingw have error
make libzip for win is OK,bug make libzip for mingw is Error!
D:\vcpkg-master>vcpkg install libzip[bzip2]:x64-mingw-static
Computing installation plan...
The following packages will be built and installed:
libzip[bzip2,core,default-aes,wincrypto]:x64-mingw-static -> 1.8.0
Detecting compiler hash for triplet x64-mingw-static...
Restored 0 packages from C:\Users\Administrator\AppData\Local\vcpkg\archives in 246 us. Use --debug to see more details.
Starting package 1/1: libzip:x64-mingw-static
Building package libzip[bzip2,core,default-aes,wincrypto]:x64-mingw-static...
-- Using community triplet x64-mingw-static. This triplet configuration is not guaranteed to succeed.
-- [COMMUNITY] Loading triplet configuration from: D:\vcpkg-master\triplets\community\x64-mingw-static.cmake
-- Using cached nih-at-libzip-v1.8.0.tar.gz.
-- Cleaning sources at D:/vcpkg-master/buildtrees/libzip/src/v1.8.0-57e7193039.clean. Use --editable to skip cleaning for the packages you specify.
-- Extracting source D:/vcpkg-master/downloads/nih-at-libzip-v1.8.0.tar.gz
-- Applying patch fix-dependency.patch
-- Using source at D:/vcpkg-master/buildtrees/libzip/src/v1.8.0-57e7193039.clean
-- Configuring x64-mingw-static
-- Building x64-mingw-static-dbg
-- Building x64-mingw-static-rel
-- Fixing pkgconfig file: D:/vcpkg-master/packages/libzip_x64-mingw-static/lib/pkgconfig/libzip.pc
-- Using cached msys-mingw-w64-i686-pkg-config-0.29.2-2-any.pkg.tar.zst.
-- Using cached msys-mingw-w64-i686-libwinpthread-git-8.0.0.5906.c9a21571-1-any.pkg.tar.zst.
-- Using msys root at D:/vcpkg-master/downloads/tools/msys2/aa5af7b2aa7e90e8
-- Fixing pkgconfig file: D:/vcpkg-master/packages/libzip_x64-mingw-static/debug/lib/pkgconfig/libzip.pc
-- Installing: D:/vcpkg-master/packages/libzip_x64-mingw-static/share/libzip/copyright
-- Performing post-build validation
-- Performing post-build validation done
Stored binary cache: C:\Users\Administrator\AppData\Local\vcpkg\archives\e7\e7367733ed889d42cc52d811dee7f40c1cb877ba7a817972ba201a41ad55e553.zip
Installing package libzip[bzip2,core,default-aes,wincrypto]:x64-mingw-static...
Elapsed time for package libzip:x64-mingw-static: 32.74 s
Total elapsed time: 35.13 s
The package libzip provides CMake targets:
find_package(libzip CONFIG REQUIRED)
target_link_libraries(main PRIVATE libzip::zip)
make libzip for mingw-x64-static file size is 277kb
error code win64/libzip.a(zip_crypto_win.c.obj):zip_crypto_win.c:(.text+0xeb5): undefined reference to `BCryptCloseAlgorithmProvider'
This issue hasn’t been updated in 3 month, if it is still an issue, please reopen this issue.
| gharchive/issue | 2021-11-24T07:56:41 | 2025-04-01T06:44:59.036247 | {
"authors": [
"JackBoosY",
"Moodsky"
],
"repo": "microsoft/vcpkg",
"url": "https://github.com/microsoft/vcpkg/issues/21643",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1168348477 | vcpkg.exe install boost:x64-mingw-static
Host Environment
OS: [e.g. Windows/Linux etc...]
Compiler: revision
To Reproduce
Steps to reproduce the behavior:
./vcpkg install xxxx
Failure logs
-Cut and paste the appropriate build messages from the console output.
-Please attach any additional failure logs mentioned in the console output.
Additional context
Add any other context about the problem here, such as what you have already tried to resolve the issue.
We don't have enough information to solve this issue, unfortunately; if you have any more information to help us solve this, please reopen!
| gharchive/issue | 2022-03-14T13:05:05 | 2025-04-01T06:44:59.039316 | {
"authors": [
"Adela0814",
"codezhy"
],
"repo": "microsoft/vcpkg",
"url": "https://github.com/microsoft/vcpkg/issues/23545",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1416074694 | [icu] build failure: Detected whitespace in root directory.
Computing installation plan...
The following packages will be built and installed:
* icu[core]:x64-windows -> 71.1
* jasper[core,default-features,opengl]:x64-windows -> 2.0.33#6
* jsoncpp[core]:x64-windows -> 1.9.5
* libharu[core]:x64-windows -> 2.4.2
* libiconv[core]:x64-windows -> 1.17
* libjpeg-turbo[core]:x64-windows -> 2.1.4
* libogg[core]:x64-windows -> 1.3.5
* libtheora[core]:x64-windows -> 1.2.0alpha1-20170719#4
* libwebp[core,libwebpmux,nearlossless,simd,unicode]:x64-windows -> 1.2.4
* libxml2[core]:x64-windows -> 2.9.14#1
* netcdf-c[core,dap,hdf5,nczarr,netcdf-4,platform-default-features]:x64-windows -> 4.8.1#2
* nlohmann-json[core]:x64-windows -> 3.11.2
* openssl[core]:x64-windows -> 3.0.5#5
pcl[core,qt,visualization,vtk]:x64-windows -> 1.12.0#6
* pcre2[core]:x64-windows -> 10.40
* pegtl-2[core]:x64-windows -> 2.8.3#1
* pkgconf[core]:x64-windows -> 1.8.0#3
* proj[core,net,tiff]:x64-windows -> 9.1.0
* pugixml[core]:x64-windows -> 1.12.1#1
* qhull[core]:x64-windows -> 8.0.2#3
* qt5-activeqt[core]:x64-windows -> 5.15.6
* qt5-base[core]:x64-windows -> 5.15.6#1
* qt5-declarative[core]:x64-windows -> 5.15.6
* qt5-imageformats[core]:x64-windows -> 5.15.6
* qt5-svg[core]:x64-windows -> 5.15.6
* qt5-tools[core]:x64-windows -> 5.15.6
* qt5-xmlpatterns[core]:x64-windows -> 5.15.6
* sqlite3[core,tool]:x64-windows -> 3.39.2
* tiff[core,jpeg,lzma,zip]:x64-windows -> 4.4.0#1
* utfcpp[core]:x64-windows -> 3.2.1#1
* vcpkg-pkgconfig-get-modules[core]:x64-windows -> 2022-02-10#1
* vtk[core,qt]:x64-windows -> 9.0.3-pv5.9.1#12
Additional packages (*) will be modified to complete this operation.
Detecting compiler hash for triplet x64-windows...
Restored 0 package(s) from C:\Users\DELL\AppData\Local\vcpkg\archives in 765 us. Use --debug to see more details.
Installing 1/32 icu:x64-windows...
Building icu[core]:x64-windows...
-- Using cached icu4c-71_1-src.tgz.
-- Cleaning sources at D:/Program Files (x86)/vcpkg/vcpkg/buildtrees/icu/src/c-71_1-src-08c83f7de1.clean. Use --editable to skip cleaning for the packages you specify.
-- Extracting source D:/Program Files (x86)/vcpkg/vcpkg/downloads/icu4c-71_1-src.tgz
-- Applying patch disable-escapestr-tool.patch
-- Applying patch remove-MD-from-configure.patch
-- Applying patch fix_parallel_build_on_windows.patch
-- Applying patch fix-extra.patch
-- Applying patch mingw-dll-install.patch
-- Applying patch disable-static-prefix.patch
-- Applying patch fix-win-build.patch
-- Applying patch check-autoconf-archive.patch
-- Using source at D:/Program Files (x86)/vcpkg/vcpkg/buildtrees/icu/src/c-71_1-src-08c83f7de1.clean
-- Found external ninja('1.10.2').
-- Getting CMake variables for x64-windows
CMake Warning at scripts/cmake/vcpkg_configure_make.cmake:203 (message):
Detected whitespace in root directory. Please move the path to one without
whitespaces! The required tools do not handle whitespaces correctly and the
build will most likely fail
Call Stack (most recent call first):
ports/icu/portfile.cmake:53 (vcpkg_configure_make)
scripts/ports.cmake:147 (include)
-- Using cached msys-gzip-1.11-1-x86_64.pkg.tar.zst.
-- Using cached msys-bash-5.1.008-1-x86_64.pkg.tar.zst.
-- Using cached msys-autoconf-2.71-3-any.pkg.tar.zst.
-- Using cached msys-autoconf-archive-2019.01.06-1-any.pkg.tar.xz.
-- Using cached msys-diffutils-3.8-2-x86_64.pkg.tar.zst.
-- Using cached msys-binutils-2.37-5-x86_64.pkg.tar.zst.
-- Using cached msys-libtool-2.4.6-9-x86_64.pkg.tar.xz.
-- Using cached msys-file-5.41-2-x86_64.pkg.tar.zst.
-- Using cached msys-zlib-1.2.11-1-x86_64.pkg.tar.xz.
-- Using cached msys-libbz2-1.0.8-3-x86_64.pkg.tar.zst.
-- Using cached msys-coreutils-8.32-2-x86_64.pkg.tar.zst.
-- Using cached msys-grep-3.0-2-x86_64.pkg.tar.xz.
-- Using cached msys-sed-4.8-2-x86_64.pkg.tar.zst.
-- Using cached msys-libpcre-8.45-1-x86_64.pkg.tar.zst.
-- Using cached msys-m4-1.4.19-2-x86_64.pkg.tar.zst.
-- Using cached msys-automake-wrapper-11-4-any.pkg.tar.zst.
-- Using cached msys-gawk-5.1.0-2-x86_64.pkg.tar.zst.
-- Using cached msys-mpfr-4.1.0-1-x86_64.pkg.tar.zst.
-- Using cached msys-gmp-6.2.1-1-x86_64.pkg.tar.zst.
-- Using cached msys-libreadline-8.1.001-1-x86_64.pkg.tar.zst.
-- Using cached msys-ncurses-6.2-2-x86_64.pkg.tar.zst.
-- Using cached msys-automake1.16-1.16.3-3-any.pkg.tar.zst.
-- Using cached msys-perl-5.32.1-2-x86_64.pkg.tar.zst.
-- Using cached msys-libcrypt-2.1-3-x86_64.pkg.tar.zst.
-- Using cached msys-pkg-config-0.29.2-4-x86_64.pkg.tar.zst.
-- Using cached msys-make-4.3-3-x86_64.pkg.tar.zst.
-- Using cached msys-findutils-4.8.0-1-x86_64.pkg.tar.zst.
-- Using cached msys-libintl-0.21-1-x86_64.pkg.tar.zst.
-- Using cached msys-libiconv-1.16-2-x86_64.pkg.tar.zst.
-- Using cached msys-gcc-libs-11.2.0-3-x86_64.pkg.tar.zst.
-- Using cached msys-msys2-runtime-3.2.0-8-x86_64.pkg.tar.zst.
-- Using cached msys-which-2.21-4-x86_64.pkg.tar.zst.
-- Using msys root at D:/Program Files (x86)/vcpkg/vcpkg/downloads/tools/msys2/16df26917335bb2d
-- Generating configure for x64-windows
-- Finished generating configure for x64-windows
-- Using cached msys-mingw-w64-i686-pkg-config-0.29.2-3-any.pkg.tar.zst.
-- Using cached msys-mingw-w64-i686-libwinpthread-git-9.0.0.6373.5be8fcd83-1-any.pkg.tar.zst.
-- Using msys root at D:/Program Files (x86)/vcpkg/vcpkg/downloads/tools/msys2/9a1ec3f33446b195
-- Configuring x64-windows-dbg
CMake Error at scripts/cmake/vcpkg_execute_required_process.cmake:96 (message):
Command failed: "D:/Program Files (x86)/vcpkg/vcpkg/downloads/tools/msys2/16df26917335bb2d/usr/bin/bash.exe" --noprofile --norc --debug -c "V=1 CPP='compile cl.exe -E' CC='compile cl.exe' CC_FOR_BUILD='compile cl.exe' CPP_FOR_BUILD='compile cl.exe -E' CXX_FOR_BUILD='compile cl.exe' CXX='compile cl.exe' RC='windres-rc rc.exe' WINDRES='windres-rc rc.exe' AR='ar-lib lib.exe' LD='link.exe -verbose' RANLIB=':' STRIP=':' NM='dumpbin.exe -symbols -headers' DLLTOOL='link.exe -verbose -dll' CCAS=':' AS=':' ./../src/c-71_1-src-08c83f7de1.clean/source/configure --build=x86_64-pc-mingw32 \"--enable-icu-build-win\" \"--disable-samples\" \"--disable-tests\" \"--disable-layoutex\" \"ac_cv_prog_ac_ct_STRIP=:\" \"gl_cv_double_slash_root=yes\" \"ac_cv_func_memmove=yes\" \"--disable-silent-rules\" \"--verbose\" \"--enable-shared\" \"--disable-static\" \"--enable-debug\" \"--disable-release\" \"--prefix=/D/Program Files (x86)/vcpkg/vcpkg/installed/x64-windows/debug\" \"--bindir=\\${prefix}/../tools/icu/debug/bin\" \"--sbindir=\\${prefix}/../tools/icu/debug/sbin\" \"--libdir=\\${prefix}/lib\" \"--includedir=\\${prefix}/../include\" \"--datarootdir=\\${prefix}/share/icu\""
Working Directory: D:/Program Files (x86)/vcpkg/vcpkg/buildtrees/icu/x64-windows-dbg
Error code: 1
See logs for more information:
D:\Program Files (x86)\vcpkg\vcpkg\buildtrees\icu\config-x64-windows-dbg-config.log
D:\Program Files (x86)\vcpkg\vcpkg\buildtrees\icu\config-x64-windows-dbg-out.log
D:\Program Files (x86)\vcpkg\vcpkg\buildtrees\icu\config-x64-windows-dbg-err.log
Call Stack (most recent call first):
scripts/cmake/vcpkg_configure_make.cmake:808 (vcpkg_execute_required_process)
ports/icu/portfile.cmake:53 (vcpkg_configure_make)
scripts/ports.cmake:147 (include)
error: building icu:x64-windows failed with: BUILD_FAILED
error: Please ensure you're using the latest port files with `git pull` and `vcpkg update`.
Then check for known issues at:
https://github.com/microsoft/vcpkg/issues?q=is%3Aissue+is%3Aopen+in%3Atitle+icu
You can submit a new issue at:
https://github.com/microsoft/vcpkg/issues/new?template=report-package-build-failure.md&title=[icu]+Build+error
Include '[icu] Build error' in your bug report title, the following version information in your bug description, and attach any relevant failure logs from above.
vcpkg-tool version: 2022-10-12-b586c2752f75bcc3f6a243749e9a0d94d0d93ccd
vcpkg-scripts version: 94ce0dab5 2022-10-19 (5 hours ago)
Please use the prefilled template from D:\Program Files (x86)\vcpkg\vcpkg\installed\vcpkg\issue_body.md when reporting your issue.
Please move vcpkg to a directory without spaces and try again. :)
does the issue still occur?
We hope your question was answered to your satisfaction; if it wasn't, you can reopen with more info.
| gharchive/issue | 2022-10-20T06:56:50 | 2025-04-01T06:44:59.043285 | {
"authors": [
"FrankXie05",
"munitioner"
],
"repo": "microsoft/vcpkg",
"url": "https://github.com/microsoft/vcpkg/issues/27344",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2426322228 | [New Port Request] libhsl
Library name
libhsl
Library description
HSL. A collection of Fortran codes for large scale scientific computation. http://www.hsl.rl.ac.uk/
Source repository URL
https://www.hsl.rl.ac.uk/catalogue/index.html
Project homepage (if different from the source repository)
https://www.hsl.rl.ac.uk/index.html
Anything else that is useful to know when adding (such as optional features the library may have that should be included)
No response
why i request
after build and install ipopt, i copied the sample
cpp_example.cpp
MyNLP.hpp
MyNLP.cpp
from vcpkg\buildtrees\coin-or-ipopt\src\30c9ada089-5859d8f5b5.clean\examples\Cpp_example.
then i changed the #include:
from
#include "IpIpoptApplication.hpp"
to
#include "coin-or/IpIpoptApplication.hpp"
but the program genarated runtime exception:
Exception of type: DYNAMIC_LIBRARY_FAILURE in file ".././../src/30c9ada089-5859d8f5b5.clean/src/Common/IpLibraryLoader.cpp" at line 59:
Exception message: Error 126 while loading DLL libhsl.dll: cannot find specified module.
EXIT: Library loading failure.
then i tried to type in cmd prompt:
C:\Users\Huang>vcpkg search hsl
The result may be outdated. Run `git pull` to get the latest results.
If your port is not listed, please open an issue at and/or consider making a pull request. - https://github.com/Microsoft/vcpkg/issues
the result is empty.
。
我遇到相同的问题并解决了:关键在于Ipopt包bin\ipopt-3.dll文件的大小。
我最初使用vcpkg安装了Ipopt,在VS和Clion上配置都出现该问题。
最后在Ipopt的git项目下载了Ipopt-3.14.16-win64-msvs2019-md.zip 包并配置后能正常运行,对比发现:vcpkg包中的bin\ipopt-3.dll文件大小仅几MB,而官网包下的对应文件100多MB。这就是问题原因。
I encountered the same issue and resolved it: the key lies in the size of the bin\ipopt-3.dll file in the Ipopt package.
Initially, I installed Ipopt using vcpkg and encountered the issue when configuring both VS and CLion. Eventually, I downloaded the Ipopt-3.14.16-win64-msvs2019-md.zip package from the Ipopt Git project, configured it, and everything worked correctly. Upon comparison, I found that the bin\ipopt-3.dll file in the vcpkg package is only a few megabytes, whereas the corresponding file in the official package is over 100 MB. This discrepancy was the root cause of the problem.
Got it !
Thank you @DuskandDawn
我遇到相同的问题并解决了:关键在于Ipopt包bin\ipopt-3.dll文件的大小。 我最初使用vcpkg安装了Ipopt,在VS和Clion上配置都出现该问题。 最后在Ipopt的git项目下载了Ipopt-3.14.16-win64-msvs2019-md.zip 包并配置后能正常运行,对比发现:vcpkg包中的bin\ipopt-3.dll文件大小仅几MB,而官网包下的对应文件100多MB。这就是问题原因。 I encountered the same issue and resolved it: the key lies in the size of the bin\ipopt-3.dll file in the Ipopt package. Initially, I installed Ipopt using vcpkg and encountered the issue when configuring both VS and CLion. Eventually, I downloaded the Ipopt-3.14.16-win64-msvs2019-md.zip package from the Ipopt Git project, configured it, and everything worked correctly. Upon comparison, I found that the bin\ipopt-3.dll file in the vcpkg package is only a few megabytes, whereas the corresponding file in the official package is over 100 MB. This discrepancy was the root cause of the problem.
@DuskandDawn 我猜是官方仓库默认静态链接了 libhsl,而 vcpkg 编译时改用动态链接。我当时没试着找 vcpkg 文件夹里有没有 libhsl.dll。现在我换了机器,还没试。
I guess the official repository statically links libhsl by default, while vcpkg uses dynamic linking during compilation. I didn't try to find libhsl.dll in the vcpkg folder at that time. Now I've switched machines and haven't tried it yet.
| gharchive/issue | 2024-07-24T00:01:38 | 2025-04-01T06:44:59.053595 | {
"authors": [
"DuskandDawn",
"HuangDuoYan",
"SKNo"
],
"repo": "microsoft/vcpkg",
"url": "https://github.com/microsoft/vcpkg/issues/40060",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
369799681 | Obscure vcpkg/CMake error when "Visual C++ tools for CMake" are not installed
I've had a bunch of libraries installed no problem, but then I tried installing openblas, and it failed on the project generation phase. vcpkg output mentioned a log file, which contained the following CMake error: Visual Studio 15 2017 could not find any instance of Visual Studio. Some googling lead me to a solution: install "Visual C++ tools for CMake" through the VS 2017 installer.
Couldn't vcpkg parse the logs to detect this specific error in order to suggest a solution, to save the users some time and effort?
@VioletGiraffe Yes, it could, vcpkg is open source and any PR, deemed useful, will be merged after CLA signing. I'm watching this space.
Now, vcpkg will install cmake automaticlly when cmake not found.
| gharchive/issue | 2018-10-13T12:54:01 | 2025-04-01T06:44:59.055749 | {
"authors": [
"JackBoosY",
"VioletGiraffe",
"degski"
],
"repo": "microsoft/vcpkg",
"url": "https://github.com/microsoft/vcpkg/issues/4470",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
538329973 | How do I ...
When asking a question please also include where you looked for an answer (so we can update the documentation if needed).
@Tsunoda1048 , thanks for your suggestion!
For the answer to the question, some documented, all docs located in https://github.com/microsoft/vcpkg/tree/master/docs. Some not, they will be tagged to 'documentation' if needed. We'd like to provide users the link to docs if there has.
| gharchive/issue | 2019-12-16T10:43:13 | 2025-04-01T06:44:59.057347 | {
"authors": [
"PhoebeHui",
"Tsunoda1048"
],
"repo": "microsoft/vcpkg",
"url": "https://github.com/microsoft/vcpkg/issues/9337",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
667124914 | [cgal] Upgrade CGAL to 5.1
This PR upgrades the CGAL pkg to the version 5.1.
For now it is only a test, as we are only at the beta stage.
@maxGimeno: I have updated the package to v5.1. I think you can un-draft the PR.
@LilyWangL: I am surprised that the CI test were fine, even if I have not changed the SHA256.
Thanks for your contribution!
| gharchive/pull-request | 2020-07-28T14:34:28 | 2025-04-01T06:44:59.059209 | {
"authors": [
"BillyONeal",
"lrineau",
"maxGimeno"
],
"repo": "microsoft/vcpkg",
"url": "https://github.com/microsoft/vcpkg/pull/12614",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
802768771 | [V8] Update to Stable version 8.8.278.14
This change also includes a change to use the
Windows SDK as reported by the WinSDK ENV variable.
Describe the pull request
Update to V8 8.8.278.14
What does your PR fix? Fixes #
May fix #15351
Which triplets are supported/not supported? Have you updated the CI baseline?
Same as before
x64-windows,
x64-windows-static,
x86-windows,
x86-windows-static,
x64-linux
Does your PR follow the maintainer guide?
Yes.
Thanks!
Thanks, that's what I wanted to know :)
Hi @JackBoosY , anything left to do for this merge? let me know.
@ras0219-msft ping for merge this PR.
@ras0219-msft Friendly poke here 😁
Depends on https://github.com/microsoft/vcpkg/pull/17341.
| gharchive/pull-request | 2021-02-06T19:11:34 | 2025-04-01T06:44:59.064073 | {
"authors": [
"JackBoosY",
"Kwizatz",
"ras0219-msft"
],
"repo": "microsoft/vcpkg",
"url": "https://github.com/microsoft/vcpkg/pull/16077",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1441811576 | [folly] Fix feature's find_dependency
Describe the pull request
What does your PR fix?
Fixes #27422, add feature option for some find_dependency, this is regression from PR #25335.
cc @jiayuehua
You must actually control CMAKE_REQUIRE_FIND_PACKAGE_ZLIB from the portfile.
I cannot see this was fixed.
Sigh, sorry, I dropped the ball there. I saw the comment in the files view but not the comment back here in 'conversation' view.
I super hate GitHub's code review tools sometimes :(
| gharchive/pull-request | 2022-11-09T10:14:19 | 2025-04-01T06:44:59.066452 | {
"authors": [
"BillyONeal",
"LilyWangLL"
],
"repo": "microsoft/vcpkg",
"url": "https://github.com/microsoft/vcpkg/pull/27728",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.