added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:36:43.439062
| 2018-09-25T21:18:10
|
363770304
|
{
"authors": [
"eliperelman",
"grsabreu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:222",
"repo": "Astrocoders/epitath",
"url": "https://github.com/Astrocoders/epitath/issues/8"
}
|
gharchive/issue
|
Can this technique be reliably used against Components?
Really interested in the technique being used here, and I am curious if this could also be extended for use against Components. I wrote a HOC to experiment with the idea, using the same demo code as the README, using Components instead of functions:
import React, { Component, cloneElement } from 'react';
import immutagen from 'immutagen';
const compose = ({ value, next }) => next
? cloneElement(value, null, values => compose(next(values)))
: value;
export default UnyieldableComponent => {
const generator = immutagen(UnyieldableComponent.prototype.render);
UnyieldableComponent.prototype.render = function() {
return compose(generator(this.props));
};
return class YieldableComponent extends Component {
render() {
return <UnyieldableComponent {...this.props} />;
}
};
};
// used as legacy decorator, but can use as HOC as well
@withGeneration
export default class App extends Component {
* render() {
console.log('Rendering again!');
const { loading, data } = yield <Query />;
const { time } = yield <Time/>;
if (loading) {
return <h1>Loading</h1>;
}
const {
values,
touched,
errors,
handleChange,
handleBlur,
handleSubmit,
isSubmitting,
} = yield (
<WrapFormik
initialValues={{
// Use data from other HOCs!
email: data.user.email,
password: '',
}}
validate={values => {
// same as above, but feel free to move this into a class method now.
let errors = {};
if (!values.email) {
errors.email = 'Required'
} else if (
!/^[A-Z0-9._%+-]+@[A-Z0-9.-]+\.[A-Z]{2,4}$/i.test(values.email)
) {
errors.email = 'Invalid email address';
}
return errors;
}}
/>
);
return (
<div className="App">
<h1>{`Hello, ${data.user.name}`}</h1>
<h2>The time is {time.toLocaleString()}!</h2>
<form onSubmit={handleSubmit}>
<input
type="email"
name="email"
onChange={handleChange}
onBlur={handleBlur}
value={values.email}
/>
{touched.email && errors.email && <div>{errors.email}</div>}
<input
type="password"
name="password"
onChange={handleChange}
onBlur={handleBlur}
value={values.password}
/>
{touched.password && errors.password && <div>{errors.password}</div>}
<button type="submit" disabled={isSubmitting}>
Submit
</button>
</form>
</div>
);
}
}
It seemed to work pretty well; do you see any concerns with this approach?
One thing that needs to be ironed out is accessing this within render. I'm sure a call(this needs to happen somewhere, but not sure where. :smile:
Yeah, this is one of the reasons I always prefer to use functional components.
I fear that it can add more margin for errors for beginners too π€
One thing that needs to be ironed out is accessing this within render.
I have solved this and updated the original post by caching the generator at first render with a binding.
|
2025-04-01T06:36:43.441042
| 2021-10-01T12:48:06
|
1013304392
|
{
"authors": [
"Astrodevil",
"ShouryaBrahmastra"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:223",
"repo": "Astrodevil/Programming-Basics",
"url": "https://github.com/Astrodevil/Programming-Basics/issues/177"
}
|
gharchive/issue
|
Finding the norm of a matrix using dynamic memory allocation
Would like do this using c
Can you please assign this
You have to make appropriate folder or choose proper directory to put this program of C
Link PR if made
https://github.com/Astrodevil/Programming-Basics/pull/289
Here it is sorry for the delay
|
2025-04-01T06:36:43.565626
| 2014-12-15T12:32:41
|
51980686
|
{
"authors": [
"gatormha",
"yemi"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:225",
"repo": "AtomLinter/Linter",
"url": "https://github.com/AtomLinter/Linter/issues/292"
}
|
gharchive/issue
|
Uncaught Error: spawn EACCES
Uncaught Error: spawn EACCES
Atom Version: 0.158.0
System: Mac OS X 10.10.1
Thrown From: linter package, v0.9.0
Steps To Reproduce
Happens everytime i open a .py file, i have linter-pep8 installed
Stack Trace
At child_process.js:1160
Error: spawn EACCES
at exports._errnoException (util.js:742:11)
at ChildProcess.spawn (child_process.js:1160:11)
at Object.exports.spawn (child_process.js:993:9)
at new BufferedProcess (/Applications/Atom.app/Contents/Resources/app/src/buffered-process.js:47:37)
at LinterPep8.Linter.lintFile (/Users/fabianrios/.atom/packages/linter/lib/linter.coffee:142:19)
at /Users/fabianrios/.atom/packages/linter/lib/linter-view.coffee:138:18
at Array.forEach (native)
at /Users/fabianrios/.atom/packages/linter/lib/linter-view.coffee:137:18
at Object.oncomplete (fs.js:93:15)
/cc @atom/core
I was having this same issue with linter-pep8 and was able to fix it by changing my config from:
'linter-pep8':
'pep8ExecutablePath': '/usr/local/bin'
to:
'linter-pep8':
'pep8ExecutablePath': '/usr/local/bin/pep8'
Hopefully that helps you!
Thanks that solved the problem!
|
2025-04-01T06:36:43.609562
| 2022-03-17T22:31:27
|
1172989629
|
{
"authors": [
"Aton-Kish",
"FreshDoktor"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:226",
"repo": "Aton-Kish/quartz-elevator",
"url": "https://github.com/Aton-Kish/quartz-elevator/issues/38"
}
|
gharchive/issue
|
change texture of elevator
Describe the solution you'd like
It would be nice if you can change the texture of the elevator to any block
Additional context
The elevator would be more fitting in the floor of the buildings
I have no plans to enable to change textures because this is the "Quartz" Elevator.
If you really want to change the textures, you can manually modify the model files.
https://minecraft.fandom.com/wiki/Model
quartz-elevator.jar
βββ assets
βββ quartzelv
βββ models
βββ block
β βββ smooth_quartz_elevator.json <- edit
β βββ quartz_elevator.json <- edit
βββ item
βββ smooth_quartz_elevator.json <- edit
βββ quartz_elevator.json <- edit
If you're bothered by particles, you can hide them. (ModMenu > Quartz Elvator > Display particles: Yes -> No)
|
2025-04-01T06:36:43.612629
| 2021-05-14T20:15:22
|
892208235
|
{
"authors": [
"scala-steward"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:227",
"repo": "Atry/scalajs-all-in-one-template",
"url": "https://github.com/Atry/scalajs-all-in-one-template/pull/121"
}
|
gharchive/pull-request
|
Update scalatest to 3.2.9
Updates org.scalatest:scalatest from 3.2.3 to 3.2.9.
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scalatest", artifactId = "scalatest" } ]
labels: test-library-update, semver-patch
Superseded by #127.
|
2025-04-01T06:36:43.626471
| 2022-01-24T07:17:21
|
1112260563
|
{
"authors": [
"Cohesion97",
"DeppMeng",
"onepeachbiubiubiu"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:228",
"repo": "Atten4Vis/ConditionalDETR",
"url": "https://github.com/Atten4Vis/ConditionalDETR/issues/14"
}
|
gharchive/issue
|
the parameters are only 43196001, instead of 43524961
I run the default Conddetr-r50, but the num of parameters is different from that in the provided log.
Also, after training for 1 epoch, the eval results are
[0.04369693586567375, 0.12083834673558262, 0.023675111814434113, 0.01864211602467282, 0.052261665895792626, 0.07171156446634068, 0.09023536974930606, 0.18654859799415718, 0.22196121793196433, 0.04610799601904764, 0.21023391350986004, 0.3797766209046455],
which is weaker (about 0.7AP) than that in the provided log
[0.0509964214370242, 0.13292741190993088, 0.030383986414032393, 0.015355903493298791, 0.05914294278060285, 0.08176101640052409, 0.10028554935230335, 0.2012481198582593, 0.23517722389597043, 0.04296950016312112, 0.23670937055006003, 0.40016568706711353].
Hi,
Did you enable the '--no_aux_loss' flag? We use aux_loss in training, and disabling this flag might cause fewer parameters as well as weaker performance. Moreover, the AP in the early training stage is unstable. +/- 0.7 AP at epoch 1 is not informative. Consistent lower performance in training (maybe epoch 1 to epoch 10) might indicate that the training has some problem.
Thanks for your answer.
I did not change the args. And I used aux loss during the training. The args are:
Namespace(aux_loss=True, backbone='resnet50', batch_size=2, bbox_loss_coef=5, clip_max_norm=0.1, cls_loss_coef=2, coco_panoptic_path=None, coco_path='/mnt/lustre/share/DSK/datasets/mscoco2017/', dataset_file='coco', dec_layers=6, device='cuda', dice_loss_coef=1, dilation=False, dim_feedforward=2048, dist_backend='nccl', dist_url='env://', distributed=True, dropout=0.1, enc_layers=6, epochs=50, eval=False, focal_alpha=0.25, frozen_weights=None, giou_loss_coef=2, gpu=0, hidden_dim=256, is_slurm_job=True, lr=0.0001, lr_backbone=1e-05, lr_drop=40, mask_loss_coef=1, masks=False, nheads=8, num_queries=300, num_workers=2, output_dir='output/default', position_embedding='sine', pre_norm=False, rank=0, remove_difficult=False, resume='', seed=42, set_cost_bbox=5, set_cost_class=2, set_cost_giou=2, start_epoch=0, tcp_port='29550', weight_decay=0.0001, world_size=8)
number of params: 43196001
BUT, after 5 epoch training, the model performs just the same as that in the provided logs. So, it might just be the performance fluctuation in the early stage caused by random seeds.
I will keep tracking the performance during the training and update the comment if some other problems happen.
I also encountered the problem of inconsistency in the amount of parameters.
|
2025-04-01T06:36:43.643223
| 2017-09-20T21:54:10
|
259313699
|
{
"authors": [
"aure",
"jldunk",
"megastep"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:229",
"repo": "AudioKit/AudioKit",
"url": "https://github.com/AudioKit/AudioKit/issues/1054"
}
|
gharchive/issue
|
XCode 9 - AudioKit 4 - Objective C
If I am missing something, I apologize, but I am seeing no visible interface for what used to work in the previous version of AudioKit.
AKAudioPlayer *player = [[AKAudioPlayer alloc] initWithFile...
There is no initWithFile now for the AKAudioPlayer object when using Objective C. I am actually getting this error with several objects like AKTimePitch, AKReverb, etc.
Mmh, I wouldn't be surprised if we're missing a few @objc qualifiers in the Swift code. I noticed that for some other classes I was using.
any ideas how to work around or do I need to wait for an AudioKit new version?
We'll likely need to make a new build, I think there's too many of these oversights. Hunting them all down might take some time though.
You could use the AudioKit develop branch and include the .xcodeproj in your project instead of the framework. Then, you could find the offensive classes, add the @objc and make a pull request with the changes. That would be amazingly helpful and get you on the right track the fastest. I can help you through screensharing if you need any help getting set up.
I would be more than happy to help. Is there any way for the time being to use the previous version I was using which was AudioKit 3.7.1 with XCode 9? I seem to remember the issue with this was around Swift 4.
Not as far as I know. iOS development is like this, constantly being shoved around by toolset, device, and operating system changes. I can only offer some commiseration!
I will start tonight with the develop branch and then do the pull request with my findings. Thanks again for your help.
I am closing this issue as I was able to compile my app last night using the develop branch xcode project adding @objc where needed. I'll do a pull request this morning, but basically found that several effects needed @objc added for the init so you could instantiate the object in obj-c.
@jldunk Are you close to making a pull request? We've got a few other bugs I'd like to address with a 4.0.1 release
Just uploaded changes.
Jason
On Sep 23, 2017, at 3:29 PM, Aurelius Prochazka<EMAIL_ADDRESS>wrote:
@jldunk Are you close to making a pull request? We've got a few other bugs I'd like to address with a 4.0.1 release
β
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub, or mute the thread.
|
2025-04-01T06:36:43.645733
| 2020-07-30T21:21:05
|
669206832
|
{
"authors": [
"aure",
"megastep"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:230",
"repo": "AudioKit/AudioKit",
"url": "https://github.com/AudioKit/AudioKit/pull/2204"
}
|
gharchive/pull-request
|
Balancer problem, making a PR to see if the issue happens on CI and other people's machines
This simple change, similar to ones that worked this morning, breaks AudioKit.h. Why?
Well I see that it fails to compile because of unknown references. Maybe it has something to do with Taylor's recent changes to make internal APIs less accessible?
Well I see that it fails to compile because of unknown references. Maybe it has something to do with Taylor's recent changes to make internal APIs less accessible?
No I don't think. I did this to 50 or 60 nodes in a previous commit without issue.
I see the issue now. The balancer hpp file was the last place that AKSoundpipeDSPBase.hpp was imported from a public header file.
I've made the AKSoundpipeDSP.hpp file into a .hpp/.mm pair but still having some issues.
|
2025-04-01T06:36:43.668437
| 2018-09-26T16:30:26
|
364114465
|
{
"authors": [
"coveralls",
"phoebemirman"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:231",
"repo": "AugurProject/augur-ui",
"url": "https://github.com/AugurProject/augur-ui/pull/2331"
}
|
gharchive/pull-request
|
fix for logout bug
Steps for repro:
on dev.augur.net
sign into account 1 in augur-ui
switch to account 2
sign back in
see it signs you out
locally
in load-account-data switch line 23 to be boolean opposite
do same steps as above
Issue: dev.augur.net doesn't use localstorage for account signed in data
Coverage decreased (-0.01%) to 63.209% when pulling 43591d20ac413cd11660dda4c5f42907d6aaebe9 on dev-logout-bug into 8107792ac6ce6e20fdf05842c7a871ee94b98beb on master.
Coverage decreased (-0.01%) to 63.209% when pulling 43591d20ac413cd11660dda4c5f42907d6aaebe9 on dev-logout-bug into 8107792ac6ce6e20fdf05842c7a871ee94b98beb on master.
|
2025-04-01T06:36:43.671151
| 2019-02-27T15:54:59
|
415195805
|
{
"authors": [
"Chwy5",
"bthaile"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:232",
"repo": "AugurProject/augur",
"url": "https://github.com/AugurProject/augur/issues/1156"
}
|
gharchive/issue
|
Order Form confirmation word change
flipping a short position:
closing position
Buying Back
New position
Buying Green, instead of long and short
Flipping a long position:
closing position
Selling Out Red
new position
Selling Red
working correctly in snpk
|
2025-04-01T06:36:43.674475
| 2019-04-27T21:02:56
|
437984029
|
{
"authors": [
"bthaile",
"phoebemirman"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:233",
"repo": "AugurProject/augur",
"url": "https://github.com/AugurProject/augur/issues/1929"
}
|
gharchive/issue
|
Portfolio Open Orders Sorting
Portfolio open orders section, it only has one sort method but two views
Sort by most recently traded market, not by the logged in user. The most recent market that has a trade should be at the top of the list.
View by Most Recently Traded Market
View by Most Recently Traded Outcome
added fixes to PR
for this view:
should individual orders still be grouped by markets?
|
2025-04-01T06:36:43.677311
| 2015-07-04T19:28:40
|
93040766
|
{
"authors": [
"carver",
"scottzer0",
"tinybike"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:234",
"repo": "AugurProject/augur",
"url": "https://github.com/AugurProject/augur/issues/94"
}
|
gharchive/issue
|
Uncaught TypeError: Cannot read property 'constructor' of null
When using client.augur.net on the main (Olympic) testnet.
Browser gets stuck at the loading screen with this at console:
//one of these:
Uncaught TypeError: Cannot read property 'constructor' of null
augur.run.augur.execute.augur.invoke @ app.js:7003
...
//followed by tons of these:
Uncaught TypeError: this.state.asset.cash.toFixed is not a function
render @ app.js:2147
...
Any suggestion on how to copy a full stack trace from chrome console without the rows smashing together?
i just fixed this. mainly an issue with the new geth account management notions changing underneath me. pushing coming soon.
You should be able to use Error().stack to get a stringified stack trace.
Sweet
|
2025-04-01T06:36:43.679778
| 2018-12-04T08:12:31
|
387165980
|
{
"authors": [
"aimm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:235",
"repo": "Augustyniak/RATreeView",
"url": "https://github.com/Augustyniak/RATreeView/issues/262"
}
|
gharchive/issue
|
Do not expand while selecting cell!
cells will auto expand while select it, but I don't want this feature, what should I do?
I'd like to expand a cell while tapped the button on this cell
|
2025-04-01T06:36:43.680699
| 2019-09-04T13:43:47
|
489169775
|
{
"authors": [
"Brainyoo",
"svetlanama"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:236",
"repo": "Augustyniak/RATreeView",
"url": "https://github.com/Augustyniak/RATreeView/pull/270"
}
|
gharchive/pull-request
|
ios13 fix
default added
This is an important fix. It helped during the migration to swift5, please merge this PR.
|
2025-04-01T06:36:43.696504
| 2018-05-16T06:22:41
|
323483164
|
{
"authors": [
"Aupajo",
"DerekFroese",
"steveb85"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:237",
"repo": "Aupajo/almanack",
"url": "https://github.com/Aupajo/almanack/issues/22"
}
|
gharchive/issue
|
Is there security?
This looks like it might solve a problem I've had for a few months!
Is there any way to require a Pre-Shared Key to open the web page or ical feed?
I'd like to aggregate all my google calendars into one for sharing to specific people to open in their GCal, but not open it to the world.
i'm interested in this as well.
@DerekFroese @steveb85 Yes! There's several options, depending on what you want to achieve:
An obfuscated URL (i.e. a pre-shared key in a URL), so you can have the feed at /my-shared-key/events.ics
HTTP Basic auth, with a username/password
OAuth (with a bit of work)
What kind of solution were you after?
Personally, I like the obfuscated URL. It will have maximal compatibility, as many consumers of ical feeds are not able to do HTTP Basic Auth much less OAuth.
i.e., you wouldn't be able to add options 2 or 3 to Google Calendar, but you could add option 1.
@DerekFroese I think both obfuscated URL and basic auth will work with Google Calendar.
Obfuscated URL
For the obfuscated URL, find the following line in your config.ru:
run Almanack::Server
Change it to:
SECRET_TOKEN = 'shhhh'
app = Rack::Builder.app do
map("/#{SECRET_TOKEN}") do
run Almanack::Server
end
end
run app
This will mount the calendar (and its feed) under /shhhh. If you want to avoid keeping the secret in your codebase (a good idea), I recommend using an environment variable:
SECRET_TOKEN = ENV.fetch('SECRET_TOKEN') { fail "Couldn't find a SECRET_TOKEN env var" }
Environment variables are available on any unix-y system. On Heroku, you can set this with:
heroku config:set SECRET_TOKEN=shhhh
If you're using the default theme, you'll need to override layout.erb to fix the paths to the stylesheet and JavaScript. (I'll fix this in a future release).
Basic Auth
I believe most calendar apps, including Google Calendar, support basic auth, through use of the optional username and password parts of a URL, i.e<EMAIL_ADDRESS>To use Basic Auth, find the following line in your config.ru:
run Almanack::Server
and change it to the following:
USERNAME = 'calendar'
PASSWORD = 'sshhhsecret'
use Rack::Auth::Basic, "My Calendar" do |given_username, given_password|
Rack::Utils.secure_compare(PASSWORD, given_password) && given_username == USERNAME
end
run Almanack::Server
This will protect the application using HTTP Basic Auth. Please serve this over SSL/TLS (i.e. HTTPS) to prevent the password being sent in the clear.
If you want to avoid keeping the secret in your codebase (a good idea), I recommend using an environment variable:
CREDENTIALS = ENV.fetch('CREDENTIALS') { fail "Couldn't find a CREDENTIALS env var" }
USERNAME, PASSWORD = credentials.split(':')
This assumes an environment variable called CREDENTIALS in the format username:password.
Environment variables are available on any system. On Heroku, you can set this with:
heroku config:set CREDENTIALS=username:password
Hope that helps!
Hi @DerekFroese. Did this solve your issue? Can I close this issue?
HI Pete,
Yes, the configurations you listed solve the issue. Thanks!
On Fri, 10 May 2019 at 13:40, Pete Nicholls<EMAIL_ADDRESS>wrote:
Hi @DerekFroese https://github.com/DerekFroese. Did this solve your
issue? Can I close this issue?
β
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/Aupajo/almanack/issues/22#issuecomment-491423422, or mute
the thread
https://github.com/notifications/unsubscribe-auth/AAD3MBVSLRRODE2JF4ECYWLPUXMSNANCNFSM4FAC2ZRA
.
--
Cheers,
Derek Froese
Great!
I found that all of these changes were overwritten when Heroku is reloaded. Is it possible to have the secret token implemented as a Config Var so it can persist?
Sorry to hear that, @DerekFroese. Can you elaborate? Changes should be made via git and pushed to the Heroku repo to persist between deploys. The above example demonstrates how to do this with a Heroku config environment variable.
Hi Aupajo,
If I understand correctly; I'd have to fork your repo and make my own in order to make changes to the code that persist across Heroku reboots and such. The problem for me is that my repo will become out-of-sync with your repo and will be an older version. I'm not sure I have the experience to keep my repo in sync with yours to have the latest version.
For my personal needs, it would be nice if the official code allowed for a config variable (set in Heroku) of an authentication token that would, if used, be required in the URL to access the calendar. But I also recognize most others may not need this and it's not fair of me to ask you to write code just for me :).
I apologize for my unfamiliarity; I have some small experience with PHP and web hosting, but Heroku is foreign to me.
Hi @DerekFroese. No you don't need to maintain a fork of this repo.
The installation steps are:
gem install almanack
alamanack new my-custom-calender
This will create a directory called my-custom-calendar, which is small Git repo containing a Gemfile that keeps you in sync with releases of this project, and a configuration file you can customise your Almanack set-up. When you run:
almanack deploy
It will create or update a Heroku app for you.
If you deployed using the βHeroku Deployβ button, then these steps were already performed for you. You can clone your existing Heroku git repository by logging in to Heroku, clicking on βSettingsβ and finding your βHeroku Git URLβ:
You can clone the Heroku repo locally:
git clone https://git.heroku.com/my-almanack-app.git
Make the changes to config.ru, and then commit push them back:
git add config.ru -m "Add authentication"
git push origin master
|
2025-04-01T06:36:43.704370
| 2022-12-22T17:05:23
|
1508237868
|
{
"authors": [
"SigmaX",
"markcoletti"
],
"license": "AFL-3.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:238",
"repo": "AureumChaos/LEAP",
"url": "https://github.com/AureumChaos/LEAP/pull/236"
}
|
gharchive/pull-request
|
Feature/issue 226
Crowding distance repaired and notebooks in sync. New notebook by Luke cleaned up for formal integration.
@markcoletti A test is failing:
|
2025-04-01T06:36:43.707956
| 2023-10-05T18:38:39
|
1928847811
|
{
"authors": [
"AuroPick",
"chrom007",
"gelones"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:239",
"repo": "AuroPick/epic-free-games",
"url": "https://github.com/AuroPick/epic-free-games/issues/21"
}
|
gharchive/issue
|
Request failed with status code 404
After updating today's games, any search you run with any region the API returns this error, I was making a command yesterday and it was working, then today after updating this week's new games it gave me this
Erro: Error: An error occurred
error: {
"message": "CatalogOffer/offerMappings: Request failed with status code 404",
"locations": [
{}
],
"correlationId": "1792c500-beb8-4854-9d69-43e4a9a30e54",
"serviceResponse": "{\"errorMessage\":\"The item or resource being requested could not be found.\",\"errorCode\":\"errors.com.epicgames.not_found\",\"numericErrorCode\":1004,\"errorStatus\":404}",
"stack": null,
"path": [
"Catalog",
"searchStore",
"elements",
3,
"offerMappings"
]
},{
"message": "CatalogNamespace/mappings: Request failed with status code 404",
"locations": [
{}
],
"correlationId": "1792c500-beb8-4854-9d69-43e4a9a30e54",
"serviceResponse": "{\"errorMessage\":\"The item or resource being requested could not be found.\",\"errorCode\":\"errors.com.epicgames.not_found\",\"numericErrorCode\":1004,\"errorStatus\":404}",
"stack": null,
"path": [
"Catalog",
"searchStore",
"elements",
3,
"catalogNs",
"mappings"
]
}
Same error...
I have just released v4.0.2. Please Install the new version.
|
2025-04-01T06:36:43.715364
| 2020-07-15T23:48:27
|
657733281
|
{
"authors": [
"BehrRiley"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:240",
"repo": "AuroraInteractive/dDiscordBot",
"url": "https://github.com/AuroraInteractive/dDiscordBot/issues/1"
}
|
gharchive/issue
|
Reaction Event - Context Requests
Would like to request one additional context tags:
[x] <context.group> - Returns the guild Group that the reaction is heard from
Would like to request some revisions:
[x] <context.user> returns the DiscordUser object; can we change this to <context.author> for consistency with other discord events?
[x] <context.emoji_id> returns the literal name of the emoji as opposed to the ID of the emoji; some emoji's from different groups can contain the same name where-as the ID is always directly associated with the correct emoji.
Would like to also add meta for this event to be called by our clone of the Meta bot if possible
Lastly, would it be difficult to add a DiscordEmoji object type that could return each of the emoji fields (Excluding the User, which is provided with the user context we have now) listed in the Emoji Object Structure?:
{
"id": "41771983429993937",
"name": "LUL",
"roles": ["41771983429993000", "41771983429993111"],
"user": {
"username": "Luigi",
"discriminator": "0002",
"id": "96008815106887111",
"avatar": "5500909a3274e1812beb4e8de6631111"
},
"require_colons": true,
"managed": false,
"animated": false
}
[x] ID | Emoji IDs would differentiate the Emoji from every other emoji regardless of name, animation or group it originated from
[x] Name | Great for formatted name of the emoji
[ ] Roles | This specifically would return the list of roles that the specified channel has whitelisted the roles to via the 'Allow Reactions' and 'Allow External Reactions' permissions
[ ] Require_colons | This may only be beneficial if we could utilize this as opposed to regex to coordinate easier message replacements of emojis
[ ] Managed | Returns if this emoji is implemented by this group or not
[x] Animated | Returns if this emoji is animated; may only be beneficial for the same reason as the require_colons field
[ ] Available | This would return true or false if the server loses access to this emoji due to Server Boost falling to a level that reduces the server's emoji count
Thanks :}
We have everything we need from this currently for the Ticket system; closing and leaving the unmarked for future potential features we may or may not need in the future.
|
2025-04-01T06:36:43.795946
| 2017-11-08T19:14:12
|
272315716
|
{
"authors": [
"jcasale",
"moodmosaic",
"ploeh",
"zvirja"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:241",
"repo": "AutoFixture/AutoFixture",
"url": "https://github.com/AutoFixture/AutoFixture/issues/919"
}
|
gharchive/issue
|
Add support for string based properties decorated with the RangeAttribute
Consider the following:
public class MyModel
{
[Required]
[Range(1, 65535)]
public string Port { get; set; }
}
var fixture = new Fixture();
var model = fixture.Create<MyModel>();
In 3.51.0 the first call the Create() worked, however subsequent calls threw, now in 4.0.0-rc1 (the version I am using), the first call throws ObjectCreationException whereas the attribute itself offers support for the use case.
Originated from https://stackoverflow.com/q/47184400/126014
@jcasale Thanks for raising the issue.
Well, as far as I noticed, AutoFixture wasn't designed to handle gracefully the Range attribute applied to non-numeric members. Therefore, I'm basically not surprised as we already have the same issue for enums. They should be fixed together and I've already started to investigate that.
I don't have too much experience with annotation attributes and at the first glance it looks a bit weird that you apply numeric range attribute to a property of string type. Could you please point to some documentation describing that such usage is perfectly valid (and what is the expected behavior)? Just want to ensure that we cover valid scenarios :blush:
P.S. :+1: for using AutoFixture v4 :wink:
To be honest, it's a convention I am used to from the ASP arena and I don't know if it's an abuse of the facility or intended. I will check out the docs and if not, the reference source for logic that indicates it's valid and expected and not simply a fragile coincidence.
Reference source implies it will handle casting a string (which it does) and the docs illustrate an example where the the value being applied is coerced and in that case, boxed. I have a workaround and I admit the use case is rare, however it does appear to be valid but certainly low priority if any.
@jcasale Thanks for the confirmation. Well, it's indeed seems that string could be a valid type. Also it looks like that Range attribute could support even other types, like DateTime. The question is whether we are going to support that in AutoFixture out-of-the-box as we clearly cannot support all those types.
Probably, it makes sense to follow the way, when we support this feature partially - recognize the Range attribute and wrap that to the RangedRequest. Later clients could add customization to handle the RangedRequest for the custom types they have. The benefit is that you will not need to handle the Range attribute directly. Also we as a project will not need to support the variety of different possible types as they might be quite rare.
@jcasale What would you say about that plan?
@moodmosaic Your opinion is also appreciated as the topic is quite tricky.
The purpose of supporting data annotations in AutoFixture is to provide a more fine-grained control over the scope of generated values.
However, some of those data annotations have a totally weird API, where you can easily do the wrong thing, as with the RangeAttribute, which has 3 constructor overloads:
RangeAttribute(Double, Double)
RangeAttribute(Int32, Int32)
RangeAttribute(Type, String, String)
And because of that pesky 3rd constructor overload accepting a Type, it is possible to do this:
[Range(1, long.MaxValue)]
public long SomeProperty { get; set; }
1 gets casted into a Double
long.MaxValue gets casted into a Double resulting to an arithmetic overflow.
So, IMHO, and AFAICT, in this case:
[Range(1, 65535)]
public string Port { get; set; }
we should throw an error. In the error message we should probably tell to the user that the right way of controlling the scope of generated strings is by doing this:
[Range(typeof(string), "1", "65535")]
public string Port { get; set; }
And then, we'd have to make sure we support not only strings, but chars, dates, and so on:
[Range(typeof(DateTime),"14-Dec-1984", "14-Dec-2099")]
[Range(typeof(char), "n", "s")]
This is one of the reasons that F# Hedgehog makes this concept explicit, and easier, through the Range type and combinators.
In the error message we should probably tell to the user that the right way of controlling the scope of generated strings is by doing this:
[Range(typeof(string), "1", "65535")]
public string Port { get; set; }
That is incorrect, it works coincidentally for values where none of the place values exceed the leading place values of the max string. For example, 1 through 6 pass, however 7 through 9 fail, 10 through 65 pass, 66 to 99 fail, 100 through to 655 pass then 656 fails etc.
A contrived example:
using System;
using System.Collections.Generic;
using System.ComponentModel.DataAnnotations;
using System.Linq;
using System.Reflection;
internal class Program
{
[Range(typeof(string), "1", "65535")]
// [Range(1, 65535)]
public string Port { get; set; }
private static void Main()
{
for (int i = 1; i < 65536; i++)
{
Program program = new Program
{
Port = i.ToString()
};
foreach (var error in program.Validate())
{
Console.WriteLine($"{i}, {error.ErrorMessage}");
}
}
}
}
public static class Extensions
{
public static IEnumerable<ValidationResult> Validate<T>(this T model)
where T : class
{
if (model == null)
{
throw new ArgumentNullException(nameof(model));
}
foreach (PropertyInfo propertyInfo in model.GetType().GetProperties())
{
object[] attributes = propertyInfo.GetCustomAttributes(typeof(ValidationAttribute), false);
if (attributes.Length == 0)
{
continue;
}
ValidationContext validationContext = new ValidationContext(model)
{
DisplayName = propertyInfo.Name
};
if (attributes.OfType<RequiredAttribute>().FirstOrDefault() is RequiredAttribute required)
{
ValidationResult result = required.GetValidationResult(propertyInfo.GetValue(model), validationContext);
if (result != null)
{
yield return result;
yield break;
}
}
foreach (ValidationAttribute attribute in attributes)
{
ValidationResult result = attribute.GetValidationResult(propertyInfo.GetValue(model), validationContext);
if (result == null)
{
continue;
}
yield return result;
}
}
}
}
@jcasale Thanks for the sample.
That's why I'd suggest to not include this feature to the AutoFixture for now as there might be a whole set of different options depending on the OperandType/MemberType combinations.
In the #920 I've introduced the generic RangedRequest in a way that later any specimen builder could decide what to do. Request offers both the OperandType (this is how the Ranged attribute refers to the type you specified) and MemberData - type of the member you applied this property to.
Later @jcasale could register it's own RangedRequest builder to handle strings in a way he wants, without a need to deal with the RangedAttribute directly.
For me that looks like a good trade-off.
@zvirja No objections, I have a workaround and realized some good takeaways from this. In reference to registering my own RangedRequest, off the top of your head do you know of any code exercising the concept that I could review to see how this is accomplished?
any code exercising the concept that I could review
Probably, not for now, as #920 is still under the review by @moodmosaic. Only after we merge the PR, we'll know its shape, so I'll be able to show you a demo.
However, the best place would be to look at the NumericRangedRequestRelay implementation (if it doesn't changes during the review) as it's a sample of builder for numeric types.
@jcasale Feel free to use the NumericRangedRequestRelay or EnumRangedRequestRelay as a sample of such customization.
This API will be available since our next release that should happen in the nearby future.
Closing this one as no further action is required so far. Feel free to ask more questions if you have π
|
2025-04-01T06:36:43.798496
| 2016-06-22T16:24:15
|
161727259
|
{
"authors": [
"caloggins",
"ploeh"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:242",
"repo": "AutoFixture/AutoFixture",
"url": "https://github.com/AutoFixture/AutoFixture/pull/661"
}
|
gharchive/pull-request
|
Add Support for FakeItEasy 2.1+ (#628)
Went with 2.1 plus, since it's the latest as of now.
I don't know why it fails to a missing assembly.
Thank you for your interest in contributing to AutoFixture! This looks promising :+1: The few comments I have are superficial, and are easy to address :smile:
Thank you for your contribution! It's now live as AutoFixture.AutoFakeItEasy2 3.48.0.
|
2025-04-01T06:36:43.883408
| 2023-07-09T21:08:33
|
1795564390
|
{
"authors": [
"hollandjg",
"younesStrittmatter"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:243",
"repo": "AutoResearch/autora-core",
"url": "https://github.com/AutoResearch/autora-core/pull/26"
}
|
gharchive/pull-request
|
Feat: general wrapper
Description
a general wrapper that accepts arguments to field mappings as arguments
Type of change
feat: A new feature
I like the look of this!
Would it be sensible to split up the wrapper into two β one which does the input name mapping, and one which does the output wrapping? Then we could use the two independently β perhaps your inner function needs to return two values (say conditions and experiment_js_code) and returns those as a Result/Delta object, but you still want a mapping. Conversely, perhaps you have a function which uses entirely standard naming for the variables, but you want to wrap the outputs.
hey @younesStrittmatter , do you think you could refactor this so that there's a wrapper function which just does the input field-name mapping? We could include that as an option in the on_state function from #33, but it needs to be an independent wrapper function first so that we can test it really extensively on its own.
I think this is covered by the newest version of the state object β closing this now.
|
2025-04-01T06:36:43.884369
| 2016-04-26T10:09:24
|
151094535
|
{
"authors": [
"MRisto"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:244",
"repo": "AutoSimDevelopers/automata-simulation",
"url": "https://github.com/AutoSimDevelopers/automata-simulation/issues/23"
}
|
gharchive/issue
|
Check collection of input words
There should be an option to check a collection of input words, for better automaton validation.
Integrated in the next version
|
2025-04-01T06:36:43.888439
| 2016-10-24T02:21:57
|
184735624
|
{
"authors": [
"avirshup",
"dionjwa"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:245",
"repo": "Autodesk/molecular-design-toolkit",
"url": "https://github.com/Autodesk/molecular-design-toolkit/issues/107"
}
|
gharchive/issue
|
You should consider upgrading via the 'pip install --upgrade pip' command.
111
It would be really great if there was a Dockerfile in the root directory that had the correct versions of everything to run and test, independent of the host setup.
Also tried to install and run on my machine. It installed ok with sudo, but could not run:
/usr/bin/python: No module named moldesign
β-1 ~/autodesk/cloud-compute-cannon [stream_std_out_err_wip L|β 3β 2]
19:26 $ which pip
/usr/local/bin/pip
β ~/autodesk/cloud-compute-cannon [stream_std_out_err_wip L|β 3β 2]
19:27 $ which python
/usr/bin/python```
Thanks @dionjwa .
It doesn't install in the docker image because it's defaulting to Python 3, and MDT is still python 2 only. However, it should have printed out an error message stating that explicitly instead of dying with a syntax error - that's a bug.
Agreed, we need a root Dockerfile. Will add it.
Local installation doesn't work: :headdesk: Python + MacOS = sadness. Probably try switching to the Homebrew python install, or, per @hainm in #32, try pip install --user.
FYI @dionjwa - actually, the easiest way to deploy, as long as you're not doing development, is to pull the official docker image:
docker run -p 8888:8888 -it autodesk/moldesign:moldesign_notebook-0.7.3
Fixed with #108
|
2025-04-01T06:36:43.892842
| 2017-03-06T15:45:59
|
212161818
|
{
"authors": [
"avirshup",
"dionjwa"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:246",
"repo": "Autodesk/molecular-simulation-tools",
"url": "https://github.com/Autodesk/molecular-simulation-tools/issues/195"
}
|
gharchive/issue
|
Front-end Widget Catalogue
If the user can add front-end widgets to apps, where are those widgets stored, and how can they be referred to?
Deliverables:
Database of front-end widgets. (These could simply be tagged URLs for iframe widgets).
API to add/remove/edit widgets from the catalogue
Catalogue must support versioning
Policy for users to add their widgets (everything whitelisted or manually curated?)
Front-end Searching and viewing widgets.
Testing out widgets?
This ticket needs to be broken down when specced out more thoroughly.
Can this just be a publicly available github repo that we control final pull requests? This then specifies submission policies, potential license agreements, versioning.
@dionjwa - I really really like the idea of contributing workfows/apps via pull request to a master repository. This is exactly how Conda Forge manages its contributions. In addition to having a clear, well-understood contribution policy, the whole github infrastructure would also take care of the problems of scaling, community management, etc.
For right now, the priority is definitely on workflow development, so let's keep most of this stuff on the backlog for now.
We will definitely need a widget catalog for workflow/app developers; but as the widget themselves will just be MST components, this doesn't need to be anything more than a JSON document in the MST repo.
|
2025-04-01T06:36:43.894153
| 2017-02-15T09:26:48
|
207748946
|
{
"authors": [
"danielholmes",
"justinmc"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:247",
"repo": "Autodesk/molecule-3d-for-react",
"url": "https://github.com/Autodesk/molecule-3d-for-react/pull/24"
}
|
gharchive/pull-request
|
Remove rendering when empty model data set on component
There was a bug where when you change the modelData to an empty instance (i.e. { atoms: [], bonds: [] }) the previously rendered model isn't removed. This can quickly be verified in npm run example.
Yup I was able to reproduce the bug. I think this looks good. I'm going to add another test or two just to prove to myself it works and then I think it's good to go.
|
2025-04-01T06:36:44.122104
| 2024-01-10T07:35:27
|
2073772337
|
{
"authors": [
"janboddez",
"jaschaurbach"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:248",
"repo": "Automattic/wordpress-activitypub",
"url": "https://github.com/Automattic/wordpress-activitypub/issues/644"
}
|
gharchive/issue
|
Option to set visibility of answers to comment
What
I would like to be able to set the visibility of answers to βnot listedβ in general and be able to set it for every comment before posting it on demand.
Why
Answering to many comments would clutter the timeline of folks just following the website.
The feature of sending answers to comments back to the fediverse is great but may be not usable in cases with a hugh ammount of interactions.
Sometimes if a comment from the fediverse gets big traction one might want to answer public in the timeline.
How
No response
I would like to be able to set the visibility of answers to βnot listedβ in general and be able to set it for every comment before posting it on demand.
Ooh, this sounds like something I may want to add to my add-on plugin.
|
2025-04-01T06:36:44.287996
| 2023-12-02T17:37:14
|
2022138286
|
{
"authors": [
"Coloryr",
"Dragorn421",
"timunie"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:249",
"repo": "AvaloniaUI/Avalonia",
"url": "https://github.com/AvaloniaUI/Avalonia/issues/13807"
}
|
gharchive/issue
|
OpenGlControlBase enters OnOpenGlInit with OpenGL errors
Describe the bug
When implement OnOpenGlInit in an OpenGlControlBase and checking gl.GetError() for any error, there is one, presumably from something in Avalonia's code.
To Reproduce
Steps to reproduce the behavior:
Clone, build and run https://github.com/Dragorn421/DragoStuff/tree/ffe12b1003c3e6cd92419d7fdaab97b3c90c5dc4
Check the debug prints:
GL1.OnOpenGlInit
GL1.CheckError 1280
GL1.OnOpenGlRender
GL1.CheckError OK
GL1.CheckError OK
GL1.OnOpenGlRender
GL1.CheckError OK
GL1.CheckError OK
Notice GL1.CheckError 1280. It is printed by https://github.com/Dragorn421/DragoStuff/blob/ffe12b1003c3e6cd92419d7fdaab97b3c90c5dc4/MyOpenGLControl.cs#L29 , before the child control does any OpenGL call, so it seems to be an error originating from something in Avalonia.
Expected behavior
gl.GetError() should always return GL_NO_ERROR on entering child/user methods like OnOpenGlInit
Environment
OS: Kubuntu 23.10 (Linux, X11)
Avalonia-Version: 11.0.5
1280 is mean GL_INVALID_ENUM
Some operations use incorrect parameters, but this is related to the version of OpenGL
Still in Avalonia 11.0.10
@Dragorn421 if you wanted to check newer version, worth to try 11.1 betas
Still in Avalonia 11.1.1
|
2025-04-01T06:36:44.290694
| 2021-09-14T17:56:30
|
996288008
|
{
"authors": [
"grokys"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:250",
"repo": "AvaloniaUI/Avalonia",
"url": "https://github.com/AvaloniaUI/Avalonia/issues/6592"
}
|
gharchive/issue
|
Win32: Doesn't receive next key after Alt or F10
Describe the bug
When the Alt or F10 key is pressed, the next keypress is not received.
To Reproduce
Run the following program and press Alt or F10, followed by another key. The key subsequent to Alt or F10 is not registered.
public class MainWindow : Window
{
public MainWindow()
{
this.InitializeComponent();
this.AttachDevTools();
}
protected override void OnKeyDown(KeyEventArgs e)
{
System.Diagnostics.Debug.WriteLine("KeyDown " + e.Key);
base.OnKeyDown(e);
}
}
Desktop (please complete the following information):
OS: Windows
Version: master
Appears to be caused by our (not) handling of WM_ENTERIDLE?
|
2025-04-01T06:36:44.294077
| 2023-04-28T16:00:08
|
1688795342
|
{
"authors": [
"BAndysc",
"MrJul",
"avaloniaui-team",
"grokys"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:251",
"repo": "AvaloniaUI/Avalonia",
"url": "https://github.com/AvaloniaUI/Avalonia/pull/11178"
}
|
gharchive/pull-request
|
Correctly remove ContentPresenter's content from its parent host
Correctly remove ContentPresenter's content from its parent host when the content is updated while being detached from the logical tree.
I've updated the existing ContentControlTests.Should_Set_Child_LogicalParent_After_Removing_And_Adding_Back_To_Logical_Tree test with asserts that failed before, and now pass with this change. This seems acceptable to me as it already tested the right thing (but only from the child β parent side, now from both sides), tell me if you prefer a whole new test instead.
Fixes #11149
You can test this PR using the following package version. 11.0.999-cibuild0034002-beta. (feed url: https://pkgs.dev.azure.com/AvaloniaUI/AvaloniaUI/_packaging/avalonia-all/nuget/v3/index.json) [PRBUILDID]
Wow, thanks for finding that - looks like this bug's been in there forever!
Yay, this fixes https://github.com/AvaloniaUI/Avalonia/issues/9940 which was very annoying, thanks!
|
2025-04-01T06:36:44.310855
| 2019-12-05T10:21:04
|
533256418
|
{
"authors": [
"Aviana",
"Slivo-fr"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:252",
"repo": "Aviana/YaHT",
"url": "https://github.com/Aviana/YaHT/issues/41"
}
|
gharchive/issue
|
Licencing and code usage
Hey !
I'm starting to write an addon aiming to help hunters with tranqshot rotation.
As I'm really new to addon, I used YaHT as a reference and inspiration for the base and announces part.
The addon should grow on his own way past a first basic release.
As our code is really similar and I found no licence there, I have to ask you if you are fine with this.
Repo at https://github.com/Slivo-fr/TranqRotate
Thanks for your time
Added a license.
Thanks <3
|
2025-04-01T06:36:44.327309
| 2024-07-01T08:30:43
|
2383197453
|
{
"authors": [
"guillaume-chervet",
"radk0s"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:253",
"repo": "AxaFrance/oidc-client",
"url": "https://github.com/AxaFrance/oidc-client/issues/1395"
}
|
gharchive/issue
|
Communication with Service Worker causes memory leak
Issue and Steps to Reproduce
Issue is reproducible on demo deployment https://icy-glacier-004ab4303.2.azurestaticapps.net/. When you log-in and app is opened for few minutes memory usage is increasing. It can be checked in Chrome Dev Tools/Memory panel.
Versions
7.22.8
Screenshots
Expected
When app is open (no actions), memory usage is not increasing over time.
Actual
I did few memory snapshots and noticed that most of new memory allocations are coming from MessagePort class. After digging into oidc-client code I found this util used for Service Worker communication which creates MessageChannel but it's never closed after promise resolution.
Additional Details
Thank you very much @radk0s
hi @radk0s version 7.22.9 contains your fix!
Great, thanks @guillaume-chervet! Already checked and issue is gone.
|
2025-04-01T06:36:44.330332
| 2023-03-06T17:38:48
|
1611914570
|
{
"authors": [
"AykutSarac",
"aaron-ballard-530"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:254",
"repo": "AykutSarac/jsoncrack-vscode",
"url": "https://github.com/AykutSarac/jsoncrack-vscode/issues/24"
}
|
gharchive/issue
|
[BUG]: No Save/Download option with VSCode
Issue description
There is no option to save the rendered mapping with the VSCode Extension. The Save and SaveAs do not seem to have any connection with the generated diagram
Media & Screenshots
No response
Operating system
Version: 1.75.1
Commit: 441438abd1ac652551dbe4d408dfcec8a499b8bf
Date: 2023-02-08T21:34:59.000Z
Electron: 19.1.9
Chromium: 102.0.5005.194
Node.js: 16.14.2
V8: <IP_ADDRESS>-electron.0
OS: Darwin x64 21.6.0
Sandboxed: No
Priority this issue should have
Low (slightly annoying)
Merging under #29
|
2025-04-01T06:36:44.361993
| 2021-06-13T05:01:05
|
919716855
|
{
"authors": [
"Amit366",
"anshyyy"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:255",
"repo": "Ayush7614/Daily-Coding-DS-ALGO-Practice",
"url": "https://github.com/Ayush7614/Daily-Coding-DS-ALGO-Practice/pull/735"
}
|
gharchive/pull-request
|
Regular Expression -anshyyy
Description
Please include a summary of the change and which issue is fixed. List any dependencies that are required for this change.
Fixes #(issue_no)
Replace issue_no with the issue number which is fixed in this PR
Type of change
Please delete options that are not relevant.
Checklist:
[1 ] My code follows the style guidelines(Clean Code) of this project
[1 ] I have performed a self-review of my own code
[1 ] I have commented my code, particularly in hard-to-understand areas
[1] I have made corresponding changes to the documentation
[1] My changes generate no new warnings
@anshyyy issue number??
@anshyyy any update??
|
2025-04-01T06:36:44.375217
| 2024-03-02T07:04:15
|
2164665203
|
{
"authors": [
"charlielye",
"miles-six"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:256",
"repo": "AztecProtocol/aztec-connect",
"url": "https://github.com/AztecProtocol/aztec-connect/pull/64"
}
|
gharchive/pull-request
|
Fix 31 issues
I have been studying this repository for a long time, and whenever I come across a typo, I fix it.
I discovered 31 typos in a total of 29 files. Then I merged them together.
The code quality is exceptionally high, so sincerely hope my PR can help make this repository more standardized.
We don't accept typo fixes directly, but thank you for your contribution.
|
2025-04-01T06:36:44.376555
| 2023-08-17T17:56:56
|
1855460861
|
{
"authors": [
"dbanks12"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:257",
"repo": "AztecProtocol/aztec-packages",
"url": "https://github.com/AztecProtocol/aztec-packages/issues/1622"
}
|
gharchive/issue
|
Temporary hack: remove public kernel checks that state updates are in correct order
Do #1616 (and maybe #1617) first
Followup: #1623
Will be closed by https://github.com/AztecProtocol/aztec-packages/pull/1685
|
2025-04-01T06:36:44.430552
| 2024-04-03T17:08:51
|
2223462329
|
{
"authors": [
"AztecBot",
"fcarreiro"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:258",
"repo": "AztecProtocol/aztec-packages",
"url": "https://github.com/AztecProtocol/aztec-packages/pull/5561"
}
|
gharchive/pull-request
|
rollback FunctionAbi isTranspiled changes
Please read contributing guidelines and remove this line.
#5561 π
master
This stack of pull requests is managed by Graphite. Learn more about stacking.
Join @fcarreiro and the rest of your teammates on Graphite
Benchmark results
Metrics with a significant change:
note_trial_decrypting_time_in_ms (32): 112 (+218%)
Detailed results
All benchmarks are run on txs on the Benchmarking contract on the repository. Each tx consists of a batch call to create_note and increment_balance, which guarantees that each tx has a private call, a nested private call, a public call, and a nested public call, as well as an emitted private note, an unencrypted log, and public storage read and write.
This benchmark source data is available in JSON format on S3 here.
Values are compared against data from master at commit a581e80d and shown if the difference exceeds 1%.
L2 block published to L1
Each column represents the number of txs on an L2 block published to L1.
Metric
8 txs
32 txs
64 txs
l1_rollup_calldata_size_in_bytes
676
676
676
l1_rollup_calldata_gas
6,424
6,424
6,412
l1_rollup_execution_gas
585,757
585,757
585,745
l2_block_processing_time_in_ms
1,337 (+1%)
4,792 (-1%)
9,106 (+1%)
note_successful_decrypting_time_in_ms
248 (+5%)
601 (-4%)
1,036 (+3%)
note_trial_decrypting_time_in_ms
53.0 (-50%)
:warning: 112 (+218%)
32.7 (+27%)
l2_block_building_time_in_ms
11,985 (+1%)
43,796 (+1%)
86,804 (+1%)
l2_block_rollup_simulation_time_in_ms
7,182 (+1%)
24,860
47,857 (-1%)
l2_block_public_tx_process_time_in_ms
4,762 (+1%)
18,807 (+1%)
38,705 (+3%)
L2 chain processing
Each column represents the number of blocks on the L2 chain where each block has 16 txs.
Metric
5 blocks
10 blocks
node_history_sync_time_in_ms
14,048 (-2%)
26,406
note_history_successful_decrypting_time_in_ms
1,272
2,470 (+4%)
note_history_trial_decrypting_time_in_ms
93.0 (+29%)
118 (+15%)
node_database_size_in_bytes
18,657,360
35,082,320 (+1%)
pxe_database_size_in_bytes
29,859
59,414
Circuits stats
Stats on running time and I/O sizes collected for every circuit run across all benchmarks.
Circuit
circuit_simulation_time_in_ms
circuit_input_size_in_bytes
circuit_output_size_in_bytes
private-kernel-init
181
44,377
26,164
private-kernel-ordering
162
50,830
39,325
base-parity
4,366
128
311
root-parity
1,171 (+1%)
1,244
311
base-rollup
14,514
116,608
861
root-rollup
49.9
4,359
725
private-kernel-inner
221 (+1%)
71,744
26,164
public-kernel-app-logic
122 (+2%)
47,695
40,661
public-kernel-tail
165 (+1%)
53,372
13,269
merge-rollup
10.1
2,568
861
public-kernel-teardown
119 (+1%)
47,695
40,661
public-kernel-setup
118 (+1%)
47,695
40,661
Tree insertion stats
The duration to insert a fixed batch of leaves into each tree type.
Metric
1 leaves
16 leaves
64 leaves
128 leaves
512 leaves
1024 leaves
2048 leaves
4096 leaves
32 leaves
batch_insert_into_append_only_tree_16_depth_ms
10.1 (+1%)
16.1
N/A
N/A
N/A
N/A
N/A
N/A
N/A
batch_insert_into_append_only_tree_16_depth_hash_count
16.8
31.6
N/A
N/A
N/A
N/A
N/A
N/A
N/A
batch_insert_into_append_only_tree_16_depth_hash_ms
0.587
0.496
N/A
N/A
N/A
N/A
N/A
N/A
N/A
batch_insert_into_append_only_tree_32_depth_ms
N/A
N/A
46.5
72.7
231 (-1%)
448
873
1,759 (+1%)
N/A
batch_insert_into_append_only_tree_32_depth_hash_count
N/A
N/A
96.0
159
543
1,055
2,079
4,127
N/A
batch_insert_into_append_only_tree_32_depth_hash_ms
N/A
N/A
0.476
0.448
0.420 (-2%)
0.418 (+1%)
0.413 (-1%)
0.421 (+1%)
N/A
batch_insert_into_indexed_tree_20_depth_ms
N/A
N/A
54.4
107 (-1%)
339 (+1%)
663
1,309 (-2%)
2,614
N/A
batch_insert_into_indexed_tree_20_depth_hash_count
N/A
N/A
105
207
691
1,363
2,707
5,395
N/A
batch_insert_into_indexed_tree_20_depth_hash_ms
N/A
N/A
0.479
0.482
0.458
0.456
0.452 (-2%)
0.454
N/A
batch_insert_into_indexed_tree_40_depth_ms
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
61.6
batch_insert_into_indexed_tree_40_depth_hash_count
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
109
batch_insert_into_indexed_tree_40_depth_hash_ms
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
0.535
Miscellaneous
Transaction sizes based on how many contract classes are registered in the tx.
Metric
0 registered classes
1 registered classes
tx_size_in_bytes
40,548
501,142
Transaction size based on fee payment method
Metric
native fee payment method
fpc_public fee payment method
fpc_private fee payment method
tx_with_fee_size_in_bytes
905
1,161
1,377
Transaction processing duration by data writes.
Metric
0 new note hashes
1 new note hashes
2 new note hashes
tx_pxe_processing_time_ms
1,751
1,098
5,635 (+2%)
Metric
1 public data writes
2 public data writes
3 public data writes
4 public data writes
5 public data writes
8 public data writes
tx_sequencer_processing_time_ms
587 (+2%)
443
1,081
604 (+1%)
1,768
597 (+1%)
|
2025-04-01T06:36:44.492099
| 2023-02-23T00:05:44
|
1596018812
|
{
"authors": [
"amahab",
"jfomhover",
"thomasp-ms"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:259",
"repo": "Azure-Samples/azure-ml-federated-learning",
"url": "https://github.com/Azure-Samples/azure-ml-federated-learning/issues/276"
}
|
gharchive/issue
|
Example pipeline (i.e HELLOWORLD pipeline) to use local kubernetes (external silo) PVC data for training
Is your feature request related to a problem? Please describe.
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
Don't find the syntax to use local k8s data in external silo for FL training. Followed these steps to create PVC - https://github.com/Azure-Samples/azure-ml-federated-learning/blob/main/docs/tutorials/read-local-data-in-k8s-silo. How do I modify the HELLOWORLD pipeline to use this local PVC data for training?
Describe the solution you'd like
A clear and concise description of what you want to happen.
Provide example pipline config files to use local k8s silo data.
Describe alternatives you've considered
A clear and concise description of any alternative solutions or features you've considered.
Searched for syntax both in this repo and Azure AML K8s repo. Only method for creating and exposing PVC is provided. Could not find yaml syntax for pointing to local mountpath,
Additional context
Add any other context or screenshots about the feature request here.
Hi @amahab !
First of all, thanks for filing an issue. We will address it and provide detailed instructions in the repo on what needs to be done to consume local data.
However, for the time being, please let me provide an answer here to unblock you as fast as possible. The key difference between reading local data vs consuming data from an Azure ML data asset is that for local data one cannot use uri_file or uri_folder input parameter like in this config file for instance. Instead, one needs to use a string input. To implement that, you will need to do 2 things:
have your "read local data" component accept a string input in its spec, and have the component code use this string as needed;
provide the value of the string parameter in the config.yml used to submit the job.
We did have a very basic example for that in the initial version of this PR, but at the last minute we decided to drop it. The good news is, the files are still available in GitHub.
The component spec can be found here; see that string input called local_data_path.
The component code can be found here; see how the input parameter value is being used .
The job config.yml is there, and the associated submission script is over there. They both use the parameter defined in the component spec.
The documentation, at that time, had a section on how to configure and run a test job, that was leveraging the files above.
Hopefully this is enough to unblock you, but if not feel free to ask follow-up questions :)
Hi,
Thanks for the response and pointing me to a version of this repo that has this example. Upon inspection of the submit script, I notice that pre-processed data is written back to the Azure cloud datastore for the silo. Is there example where all data remains local on the pvc (r/w mount) and only trained model parameters are written to the silo Azure datastore?
The benefit of external k8s silo is avoiding costly data movement to the cloud. E.g. A scenarios where multiple edge geo sites are generating data. The ability to train the data locally without data movement to Azure and aggregating only model weights via FL is useful.
Appreciate the follow up :-)
No, we don't have that other example currently @amahab .
@amahab as @thomasp-ms mentioned, we don't have another example like that. Overall, you could apply the same kind of guidance for the preprocessing step, although there might be some complications.
One question: how would you expect the interactions between the preprocessing and training when using local mount points in k8s?
For instance:
do you expect the preprocessing to write the data in a unique path (ex: use a run id as subfolder) in the local, then give this unique path down to the training to locate the preprocessed data?
or do you expect the preprocessing to just write in some hardcoded location?
@jfomhover
Preprocessing can write data to a runid subfolder on local and give that part for training.
The whole use case i'm evaluating is that data is generated at the data center/edge locations. These locations train the data locally without moving data to the cloud. Only FL model weights are sent to Azure Workspace for updating the wights. New data will come into the locations at a later time and model gets re-trained locally with updated model sent from cloud.
|
2025-04-01T06:36:44.493577
| 2024-04-06T14:36:08
|
2229276242
|
{
"authors": [
"pamelafox"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:260",
"repo": "Azure-Samples/azure-postgres-pgvector-python",
"url": "https://github.com/Azure-Samples/azure-postgres-pgvector-python/pull/9"
}
|
gharchive/pull-request
|
Use create_all for index creation
This PR uses the already existing create_all call to create indexes, so no additional create call is needed.
Also added missing index to async example.
@nachoalonsoportillo I think you mentioned the missing index in async, this fixes that.
|
2025-04-01T06:36:44.500968
| 2023-05-09T08:16:15
|
1701583946
|
{
"authors": [
"FunkyDialUpDude",
"IhebGhazala"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:261",
"repo": "Azure-Samples/azure-search-openai-demo",
"url": "https://github.com/Azure-Samples/azure-search-openai-demo/issues/184"
}
|
gharchive/issue
|
ERROR: deployment failed: error deploying infrastructure: deploying to subscription:
Please provide us with the following information:
This issue is for a: (mark with an x)
- [x] bug report -> please search issues before submitting
- [ ] feature request
- [x] documentation issue or request
- [ ] regression (a behavior that used to work and stopped in a new release)
Minimal steps to reproduce
running azd up
Any log messages given by the failure
ERROR: deployment failed: error deploying infrastructure: deploying to subscription:
Deployment Error Details:
Conflict: No available instances to satisfy this request. App Service is attempting to increase capacity. Please retry your request later. If urgent, this can be mitigated by deploying this to a new resource group.
No available instances to satisfy this request. App Service is attempting to increase capacity. Please retry your request later. If urgent, this can be mitigated by deploying this to a new resource group.
Expected/desired behavior
the app service plan is expected to deploy with no error
OS and Version?
Windows 11
Versions
22H2
Mention any other details that might be useful
nothing to mention the error detail provides all the info,also i tried making a new resource group just as it recommended but i got the same error
Thanks! We'll be in touch soon.
I am receiving the same error when trying to deploy to East US. Any update on this issue?
|
2025-04-01T06:36:44.508615
| 2023-06-13T22:18:20
|
1755730242
|
{
"authors": [
"GregLin0420",
"TimoRiikonenIF",
"bhasampa",
"clojosh",
"jjczopek",
"jongio",
"paprocki-r",
"silvio-machado",
"tickx-cegeka"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:262",
"repo": "Azure-Samples/azure-search-openai-demo",
"url": "https://github.com/Azure-Samples/azure-search-openai-demo/issues/307"
}
|
gharchive/issue
|
Installtion Setup
Dear Team,
When I am planning to execute the setup getting the following error:
ERROR: deployment failed: failing invoking action 'provision', error deploying infrastructure: deploying to subscription:
Deployment Error Details:
InvalidTemplateDeployment: The template deployment 'openai' is not valid according to the validation procedure. The tracking id is '6777928d-1567-4630-b286-8f77cbf7cacf'. See inner errors for details.
InsufficientQuota: The specified capacity '120' of account deployment is bigger than available capacity '0' for UsageName 'Tokens Per Minute (thousands) - GPT-35-Turbo'.
TraceID: 09f9bd8b6cdac42e96b5848442b7a646
But here I am trying to use the existing the deployed gpt & chat model in spite of why it is going to create new deployment model?
AZURE_OPENAI_CHATGPT_DEPLOYMENT="davinci"
AZURE_OPENAI_GPT_DEPLOYMENT="chat"
AZURE_OPENAI_RESOURCE_GROUP="XXXX"
AZURE_OPENAI_SERVICE="XXXX"
am I missing anything can you please help on the same
Thanks,
Bhaskar
I am getting the same error and I was able to successfully deploy last Friday.
The same here.
same
same
I have the same issue since this week indeed.
If you need 120 for example, but it is currently taking 120/120 , it still demands another 120 units to be free, although no extra units would be needed.
Temporary fix for us is to go to quotas and set all current deployments to a very low number such that you have enough free space again. You'd have to do this for every deployment, so I am also waiting for a fix.
I have the same issue since this week indeed. If you need 120 for example and it is currently taking 120/120 , it still demands another 120 units to be free, although no extra units would be needed in practice. Temporary fix for us is to go to quotas and set all current deployments to a very low number such that you have enough free space again. You'd have to do this for every deployment, so I am also waiting for a fix.
I'm having trouble changing the quotas, where is it?
Thanks for advise I had updated the quota in model the deployment is going fine.. but eventually we need to wait for fix form MSFT.
so, temporary solution is to go to Azure AI studio, Deployments, and for each Edit deployment, Advanced, set "Tokens per Minute Rate Limit" from 120 to e.g. 2. Then deploy your accelerator
This is very inconvenient - having to change quotas manually, especially if the infra is deployed frequently. I hope it's gonna get fixed soon.
How do you decrease the quota for the service you are deploying?
Let's presume that you are using 120/120.
You decrease it to 1/120.
Then it tries to deploy another 120 and complains that it is more than the free quota which is 119.
This https://github.com/Azure-Samples/azure-search-openai-demo/commit/f5a7095e44bd45e224c3200874be36e93e4e0579 should help. Can you either get that patch into your current code or start from scratch and try again?
|
2025-04-01T06:36:44.512839
| 2024-06-27T15:34:01
|
2378483692
|
{
"authors": [
"croblesm",
"seantleonard"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:263",
"repo": "Azure-Samples/azure-sql-library-app-blazor",
"url": "https://github.com/Azure-Samples/azure-sql-library-app-blazor/issues/12"
}
|
gharchive/issue
|
Azure SQL -> Should be SQL Server, there's no Azure SQL used in this sample project.
Instances of Azure SQL mentioned in README and sample title should reflect that this sample uses SQL Server 2022 (Linux). I'm missing where Azure SQL is used. I supposed you'd need to update the Readme to indicate that you could publish the dacpac to Azure SQL if desired. Then document where and how to update the DB connection string in the sample to point to Azure SQL and not localhost.
Thanks for the feedback! This issue was partially fixed on this commit: 06ff3078ffe49284f0d95fdcc66fcd6b20f5e994
The SQL Database project is created for an Azure SQL Database target, thus the database schema is Azure SQL Database compatible. I added a note to explain this situation here.
|
2025-04-01T06:36:44.514372
| 2024-05-16T12:40:38
|
2300299845
|
{
"authors": [
"MoonGoon72",
"yulin-li"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:264",
"repo": "Azure-Samples/cognitive-services-speech-sdk",
"url": "https://github.com/Azure-Samples/cognitive-services-speech-sdk/pull/2378"
}
|
gharchive/pull-request
|
Remove unnecessary semicolon
Purpose
This remove unnecessary semicolon in Swift.
Pull Request Type
What kind of change does this Pull Request introduce?
[ ] Bugfix
[ ] Feature
[x] Code style update (formatting, local variables)
[ ] Refactoring (no functional changes, no api changes)
[ ] Documentation content changes
[ ] Other... Please describe:
/azp run
|
2025-04-01T06:36:44.522400
| 2023-08-07T13:06:36
|
1839422552
|
{
"authors": [
"Tanmai2002",
"anfibiacreativa",
"juliamuiruri4",
"mzbhewtee"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:265",
"repo": "Azure-Samples/contoso-real-estate",
"url": "https://github.com/Azure-Samples/contoso-real-estate/issues/264"
}
|
gharchive/issue
|
HackTogether: Fix feature listing for registered users
Are you accessing the project from the GitHub Codespaces?
[ ] No, I am running the project on my local machine (I will fill in the OS section below)
[x] Yes, I am using GitHub Codespaces.
Note: The project is optimized to run on Codespaces. We welcome contributions to enhance it for cross-platform local development, but it's not something we will prioritize right now. Thank you!
Describe the bug
The bookmark icon is too close to the image.
The Property description is overlapping the bookmark icon.
The short description for the listed property could be removed in order to create a minimalistic design. This description could be assessed when the user wants to know more about the property.
To Reproduce
Steps to reproduce the behavior
Go to home page after login
Expected behavior
I expect the bookmark icon to be properly aligned/placed
The property name/description should not overlap the bookmark icon
A more aesthetic property listing
Screenshots
OS (please complete the following information):
OS:
OS Version:
npm version:
Node.js version:
output of azd version:
Additional context
Having the long description removed from the featured listing will let each card have equal height.
Hi @mzbhewtee Thanks for reporting! We suggest you investigate this issue as related to Angular material design. Thank you for participating. We look forward to your PR!
Hi @anfibiacreativa Thank you, i will investigate the issue.
Suggestion to tackle with issue #242
Removing long description is not ideal recommendation. we can properly customize scss for having responsive cards with appropriate dimensions.
I am also willing to contribute and investigate together @mzbhewtee
Do we have any updates? Thank you!
|
2025-04-01T06:36:44.525405
| 2024-03-11T13:57:44
|
2179206035
|
{
"authors": [
"juliamuiruri4",
"manekinekko"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:266",
"repo": "Azure-Samples/contoso-real-estate",
"url": "https://github.com/Azure-Samples/contoso-real-estate/pull/452"
}
|
gharchive/pull-request
|
fix: updated command for package installs for the notification server (ReadMe)
Issue #451
Readme instruction
Current output
Target output
Hi @juliamuiruri4 your change LGTM. Can you change your commit message to follow these guidelines.
Done @manekinekko
Let me know if I should change anything else. Thanks
|
2025-04-01T06:36:44.529326
| 2021-05-11T16:50:40
|
887863890
|
{
"authors": [
"derisen",
"kalyankrishna1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:267",
"repo": "Azure-Samples/ms-identity-javascript-react-tutorial",
"url": "https://github.com/Azure-Samples/ms-identity-javascript-react-tutorial/pull/10"
}
|
gharchive/pull-request
|
Adding pop tokens sample
Purpose
Adds a new sample where React SPA acquiring PoP tokens to call Node.js web API which validates the token using jose
cc @hectormmg
Does this introduce a breaking change?
[ ] Yes
[ x ] No
Pull Request Type
What kind of change does this Pull Request introduce?
[ ] Bugfix
[ ] Feature
[ ] Code style update (formatting, local variables)
[ ] Refactoring (no functional changes, no api changes)
[ ] Documentation content changes
[ x ] Other... Please describe: **new sample**
How to Test
Get the code
git clone https://github.com/Azure-Samples/ms-identity-javascript-react-tutorial.git
cd ms-identity-javascript-react-tutorial
git checkout pop-sample
cd 6-AdvancedScenarios/2-call-api-pop/API
npm install
cd ../SPA
npm install
What to Check
Verify that the following are valid
Test in Chrome, Edge, Firefox
Test in private mode
"dependencies": {
delete this file?
Refers to: 3-Authorization-II/1-call-api/SPA/package-lock.json:6 in 3e8d6a7. [](commit_id = 3e8d6a7d8c3a8f054648fc291b1fd33a202924dc, deletion_comment = False)
|
2025-04-01T06:36:44.536725
| 2022-12-07T15:58:06
|
1482257365
|
{
"authors": [
"Gordonby",
"MattLeach25",
"romerve"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:268",
"repo": "Azure/AKS-Construction",
"url": "https://github.com/Azure/AKS-Construction/issues/482"
}
|
gharchive/issue
|
Deployment fails when using Custom Networking and when using custom Address space
Describe the bug
When using AKS Construction helper, the deployment fails with the following error:
{"status":"Failed","error":{"code":"DeploymentFailed","message":"At least one resource deployment operation failed. Please list deployment operations for details. Please see https://aka.ms/DeployOperations for usage details.","details":[{"code":"Conflict","message":"{\r\n \"status\": \"Failed\",\r\n \"error\": {\r\n \"code\": \"ResourceDeploymentFailure\",\r\n \"message\": \"The resource operation completed with terminal provisioning state 'Failed'.\",\r\n \"details\": [\r\n {\r\n \"code\": \"DeploymentFailed\",\r\n \"message\": \"At least one resource deployment operation failed. Please list deployment operations for details. Please see https://aka.ms/DeployOperations for usage details.\",\r\n \"details\": [\r\n {\r\n \"code\": \"BadRequest\",\r\n \"message\": \"{\\r\\n \\\"error\\\": {\\r\\n \\\"code\\\": \\\"NetcfgInvalidSubnet\\\",\\r\\n \\\"message\\\": \\\"Subnet 'AzureFirewallManagementSubnet' is not valid in virtual network 'vnet-aks-rv-test'.\\\",\\r\\n \\\"details\\\": []\\r\\n }\\r\\n}\"\r\n }\r\n ]\r\n }\r\n ]\r\n }\r\n}"}]}}
It appears that even when no Azure Firewall is requested in the helper, the helper attempts to create an AzureFirewallManagementSubnet
To Reproduce
Steps to reproduce the behavior:
Go to Networking Details tab
Under Network plugin, select "Use CNI..."
Under Default or Custom VNET, select Custom Networking
VNET Address space, anything, but the default space. EX: <IP_ADDRESS>/16
AKS Nodes Subnet, anything within the range.
I think this could have been fixed in #480 and just needs released.
@MattLeach25 - thoughts?
I think this could have been fixed in #480 and just needs released. @MattLeach25 - thoughts?
Yeah this is fixed with the new release.
We'll aim to cut a release tomorrow.
@romerve - if you clone the repo and target bicep/main.bicep instead of the release - then you should be good.
@MattLeach25 created a release today, so this should be fixed.
|
2025-04-01T06:36:44.539255
| 2022-06-07T07:29:06
|
1262848073
|
{
"authors": [
"Gordonby",
"khowling"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:269",
"repo": "Azure/AKS-Construction",
"url": "https://github.com/Azure/AKS-Construction/pull/302"
}
|
gharchive/pull-request
|
Setting codespaces memory requirement
PR Summary
Closes #289
PR Checklist
[x] PR has a meaningful title
[x] Summarized changes
[x] This PR is ready to merge and is not Work in Progress
[x] Link to a filed issue
approve, but no way react dev takes 8GB, only takes 500MB locally
|
2025-04-01T06:36:44.548504
| 2023-03-13T15:55:37
|
1621778907
|
{
"authors": [
"JamJarchitect"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:270",
"repo": "Azure/ALZ-Bicep",
"url": "https://github.com/Azure/ALZ-Bicep/pull/476"
}
|
gharchive/pull-request
|
26808 - Feature - Add support to allow Private DNS Zone VNet Linking for Spoke VNets
Overview/Summary
Adding support to allow Private DNS Zone VNet Linking for Spoke VNets through HubPeeredSpoke orchestration module and a new Private Dns Zone Linking module.
AB#26808
This PR fixes/adds/changes/removes
Added feature to HubPeeredSpoke orchestration module to allow Private DNS Zone Linking to Spoke VNets.
Added new module for Private DNS Zone Linking.
Breaking Changes
N/A
Testing Evidence
Evidence to be posted.
As part of this Pull Request I have
[X] Read the Contribution Guide and ensured this PR is compliant with the guide
[X] Ensured the resource API versions in .bicep file/s I am adding/editing are using the latest API version possible
[X] Checked for duplicate Pull Requests
[X] Associated it with relevant GitHub Issues
[X] (ALZ Bicep Core Team Only) Associated it with relevant ADO Items
[X] Ensured my code/branch is up-to-date with the latest changes in the main branch
[X] Performed testing and provided evidence.
[X] Updated one or more of the following tests (if required)
Unit
Linting
E2E (End-To-End)
ValidateAzCloud (Base validation in Azure Cloud)
ValidateMcCloud (Base validation in Azure China Cloud)
[X] Updated relevant and associated documentation (e.g. Contribution Guide, Module READMEs, Wiki Docs etc.)
[X] If relevant, created or updated Code Tours here
New module working - orchestration module test to be done
/azp run validateazcloud
/azp run e2e
/azp run validateazcloud
|
2025-04-01T06:36:44.552595
| 2022-04-29T17:34:44
|
1221331356
|
{
"authors": [
"cadenmarchese",
"rogbas"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:271",
"repo": "Azure/ARO-RP",
"url": "https://github.com/Azure/ARO-RP/pull/2096"
}
|
gharchive/pull-request
|
Add NRMS/Hypernet tags to cluster resource group in CI mode
Which issue this PR addresses:
The RP fails cluster installs if NSGs are attached to subnets. Simply Secure v1 applies NSGs in our subscriptions when these tags aren't present, resulting in a race condition.
What this PR does / why we need it:
Adds the tags needed to stop SSv1 from applying NSGs to subnets: https://eng.ms/docs/security-compliance-identity-and-management-scim/security/azure-security/security-health-analytics/network-isolation/netiso-teamdocs/netiso-program-overview/manage/hypernet-and-nrms-simply-secure-v1-network-security-rules
Test plan for issue:
Deploy this commit to central us euap (where the issue is faced)
Is there any documentation that needs to be updated for this PR?
No
/approve
hold for now as tagging the resource group didn't have the desired effect. We'll need to update these tags with a temporary SSv1 exception number and re-test.
@ross-bryan we can open a new PR with exception tags later if needed. Closing this one.
|
2025-04-01T06:36:44.555306
| 2023-03-08T20:00:04
|
1615840762
|
{
"authors": [
"cadenmarchese"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:272",
"repo": "Azure/ARO-RP",
"url": "https://github.com/Azure/ARO-RP/pull/2766"
}
|
gharchive/pull-request
|
Fix Purge MSFT CI Subscription
Which issue this PR addresses:
Fixes
What this PR does / why we need it:
Test plan for issue:
Is there any documentation that needs to be updated for this PR?
Also fixes Purge Red Hat Dev-Gratis: https://msazure.visualstudio.com/AzureRedHatOpenShift/_build/results?buildId=69797295&view=logs&j=a4f1910f-c367-5697-edcd-724d81cde23b&t=2152b647-0197-5277-e95c-5d58fca9c7d3
|
2025-04-01T06:36:44.558564
| 2023-10-02T20:37:16
|
1922661909
|
{
"authors": [
"bennerv",
"niontive"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:273",
"repo": "Azure/ARO-RP",
"url": "https://github.com/Azure/ARO-RP/pull/3197"
}
|
gharchive/pull-request
|
Set maintenace task to "" after PUCM pending operation
Which issue this PR addresses:
Fixes bug introduced by https://github.com/Azure/ARO-RP/pull/3194
What this PR does / why we need it:
We never unset PUCM pending maintenance task
If we do admin update with no maintenance task set, we're suppose to do a complete admin update, but instead we get PUCM pending task completed again
Test plan for issue:
Unit
INT
Is there any documentation that needs to be updated for this PR?
No - bug fix
/azp run e2e
|
2025-04-01T06:36:44.571438
| 2023-04-04T22:40:39
|
1654669573
|
{
"authors": [
"AlisonB319",
"coveralls"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:274",
"repo": "Azure/AgentBaker",
"url": "https://github.com/Azure/AgentBaker/pull/2976"
}
|
gharchive/pull-request
|
refactor: lint errors; strings, undefinied nums, global vars
What type of PR is this?
What this PR does / why we need it:
Which issue(s) this PR fixes:
Fixes #
Requirements:
[ ] uses conventional commit messages
[ ] includes documentation
[ ] adds unit tests
[ ] tested upgrade from previous version
Special notes for your reviewer:
Release note:
none
Pull Request Test Coverage Report for Build<PHONE_NUMBER>
44 of 126 (34.92%) changed or added relevant lines in 8 files are covered.
2 unchanged lines in 2 files lost coverage.
Overall coverage decreased (-2.3%) to 77.079%
Changes Missing Coverage
Covered Lines
Changed/Added Lines
%
pkg/agent/utils.go
9
13
69.23%
pkg/agent/datamodel/types.go
12
17
70.59%
pkg/agent/datamodel/sig_config.go
0
73
0.0%
Files with Coverage Reduction
New Missed Lines
%
pkg/agent/datamodel/sig_config.go
1
42.29%
pkg/agent/datamodel/types.go
1
80.43%
Totals
Change from base Build<PHONE_NUMBER>:
-2.3%
Covered Lines:
2011
Relevant Lines:
2609
π - Coveralls
|
2025-04-01T06:36:44.579595
| 2024-03-18T05:31:51
|
2191269333
|
{
"authors": [
"coveralls",
"junjiezhang1997"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:275",
"repo": "Azure/AgentBaker",
"url": "https://github.com/Azure/AgentBaker/pull/4183"
}
|
gharchive/pull-request
|
docs: add Windows image release notes for 2024-3b
What type of PR is this?
/kind documentation
What this PR does / why we need it:
add Windows image release notes for 2024-3b. Reference: #4175
Which issue(s) this PR fixes:
Fixes #
Requirements:
[x] uses conventional commit messages
[ ] includes documentation
[ ] adds unit tests
[ ] tested upgrade from previous version
Special notes for your reviewer:
Release note:
none
Pull Request Test Coverage Report for Build<PHONE_NUMBER>
Details
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage remained the same at 79.343%
Totals
Change from base Build<PHONE_NUMBER>:
0.0%
Covered Lines:
2293
Relevant Lines:
2890
π - Coveralls
|
2025-04-01T06:36:44.584928
| 2019-09-19T20:07:16
|
496001586
|
{
"authors": [
"jimmyca15",
"jmsalvo",
"nhart12",
"zhenlan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:276",
"repo": "Azure/AppConfiguration",
"url": "https://github.com/Azure/AppConfiguration/issues/142"
}
|
gharchive/issue
|
Event grid subscriptions no longer working
I had created some EventGrid subscriptions to a preview app configuration instance in East US. Today I can no longer see the subscription and receive the following error when clicking on Events
Are event grid subscriptions really no longer available in East US or is this a UI bug?
@nhart12 This is a UI bug, working on resolving it. Thanks for notifying us!
This has been resolved.
Thanks @jimmyca15 the UI seems to work now, the events however are not firing consistently in our app configuration instance
@nhart12. Is this happening in a single region with a single store? Do you mean events are firing sometimes, and then not firing other times or that they're not firing at all?
@jimmyca15 So this is happening in our East US store. Just tried again this morning, created a couple settings, updated them several times and then deleted. I have an EventGrid subscription setup for key-value Modified and Deleted. Only the deleted events fired
There was a bug in event grid publishing identified last week. We are in the process of rolling out our fix for that. That may be causing your issue. When the fix is rolled out I will let you know.
Ok perfect thanks for the info @jimmyca15 !
@jimmyca15 Can you provide an update regarding when you anticipate the event grid publishing issues may be resolved? Also, is a GA estimate available...the most recent discussion around this mentioned October timeframe, is that still a possibility?
@jmsalvo event grid publishing changes are getting rolled out region by region. Expected to all regions by thursday eod.
Estimated GA is still end of October. Thanks.
Yesterday we deployed a fix to event grid to improve some event publishing reliability issues we were facing. Please open another issue if you are still experiencing problems.
Hi @nhart12 and @jmsalvo, I want to give you an update. We just posted our GA and pricing plan. Please check it out.
https://github.com/Azure/AppConfiguration-Announcements/issues/2
|
2025-04-01T06:36:44.596575
| 2023-06-15T18:26:48
|
1759342230
|
{
"authors": [
"AndrewCS149",
"mbrat2005"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:277",
"repo": "Azure/AzLoadBalancerMigration",
"url": "https://github.com/Azure/AzLoadBalancerMigration/issues/65"
}
|
gharchive/issue
|
πͺ² Bug Report - Upgrade Fails for SF VMSS with KeyVault Extension
Describe the bug
Upgrade fails due to KV VM extension timeout:
2023-06-14T21:07:57+00 [Information]:############################## Initializing Start-AzBasicLoadBalancerUpgrade ##############################
2023-06-14T21:07:57+00 [Information]:[Start-AzBasicLoadBalancerUpgrade] PowerShell Version: **7.3.4**
2023-06-14T21:07:57+00 [Information]:[Start-AzBasicLoadBalancerUpgrade] AzureBasicLoadBalancerUpgrade **Version: 2.0.19**
...
2023-06-14T21:08:00+00 [Information]:[Test-SupportedMigrationScenario] Checking whether VMSS scale set 'quotavmssdevbn' is a Service Fabric cluster...
WARNING: 2023-06-14T21:08:00+00 [Warning]:[Test-SupportedMigrationScenario] **VMSS appears to be a Service Fabric** cluster based on extension profile. SF Clusters experienced potentically significant downtime during migration using this PowerShell module. In testing, a 5-node Bronze cluster was unavailable for about 30 minutes and a 5-node Silver cluster was unavailabile for about 45 minutes. Shutting down the cluster VMSS prior to initiating migration will result in a more consistent experience of about 5 minutes to complete the LB migration. For Service Fabric clusters that require minimal / no connectivity downtime, adding a new nodetype with standard load balancer and IP resources is a better solution.
Do you want to proceed with the migration of your Service Fabric Cluster's Load Balancer?
...
2023-06-14T21:08:00+00 [Information]:[PublicLBMigration] **Public Load Balancer Detected**. Initiating Public Load Balancer Migration
...
2023-06-14T21:24:34+00 [Information]:[NatRulesMigration] Waiting for saving standard load balancer LB-quota-cluster-dev-bn job to complete...
2023-06-14T21:24:34+00 [Information]:[NatRulesMigration] Nat Rules Migration Completed
2023-06-14T21:24:34+00 [Information]:[InboundNatPoolsMigration] Initiating Inbound NAT Pools Migration
2023-06-14T21:24:34+00 [Information]:[InboundNatPoolsMigration] Adding Inbound NAT Pool LoadBalancerBEAddressNatPool to Standard Load Balancer
2023-06-14T21:24:34+00 [Information]:[InboundNatPoolsMigration] Saving Standard Load Balancer LB-quota-cluster-dev-bn
2023-06-14T21:24:49+00 [Information]:[InboundNatPoolsMigration] Waiting for saving standard load balancer LB-quota-cluster-dev-bn job to complete...
2023-06-14T21:24:49+00 [Information]:[GetVmssFromBasicLoadBalancer] Initiating GetVmssFromBasicLoadBalancer
2023-06-14T21:24:49+00 [Information]:[GetVmssFromBasicLoadBalancer] Getting VMSS object '/subscriptions/.../resourcegroups/azure-quota-dev-eastus2/providers/microsoft.compute/virtualmachinescalesets/quotavmssdevbn' from Azure
2023-06-14T21:24:49+00 [Information]:[GetVmssFromBasicLoadBalancer] VMSS loaded Name quotavmssdevbn from RG azure-quota-dev-eastus2
2023-06-14T21:24:49+00 [Information]:[_MigrateNetworkInterfaceConfigurations] Adding InboundNATPool to VMSS quotavmssdevbn
2023-06-14T21:24:49+00 [Information]:[_MigrateNetworkInterfaceConfigurations] Checking if VMSS 'quotavmssdevbn' NIC 'NIC-azure-quota-dev-eastus2' IPConfig 'NIC-azure-quota-dev-eastus2' should be associated with NAT Pool 'LoadBalancerBEAddressNatPool'
2023-06-14T21:24:49+00 [Information]:[_MigrateNetworkInterfaceConfigurations] Adding NAT Pool 'LoadBalancerBEAddressNatPool' to IPConfig 'NIC-azure-quota-dev-eastus2'
2023-06-14T21:24:49+00 [Information]:[_MigrateNetworkInterfaceConfigurations] Migrate NetworkInterface Configurations completed
2023-06-14T21:24:49+00 [Information]:[InboundNatPoolsMigration] Saving VMSS quotavmssdevbn
2023-06-14T21:24:49+00 [Information]:[UpdateVmss] Updating configuration of VMSS 'quotavmssdevbn'
2023-06-14T21:25:04+00 [Information]:[UpdateVmss] Waiting for job (id: '5') updating VMSS 'quotavmssdevbn' to complete...
...
2023-06-14T23:10:50+00 [Information]:[UpdateVmss] Waiting for job (id: '5') updating VMSS 'quotavmssdevbn' to complete...
InvalidOperation: Long running operation failed with status 'Failed'. Additional Info:'Provisioning of VM extension **KvVmExtension** has timed out. Extension provisioning has taken too long to complete. The extension did not report a message. More information on troubleshooting is available at https://aka.ms/vmextensionwindowstroubleshoot'
ErrorCode: VMExtensionProvisioningTimeout
ErrorMessage: Provisioning of VM extension KvVmExtension has timed out. Extension provisioning has taken too long to complete. The extension did not report a message. More information on troubleshooting is available at https://aka.ms/vmextensionwindowstroubleshoot
ErrorTarget: 0
StartTime: 6/14/2023 9:24:52 PM
EndTime: 6/14/2023 11:10:27 PM
OperationID: 85ee53b5-9ce3-4458-9edd-f46e8c7baf02
Status: Failed
Write-Error: 2023-06-14T23:10:50+00 [Error]:[InboundNatPoolsMigration] An error occured when attempting to update VMSS network config on the new Standard LB backend pool membership. To recover address
the following error, and try again specifying the -FailedMigrationRetryFilePath parameter and Basic Load Balancer backup State file located either in this directory or the directory
specified with -RecoveryBackupPath
To Reproduce
Steps to reproduce the behavior:
VMSS
Public LB
KVVMExtension [this case, extension adds a cert to local store, auto upgrade disabled]
SF Cluster [?]
Additional context - please include:
See log
Trying to test this scenario, I can't get the LB migration to even start due to:
InvalidOperation: Long running operation failed with status 'Failed'. Additional Info:'VM has reported a failure when processing extension 'vmss-01'. Error message: "Failed to parse the configuration settings with: 'not a string'"
More information on troubleshooting is available at https://aka.ms/vmextensionwindowstroubleshoot '
ErrorCode: VMExtensionProvisioningError
ErrorMessage: VM has reported a failure when processing extension 'vmss-01'. Error message: "Failed to parse the configuration settings with: 'not a string'"
This issue is reportedly intermittent...still working to repro
Closing due to lack of activity and reproducibility
@mbrat2005 Im experiencing the same issue. Did you ever find a solution?
Hi Andrew,
I haven't made progress on this one, since I couldn't seem to repro it. Would you be able to share your upgrade log for details? Also, are you upgrading a basic LB for a Service Fabric Cluster?
Thanks!
Matthew
From: Andrew Smith @.>
Sent: Wednesday, July 3, 2024 13:29
To: Azure/AzLoadBalancerMigration @.>
Cc: Mention @.>; Author @.>; Comment @.>; Assign @.>; State change @.***>
Subject: Re: [Azure/AzLoadBalancerMigration] πͺ² Bug Report - Upgrade Fails for SF VMSS with KeyVault Extension (Issue #65)
@mbrat2005https://github.com/mbrat2005 Im experiencing the same issue. Did you ever find a solution?
β
Reply to this email directly, view it on GitHubhttps://github.com/Azure/AzLoadBalancerMigration/issues/65#issuecomment-2207038571 or unsubscribehttps://github.com/notifications/unsubscribe-auth/AGBW6WHTSMOHXES4YPWH4F3ZKRGLFBFKMF2HI4TJMJ2XIZLTSWBKK5TBNR2WLJDUOJ2WLJDOMFWWLO3UNBZGKYLEL5YGC4TUNFRWS4DBNZ2F6YLDORUXM2LUPGBKK5TBNR2WLJDUOJ2WLJDOMFWWLLTXMF2GG2C7MFRXI2LWNF2HTAVFOZQWY5LFUVUXG43VMWSG4YLNMWVXI2DSMVQWIX3UPFYGLAVFOZQWY5LFVI2DKNRTHEYTONZQHGSG4YLNMWUWQYLTL5WGCYTFNSBKK5TBNR2WLKRVGU2DSOJZGYZDQM5ENZQW2ZNJNBQXGX3MMFRGK3FMON2WE2TFMN2F65DZOBS2YSLTON2WKQ3PNVWWK3TUUZ2G64DJMNZZJAVEOR4XAZNKOJSXA33TNF2G64TZUV3GC3DVMWUTKMZZHE4TSOJYG6BKI5DZOBS2K2LTON2WLJLWMFWHKZNKGE3TKOJTGQZDEMZQQKSHI6LQMWSWYYLCMVWKK5TBNR2WLKRUGU3DGOJRG43TAOMCUR2HS4DFUVWGCYTFNSSXMYLMOVS2UNJVGQ4TSOJWGI4DHJ3UOJUWOZ3FOKTGG4TFMF2GK.
You are receiving this email because you were mentioned.
Triage notifications on the go with GitHub Mobile for iOShttps://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675 or Androidhttps://play.google.com/store/apps/details?id=com.github.android&referrer=utm_campaign%3Dnotification-email%26utm_medium%3Demail%26utm_source%3Dgithub.
@AndrewCS149 I haven't made progress on this one, since I couldn't seem to repro it. Would you be able to share your upgrade log for details? Also, are you upgrading a basic LB for a Service Fabric Cluster?
|
2025-04-01T06:36:44.609120
| 2024-03-27T10:10:15
|
2210379760
|
{
"authors": [
"jayeshprajapaticrest",
"v-rusraut"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:278",
"repo": "Azure/Azure-Sentinel",
"url": "https://github.com/Azure/Azure-Sentinel/issues/10208"
}
|
gharchive/issue
|
Error while clicking Manage button from the already installed BitSight solution under the Content Hub
Installed the BitSight solution from the Content Hub from the azure portal. After successfully installed when I go to manage it by clicking on the Manage button from the installed solution page from the Content Hub getting the "not found" error as below:
To Reproduce
Steps to reproduce the behavior:
Login to Azure Portal
Go to Microsoft Sentinel Workspace
Go to the Content Hub from the side menu and search for the BitSight solution
Install that solution by providing the required information of Subscription, Resource Group etc...
Again go to the BitSight solution from the content hub and click on the Manage button
Expected behavior
When clicking on the Manage button from the BitSight solution under the content hub it should display BitSight Solution Manage screen which all the available component of that solution.
Desktop (please complete the following information):
OS: Windows
Browser: chrome
Version: 122.0.6261.129 (Official Build) (64-bit)
Hi @jayeshprajapaticrest,
Thanks for flagging this issue, we will investigate this issue and get back to you with some updates by 03 Apr 2024. Thanks!
@v-rusraut
Understood. Thanks
Hi @jayeshprajapaticrest,
Please check with below link , if you still facing issue,please let us know.
https://ms.portal.azure.com/
Thanks
@v-rusraut
Thanks
Still getting the same error.
@v-rusraut
I checked today and the issue get resolved.
Thanks
Just to know Can you please let me know the cause of this issue?
|
2025-04-01T06:36:44.621382
| 2022-11-04T04:01:59
|
1435515130
|
{
"authors": [
"theomsoc",
"v-dvedak",
"v-spadarthi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:279",
"repo": "Azure/Azure-Sentinel",
"url": "https://github.com/Azure/Azure-Sentinel/pull/6556"
}
|
gharchive/pull-request
|
Create Theom Solution
Required items, please complete
Change(s):
Adding Solution for Theom
Added 1 Workbook, 20 Analytics Rules
Reason for Change(s):
New
Version Updated:
Yes
Testing Completed:
Yes
Checked that the validations are passing and have addressed any issues that are present:
Yes
@theomsoc :
Logo is looking good
Workbook is looking good
Could you please create a solution package by using V2 tool
Please follow the below URL and do the package creation. Thanks!
Azure-Sentinel/README.md at master Β· Azure/Azure-Sentinel (github.com)
@v-spadarthi should I use my branch as input for Logo and Basepath in input file at https://github.com/Azure/Azure-Sentinel/blob/master/Tools/Create-Azure-Sentinel-Solution/V2/README.md#create-input-file
@v-spadarthi added solution package, please review
@v-spadarthi thanks for the review. Addressed comments and rebuilt package. Shortcut names and atypical are as expected, to match the wording on our application.
@theomsoc : Data connector is looking fine
CreateUI definition
Main Template
Arm-ttk also running fine
@aprakash13 : Please review the analytics and provide your signoff. Thanks!
Once Analytics rules reviewed by @aprakash13 we are going to approve this PR.
@aprakash13 addressed comments and re-published new package
@aprakash13 @v-spadarthi could you please review and merge PR soon, as we are waiting on publishing solution.
Thanks @aprakash13 !! @v-spadarthi please do the needful.
@theomsoc : Please add the solution input file for that create Data folder under Solutions/Theom/ and keep it here.
@theomsoc : Thanks for adding solution input file.
Plz remove the empty properties from solution input file like parsers and playbooks
@theomsoc : Please add the solution input file for that create Data folder under Solutions/Theom/ and keep it here.
@v-spadarthi Added input file, resolved conflicts and removed empty properties.
If all looks good, please merge PR, thanks!
@v-spadarthi could you please check and merge PR soon, so we are able to publish today ?
Create UI definition is looking good
Arm ttk also running fine
All are looking good
Approved and merged since Suresh's has approved the PR
|
2025-04-01T06:36:44.627214
| 2017-03-22T17:49:38
|
216154443
|
{
"authors": [
"larisaborodina",
"msftclas"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:280",
"repo": "Azure/AzureStack-Tools",
"url": "https://github.com/Azure/AzureStack-Tools/pull/122"
}
|
gharchive/pull-request
|
Add-GalleryItems from local folder; Get-GalleryItem; Remove-GalleryItβ¦
Add-GalleryItems from local folder; Get-GalleryItem; Remove-GalleryItem in 1 script
@larisaborodina,
Thanks for your contribution as a Microsoft full-time employee or intern. You do not need to sign a CLA.
Thanks,
Microsoft Pull Request Bot
|
2025-04-01T06:36:44.641187
| 2024-07-29T14:30:04
|
2435588234
|
{
"authors": [
"marteltran",
"santhoshb-msft"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:281",
"repo": "Azure/Commercial-Marketplace-SaaS-Accelerator",
"url": "https://github.com/Azure/Commercial-Marketplace-SaaS-Accelerator/issues/740"
}
|
gharchive/issue
|
Update script ERROR
Describe the bug
Run the update script to update existing SaaS-Accelerator in Azure, but get error
To Reproduce
Steps to reproduce the behavior:
Fork the SaaS-Accelerator code with latest release 7.6.3
Logon to Azure Portal > Open Cloud Shell and run below update-script
wget https://dotnet.microsoft.com/download/dotnet/scripts/v1/dotnet-install.sh; chmod +x dotnet-install.sh;
./dotnet-install.sh -version 6.0.423; $ENV:PATH="$HOME/.dotnet:$ENV:PATH";
dotnet tool install --global dotnet-ef --version 6.0.1; git clone https://github.com/xxxx/Commercial-Marketplace-SaaS-Accelerator.git; cd ./Commercial-Marketplace-SaaS-Accelerator/deployment;
.\Upgrade.ps1 -WebAppNamePrefix "xxxxx"
-ResourceGroupForDeployment "xxxx"
See the errors:
Line |
32 | Param(
| ~~~~~
| The term 'Param' is not recognized as a name of a cmdlet, function, script file, or executable program. Check the spelling of the name, or if a path was
| included, verify that the path is correct and try again.
Deploying new database
ERROR: argument --vault-name: expected one argument
String-Between: /home/martel/Commercial-Marketplace-SaaS-Accelerator/deployment/Upgrade.ps1:64
Line |
64 | $Server = String-Between -source $ConnectionString -start "Data Sourc β¦
| ~~~~~~~~~~~~~~~~~
| Cannot bind argument to parameter 'Source' because it is an empty string.
Expected behavior
Expect to see successful message
Environment (please complete the following information):
Are you using the CloudShell? YES
If not using the CloudShell:
OS: [e.g. iOS] Window
PowerShell version
Modules installed: Pre installed SaaS-Accelerator 6.5
PowerShell Version Used: 7.4.3
The latest PR should have fixed this. @code4clouds can you see if itβs still an issue please
I got the fix and rerun the Upgrade.ps1 script, but get error with MSI:
Deploying new database
ERROR: Failed to connect to MSI. Please make sure MSI is configured correctly.
Get Token request returned: <Response [400]>
Upgrade.ps1:
Line |
8 | .\Upgrade.ps1 `
| ~~~~~~~~~~~~~~~
| Cannot bind argument to parameter 'Source' because it is an empty string.
@marteltran if you have already installed latest any reason why you are upgrading?
But if you are anyway, please make sure to remove vnet protection for sql, keyvault and app services during the upgrade and add it back again after the upgrade
I have installed the old version (6.5.x) so I want to upgrade to the latest (7.4.3).
Just confirm:
Do I have to DELETE the key-vault and 2 App services (Admin & Portal) before running the upgrade script? (I do not use VNET).
How to add the key-vault after all?
@marteltran - lets take a step back, but the latest is 7.6.2 not 7.4.3. I dont even see 7.4.3 anywhere am I missing something?
Also there is no comment about deleting anything, so please dont delete anything unless you are aware of what you are doing.
I confuse. When I fork the main branch, What is current version? and How to make sure that my fork contain 7.6.2?
|
2025-04-01T06:36:44.649221
| 2019-04-30T05:02:14
|
476450890
|
{
"authors": [
"jmbrunskill"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:282",
"repo": "Azure/Industrial-IoT",
"url": "https://github.com/Azure/Industrial-IoT/issues/55"
}
|
gharchive/issue
|
'BadEncodingLimitsExceeded ' received when creating subscriptions
We have OPC Publisher 2.3.3 running on iotedge <IP_ADDRESS> (3fa6cbef8b7fc3c55a49a622735eb1021b8a5963)
On startup up of the Publisher we are getting a 'BadEncodingLimitsExceeded ' error.
How should we go about debugging/resolving this issue?
[04:34:04 INF] Publisher is running. Press CTRL-C to quit.
[04:34:04 INF] Connect and monitor session and nodes on endpoint 'opc.tcp://<IP_ADDRESS>:49320'.
[04:34:06 INF] Create unsecured session for endpoint URI 'opc.tcp://<IP_ADDRESS>:49320' with timeout of 10000 ms.
[04:34:06 INF] Session successfully created with Id ns=1;g=73c5b05f-456f-49ec-a62b-a555657aaa6d.
[04:34:06 INF] The session to endpoint 'opc.tcp://<IP_ADDRESS>:49320/' has 3 entries in its namespace array:
[04:34:06 INF] Namespace index 0: http://opcfoundation.org/UA/
[04:34:06 INF] Namespace index 1: urn:NZS070HSTOPC01X.prod01.inroot.net:PTC.ThingWorxIC.1:UA%20Server
[04:34:06 INF] Namespace index 2: ThingWorx Industrial Connectivity
[04:34:06 INF] Created subscription with id 8 on endpoint 'opc.tcp://<IP_ADDRESS>:49320'
[04:34:06 INF] Create subscription on endpoint 'opc.tcp://<IP_ADDRESS>:49320' requested OPC publishing interval is 30000 ms. (revised: 30000 ms)
[04:34:06 INF] Start monitoring items on endpoint 'opc.tcp://<IP_ADDRESS>:49320'. Currently monitoring 0 items.
[04:34:06 INF] Now monitoring 10000 items in subscription with id '8'
[04:34:07 INF] Now monitoring 20000 items in subscription with id '8'
[04:34:08 INF] Now monitoring 30000 items in subscription with id '8'
[04:34:10 INF] Now monitoring 40000 items in subscription with id '8'
[04:34:14 INF] Now monitoring 50000 items in subscription with id '8'
[04:34:14 ERR] Unhandled OPC UA ServiceResultException 'BadEncodingLimitsExceeded 'Could not send request to server.'' when monitoring node 'ns=2;s=NZ070_Darfield_NZS070SRV01X.C270.NZ070_B05VA0506_AlmFault' on endpoint 'opc.tcp://<IP_ADDRESS>:49320'. Continue.
[04:34:14 INF] Now monitoring 60000 items in subscription with id '8'
[04:34:14 ERR] Unhandled OPC UA ServiceResultException 'BadEncodingLimitsExceeded 'Could not send request to server.'' when monitoring node 'ns=2;s=NZ070_Darfield_NZS070SRV01X.C801.NZ070_S1PW1P01_Msg1' on endpoint 'opc.tcp://<IP_ADDRESS>:49320'. Continue.
[04:34:15 ERR] Exception
Opc.Ua.ServiceResultException: Could not send request to server.
at Opc.Ua.Bindings.ChannelAsyncOperation`1.End(Int32 timeout, Boolean throwOnError)
at Opc.Ua.Bindings.UaSCUaBinaryClientChannel.EndSendRequest(IAsyncResult result)
at Opc.Ua.Bindings.UaSCUaBinaryTransportChannel.EndSendRequest(IAsyncResult result)
at Opc.Ua.Bindings.UaSCUaBinaryTransportChannel.SendRequest(IServiceRequest request)
at Opc.Ua.SessionClient.CreateMonitoredItems(RequestHeader requestHeader, UInt32 subscriptionId, TimestampsToReturn timestampsToReturn, MonitoredItemCreateRequestCollection itemsToCreate, MonitoredItemCreateResultCollection& results, DiagnosticInfoCollection& diagnosticInfos)
at Opc.Ua.Client.Subscription.CreateItems()
at Opc.Ua.Client.Subscription.ApplyChanges()
at OpcPublisher.OpcSession.MonitorNodesAsync(CancellationToken ct) in /app/opcpublisher/OpcSession.cs:line 561
[04:34:25 INF] Start monitoring items on endpoint 'opc.tcp://<IP_ADDRESS>:49320'. Currently monitoring 66541 items.
[04:34:25 ERR] Unhandled OPC UA ServiceResultException 'BadEncodingLimitsExceeded 'Could not send request to server.'' when monitoring node 'ns=2;s=NZ070_Darfield_NZS070SRV01X.C270.NZ070_B05VA0506_AlmFault' on endpoint 'opc.tcp://<IP_ADDRESS>:49320'. Continue.
[04:34:25 ERR] Exception
Opc.Ua.ServiceResultException: Could not send request to server.
at Opc.Ua.Bindings.ChannelAsyncOperation`1.End(Int32 timeout, Boolean throwOnError)
at Opc.Ua.Bindings.UaSCUaBinaryClientChannel.EndSendRequest(IAsyncResult result)
at Opc.Ua.Bindings.UaSCUaBinaryTransportChannel.EndSendRequest(IAsyncResult result)
at Opc.Ua.Bindings.UaSCUaBinaryTransportChannel.SendRequest(IServiceRequest request)
at Opc.Ua.SessionClient.CreateMonitoredItems(RequestHeader requestHeader, UInt32 subscriptionId, TimestampsToReturn timestampsToReturn, MonitoredItemCreateRequestCollection itemsToCreate, MonitoredItemCreateResultCollection& results, DiagnosticInfoCollection& diagnosticInfos)
at Opc.Ua.Client.Subscription.CreateItems()
at Opc.Ua.Client.Subscription.ApplyChanges()
at OpcPublisher.OpcSession.MonitorNodesAsync(CancellationToken ct) in /app/opcpublisher/OpcSession.cs:line 561
I have tried removing the tags that are reporting errors but it appears to just report the next tag as error'd after this.
This error is also mentioned on 2 other OPC github projects.
https://github.com/OPCFoundation/UA-.NET-Legacy/issues/107
https://github.com/open62541/open62541/issues/434
Neither is very clear to me why this might be happening, but sounds like it could be something going over the size limit for an OPCUA message?
@marcschier Yes, this use case is working now that we have the correct path.
We are seeing some performance challenges, I think it would be good to consider batching the OPC UA subscriptions if the resulting OPC UA group would result in a really large subscription. Would be interesting to see if this would improve performance or reliability.
|
2025-04-01T06:36:44.658061
| 2024-08-28T22:48:18
|
2493107915
|
{
"authors": [
"DanielleCogs",
"DrLeeroyPhD"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:283",
"repo": "Azure/LogicAppsUX",
"url": "https://github.com/Azure/LogicAppsUX/issues/5544"
}
|
gharchive/issue
|
WYSIWYG Formatting Issues
Describe the Bug with repro steps
Update any WYSIWYG text fields in new Logic App Editor
Save
On reload, line breaks of edited fields tend to have been removed
In my organization, this has been observed especially with Outlook Email "Message" field and Azure DevOps > Create a Work Item "Description" field
What type of Logic App Is this happening in?
Consumption (Portal)
Which operating system are you using?
Windows, Linux
Are you using new designer or old designer
New Designer
Did you refer to the TSG before filing this issue? https://aka.ms/lauxtsg
Yes
Workflow JSON
No response
Screenshots or Videos
Formatting before saving changes
Formatting after saving changes
Browser
Microsoft Edge, Chrome, Firefox
Additional context
This has occurred in 4 of our logic apps that I'm aware of
This has occurred under at least two different users on two different computers
This seems to generally occur on larger logic apps
Only occurs on New Designer, remediated by using Old Designer, which is slated for retirement
Also confirmed issue on Firefox from Fedora Linux
This has been fixed but not deployed, will start rolling out deployment for this on Monday https://github.com/Azure/LogicAppsUX/pull/5392
Right on, thanks a ton for the heads up, I'll look forward to it!
Actually Tuesday! Monday holiday in US
In case anyone finds this, deployment to prod took effect today 9/10/2024
Thanks a ton for all of your help!
|
2025-04-01T06:36:44.659909
| 2017-06-05T22:52:18
|
233733124
|
{
"authors": [
"msftclas",
"suhaildawood"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:284",
"repo": "Azure/acs-demos",
"url": "https://github.com/Azure/acs-demos/pull/26"
}
|
gharchive/pull-request
|
Update Kubernetes Spark Install Demo
Update the Kubernetes + Spark install demo to work with the SimDem update. Adds cloning the Kubernetes GitHub repository to the home directory as part of the preparation as it takes a while.
@suhaildawood,
Thanks for your contribution as a Microsoft full-time employee or intern. You do not need to sign a CLA.
Thanks,
Microsoft Pull Request Bot
|
2025-04-01T06:36:44.666600
| 2024-07-09T13:15:45
|
2398197922
|
{
"authors": [
"weeyin83"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:285",
"repo": "Azure/arc_jumpstart_drops",
"url": "https://github.com/Azure/arc_jumpstart_drops/pull/119"
}
|
gharchive/pull-request
|
Drop Azure Arc ESU Dashboard
Add in a drop referencing an Azure Arc Windows ESU Dashboard that people can use to view their Arc and ESU status.
Thanks for the feedback @fcabrera23, I've made the necessary changes. Let me know your thoughts. Thanks!
|
2025-04-01T06:36:44.672960
| 2021-08-27T09:57:15
|
981078307
|
{
"authors": [
"marvinbuss",
"nielsvdc",
"t-dedah"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:286",
"repo": "Azure/arm-deploy",
"url": "https://github.com/Azure/arm-deploy/issues/80"
}
|
gharchive/issue
|
Redeploying Bastion via template fails (Subnet AzureBastionSubnet is in use)
I'm not sure if this belongs here, but I'll give it a try.
I've created a bicep template to deploy a vnet resource with a Bastion resource. The first time deploying this template works as expected and all resources are created.
But when redeploying the template, I get the error "Subnet AzureBastionSubnet is in use by /subscriptions/..../resourceGroups/rg-test/providers/Microsoft.Network/bastionHosts/avd-vnet-bastion/bastionHostIp
Configurations/IpConf and cannot be deleted. In order to delete the subnet, delete all the resources within the subnet."
What I'm expecting is that when redeploying the template, it would detected that the resources exist and no changes are made to the resources. This is how other templates work when deploying other resources. I don't understand why it's trying to delete the AzureBastionSubnet.
See attached file that is the bicep template.
bastion.txt
@nielsvdc This is related to this behavior stated here: https://stackoverflow.com/questions/55901747/azure-arm-code-inusesubnetcannotbedeleted-when-trying-to-update-vnet
When the vnet gets deployed and no subnets are specified in the properties section, the vnet tries to delete the subnets. However, services are connected and hence, the deployment fails. Just the first deployment works fine, as there is no subnet that is in use.
Sorry people, I was unavailable for a moment. I tested with redeployment with CLI and also via Azure GUI. These also both failed with error message "Subnet AzureBastionSubnet is in use by /subscriptions/.../resourceGroups/test/providers/Microsoft.Network/bastionHosts/avd-vnet-bastion-eus/bastionHostIpConfigurations/IpConf and cannot be deleted. In order to delete the subnet, delete all the resources within the subnet."
Hi @nielsvdc, if we try to run az deployment group what-if ... we get the below outcome which clearly shows that they are trying to delete IpConf. So this looks like an expected behaviour.
This is a possibility when different resources doesnt have context about other resources as explained in this article
If you still feel this should not be an expected behaviour then as we use azure-cli internally yo make these calls, they will be able to help you out. Can you please raise an issue here? https://github.com/Azure/azure-cli
Note: The result may contain false positive predictions (noise).
You can help us improve the accuracy of the result by opening an issue here: https://aka.ms/WhatIfIssues.
Resource and property changes are indicated with these symbols:
- Delete
~ Modify
= Nochange
* Ignore
The deployment will update the following scope:
Scope: /subscriptions/c00d16c7-6c1f-4c03-9be1-6934a4c49682/resourceGroups/deepak-rg
~ Microsoft.Network/bastionHosts/avd-vnet-bastion [2021-02-01]
- properties.dnsName: "bst-fffc8494-9771-467e-979f-389868569805.bastion.azure.com"
- properties.scaleUnits: 2
~ properties.ipConfigurations: [
~ 0:
- id: "/subscriptions/c00d16c7-6c1f-4c03-9be1-6934a4c49682/resourceGroups/deepak-rg/providers/Microsoft.Network/bastionHosts/avd-vnet-bastion/bastionHostIpConfigurations/IpConf"
]
~ sku.name: "Basic" => "Standard"
~ Microsoft.Network/publicIPAddresses/avd-vnet-bastion-ip [2021-02-01]
- sku.tier: "Regional"
= Microsoft.Network/virtualNetworks/myVnet [2021-02-01]
= Microsoft.Network/virtualNetworks/myVnet/subnets/AzureBastionSubnet [2021-02-01]
* Microsoft.Network/networkSecurityGroups/NRMS-mqpdnhjdzis7mmyVnet
* Microsoft.Network/networkSecurityGroups/NRMS-wlg2nxyjuzp7ymyVnet-AzureBastionSubnet
* Microsoft.Storage/storageAccounts/deepak21
* Microsoft.Storage/storageAccounts/deepak2121
* Microsoft.Storage/storageAccounts/devstore4z32rb65cghuc
* Microsoft.Web/serverFarms/AppServicePlan-GEN-UNIQUE
* Microsoft.Web/serverFarms/devplan
* Microsoft.Web/sites/devapp4z32rb65cghuc
Resource changes: 2 to modify, 2 no change, 8 to ignore.
Hi @nielsvdc are you still facing this issue? If no, then should we close this?
|
2025-04-01T06:36:44.678427
| 2022-08-01T05:20:41
|
1323839318
|
{
"authors": [
"Yao725",
"m-nash"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:287",
"repo": "Azure/autorest.csharp",
"url": "https://github.com/Azure/autorest.csharp/pull/2520"
}
|
gharchive/pull-request
|
Ensure base class is absrtact in polymorphism
Description
Regen PR in sdk repo: https://github.com/Azure/azure-sdk-for-net/pull/30235
Checklist
To ensure a quick review and merge, please ensure:
[ ] The PR has a understandable title and description explaining the why and what.
[ ] The PR is opened in draft if not ready for review yet.
If opened in draft, please allocate sufficient time (24 hours) after moving out of draft for review
[ ] The branch is recent enough to not have merge conflicts upon creation.
Ready to Land?
[ ] Build is completely green
Submissions with test failures require tracking issue and approval of a CODEOWNER
[ ] At least one +1 review by a CODEOWNER
[ ] All -1 reviews are confirmed resolved by the reviewer
Override/Marking reviews stale must be discussed with CODEOWNERS first
We might need to make exceptions for this in GA'd types as its a breaking change. Technically if someone is using that ctor today it wouldn't work so it might be an acceptable break I have asked @KrzysztofCwalina to weigh in before we move forward.
|
2025-04-01T06:36:44.681663
| 2023-02-02T13:57:56
|
1568110280
|
{
"authors": [
"alzimmermsft",
"weidongxu-microsoft"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:288",
"repo": "Azure/autorest.java",
"url": "https://github.com/Azure/autorest.java/issues/1942"
}
|
gharchive/issue
|
Reduce Number of Methods Generates in Clients
Currently, depending on the configuration setup, in ClientMethodMapper we can end up generating overloads of methods that'll never be used. For example, if Context is enabled we'll end up with api(params) and api(paramsPlusContext) and if this is being used with a hand-written SDK the non-Context overload will almost never be called as the written layer will handle either propagating a Context or use an empty Context. This has a few downsides, it increases the size of JARs, results in longer compilation times, and skews test code coverage, given that we should look into the following ideas/improvements:
Only generate a Context overload if the generated client will be in implementation. The service client being in implementation is a good indicator on whether the generated code is being used in a hand-written SDK.
Only generate Response-based APIs if the generated client will be in implementation. The hand-written layer can handle cracking open the Response to return T.
If the client is generated only for implementation, one might actually use the sync-methods=NONE to only generate the Response methods (that likely be the minimum), and wrapper client can do whatever they want?
|
2025-04-01T06:36:44.698919
| 2017-12-09T10:17:14
|
280707394
|
{
"authors": [
"jafreck",
"willferreira"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:289",
"repo": "Azure/aztk",
"url": "https://github.com/Azure/aztk/issues/260"
}
|
gharchive/issue
|
global configuration files created with aztk spark init --global are not found in other dirs
When I run:
aztk spark init --global
a .aztk/ directory is correctly created in my $HOME dir. I edit the secrets.yaml file to configure keys, account names etc. But when I try and create a cluster in another location using the command:
aztk spark cluster create --id spark --vm-size standard_a2 --size 4
I get the message:
Batch account key is not set in secrets.yaml config
The workaround is to copy the directory ./aztk/ in my $HOME dir to the location I'm trying to create the cluster. I'm using the latest release of aztk with python 3.6.2 on MacOS High Sierra 10.13.1
conda list gives:
azure-batch 3.0.0
azure-common 1.1.8
azure-nspkg 2.0.0
azure-storage 0.33.0
I believe the wrong file path was being used to read the global secrets file. Just put out a PR to fix it. Nice catch!
|
2025-04-01T06:36:44.702430
| 2021-10-13T16:12:12
|
1025442953
|
{
"authors": [
"alexeldeib",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:290",
"repo": "Azure/azure-cli-extensions",
"url": "https://github.com/Azure/azure-cli-extensions/pull/3958"
}
|
gharchive/pull-request
|
bump aks-preview to 0.5.35 for mig support
This checklist is used to make sure that common guidelines for a pull request are followed.
For new extensions:
[ ] My extension description/summary conforms to the Extension Summary Guidelines.
About Extension Publish
There is a pipeline to automatically build, upload and publish extension wheels.
Once your PR is merged into master branch, a new PR will be created to update src/index.json automatically.
The precondition is to put your code inside this repo and upgrade the version in the PR but do not modify src/index.json.
forgot to do this in #3895
aks-preview
|
2025-04-01T06:36:44.707193
| 2022-09-19T23:07:06
|
1378621466
|
{
"authors": [
"vthiebaut10",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:291",
"repo": "Azure/azure-cli-extensions",
"url": "https://github.com/Azure/azure-cli-extensions/pull/5366"
}
|
gharchive/pull-request
|
[ssh] Bug Fix: Change how ssh vm reads the ssh client logs
Fix: https://github.com/Azure/azure-cli-extensions/issues/4915#issuecomment-1243939019
ssh vm reads the logs from ssh client and ssh proxy to delete sensitive information when it is no longer needed and print helpful error messages for known errors. Our approach to do that was causing ssh banners to be printed after the ssh process was terminated, and not before authentication.
Known issue: we are no longer able to read ssh proxy error channel when the user is running the command from a linux machine to connect to a local user on an arc machine. That is not a breaking change, but users in this particular situation will not get a helpful error message if there is an error with the proxy.
This checklist is used to make sure that common guidelines for a pull request are followed.
For new extensions:
[x] My extension description/summary conforms to the Extension Summary Guidelines.
About Extension Publish
There is a pipeline to automatically build, upload and publish extension wheels.
Once your pull request is merged into main branch, a new pull request will be created to update src/index.json automatically.
The precondition is to put your code inside this repository and upgrade the version in the pull request but do not modify src/index.json.
ssh
|
2025-04-01T06:36:44.718264
| 2023-04-03T05:58:49
|
1651387103
|
{
"authors": [
"navba-MSFT",
"yonzhan",
"zhoxing-ms"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:292",
"repo": "Azure/azure-cli-extensions",
"url": "https://github.com/Azure/azure-cli-extensions/pull/6157"
}
|
gharchive/pull-request
|
{AzureAppService} Fixing the supported api-version for get_kube_client() function
Fixes Azure/azure-cli-extensions#6155
We get the below error while running CLI command: az appservice kube show
Enviornment:
Windows-10-10.0.20348-SP0
Python 3.10.10
Installer: MSI
azure-cli 2.47.0
Extensions:
appservice-kube 0.1.7
Error:
File "C:\Users\Administrator\.azure\cliextensions\appservice-kube\azext_appservice_kube\custom.py", line 305, in show_kube_environments
client = _get_kube_client(cmd)
File "C:\Users\Administrator\.azure\cliextensions\appservice-kube\azext_appservice_kube\custom.py", line 301, in _get_kube_client
return client.kube_environments
File "D:\a\_work\1\s\build_scripts\windows\artifacts\cli\Lib\site-packages\azure/mgmt/web/_web_site_management_client.py", line 627, in kube_environments
ModuleNotFoundError: No module named 'azure.mgmt.web.v2021_01_01'
This error is because API version 2021-01-01 is trimmed from azure-mgmt-web SDK by scripts/trim_sdk.py to reduce MSI package size (#23946)
but appservice-kube extension hardcodes 2021-01-01:
https://github.com/Azure/azure-cli-extensions/blob/f4033ee2830c68668f340e12c39d76b6d2b7ee3f/src/appservice-kube/azext_appservice_kube/custom.py#L300
appservice-kube extension can use the latest API version api_version="2022-03-01" defined at:
https://github.com/Azure/azure-cli/blob/8c45b0feeaaa9a8ca3ed53c5cead1a8435b04fa6/src/azure-cli-core/azure/cli/core/profiles/_shared.py#L243
This checklist is used to make sure that common guidelines for a pull request are followed.
For new extensions:
[ ] My extension description/summary conforms to the Extension Summary Guidelines.
About Extension Publish
There is a pipeline to automatically build, upload and publish extension wheels.
Once your pull request is merged into main branch, a new pull request will be created to update src/index.json automatically.
You only need to update the version information in file setup.py and historical information in file HISTORY.rst in your PR but do not modify src/index.json.
AzureAppService
@StrawnSC Could you please review this once you get a chance ? Thanks in advance.
ERROR cli.azure.cli.core.azclierror:azlogging.py:212 Can't overwrite existing cassette ('/mnt/vss/_work/1/s/src/appservice-kube/azext_appservice_kube/tests/latest/recordings/test_win_webapp_quick_create_runtime_kube.yaml') in your current record mode ('once').
No match for the request (<Request (GET) https://management.azure.com/subscriptions/00000000-0000-0000-0000-000000000000/resourcegroups/clitest000001?api-version=2022-09-01>) was found.
Found 2 similar requests with 1 different matcher(s) :
1 - (<Request (GET) https://management.azure.com/subscriptions/00000000-0000-0000-0000-000000000000/resourcegroups/clitest000001?api-version=2021-04-01>)..)
Matchers succeeded : ['method', 'scheme', 'host', 'port', 'path']
Matchers failed :
_custom_request_query_matcher - assertion failure :
None
The CI issues are caused by the bumping api-version of Resource RP from 2021-04-01 to 2022-09-01 in azure-cli-core, so you need to pull the latest code from the remote dev branch of Azure/azure-cli repo and then re-record these failed tests in live mode.
|
2025-04-01T06:36:44.735539
| 2023-09-15T02:30:40
|
1897597343
|
{
"authors": [
"jonathanhe-msft",
"kairu-ms",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:293",
"repo": "Azure/azure-cli-extensions",
"url": "https://github.com/Azure/azure-cli-extensions/pull/6765"
}
|
gharchive/pull-request
|
Upgrade az vmware extension to use 2023-03-01 API version
az vmware: update to AVS 2023-03-01 API
az vmware private-cloud create: Add new --ext-nw-blocks parameter
az vmware private-cloud update: Add new --ext-nw-blocks parameter
This checklist is used to make sure that common guidelines for a pull request are followed.
For new extensions:
[ ] My extension description/summary conforms to the Extension Summary Guidelines.
About Extension Publish
There is a pipeline to automatically build, upload and publish extension wheels.
Once your pull request is merged into main branch, a new pull request will be created to update src/index.json automatically.
You only need to update the version information in file setup.py and historical information in file HISTORY.rst in your PR but do not modify src/index.json.
Thank you for your contribution! We will review the pull request and get back to you soon.
Hi @jonathanhe-msft and @cataggar. I've changed the related the code. Could you help update the test recordings in your mock servers. Thanks.
az vmware placement-policy vm-host and az vmware placement-policy vm aren't working properly as they're missing a required affinityType parameter: https://github.com/Azure/azure-rest-api-specs/blob/main/specification/vmware/resource-manager/Microsoft.AVS/stable/2023-03-01/vmware.json#L7718
az vmware hcx-enterprise-site-create failing due to bad request. the request looks fine to me, but the mock server isn't accepting the values. I think it's an error with the mock, but can you confirm everything looks ok from your end?
vmware private-cloud add-identity-source fails with ResourceNotFoundError, even though i see the calls made successfully to our mock server:
aren't working properly as they
jonathanhe-msft Could you use the --debug argument to check the api response. It seems the affinityType isn't returned in the api response of Get method. The update command will GET the resource first. And as affinityType is not exposed as an argument, the value will be inherent from the GET response. According to api specs, that property should be included in api response.
az vmware hcx-enterprise-site-create failing due to bad request. the request looks fine to me, but the mock server isn't accepting the values. I verified with the old implementation (pre-migration) and it is working as expected. can you take a look?
Fixed in this commit(https://github.com/Azure/azure-cli-extensions/pull/6765/commits/7d7faac452443a57c953f1806df6300074db8cd3)
vmware private-cloud add-identity-source fails with ResourceNotFoundError, even though i see the calls made successfully to our mock server:
That's because the
vmware private-cloud add-identity-source fails with ResourceNotFoundError, even though i see the calls made successfully to our mock server:
That's because the newly added identity source is missed in the response of your mock server. I also find this issue in old recordings and I manually fixed the server response in previous migration PR.
az vmware placement-policy vm-host and az vmware placement-policy vm aren't working properly as they're missing a required affinityType parameter: https://github.com/Azure/azure-rest-api-specs/blob/main/specification/vmware/resource-manager/Microsoft.AVS/stable/2023-03-01/vmware.json#L7718
jonathanhe-msft Could you use the --debug argument to check the api response. It seems the affinityType isn't returned in the api response of Get method. The update command will GET the resource first. And as affinityType is not exposed as an argument, the value will be inherent from the GET response. According to api specs, that property should be included in api response.
confirmed get response is missing the affinityType property. working on fixing and will update live tests once complete.
@zhoxing-ms Please help review this PR.
|
2025-04-01T06:36:44.740967
| 2023-10-18T22:55:24
|
1950838527
|
{
"authors": [
"9lash",
"yanzhudd",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:294",
"repo": "Azure/azure-cli-extensions",
"url": "https://github.com/Azure/azure-cli-extensions/pull/6873"
}
|
gharchive/pull-request
|
Update API for Aks Edge Essentials
This PR updates the get calls for annotations to provide distribution names for aks-edge-k3s and aks-edge-k8s.
This checklist is used to make sure that common guidelines for a pull request are followed.
About Extension Publish
There is a pipeline to automatically build, upload and publish extension wheels.
Once your pull request is merged into main branch, a new pull request will be created to update src/index.json automatically.
You only need to update the version information in file setup.py and historical information in file HISTORY.rst in your PR but do not modify src/index.json.
Draft
Please add test for this new feature.
Please add test for this new feature.
When you do kubectl get nodes on that infrastructure, you get an api response. In the infrastructure, the distro name is aks-edge-k3s. The test is dependent on the underlying infrastructure. We worked with the customer on validating this feature on that particular infrastructure. It passed the validation for the aks-edge-essentials scenario where this feature was tested on on-prem server with aks-edge-k3s/k8s configuration.
|
2025-04-01T06:36:44.754809
| 2019-12-20T18:19:11
|
541131612
|
{
"authors": [
"btardif",
"chenlomis",
"jiasli",
"panchagnula",
"qubitron",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:295",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/11634"
}
|
gharchive/issue
|
az webapp: Magic .azure files created in my development environment during deploy
az feedback auto-generates most of the information requested below, as of CLI version 2.0.62
Describe the bug
During the deployment two magic files were created in my development environment β .azure and .deployment. These files didn't have comments or links to any documentation regarding what these files were for, what they contained, or if the options were configurable
To Reproduce
Steps to reproduce the behavior.
Expected behavior
A clear and concise description of what you expected to happen.
Environment summary
Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (az --version) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows)
Additional context
Add any other context about the problem here.
Running az webapp up configures local context for which is included in the .Azure files see https://docs.microsoft.com/en-us/cli/azure/reference-index?view=azure-cli-latest#az-configure for more details.
Adding @btardif as FYI.
We should add a comment to the content of the files with links to more info or explanation of what they are and what they are used for... re-opening this and assigning to me to come up with plan
move to S166.
@btardif any update for this issue?
move to S167
@yonzhan , @btardif and I just had a quick chat about this issue and the webapp team is suspecting that the magical .azure file is being created via the CLI rather than the az webapp up command. There's confusion on what this file is and how to make use of it. It'd be great if we could investigate and confirm on what's happening and the next steps
@Juliehzl please help to investigate the issue mentioned above.
move to S169
@chenlomis this is the same issue being discussed here: https://dev.azure.com/azuretracking/Azure tracking/_workitems/edit/680
add to S173
.azure should be a folder instead of a file, created by Azure CLI as the configuration folder.
https://github.com/Azure/azure-cli/blob/3b4d10d473a8d72dc2a1607478beb14424ce2004/src/azure-cli-core/azure/cli/core/_environment.py#L12
It is by-default created under the user's home folder, unless
AZURE_EXTENSION_DIR is set
az configure --scope local is used
local-context is turn on
.deployment seems to be created by botservice:
https://github.com/Azure/azure-cli/blob/9370ec21d5febcb0e7099f1a3969a42d04c4ceb1/src/azure-cli/azure/cli/command_modules/botservice/custom.py#L560
Please share the full deployment script for us to further analyze the issue.
The deployment path used is here:
https://docs.microsoft.com/en-us/azure/app-service/containers/quickstart-python?tabs=bash
This generates a .azure/config file with no comment explaining what it is for.
The .deployment file comes from VS Code deploy so that is a separate issue.
We have a document Azure CLI configuration explaining what .azure/config is.
Indeed, we can put a comments in .azure/config explaining what it is. @yonzhan @achandmsft @chenlomis, any thoughts?
Sounds like a plan to me
Something along the lines of:" this is where your CLI configs are stored"
@qubitron that aligns with your proposal right?
Yep something like "This file contains saved configurations for your CLI, read more at: aka.ms/1234".
We probably want an aka link in case we change where this points to.
@jiasli @chenlomis should we re-open the issue?
Created issue https://github.com/Azure/azure-cli/issues/15001 to track.
|
2025-04-01T06:36:44.791317
| 2020-01-28T22:46:53
|
556526714
|
{
"authors": [
"Juliehzl",
"hulseware",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:296",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/11969"
}
|
gharchive/issue
|
Very ugly error message when using az account logging update command on a Storage Account of kind BlobStorage or sku.tier Premium
If you try to exeute az account logging update with a Premium storage account or with a BlobStorage storage account, you will get an error similar to the following:
az>> az storage logging show --account-name xxxxxxxxxxx --subscription xx-xxx-xxx
The command failed with an unexpected error. Here is the traceback:
HTTPSConnectionPool(host='xxxxxxxxxx.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x00DFC830>: Failed to establish a new connection: [Errno 11002] getaddrinfo failed',))
Traceback (most recent call last):
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connection.py", line 159, in _new_conn
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\util\connection.py", line 57, in create_connection
File "C:\Program Files (x86)\Microsoft SDKs\Azure\CLI2\lib\socket.py", line 745, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno 11002] getaddrinfo failed
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connectionpool.py", line 600, in urlopen
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connectionpool.py", line 343, in _make_request
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connectionpool.py", line 839, in _validate_conn
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connection.py", line 301, in connect
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connection.py", line 168, in _new_conn
urllib3.exceptions.NewConnectionError: <urllib3.connection.VerifiedHTTPSConnection object at 0x00DFC830>: Failed to establish a new connection: [Errno 11002] getaddrinfo failed
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\requests\requests\adapters.py", line 449, in send
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\connectionpool.py", line 638, in urlopen
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-hb486cm7\urllib3\urllib3\util\retry.py", line 399, in increment
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='xxxxxxxxxxx.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x00DFC830>: Failed to establish a new connection: [Errno 11002] getaddrinfo failed',))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-multiapi-storage\azure\multiapi\storage\v2018_11_09\common\storageclient.py", line 321, in _perform_request
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-multiapi-storage\azure\multiapi\storage\v2018_11_09\common_http\httpclient.py", line 92, in perform_request
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\requests\requests\sessions.py", line 533, in request
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\requests\requests\sessions.py", line 646, in send
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\requests\requests\adapters.py", line 516, in send
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='xxxxxxxxx.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x00DFC830>: Failed to establish a new connection: [Errno 11002] getaddrinfo failed',))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Users\bhulse.azure\cliextensions\interactive\azext_interactive\azclishell\app.py", line 662, in cli_execute
result = invocation.execute(args)
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands_init_.py", line 603, in execute
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands_init_.py", line 661, in run_jobs_serially
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands_init.py", line 652, in run_job
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands\arm.py", line 750, in show_exception_handler
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands_init.py", line 631, in run_job
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core\commands_init.py", line 306, in call
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli-core\azure\cli\core_init_.py", line 485, in default_command_handler
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli\azure\cli\command_modules\storage\operations\logging.py", line 15, in get_logging
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-cli\azure\cli\command_modules\storage\services_wrapper.py", line 29, in get_logging
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-multiapi-storage\azure\multiapi\storage\v2018_11_09\queue\queueservice.py", line 372, in get_queue_service_properties
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-multiapi-storage\azure\multiapi\storage\v2018_11_09\common\storageclient.py", line 430, in _perform_request
File "C:\Users\VSSADM~1\AppData\Local\Temp\pip-install-umfcm_n2\azure-multiapi-storage\azure\multiapi\storage\v2018_11_09\common\storageclient.py", line 361, in _perform_request
azure.common.AzureException: HTTPSConnectionPool(host='xxxxxxxxxxx.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x00DFC830>: Failed to establish a new connection: [Errno 11002] getaddrinfo failed',))
I would have expected a simple error message along the lines of:
Premium Storage accounts do not support logging
BlobStorage storage accounts do not support logging.
I though I was having a problem with api quotas.
Document Details
β Do not edit this section. It is required for docs.microsoft.com β GitHub issue linking.
ID: c237378a-a708-8e2f-3c8c-0bc8b623d21c
Version Independent ID: c4e27b67-9605-60f5-899d-8ffd727079b2
Content: az storage logging
Content Source: src/azure-cli/azure/cli/command_modules/storage/_help.py
Service: storage
GitHub Login: @rloutlaw
Microsoft Alias: routlaw
add to S168
Hi @hulseware, sorry for late response first.
There are several points to clarify here:
Are you using az interactive to run cli commands?
When I try to reproduce your issue with normal cli (not in interactive mode), I only get one exception not four exceptions like what in your scenario. Here is what I get:
HTTPSConnectionPool(host='zuhblob.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.HTTPSConnect
ion object at 0x000001F15CE28288>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed'))
Your suggestion is very valuable to us. As you mentioned, the error occurs because of specific storage account type. Actually it is caused by default value (bqt) in --services parameter:
BlobStorage doesn't support logging for queue service;
PremiumStorage doesn't support logging for blob/queue/table services.
I will refine the error message to make it point to wrong storage account type.
Yes, I use az interactive while developing the right cli commands to add to utilities targeted at maintaining better security across the subscriptions I am managing for my team. So the extra tracebacks were coming from the az interactive tool.
Yes, I believe that message is what I see when running via the normal command line.
Thanks for working to make the error message more descriptive of the real problem!
From: Zunli Hu<EMAIL_ADDRESS>Sent: Thursday, May 21, 2020 2:11 AM
To: Azure/azure-cli<EMAIL_ADDRESS>Cc: Hulse, Bruce<EMAIL_ADDRESS>Mention<EMAIL_ADDRESS>Subject: Re: [Azure/azure-cli] Very ugly error message when using az account logging update command on a Storage Account of kind BlobStorage or sku.tier Premium (#11969)
External email from: noreply<EMAIL_ADDRESS>Hi @hulsewarehttps://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fhulseware&data=02|01|bhulse%40ptc.com|f25ae0dc795942b8432f08d7fd4dd671|b9921086ff774d0d828acb3381f678e2|0|0|637256383063474220&sdata=Zo%2FTrRWewqd33HZtjCS%2FtKAUGF%2FCs%2F2qoIMW0OlqI28%3D&reserved=0, sorry for late response first.
There are several points to clarify here:
Are you using az interactive to run cli commands?
When I try to reproduce your issue with normal cli (not in interactive mode), I only get one exception not four exceptions like what in your scenario. Here is what I get:
HTTPSConnectionPool(host='zuhblob.queue.core.windows.net', port=443): Max retries exceeded with url: /?restype=service&comp=properties (Caused by NewConnectionError('<urllib3.connection.HTTPSConnect
ion object at 0x000001F15CE28288>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed'))
Your suggestion is very valuable to us. As you mentioned, the error occurs because of specific storage account type. Actually it is caused by default value (bqt) in --services parameter:
BlobStorage doesn't support logging for queue service;
PremiumStorage doesn't support logging for blob/queue/table services.
I will refine the error message to make it point to wrong storage account type.
β
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHubhttps://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2FAzure%2Fazure-cli%2Fissues%2F11969%23issuecomment-631902964&data=02|01|bhulse%40ptc.com|f25ae0dc795942b8432f08d7fd4dd671|b9921086ff774d0d828acb3381f678e2|0|0|637256383063479216&sdata=HCtaDxvR6agdww5TSxgEOE6WmNgJ4eTYKG3mk9%2Bpw00%3D&reserved=0, or unsubscribehttps://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FACLZFQQTT6JLU27D6HU6GFLRSTAYXANCNFSM4KM3LHYQ&data=02|01|bhulse%40ptc.com|f25ae0dc795942b8432f08d7fd4dd671|b9921086ff774d0d828acb3381f678e2|0|0|637256383063484198&sdata=Dj9dgmYAByp4XMcbQz9VIO6FfXbVtopi33iJZVX1u9A%3D&reserved=0.
|
2025-04-01T06:36:44.799367
| 2020-05-25T04:16:32
|
624033610
|
{
"authors": [
"cuzzlor",
"fengzhou-msft",
"panchagnula",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:297",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/13660"
}
|
gharchive/issue
|
az cli fails with ModuleNotFoundError
This is autogenerated. Please review and update as needed.
Describe the bug
Command Name
az webapp create
Errors:
No module named 'decorator'
Traceback (most recent call last):
python3/dist-packages/fabric/connection.py, ln 5, in <module>
from invoke.vendor.six import StringIO
ModuleNotFoundError: No module named 'invoke.vendor.six'
...
python3/dist-packages/fabric/connection.py, ln 10, in <module>
from decorator import decorator
ModuleNotFoundError: No module named 'decorator'
To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
Install Ubuntu 20.4 LTS on Windows WSL
Install az cli following these steps: https://docs.microsoft.com/en-us/cli/azure/install-azure-cli-apt?view=azure-cli-latest#set-release
az login
az webapp create --name --resource-group --plan --runtime {}
Expected Behavior
az commands work
Environment Summary
Linux-4.4.0-18362-Microsoft-x86_64-with-glibc2.29
Python 3.8.2
Shell: bash
azure-cli 2.0.81
Extensions:
azure-devops 0.17.0
Additional Context
webapp
No related to app service - see https://github.com/Azure/azure-cli/issues/13644
I read all that at the time, I still couldnβt make it work on a fresh install following all the instructions. Care to elaborate? Someone elseβs job?
@fengzhou-msft could you help with this?
@cuzzlor you need to first remove the packages installed by the old 2.0.81 azure-cli that comes with the Ubuntu 20.4 LTS. Try sudo apt remove azure-cli and make sure /usr/lib/python3/dist-packages/azure is deleted.
Ok. At the time it was a brand new install of WSL and the azure cli install script. Iβll try again maybe something has changed. Thanks.
You may need to run sudo apt autoremove after sudo apt remove azure-cli to remove the old azure packages that azure-cli 2.0.81 depends on.
Duplicate with #14011
|
2025-04-01T06:36:44.806509
| 2020-08-29T01:49:55
|
688434804
|
{
"authors": [
"arde0708",
"fengzhou-msft",
"nachoalonsoportillo",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:298",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/14968"
}
|
gharchive/issue
|
How to pass an argument parameter without defining a value?
I am trying to achieve three scenarios for a command. Is there a way to do this ?
az postgres server create (no argument parameter passed)
az postgres server create --public-access (no value passed for argument parameter --public-access)
az postgres server create --public access ( a valid value is passed to argument parameter --public-access).
Is there a way to do (2) that is pass a parameter without a value ? I am running into the below error :
az postgres server create: error: argument --public-access: expected one argument.
If this is do-able, how should I differentiate between (1) and (2) that is whether the parameter is completely missing or is present with a missing value ?
postgres
@arde0708 what are the sample values of --public-access? What is the behavior for 2?
We have get_three_state_flag arg_type that can achieve 2 and 3, but it only support boolean values.
We would want to define three kind of values for --public-access
No value. Just the parameter is passed. Ex : 'az postgres flexible-server create --public-access' ( Adds neither a Vnet, nor a firewall rule)
Single IP. Ex : 'az postgres flexible-server create --public-access <IP_ADDRESS>' (Behavior : Adds single firewall rule)
Multiple IP. Ex : 'az postgres flexible-server create --public-access <IP_ADDRESS>-<IP_ADDRESS>' (Behavior : Adds range of firewall rules)
Don't pass the parameter itself 'Ex : az postgres flexible-server create' (Behavior : Adds Vnet, Subnet)
@yonzhan Single Server is being retired. We won't be fixing this at this point in time. Please, close this issue. Thanks!
|
2025-04-01T06:36:44.813247
| 2022-06-23T21:03:12
|
1282938468
|
{
"authors": [
"navba-MSFT",
"roslynlu",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:299",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/22990"
}
|
gharchive/issue
|
[Stream Analytics] Bugs related to clusterid property
az feedback auto-generates most of the information requested below, as of CLI version 2.0.62
Related command
az stream-analytics job update
In our streaming job definition, we have the following nested cluster and id properties:
"properties": {
"cluster": {
"id": "subscriptions/{mysub}/resourceGroup/{myrg}/etc..."
}
}
Describe the bug
In the CLI, this cluster id property gets translated into just id:
o az stream-analytics job update --job-name 'asa-fleide-large396' --resource-group $rg --id $clusterId
We think this is a possible source of confusion for the user -- because this command is an operation on the job, it's reasonable to infer that id is the job id, not the cluster id. Could we possible change this parameter to clusterid?
Another possible scenario is that the user should be able to set this id value to null. I tried to do this via CLI (setting the value to null or empty string), but got the following error:
(LinkedInvalidPropertyId) Property id '' at path 'properties.cluster.id' is invalid. Expect fully qualified resource Id that start with '/subscriptions/{subscriptionId}' or '/providers/{resourceProviderNamespace}/'.
Code: LinkedInvalidPropertyId
Message: Property id '' at path 'properties.cluster.id' is invalid. Expect fully qualified resource Id that start with '/subscriptions/{subscriptionId}' or '/providers/{resourceProviderNamespace}/'.
Is there another way to set it this property to null? If not, can we remove this validation?
To Reproduce
az stream-analytics job update --job-name 'asa-fleide-large396' --resource-group $rg --id $clusterId
az stream-analytics job update --job-name 'asa-fleide-large396' --resource-group $rg --id ""
Expected behavior
property name is clusterid instead of id
user able to set id value to null
Environment summary
Additional context
route to CXP team
Noncustomer-reported issue. Adding Service team to look in this.
|
2025-04-01T06:36:44.821956
| 2023-02-13T14:21:25
|
1582443192
|
{
"authors": [
"arambazamba",
"navba-MSFT",
"seligj95",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:300",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/25443"
}
|
gharchive/issue
|
az webapp up does re-create existing appservice pan and does not upload published .net app
az feedback auto-generates most of the information requested below, as of CLI version 2.0.62
Related command
az webapp up -n $app -g $grp -p $plan --sku F1 -l $loc -r "DOTNET|6.0"
Describe the bug
Azure cli 2.45.0 az webapp up does re-create existing appservice pan insted of using it and does not upload published .net app - instead the raw source files are uploaded
To Reproduce
env=$RANDOM
grp=appservices-$env
loc=westeurope
plan=appservices-$env
app=cli-api-$env
az group create -n $grp -l $loc
az appservice plan create -n $plan -g $grp --sku S1
az webapp create -n $app -g $grp --plan $plan --runtime "DOTNET|6.0"
dotnet new webapi -n cli-api --framework net6.0
cd cli-api
az webapp up -n $app -g $grp -p $plan --sku F1 -l $loc -r "DOTNET|6.0"
cd ..
Expected behavior
the existing app service plan should be used and the app that has been processed by dotnet publish should have been uploaded
Environment summary
Windows 11, WSL, Azure cli 2.45.0
Additional context
route to CXP team
Adding Service team to look into this.
Unable to repro. Closing for now as this has been open for a while with no recent activity. Please update to the latest CLI version and let us know if this is still an issue.
Also, when running az webapp up after already creating the app using az webapp create, you don't need to specify the runtime/location/SKU again in az webapp up. The cli output when running the command might say "Creating AppServicePlan..." even though it already exists. We are aware of that bug and are addressing that separately. See https://github.com/Azure/azure-cli/issues/22767 for more details on that issue.
|
2025-04-01T06:36:44.829107
| 2023-08-31T10:47:06
|
1875271598
|
{
"authors": [
"kamperiadis",
"luis261",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:301",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/27297"
}
|
gharchive/issue
|
Implement slot parameter for az functionapp function show
Related command
az functionapp function show
Is your feature request related to a problem? Please describe.
I can't find a way to run az functionapp function show and target a deployment slot (besides the default/production one).
It apparently does not offer a --slot parameter: https://learn.microsoft.com/en-us/cli/azure/functionapp/function?view=azure-cli-latest#az-functionapp-function-show
Describe the solution you'd like
Please consider adding a --slot parameter to az functionapp function show
Describe alternatives you've considered
I tried a workaround via modifying --name to match the test slots name, but that does not seem to work
Additional context
This is the context in which I would use the new parameter: https://github.com/luis261/aztraphile/blob/reintroduce-slot-bound-http-invocations/utils/aztra-utils.ps1#L334
Thank you for opening this issue, we will look into it.
Hello @luis261 Thank you for your feedback. This is a feature request I just added to our backlog. In the meantime, you would just have to make direct API calls to https://management.azure.com/subscriptions/<subscriptionId>/resourceGroups/<resourceGroup>/providers/Microsoft.Web/sites/<functionAppName>/slots/<slotName>/functions/<functionName>?api-version=2022-03-01.
Hi @kamperiadis - glad to know it made the backlog! Thanks for keeping me in the loop. Also good to know the alternative via the direct API call (:
|
2025-04-01T06:36:44.834652
| 2017-08-11T14:59:03
|
249661833
|
{
"authors": [
"nneul",
"superminiek",
"tjprescott",
"yugangw-msft"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:302",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/4202"
}
|
gharchive/issue
|
Add a VM to an AvailabilitySet after creation.
Description
Is there an azure-cli equivalent of this powershell cmdlet? Described here:
https://www.petri.com/understanding-azure-availability-sets
https://docs.microsoft.com/en-us/powershell/module/azure/set-azureavailabilityset?view=azuresmps-4.0.0`
The first link has an example like this:
Get-AzureVM -ServiceName contosodc1 -Name contosodc1 | Set-AzureAvailabilitySet -AvailabilitySetName contosodcs | Update-AzureVM
Environment summary
Install Method: nightly
CLI Version: azure-cli (2.0.12+1.dev20170808)
OS Version: ubuntu 16.04 x64
Shell Type: bash
Hi @nneul there is no dedicated command for this. I tried to accomplish this via generic update and had no luck. It seems you must specify the availability set when you create the VM (at least currently in 2.0) We have an open project to improve the adding of VMs to load balancers and application gateway pools.
I'll investigate some potential solutions for this next sprint and see if we can get something into the next release. Ideas would be having an availability-set add-vm command, or adding availability set as strongly typed parameter on vm update.
Moving this into the same project as improving LB/AG experience.
Anything new here?
The item is on our backlog but not planned for upcoming work. @yugangw-msft does this fit into your upcoming sprint plans?
@tjprescott, I will take a look in the next sprint
I have confirmed with Compute team that adding an existing VM to an existing av-set only works with RDFE(cloud service). For Compute RP, the only time you can add to an av-set is during VM create.
|
2025-04-01T06:36:44.849438
| 2018-09-11T11:59:50
|
359013990
|
{
"authors": [
"holgerbrandl",
"mattchenderson",
"panchagnula",
"swm7",
"tjprescott"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:303",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/issues/7282"
}
|
gharchive/issue
|
az webapp deployment source config-zip fails from jenkins pipeline
Describe the bug
We are using Jenkins integration with Azure. While trying to run az deploy webapp from jenkins pipeline it fails with the following error:
az webapp deployment source config-zip --debug --resource-group st-worker-test --name st-worker-test --slot stagedeploy --src<EMAIL_ADDRESS>ERROR: Expecting value: line 1 column 1 (char 0)
Traceback (most recent call last):
File "/opt/az/lib/python3.6/site-packages/knack/cli.py", line 197, in invoke
cmd_result = self.invocation.execute(args)
File "/opt/az/lib/python3.6/site-packages/azure/cli/core/commands/init.py", line 369, in execute
six.reraise(*sys.exc_info())
File "/opt/az/lib/python3.6/site-packages/six.py", line 693, in reraise
raise value
File "/opt/az/lib/python3.6/site-packages/azure/cli/core/commands/init.py", line 343, in execute
result = cmd(params)
File "/opt/az/lib/python3.6/site-packages/azure/cli/core/commands/init.py", line 182, in call
return self.handler(*args, **kwargs)
File "/opt/az/lib/python3.6/site-packages/azure/cli/core/init.py", line 436, in default_command_handler
result = op(**command_args)
File "/opt/az/lib/python3.6/site-packages/azure/cli/command_modules/appservice/custom.py", line 188, in enable_zip_deploy
response = response.json()
File "/opt/az/lib/python3.6/site-packages/requests/models.py", line 896, in json
return complexjson.loads(self.text, **kwargs)
File "/opt/az/lib/python3.6/json/init.py", line 354, in loads
return _default_decoder.decode(s)
File "/opt/az/lib/python3.6/json/decoder.py", line 339, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/opt/az/lib/python3.6/json/decoder.py", line 357, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
To Reproduce
Set to Azure serviceprincipal credentials
use Azure cli commands for the below:
login to Azure with service principal
set to the azure subscription
run az webapp deployment source config-zip --debug --resource-group st-worker-test --name st-worker-test --slot stagedeploy --src<EMAIL_ADDRESS>The command above fails with the error mentioned.
Expected behavior
It should deploy the zip file to Azure webapp sucessfully
Environment summary
Install Method : sudo apt-get / CLI version (azure-cli (2.0.45)) / Ubuntu 16.0.4 / Shell Type (bash)
Additional context
If i run the same command manually by logging in to the Jenkins slave it works successfully. We need to get this working from Jenkins pipeline and need help in fixing the issue.
Also seeing that the deployment hangs sometimes. Attached is the debug log. I really need help on this. Could someone please look into it.
azure-deploy-log.docx
@panchagnula
@panchagnula I think the right fix here is to make the timeout configurable with an optional flag. The current behavior is to be assumed as the default if no timeout is provided.
Same here using on a bash terminal under macos v10.14.1. When deploying the demo project under https://docs.microsoft.com/en-us/learn/modules/store-app-data-with-azure-blob-storage/7-blob-uploads-and-downloads it fails with a highly similar error:
az webapp deployment source config-zip --resource-group ${AZ_RES_GROUP} --name ${AZ_WEBAPP_NAME} --src ../site.zip
Expecting value: line 1 column 1 (char 0)
Traceback (most recent call last):
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/knack/cli.py", line 197, in invoke
cmd_result = self.invocation.execute(args)
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/azure/cli/core/commands/__init__.py", line 369, in execute
six.reraise(*sys.exc_info())
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/six.py", line 693, in reraise
raise value
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/azure/cli/core/commands/__init__.py", line 343, in execute
result = cmd(params)
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/azure/cli/core/commands/__init__.py", line 182, in __call__
return self.handler(*args, **kwargs)
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/azure/cli/core/__init__.py", line 436, in default_command_handler
result = op(**command_args)
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/azure/cli/command_modules/appservice/custom.py", line 262, in enable_zip_deploy
response = response.json()
File "/usr/local/Cellar/azure-cli/2.0.48/libexec/lib/python3.7/site-packages/requests/models.py", line 892, in json
return complexjson.loads(self.text, **kwargs)
File "/Users/brandl/anaconda3/lib/python3.6/json/__init__.py", line 354, in loads
return _default_decoder.decode(s)
File "/Users/brandl/anaconda3/lib/python3.6/json/decoder.py", line 339, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/Users/brandl/anaconda3/lib/python3.6/json/decoder.py", line 357, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
My zip file seems legit, and my CLI version is azure-cli (2.0.48)
@holgerbrandl - what is the python version that is being used here? is it possible for you to upgrade the python version & try? Do Ptyhon --version, this is the version i have installed "Python 3.7.0". thank you.
Since I had deleted the corresponding resource group already, I've created the complete demo setup again (group, storage account, wegapp) and tried to reproduce the error. However now it seems fine and deployment works as expected.
I'm not aware of any changes in my local setup. My python version has not changed since I have reported the ticket and is 3.6.3.
Not sure if this is compatible with the stacktrace, but it may have been also some kind of connection timeout?
@holgerbrandl , thanks for the update. with the new setup has the version of azure-cli changed, especially since we have a newer version available now & we have had some code changes to the zip deployment command, so there might be some bugfixes that would have infact fixed the problem you are seeing.
|
2025-04-01T06:36:44.854852
| 2022-09-09T07:29:28
|
1367411444
|
{
"authors": [
"jiasli",
"yonzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:304",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/pull/23828"
}
|
gharchive/pull-request
|
[Core] PREVIEW: Support Web Account Manager (WAM) login on Windows
Related command
az login
Description
Close #21201
Require https://github.com/AzureAD/microsoft-authentication-library-for-python/pull/415
Rework https://github.com/Azure/azure-cli/pull/22774
Testing Guide
On Windows, users can opt in WAM login using:
az config set core.allow_broker=true
Then run
az login
Instead of launching a web browser, an account selector will be shown to ask the user to select a Windows account:
History Notes
[Core] PREVIEW: Support Web Account Manager (WAM) login on Windows. To opt in, run az config set core.allow_broker=true
Additional information
WAM will be enabled by default in year 2023.
WAM support
|
2025-04-01T06:36:44.869014
| 2017-03-07T00:07:58
|
212290280
|
{
"authors": [
"codecov-io",
"derekbekoe"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:305",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/pull/2398"
}
|
gharchive/pull-request
|
Add profile switching params and profile listing command
Modify βaz cloud registerβ and βaz cloud updateβ to include the ββprofileβ parameter
Add ability to view all supported API versions without querying a cloud
note: With this change, switching profile will not change anything as they are just example profiles.
FYI @lmazuel For the _shared.py file change. This would contain the resource type to api version map.
Closes https://github.com/Azure/azure-cli/issues/2280
Codecov Report
Merging #2398 into api-profile-support will increase coverage by <.01%.
The diff coverage is 77.14%.
@@ Coverage Diff @@
## api-profile-support #2398 +/- ##
=======================================================
+ Coverage 72.33% 72.33% +<.01%
=======================================================
Files 323 325 +2
Lines 18273 18304 +31
Branches 2701 2707 +6
=======================================================
+ Hits 13217 13240 +23
- Misses 4223 4231 +8
Partials 833 833
Impacted Files
Coverage Ξ
...cli-cloud/azure/cli/command_modules/cloud/_help.py
100% <100%> (ΓΈ)
:white_check_mark:
src/azure-cli-core/azure/cli/core/cloud.py
69.08% <100%> (+2.58%)
:white_check_mark:
...-cloud/azure/cli/command_modules/cloud/commands.py
100% <100%> (ΓΈ)
:white_check_mark:
...azure-cli-core/azure/cli/core/profiles/init.py
100% <100%> (ΓΈ)
...i-cloud/azure/cli/command_modules/cloud/_params.py
93.33% <100%> (+1.02%)
:white_check_mark:
src/azure-cli-core/azure/cli/core/_profile.py
83.78% <100%> (-0.15%)
:x:
.../azure-cli-core/azure/cli/core/profiles/_shared.py
100% <100%> (ΓΈ)
...li-cloud/azure/cli/command_modules/cloud/custom.py
17.24% <11.11%> (-1.13%)
:x:
Continue to review full report at Codecov.
Legend - Click here to learn more
Ξ = absolute <relative> (impact), ΓΈ = not affected, ? = missing data
Powered by Codecov. Last update 4c0fdcf...52e3356. Read the comment docs.
|
2025-04-01T06:36:44.875038
| 2022-12-05T04:29:12
|
1475637951
|
{
"authors": [
"jiasli"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:306",
"repo": "Azure/azure-cli",
"url": "https://github.com/Azure/azure-cli/pull/24804"
}
|
gharchive/pull-request
|
{Style} Use light theme by default on MacOS
Issue
We specify the color for warnings as ANSI color Bright Yellow without any specification of the actual RGB value:
https://github.com/Azure/azure-cli/blob/8d8d71707d6c60083d008062667cab18d0adaa25/src/azure-cli-core/azure/cli/core/style.py#L73
According to https://github.com/Azure/azure-cli/issues/18298#issuecomment-1263529261, MacOS's terminal by default uses white background and Bright Yellow is not readable on that white background.
Something like:
Personally, I think it should the terminal's theme's responsibility to make all ANSI colors readable on its background.
Change
Since MacOS's terminal by default uses white background, we by default uses light theme which uses (dark) Yellow for warnings:
https://github.com/Azure/azure-cli/blob/8d8d71707d6c60083d008062667cab18d0adaa25/src/azure-cli-core/azure/cli/core/style.py#L85
Cons
However, a disadvantage is that if the user manually chooses a dark background, or use an IDE like PyCharm with dark theme. There may be other readability issues if we default to light theme, such as blue, magenta.
As there isn't a way to detect the background color, there is no optimal solution. Users still need to run az config set core.theme=light/dark to select the color theme or totally disable color with az config set core.no_color=true.
https://stackoverflow.com/questions/58186391/text-displaying-as-bright-yellow-in-terminal
This is the interface for configuring ANSI color's real rendered RGB:
|
2025-04-01T06:36:44.876873
| 2018-03-18T00:32:24
|
306202432
|
{
"authors": [
"frenchyjef",
"sharmasushant"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:307",
"repo": "Azure/azure-container-networking",
"url": "https://github.com/Azure/azure-container-networking/issues/104"
}
|
gharchive/issue
|
Failed to create File Store directory Error:mkdir : The system cannot find the path specified.
Running the new azure-vnet-plugin.exe 1.0.3 on my Windows 10 Azure VM gives me the below error...
C:\temp\azure-vnet-cnm-windows-amd64-v1.0.3>azure-vnet-plugin.exe
Failed to create File Store directory Error:mkdir : The system cannot find the path specified.
Also, am I supposed to set the --api-url argument? What value should go there?
Thanks,
J-F
Closing as this is now fixed in latest release.
|
2025-04-01T06:36:44.879232
| 2016-01-24T06:28:05
|
128376807
|
{
"authors": [
"andyjballgit",
"azurecla",
"ggailey777",
"tfitzmac"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:308",
"repo": "Azure/azure-content",
"url": "https://github.com/Azure/azure-content/pull/5442"
}
|
gharchive/pull-request
|
Update resource-group-move-resources.md
Added Express Route as not support moving a resource
Hi @andyjballgit, I'm your friendly neighborhood Azure Pull Request Bot (You can call me AZPRBOT). Thanks for your contribution!
This seems like a small (but important) contribution, so no contribution license agreement is required at this point. Real humans will now evaluate your PR.
TTYL, AZPRBOT;
@ggailey777 - I have double-checked this change. This pull request in azure-content can be merged.
Thanks for helping out with the Azure docs
|
2025-04-01T06:36:44.882896
| 2016-06-14T18:02:24
|
160245572
|
{
"authors": [
"PRmerger",
"azurecla",
"cjgronlund",
"victorar"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:309",
"repo": "Azure/azure-content",
"url": "https://github.com/Azure/azure-content/pull/6822"
}
|
gharchive/pull-request
|
fixed minor typo
fixed minor typo
Hi @victorar, I'm your friendly neighborhood Azure Pull Request Bot (You can call me AZPRBOT). Thanks for your contribution!
It looks like you're working at Microsoft (victorar). If you're full-time, we DON'T require a contribution license agreement.
If you are a vendor, DO please sign the electronic contribution license agreement. It will take 2 minutes and there's no faxing! https://cla.azure.com.
TTYL, AZPRBOT;
@victorar : Thanks for your contribution to the Azure documentation! The author, @telmosampaio, has been notified to review your proposed change.
@telmosampaio
Easy typo. I'm accepting this PR. Thanks!
--Carolyn
@victorar
Nice catch! I'll accept this PR. Thank you!
--Carolyn
|
2025-04-01T06:36:44.970964
| 2023-01-07T00:45:22
|
1523359111
|
{
"authors": [
"azure-sdk",
"weikanglim"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:310",
"repo": "Azure/azure-dev",
"url": "https://github.com/Azure/azure-dev/pull/1329"
}
|
gharchive/pull-request
|
Add pid to debug attach printout
Add pid to debug attach printout. Since attaching debugger to the process requires the PID, having it displayed saves an extra step and avoids confusion when multiple instances of azd is running.
Azure Dev CLI Install Instructions
Install scripts
MacOS/Linux
May elevate using sudo on some platforms and configurations
bash:
curl -fsSL https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/uninstall-azd.sh | bash;
curl -fsSL https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/install-azd.sh | bash -s -- --base-url https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329 --version '' --verbose
pwsh:
Invoke-RestMethod 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/uninstall-azd.ps1' -OutFile uninstall-azd.ps1; ./uninstall-azd.ps1
Invoke-RestMethod 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/install-azd.ps1' -OutFile install-azd.ps1; ./install-azd.ps1 -BaseUrl 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329' -Version '' -Verbose
Windows
PowerShell install
powershell -c "Set-ExecutionPolicy Bypass Process; irm 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/uninstall-azd.ps1' > uninstall-azd.ps1; ./uninstall-azd.ps1;"
powershell -c "Set-ExecutionPolicy Bypass Process; irm 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/install-azd.ps1' > install-azd.ps1; ./install-azd.ps1 -BaseUrl 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329' -Version '' -Verbose;"
MSI install
powershell -c "irm 'https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/azd-windows-amd64.msi' -OutFile azd-windows-amd64.msi; msiexec /i azd-windows-amd64.msi /qn"
Standalone Binary
Linux - https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/azd-linux-amd64.tar.gz
MacOS - https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/azd-darwin-amd64.zip
Windows - https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/azd-windows-amd64.zip
MSI
https://azuresdkreleasepreview.blob.core.windows.net/azd/standalone/pr/1329/azd-windows-amd64.msi
Container
docker run -it azdevcliextacr.azurecr.io/azure-dev:pr-1329
Documentation
learn.microsoft.com documentation
title: Azure Developer CLI Preview reference
description: This article explains the syntax and parameters for the various Azure Developer CLI Preview commands.
author: puichan
ms.author: puichan
ms.date: 01/07/23
ms.topic: conceptual
ms.custom: devx-track-azdevcli
ms.prod: azure
Azure Developer CLI Preview reference
This article explains the syntax and parameters for the various Azure Developer CLI Preview commands.
azd
Azure Developer CLI is a command-line interface for developers who build Azure solutions.
Synopsis
Azure Developer CLI is a command-line interface for developers who build Azure solutions.
To begin working with Azure Developer CLI, run the azd up command by supplying a sample template in an empty directory:
azd up β-template todo-nodejs-mongo
You can pick a template by running azd template listand then supplying the repo name as a value to --template.
The most common next commands are:
azd pipeline config
azd deploy
azd monitor --overview
For more information, visit the Azure Developer CLI Dev Hub: https://aka.ms/azure-dev/devhub.
Options
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
-h, --help Gets help for azd.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
azd deploy: Deploy the application's code to Azure.
azd down: Delete Azure resources for an application.
azd env: Manage environments.
azd infra: Manage Azure resources.
azd init: Initialize a new application.
azd login: Log in to Azure.
azd logout: Log out of Azure
azd monitor: Monitor a deployed application.
azd pipeline: Manage GitHub Actions pipelines.
azd provision: Provision the Azure resources for an application.
azd restore: Restore application dependencies.
azd template: Manage templates.
azd up: Initialize application, provision Azure resources, and deploy your project with a single command.
azd version: Print the version number of Azure Developer CLI.
azd config
Manage Azure Developer CLI configuration
Synopsis
Manage the Azure Developer CLI user configuration, which includes your default Azure subscription and location.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable.
Options
-h, --help Gets help for config.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config get: Gets a configuration
azd config list: Lists all configuration values
azd config reset: Resets configuration to default
azd config set: Sets a configuration
azd config unset: Unsets a configuration
Back to top
azd config get
Gets a configuration
Synopsis
Gets a configuration in the configuration path.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable.
azd config get <path> [flags]
Options
-h, --help Gets help for get.
-o, --output string The output format (the supported formats are json). (default "json")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
Back to top
azd config list
Lists all configuration values
Synopsis
Lists all configuration values in the configuration path.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable.
azd config list [flags]
Options
-h, --help Gets help for list.
-o, --output string The output format (the supported formats are json). (default "json")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
Back to top
azd config reset
Resets configuration to default
Synopsis
Resets all configuration in the configuration path.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable to the default.
azd config reset [flags]
Options
-h, --help Gets help for reset.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
Back to top
azd config set
Sets a configuration
Synopsis
Sets a configuration in the configuration path.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable.
azd config set <path> <value> [flags]
Examples
azd config set defaults.subscription <yourSubscriptionID>
azd config set defaults.location eastus
Options
-h, --help Gets help for set.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
Back to top
azd config unset
Unsets a configuration
Synopsis
Removes a configuration in the configuration path.
The default value of the config directory is:
$HOME/.azd on Linux and MacOS
%USERPROFILE%.azd on Windows
The configuration directory can be overridden by specifying a path in the AZD_CONFIG_DIR environment variable.
azd config unset <path> [flags]
Examples
azd config unset defaults.location
Options
-h, --help Gets help for unset.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd config: Manage Azure Developer CLI configuration
Back to top
azd deploy
Deploy the application's code to Azure.
Synopsis
Deploy the application's code to Azure.
When no --service value is specified, all services in the azure.yaml file (found in the root of your project) are deployed.
Examples:
azd deploy
azd deploy --service api
azd deploy --service web
After the deployment is complete, the endpoint is printed. To start the service, select the endpoint or paste it in a browser.
azd deploy [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for deploy.
-o, --output string The output format (the supported formats are json, none). (default "none")
--service string Deploys a specific service (when the string is unspecified, all services that are listed in the azure.yaml file are deployed).
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd down
Delete Azure resources for an application.
azd down [flags]
Options
-e, --environment string The name of the environment to use.
--force Does not require confirmation before it deletes resources.
-h, --help Gets help for down.
-o, --output string The output format (the supported formats are json, none). (default "none")
--purge Does not require confirmation before it permanently deletes resources that are soft-deleted by default (for example, key vaults).
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd env
Manage environments.
Synopsis
Manage environments.
With this command group, you can create a new environment or get, set, and list your application environments. An application can have multiple environments (for example, dev, test, prod), each with a different configuration (that is, connectivity information) for accessing Azure resources.
You can find all environment configurations under the .azure<environment-name> folder. The environment name is stored as the AZURE_ENV_NAME environment variable in the .azure<environment-name>\folder.env file.
Options
-h, --help Gets help for env.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
Back to top
azd env get-values
Get all environment values.
azd env get-values [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for get-values.
-o, --output string The output format (the supported formats are json, dotenv). (default "dotenv")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd env list
List environments
azd env list [flags]
Options
-h, --help Gets help for list.
-o, --output string The output format (the supported formats are json, table). (default "table")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd env new
Create a new environment.
azd env new <environment> [flags]
Options
-h, --help Gets help for new.
-l, --location string Azure location for the new environment
--subscription string Name or ID of an Azure subscription to use for the new environment
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd env refresh
Refresh environment settings by using information from a previous infrastructure provision.
azd env refresh [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for refresh.
-o, --output string The output format (the supported formats are json, none). (default "none")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd env select
Set the default environment.
azd env select <environment> [flags]
Options
-h, --help Gets help for select.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd env set
Set a value in the environment.
azd env set <key> <value> [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for set.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd env: Manage environments.
Back to top
azd infra
Manage Azure resources.
Options
-h, --help Gets help for infra.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd infra create: Create Azure resources for an application.
azd infra delete: Delete Azure resources for an application.
Back to top
azd infra create
Create Azure resources for an application.
azd infra create [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for create.
--no-progress Suppresses progress information.
-o, --output string The output format (the supported formats are json, none). (default "none")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd infra: Manage Azure resources.
Back to top
azd infra delete
Delete Azure resources for an application.
azd infra delete [flags]
Options
-e, --environment string The name of the environment to use.
--force Does not require confirmation before it deletes resources.
-h, --help Gets help for delete.
--purge Does not require confirmation before it permanently deletes resources that are soft-deleted by default (for example, key vaults).
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd infra: Manage Azure resources.
Back to top
azd init
Initialize a new application.
Synopsis
Initialize a new application.
When no template is supplied, you can optionally select an Azure Developer CLI template for cloning. Otherwise, azd init initializes the current directory and creates resources so that your project is compatible with Azure Developer CLI.
When a template is provided, the sample code is cloned to the current directory.
azd init [flags]
Options
-b, --branch string The template branch to initialize from.
-e, --environment string The name of the environment to use.
-h, --help Gets help for init.
-l, --location string Azure location for the new environment
--subscription string Name or ID of an Azure subscription to use for the new environment
-t, --template string The template to use when you initialize the project. You can use Full URI, <owner>/<repository>, or <repository> if it's part of the azure-samples organization.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd login
Log in to Azure.
Synopsis
Log in to Azure.
When run without any arguments, log in interactively using a browser. To log in using a device code, pass
--device-code.
To log in as a service principal, pass --client-id and --tenant-id as well as one of --client-secret,
--client-certificate, --client-credential or --client-credential-provider.
azd login [flags]
Options
--check-status Checks the log-in status instead of logging in.
--client-certificate string The path to the client certificate for the service principal to authenticate with.
--client-id string The client id for the service principal to authenticate with.
--client-secret string The client secret for the service principal to authenticate with. Set to the empty string to read the value from the console.
--federated-credential string The federated token for the service principal to authenticate with. Set to the empty string to read the value from the console.
--federated-credential-provider string The provider to use to acquire a federated token to authenticate with.
-h, --help Gets help for login.
-o, --output string The output format (the supported formats are json, none). (default "none")
--redirect-port int Choose the port to be used as part of the redirect URI during interactive login.
--tenant-id string The tenant id for the service principal to authenticate with.
--use-device-code When true, log in by using a device code instead of a browser.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd logout
Log out of Azure
Synopsis
Log out of Azure
azd logout [flags]
Options
-h, --help Gets help for logout.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd monitor
Monitor a deployed application.
Synopsis
Monitor a deployed application.
Examples:
azd monitor --overview
azd monitor -βlive
azd monitor --logs
For more information, go to https://aka.ms/azure-dev/monitor.
azd monitor [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for monitor.
--live Open a browser to Application Insights Live Metrics. Live Metrics is currently not supported for Python applications.
--logs Open a browser to Application Insights Logs.
--overview Open a browser to Application Insights Overview Dashboard.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd pipeline
Manage GitHub Actions pipelines.
Synopsis
Manage GitHub Actions pipelines.
The Azure Developer CLI template includes a GitHub Actions pipeline configuration file (in the .github/workflows folder) that deploys your application whenever code is pushed to the main branch.
For more information, go to https://aka.ms/azure-dev/pipeline.
Options
-h, --help Gets help for pipeline.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd pipeline config: Create and configure your deployment pipeline by using GitHub Actions or Azure Pipelines.
Back to top
azd pipeline config
Create and configure your deployment pipeline by using GitHub Actions or Azure Pipelines.
Synopsis
Create and configure your deployment pipeline by using GitHub Actions or Azure Pipelines.
For more information, go to https://aka.ms/azure-dev/pipeline.
azd pipeline config [flags]
Options
--auth-type string The authentication type used between the pipeline provider and Azure for deployment (Only valid for GitHub provider)
-e, --environment string The name of the environment to use.
-h, --help Gets help for config.
--principal-name string The name of the service principal to use to grant access to Azure resources as part of the pipeline.
--principal-role string The role to assign to the service principal. (default "contributor")
--provider string The pipeline provider to use (github for Github Actions and azdo for Azure Pipelines). (default "github")
--remote-name string The name of the git remote to configure the pipeline to run on. (default "origin")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd pipeline: Manage GitHub Actions pipelines.
Back to top
azd provision
Provision the Azure resources for an application.
Synopsis
Provision the Azure resources for an application.
The command prompts you for the following:
Environment name: The name of your environment.
Azure location: The Azure location where your resources will be deployed.
Azure subscription: The Azure subscription where your resources will be deployed.
Depending on what Azure resources are created, running this command might take a while. To view progress, go to the Azure portal and search for the resource group that contains your environment name.
azd provision [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for provision.
--no-progress Suppresses progress information.
-o, --output string The output format (the supported formats are json, none). (default "none")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd restore
Restore application dependencies.
Synopsis
Restore application dependencies.
Run this command to download and install all the required libraries so that you can build, run, and debug the application locally.
For the best local run and debug experience, go to https://aka.ms/azure-dev/vscode to learn how to use the Visual Studio Code extension.
azd restore [flags]
Options
-e, --environment string The name of the environment to use.
-h, --help Gets help for restore.
--service string Restores a specific service (when the string is unspecified, all services that are listed in the azure.yaml file are restored).
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd template
Manage templates.
Options
-h, --help Gets help for template.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd template list: List templates.
azd template show: Show the template details.
Back to top
azd template list
List templates.
azd template list [flags]
Options
-h, --help Gets help for list.
-o, --output string The output format (the supported formats are json, table). (default "table")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd template: Manage templates.
Back to top
azd template show
Show the template details.
azd template show <template> [flags]
Options
-h, --help Gets help for show.
-o, --output string The output format (the supported formats are json, table). (default "table")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
azd template: Manage templates.
Back to top
azd up
Initialize application, provision Azure resources, and deploy your project with a single command.
Synopsis
Initialize the project (if the project folder has not been initialized or cloned from a template), provision Azure resources, and deploy your project with a single command.
This command executes the following in one step:
azd init
azd provision
azd deploy
When no template is supplied, you can optionally select an Azure Developer CLI template for cloning. Otherwise, running azd up initializes the current directory so that your project is compatible with Azure Developer CLI.
azd up [flags]
Options
-b, --branch string The template branch to initialize from.
-e, --environment string The name of the environment to use.
-h, --help Gets help for up.
-l, --location string Azure location for the new environment
--no-progress Suppresses progress information.
-o, --output string The output format (the supported formats are json, none). (default "none")
--service string Deploys a specific service (when the string is unspecified, all services that are listed in the azure.yaml file are deployed).
--subscription string Name or ID of an Azure subscription to use for the new environment
-t, --template string The template to use when you initialize the project. You can use Full URI, <owner>/<repository>, or <repository> if it's part of the azure-samples organization.
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
azd version
Print the version number of Azure Developer CLI.
azd version [flags]
Options
-h, --help Gets help for version.
-o, --output string The output format (the supported formats are json, none). (default "none")
Options inherited from parent commands
-C, --cwd string Sets the current working directory.
--debug Enables debugging and diagnostics logging.
--no-prompt Accepts the default value instead of prompting, or it fails if there is no default.
See also
Back to top
|
2025-04-01T06:36:45.007411
| 2022-08-25T14:57:47
|
1351038098
|
{
"authors": [
"Ultre00",
"brettsam",
"kshyju"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:311",
"repo": "Azure/azure-functions-dotnet-worker",
"url": "https://github.com/Azure/azure-functions-dotnet-worker/issues/1005"
}
|
gharchive/issue
|
[HttpTrigger] Receiving 403 when not sending Expect: 100-continue
We are hosting an Azure Function V4 with runtime type dotnet-isolated.
The package versions are:
Microsoft.Azure.Functions.Worker "1.6.0"
Microsoft.Azure.Functions.Worker.Extensions.Http "3.0.13"
When the Function is deployed to Azure we are facing some issues with larger requests where the request body is over let's say 100 kb. We immediately get a 403 error. This doesn't even log a requests so we don't have any logging we can view. The only way we get it to work is by adding a Expect: 100-continue header. However we don't expect our users to always send this. How can we disable this feature?
@Ultre00 Is this a V3 app or V4? You can check the AzureFunctionsVersion element value in your .csproj file. Do you have a minimal repro you could share? (A github repo will be great)
@kshyju as mentioned this is a V4 function. I also figured out when this happens. It is only in combination with mTLS where the option "clientCertEnabled" is set to true. For now I simply set this to false since we are no longer using mTLS here.
@Ultre00 -- where did you have these configurations set? Is this somewhere in App Service?
@brettsam yes you can set that in the Function App. If you use an ARM template the properties are called
clientCertEnabled in combination with clientCertMode but if you use the UI then you can also do this under configuration > general settings > Client certificate mode
|
2025-04-01T06:36:45.013400
| 2021-04-21T14:13:59
|
863938687
|
{
"authors": [
"fabiocav",
"ondrejdurica"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:312",
"repo": "Azure/azure-functions-dotnet-worker",
"url": "https://github.com/Azure/azure-functions-dotnet-worker/issues/406"
}
|
gharchive/issue
|
Not able to debug AzF
Hi, I'm not able to debug. Once I attached written PID, I'm getting error on System.PrivateCore.Lib Can't write the message because the call is complete.
When I look inside, I'm gettting this:
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.GetResult()
at Microsoft.Azure.Functions.Worker.GrpcWorker.d__14.MoveNext() in D:\a\1\s\src\DotNetWorker.Grpc\GrpcWorker.cs:line 86
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.GetResult()
at Microsoft.Azure.Functions.Worker.WorkerHostedService.d__3.MoveNext() in D:\a\1\s\src\DotNetWorker.Core\WorkerHostedService.cs:line 27
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at Microsoft.Extensions.Hosting.Internal.Host.d__9.MoveNext()
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.ConfiguredTaskAwaitable.ConfiguredTaskAwaiter.GetResult()
at Microsoft.Extensions.Hosting.HostingAbstractionsHostExtensions.d__4.MoveNext()
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at Microsoft.Extensions.Hosting.HostingAbstractionsHostExtensions.d__4.MoveNext()
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.GetResult()
at xxx.Program.d__0.MoveNext() in
Do you know, what could be an issue?
@ondrejdurica can you share more details on how you're launching the worker? Is this from VS? Direct command line? If from VS, what does your project configuration look like for the debugging properties?
Thanks!
|
2025-04-01T06:36:45.059688
| 2022-06-08T13:34:29
|
1264762713
|
{
"authors": [
"MrVhek",
"anthonyvercolano",
"cartertinney",
"raowaqas72"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:313",
"repo": "Azure/azure-iot-sdk-python",
"url": "https://github.com/Azure/azure-iot-sdk-python/issues/1009"
}
|
gharchive/issue
|
create_from_edge_environment does not send messages to IoT Hub
Hello,
I've been facing a weird issue using:
iotedge 1.2.10
edgeAgent/edgeHub <IP_ADDRESS>800403
azure-iot-device 2.11.0
When I create the client module using create_from_edge_environment function within my IoT Edge Module, the module is able to correctly push and pull twin, but any message sent is never received on the hub.
If I use the from_connection_string and use a connection string within the same Docker, it works...
I normally have a function app that will use the fallback route to receive messages sent.
Any way to investiguate it further on IoT Hub side ? Or some additional logs to investigate on python side ?
Thanks
@MrVhek have u raised this issue in the Edge repo?
To get more debug logs add:
import logging
logging.basicConfig(level=logging.DEBUG)
To this at the beginning of your application
@MrVhek have u raised this issue in the Edge repo?
Not yet, not sure if this can be related to your SDK or to Azure IoT Edge directly ?
To get more debug logs add:
import logging logging.basicConfig(level=logging.DEBUG)
At the beginning of your application.
Thanks, I've added it but don't see more python logs than those that I posted above.
I need to see if I can have more logs on Azure side.
i am facing same problem
@MrVhek can you try reverting to IoT Edge 1.1? There are known issues with using Edge 1.2 with the SDK.
Sorry for the late reply, I finally found the issue, it was related to the routing of send_message.
Be sure to not have an output after your module name (here GWApp or ToolsApp).
|
2025-04-01T06:36:45.061183
| 2021-12-10T20:58:31
|
1077223829
|
{
"authors": [
"cartertinney"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:314",
"repo": "Azure/azure-iot-sdk-python",
"url": "https://github.com/Azure/azure-iot-sdk-python/pull/920"
}
|
gharchive/pull-request
|
chore: Updated language classifiers
Formally indicated support for Python 3.9 and 3.10
Formally indicated support for Python 3.9 and 3.10
|
2025-04-01T06:36:45.151597
| 2022-05-22T17:10:44
|
1244313941
|
{
"authors": [
"iwate",
"justinpenguin45",
"navba-MSFT",
"stephanwehr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:315",
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/issues/19173"
}
|
gharchive/issue
|
Container Instances should support environment variable names containing dot '.'
System Info
OS: Windows 10
Shell: Ubuntu 20.04 on Windows
Azure CLI: 2.36.0
Expected Behavior
Executing the following Azure CLI command should create a container instance running Elasticsearch:
az container create --image elasticsearch:7.3.2 --name es -g test-rg --memory 2 --cpu 1 --ports 9200 --environment-variables "discovery.type=single-node"
Using environment variable names containing dot '.' should work.
Current Behavior
Executing the Azure CLI command exit with an error:
(InvalidContainerEnvironmentVariable) The environment variable name in container 'es' of container group 'es' is invalid. A valid environment variable name must start with alphabetic character or '_', followed by a string of alphanumeric characters or '_' (e.g. 'my_name', or 'MY_NAME', or 'MyName').
Code: InvalidContainerEnvironmentVariable
Message: The environment variable name in container 'es' of container group 'es' is invalid. A valid environment variable name must start with alphabetic character or '_', followed by a string of alphanumeric characters or '_' (e.g. 'my_name', or 'MY_NAME', or 'MyName').
@stephanwehr Thanks for reaching out to us and sharing this feedback. I was able to reproduce this issue from AzCLI as well as Powershell. Post isolating this issue, I found that this issue isn't related to the Container Instance but the underlying image OS type.
The bash manual mentions this clearly (here)
name
A word consisting solely of letters, numbers, and underscores, and beginning with a letter or underscore. Names are used as shell variable and function names. Also referred to as an identifier.
The elasticSearch image seems to be using Linux OSType. So the fix / workaround has to come from the OS side and not from the AzCLI / REST API Spec. Hope this helps.
@navba-MSFT you are correct, Bash can't but Elasticsearch can.
With the "env" command it is possible to set environment variables containing special characters.
Here is an example:
# env -i "cluster.initial_master_nodes=elelastico" "my:test:var=value" /bin/bash -c "env"
my:test:var=value
cluster.initial_master_nodes=elelastico
PWD=/root
SHLVL=0
_=/usr/bin/env
In Docker compose the environment can be set like this:
version: '2.2'
services:
elasticsearch:
image: elasticsearch:7.3.2
environment:
- "ES_JAVA_OPTS=-Xms512m -Xmx512m"
- discovery.type=single-node
ports:
- 9200:9200
I've dug a little bit deeper on this issues and it seems Linux executes bash and other programs with execve().
With environ() an initial environment can be set.
How Linux execute programs can be seen using "strace":
$ strace env -i "my:test:var=some text" /bin/bash -c ":" 2>&1 | grep "execve"
execve("/usr/bin/env", ["env", "-i", "my:test:var=some text", "/bin/bash", "-c", ":"], 0x7ffd31dcb018 /* 21 vars */) = 0
execve("/bin/bash", ["/bin/bash", "-c", ":"], 0x5577895c20f0 /* 1 var */) = 0
As I understand, on Linux container environments like Docker for example set the initial environment before executing the 1st program inside the container. There is a process environment and bash environment which both can be retrieved using commands like "env" or "printenv".
I assume referring to this error message that the API performs a syntax check on variable names which should not be there:
{"error":{"code":"InvalidContainerEnvironmentVariable","message":"The environment variable name in container 'es' of container group 'es' is invalid. A valid environment variable name must start with alphabetic character or '_', followed by a string of alphanumeric characters or '_' (e.g. 'my_name', or 'MY_NAME', or 'MyName')."}}
@stephanwehr I just wanted to confirm if you are able to get this working with elasticsearch image with Linux OStype ? Awaiting your reply.
@navba-MSFT I'm not able to get Elasticsearch to work with Linux OStype because environment variables containing dot '.' are not supported.
@stephanwehr AFAIK, this ask is something which violates the IEEE standard for environment variables.
Environment variable names used by the utilities in the Shell and Utilities volume of IEEE Std 1003.1-2001 consist solely of uppercase letters, digits, and the '_' (underscore)
Refer this.
@navba-MSFT as stated this applies to the Shell and Utilities and not necessarily to other programs like Elasticsearch.
There is no hard requirement under Linux or inside a Docker container to use a shell to execute programs.
Linux itself allows variable names of type char:
#include <stdlib.h>
int setenv(const char *name, const char *value, int overwrite);
int unsetenv(const char *name);
Reference
@stephanwehr Thanks for clarifying. I have added the Service Team to look into this feature request ask.
I've tried hosting ElasticSearch on ACI and got this issue.
Are there any updates?
Same for me tried to use ElasticSearch with ACI and I get the same error with dotted env variable names. Why is this closed? There is no fix or workaround
|
2025-04-01T06:36:45.159719
| 2018-12-04T12:38:55
|
496459502
|
{
"authors": [
"PaulVrugt",
"m1dst",
"rudfoss",
"simonethreechillies"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:316",
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/issues/7276"
}
|
gharchive/issue
|
webapp:New-AzResourceGroupDeployment for web site with slot fails if template contains "cloningInfo"
Description
I'm trying to deploy a web app with two slots (production, deploytarget). Everything works fine the first time, but subsequent runs with the web app already present fail with an error indicating the site already exists. I was under the impression that ARM templates should always be able to run even if the resources already exist? If not is there a quick way to tell ARM that it should not touch the slot if it exists?
The problem goes away if I remove the "cloningInfo" options from the template. It appears there is a problem with cloning onto an existing app service.
"cloningInfo": {
"sourceWebAppId": "[resourceId('Microsoft.Web/sites', variables('webAppName'))]",
"overwrite": true
}
Script/Steps for Reproduction
New-AzResourceGroupDeployment # with provided template
This is the template that is used:
{
"$schema": "https://schema.management.azure.com/schemas/2015-01-01/deploymentTemplate.json#",
"contentVersion": "<IP_ADDRESS>",
"parameters": {
"environment": {
"type": "string",
"allowedValues": [
"devtest",
"test",
"stag",
"prod"
]
},
"skuName": {
"type": "string",
"defaultValue": "S1"
},
"skuCapacity": {
"type": "int",
"defaultValue": 1
},
"nodeVersion": {
"type": "string",
"defaultValue": "10.6.0"
},
"location": {
"type": "string",
"defaultValue": "[resourceGroup().location]"
},
"deployTargetSlotName": {
"type": "string",
"defaultValue": "deploytarget"
}
},
"variables": {
"appServiceName": "[concat('TestAppServiceName-ASP-', parameters('environment'))]",
"webAppName": "[concat('TestAppServiceName-APP-', parameters('environment'))]"
},
"resources": [
{
"name": "[variables('appServiceName')]",
"type": "Microsoft.Web/serverfarms",
"kind": "app",
"sku": {
"name": "[parameters('skuName')]",
"capacity": "[parameters('skuCapacity')]"
},
"apiVersion": "2016-09-01",
"location": "[parameters('location')]",
"scale": null,
"properties": {
"name": "[variables('appServiceName')]"
}
},
{
"name": "[variables('webAppName')]",
"type": "Microsoft.Web/sites",
"apiVersion": "2018-02-01",
"location": "[parameters('location')]",
"tags": {
"displayName": "[variables('webAppName')]"
},
"properties": {
"name": "[variables('webAppName')]",
"serverFarmId": "[resourceId('Microsoft.Web/serverfarms', variables('appServiceName'))]",
"clientAffinityEnabled": false,
"httpsOnly": true,
"siteConfig": {
"alwaysOn": true,
"phpVersion": ""
}
},
"dependsOn": [
"[variables('appServiceName')]"
],
"resources": [
{
"name": "[parameters('deployTargetSlotName')]",
"type": "slots",
"apiVersion": "2018-02-01",
"dependsOn": [
"[variables('webAppName')]"
],
"location": "[parameters('location')]",
"properties": {
"enabled": false,
"serverFarmId": "[resourceId('Microsoft.Web/serverfarms', variables('appServiceName'))]",
"clientAffinityEnabled": false,
"httpsOnly": true,
"cloningInfo": {
"sourceWebAppId": "[resourceId('Microsoft.Web/sites', variables('webAppName'))]",
"overwrite": true
}
}
}
]
}
],
"outputs": {
"webappName": {
"type": "string",
"value": "[variables('webAppName')]"
}
}
}
Module Version
ModuleType Version Name ExportedCommands
---------- ------- ---- ----------------
Script 0.6.1 Az.Profile {Add-AzEnvironment, Clear-AzContext, Clear-AzDefault, Connect-AzAccount...}
Script 0.6.1 Az.Storage {Add-AzRmStorageContainerLegalHold, Add-AzStorageAccountNetworkRule, Disable-AzStorageDeleteRetentionPolicy, Enable-AzStorageDeleteRetentionPolicy...}
Manifest <IP_ADDRESS> Microsoft.PowerShell.Management {Add-Computer, Add-Content, Checkpoint-Computer, Clear-Content...}
Manifest <IP_ADDRESS> Microsoft.PowerShell.Security {ConvertFrom-SecureString, ConvertTo-SecureString, Get-Acl, Get-AuthenticodeSignature...}
Manifest <IP_ADDRESS> Microsoft.PowerShell.Utility {Add-Member, Add-Type, Clear-Variable, Compare-Object...}
Manifest <IP_ADDRESS> Microsoft.WSMan.Management {Connect-WSMan, Disable-WSManCredSSP, Disconnect-WSMan, Enable-WSManCredSSP...}
Script 1.2 PSReadline {Get-PSReadlineKeyHandler, Get-PSReadlineOption, Remove-PSReadlineKeyHandler, Set-PSReadlineKeyHandler...}
Environment Data
Name Value
---- -----
PSVersion 5.1.17134.407
PSEdition Desktop
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...}
BuildVersion 10.0.17134.407
CLRVersion 4.0.30319.42000
WSManStackVersion 3.0
PSRemotingProtocolVersion 2.3
SerializationVersion <IP_ADDRESS>
Debug Output
13:23:59 - Resource Microsoft.Web/sites/slots 'TestAppServiceName-ASP-devtest/deploytarget' failed with message '{
"status": "Failed",
"error": {
"code": "ResourceDeploymentFailure",
"message": "The resource operation completed with terminal provisioning state 'Failed'.",
"details": [
{
"message": "Conflict\r\nDetail: Site with name TestAppServiceName-ASP-devtest(deploytarget) already exists.\r\nExtendedCode: 04008\r\n at Microsoft.Web.Hosting.Administration.Csm.Common.Clients.GeoClientBase.GetResponse[T](HttpWebRequest request, Boolean reThrowOnHostUnavailability, HttpStatusCode[]
supressLogForHttpCodes)\r\n at Microsoft.Web.Hosting.Administration.GeoScale.Sql.WebSiteCloneManager.StartSiteRestore(SiteClone clone, Operation cloneOperation)\r\n"
}
]
}
}'
Is there any update in regards to this issue?
@panchagnula so its been over 2 years now that this bug was reported. You reassigned this over a year ago, but nothing happened.
@zikalino any update?
@maggiepint you maybe?
Can anyone from microsoft please give a status update?
@rudfoss did you ever find a workaround?
this issue seems already fixed
Is this issue ever going to be fixed? It is impossible to redeploy a template when using cloningInfo unless you know ahead of time if the resource already exists and set the overwrite property accordingly. This goes against the whole premise of ARM.
|
2025-04-01T06:36:45.175375
| 2022-05-02T12:20:50
|
1222819881
|
{
"authors": [
"pratimaupadhyay02",
"xiaoxuqi-ms"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:317",
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/18875"
}
|
gharchive/pull-request
|
Fixing swagger correctness items RSV
MSFT employees can try out our new experience at OpenAPI Hub - one location for using our validation tools and finding your workflow.
Changelog
Add a changelog entry for this PR by answering the following questions:
What's the purpose of the update?
[ ] new service onboarding
[ ] new API version
[ ] update existing version for new feature
[ ] update existing version to fix swagger quality issue in s360
[ ] Other, please clarify
When are you targeting to deploy the new service/feature to public regions? Please provide the date or, if the date is not yet available, the month.
When do you expect to publish the swagger? Please provide date or, the the date is not yet available, the month.
If updating an existing version, please select the specific language SDKs and CLIs that must be refreshed after the swagger is published.
[ ] SDK of .NET (need service team to ensure code readiness)
[ ] SDK of Python
[ ] SDK of Java
[ ] SDK of Js
[ ] SDK of Go
[ ] PowerShell
[ ] CLI
[ ] Terraform
[ ] No refresh required for updates in this PR
Contribution checklist:
[ ] I commit to follow the Breaking Change Policy of "no breaking changes"
[ ] I have reviewed the documentation for the workflow.
[ ] Validation tools were run on swagger spec(s) and errors have all been fixed in this PR. How to fix?
If any further question about AME onboarding or validation tools, please view the FAQ.
ARM API Review Checklist
Applicability: :warning:
If your changes encompass only the following scenarios, you should SKIP this section, as these scenarios do not require ARM review.
Change to data plane APIs
Adding new properties
All removals
Otherwise your PR may be subject to ARM review requirements. Complete the following:
[ ] Check this box if any of the following apply to the PR so that label "WaitForARMFeedback" will be added automatically to begin ARM API Review. Failure to comply may result in delays to the manifest.
Adding a new service
Adding new API(s)
Adding a new API version
-[ ] To review changes efficiently, ensure you are using OpenAPIHub to initialize the PR for adding a new version. More details, refer to the wiki.
[ ] Ensure you've reviewed following guidelines including ARM resource provider contract and REST guidelines. Estimated time (4 hours). This is required before you can request review from ARM API Review board.
[ ] If you are blocked on ARM review and want to get the PR merged with urgency, please get the ARM oncall for reviews (RP Manifest Approvers team under Azure Resource Manager service) from IcM and reach out to them.
Breaking Change Review Checklist
If any of the following scenarios apply to the PR, request approval from the Breaking Change Review Board as defined in the Breaking Change Policy.
[ ] Removing API(s) in a stable version
[ ] Removing properties in a stable version
[ ] Removing API version(s) in a stable version
[ ] Updating API in a stable or public preview version with Breaking Change Validation errors
[ ] Updating API(s) in public preview over 1 year (refer to Retirement of Previews)
Action: to initiate an evaluation of the breaking change, create a new intake using the template for breaking changes. Addition details on the process and office hours are on the Breaking change Wiki.
Please follow the link to find more details on PR review process.
Hi @pratimaupadhyay02, please add the new properties to new version and fix the linter errors.
|
2025-04-01T06:36:45.193617
| 2022-05-11T20:17:51
|
1233138482
|
{
"authors": [
"ChongTang",
"heaths"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:318",
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/19030"
}
|
gharchive/pull-request
|
Copy QuestionAnswering definitions to common.json and refer them
MSFT employees can try out our new experience at OpenAPI Hub - one location for using our validation tools and finding your workflow.
Changelog
Add a changelog entry for this PR by answering the following questions:
What's the purpose of the update?
[ ] new service onboarding
[ ] new API version
[ ] update existing version for new feature
[ ] update existing version to fix swagger quality issue in s360
[ ] Other, please clarify
When are you targeting to deploy the new service/feature to public regions? Please provide the date or, if the date is not yet available, the month.
When do you expect to publish the swagger? Please provide date or, the the date is not yet available, the month.
If updating an existing version, please select the specific language SDKs and CLIs that must be refreshed after the swagger is published.
[ ] SDK of .NET (need service team to ensure code readiness)
[ ] SDK of Python
[ ] SDK of Java
[ ] SDK of Js
[ ] SDK of Go
[ ] PowerShell
[ ] CLI
[ ] Terraform
[ ] No refresh required for updates in this PR
Contribution checklist:
[ ] I commit to follow the Breaking Change Policy of "no breaking changes"
[ ] I have reviewed the documentation for the workflow.
[ ] Validation tools were run on swagger spec(s) and errors have all been fixed in this PR. How to fix?
If any further question about AME onboarding or validation tools, please view the FAQ.
ARM API Review Checklist
Applicability: :warning:
If your changes encompass only the following scenarios, you should SKIP this section, as these scenarios do not require ARM review.
Change to data plane APIs
Adding new properties
All removals
Otherwise your PR may be subject to ARM review requirements. Complete the following:
[ ] Check this box if any of the following appy to the PR so that the label "ARMReview" and "WaitForARMFeedback" will be added by bot to kick off ARM API Review. Missing to check this box in the following scenario may result in delays to the ARM manifest review and deployment.
Adding a new service
Adding new API(s)
Adding a new API version
-[ ] To review changes efficiently, ensure you are using OpenAPIHub to initialize the PR for adding a new version. More details, refer to the wiki.
[ ] Ensure you've reviewed following guidelines including ARM resource provider contract and REST guidelines. Estimated time (4 hours). This is required before you can request review from ARM API Review board.
[ ] If you are blocked on ARM review and want to get the PR merged with urgency, please get the ARM oncall for reviews (RP Manifest Approvers team under Azure Resource Manager service) from IcM and reach out to them.
Breaking Change Review Checklist
If any of the following scenarios apply to the PR, request approval from the Breaking Change Review Board as defined in the Breaking Change Policy.
[ ] Removing API(s) in a stable version
[ ] Removing properties in a stable version
[ ] Removing API version(s) in a stable version
[ ] Updating API in a stable or public preview version with Breaking Change Validation errors
[ ] Updating API(s) in public preview over 1 year (refer to Retirement of Previews)
Action: to initiate an evaluation of the breaking change, create a new intake using the template for breaking changes. Addition details on the process and office hours are on the Breaking change Wiki.
Please follow the link to find more details on PR review process.
@lmazuel, the only breaking change is our old friend the now-required error property. Could you merge regardless?
Hopefully once this GAs to main this specific issue will be a thing of the past.
@lmazuel, the only breaking change is our old friend the now-required error property. Could you merge regardless?
Hopefully once this GAs to main this specific issue will be a thing of the past.
Hi @lmazuel , can you help us merge it? Thanks!
|
2025-04-01T06:36:45.209048
| 2024-07-01T17:37:30
|
2384396672
|
{
"authors": [
"dpwatrous",
"raych1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:319",
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/29656"
}
|
gharchive/pull-request
|
[Batch] Switch contentLength from int64 to string
This is a breaking change, but accurately reflects what the Batch service has always returned for this property.
Data Plane API Specification Update Pull Request
[!TIP]
Overwhelmed by all this guidance? See the Getting help section at the bottom of this PR description.
PR review workflow diagram
Please understand this diagram before proceeding. It explains how to get your PR approved & merged.
API Info: The Basics
Most of the information about your service should be captured in the issue that serves as your API Spec engagement record.
Link to API Spec engagement record issue:
Is this review for (select one):
[ ] a private preview
[x] a public preview
[ ] GA release
Change Scope
This section will help us focus on the specific parts of your API that are new or have been modified. Please share a link to the design document for the new APIs, a link to the previous API Spec document (if applicable), and the root paths that have been updated.
Design Document:
Previous API Spec Doc:
Updated paths:
Viewing API changes
For convenient view of the API changes made by this PR, refer to the URLs provided in the table
in the Generated ApiView comment added to this PR. You can use ApiView to show API versions diff.
Suppressing failures
If one or multiple validation error/warning suppression(s) is detected in your PR, please follow the
Swagger-Suppression-Process
to get approval.
βGot questions? Need additional info?? We are here to help!
Contact us!
The Azure API Review Board is dedicated to helping you create amazing APIs. You can read about our mission and learn more about our process on our wiki.
π¬ Teams Channel
π email
Click here for links to tools, specs, guidelines & other good stuff
Tooling
Open API validation tools were run on this PR. Go here to see how to fix errors
Spectral Linting
Guidelines & Specifications
Azure REST API Guidelines
OpenAPI Style Guidelines
Azure Breaking Change Policy
Helpful Links
Schedule a data plane REST API spec review
Getting help
First, please carefully read through this PR description, from top to bottom.
If you don't have permissions to remove or add labels to the PR, request write access per aka.ms/azsdk/access#request-access-to-rest-api-or-sdk-repositories
To understand what you must do next to merge this PR, see the Next Steps to Merge comment. It will appear within few minutes of submitting this PR and will continue to be up-to-date with current PR state.
For guidance on fixing this PR CI check failures, see the hyperlinks provided in given failure
and https://aka.ms/ci-fix.
If the PR CI checks appear to be stuck in queued state, please add a comment with contents /azp run.
This should result in a new comment denoting a PR validation pipeline has started and the checks should be updated after few minutes.
If the help provided by the previous points is not enough, post to https://aka.ms/azsdk/support/specreview-channel and link to this PR.
Note that this TypeSpec is currently in a preview state.
/azp run
@raych1 is there anything blocking this from being merged? I'm not sure what the VersionReviewRequired tag is, but if it's regarding making changes to an existing API version, I think we should be fine. The TypeSpec in this PR is in a public preview state, not GA.
/azp run
/azp run
|
2025-04-01T06:36:45.213232
| 2019-03-22T18:01:01
|
424334140
|
{
"authors": [
"AutorestCI",
"azuresdkci",
"dsgouda",
"nschonni"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:320",
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/5458"
}
|
gharchive/pull-request
|
typo: consumption/resource-manager/Microsoft.Consumption
chages -> charges
Cylce -> Cycle
Can one of the admins verify this patch?
Automation for azure-sdk-for-js
Automation for azure-sdk-for-python
Automation for azure-sdk-for-ruby
Automation for azure-sdk-for-node
Automation for azure-sdk-for-go
Automation for azure-sdk-for-java
@sandeepkatumalla could you take a look since you worked on this version last?
I think the build failure is actually a message issue https://github.com/Azure/avocado/issues/11
|
2025-04-01T06:36:45.219036
| 2019-10-17T03:09:02
|
508205631
|
{
"authors": [
"AutorestCI",
"azuresdkci",
"seanmcc-msft"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:321",
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/7534"
}
|
gharchive/pull-request
|
Updated ContentMD5 parameter in DataLake swagger
Latest improvements:
MSFT employees can try out our new experience at OpenAPI Hub - one location for using our validation tools and finding your workflow.
Contribution checklist:
[ ] I have reviewed the documentation for the workflow.
[ ] Validation tools were run on swagger spec(s) and have all been fixed in this PR.
[ ] The OpenAPI Hub was used for checking validation status and next steps.
ARM API Review Checklist
[ ] Service team MUST add the "WaitForARMFeedback" label if the management plane API changes fall into one of the below categories.
adding/removing APIs.
adding/removing properties.
adding/removing API-version.
adding a new service in Azure.
Failure to comply may result in delays for manifest application. Note this does not apply to data plane APIs.
[ ] If you are blocked on ARM review and want to get the PR merged urgently, please get the ARM oncall for reviews (RP Manifest Approvers team under Azure Resource Manager service) from IcM and reach out to them.
Please follow the link to find more details on API review process.
Can one of the admins verify this patch?
Automation for azure-sdk-for-python
Automation for azure-sdk-for-go
|
2025-04-01T06:36:45.232302
| 2020-04-01T14:39:52
|
591968682
|
{
"authors": [
"AutorestCI",
"MilanBrkicFON",
"NullMDR",
"allenjzhang",
"azuresdkci",
"pilor"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:322",
"repo": "Azure/azure-rest-api-specs",
"url": "https://github.com/Azure/azure-rest-api-specs/pull/8911"
}
|
gharchive/pull-request
|
Managed Instance failover swagger
Latest improvements:
MSFT employees can try out our new experience at OpenAPI Hub - one location for using our validation tools and finding your workflow.
Contribution checklist:
[x] I have reviewed the documentation for the workflow.
[x] Validation tools were run on swagger spec(s) and have all been fixed in this PR.
[ ] The OpenAPI Hub was used for checking validation status and next steps.
ARM API Review Checklist
[x] Service team MUST add the "WaitForARMFeedback" label if the management plane API changes fall into one of the below categories.
adding/removing APIs.
adding/removing properties.
adding/removing API-version.
adding a new service in Azure.
Failure to comply may result in delays for manifest application. Note this does not apply to data plane APIs.
[ ] If you are blocked on ARM review and want to get the PR merged urgently, please get the ARM oncall for reviews (RP Manifest Approvers team under Azure Resource Manager service) from IcM and reach out to them.
Please follow the link to find more details on API review process.
/azp run automation - sdk
Can one of the admins verify this patch?
/azp run automation - sdk
/azp run automation - sdk
/azp run automation - sdk
/azp run automation - sdk
/azp run
@MilanBrkicFON Any update on this PR?
@NullMDR Last week changes landed in production, but I didn't have time to update this pr.
Now is ready for review!
/azp run
/azp run automation - sdk
Hi @NullMDR are you ok with the changes? Can we merge this PR?
/azp run
@MilanBrkicFON You need to wait for ARM review. I've emailed them.
/azp run automation - sdk
/azp run
@pilor @allenjzhang Can I have this PR reviewed please?
@pilor @allenjzhang Can I have this PR reviewed please?
API reviews are a weekly rotation described here. Please reach out to this week or next week's reviewer
@MilanBrkicFON, pretty straightforward PR, though I have noticed consistent pattern issues with SQL swagger files. Please take a look at these easy fixes. Will loop back to work with you to cleanup if these are manually created. If these are generated, please incorporate feedback and make appropriate changes to generate higher quality swagger in the future.
BTW, this is SDK team review. As per policy, ARM folks need to sign off for new APIs. Giving the low complexity of this PR, it should be a quick sign off. Please engage ARM folks per @pilor's instruction.
/azp run automation - sdk
/azp run automation - sdk
@allenjzhang Thanks for the review. I think I've addressed all of your comments. Can you please check again?
I will also update internal docs on ARM development to warn engineers what we are doing wrong and to avoid same mistakes in the future, but I think the right solution here would be working auto-generation script.
/azp run automation - sdk
/azp run
/azp run automation - sdk
/azp run automation - sdk
@pilor @allenjzhang @NullMDR Can someone merge this PR please? I am not authorized to do it.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.