Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
433,630 | 12,508,038,527 | IssuesEvent | 2020-06-02 14:58:49 | canonical-web-and-design/build.snapcraft.io | https://api.github.com/repos/canonical-web-and-design/build.snapcraft.io | closed | Should not build i386 snaps with base: core20 | Priority: High | ## Summary
With the release of Ubuntu 20.04, snaps that specify `base: core20` should not auto-trigger builds on the i386 architecture because that will not be possible since there is not an Ubuntu 20.04 for i386, and thus there is not a core20 base snap for i386.
## Process
Create a snapcraft.yaml with `base: core20` such as this one: https://github.com/anonymouse64/ubuntu-core-initramfs-snap
## Current and expected result
I expected build.snapcraft.io to not try to build for i386 automatically when specifying a snap with `base: core20` in the snapcraft.yaml of a GitHub issue.
Instead it triggered a i386 build, which failed on LP like this:
```
linux32: failed to execute snap: No such file or directory
Install failed
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/lpbuildd/target/build_snap.py", line 259, in run
self.install()
File "/usr/lib/python2.7/dist-packages/lpbuildd/target/build_snap.py", line 163, in install
snap_name])
File "/usr/lib/python2.7/dist-packages/lpbuildd/target/lxd.py", line 536, in run
subprocess.check_call(cmd, **kwargs)
File "/usr/lib/python2.7/subprocess.py", line 541, in check_call
raise CalledProcessError(retcode, cmd)
CalledProcessError: Command '['lxc', 'exec', 'lp-focal-i386', '--', 'linux32', 'snap', 'install', '--channel=stable', 'core20']' returned non-zero exit status 127
```
See also https://bugs.launchpad.net/launchpad/+bug/1862258 for LP side | 1.0 | Should not build i386 snaps with base: core20 - ## Summary
With the release of Ubuntu 20.04, snaps that specify `base: core20` should not auto-trigger builds on the i386 architecture because that will not be possible since there is not an Ubuntu 20.04 for i386, and thus there is not a core20 base snap for i386.
## Process
Create a snapcraft.yaml with `base: core20` such as this one: https://github.com/anonymouse64/ubuntu-core-initramfs-snap
## Current and expected result
I expected build.snapcraft.io to not try to build for i386 automatically when specifying a snap with `base: core20` in the snapcraft.yaml of a GitHub issue.
Instead it triggered a i386 build, which failed on LP like this:
```
linux32: failed to execute snap: No such file or directory
Install failed
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/lpbuildd/target/build_snap.py", line 259, in run
self.install()
File "/usr/lib/python2.7/dist-packages/lpbuildd/target/build_snap.py", line 163, in install
snap_name])
File "/usr/lib/python2.7/dist-packages/lpbuildd/target/lxd.py", line 536, in run
subprocess.check_call(cmd, **kwargs)
File "/usr/lib/python2.7/subprocess.py", line 541, in check_call
raise CalledProcessError(retcode, cmd)
CalledProcessError: Command '['lxc', 'exec', 'lp-focal-i386', '--', 'linux32', 'snap', 'install', '--channel=stable', 'core20']' returned non-zero exit status 127
```
See also https://bugs.launchpad.net/launchpad/+bug/1862258 for LP side | priority | should not build snaps with base summary with the release of ubuntu snaps that specify base should not auto trigger builds on the architecture because that will not be possible since there is not an ubuntu for and thus there is not a base snap for process create a snapcraft yaml with base such as this one current and expected result i expected build snapcraft io to not try to build for automatically when specifying a snap with base in the snapcraft yaml of a github issue instead it triggered a build which failed on lp like this failed to execute snap no such file or directory install failed traceback most recent call last file usr lib dist packages lpbuildd target build snap py line in run self install file usr lib dist packages lpbuildd target build snap py line in install snap name file usr lib dist packages lpbuildd target lxd py line in run subprocess check call cmd kwargs file usr lib subprocess py line in check call raise calledprocesserror retcode cmd calledprocesserror command returned non zero exit status see also for lp side | 1 |
61,674 | 25,598,549,355 | IssuesEvent | 2022-12-01 18:05:05 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | [Enhancement]: Add ECR Registry Permissions Resource | enhancement new-resource service/ecr | ### Description
I would like to be able to define the registry permissions JSON, which is necessary for cross account replication on the destination end. So https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/ecr_replication_configuration covers the origin part, but not the destination one.
### Affected Resource(s) and/or Data Source(s)
* aws_ecr_replication_configuration (not per se affected, but related)
### Potential Terraform Configuration
_No response_
### References
_No response_
### Would you like to implement a fix?
_No response_ | 1.0 | [Enhancement]: Add ECR Registry Permissions Resource - ### Description
I would like to be able to define the registry permissions JSON, which is necessary for cross account replication on the destination end. So https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/ecr_replication_configuration covers the origin part, but not the destination one.
### Affected Resource(s) and/or Data Source(s)
* aws_ecr_replication_configuration (not per se affected, but related)
### Potential Terraform Configuration
_No response_
### References
_No response_
### Would you like to implement a fix?
_No response_ | non_priority | add ecr registry permissions resource description i would like to be able to define the registry permissions json which is necessary for cross account replication on the destination end so covers the origin part but not the destination one affected resource s and or data source s aws ecr replication configuration not per se affected but related potential terraform configuration no response references no response would you like to implement a fix no response | 0 |
83,355 | 3,633,955,882 | IssuesEvent | 2016-02-11 16:21:36 | rsanchez-wsu/sp16-ceg3120 | https://api.github.com/repos/rsanchez-wsu/sp16-ceg3120 | closed | Fix checkstyle issues with team 6 branch. | priority-high state-inprogress team-6 | Code submission works on individual workstations but fails when Jenkins attempts to build it. | 1.0 | Fix checkstyle issues with team 6 branch. - Code submission works on individual workstations but fails when Jenkins attempts to build it. | priority | fix checkstyle issues with team branch code submission works on individual workstations but fails when jenkins attempts to build it | 1 |
6,720 | 6,609,341,277 | IssuesEvent | 2017-09-19 14:15:23 | ekylibre/ekylibre | https://api.github.com/repos/ekylibre/ekylibre | closed | Infinite map = infinite cultivable zone | Bug Cartography Security | When you create a new zone or modify an existent one, there is no limit on the size of it. The problem is the map is not a loop, it's a patern which repeat as many time as you want.
If the cultivable zone is 2x earth size there is a problem, an error is send and you have a zone which is at the same place on multiple zone on the map.
| True | Infinite map = infinite cultivable zone - When you create a new zone or modify an existent one, there is no limit on the size of it. The problem is the map is not a loop, it's a patern which repeat as many time as you want.
If the cultivable zone is 2x earth size there is a problem, an error is send and you have a zone which is at the same place on multiple zone on the map.
| non_priority | infinite map infinite cultivable zone when you create a new zone or modify an existent one there is no limit on the size of it the problem is the map is not a loop it s a patern which repeat as many time as you want if the cultivable zone is earth size there is a problem an error is send and you have a zone which is at the same place on multiple zone on the map | 0 |
65,988 | 16,518,013,592 | IssuesEvent | 2021-05-26 11:49:44 | angular/angular-cli | https://api.github.com/repos/angular/angular-cli | closed | ng e2e does not use --proxy-config for webdriver-manager update | comp: devkit/build-angular type: feature | <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request (mark with an `x`)
```
- [x] bug report -> please search issues before submitting
- [ ] feature request
```
### Versions.
<!--
Output from: `ng --version`.
If nothing, output from: `node --version` and `npm --version`.
Windows (7/8/10). Linux (incl. distribution). macOS (El Capitan? Sierra?)
-->
@angular/cli: 1.0.0
node: 6.10.1
os: win32 x64
@angular/common: 4.1.2
@angular/compiler: 4.1.2
@angular/core: 4.1.2
@angular/forms: 4.1.2
@angular/http: 4.1.2
@angular/platform-browser: 4.1.2
@angular/platform-browser-dynamic: 4.1.2
@angular/router: 4.1.2
@angular/cli: 1.0.0
@angular/compiler-cli: 4.1.2
### Repro steps.
<!--
Simple steps to reproduce this bug.
Please include: commands run, packages added, related code changes.
A link to a sample repo would help too.
-->
1. Must have an out of date web driver or delete it from `node_modules/protractor/node_modules/webdriver-manager/selenium/`. For example, if using Chrome webdriver, delete all `chrome*` files in this selenium folder.
2. Must be behind a proxy, but without HTTPS_PROXY or https_proxy environment variables set.
3. Using the proxy guide described in [the cli wiki](https://github.com/angular/angular-cli/wiki/stories-using-corporate-proxy), create a `proxy.conf.js` file to pass to `ng e2e`.
4. Run `ng e2e --proxy-config proxy.conf.js`.
If you are behind a proxy and https_proxy is not set, the command will fail after compiling the app. This is because by default, `ng e2e` will run `webdriver-manager update` to fetch the right browser drivers, but does not use the proxy config to do so.
### The log given by the failure.
<!-- Normally this include a stack trace and some more information. -->
```
$ ng e2e --proxy-config proxy.conf.js
** NG Live Development Server is running on http://localhost:49152 **
Hash: 95e0fe13c75ed4802b8b
Time: 17998ms
chunk {0} 0.chunk.js, 0.chunk.js.map 491 kB {3} [rendered]
chunk {1} polyfills.bundle.js, polyfills.bundle.js.map (polyfills) 237 kB {5} [initial] [rendered]
chunk {2} styles.bundle.js, styles.bundle.js.map (styles) 226 kB {5} [initial] [rendered]
chunk {3} main.bundle.js, main.bundle.js.map (main) 8.25 kB {4} [initial] [rendered]
chunk {4} vendor.bundle.js, vendor.bundle.js.map (vendor) 3.12 MB [initial] [rendered]
chunk {5} inline.bundle.js, inline.bundle.js.map (inline) 0 bytes [entry] [rendered]
webpack: Compiled successfully.
events.js:160
throw er; // Unhandled 'error' event
^
Error: connect ECONNREFUSED 172.217.8.16:443
at Object.exports._errnoException (util.js:1018:11)
at exports._exceptionWithHostPort (util.js:1041:20)
at TCPConnectWrap.afterConnect [as oncomplete] (net.js:1090:14)
```
### Desired functionality.
<!--
What would like to see implemented?
What is the usecase?
-->
`ng e2e` should pass along proxy configuration to webdriver-manager when the `--webdriver-update` flag is set to true. For webdriver-manager, the flag to use is `--proxy`.
### Mention any other details that might be useful.
<!-- Please include a link to the repo if this is related to an OSS project. -->
The code that calls webdriver-manager is at Line 65 in [e2e.ts](https://github.com/angular/angular-cli/blob/master/packages/%40angular/cli/tasks/e2e.ts#L65).
| 1.0 | ng e2e does not use --proxy-config for webdriver-manager update - <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request (mark with an `x`)
```
- [x] bug report -> please search issues before submitting
- [ ] feature request
```
### Versions.
<!--
Output from: `ng --version`.
If nothing, output from: `node --version` and `npm --version`.
Windows (7/8/10). Linux (incl. distribution). macOS (El Capitan? Sierra?)
-->
@angular/cli: 1.0.0
node: 6.10.1
os: win32 x64
@angular/common: 4.1.2
@angular/compiler: 4.1.2
@angular/core: 4.1.2
@angular/forms: 4.1.2
@angular/http: 4.1.2
@angular/platform-browser: 4.1.2
@angular/platform-browser-dynamic: 4.1.2
@angular/router: 4.1.2
@angular/cli: 1.0.0
@angular/compiler-cli: 4.1.2
### Repro steps.
<!--
Simple steps to reproduce this bug.
Please include: commands run, packages added, related code changes.
A link to a sample repo would help too.
-->
1. Must have an out of date web driver or delete it from `node_modules/protractor/node_modules/webdriver-manager/selenium/`. For example, if using Chrome webdriver, delete all `chrome*` files in this selenium folder.
2. Must be behind a proxy, but without HTTPS_PROXY or https_proxy environment variables set.
3. Using the proxy guide described in [the cli wiki](https://github.com/angular/angular-cli/wiki/stories-using-corporate-proxy), create a `proxy.conf.js` file to pass to `ng e2e`.
4. Run `ng e2e --proxy-config proxy.conf.js`.
If you are behind a proxy and https_proxy is not set, the command will fail after compiling the app. This is because by default, `ng e2e` will run `webdriver-manager update` to fetch the right browser drivers, but does not use the proxy config to do so.
### The log given by the failure.
<!-- Normally this include a stack trace and some more information. -->
```
$ ng e2e --proxy-config proxy.conf.js
** NG Live Development Server is running on http://localhost:49152 **
Hash: 95e0fe13c75ed4802b8b
Time: 17998ms
chunk {0} 0.chunk.js, 0.chunk.js.map 491 kB {3} [rendered]
chunk {1} polyfills.bundle.js, polyfills.bundle.js.map (polyfills) 237 kB {5} [initial] [rendered]
chunk {2} styles.bundle.js, styles.bundle.js.map (styles) 226 kB {5} [initial] [rendered]
chunk {3} main.bundle.js, main.bundle.js.map (main) 8.25 kB {4} [initial] [rendered]
chunk {4} vendor.bundle.js, vendor.bundle.js.map (vendor) 3.12 MB [initial] [rendered]
chunk {5} inline.bundle.js, inline.bundle.js.map (inline) 0 bytes [entry] [rendered]
webpack: Compiled successfully.
events.js:160
throw er; // Unhandled 'error' event
^
Error: connect ECONNREFUSED 172.217.8.16:443
at Object.exports._errnoException (util.js:1018:11)
at exports._exceptionWithHostPort (util.js:1041:20)
at TCPConnectWrap.afterConnect [as oncomplete] (net.js:1090:14)
```
### Desired functionality.
<!--
What would like to see implemented?
What is the usecase?
-->
`ng e2e` should pass along proxy configuration to webdriver-manager when the `--webdriver-update` flag is set to true. For webdriver-manager, the flag to use is `--proxy`.
### Mention any other details that might be useful.
<!-- Please include a link to the repo if this is related to an OSS project. -->
The code that calls webdriver-manager is at Line 65 in [e2e.ts](https://github.com/angular/angular-cli/blob/master/packages/%40angular/cli/tasks/e2e.ts#L65).
| non_priority | ng does not use proxy config for webdriver manager update if you don t fill out the following information your issue might be closed without investigating bug report or feature request mark with an x bug report please search issues before submitting feature request versions output from ng version if nothing output from node version and npm version windows linux incl distribution macos el capitan sierra angular cli node os angular common angular compiler angular core angular forms angular http angular platform browser angular platform browser dynamic angular router angular cli angular compiler cli repro steps simple steps to reproduce this bug please include commands run packages added related code changes a link to a sample repo would help too must have an out of date web driver or delete it from node modules protractor node modules webdriver manager selenium for example if using chrome webdriver delete all chrome files in this selenium folder must be behind a proxy but without https proxy or https proxy environment variables set using the proxy guide described in create a proxy conf js file to pass to ng run ng proxy config proxy conf js if you are behind a proxy and https proxy is not set the command will fail after compiling the app this is because by default ng will run webdriver manager update to fetch the right browser drivers but does not use the proxy config to do so the log given by the failure ng proxy config proxy conf js ng live development server is running on hash time chunk chunk js chunk js map kb chunk polyfills bundle js polyfills bundle js map polyfills kb chunk styles bundle js styles bundle js map styles kb chunk main bundle js main bundle js map main kb chunk vendor bundle js vendor bundle js map vendor mb chunk inline bundle js inline bundle js map inline bytes webpack compiled successfully events js throw er unhandled error event error connect econnrefused at object exports errnoexception util js at exports exceptionwithhostport util js at tcpconnectwrap afterconnect net js desired functionality what would like to see implemented what is the usecase ng should pass along proxy configuration to webdriver manager when the webdriver update flag is set to true for webdriver manager the flag to use is proxy mention any other details that might be useful the code that calls webdriver manager is at line in | 0 |
31,044 | 2,731,323,356 | IssuesEvent | 2015-04-16 19:41:20 | Theano/Theano | https://api.github.com/repos/Theano/Theano | opened | Sparse-aware addition for combining gradients | Low Priority Sparse | Reported in https://groups.google.com/d/topic/theano-users/aa-Ydpy6_0A/discussion
During gradient computation, when summing the contribution of different gradient path for the same variable, `tensor.add` can get called with one sparse and one dense input, which lead to a crash.
We could either:
- have an Op for that, that then gets optimized into sparse/sparse or dense/dense addition, or optimized away (if one of the inputs is all zeros, for instance)
- explicitly cast the sparse matrix into a dense one inside `theano.grad`, before the addition. I don't think we should fear a memory blow-up here, since we already have a dense matrix of the same dimension. | 1.0 | Sparse-aware addition for combining gradients - Reported in https://groups.google.com/d/topic/theano-users/aa-Ydpy6_0A/discussion
During gradient computation, when summing the contribution of different gradient path for the same variable, `tensor.add` can get called with one sparse and one dense input, which lead to a crash.
We could either:
- have an Op for that, that then gets optimized into sparse/sparse or dense/dense addition, or optimized away (if one of the inputs is all zeros, for instance)
- explicitly cast the sparse matrix into a dense one inside `theano.grad`, before the addition. I don't think we should fear a memory blow-up here, since we already have a dense matrix of the same dimension. | priority | sparse aware addition for combining gradients reported in during gradient computation when summing the contribution of different gradient path for the same variable tensor add can get called with one sparse and one dense input which lead to a crash we could either have an op for that that then gets optimized into sparse sparse or dense dense addition or optimized away if one of the inputs is all zeros for instance explicitly cast the sparse matrix into a dense one inside theano grad before the addition i don t think we should fear a memory blow up here since we already have a dense matrix of the same dimension | 1 |
56,283 | 23,743,020,559 | IssuesEvent | 2022-08-31 13:55:01 | miranda-ng/miranda-ng | https://api.github.com/repos/miranda-ng/miranda-ng | closed | VoiceService: некоторые надписи в окне вызова не переводятся | bug VoiceService | ERROR: type should be string, got "https://github.com/miranda-ng/miranda-ng/blob/master/plugins/VoiceService/src/VoiceCall.cpp#L222\r\n\r\n```\r\n\tcase VOICE_STATE_RINGING:\r\n\t\tincoming = true;\r\n\t\tSetCaption(L\"Incoming call\");\r\n\t\tm_btnAnswer.Enable(true);\r\n\t\tm_lblStatus.SetText(L\"Ringing\");\r\n\t\tSetWindowPos(GetHwnd(), HWND_TOPMOST, 0, 0, 0, 0, SWP_NOMOVE | SWP_NOSIZE | SWP_SHOWWINDOW);\r\n\t\tSetWindowPos(GetHwnd(), HWND_NOTOPMOST, 0, 0, 0, 0, SWP_NOMOVE | SWP_NOSIZE | SWP_SHOWWINDOW);\r\n\t\tbreak;\r\n\tcase VOICE_STATE_CALLING:\r\n\t\tincoming = false;\r\n\t\tSetCaption(L\"Outgoing call\");\r\n\t\tm_lblStatus.SetText(L\"Calling\");\r\n\t\tm_btnAnswer.Enable(false);\r\n\t\tbreak;\r\n\tcase VOICE_STATE_ON_HOLD:\r\n\t\tm_lblStatus.SetText(L\"Holded\");\r\n\t\tm_btnAnswer.Enable(true);\r\n\t\tm_btnAnswer.SetText(L\"Unhold\");\r\n\t\tbreak;\r\n\tcase VOICE_STATE_ENDED:\r\n\t\tm_calltimer.Stop();\r\n\t\tmir_snwprintf(text, _countof(text), L\"Call ended %s\", m_lblStatus.GetText());\r\n\t\tm_lblStatus.SetText(text);\r\n\t\tm_btnAnswer.Enable(false);\r\n\t\tm_btnDrop.SetText(L\"Close\");\r\n\t\tbreak;\r\n\tcase VOICE_STATE_BUSY:\r\n\t\tm_lblStatus.SetText(L\"Busy\");\r\n\t\tm_btnAnswer.Enable(false);\r\n\t\tm_btnDrop.SetText(L\"Close\");\r\n\t\tbreak;\r\n\tdefault:\r\n\t\tm_lblStatus.SetText(L\"Unknown state\");\r\n\t\tbreak;\r\n\t}\r\n```" | 1.0 | VoiceService: некоторые надписи в окне вызова не переводятся - https://github.com/miranda-ng/miranda-ng/blob/master/plugins/VoiceService/src/VoiceCall.cpp#L222
```
case VOICE_STATE_RINGING:
incoming = true;
SetCaption(L"Incoming call");
m_btnAnswer.Enable(true);
m_lblStatus.SetText(L"Ringing");
SetWindowPos(GetHwnd(), HWND_TOPMOST, 0, 0, 0, 0, SWP_NOMOVE | SWP_NOSIZE | SWP_SHOWWINDOW);
SetWindowPos(GetHwnd(), HWND_NOTOPMOST, 0, 0, 0, 0, SWP_NOMOVE | SWP_NOSIZE | SWP_SHOWWINDOW);
break;
case VOICE_STATE_CALLING:
incoming = false;
SetCaption(L"Outgoing call");
m_lblStatus.SetText(L"Calling");
m_btnAnswer.Enable(false);
break;
case VOICE_STATE_ON_HOLD:
m_lblStatus.SetText(L"Holded");
m_btnAnswer.Enable(true);
m_btnAnswer.SetText(L"Unhold");
break;
case VOICE_STATE_ENDED:
m_calltimer.Stop();
mir_snwprintf(text, _countof(text), L"Call ended %s", m_lblStatus.GetText());
m_lblStatus.SetText(text);
m_btnAnswer.Enable(false);
m_btnDrop.SetText(L"Close");
break;
case VOICE_STATE_BUSY:
m_lblStatus.SetText(L"Busy");
m_btnAnswer.Enable(false);
m_btnDrop.SetText(L"Close");
break;
default:
m_lblStatus.SetText(L"Unknown state");
break;
}
``` | non_priority | voiceservice некоторые надписи в окне вызова не переводятся case voice state ringing incoming true setcaption l incoming call m btnanswer enable true m lblstatus settext l ringing setwindowpos gethwnd hwnd topmost swp nomove swp nosize swp showwindow setwindowpos gethwnd hwnd notopmost swp nomove swp nosize swp showwindow break case voice state calling incoming false setcaption l outgoing call m lblstatus settext l calling m btnanswer enable false break case voice state on hold m lblstatus settext l holded m btnanswer enable true m btnanswer settext l unhold break case voice state ended m calltimer stop mir snwprintf text countof text l call ended s m lblstatus gettext m lblstatus settext text m btnanswer enable false m btndrop settext l close break case voice state busy m lblstatus settext l busy m btnanswer enable false m btndrop settext l close break default m lblstatus settext l unknown state break | 0 |
313,548 | 9,564,542,184 | IssuesEvent | 2019-05-05 04:40:06 | sevenzk/SJKTCI | https://api.github.com/repos/sevenzk/SJKTCI | closed | [Master Harga] Pembulatan 1000 ke atas atau ke bawah di Harga Setelah Pajak | Priority enhancement fixed | Ada pembulatan 1000 ke atas atau ke bawah (yang mendekati) di Harga Setelah Pajak seperti gambar di bawah:

| 1.0 | [Master Harga] Pembulatan 1000 ke atas atau ke bawah di Harga Setelah Pajak - Ada pembulatan 1000 ke atas atau ke bawah (yang mendekati) di Harga Setelah Pajak seperti gambar di bawah:

| priority | pembulatan ke atas atau ke bawah di harga setelah pajak ada pembulatan ke atas atau ke bawah yang mendekati di harga setelah pajak seperti gambar di bawah | 1 |
118,389 | 4,744,343,449 | IssuesEvent | 2016-10-21 00:29:54 | FeraGroup/FTCVortexScoreCounter | https://api.github.com/repos/FeraGroup/FTCVortexScoreCounter | closed | Large numbers do not fit | enhancement Low Priority | Any score higher then 99 will not fit completely in the box that displays the scores. This is only an issue while looking at the score using the smaller counter. | 1.0 | Large numbers do not fit - Any score higher then 99 will not fit completely in the box that displays the scores. This is only an issue while looking at the score using the smaller counter. | priority | large numbers do not fit any score higher then will not fit completely in the box that displays the scores this is only an issue while looking at the score using the smaller counter | 1 |
580,262 | 17,214,353,905 | IssuesEvent | 2021-07-19 09:33:46 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Pin the GitHub actions we use on this repo to a full length commit SHA | priority-3-normal status:blocked type:refactor | **What would you like Renovate to be able to do?**
<!-- Tell us what requirements you need solving, and be sure to mention too if this is part of any "bigger" problem you're trying to solve. -->
@rarkins and @viceice now that PR #10835 is merged, we can start thinking about pinning our GitHub Actions to the current full length Git commit SHA?
**Did you already have any implementation ideas?**
<!-- In case you've already dug into existing options or source code and have ideas, mention them here. Try to keep implementation ideas separate from *requirements* above -->
<!-- Please also mention here in case this is a feature you'd be interested in writing yourself, so you can be assigned it. -->
@viceice can you explain what pattern we need to follow to make things work properly?
I think you can use a comment to say what "tagged version" you're at???
As an example I've grabbed the commit to which `actions/setup-node@v2.2.0` points right now.
```yml
- name: Set up Node.js ${{ env.NODE_VERSION }}
uses: actions/setup-node@38d90ce44d5275ad62cc48384b3d8a58c500bb5f
with:
node-version: ${{ env.NODE_VERSION }}
cache: yarn
```
Or maybe we can add a comment which says what tag we're "following".
```yml
- name: Set up Node.js ${{ env.NODE_VERSION }}
uses: actions/setup-node@38d90ce44d5275ad62cc48384b3d8a58c500bb5f # setup-node: tag=v2
with:
node-version: ${{ env.NODE_VERSION }}
cache: yarn
```
We'll also need to update the `default.json` over on the `renovate/.github` repository, to make use of the new feature. | 1.0 | Pin the GitHub actions we use on this repo to a full length commit SHA - **What would you like Renovate to be able to do?**
<!-- Tell us what requirements you need solving, and be sure to mention too if this is part of any "bigger" problem you're trying to solve. -->
@rarkins and @viceice now that PR #10835 is merged, we can start thinking about pinning our GitHub Actions to the current full length Git commit SHA?
**Did you already have any implementation ideas?**
<!-- In case you've already dug into existing options or source code and have ideas, mention them here. Try to keep implementation ideas separate from *requirements* above -->
<!-- Please also mention here in case this is a feature you'd be interested in writing yourself, so you can be assigned it. -->
@viceice can you explain what pattern we need to follow to make things work properly?
I think you can use a comment to say what "tagged version" you're at???
As an example I've grabbed the commit to which `actions/setup-node@v2.2.0` points right now.
```yml
- name: Set up Node.js ${{ env.NODE_VERSION }}
uses: actions/setup-node@38d90ce44d5275ad62cc48384b3d8a58c500bb5f
with:
node-version: ${{ env.NODE_VERSION }}
cache: yarn
```
Or maybe we can add a comment which says what tag we're "following".
```yml
- name: Set up Node.js ${{ env.NODE_VERSION }}
uses: actions/setup-node@38d90ce44d5275ad62cc48384b3d8a58c500bb5f # setup-node: tag=v2
with:
node-version: ${{ env.NODE_VERSION }}
cache: yarn
```
We'll also need to update the `default.json` over on the `renovate/.github` repository, to make use of the new feature. | priority | pin the github actions we use on this repo to a full length commit sha what would you like renovate to be able to do rarkins and viceice now that pr is merged we can start thinking about pinning our github actions to the current full length git commit sha did you already have any implementation ideas viceice can you explain what pattern we need to follow to make things work properly i think you can use a comment to say what tagged version you re at as an example i ve grabbed the commit to which actions setup node points right now yml name set up node js env node version uses actions setup node with node version env node version cache yarn or maybe we can add a comment which says what tag we re following yml name set up node js env node version uses actions setup node setup node tag with node version env node version cache yarn we ll also need to update the default json over on the renovate github repository to make use of the new feature | 1 |
716,982 | 24,656,113,350 | IssuesEvent | 2022-10-17 23:47:14 | ApplETS/Notre-Dame | https://api.github.com/repos/ApplETS/Notre-Dame | closed | Golden files not updated - CI tests failing | bug CI priority: high | **Describe the bug**
Golden files have not been updated correctly in a recent PR and CI tests are failing.
The files that are throwing the error are `gradesDetailsView_1.png`, `gradesDetailsView_2.png`, `gradesDetailsView_evaluation_not_completed.png` (located in `goldenFiles/`)
**To Reproduce**
Steps to reproduce the behavior:
1. Run CI tests on any valid PR
2. Golden files will throw errors and the tests will fail
**Expected behavior**
Golden files tests should not fail.
| 1.0 | Golden files not updated - CI tests failing - **Describe the bug**
Golden files have not been updated correctly in a recent PR and CI tests are failing.
The files that are throwing the error are `gradesDetailsView_1.png`, `gradesDetailsView_2.png`, `gradesDetailsView_evaluation_not_completed.png` (located in `goldenFiles/`)
**To Reproduce**
Steps to reproduce the behavior:
1. Run CI tests on any valid PR
2. Golden files will throw errors and the tests will fail
**Expected behavior**
Golden files tests should not fail.
| priority | golden files not updated ci tests failing describe the bug golden files have not been updated correctly in a recent pr and ci tests are failing the files that are throwing the error are gradesdetailsview png gradesdetailsview png gradesdetailsview evaluation not completed png located in goldenfiles to reproduce steps to reproduce the behavior run ci tests on any valid pr golden files will throw errors and the tests will fail expected behavior golden files tests should not fail | 1 |
269,463 | 8,435,892,153 | IssuesEvent | 2018-10-17 14:12:52 | smartdevicelink/sdl_core | https://api.github.com/repos/smartdevicelink/sdl_core | closed | Adjust code to accomodate new JsonCPP version | Bug Contributor priority 1: High | ### Bug Report
Adjust code to accomodate new JsonCPP version
##### Expected Behavior
Need to upgrade the third-party JsonCpp library.
The SDL library is currently using an old release candidate version of JsonCpp (0.6.0-rc2). This should be updated to an actually released version.
##### OS & Version Information
* OS/Version:
* SDL Core Version:
* Testing Against: | 1.0 | Adjust code to accomodate new JsonCPP version - ### Bug Report
Adjust code to accomodate new JsonCPP version
##### Expected Behavior
Need to upgrade the third-party JsonCpp library.
The SDL library is currently using an old release candidate version of JsonCpp (0.6.0-rc2). This should be updated to an actually released version.
##### OS & Version Information
* OS/Version:
* SDL Core Version:
* Testing Against: | priority | adjust code to accomodate new jsoncpp version bug report adjust code to accomodate new jsoncpp version expected behavior need to upgrade the third party jsoncpp library the sdl library is currently using an old release candidate version of jsoncpp this should be updated to an actually released version os version information os version sdl core version testing against | 1 |
455,448 | 13,127,063,960 | IssuesEvent | 2020-08-06 09:40:16 | phovea/generator-phovea | https://api.github.com/repos/phovea/generator-phovea | opened | Update build.js after moving deploy scripts from app to product | priority: high type: bug | * Release number or git hash: 2feaac301b3ccbaffb9af5bbac124be3832218b1
* OS: Linux
* Environment (local or deployed): both
### Steps to reproduce
1. build a product containing more deployment configurations than web and api
### Observed behavior
* only web and api are accepted options for Dockerfiles (seee https://github.com/phovea/generator-phovea/blob/master/generators/init-product/templates/plain/build.js#L626)
* additional folders like dtiles or benchmark are ignored
### Expected behavior
* additional Dockerfiles are copied as well as long as they are available in templates/<name>/deploy/<name>
| 1.0 | Update build.js after moving deploy scripts from app to product - * Release number or git hash: 2feaac301b3ccbaffb9af5bbac124be3832218b1
* OS: Linux
* Environment (local or deployed): both
### Steps to reproduce
1. build a product containing more deployment configurations than web and api
### Observed behavior
* only web and api are accepted options for Dockerfiles (seee https://github.com/phovea/generator-phovea/blob/master/generators/init-product/templates/plain/build.js#L626)
* additional folders like dtiles or benchmark are ignored
### Expected behavior
* additional Dockerfiles are copied as well as long as they are available in templates/<name>/deploy/<name>
| priority | update build js after moving deploy scripts from app to product release number or git hash os linux environment local or deployed both steps to reproduce build a product containing more deployment configurations than web and api observed behavior only web and api are accepted options for dockerfiles seee additional folders like dtiles or benchmark are ignored expected behavior additional dockerfiles are copied as well as long as they are available in templates deploy | 1 |
450,454 | 31,925,465,711 | IssuesEvent | 2023-09-19 01:10:23 | ICEI-PUC-Minas-PMV-ADS/pmv-ads-2023-2-e3-proj-mov-t1-entre-time | https://api.github.com/repos/ICEI-PUC-Minas-PMV-ADS/pmv-ads-2023-2-e3-proj-mov-t1-entre-time | closed | Justificativa e Público alvo(01- Documentação de contexto)- H11a-ADS-CST | documentation | H11a-ADS-CST - Compreender os usuários e definir uma proposta de solução: definir o problema de forma clara e objetiva, apresentando os objetivos, a justificativa e a motivação da escolha. | 1.0 | Justificativa e Público alvo(01- Documentação de contexto)- H11a-ADS-CST - H11a-ADS-CST - Compreender os usuários e definir uma proposta de solução: definir o problema de forma clara e objetiva, apresentando os objetivos, a justificativa e a motivação da escolha. | non_priority | justificativa e público alvo documentação de contexto ads cst ads cst compreender os usuários e definir uma proposta de solução definir o problema de forma clara e objetiva apresentando os objetivos a justificativa e a motivação da escolha | 0 |
50,728 | 12,549,929,659 | IssuesEvent | 2020-06-06 09:07:38 | GoogleCloudPlatform/python-docs-samples | https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples | closed | appengine.flexible.tasks.snippets_test: test_pause_queue failed | buildcop: issue priority: p1 type: bug | This test failed!
To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop).
If I'm commenting on this issue too often, add the `buildcop: quiet` label and
I will stop commenting.
---
commit: cc68a07af4cab7b48233680996d2913fb0ba5e24
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/3ddf5a57-41f5-46d8-812c-bc99c4adc730), [Sponge](http://sponge2/3ddf5a57-41f5-46d8-812c-bc99c4adc730)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 57, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 826, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 729, in _end_unary_response_blocking
raise _InactiveRpcError(state)
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.NOT_FOUND
details = "Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize."
debug_error_string = "{"created":"@1591387480.150681677","description":"Error received from peer ipv4:172.217.212.95:443","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize.","grpc_status":5}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/appengine/flexible/tasks/snippets_test.py", line 94, in test_pause_queue
TEST_PROJECT_ID, TEST_LOCATION, QUEUE_NAME_1)
File "/workspace/appengine/flexible/tasks/snippets.py", line 218, in pause_queue
response = client.pause_queue(queue_path)
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/tasks_v2/gapic/cloud_tasks_client.py", line 832, in pause_queue
request, retry=retry, timeout=timeout, metadata=metadata
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 143, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 286, in retry_wrapped_func
on_error=on_error,
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 184, in retry_target
return target()
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/timeout.py", line 214, in func_with_timeout
return func(*args, **kwargs)
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 59, in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
File "<string>", line 3, in raise_from
google.api_core.exceptions.NotFound: 404 Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize.</pre></details> | 1.0 | appengine.flexible.tasks.snippets_test: test_pause_queue failed - This test failed!
To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop).
If I'm commenting on this issue too often, add the `buildcop: quiet` label and
I will stop commenting.
---
commit: cc68a07af4cab7b48233680996d2913fb0ba5e24
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/3ddf5a57-41f5-46d8-812c-bc99c4adc730), [Sponge](http://sponge2/3ddf5a57-41f5-46d8-812c-bc99c4adc730)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 57, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 826, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 729, in _end_unary_response_blocking
raise _InactiveRpcError(state)
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.NOT_FOUND
details = "Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize."
debug_error_string = "{"created":"@1591387480.150681677","description":"Error received from peer ipv4:172.217.212.95:443","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize.","grpc_status":5}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/appengine/flexible/tasks/snippets_test.py", line 94, in test_pause_queue
TEST_PROJECT_ID, TEST_LOCATION, QUEUE_NAME_1)
File "/workspace/appengine/flexible/tasks/snippets.py", line 218, in pause_queue
response = client.pause_queue(queue_path)
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/tasks_v2/gapic/cloud_tasks_client.py", line 832, in pause_queue
request, retry=retry, timeout=timeout, metadata=metadata
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 143, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 286, in retry_wrapped_func
on_error=on_error,
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 184, in retry_target
return target()
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/timeout.py", line 214, in func_with_timeout
return func(*args, **kwargs)
File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 59, in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
File "<string>", line 3, in raise_from
google.api_core.exceptions.NotFound: 404 Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize.</pre></details> | non_priority | appengine flexible tasks snippets test test pause queue failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace appengine flexible tasks nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace appengine flexible tasks nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace appengine flexible tasks nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode not found details queue does not exist if you just created the queue wait at least a minute for the queue to initialize debug error string created description error received from peer file src core lib surface call cc file line grpc message queue does not exist if you just created the queue wait at least a minute for the queue to initialize grpc status the above exception was the direct cause of the following exception traceback most recent call last file workspace appengine flexible tasks snippets test py line in test pause queue test project id test location queue name file workspace appengine flexible tasks snippets py line in pause queue response client pause queue queue path file workspace appengine flexible tasks nox py lib site packages google cloud tasks gapic cloud tasks client py line in pause queue request retry retry timeout timeout metadata metadata file workspace appengine flexible tasks nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace appengine flexible tasks nox py lib site packages google api core retry py line in retry wrapped func on error on error file workspace appengine flexible tasks nox py lib site packages google api core retry py line in retry target return target file workspace appengine flexible tasks nox py lib site packages google api core timeout py line in func with timeout return func args kwargs file workspace appengine flexible tasks nox py lib site packages google api core grpc helpers py line in error remapped callable six raise from exceptions from grpc error exc exc file line in raise from google api core exceptions notfound queue does not exist if you just created the queue wait at least a minute for the queue to initialize | 0 |
786,415 | 27,645,769,221 | IssuesEvent | 2023-03-10 22:49:58 | briandfoy/cpan-audit | https://api.github.com/repos/briandfoy/cpan-audit | closed | Feature: example drop-in cpan-audit.t file | Type: enhancement Priority: low | I think it would be great to offer a drop-in cpan-audit.t file for users to drop into their t/ directory, which does an audit and reports any security advisories on whatever is in cpanfile.snapshot, for example. | 1.0 | Feature: example drop-in cpan-audit.t file - I think it would be great to offer a drop-in cpan-audit.t file for users to drop into their t/ directory, which does an audit and reports any security advisories on whatever is in cpanfile.snapshot, for example. | priority | feature example drop in cpan audit t file i think it would be great to offer a drop in cpan audit t file for users to drop into their t directory which does an audit and reports any security advisories on whatever is in cpanfile snapshot for example | 1 |
74,821 | 14,346,379,831 | IssuesEvent | 2020-11-29 00:09:40 | Arcanorum/dungeonz | https://api.github.com/repos/Arcanorum/dungeonz | opened | Spatial audio | code | **Task description:**
Need to figure out and implement a way to associate playing certain sounds with an entity on screen, and adjust the volume to be appropriate for the distance from the player.
**References/notes:**
Some discussion around the topic.
https://phaser.discourse.group/t/sound-in-particular-place/2547/2
**Acceptance criteria:**
AC: A given sound can be associated with an entity on screen.
AC: If the player moves towards/away from that entity, the volume of that sound will increase/decrease based on distance.
AC: If the entity moves towards/away from the player, the volume of that sound will increase/decrease based on distance. | 1.0 | Spatial audio - **Task description:**
Need to figure out and implement a way to associate playing certain sounds with an entity on screen, and adjust the volume to be appropriate for the distance from the player.
**References/notes:**
Some discussion around the topic.
https://phaser.discourse.group/t/sound-in-particular-place/2547/2
**Acceptance criteria:**
AC: A given sound can be associated with an entity on screen.
AC: If the player moves towards/away from that entity, the volume of that sound will increase/decrease based on distance.
AC: If the entity moves towards/away from the player, the volume of that sound will increase/decrease based on distance. | non_priority | spatial audio task description need to figure out and implement a way to associate playing certain sounds with an entity on screen and adjust the volume to be appropriate for the distance from the player references notes some discussion around the topic acceptance criteria ac a given sound can be associated with an entity on screen ac if the player moves towards away from that entity the volume of that sound will increase decrease based on distance ac if the entity moves towards away from the player the volume of that sound will increase decrease based on distance | 0 |
640,842 | 20,810,232,190 | IssuesEvent | 2022-03-18 01:10:45 | monarch-initiative/mondo | https://api.github.com/repos/monarch-initiative/mondo | closed | MONDO:0005755 equine infectious anemia; non-human disease [Revise subclass] | Revise subclass high priority | **Mondo term (ID and Label)**
MONDO:0005755 equine infectious anemia
**Suggested revision and reasons**
I think this belongs under MONDO:0005583 "non-human animal disease"
NCI definition from UMLS CUI C0014661: "A horse disease caused by a retrovirus which is transmitted by biting flies. The acute phase symptoms include high fever, anemia, weakness, and swelling of the legs. The subacute phase symptoms include splenomegaly, anemia, and weight loss. The chronic phase symptoms include recurrent fevers and anemia. (NCI)"
**Your nano-attribution (ORCID)**
If you don't have an ORCID, you can sign up for one [here](https://orcid.org/)
| 1.0 | MONDO:0005755 equine infectious anemia; non-human disease [Revise subclass] - **Mondo term (ID and Label)**
MONDO:0005755 equine infectious anemia
**Suggested revision and reasons**
I think this belongs under MONDO:0005583 "non-human animal disease"
NCI definition from UMLS CUI C0014661: "A horse disease caused by a retrovirus which is transmitted by biting flies. The acute phase symptoms include high fever, anemia, weakness, and swelling of the legs. The subacute phase symptoms include splenomegaly, anemia, and weight loss. The chronic phase symptoms include recurrent fevers and anemia. (NCI)"
**Your nano-attribution (ORCID)**
If you don't have an ORCID, you can sign up for one [here](https://orcid.org/)
| priority | mondo equine infectious anemia non human disease mondo term id and label mondo equine infectious anemia suggested revision and reasons i think this belongs under mondo non human animal disease nci definition from umls cui a horse disease caused by a retrovirus which is transmitted by biting flies the acute phase symptoms include high fever anemia weakness and swelling of the legs the subacute phase symptoms include splenomegaly anemia and weight loss the chronic phase symptoms include recurrent fevers and anemia nci your nano attribution orcid if you don t have an orcid you can sign up for one | 1 |
49,621 | 3,003,711,772 | IssuesEvent | 2015-07-25 05:58:51 | jayway/powermock | https://api.github.com/repos/jayway/powermock | opened | mockin org.apache.http.impl.client.DefaultHttpClient class | bug imported Priority-Medium | _From [daghana...@gmail.com](https://code.google.com/u/104674216580764072044/) on July 22, 2014 06:55:51_
What steps will reproduce the problem? 1. create a constructor mock of .DefaultHttpClient using Powermock_V1.5.5 Mockito_V1.9.5 and Junit_V4.1 ektorp_V1.4.1
2.run the test call new StdHttpClient.Builder().url(dburl.toString()).caching(true) .cleanupIdleConnections(true).socketTimeout(Integer.parseInt(socketTimeout)).build() What is the expected output? What do you see instead? 1- connection established using mocked HttpClient
2- SSLContext.getInstance("TLS") throws NoSuchAlgorithmException What version of the product are you using? On what operating system? OS Ubuntu 14.04 Please provide any additional information below. Test code
@RunWith(PowerMockRunner.class)
@PrepareForTest(DefaultHttpClient.class)
public class App_UnitTest {
protected static InputStream mockUrlConnectionInputStream;
protected static OutputStream mockUrlConnectionOutputStream;
@InjectMocks
private final App appUnderTest = new App();
@Spy
private final CouchConnection couchConnection = new CouchConnection();
@Before
public void setupOnce() throws Exception {
MockitoAnnotations.initMocks(this);
couchConnection.setDbUrl("https://test.couch.url.com");
couchConnection.setDbName("testCouchDB");
couchConnection.setPort("12211");
couchConnection.setSocketTimeout("10000");
final DefaultHttpClient mockDefaultHttpClient = PowerMockito.mock(DefaultHttpClient.class);
PowerMockito.whenNew(DefaultHttpClient.class).withAnyArguments().thenReturn(mockDefaultHttpClient);
try {
URL.setURLStreamHandlerFactory(new MockURLStreamHandler());
} catch (final Error e) {
// Do nothing, this will be called by all the extending test class but should only be called once for JVM so just
// ignore
}
}
@Test
public void testifJobExists() throws IOException {
final HttpClient authenticatedHttpClient = new StdHttpClient.Builder().url(dburl.toString()).caching(true) .cleanupIdleConnections(true).socketTimeout(Integer.parseInt(socketTimeout)).build();
final CouchDbInstance couchInstance = new StdCouchDbInstance(authenticatedHttpClient);
couchConnector = couchInstance.createConnector(dbName, true);
}
public class MockURLStreamHandler extends URLStreamHandler implements URLStreamHandlerFactory {
// *** URLStreamHandler
@Override
protected URLConnection openConnection(final URL u) throws IOException {
return new MockHttpsURLConnection(u);
}
// *** URLStreamHandlerFactory
@Override
public URLStreamHandler createURLStreamHandler(final String protocol) {
// this will only override the chosen protocol so that file and other protocols used by the Spring and other
// classes will work as usual.
// WARNING this will only mock https connections, so your URL has to start with https e.g.
// https:\\something.etc.com
if (protocol.equalsIgnoreCase("https")) {
return this;
} else {
return null;
}
}
}
public class MockHttpsURLConnection extends HttpsURLConnection {
protected MockHttpsURLConnection(final URL url) {
super(url);
}
// *** HttpURLConnection
@Override
public InputStream getInputStream() throws IOException {
return mockUrlConnectionInputStream;
}
@Override
public void connect() throws IOException {
}
@Override
public void disconnect() {
}
@Override
public boolean usingProxy() {
return false;
}
@Override
public String getCipherSuite() {
return null;
}
@Override
public Certificate[] getLocalCertificates() {
return null;
}
@Override
public Certificate[] getServerCertificates() throws SSLPeerUnverifiedException {
return null;
}
@Override
public OutputStream getOutputStream() throws IOException {
return mockUrlConnectionOutputStream;
}
}
}
_Original issue: http://code.google.com/p/powermock/issues/detail?id=507_ | 1.0 | mockin org.apache.http.impl.client.DefaultHttpClient class - _From [daghana...@gmail.com](https://code.google.com/u/104674216580764072044/) on July 22, 2014 06:55:51_
What steps will reproduce the problem? 1. create a constructor mock of .DefaultHttpClient using Powermock_V1.5.5 Mockito_V1.9.5 and Junit_V4.1 ektorp_V1.4.1
2.run the test call new StdHttpClient.Builder().url(dburl.toString()).caching(true) .cleanupIdleConnections(true).socketTimeout(Integer.parseInt(socketTimeout)).build() What is the expected output? What do you see instead? 1- connection established using mocked HttpClient
2- SSLContext.getInstance("TLS") throws NoSuchAlgorithmException What version of the product are you using? On what operating system? OS Ubuntu 14.04 Please provide any additional information below. Test code
@RunWith(PowerMockRunner.class)
@PrepareForTest(DefaultHttpClient.class)
public class App_UnitTest {
protected static InputStream mockUrlConnectionInputStream;
protected static OutputStream mockUrlConnectionOutputStream;
@InjectMocks
private final App appUnderTest = new App();
@Spy
private final CouchConnection couchConnection = new CouchConnection();
@Before
public void setupOnce() throws Exception {
MockitoAnnotations.initMocks(this);
couchConnection.setDbUrl("https://test.couch.url.com");
couchConnection.setDbName("testCouchDB");
couchConnection.setPort("12211");
couchConnection.setSocketTimeout("10000");
final DefaultHttpClient mockDefaultHttpClient = PowerMockito.mock(DefaultHttpClient.class);
PowerMockito.whenNew(DefaultHttpClient.class).withAnyArguments().thenReturn(mockDefaultHttpClient);
try {
URL.setURLStreamHandlerFactory(new MockURLStreamHandler());
} catch (final Error e) {
// Do nothing, this will be called by all the extending test class but should only be called once for JVM so just
// ignore
}
}
@Test
public void testifJobExists() throws IOException {
final HttpClient authenticatedHttpClient = new StdHttpClient.Builder().url(dburl.toString()).caching(true) .cleanupIdleConnections(true).socketTimeout(Integer.parseInt(socketTimeout)).build();
final CouchDbInstance couchInstance = new StdCouchDbInstance(authenticatedHttpClient);
couchConnector = couchInstance.createConnector(dbName, true);
}
public class MockURLStreamHandler extends URLStreamHandler implements URLStreamHandlerFactory {
// *** URLStreamHandler
@Override
protected URLConnection openConnection(final URL u) throws IOException {
return new MockHttpsURLConnection(u);
}
// *** URLStreamHandlerFactory
@Override
public URLStreamHandler createURLStreamHandler(final String protocol) {
// this will only override the chosen protocol so that file and other protocols used by the Spring and other
// classes will work as usual.
// WARNING this will only mock https connections, so your URL has to start with https e.g.
// https:\\something.etc.com
if (protocol.equalsIgnoreCase("https")) {
return this;
} else {
return null;
}
}
}
public class MockHttpsURLConnection extends HttpsURLConnection {
protected MockHttpsURLConnection(final URL url) {
super(url);
}
// *** HttpURLConnection
@Override
public InputStream getInputStream() throws IOException {
return mockUrlConnectionInputStream;
}
@Override
public void connect() throws IOException {
}
@Override
public void disconnect() {
}
@Override
public boolean usingProxy() {
return false;
}
@Override
public String getCipherSuite() {
return null;
}
@Override
public Certificate[] getLocalCertificates() {
return null;
}
@Override
public Certificate[] getServerCertificates() throws SSLPeerUnverifiedException {
return null;
}
@Override
public OutputStream getOutputStream() throws IOException {
return mockUrlConnectionOutputStream;
}
}
}
_Original issue: http://code.google.com/p/powermock/issues/detail?id=507_ | priority | mockin org apache http impl client defaulthttpclient class from on july what steps will reproduce the problem create a constructor mock of defaulthttpclient using powermock mockito and junit ektorp run the test call new stdhttpclient builder url dburl tostring caching true cleanupidleconnections true sockettimeout integer parseint sockettimeout build what is the expected output what do you see instead connection established using mocked httpclient sslcontext getinstance tls throws nosuchalgorithmexception what version of the product are you using on what operating system os ubuntu please provide any additional information below test code runwith powermockrunner class preparefortest defaulthttpclient class public class app unittest protected static inputstream mockurlconnectioninputstream protected static outputstream mockurlconnectionoutputstream injectmocks private final app appundertest new app spy private final couchconnection couchconnection new couchconnection before public void setuponce throws exception mockitoannotations initmocks this couchconnection setdburl couchconnection setdbname testcouchdb couchconnection setport couchconnection setsockettimeout final defaulthttpclient mockdefaulthttpclient powermockito mock defaulthttpclient class powermockito whennew defaulthttpclient class withanyarguments thenreturn mockdefaulthttpclient try url seturlstreamhandlerfactory new mockurlstreamhandler catch final error e do nothing this will be called by all the extending test class but should only be called once for jvm so just ignore test public void testifjobexists throws ioexception final httpclient authenticatedhttpclient new stdhttpclient builder url dburl tostring caching true cleanupidleconnections true sockettimeout integer parseint sockettimeout build final couchdbinstance couchinstance new stdcouchdbinstance authenticatedhttpclient couchconnector couchinstance createconnector dbname true public class mockurlstreamhandler extends urlstreamhandler implements urlstreamhandlerfactory urlstreamhandler override protected urlconnection openconnection final url u throws ioexception return new mockhttpsurlconnection u urlstreamhandlerfactory override public urlstreamhandler createurlstreamhandler final string protocol this will only override the chosen protocol so that file and other protocols used by the spring and other classes will work as usual warning this will only mock https connections so your url has to start with https e g https something etc com if protocol equalsignorecase https return this else return null public class mockhttpsurlconnection extends httpsurlconnection protected mockhttpsurlconnection final url url super url httpurlconnection override public inputstream getinputstream throws ioexception return mockurlconnectioninputstream override public void connect throws ioexception override public void disconnect override public boolean usingproxy return false override public string getciphersuite return null override public certificate getlocalcertificates return null override public certificate getservercertificates throws sslpeerunverifiedexception return null override public outputstream getoutputstream throws ioexception return mockurlconnectionoutputstream original issue | 1 |
43,168 | 5,529,972,604 | IssuesEvent | 2017-03-21 00:30:20 | easydigitaldownloads/easy-digital-downloads | https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads | closed | Multiple EDD_Payments_Query's affect each other. | Bug Has PR Needs Testing Payments | If you instantiate 2 objects of EDD_Payments_Query, because the values are hooked to edd_pre_get_payments, any custom values you set up for the first object continue to be hooked for any subsequent objects.
**For example:**
The following code snippet will cause the payment history page in the WordPress dashboard to show no results (notice how we don't return any values or do anything with our returned data here - in theory, because of the word "get", this shouldn't be affecting anything):
```
function mess_with_edd_get_payments() {
$args = array(
'end_date' => time()
);
$payments = edd_get_payments( $args );
}
add_action( 'admin_init', 'mess_with_edd_get_payments' );
```
The problem is that in the EDD_Payments_Query class, the values are hooked to edd_pre_get_payments. So when the payment history page runs this in a separately instatiated object, the hook still runs which sets the "end_date" - even though we had no intention of setting the end_date for the Payment History page.
**Proposed Solution:**
Instead of hooking the custom values to edd_pre_get_payments, we should simply call each method from within the get_payments method in EDD_Payments_Query. | 1.0 | Multiple EDD_Payments_Query's affect each other. - If you instantiate 2 objects of EDD_Payments_Query, because the values are hooked to edd_pre_get_payments, any custom values you set up for the first object continue to be hooked for any subsequent objects.
**For example:**
The following code snippet will cause the payment history page in the WordPress dashboard to show no results (notice how we don't return any values or do anything with our returned data here - in theory, because of the word "get", this shouldn't be affecting anything):
```
function mess_with_edd_get_payments() {
$args = array(
'end_date' => time()
);
$payments = edd_get_payments( $args );
}
add_action( 'admin_init', 'mess_with_edd_get_payments' );
```
The problem is that in the EDD_Payments_Query class, the values are hooked to edd_pre_get_payments. So when the payment history page runs this in a separately instatiated object, the hook still runs which sets the "end_date" - even though we had no intention of setting the end_date for the Payment History page.
**Proposed Solution:**
Instead of hooking the custom values to edd_pre_get_payments, we should simply call each method from within the get_payments method in EDD_Payments_Query. | non_priority | multiple edd payments query s affect each other if you instantiate objects of edd payments query because the values are hooked to edd pre get payments any custom values you set up for the first object continue to be hooked for any subsequent objects for example the following code snippet will cause the payment history page in the wordpress dashboard to show no results notice how we don t return any values or do anything with our returned data here in theory because of the word get this shouldn t be affecting anything function mess with edd get payments args array end date time payments edd get payments args add action admin init mess with edd get payments the problem is that in the edd payments query class the values are hooked to edd pre get payments so when the payment history page runs this in a separately instatiated object the hook still runs which sets the end date even though we had no intention of setting the end date for the payment history page proposed solution instead of hooking the custom values to edd pre get payments we should simply call each method from within the get payments method in edd payments query | 0 |
47,853 | 7,354,063,348 | IssuesEvent | 2018-03-09 04:24:51 | Naoghuman/lib-validation | https://api.github.com/repos/Naoghuman/lib-validation | opened | [doc] Update ReadMe.md to 0.3.0. | documentation refactoring | [doc] Update ReadMe.md to 0.3.0.
* New UML image for the section `Intention`.
* Dependencies, Download... | 1.0 | [doc] Update ReadMe.md to 0.3.0. - [doc] Update ReadMe.md to 0.3.0.
* New UML image for the section `Intention`.
* Dependencies, Download... | non_priority | update readme md to update readme md to new uml image for the section intention dependencies download | 0 |
242,138 | 7,838,626,663 | IssuesEvent | 2018-06-18 10:56:19 | minishift/minishift-addons | https://api.github.com/repos/minishift/minishift-addons | opened | Add scenario for removal of Che addon | kind/task priority/major | Since PR #123 for removal of Che was merged there should be also a test case for the removal to cover the functionality in the future. | 1.0 | Add scenario for removal of Che addon - Since PR #123 for removal of Che was merged there should be also a test case for the removal to cover the functionality in the future. | priority | add scenario for removal of che addon since pr for removal of che was merged there should be also a test case for the removal to cover the functionality in the future | 1 |
612,926 | 19,059,447,170 | IssuesEvent | 2021-11-26 04:32:14 | tomusborne/generatepress | https://api.github.com/repos/tomusborne/generatepress | opened | Add missing wp_set_script_translations() functions | type: bug priority: medium | We're missing the needed `wp_set_script_translations( 'handle', 'generatepress' )` functions wherever we're adding `wp-i18n` as a dependency right now, which is preventing translations from working. | 1.0 | Add missing wp_set_script_translations() functions - We're missing the needed `wp_set_script_translations( 'handle', 'generatepress' )` functions wherever we're adding `wp-i18n` as a dependency right now, which is preventing translations from working. | priority | add missing wp set script translations functions we re missing the needed wp set script translations handle generatepress functions wherever we re adding wp as a dependency right now which is preventing translations from working | 1 |
830,349 | 32,003,233,796 | IssuesEvent | 2023-09-21 13:28:57 | dag-hammarskjold-library/dlx-rest | https://api.github.com/repos/dag-hammarskjold-library/dlx-rest | closed | Display and sorting in browse indexes by subfield when not in order | type: enhancement priority: high function: search sort | It looks like the sorting in the browse indexes is not by the alphabetical order of the subfields in the record, but rather by the order in which they display in the field? Here is an example:

This is what the second record in the list looks like - the subfields in 191 are not in order, but the index seems to read them as they are in the record and thus puts it in the wrong place in the browse index:

It should appear here:

| 1.0 | Display and sorting in browse indexes by subfield when not in order - It looks like the sorting in the browse indexes is not by the alphabetical order of the subfields in the record, but rather by the order in which they display in the field? Here is an example:

This is what the second record in the list looks like - the subfields in 191 are not in order, but the index seems to read them as they are in the record and thus puts it in the wrong place in the browse index:

It should appear here:

| priority | display and sorting in browse indexes by subfield when not in order it looks like the sorting in the browse indexes is not by the alphabetical order of the subfields in the record but rather by the order in which they display in the field here is an example this is what the second record in the list looks like the subfields in are not in order but the index seems to read them as they are in the record and thus puts it in the wrong place in the browse index it should appear here | 1 |
452,720 | 13,058,468,105 | IssuesEvent | 2020-07-30 09:04:12 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | Create standard/curved shader with tinting | Category: Tech Priority: Low | Copy Curved/Avatar, could possibly just replace it with generic version.
Would allow two colours, and blending between them. Could be extended to support more colours with less blending accuracy.
Can be used for:
* Avatar skin/clothing colour
* Colouring vehicles
* Colouring world objects (good use would be civics objects with nation colours) | 1.0 | Create standard/curved shader with tinting - Copy Curved/Avatar, could possibly just replace it with generic version.
Would allow two colours, and blending between them. Could be extended to support more colours with less blending accuracy.
Can be used for:
* Avatar skin/clothing colour
* Colouring vehicles
* Colouring world objects (good use would be civics objects with nation colours) | priority | create standard curved shader with tinting copy curved avatar could possibly just replace it with generic version would allow two colours and blending between them could be extended to support more colours with less blending accuracy can be used for avatar skin clothing colour colouring vehicles colouring world objects good use would be civics objects with nation colours | 1 |
335,417 | 10,153,354,408 | IssuesEvent | 2019-08-06 04:07:06 | gitcoinco/web | https://api.github.com/repos/gitcoinco/web | opened | Team Payout -Build - Milestone I - As a funder on Gitcion - I'd like the ability to manage permissions and payouts with my team members. | 8 priority: backlog sprint 16 | <!--
Hello Gitcoiner!
Please use the template below for issue ideas or bugs found within Gitcoin.
If it is general support you need, reach out to us at
gitcoin.co/slack
Provide a general summary of the issue in the title above and use relevant fields below to define the problem.
-->
### User Story
As a funder on Gitcion - I'd like the ability to manage permissions and payouts with my team members.
### Assumptions
### What
- The Gitcoin platform syncs repo data from Github related to collaborators and team members, mirroring their permissions model. TBD: what level of access will allow users to be added as "Issuers", who have full control, and "Approvers", who can accept submissions and issue payout.
### Why
Funders have requested this! | 1.0 | Team Payout -Build - Milestone I - As a funder on Gitcion - I'd like the ability to manage permissions and payouts with my team members. - <!--
Hello Gitcoiner!
Please use the template below for issue ideas or bugs found within Gitcoin.
If it is general support you need, reach out to us at
gitcoin.co/slack
Provide a general summary of the issue in the title above and use relevant fields below to define the problem.
-->
### User Story
As a funder on Gitcion - I'd like the ability to manage permissions and payouts with my team members.
### Assumptions
### What
- The Gitcoin platform syncs repo data from Github related to collaborators and team members, mirroring their permissions model. TBD: what level of access will allow users to be added as "Issuers", who have full control, and "Approvers", who can accept submissions and issue payout.
### Why
Funders have requested this! | priority | team payout build milestone i as a funder on gitcion i d like the ability to manage permissions and payouts with my team members hello gitcoiner please use the template below for issue ideas or bugs found within gitcoin if it is general support you need reach out to us at gitcoin co slack provide a general summary of the issue in the title above and use relevant fields below to define the problem user story as a funder on gitcion i d like the ability to manage permissions and payouts with my team members assumptions what the gitcoin platform syncs repo data from github related to collaborators and team members mirroring their permissions model tbd what level of access will allow users to be added as issuers who have full control and approvers who can accept submissions and issue payout why funders have requested this | 1 |
703,234 | 24,150,477,623 | IssuesEvent | 2022-09-21 23:47:45 | googleapis/nodejs-ai-platform | https://api.github.com/repos/googleapis/nodejs-ai-platform | closed | AI platform get hyperparameter tuning job: should get the specified hyperparameter tuning job failed | type: bug priority: p1 flakybot: issue api: vertex-ai | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: e1c5cd6b5d03afb03911ba9aa685457aa359a602
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/69dc3b23-298d-461c-9279-cd7551a90fc9), [Sponge](http://sponge2/69dc3b23-298d-461c-9279-cd7551a90fc9)
status: failed
<details><summary>Test output</summary><br><pre>Command failed: node ./get-hyperparameter-tuning-job.js 2216298782247616512 ucaip-sample-tests us-central1
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
Error: Command failed: node ./get-hyperparameter-tuning-job.js 2216298782247616512 ucaip-sample-tests us-central1
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
at checkExecSyncError (child_process.js:635:11)
at Object.execSync (child_process.js:671:15)
at execSync (test/get-hyperparameter-tuning-job.test.js:23:28)
at Context.<anonymous> (test/get-hyperparameter-tuning-job.test.js:31:20)
at processImmediate (internal/timers.js:461:21)</pre></details> | 1.0 | AI platform get hyperparameter tuning job: should get the specified hyperparameter tuning job failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: e1c5cd6b5d03afb03911ba9aa685457aa359a602
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/69dc3b23-298d-461c-9279-cd7551a90fc9), [Sponge](http://sponge2/69dc3b23-298d-461c-9279-cd7551a90fc9)
status: failed
<details><summary>Test output</summary><br><pre>Command failed: node ./get-hyperparameter-tuning-job.js 2216298782247616512 ucaip-sample-tests us-central1
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
Error: Command failed: node ./get-hyperparameter-tuning-job.js 2216298782247616512 ucaip-sample-tests us-central1
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
at checkExecSyncError (child_process.js:635:11)
at Object.execSync (child_process.js:671:15)
at execSync (test/get-hyperparameter-tuning-job.test.js:23:28)
at Context.<anonymous> (test/get-hyperparameter-tuning-job.test.js:31:20)
at processImmediate (internal/timers.js:461:21)</pre></details> | priority | ai platform get hyperparameter tuning job should get the specified hyperparameter tuning job failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output command failed node get hyperparameter tuning job js ucaip sample tests us unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see error command failed node get hyperparameter tuning job js ucaip sample tests us unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at checkexecsyncerror child process js at object execsync child process js at execsync test get hyperparameter tuning job test js at context test get hyperparameter tuning job test js at processimmediate internal timers js | 1 |
73,471 | 7,335,034,949 | IssuesEvent | 2018-03-06 01:44:21 | istio/istio | https://api.github.com/repos/istio/istio | closed | Test Failure: security/tests/integration/certificateRotationTest | kind/fixit kind/test-failure | From: https://k8s-gubernator.appspot.com/build/istio-prow/pull/istio_istio/3913/istio-presubmit/6070/
```
I0302 22:34:20.343] 2018-03-02T22:34:20.341870Z error failed to create test namespace: failed to create a role (error: failed to create role (error: roles.rbac.authorization.k8s.io "istio-ca-role" is forbidden: attempt to grant extra privileges: [PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["create"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["update"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["create"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["update"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["list"]}] user=&{istio-prow-test-job@istio-testing.iam.gserviceaccount.com [system:authenticated] map[authenticator:[GKE]]} ownerrules=[PolicyRule{Resources:["selfsubjectaccessreviews" "selfsubjectrulesreviews"], APIGroups:["authorization.k8s.io"], Verbs:["create"]} PolicyRule{NonResourceURLs:["/api" "/api/*" "/apis" "/apis/*" "/healthz" "/swagger-2.0.0.pb-v1" "/swagger.json" "/swaggerapi" "/swaggerapi/*" "/version"], Verbs:["get"]}] ruleResolutionErrors=[]))
I0302 22:34:20.344] 2018-03-02T22:34:20.341877Z error test environment creation failure
I0302 22:34:20.344] FAIL istio.io/istio/security/tests/integration/certificateRotationTest 2.535s
E0302 22:34:20.355] Build failed
I0302 22:34:20.355] process 509 exited with code 1 after 15.7m
E0302 22:34:20.355] FAIL: istio-presubmit
```
<!--
Please see https://istio.io/help and if you are a user of Istio, please file issues in
https://github.com/istio/issues/issues instead of here.
Only confirmed, triaged and labelled issues should be filed here.
Please add the correct labels and epics (and priority and milestones if you have that information)
-->
| 1.0 | Test Failure: security/tests/integration/certificateRotationTest - From: https://k8s-gubernator.appspot.com/build/istio-prow/pull/istio_istio/3913/istio-presubmit/6070/
```
I0302 22:34:20.343] 2018-03-02T22:34:20.341870Z error failed to create test namespace: failed to create a role (error: failed to create role (error: roles.rbac.authorization.k8s.io "istio-ca-role" is forbidden: attempt to grant extra privileges: [PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["create"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["update"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["create"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["update"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["list"]}] user=&{istio-prow-test-job@istio-testing.iam.gserviceaccount.com [system:authenticated] map[authenticator:[GKE]]} ownerrules=[PolicyRule{Resources:["selfsubjectaccessreviews" "selfsubjectrulesreviews"], APIGroups:["authorization.k8s.io"], Verbs:["create"]} PolicyRule{NonResourceURLs:["/api" "/api/*" "/apis" "/apis/*" "/healthz" "/swagger-2.0.0.pb-v1" "/swagger.json" "/swaggerapi" "/swaggerapi/*" "/version"], Verbs:["get"]}] ruleResolutionErrors=[]))
I0302 22:34:20.344] 2018-03-02T22:34:20.341877Z error test environment creation failure
I0302 22:34:20.344] FAIL istio.io/istio/security/tests/integration/certificateRotationTest 2.535s
E0302 22:34:20.355] Build failed
I0302 22:34:20.355] process 509 exited with code 1 after 15.7m
E0302 22:34:20.355] FAIL: istio-presubmit
```
<!--
Please see https://istio.io/help and if you are a user of Istio, please file issues in
https://github.com/istio/issues/issues instead of here.
Only confirmed, triaged and labelled issues should be filed here.
Please add the correct labels and epics (and priority and milestones if you have that information)
-->
| non_priority | test failure security tests integration certificaterotationtest from error failed to create test namespace failed to create a role error failed to create role error roles rbac authorization io istio ca role is forbidden attempt to grant extra privileges apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs user istio prow test job istio testing iam gserviceaccount com map ownerrules apigroups verbs policyrule nonresourceurls verbs ruleresolutionerrors error test environment creation failure fail istio io istio security tests integration certificaterotationtest build failed process exited with code after fail istio presubmit please see and if you are a user of istio please file issues in instead of here only confirmed triaged and labelled issues should be filed here please add the correct labels and epics and priority and milestones if you have that information | 0 |
44,335 | 12,101,453,469 | IssuesEvent | 2020-04-20 15:14:09 | codesmithtools/Templates | https://api.github.com/repos/codesmithtools/Templates | closed | Join Table w/ Dependent Foreign Key | Framework-NHibernate Type-Defect auto-migrated | ```
Update the IsManyToMany logic and add a constraint on having no dependent
foreign keys.
http://community.codesmithtools.com/forums/t/10071.aspx
```
Original issue reported on code.google.com by `tdupont...@gmail.com` on 24 Aug 2009 at 4:13
| 1.0 | Join Table w/ Dependent Foreign Key - ```
Update the IsManyToMany logic and add a constraint on having no dependent
foreign keys.
http://community.codesmithtools.com/forums/t/10071.aspx
```
Original issue reported on code.google.com by `tdupont...@gmail.com` on 24 Aug 2009 at 4:13
| non_priority | join table w dependent foreign key update the ismanytomany logic and add a constraint on having no dependent foreign keys original issue reported on code google com by tdupont gmail com on aug at | 0 |
29,666 | 2,716,767,477 | IssuesEvent | 2015-04-10 21:15:18 | CruxFramework/crux | https://api.github.com/repos/CruxFramework/crux | closed | DataProvider clearChanges method is not working | bug imported Milestone-M14-C4 Priority-Medium TargetVersion-5.3.0 | _From [trbustam...@gmail.com](https://code.google.com/u/117925048001886933493/) on September 19, 2014 14:54:16_
The clearChanges method is not undoing editions on the dataprovider. It is only cleaning the change logs.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=537_ | 1.0 | DataProvider clearChanges method is not working - _From [trbustam...@gmail.com](https://code.google.com/u/117925048001886933493/) on September 19, 2014 14:54:16_
The clearChanges method is not undoing editions on the dataprovider. It is only cleaning the change logs.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=537_ | priority | dataprovider clearchanges method is not working from on september the clearchanges method is not undoing editions on the dataprovider it is only cleaning the change logs original issue | 1 |
824,309 | 31,149,247,347 | IssuesEvent | 2023-08-16 08:49:17 | fossasia/open-event-frontend | https://api.github.com/repos/fossasia/open-event-frontend | opened | Speakers getting disconnected from sessions when editing social media profile | bug Priority: Urgent | When a speaker adds social media entries to their account they get disconnected from some sessions if they have several sessions. | 1.0 | Speakers getting disconnected from sessions when editing social media profile - When a speaker adds social media entries to their account they get disconnected from some sessions if they have several sessions. | priority | speakers getting disconnected from sessions when editing social media profile when a speaker adds social media entries to their account they get disconnected from some sessions if they have several sessions | 1 |
657,885 | 21,870,262,182 | IssuesEvent | 2022-05-19 03:59:34 | pytorch/data | https://api.github.com/repos/pytorch/data | closed | Multiprocessing with any DataPipe writing to local file | bug good first issue help wanted high priority | ### 🐛 Describe the bug
We need to take extra care all DataPipe that would write to file system when DataLoader2 triggered multiprocessing. If the file name on the local file system is same across multiple processes, it would be a racing condition.
This is found when TorchText team is using `on_disk_cache` to cache file.
DataLoader needs to know such DataPipe must be sharded with multiprocessing or enforce it into single process.
As a workaround, users have to download the file to local file system to prevent writing within DataPipe.
### Versions
main branch | 1.0 | Multiprocessing with any DataPipe writing to local file - ### 🐛 Describe the bug
We need to take extra care all DataPipe that would write to file system when DataLoader2 triggered multiprocessing. If the file name on the local file system is same across multiple processes, it would be a racing condition.
This is found when TorchText team is using `on_disk_cache` to cache file.
DataLoader needs to know such DataPipe must be sharded with multiprocessing or enforce it into single process.
As a workaround, users have to download the file to local file system to prevent writing within DataPipe.
### Versions
main branch | priority | multiprocessing with any datapipe writing to local file 🐛 describe the bug we need to take extra care all datapipe that would write to file system when triggered multiprocessing if the file name on the local file system is same across multiple processes it would be a racing condition this is found when torchtext team is using on disk cache to cache file dataloader needs to know such datapipe must be sharded with multiprocessing or enforce it into single process as a workaround users have to download the file to local file system to prevent writing within datapipe versions main branch | 1 |
50,182 | 3,006,232,964 | IssuesEvent | 2015-07-27 09:03:14 | Itseez/opencv | https://api.github.com/repos/Itseez/opencv | opened | Incorrect window size on Mac OS X Lion | auto-transferred bug category: highgui-gui priority: normal | Transferred from http://code.opencv.org/issues/2189
```
|| Jan Dlabal on 2012-07-24 17:26
|| Priority: Normal
|| Affected: None
|| Category: highgui-gui
|| Tracker: Bug
|| Difficulty: None
|| PR: None
|| Platform: None / None
```
Incorrect window size on Mac OS X Lion
-----------
```
See http://stackoverflow.com/questions/11635842/opencv-not-filling-entire-image.
Basically this:
@
cv::Mat cvSideDepthImage1(150, 150, CV_8UC1, cv::Scalar(100));
cv::imshow("side1", cvSideDepthImage1);
@
Creates this 200x150px window:
!http://i.stack.imgur.com/HetAA.png!
When it should just show a gray 150x150 square.
This is OpenCV 2.4.9 on OS X Lion (all updates installed).
```
History
-------
##### Marina Kolpakova on 2012-07-24 18:05
```
- Category set to highgui-images
```
##### Andrey Kamaev on 2012-08-15 13:42
```
- Assignee set to Vadim Pisarevsky
```
##### Andrey Kamaev on 2012-08-16 15:42
```
- Category changed from highgui-images to highgui-gui
``` | 1.0 | Incorrect window size on Mac OS X Lion - Transferred from http://code.opencv.org/issues/2189
```
|| Jan Dlabal on 2012-07-24 17:26
|| Priority: Normal
|| Affected: None
|| Category: highgui-gui
|| Tracker: Bug
|| Difficulty: None
|| PR: None
|| Platform: None / None
```
Incorrect window size on Mac OS X Lion
-----------
```
See http://stackoverflow.com/questions/11635842/opencv-not-filling-entire-image.
Basically this:
@
cv::Mat cvSideDepthImage1(150, 150, CV_8UC1, cv::Scalar(100));
cv::imshow("side1", cvSideDepthImage1);
@
Creates this 200x150px window:
!http://i.stack.imgur.com/HetAA.png!
When it should just show a gray 150x150 square.
This is OpenCV 2.4.9 on OS X Lion (all updates installed).
```
History
-------
##### Marina Kolpakova on 2012-07-24 18:05
```
- Category set to highgui-images
```
##### Andrey Kamaev on 2012-08-15 13:42
```
- Assignee set to Vadim Pisarevsky
```
##### Andrey Kamaev on 2012-08-16 15:42
```
- Category changed from highgui-images to highgui-gui
``` | priority | incorrect window size on mac os x lion transferred from jan dlabal on priority normal affected none category highgui gui tracker bug difficulty none pr none platform none none incorrect window size on mac os x lion see basically this cv mat cv cv scalar cv imshow creates this window when it should just show a gray square this is opencv on os x lion all updates installed history marina kolpakova on category set to highgui images andrey kamaev on assignee set to vadim pisarevsky andrey kamaev on category changed from highgui images to highgui gui | 1 |
269,879 | 8,444,066,087 | IssuesEvent | 2018-10-18 17:19:44 | poanetwork/metamask-extension | https://api.github.com/repos/poanetwork/metamask-extension | closed | (Bug) Token info isn't displayed if switch the network to localhost and back | high priority logical bug ready for release | Steps:
1. Set Sokol network
2. Add any token
3. Switch network to localhost
4. Switch network back to Sokol
Expected result:
- token info should be loaded and properly displayed
Actual result:
- token info isn't loaded
<img width="1440" alt="screen shot 2018-09-10 at 11 47 41 pm" src="https://user-images.githubusercontent.com/14067096/45345149-54408a00-b55a-11e8-8244-745f1a927606.png">
| 1.0 | (Bug) Token info isn't displayed if switch the network to localhost and back - Steps:
1. Set Sokol network
2. Add any token
3. Switch network to localhost
4. Switch network back to Sokol
Expected result:
- token info should be loaded and properly displayed
Actual result:
- token info isn't loaded
<img width="1440" alt="screen shot 2018-09-10 at 11 47 41 pm" src="https://user-images.githubusercontent.com/14067096/45345149-54408a00-b55a-11e8-8244-745f1a927606.png">
| priority | bug token info isn t displayed if switch the network to localhost and back steps set sokol network add any token switch network to localhost switch network back to sokol expected result token info should be loaded and properly displayed actual result token info isn t loaded img width alt screen shot at pm src | 1 |
114,876 | 17,266,880,013 | IssuesEvent | 2021-07-22 14:44:56 | turkdevops/php-src | https://api.github.com/repos/turkdevops/php-src | closed | CVE-2019-11041 (High) detected in php-srcphp-7.1.0RC3 - autoclosed | security vulnerability | ## CVE-2019-11041 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>php-srcphp-7.1.0RC3</b></p></summary>
<p>
<p>The PHP Interpreter</p>
<p>Library home page: <a href=https://github.com/php/php-src.git>https://github.com/php/php-src.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/php-src/commit/ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2">ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2</a></p>
<p>Found in base branch: <b>microseconds</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>php-src/ext/exif/exif.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When PHP EXIF extension is parsing EXIF information from an image, e.g. via exif_read_data() function, in PHP versions 7.1.x below 7.1.31, 7.2.x below 7.2.21 and 7.3.x below 7.3.8 it is possible to supply it with data what will cause it to read past the allocated buffer. This may lead to information disclosure or crash.
<p>Publish Date: 2019-08-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11041>CVE-2019-11041</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11041">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11041</a></p>
<p>Release Date: 2019-08-09</p>
<p>Fix Resolution: 7.1.31,7.2.21,7.3.8 </p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-11041 (High) detected in php-srcphp-7.1.0RC3 - autoclosed - ## CVE-2019-11041 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>php-srcphp-7.1.0RC3</b></p></summary>
<p>
<p>The PHP Interpreter</p>
<p>Library home page: <a href=https://github.com/php/php-src.git>https://github.com/php/php-src.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/php-src/commit/ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2">ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2</a></p>
<p>Found in base branch: <b>microseconds</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>php-src/ext/exif/exif.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When PHP EXIF extension is parsing EXIF information from an image, e.g. via exif_read_data() function, in PHP versions 7.1.x below 7.1.31, 7.2.x below 7.2.21 and 7.3.x below 7.3.8 it is possible to supply it with data what will cause it to read past the allocated buffer. This may lead to information disclosure or crash.
<p>Publish Date: 2019-08-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11041>CVE-2019-11041</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11041">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11041</a></p>
<p>Release Date: 2019-08-09</p>
<p>Fix Resolution: 7.1.31,7.2.21,7.3.8 </p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in php srcphp autoclosed cve high severity vulnerability vulnerable library php srcphp the php interpreter library home page a href found in head commit a href found in base branch microseconds vulnerable source files php src ext exif exif c vulnerability details when php exif extension is parsing exif information from an image e g via exif read data function in php versions x below x below and x below it is possible to supply it with data what will cause it to read past the allocated buffer this may lead to information disclosure or crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
125,225 | 26,620,774,968 | IssuesEvent | 2023-01-24 11:03:02 | sourcegraph/sourcegraph | https://api.github.com/repos/sourcegraph/sourcegraph | opened | insights: use code insight title for data export filename | team/code-insights backend | follow up on https://github.com/sourcegraph/sourcegraph/pull/46662#discussion_r1084634856
escape illegal characters | 1.0 | insights: use code insight title for data export filename - follow up on https://github.com/sourcegraph/sourcegraph/pull/46662#discussion_r1084634856
escape illegal characters | non_priority | insights use code insight title for data export filename follow up on escape illegal characters | 0 |
111,739 | 11,741,181,964 | IssuesEvent | 2020-03-11 21:09:49 | ISPPNightTurn/Clubby | https://api.github.com/repos/ISPPNightTurn/Clubby | closed | Diseñar presentación del 11/03/2020 | desing documentation | Es necesario preparar la presentación para la clase del próximo miércoles. | 1.0 | Diseñar presentación del 11/03/2020 - Es necesario preparar la presentación para la clase del próximo miércoles. | non_priority | diseñar presentación del es necesario preparar la presentación para la clase del próximo miércoles | 0 |
79,357 | 10,120,685,146 | IssuesEvent | 2019-07-31 14:12:42 | kids-first/kf-api-release-coordinator | https://api.github.com/repos/kids-first/kf-api-release-coordinator | closed | Add sphinx docs site | documentation | We should update existing documentation to use a sphinx docs site like most other code bases. | 1.0 | Add sphinx docs site - We should update existing documentation to use a sphinx docs site like most other code bases. | non_priority | add sphinx docs site we should update existing documentation to use a sphinx docs site like most other code bases | 0 |
490,825 | 14,140,593,105 | IssuesEvent | 2020-11-10 11:25:38 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | SHUTDOWN should not disable SHUTDOWN_HARD | Category: Core & System Community Priority: Normal Status: Accepted Type: Feature | **Description**
The most annoying issue of opennebula.
**To Reproduce**
Regarding:
- Shutdown
- Power Off
- Reset
Try to shut down a VM. Find out it won't shut down.
Now sit and wait a few minutes so you can force shutdown it.
Optionally, be educated that you want to do something invalid,
while in reality SW failed.
Also applies to trying a hard reset if a VM didn't boot off i.e. PXE.
**Expected behavior**
More intrusive "power actions" are not blocked by lighter ones.
I can pull the power cable on a server if it hangs during shut down.
There's nothing that will first block me from doing that and suddenly, without notice it will allow it.
**Details**
- Affected Component: [e.g. Sunstone, Scheduler or Storage]
- Hypervisor: [e.g. KVM]
- Version: [e.g. 5.4.12 or development]
**Additional context**
Add any other context about the problem here.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| 1.0 | SHUTDOWN should not disable SHUTDOWN_HARD - **Description**
The most annoying issue of opennebula.
**To Reproduce**
Regarding:
- Shutdown
- Power Off
- Reset
Try to shut down a VM. Find out it won't shut down.
Now sit and wait a few minutes so you can force shutdown it.
Optionally, be educated that you want to do something invalid,
while in reality SW failed.
Also applies to trying a hard reset if a VM didn't boot off i.e. PXE.
**Expected behavior**
More intrusive "power actions" are not blocked by lighter ones.
I can pull the power cable on a server if it hangs during shut down.
There's nothing that will first block me from doing that and suddenly, without notice it will allow it.
**Details**
- Affected Component: [e.g. Sunstone, Scheduler or Storage]
- Hypervisor: [e.g. KVM]
- Version: [e.g. 5.4.12 or development]
**Additional context**
Add any other context about the problem here.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| priority | shutdown should not disable shutdown hard description the most annoying issue of opennebula to reproduce regarding shutdown power off reset try to shut down a vm find out it won t shut down now sit and wait a few minutes so you can force shutdown it optionally be educated that you want to do something invalid while in reality sw failed also applies to trying a hard reset if a vm didn t boot off i e pxe expected behavior more intrusive power actions are not blocked by lighter ones i can pull the power cable on a server if it hangs during shut down there s nothing that will first block me from doing that and suddenly without notice it will allow it details affected component hypervisor version additional context add any other context about the problem here progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches | 1 |
239,848 | 18,285,901,013 | IssuesEvent | 2021-10-05 10:13:32 | girlscript/winter-of-contributing | https://api.github.com/repos/girlscript/winter-of-contributing | closed | Competitive Programming : : Sliding window maximum (documentation) | documentation GWOC21 Assigned Competitive Programming | <hr>
## Description 📜
I would like to provide documentation about the Sliding window maximum problem.
<hr>
## Domain of Contribution 📊
<!----Please delete options that are not relevant.And in order to tick the check box just but x inside them for example [x] like this----->
- [x] Competitive Programming
<hr>
| 1.0 | Competitive Programming : : Sliding window maximum (documentation) - <hr>
## Description 📜
I would like to provide documentation about the Sliding window maximum problem.
<hr>
## Domain of Contribution 📊
<!----Please delete options that are not relevant.And in order to tick the check box just but x inside them for example [x] like this----->
- [x] Competitive Programming
<hr>
| non_priority | competitive programming sliding window maximum documentation description 📜 i would like to provide documentation about the sliding window maximum problem domain of contribution 📊 competitive programming | 0 |
250,619 | 7,979,201,252 | IssuesEvent | 2018-07-17 20:52:53 | neurosynth/neurosynth-web | https://api.github.com/repos/neurosynth/neurosynth-web | closed | Update code page to include this repo | enhancement priority:med | The Code page needs to add a link to and description of this repository.
| 1.0 | Update code page to include this repo - The Code page needs to add a link to and description of this repository.
| priority | update code page to include this repo the code page needs to add a link to and description of this repository | 1 |
107,692 | 16,762,159,331 | IssuesEvent | 2021-06-14 01:03:34 | ioana-nicolae/first | https://api.github.com/repos/ioana-nicolae/first | closed | WS-2018-0209 (Medium) detected in morgan-1.8.0.tgz - autoclosed | security vulnerability | ## WS-2018-0209 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>morgan-1.8.0.tgz</b></p></summary>
<p>HTTP request logger middleware for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.8.0.tgz">https://registry.npmjs.org/morgan/-/morgan-1.8.0.tgz</a></p>
<p>Path to dependency file: first/angular.js-master/angular.js-master/yarn.lock</p>
<p>Path to vulnerable library: first/angular.js-master/angular.js-master/yarn.lock</p>
<p>
Dependency Hierarchy:
- grunt-contrib-connect-1.0.2.tgz (Root Library)
- :x: **morgan-1.8.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/first/commit/6e040516f98074b37f69996b76c1c99015c1d685">6e040516f98074b37f69996b76c1c99015c1d685</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.
<p>Publish Date: 2018-11-25
<p>URL: <a href=https://github.com/Twipped/ircdkit/issues/1>WS-2018-0209</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/735">https://www.npmjs.com/advisories/735</a></p>
<p>Release Date: 2019-04-08</p>
<p>Fix Resolution: 1.9.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"morgan","packageVersion":"1.8.0","packageFilePaths":["/angular.js-master/angular.js-master/yarn.lock"],"isTransitiveDependency":true,"dependencyTree":"grunt-contrib-connect:1.0.2;morgan:1.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2018-0209","vulnerabilityDetails":"Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.","vulnerabilityUrl":"https://github.com/Twipped/ircdkit/issues/1","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> --> | True | WS-2018-0209 (Medium) detected in morgan-1.8.0.tgz - autoclosed - ## WS-2018-0209 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>morgan-1.8.0.tgz</b></p></summary>
<p>HTTP request logger middleware for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.8.0.tgz">https://registry.npmjs.org/morgan/-/morgan-1.8.0.tgz</a></p>
<p>Path to dependency file: first/angular.js-master/angular.js-master/yarn.lock</p>
<p>Path to vulnerable library: first/angular.js-master/angular.js-master/yarn.lock</p>
<p>
Dependency Hierarchy:
- grunt-contrib-connect-1.0.2.tgz (Root Library)
- :x: **morgan-1.8.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/first/commit/6e040516f98074b37f69996b76c1c99015c1d685">6e040516f98074b37f69996b76c1c99015c1d685</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.
<p>Publish Date: 2018-11-25
<p>URL: <a href=https://github.com/Twipped/ircdkit/issues/1>WS-2018-0209</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/735">https://www.npmjs.com/advisories/735</a></p>
<p>Release Date: 2019-04-08</p>
<p>Fix Resolution: 1.9.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"morgan","packageVersion":"1.8.0","packageFilePaths":["/angular.js-master/angular.js-master/yarn.lock"],"isTransitiveDependency":true,"dependencyTree":"grunt-contrib-connect:1.0.2;morgan:1.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2018-0209","vulnerabilityDetails":"Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.","vulnerabilityUrl":"https://github.com/Twipped/ircdkit/issues/1","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> --> | non_priority | ws medium detected in morgan tgz autoclosed ws medium severity vulnerability vulnerable library morgan tgz http request logger middleware for node js library home page a href path to dependency file first angular js master angular js master yarn lock path to vulnerable library first angular js master angular js master yarn lock dependency hierarchy grunt contrib connect tgz root library x morgan tgz vulnerable library found in head commit a href found in base branch master vulnerability details morgan before is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt contrib connect morgan isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier ws vulnerabilitydetails morgan before is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack vulnerabilityurl | 0 |
162,834 | 6,176,610,319 | IssuesEvent | 2017-07-01 15:22:03 | Placeholder-Software/Dissonance | https://api.github.com/repos/Placeholder-Software/Dissonance | closed | Reconnection Error | Priority: High Status: Awaiting User Feedback Type: Bug | Hi,
we've bought your asset a couple of days ago and trying out your HLAPI as well as LLAPI integrations. Unfortunately LLAPI has some issues I'm not sure how to resolve them and that seem unusual.
I'm currently working with Unity 5.6.0f3. Empty Project, Dissonance Package imported as well asLLAPI integration.
## Steps
0) Open DissonanceLlapiDemo
1) Build project, x64, development build
2) run build as client/server
3) run unity editor and connect (2-3 may be flipped)
4) wait a sec, maybe talk
5) disconnect client
6) wait a sec
7) reconnect
sometimes when reconnecting/disconnecting something goes wrong and the following errors appear. After that I'm not able to connect to the server anymore and have to restart the server.
Server side error:
```
host id {0} has been already deleted
UnityEngine.Networking.NetworkTransport:ReceiveFromHost(Int32, Int32&, Int32&, Byte[], Int32, Int32&, Byte&)
Dissonance.Integrations.UNet_LLAPI.UNetServer:ReadMessages() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetServer.cs:47)
Dissonance.Networking.BaseServer`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseServer.cs:78)
Dissonance.Networking.BaseCommsNetwork`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:124)
[Dissonance:Network] UNetServer: Error reading server socket: WrongHost
UnityEngine.Debug:LogError(Object)
Dissonance.Logs:SendLogMessage(String, Action`1) (at Assets/Plugins/Dissonance/Core/Log.cs:114)
Dissonance.Log:WriteLog(LogLevel, String) (at Assets/Plugins/Dissonance/Core/Log.cs:167)
Dissonance.Log:WriteLogFormat(LogLevel, String, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:190)
Dissonance.Log:Error(String, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:340)
Dissonance.Integrations.UNet_LLAPI.UNetServer:ReadMessages() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetServer.cs:50)
Dissonance.Networking.BaseServer`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseServer.cs:78)
Dissonance.Networking.BaseCommsNetwork`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:124)
```
Client-side error
```
[Dissonance:Network] UNetClient: Failed to cleanly disconnect from Dissonance server at 127.0.0.1:5889, Error WrongConnection
UnityEngine.Debug:LogError(Object)
Dissonance.Logs:SendLogMessage(String, Action`1) (at Assets/Plugins/Dissonance/Core/Log.cs:114)
Dissonance.Log:WriteLog(LogLevel, String) (at Assets/Plugins/Dissonance/Core/Log.cs:167)
Dissonance.Log:WriteLogFormat(LogLevel, String, String, UInt16, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:208)
Dissonance.Log:Error(String, String, UInt16, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:352)
Dissonance.Integrations.UNet_LLAPI.UNetClient:Disconnect() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetClient.cs:40)
Dissonance.Networking.BaseCommsNetwork`3:StopClient() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:186)
Dissonance.Integrations.UNet_LLAPI.Demo.LlapiManager:OnGUI() (at Assets/Dissonance/Integrations/UNet_LLAPI/Demo/LlapiManager.cs:45)
```
When restarting the server (full restart, close application, reopening the application) and do not stop the client and try to reconnect I get:
```
[Dissonance:Network] UNetClient: Exception invoking event handler: System.ArgumentNullException: Argument cannot be null.
Parameter name: key
at System.Collections.Generic.Dictionary`2[System.String,Dissonance.VoicePlayerState].TryGetValue (System.String key, Dissonance.VoicePlayerState& value) [0x000ba] in /Users/builduser/buildslave/mono/build/mcs/class/corlib/System.Collections.Generic/Dictionary.cs:608
at Dissonance.DissonanceComms.Net_PlayerStartedSpeaking (System.String player) [0x0000a] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\DissonanceComms.cs:417
at Dissonance.Networking.BaseCommsNetwork`3[Dissonance.Integrations.UNet_LLAPI.UNetServer,Dissonance.Integrations.UNet_LLAPI.UNetClient,System.Int32].OnPlayerStartedSpeaking (System.String obj) [0x00010] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseCommsNetwork.cs:296
at Dissonance.Networking.BaseClient`3[Dissonance.Integrations.UNet_LLAPI.UNetServer,Dissonance.Integrations.UNet_LLAPI.UNetClient,System.Int32].InvokeEvent[String] (System.String& arg, System.Action`1 handler) [0x0000f] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:316
UnityEngine.DebugLogHandler:Internal_Log(LogType, String, Object)
UnityEngine.DebugLogHandler:LogFormat(LogType, Object, String, Object[])
UnityEngine.Logger:Log(LogType, Object)
UnityEngine.Debug:LogError(Object)
Dissonance.Logs:SendLogMessage(String, Action`1) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:114)
Dissonance.Log:WriteLog(LogLevel, String) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:167)
Dissonance.Log:WriteLogFormat(LogLevel, String, Exception) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:190)
Dissonance.Log:Error(String, Exception) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:340)
Dissonance.Networking.BaseClient`3:InvokeEvent(String&, Action`1) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:320)
Dissonance.Networking.BaseClient`3:DispatchEvents() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:282)
Dissonance.Networking.BaseClient`3:Update() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:174)
Dissonance.Networking.BaseCommsNetwork`3:Update() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseCommsNetwork.cs:128)
```
The dicionary error may be resolved by just adding a check if state already exists for this key and if so just update/replace the state.I already tried that but it doesn't resolve the initial error.
Can you reproduce the error? If not I can upload you my test project. Or if you need anything else let me know.
Would be great if you could give a quick hint if this is a known issue and/or being worked on or not so that I can plan ahead.
Best
Marcel | 1.0 | Reconnection Error - Hi,
we've bought your asset a couple of days ago and trying out your HLAPI as well as LLAPI integrations. Unfortunately LLAPI has some issues I'm not sure how to resolve them and that seem unusual.
I'm currently working with Unity 5.6.0f3. Empty Project, Dissonance Package imported as well asLLAPI integration.
## Steps
0) Open DissonanceLlapiDemo
1) Build project, x64, development build
2) run build as client/server
3) run unity editor and connect (2-3 may be flipped)
4) wait a sec, maybe talk
5) disconnect client
6) wait a sec
7) reconnect
sometimes when reconnecting/disconnecting something goes wrong and the following errors appear. After that I'm not able to connect to the server anymore and have to restart the server.
Server side error:
```
host id {0} has been already deleted
UnityEngine.Networking.NetworkTransport:ReceiveFromHost(Int32, Int32&, Int32&, Byte[], Int32, Int32&, Byte&)
Dissonance.Integrations.UNet_LLAPI.UNetServer:ReadMessages() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetServer.cs:47)
Dissonance.Networking.BaseServer`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseServer.cs:78)
Dissonance.Networking.BaseCommsNetwork`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:124)
[Dissonance:Network] UNetServer: Error reading server socket: WrongHost
UnityEngine.Debug:LogError(Object)
Dissonance.Logs:SendLogMessage(String, Action`1) (at Assets/Plugins/Dissonance/Core/Log.cs:114)
Dissonance.Log:WriteLog(LogLevel, String) (at Assets/Plugins/Dissonance/Core/Log.cs:167)
Dissonance.Log:WriteLogFormat(LogLevel, String, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:190)
Dissonance.Log:Error(String, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:340)
Dissonance.Integrations.UNet_LLAPI.UNetServer:ReadMessages() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetServer.cs:50)
Dissonance.Networking.BaseServer`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseServer.cs:78)
Dissonance.Networking.BaseCommsNetwork`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:124)
```
Client-side error
```
[Dissonance:Network] UNetClient: Failed to cleanly disconnect from Dissonance server at 127.0.0.1:5889, Error WrongConnection
UnityEngine.Debug:LogError(Object)
Dissonance.Logs:SendLogMessage(String, Action`1) (at Assets/Plugins/Dissonance/Core/Log.cs:114)
Dissonance.Log:WriteLog(LogLevel, String) (at Assets/Plugins/Dissonance/Core/Log.cs:167)
Dissonance.Log:WriteLogFormat(LogLevel, String, String, UInt16, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:208)
Dissonance.Log:Error(String, String, UInt16, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:352)
Dissonance.Integrations.UNet_LLAPI.UNetClient:Disconnect() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetClient.cs:40)
Dissonance.Networking.BaseCommsNetwork`3:StopClient() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:186)
Dissonance.Integrations.UNet_LLAPI.Demo.LlapiManager:OnGUI() (at Assets/Dissonance/Integrations/UNet_LLAPI/Demo/LlapiManager.cs:45)
```
When restarting the server (full restart, close application, reopening the application) and do not stop the client and try to reconnect I get:
```
[Dissonance:Network] UNetClient: Exception invoking event handler: System.ArgumentNullException: Argument cannot be null.
Parameter name: key
at System.Collections.Generic.Dictionary`2[System.String,Dissonance.VoicePlayerState].TryGetValue (System.String key, Dissonance.VoicePlayerState& value) [0x000ba] in /Users/builduser/buildslave/mono/build/mcs/class/corlib/System.Collections.Generic/Dictionary.cs:608
at Dissonance.DissonanceComms.Net_PlayerStartedSpeaking (System.String player) [0x0000a] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\DissonanceComms.cs:417
at Dissonance.Networking.BaseCommsNetwork`3[Dissonance.Integrations.UNet_LLAPI.UNetServer,Dissonance.Integrations.UNet_LLAPI.UNetClient,System.Int32].OnPlayerStartedSpeaking (System.String obj) [0x00010] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseCommsNetwork.cs:296
at Dissonance.Networking.BaseClient`3[Dissonance.Integrations.UNet_LLAPI.UNetServer,Dissonance.Integrations.UNet_LLAPI.UNetClient,System.Int32].InvokeEvent[String] (System.String& arg, System.Action`1 handler) [0x0000f] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:316
UnityEngine.DebugLogHandler:Internal_Log(LogType, String, Object)
UnityEngine.DebugLogHandler:LogFormat(LogType, Object, String, Object[])
UnityEngine.Logger:Log(LogType, Object)
UnityEngine.Debug:LogError(Object)
Dissonance.Logs:SendLogMessage(String, Action`1) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:114)
Dissonance.Log:WriteLog(LogLevel, String) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:167)
Dissonance.Log:WriteLogFormat(LogLevel, String, Exception) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:190)
Dissonance.Log:Error(String, Exception) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:340)
Dissonance.Networking.BaseClient`3:InvokeEvent(String&, Action`1) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:320)
Dissonance.Networking.BaseClient`3:DispatchEvents() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:282)
Dissonance.Networking.BaseClient`3:Update() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:174)
Dissonance.Networking.BaseCommsNetwork`3:Update() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseCommsNetwork.cs:128)
```
The dicionary error may be resolved by just adding a check if state already exists for this key and if so just update/replace the state.I already tried that but it doesn't resolve the initial error.
Can you reproduce the error? If not I can upload you my test project. Or if you need anything else let me know.
Would be great if you could give a quick hint if this is a known issue and/or being worked on or not so that I can plan ahead.
Best
Marcel | priority | reconnection error hi we ve bought your asset a couple of days ago and trying out your hlapi as well as llapi integrations unfortunately llapi has some issues i m not sure how to resolve them and that seem unusual i m currently working with unity empty project dissonance package imported as well asllapi integration steps open dissonancellapidemo build project development build run build as client server run unity editor and connect may be flipped wait a sec maybe talk disconnect client wait a sec reconnect sometimes when reconnecting disconnecting something goes wrong and the following errors appear after that i m not able to connect to the server anymore and have to restart the server server side error host id has been already deleted unityengine networking networktransport receivefromhost byte byte dissonance integrations unet llapi unetserver readmessages at assets dissonance integrations unet llapi unetserver cs dissonance networking baseserver update at assets plugins dissonance core networking baseserver cs dissonance networking basecommsnetwork update at assets plugins dissonance core networking basecommsnetwork cs unetserver error reading server socket wronghost unityengine debug logerror object dissonance logs sendlogmessage string action at assets plugins dissonance core log cs dissonance log writelog loglevel string at assets plugins dissonance core log cs dissonance log writelogformat loglevel string networkerror at assets plugins dissonance core log cs dissonance log error string networkerror at assets plugins dissonance core log cs dissonance integrations unet llapi unetserver readmessages at assets dissonance integrations unet llapi unetserver cs dissonance networking baseserver update at assets plugins dissonance core networking baseserver cs dissonance networking basecommsnetwork update at assets plugins dissonance core networking basecommsnetwork cs client side error unetclient failed to cleanly disconnect from dissonance server at error wrongconnection unityengine debug logerror object dissonance logs sendlogmessage string action at assets plugins dissonance core log cs dissonance log writelog loglevel string at assets plugins dissonance core log cs dissonance log writelogformat loglevel string string networkerror at assets plugins dissonance core log cs dissonance log error string string networkerror at assets plugins dissonance core log cs dissonance integrations unet llapi unetclient disconnect at assets dissonance integrations unet llapi unetclient cs dissonance networking basecommsnetwork stopclient at assets plugins dissonance core networking basecommsnetwork cs dissonance integrations unet llapi demo llapimanager ongui at assets dissonance integrations unet llapi demo llapimanager cs when restarting the server full restart close application reopening the application and do not stop the client and try to reconnect i get unetclient exception invoking event handler system argumentnullexception argument cannot be null parameter name key at system collections generic dictionary trygetvalue system string key dissonance voiceplayerstate value in users builduser buildslave mono build mcs class corlib system collections generic dictionary cs at dissonance dissonancecomms net playerstartedspeaking system string player in d tests audioplugintest assets plugins dissonance dissonancecomms cs at dissonance networking basecommsnetwork onplayerstartedspeaking system string obj in d tests audioplugintest assets plugins dissonance core networking basecommsnetwork cs at dissonance networking baseclient invokeevent system string arg system action handler in d tests audioplugintest assets plugins dissonance core networking baseclient cs unityengine debugloghandler internal log logtype string object unityengine debugloghandler logformat logtype object string object unityengine logger log logtype object unityengine debug logerror object dissonance logs sendlogmessage string action at d tests audioplugintest assets plugins dissonance core log cs dissonance log writelog loglevel string at d tests audioplugintest assets plugins dissonance core log cs dissonance log writelogformat loglevel string exception at d tests audioplugintest assets plugins dissonance core log cs dissonance log error string exception at d tests audioplugintest assets plugins dissonance core log cs dissonance networking baseclient invokeevent string action at d tests audioplugintest assets plugins dissonance core networking baseclient cs dissonance networking baseclient dispatchevents at d tests audioplugintest assets plugins dissonance core networking baseclient cs dissonance networking baseclient update at d tests audioplugintest assets plugins dissonance core networking baseclient cs dissonance networking basecommsnetwork update at d tests audioplugintest assets plugins dissonance core networking basecommsnetwork cs the dicionary error may be resolved by just adding a check if state already exists for this key and if so just update replace the state i already tried that but it doesn t resolve the initial error can you reproduce the error if not i can upload you my test project or if you need anything else let me know would be great if you could give a quick hint if this is a known issue and or being worked on or not so that i can plan ahead best marcel | 1 |
33,935 | 7,302,940,520 | IssuesEvent | 2018-02-27 11:21:05 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | generic HazelcastException thrown, wrapping com.hazelcast.core.MemberLeftException | Team: Core Type: Defect |
exception
```
com.hazelcast.core.HazelcastException: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster!
at com.hazelcast.util.ExceptionUtil$1.create(ExceptionUtil.java:40)
at com.hazelcast.util.ExceptionUtil.peel(ExceptionUtil.java:116)
at com.hazelcast.util.ExceptionUtil.peel(ExceptionUtil.java:69)
at com.hazelcast.util.ExceptionUtil.rethrow(ExceptionUtil.java:129)
at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.getQueryResults(MapPublisherCreateWithValueMessageTask.java:98)
at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.call(MapPublisherCreateWithValueMessageTask.java:68)
at com.hazelcast.client.impl.protocol.task.AbstractCallableMessageTask.processMessage(AbstractCallableMessageTask.java:35)
at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:123)
at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:103)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
at java.lang.Thread.run(Thread.java:748)
at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64)
at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80)
at ------ submitted from ------.(Unknown Source)
at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:96)
at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:33)
at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:155)
at com.hazelcast.client.impl.querycache.subscriber.ClientInvokerWrapper.invoke(ClientInvokerWrapper.java:89)
at com.hazelcast.client.impl.querycache.subscriber.ClientQueryCacheEndToEndConstructor.createPublishAccumulatorWithIncludeValue(ClientQueryCacheEndToEndConstructor.java:68)
at com.hazelcast.client.impl.querycache.subscriber.ClientQueryCacheEndToEndConstructor.createPublisherAccumulator(ClientQueryCacheEndToEndConstructor.java:49)
at com.hazelcast.map.impl.querycache.subscriber.AbstractQueryCacheEndToEndConstructor.createNew(AbstractQueryCacheEndToEndConstructor.java:90)
at com.hazelcast.map.impl.querycache.subscriber.AbstractQueryCacheEndToEndConstructor.createNew(AbstractQueryCacheEndToEndConstructor.java:40)
at com.hazelcast.map.impl.querycache.subscriber.QueryCacheEndToEndProvider.getOrCreateQueryCache(QueryCacheEndToEndProvider.java:60)
at com.hazelcast.client.proxy.ClientMapProxy.createQueryCache(ClientMapProxy.java:1494)
at com.hazelcast.client.proxy.ClientMapProxy.getQueryCacheInternal(ClientMapProxy.java:1487)
at com.hazelcast.client.proxy.ClientMapProxy.getQueryCache(ClientMapProxy.java:1464)
at hzcmd.map.cqc.CreateUseDestroy.timeStep(CreateUseDestroy.java:16)
at remote.bench.marker.MetricsMarker.flatOut(MetricsMarker.java:53)
at remote.bench.marker.MetricsMarker.bench(MetricsMarker.java:40)
at remote.bench.BenchThread.call(BenchThread.java:38)
at remote.bench.BenchThread.call(BenchThread.java:12)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:473)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
at java.lang.Thread.run(Thread.java:748)
Caused by: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster!
at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$OnMemberLeftTask.onTargetLoss(InvocationMonitor.java:392)
at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$OnMemberLeftTask.run0(InvocationMonitor.java:359)
at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$MonitorTask.run(InvocationMonitor.java:240)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:473)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
at java.lang.Thread.run(Thread.java:748)
at ------ submitted from ------.(Unknown Source)
at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolve(InvocationFuture.java:127)
at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolveAndThrowIfException(InvocationFuture.java:79)
at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:155)
at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.getQueryResults(MapPublisherCreateWithValueMessageTask.java:96)
at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.call(MapPublisherCreateWithValueMessageTask.java:68)
at com.hazelcast.client.impl.protocol.task.AbstractCallableMessageTask.processMessage(AbstractCallableMessageTask.java:35)
at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:123)
at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:103)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
at java.lang.Thread.run(Thread.java:748)
at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64)
at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80)
```
from
https://hazelcast-l337.ci.cloudbees.com/view/kill/job/kill-All/33/console
/disk1/jenkins/workspace/kill-All/3.10-SNAPSHOT/2018_02_21-15_05_39/stable/destroy/create-use-destroy Failed
```
fail HzClient3HZ _create-use-destroy_createUseDestroy_mapBak1HD_queryCache hzcmd.map.cqc.CreateUseDestroy threadId=0 com.hazelcast.core.HazelcastException: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster!
```
http://54.82.84.143/~jenkins/workspace/kill-All/3.10-SNAPSHOT/2018_02_21-15_05_39/stable/destroy/create-use-destroy | 1.0 | generic HazelcastException thrown, wrapping com.hazelcast.core.MemberLeftException -
exception
```
com.hazelcast.core.HazelcastException: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster!
at com.hazelcast.util.ExceptionUtil$1.create(ExceptionUtil.java:40)
at com.hazelcast.util.ExceptionUtil.peel(ExceptionUtil.java:116)
at com.hazelcast.util.ExceptionUtil.peel(ExceptionUtil.java:69)
at com.hazelcast.util.ExceptionUtil.rethrow(ExceptionUtil.java:129)
at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.getQueryResults(MapPublisherCreateWithValueMessageTask.java:98)
at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.call(MapPublisherCreateWithValueMessageTask.java:68)
at com.hazelcast.client.impl.protocol.task.AbstractCallableMessageTask.processMessage(AbstractCallableMessageTask.java:35)
at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:123)
at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:103)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
at java.lang.Thread.run(Thread.java:748)
at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64)
at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80)
at ------ submitted from ------.(Unknown Source)
at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:96)
at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:33)
at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:155)
at com.hazelcast.client.impl.querycache.subscriber.ClientInvokerWrapper.invoke(ClientInvokerWrapper.java:89)
at com.hazelcast.client.impl.querycache.subscriber.ClientQueryCacheEndToEndConstructor.createPublishAccumulatorWithIncludeValue(ClientQueryCacheEndToEndConstructor.java:68)
at com.hazelcast.client.impl.querycache.subscriber.ClientQueryCacheEndToEndConstructor.createPublisherAccumulator(ClientQueryCacheEndToEndConstructor.java:49)
at com.hazelcast.map.impl.querycache.subscriber.AbstractQueryCacheEndToEndConstructor.createNew(AbstractQueryCacheEndToEndConstructor.java:90)
at com.hazelcast.map.impl.querycache.subscriber.AbstractQueryCacheEndToEndConstructor.createNew(AbstractQueryCacheEndToEndConstructor.java:40)
at com.hazelcast.map.impl.querycache.subscriber.QueryCacheEndToEndProvider.getOrCreateQueryCache(QueryCacheEndToEndProvider.java:60)
at com.hazelcast.client.proxy.ClientMapProxy.createQueryCache(ClientMapProxy.java:1494)
at com.hazelcast.client.proxy.ClientMapProxy.getQueryCacheInternal(ClientMapProxy.java:1487)
at com.hazelcast.client.proxy.ClientMapProxy.getQueryCache(ClientMapProxy.java:1464)
at hzcmd.map.cqc.CreateUseDestroy.timeStep(CreateUseDestroy.java:16)
at remote.bench.marker.MetricsMarker.flatOut(MetricsMarker.java:53)
at remote.bench.marker.MetricsMarker.bench(MetricsMarker.java:40)
at remote.bench.BenchThread.call(BenchThread.java:38)
at remote.bench.BenchThread.call(BenchThread.java:12)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:473)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
at java.lang.Thread.run(Thread.java:748)
Caused by: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster!
at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$OnMemberLeftTask.onTargetLoss(InvocationMonitor.java:392)
at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$OnMemberLeftTask.run0(InvocationMonitor.java:359)
at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$MonitorTask.run(InvocationMonitor.java:240)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:473)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
at java.lang.Thread.run(Thread.java:748)
at ------ submitted from ------.(Unknown Source)
at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolve(InvocationFuture.java:127)
at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolveAndThrowIfException(InvocationFuture.java:79)
at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:155)
at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.getQueryResults(MapPublisherCreateWithValueMessageTask.java:96)
at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.call(MapPublisherCreateWithValueMessageTask.java:68)
at com.hazelcast.client.impl.protocol.task.AbstractCallableMessageTask.processMessage(AbstractCallableMessageTask.java:35)
at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:123)
at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:103)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
at java.lang.Thread.run(Thread.java:748)
at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64)
at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80)
```
from
https://hazelcast-l337.ci.cloudbees.com/view/kill/job/kill-All/33/console
/disk1/jenkins/workspace/kill-All/3.10-SNAPSHOT/2018_02_21-15_05_39/stable/destroy/create-use-destroy Failed
```
fail HzClient3HZ _create-use-destroy_createUseDestroy_mapBak1HD_queryCache hzcmd.map.cqc.CreateUseDestroy threadId=0 com.hazelcast.core.HazelcastException: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster!
```
http://54.82.84.143/~jenkins/workspace/kill-All/3.10-SNAPSHOT/2018_02_21-15_05_39/stable/destroy/create-use-destroy | non_priority | generic hazelcastexception thrown wrapping com hazelcast core memberleftexception exception com hazelcast core hazelcastexception com hazelcast core memberleftexception member has left cluster at com hazelcast util exceptionutil create exceptionutil java at com hazelcast util exceptionutil peel exceptionutil java at com hazelcast util exceptionutil peel exceptionutil java at com hazelcast util exceptionutil rethrow exceptionutil java at com hazelcast client impl protocol task map mappublishercreatewithvaluemessagetask getqueryresults mappublishercreatewithvaluemessagetask java at com hazelcast client impl protocol task map mappublishercreatewithvaluemessagetask call mappublishercreatewithvaluemessagetask java at com hazelcast client impl protocol task abstractcallablemessagetask processmessage abstractcallablemessagetask java at com hazelcast client impl protocol task abstractmessagetask initializeandprocessmessage abstractmessagetask java at com hazelcast client impl protocol task abstractmessagetask run abstractmessagetask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast util executor hazelcastmanagedthread run hazelcastmanagedthread java at submitted from unknown source at com hazelcast client spi impl clientinvocationfuture resolveandthrowifexception clientinvocationfuture java at com hazelcast client spi impl clientinvocationfuture resolveandthrowifexception clientinvocationfuture java at com hazelcast spi impl abstractinvocationfuture get abstractinvocationfuture java at com hazelcast client impl querycache subscriber clientinvokerwrapper invoke clientinvokerwrapper java at com hazelcast client impl querycache subscriber clientquerycacheendtoendconstructor createpublishaccumulatorwithincludevalue clientquerycacheendtoendconstructor java at com hazelcast client impl querycache subscriber clientquerycacheendtoendconstructor createpublisheraccumulator clientquerycacheendtoendconstructor java at com hazelcast map impl querycache subscriber abstractquerycacheendtoendconstructor createnew abstractquerycacheendtoendconstructor java at com hazelcast map impl querycache subscriber abstractquerycacheendtoendconstructor createnew abstractquerycacheendtoendconstructor java at com hazelcast map impl querycache subscriber querycacheendtoendprovider getorcreatequerycache querycacheendtoendprovider java at com hazelcast client proxy clientmapproxy createquerycache clientmapproxy java at com hazelcast client proxy clientmapproxy getquerycacheinternal clientmapproxy java at com hazelcast client proxy clientmapproxy getquerycache clientmapproxy java at hzcmd map cqc createusedestroy timestep createusedestroy java at remote bench marker metricsmarker flatout metricsmarker java at remote bench marker metricsmarker bench metricsmarker java at remote bench benchthread call benchthread java at remote bench benchthread call benchthread java at java util concurrent futuretask run futuretask java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by com hazelcast core memberleftexception member has left cluster at com hazelcast spi impl operationservice impl invocationmonitor onmemberlefttask ontargetloss invocationmonitor java at com hazelcast spi impl operationservice impl invocationmonitor onmemberlefttask invocationmonitor java at com hazelcast spi impl operationservice impl invocationmonitor monitortask run invocationmonitor java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at submitted from unknown source at com hazelcast spi impl operationservice impl invocationfuture resolve invocationfuture java at com hazelcast spi impl operationservice impl invocationfuture resolveandthrowifexception invocationfuture java at com hazelcast spi impl abstractinvocationfuture get abstractinvocationfuture java at com hazelcast client impl protocol task map mappublishercreatewithvaluemessagetask getqueryresults mappublishercreatewithvaluemessagetask java at com hazelcast client impl protocol task map mappublishercreatewithvaluemessagetask call mappublishercreatewithvaluemessagetask java at com hazelcast client impl protocol task abstractcallablemessagetask processmessage abstractcallablemessagetask java at com hazelcast client impl protocol task abstractmessagetask initializeandprocessmessage abstractmessagetask java at com hazelcast client impl protocol task abstractmessagetask run abstractmessagetask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast util executor hazelcastmanagedthread run hazelcastmanagedthread java from jenkins workspace kill all snapshot stable destroy create use destroy failed fail create use destroy createusedestroy querycache hzcmd map cqc createusedestroy threadid com hazelcast core hazelcastexception com hazelcast core memberleftexception member has left cluster | 0 |
240,153 | 26,254,327,648 | IssuesEvent | 2023-01-05 22:33:07 | TreyM-WSS/terra-clinical | https://api.github.com/repos/TreyM-WSS/terra-clinical | opened | CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz | security vulnerability | ## CVE-2021-3918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary>
<p>JSON Schema validation and specifications</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/json-schema/package.json</p>
<p>
Dependency Hierarchy:
- terra-toolkit-6.0.0.tgz (Root Library)
- node-sass-4.14.0.tgz
- request-2.88.2.tgz
- http-signature-1.2.0.tgz
- jsprim-1.4.1.tgz
- :x: **json-schema-0.2.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TreyM-WSS/terra-clinical/commit/0940364474cdf2812b34bb24824139d7e3f8060b">0940364474cdf2812b34bb24824139d7e3f8060b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-11-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3918>CVE-2021-3918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p>
<p>Release Date: 2021-11-13</p>
<p>Fix Resolution (json-schema): 0.4.0</p>
<p>Direct dependency fix Resolution (terra-toolkit): 6.1.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | True | CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - ## CVE-2021-3918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary>
<p>JSON Schema validation and specifications</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/json-schema/package.json</p>
<p>
Dependency Hierarchy:
- terra-toolkit-6.0.0.tgz (Root Library)
- node-sass-4.14.0.tgz
- request-2.88.2.tgz
- http-signature-1.2.0.tgz
- jsprim-1.4.1.tgz
- :x: **json-schema-0.2.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TreyM-WSS/terra-clinical/commit/0940364474cdf2812b34bb24824139d7e3f8060b">0940364474cdf2812b34bb24824139d7e3f8060b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-11-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3918>CVE-2021-3918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p>
<p>Release Date: 2021-11-13</p>
<p>Fix Resolution (json-schema): 0.4.0</p>
<p>Direct dependency fix Resolution (terra-toolkit): 6.1.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | non_priority | cve high detected in json schema tgz cve high severity vulnerability vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file package json path to vulnerable library node modules json schema package json dependency hierarchy terra toolkit tgz root library node sass tgz request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in head commit a href found in base branch master vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution json schema direct dependency fix resolution terra toolkit rescue worker helmet automatic remediation is available for this issue | 0 |
14,859 | 9,546,931,024 | IssuesEvent | 2019-05-01 21:25:23 | mercycorps/TolaActivity | https://api.github.com/repos/mercycorps/TolaActivity | closed | Add meaningful error messages for failed Django login on login page | Deploy Ready Verified usability | Currently, failing to enter a valid u/p combination on the Django login returns me silently to the login page. We need to add:
- [ ] ~link to reset password or other help]~ _(no longer addressing in this ticket)_
- [x] error message(s) | True | Add meaningful error messages for failed Django login on login page - Currently, failing to enter a valid u/p combination on the Django login returns me silently to the login page. We need to add:
- [ ] ~link to reset password or other help]~ _(no longer addressing in this ticket)_
- [x] error message(s) | non_priority | add meaningful error messages for failed django login on login page currently failing to enter a valid u p combination on the django login returns me silently to the login page we need to add link to reset password or other help no longer addressing in this ticket error message s | 0 |
604 | 3,003,590,670 | IssuesEvent | 2015-07-25 02:32:52 | mesosphere/marathon | https://api.github.com/repos/mesosphere/marathon | closed | docker: hostPath in volumes seems to be ignored | bug OKR Usability service | mesos: 0.22.1
marathon: 0.8.1
docker: 1.6.2
I define:
```
"volumes": [{
"containerPath": "/var/log/vimana",
"hostPath": "/var/log/vimana"
}]
```
But the mounts I get are:
```
"Volumes": {
"/mnt/mesos/sandbox": "/tmp/mesos/slaves/20150522-122903-2693333002-5050-7694-S9/frameworks/20150522-085224-378208266-5050-5138-0000/executors
"/var/log/vimana": "/var/lib/docker/vfs/dir/38cf9ddf3c342baec70ade21a9cce2c267f909986f06cc101f5ecb672ae99f40"
},
```
The volume is declared in the docker file, so this behavior seems to as if the containerizer just dropped my mount. | 1.0 | docker: hostPath in volumes seems to be ignored - mesos: 0.22.1
marathon: 0.8.1
docker: 1.6.2
I define:
```
"volumes": [{
"containerPath": "/var/log/vimana",
"hostPath": "/var/log/vimana"
}]
```
But the mounts I get are:
```
"Volumes": {
"/mnt/mesos/sandbox": "/tmp/mesos/slaves/20150522-122903-2693333002-5050-7694-S9/frameworks/20150522-085224-378208266-5050-5138-0000/executors
"/var/log/vimana": "/var/lib/docker/vfs/dir/38cf9ddf3c342baec70ade21a9cce2c267f909986f06cc101f5ecb672ae99f40"
},
```
The volume is declared in the docker file, so this behavior seems to as if the containerizer just dropped my mount. | non_priority | docker hostpath in volumes seems to be ignored mesos marathon docker i define volumes containerpath var log vimana hostpath var log vimana but the mounts i get are volumes mnt mesos sandbox tmp mesos slaves frameworks executors var log vimana var lib docker vfs dir the volume is declared in the docker file so this behavior seems to as if the containerizer just dropped my mount | 0 |
697,530 | 23,942,680,719 | IssuesEvent | 2022-09-12 02:25:29 | jrsteensen/OpenHornet | https://api.github.com/repos/jrsteensen/OpenHornet | closed | Update Native F360 Stick Model | Type: Enhancement Category: MCAD Priority: Normal | - [x] Add new hall sensors to pitch axis
- [x] Add new hall sensor to roll axis
- [x] Mount Electronics to base
- [x] Rename everything to OH PNs
- [x] Update File Properties on all subcomponents.
- [ ] Design an enclosure for the controller PCB to give it some protection. | 1.0 | Update Native F360 Stick Model - - [x] Add new hall sensors to pitch axis
- [x] Add new hall sensor to roll axis
- [x] Mount Electronics to base
- [x] Rename everything to OH PNs
- [x] Update File Properties on all subcomponents.
- [ ] Design an enclosure for the controller PCB to give it some protection. | priority | update native stick model add new hall sensors to pitch axis add new hall sensor to roll axis mount electronics to base rename everything to oh pns update file properties on all subcomponents design an enclosure for the controller pcb to give it some protection | 1 |
243,141 | 7,853,763,521 | IssuesEvent | 2018-06-20 18:29:12 | canmet-energy/btap_tasks | https://api.github.com/repos/canmet-energy/btap_tasks | closed | Merge nrcan into develop | Priority High Standards | The schedule has moved up and we need to get our code into NREL's for the next release
Due Monday the 11th of June. | 1.0 | Merge nrcan into develop - The schedule has moved up and we need to get our code into NREL's for the next release
Due Monday the 11th of June. | priority | merge nrcan into develop the schedule has moved up and we need to get our code into nrel s for the next release due monday the of june | 1 |
263,229 | 28,029,746,253 | IssuesEvent | 2023-03-28 11:35:20 | RG4421/ampere-centos-kernel | https://api.github.com/repos/RG4421/ampere-centos-kernel | reopened | CVE-2022-1974 (Medium) detected in linuxv5.2 | Mend: dependency security vulnerability | ## CVE-2022-1974 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A use-after-free flaw was found in the Linux kernel's NFC core functionality due to a race condition between kobject creation and delete. This vulnerability allows a local attacker with CAP_NET_ADMIN privilege to leak kernel information.
<p>Publish Date: 2022-08-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1974>CVE-2022-1974</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1974">https://www.linuxkernelcves.com/cves/CVE-2022-1974</a></p>
<p>Release Date: 2022-08-31</p>
<p>Fix Resolution: v4.9.313,v4.14.278,v4.19.242,v5.4.193,v5.10.115,v5.15.39,v5.17.7</p>
</p>
</details>
<p></p>
| True | CVE-2022-1974 (Medium) detected in linuxv5.2 - ## CVE-2022-1974 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A use-after-free flaw was found in the Linux kernel's NFC core functionality due to a race condition between kobject creation and delete. This vulnerability allows a local attacker with CAP_NET_ADMIN privilege to leak kernel information.
<p>Publish Date: 2022-08-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1974>CVE-2022-1974</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1974">https://www.linuxkernelcves.com/cves/CVE-2022-1974</a></p>
<p>Release Date: 2022-08-31</p>
<p>Fix Resolution: v4.9.313,v4.14.278,v4.19.242,v5.4.193,v5.10.115,v5.15.39,v5.17.7</p>
</p>
</details>
<p></p>
| non_priority | cve medium detected in cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files vulnerability details a use after free flaw was found in the linux kernel s nfc core functionality due to a race condition between kobject creation and delete this vulnerability allows a local attacker with cap net admin privilege to leak kernel information publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
178,605 | 6,612,727,926 | IssuesEvent | 2017-09-20 06:02:28 | arquillian/smart-testing | https://api.github.com/repos/arquillian/smart-testing | closed | NPE is thrown is strategy is mispelled | Component: Maven Priority: High Type: Bug | ##### Issue Overview
NPE is thrown if you set a strategy incorrectly
##### Expected Behaviour
Throw a meaningful exception instead of NPE.
##### Current Behaviour
NPE
| 1.0 | NPE is thrown is strategy is mispelled - ##### Issue Overview
NPE is thrown if you set a strategy incorrectly
##### Expected Behaviour
Throw a meaningful exception instead of NPE.
##### Current Behaviour
NPE
| priority | npe is thrown is strategy is mispelled issue overview npe is thrown if you set a strategy incorrectly expected behaviour throw a meaningful exception instead of npe current behaviour npe | 1 |
366,200 | 25,572,768,437 | IssuesEvent | 2022-11-30 19:08:57 | Westlake-AI/openmixup | https://api.github.com/repos/Westlake-AI/openmixup | opened | Release Models of Mixups and MogaNet and Update Features in V0.2.6 | documentation enhancement update | Updateing new features:
1. Fix the classification heads and update implementations and config files of [AlexNet](https://dl.acm.org/doi/10.1145/3065386) and [InceptionV3](https://arxiv.org/abs/1512.00567).
Uploading Benchmark Results (release):
1. Release pre-trained models and logs of mixup benchmarks on ImageNet as provided in [AutoMix](https://arxiv.org/abs/2103.13027) and update results in [Model_Zoo_sup.md](https://github.com/Westlake-AI/openmixup/tree/main/docs/en/model_zoos/Model_Zoo_sup.md).
2. Release pre-trained models and logs of [MogaNet](https://arxiv.org/abs/2211.03295).
Updating documents:
1. Update documents of mixup benchmarks on ImageNet with new backbones in [Model_Zoo_sup.md](https://github.com/Westlake-AI/openmixup/tree/main/docs/en/model_zoos/Model_Zoo_sup.md).
2. Update awesome lists in [Awesome Mixups](docs/en/awesome_selfsup/MIM.md) and [Awesome MIM](docs/en/awesome_selfsup/MIM.md) and provide teaser figures of most papers as illustrations. | 1.0 | Release Models of Mixups and MogaNet and Update Features in V0.2.6 - Updateing new features:
1. Fix the classification heads and update implementations and config files of [AlexNet](https://dl.acm.org/doi/10.1145/3065386) and [InceptionV3](https://arxiv.org/abs/1512.00567).
Uploading Benchmark Results (release):
1. Release pre-trained models and logs of mixup benchmarks on ImageNet as provided in [AutoMix](https://arxiv.org/abs/2103.13027) and update results in [Model_Zoo_sup.md](https://github.com/Westlake-AI/openmixup/tree/main/docs/en/model_zoos/Model_Zoo_sup.md).
2. Release pre-trained models and logs of [MogaNet](https://arxiv.org/abs/2211.03295).
Updating documents:
1. Update documents of mixup benchmarks on ImageNet with new backbones in [Model_Zoo_sup.md](https://github.com/Westlake-AI/openmixup/tree/main/docs/en/model_zoos/Model_Zoo_sup.md).
2. Update awesome lists in [Awesome Mixups](docs/en/awesome_selfsup/MIM.md) and [Awesome MIM](docs/en/awesome_selfsup/MIM.md) and provide teaser figures of most papers as illustrations. | non_priority | release models of mixups and moganet and update features in updateing new features fix the classification heads and update implementations and config files of and uploading benchmark results release release pre trained models and logs of mixup benchmarks on imagenet as provided in and update results in release pre trained models and logs of updating documents update documents of mixup benchmarks on imagenet with new backbones in update awesome lists in docs en awesome selfsup mim md and docs en awesome selfsup mim md and provide teaser figures of most papers as illustrations | 0 |
587,294 | 17,612,267,406 | IssuesEvent | 2021-08-18 04:08:24 | goplus/gop | https://api.github.com/repos/goplus/gop | closed | repl continueMode bug | bug priority:low | 
Cant input multiline code.
go version
go version go1.14.3 darwin/amd64
os: macOS Catalina 10.5.7 | 1.0 | repl continueMode bug - 
Cant input multiline code.
go version
go version go1.14.3 darwin/amd64
os: macOS Catalina 10.5.7 | priority | repl continuemode bug cant input multiline code go version go version darwin os macos catalina | 1 |
61,161 | 17,023,621,376 | IssuesEvent | 2021-07-03 02:58:09 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | search from IOS is crashing | Component: nominatim Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 10.44am, Wednesday, 4th August 2010]**
Trying to do a search containing a "special keyword" (POI) like:
http://nominatim.openstreetmap.org/search/airport
from Safari (or any other web browser) on a IOS device (iPhone/iPad) does result in:
string(354) "select place_id,false as in_small,false as in_large from search_name where name_vector @> ARRAY[] and ST_DWithin(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326), 0.1) order by ST_Distance(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326)) ASC, ST_Distance(ST_SetSRID(ST_Point(9.584734,47.659290),4326), centroid) asc, search_rank ASC limit 10"
object(DB_Error)#2 (8) {
["error_message_prefix"]=>
string(0) ""
["mode"]=>
int(1)
["level"]=>
int(1024)
["code"]=>
int(-2)
["message"]=>
string(22) "DB Error: syntax error"
["userinfo"]=>
string(541) "select place_id,false as in_small,false as in_large from search_name where name_vector @> ARRAY[] and ST_DWithin(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326), 0.1) order by ST_Distance(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326)) ASC, ST_Distance(ST_SetSRID(ST_Point(9.584734,47.659290),4326), centroid) asc, search_rank ASC limit 10 [nativecode=ERROR: syntax error at or near "]"
LINE 1: ...large from search_name where name_vector @> ARRAY[] and ST_D...
...
...
... | 1.0 | search from IOS is crashing - **[Submitted to the original trac issue database at 10.44am, Wednesday, 4th August 2010]**
Trying to do a search containing a "special keyword" (POI) like:
http://nominatim.openstreetmap.org/search/airport
from Safari (or any other web browser) on a IOS device (iPhone/iPad) does result in:
string(354) "select place_id,false as in_small,false as in_large from search_name where name_vector @> ARRAY[] and ST_DWithin(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326), 0.1) order by ST_Distance(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326)) ASC, ST_Distance(ST_SetSRID(ST_Point(9.584734,47.659290),4326), centroid) asc, search_rank ASC limit 10"
object(DB_Error)#2 (8) {
["error_message_prefix"]=>
string(0) ""
["mode"]=>
int(1)
["level"]=>
int(1024)
["code"]=>
int(-2)
["message"]=>
string(22) "DB Error: syntax error"
["userinfo"]=>
string(541) "select place_id,false as in_small,false as in_large from search_name where name_vector @> ARRAY[] and ST_DWithin(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326), 0.1) order by ST_Distance(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326)) ASC, ST_Distance(ST_SetSRID(ST_Point(9.584734,47.659290),4326), centroid) asc, search_rank ASC limit 10 [nativecode=ERROR: syntax error at or near "]"
LINE 1: ...large from search_name where name_vector @> ARRAY[] and ST_D...
...
...
... | non_priority | search from ios is crashing trying to do a search containing a special keyword poi like from safari or any other web browser on a ios device iphone ipad does result in string select place id false as in small false as in large from search name where name vector array and st dwithin centroid st setsrid st point order by st distance centroid st setsrid st point asc st distance st setsrid st point centroid asc search rank asc limit object db error string int int int string db error syntax error string select place id false as in small false as in large from search name where name vector array and st dwithin centroid st setsrid st point order by st distance centroid st setsrid st point asc st distance st setsrid st point centroid asc search rank asc limit line large from search name where name vector array and st d | 0 |
50,536 | 13,539,630,516 | IssuesEvent | 2020-09-16 13:42:28 | cniweb/missing-link | https://api.github.com/repos/cniweb/missing-link | opened | CVE-2020-1945 (Medium) detected in ant-1.8.2.jar | security vulnerability | ## CVE-2020-1945 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ant-1.8.2.jar</b></p></summary>
<p>master POM</p>
<p>Path to vulnerable library: missing-link/ant-props/lib/apache-ant/apache-ant-1.8.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **ant-1.8.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/cniweb/missing-link/commit/e22e6226faae42f8975654b836d8cd2d6de9967c">e22e6226faae42f8975654b836d8cd2d6de9967c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Ant 1.1 to 1.9.14 and 1.10.0 to 1.10.7 uses the default temporary directory identified by the Java system property java.io.tmpdir for several tasks and may thus leak sensitive information. The fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process.
<p>Publish Date: 2020-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1945>CVE-2020-1945</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://ant.apache.org/security.html">https://ant.apache.org/security.html</a></p>
<p>Release Date: 2020-05-14</p>
<p>Fix Resolution: org.apache.ant:ant-junitlauncher:1.10.8;org.apache.ant:ant:1.9.15,1.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-1945 (Medium) detected in ant-1.8.2.jar - ## CVE-2020-1945 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ant-1.8.2.jar</b></p></summary>
<p>master POM</p>
<p>Path to vulnerable library: missing-link/ant-props/lib/apache-ant/apache-ant-1.8.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **ant-1.8.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/cniweb/missing-link/commit/e22e6226faae42f8975654b836d8cd2d6de9967c">e22e6226faae42f8975654b836d8cd2d6de9967c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Ant 1.1 to 1.9.14 and 1.10.0 to 1.10.7 uses the default temporary directory identified by the Java system property java.io.tmpdir for several tasks and may thus leak sensitive information. The fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process.
<p>Publish Date: 2020-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1945>CVE-2020-1945</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://ant.apache.org/security.html">https://ant.apache.org/security.html</a></p>
<p>Release Date: 2020-05-14</p>
<p>Fix Resolution: org.apache.ant:ant-junitlauncher:1.10.8;org.apache.ant:ant:1.9.15,1.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in ant jar cve medium severity vulnerability vulnerable library ant jar master pom path to vulnerable library missing link ant props lib apache ant apache ant jar dependency hierarchy x ant jar vulnerable library found in head commit a href found in base branch master vulnerability details apache ant to and to uses the default temporary directory identified by the java system property java io tmpdir for several tasks and may thus leak sensitive information the fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache ant ant junitlauncher org apache ant ant step up your open source security game with whitesource | 0 |
230,392 | 7,609,805,441 | IssuesEvent | 2018-05-01 03:11:06 | Gamebuster19901/InventoryDecrapifier | https://api.github.com/repos/Gamebuster19901/InventoryDecrapifier | closed | Items are picked up one at a time | Bug Priority - Normal ↓ Side - Client Side - Server | Items are picked up from the ground one at a time at a rate of one per tick, instead of multiple per tick. | 1.0 | Items are picked up one at a time - Items are picked up from the ground one at a time at a rate of one per tick, instead of multiple per tick. | priority | items are picked up one at a time items are picked up from the ground one at a time at a rate of one per tick instead of multiple per tick | 1 |
713,347 | 24,525,461,879 | IssuesEvent | 2022-10-11 12:49:58 | quadratic-funding/mpc-phase2-suite | https://api.github.com/repos/quadratic-funding/mpc-phase2-suite | closed | Get access to cloud resource | DevOps ⚙ High Priority 🔥 | ### Description
We need to get access for GCP and AWS cloud resources in order to switch from personal billing account to EF one. We need access to Firebase and GCP Cloud Functions + Compute Engine, AWS S3 (possibly AWS EC2). | 1.0 | Get access to cloud resource - ### Description
We need to get access for GCP and AWS cloud resources in order to switch from personal billing account to EF one. We need access to Firebase and GCP Cloud Functions + Compute Engine, AWS S3 (possibly AWS EC2). | priority | get access to cloud resource description we need to get access for gcp and aws cloud resources in order to switch from personal billing account to ef one we need access to firebase and gcp cloud functions compute engine aws possibly aws | 1 |
48,885 | 13,184,766,840 | IssuesEvent | 2020-08-12 20:03:21 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | buggy nutau (Trac #383) | Incomplete Migration Migrated from Trac combo simulation defect | <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/383
, reported by olivas and owned by olivas_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-22T18:26:26",
"description": "it seems to me that in the nutau nugen dataset 6539 the light emission\nfrom muons (if present in the event) is missing. I attach a scatter plot\nof muon energy vs muon track length (LDir) in IceCube. This should depend\nonly on the muon energy (and vertex position), but not on the primary\ntype. However, the reconstructed track length is aleays around 100m for\nnu-tau primary. That value can be reached by the cascade at sufficiently\nhigh energy.\n\nIt looks like there are two populations of muons in the tau MC, as I would\nexpect: one at very low muon energy (below 10 GeV) originating in the\ncascade. The other component is at higher energies (>100 GeV) and is\noriginating in the 17% tau->muon decay channel. So the particle phyiscs\nseems okay in that MC, but it was probably missed to simulate the light\nfrom the muon. This makes dataset 6539 unusable. We need urgently a fixed\nnutau dataset.\n\n Cheers,\n Andreas\n\n\nThis is likely due to the way nugen uses the propagators. For NuTau it uses the tau\npropagator and would not propagate muons in the final state.",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1416680786826176",
"component": "combo simulation",
"summary": "buggy nutau",
"priority": "normal",
"keywords": "",
"time": "2012-03-23T19:23:00",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| 1.0 | buggy nutau (Trac #383) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/383
, reported by olivas and owned by olivas_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-22T18:26:26",
"description": "it seems to me that in the nutau nugen dataset 6539 the light emission\nfrom muons (if present in the event) is missing. I attach a scatter plot\nof muon energy vs muon track length (LDir) in IceCube. This should depend\nonly on the muon energy (and vertex position), but not on the primary\ntype. However, the reconstructed track length is aleays around 100m for\nnu-tau primary. That value can be reached by the cascade at sufficiently\nhigh energy.\n\nIt looks like there are two populations of muons in the tau MC, as I would\nexpect: one at very low muon energy (below 10 GeV) originating in the\ncascade. The other component is at higher energies (>100 GeV) and is\noriginating in the 17% tau->muon decay channel. So the particle phyiscs\nseems okay in that MC, but it was probably missed to simulate the light\nfrom the muon. This makes dataset 6539 unusable. We need urgently a fixed\nnutau dataset.\n\n Cheers,\n Andreas\n\n\nThis is likely due to the way nugen uses the propagators. For NuTau it uses the tau\npropagator and would not propagate muons in the final state.",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1416680786826176",
"component": "combo simulation",
"summary": "buggy nutau",
"priority": "normal",
"keywords": "",
"time": "2012-03-23T19:23:00",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| non_priority | buggy nutau trac migrated from reported by olivas and owned by olivas json status closed changetime description it seems to me that in the nutau nugen dataset the light emission nfrom muons if present in the event is missing i attach a scatter plot nof muon energy vs muon track length ldir in icecube this should depend nonly on the muon energy and vertex position but not on the primary ntype however the reconstructed track length is aleays around for nnu tau primary that value can be reached by the cascade at sufficiently nhigh energy n nit looks like there are two populations of muons in the tau mc as i would nexpect one at very low muon energy below gev originating in the ncascade the other component is at higher energies gev and is noriginating in the tau muon decay channel so the particle phyiscs nseems okay in that mc but it was probably missed to simulate the light nfrom the muon this makes dataset unusable we need urgently a fixed nnutau dataset n n cheers n andreas n n nthis is likely due to the way nugen uses the propagators for nutau it uses the tau npropagator and would not propagate muons in the final state reporter olivas cc resolution fixed ts component combo simulation summary buggy nutau priority normal keywords time milestone owner olivas type defect | 0 |
142,936 | 11,500,952,658 | IssuesEvent | 2020-02-12 16:21:52 | wazuh/wazuh-qa | https://api.github.com/repos/wazuh/wazuh-qa | opened | FIM System tests: Create common tasks to verify alerts on alerts.json and Elasticsearch | fim-system-tests | ### Objective
The purpose of this issue is to keep track of the progress of tasks that parse the generated file from actions tasks at https://github.com/wazuh/wazuh-qa/issues/444 and ensures that an alert has been created for every file creation/modification/deletion
### Tasks
- [ ] Create tasks to compare generated files with the alerts at `alerts.json` file. (assigned @Zenidd )
- [ ] Create tasks to compare generated files with the alerts at Elasticsearch index.
| 1.0 | FIM System tests: Create common tasks to verify alerts on alerts.json and Elasticsearch - ### Objective
The purpose of this issue is to keep track of the progress of tasks that parse the generated file from actions tasks at https://github.com/wazuh/wazuh-qa/issues/444 and ensures that an alert has been created for every file creation/modification/deletion
### Tasks
- [ ] Create tasks to compare generated files with the alerts at `alerts.json` file. (assigned @Zenidd )
- [ ] Create tasks to compare generated files with the alerts at Elasticsearch index.
| non_priority | fim system tests create common tasks to verify alerts on alerts json and elasticsearch objective the purpose of this issue is to keep track of the progress of tasks that parse the generated file from actions tasks at and ensures that an alert has been created for every file creation modification deletion tasks create tasks to compare generated files with the alerts at alerts json file assigned zenidd create tasks to compare generated files with the alerts at elasticsearch index | 0 |
246,106 | 20,822,888,405 | IssuesEvent | 2022-03-18 17:10:32 | Graylog2/graylog2-server | https://api.github.com/repos/Graylog2/graylog2-server | opened | Missing copy shareable link button | bug test-day | Opening the share entity dialog does not always render the copy shareable URL to clipboard button. The input is also showing a different size:
https://user-images.githubusercontent.com/716185/159049927-8b9e5cf6-c7a2-4854-ba88-2e25a018aead.mp4
This is not exclusive of saved searches, although I first noticed it there. There are no errors in the browser console and I could not find any patterns when it happens but it seems to be fixed when opening the modal for the same entity a second time.
## Expected Behavior
Button to copy shareable URL to clipboard should be in share dialog.
## Current Behavior
Button to copy shareable URL to clipboard is not always in share dialog.
## Steps to Reproduce (for bugs)
1. See video above
## Your Environment
* Graylog Version: Graylog 4.3.0-beta.2+a61a34c
* Browser version: Firefox Developer Edition 99.0b4 (64-bit), Vivaldi 5.1.2567.66 (Stable channel) (x86_64)
| 1.0 | Missing copy shareable link button - Opening the share entity dialog does not always render the copy shareable URL to clipboard button. The input is also showing a different size:
https://user-images.githubusercontent.com/716185/159049927-8b9e5cf6-c7a2-4854-ba88-2e25a018aead.mp4
This is not exclusive of saved searches, although I first noticed it there. There are no errors in the browser console and I could not find any patterns when it happens but it seems to be fixed when opening the modal for the same entity a second time.
## Expected Behavior
Button to copy shareable URL to clipboard should be in share dialog.
## Current Behavior
Button to copy shareable URL to clipboard is not always in share dialog.
## Steps to Reproduce (for bugs)
1. See video above
## Your Environment
* Graylog Version: Graylog 4.3.0-beta.2+a61a34c
* Browser version: Firefox Developer Edition 99.0b4 (64-bit), Vivaldi 5.1.2567.66 (Stable channel) (x86_64)
| non_priority | missing copy shareable link button opening the share entity dialog does not always render the copy shareable url to clipboard button the input is also showing a different size this is not exclusive of saved searches although i first noticed it there there are no errors in the browser console and i could not find any patterns when it happens but it seems to be fixed when opening the modal for the same entity a second time expected behavior button to copy shareable url to clipboard should be in share dialog current behavior button to copy shareable url to clipboard is not always in share dialog steps to reproduce for bugs see video above your environment graylog version graylog beta browser version firefox developer edition bit vivaldi stable channel | 0 |
367,282 | 25,730,921,407 | IssuesEvent | 2022-12-07 20:15:24 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | [Docs]: | documentation service/ec2 needs-triage | ### Documentation Link
https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/ami
### Description
As per documentation,
[public](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/ami#public) - true if the image has public launch permissions.
if the public flag is used as a filter then seeing below error message. Is `public` is still valid ? if i change `public` to `is-public` then its working as expected
code snippet
------------
filter {
name = "public"
values = [true]
}
error message
------------
Error: InvalidParameterValue: The filter 'public' is invalid
status code: 400, request id: a7f03df8-0b25-425e-a828-21b8d7a677b1
with data.aws_ami.debian,
on main.tf line 63, in data "aws_ami" "debian":
63: data "aws_ami" "debian" {
### References
_No response_
### Would you like to implement a fix?
None | 1.0 | [Docs]: - ### Documentation Link
https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/ami
### Description
As per documentation,
[public](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/ami#public) - true if the image has public launch permissions.
if the public flag is used as a filter then seeing below error message. Is `public` is still valid ? if i change `public` to `is-public` then its working as expected
code snippet
------------
filter {
name = "public"
values = [true]
}
error message
------------
Error: InvalidParameterValue: The filter 'public' is invalid
status code: 400, request id: a7f03df8-0b25-425e-a828-21b8d7a677b1
with data.aws_ami.debian,
on main.tf line 63, in data "aws_ami" "debian":
63: data "aws_ami" "debian" {
### References
_No response_
### Would you like to implement a fix?
None | non_priority | documentation link description as per documentation true if the image has public launch permissions if the public flag is used as a filter then seeing below error message is public is still valid if i change public to is public then its working as expected code snippet filter name public values error message error invalidparametervalue the filter public is invalid status code request id with data aws ami debian on main tf line in data aws ami debian data aws ami debian references no response would you like to implement a fix none | 0 |
304,251 | 9,329,465,887 | IssuesEvent | 2019-03-28 02:30:23 | rubrikinc/use-case-aws-cloudformation-template-cloudcluster | https://api.github.com/repos/rubrikinc/use-case-aws-cloudformation-template-cloudcluster | closed | Link for sharing CF Template | exp-beginner priority-p1 | The template needs to be uploaded and shared out via an AWS link like the rest of the CF templates.
For example: https://s3-us-west-1.amazonaws.com/cloudformation-templates-rubrik-prod/rubrik_cloudon.template | 1.0 | Link for sharing CF Template - The template needs to be uploaded and shared out via an AWS link like the rest of the CF templates.
For example: https://s3-us-west-1.amazonaws.com/cloudformation-templates-rubrik-prod/rubrik_cloudon.template | priority | link for sharing cf template the template needs to be uploaded and shared out via an aws link like the rest of the cf templates for example | 1 |
357,091 | 10,601,825,455 | IssuesEvent | 2019-10-10 13:07:40 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | Create a switch/variable for data discovery with PhEDEx x Rucio | Medium Priority New Feature ReqMgr2 Rucio Transition WorkQueue | For central services.
Related to: https://its.cern.ch/jira/browse/CMSRUCIO-104 | 1.0 | Create a switch/variable for data discovery with PhEDEx x Rucio - For central services.
Related to: https://its.cern.ch/jira/browse/CMSRUCIO-104 | priority | create a switch variable for data discovery with phedex x rucio for central services related to | 1 |
84,050 | 7,888,578,955 | IssuesEvent | 2018-06-27 22:45:02 | linnovate/root | https://api.github.com/repos/linnovate/root | closed | translate new entity has "type your text" | 2.0.4 testing week | 
translate new entity has "type your text" to "הכנס כותרת"
and design according to original design (see http://root.203.projects.linnovate.net for referece) | 1.0 | translate new entity has "type your text" - 
translate new entity has "type your text" to "הכנס כותרת"
and design according to original design (see http://root.203.projects.linnovate.net for referece) | non_priority | translate new entity has type your text translate new entity has type your text to הכנס כותרת and design according to original design see for referece | 0 |
314,163 | 9,593,467,818 | IssuesEvent | 2019-05-09 11:37:59 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.rockstargames.com - site is not usable | browser-firefox-mobile engine-gecko priority-normal | <!-- @browser: Firefox Mobile 67.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:67.0) Gecko/67.0 Firefox/67.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.rockstargames.com/GTAOnline/restricted-content/agegate/form?redirect=https%3A%2F%2Fwww.rockstargames.com%2FGTAOnline%2Fnews&options=&locale=en_us
**Browser / Version**: Firefox Mobile 67.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: entering birthday to access page gets stuck in loop, can't enter page
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190506235559</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.rockstargames.com - site is not usable - <!-- @browser: Firefox Mobile 67.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:67.0) Gecko/67.0 Firefox/67.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.rockstargames.com/GTAOnline/restricted-content/agegate/form?redirect=https%3A%2F%2Fwww.rockstargames.com%2FGTAOnline%2Fnews&options=&locale=en_us
**Browser / Version**: Firefox Mobile 67.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: entering birthday to access page gets stuck in loop, can't enter page
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190506235559</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description entering birthday to access page gets stuck in loop can t enter page steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta from with ❤️ | 1 |
602,643 | 18,492,048,092 | IssuesEvent | 2021-10-19 02:27:59 | AY2122S1-CS2113T-T12-3/tp | https://api.github.com/repos/AY2122S1-CS2113T-T12-3/tp | closed | Add function to clear all entries being tracked | priority.Medium | So that we can have a easier time testing code. Users might also find clear all function handy if they want to start afresh. | 1.0 | Add function to clear all entries being tracked - So that we can have a easier time testing code. Users might also find clear all function handy if they want to start afresh. | priority | add function to clear all entries being tracked so that we can have a easier time testing code users might also find clear all function handy if they want to start afresh | 1 |
186,103 | 14,394,638,184 | IssuesEvent | 2020-12-03 01:46:13 | github-vet/rangeclosure-findings | https://api.github.com/repos/github-vet/rangeclosure-findings | closed | tengteng/Guava: _vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go; 3 LoC | fresh test tiny |
Found a possible issue in [tengteng/Guava](https://www.github.com/tengteng/Guava) at [_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to test at line 106 may start a goroutine
[Click here to see the code in its original context.](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range importerTests {
runImporterTest(t, imp, initmap, &test)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: f44ca584d2dc0fe32182990065cfc4fd0e6cebe8
| 1.0 | tengteng/Guava: _vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go; 3 LoC -
Found a possible issue in [tengteng/Guava](https://www.github.com/tengteng/Guava) at [_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to test at line 106 may start a goroutine
[Click here to see the code in its original context.](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range importerTests {
runImporterTest(t, imp, initmap, &test)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: f44ca584d2dc0fe32182990065cfc4fd0e6cebe8
| non_priority | tengteng guava vendor src golang org x tools go gccgoimporter importer test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to test at line may start a goroutine click here to show the line s of go which triggered the analyzer go for test range importertests runimportertest t imp initmap test leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 0 |
545,022 | 15,934,230,699 | IssuesEvent | 2021-04-14 08:26:14 | NamanhTran/Syntax-Analyzer | https://api.github.com/repos/NamanhTran/Syntax-Analyzer | closed | Create Parsing Table Algorithm | Hard Priority | ## Overview
We need to implement the algorithm to read from the parse table, keep track of all production rules for each lexeme, and handle errors.
## Action Items
- [ ] Implement the algorithm
- [ ] Generate meaningful error message when syntax error identified.
- [ ] Test the table parser algorithm
## Assumptions for implementation
## Resources/Instructions
- [Youtube video on the parsing algorithm](https://www.youtube.com/watch?v=KX_RU4rVTGk&t=733s). | 1.0 | Create Parsing Table Algorithm - ## Overview
We need to implement the algorithm to read from the parse table, keep track of all production rules for each lexeme, and handle errors.
## Action Items
- [ ] Implement the algorithm
- [ ] Generate meaningful error message when syntax error identified.
- [ ] Test the table parser algorithm
## Assumptions for implementation
## Resources/Instructions
- [Youtube video on the parsing algorithm](https://www.youtube.com/watch?v=KX_RU4rVTGk&t=733s). | priority | create parsing table algorithm overview we need to implement the algorithm to read from the parse table keep track of all production rules for each lexeme and handle errors action items implement the algorithm generate meaningful error message when syntax error identified test the table parser algorithm assumptions for implementation resources instructions | 1 |
11,452 | 4,227,265,273 | IssuesEvent | 2016-07-03 02:40:56 | ac21/sherlock | https://api.github.com/repos/ac21/sherlock | closed | Fix "Rubocop/Lint/UnusedBlockArgument" issue in lib/api/v1/defaults.rb | code_climate | Unused block argument - `e`. You can omit the argument if you don't care about it.
https://codeclimate.com/github/ac21/sherlock/lib/api/v1/defaults.rb#issue_5771abba9591a1000110c0f3 | 1.0 | Fix "Rubocop/Lint/UnusedBlockArgument" issue in lib/api/v1/defaults.rb - Unused block argument - `e`. You can omit the argument if you don't care about it.
https://codeclimate.com/github/ac21/sherlock/lib/api/v1/defaults.rb#issue_5771abba9591a1000110c0f3 | non_priority | fix rubocop lint unusedblockargument issue in lib api defaults rb unused block argument e you can omit the argument if you don t care about it | 0 |
751,717 | 26,254,831,555 | IssuesEvent | 2023-01-05 23:09:13 | Automattic/woocommerce-payments | https://api.github.com/repos/Automattic/woocommerce-payments | closed | Non-prorated synced subscriptions don't result in synced WC Pay Subscription | type: bug priority: low component: wcpay subscriptions category: core | ### Describe the bug
<!-- A clear and concise description of what the bug is. Please be as descriptive as possible. -->
When a merchant selects the **Never (charge the full recurring amount at sign-up)** synchronisation setting, synced subscription products aren't being synced to their anchor date in Stripe.

### To Reproduce
<!-- Describe the steps to reproduce the behavior. -->
1. With WC Subscription (the extension inactive).
2. Go to **WooCommerce > Settings > Subscriptions**
1. Enable synchronisation
2. Select `Never (charge the full recurring amount at sign-up)` **Prorate first payment** setting.
4. Create a synced subscription product (See example screenshot below).
5. Purchase that product and note the subscriptions first payment date.
6. In the Stripe dashboard, view the subscription's billing period and note the next payment date isn't synced.
<img width="1247" alt="Screen Shot 2022-08-11 at 11 09 02 am" src="https://user-images.githubusercontent.com/8490476/184048395-d97453ed-fb0d-410c-b82c-1a4e87775771.png">
### Actual behavior
<!-- A clear and concise description of what actually happens. -->
With the **Never (charge the full recurring amount at sign-up)** setting enabled we incorrectly determine that the product "isn't" synchronised. This means we don't send the `trial_end` arg to Stripe when creating the subscription.
From my short investigation I believe the problem stems from this [logic](https://github.com/automattic/woocommerce-payments/blob/4.5.1/includes/subscriptions/class-wc-payments-subscription-service.php#L712-L714).
Because there is an upfront payment (non-proration), that `self::has_delayed_payment( $subscription )` returns `false` and so no trial end date is sent.
### Screenshots
<!-- If applicable, add screenshots to help explain your problem. -->
<img width="1842" alt="Screen Shot 2022-08-11 at 11 03 02 am" src="https://user-images.githubusercontent.com/8490476/184047929-5bbeb015-da6f-48e3-bd15-5b084ac2623b.png">
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
In this situation a `trial_end` value should be sent to Stripe when creating the subscription to ensure the subscription created in Stripe is synced.
### Additional information
- Original report in Slack: p1660131760533269/1659720769.614509-slack-CGGCLBN58 | 1.0 | Non-prorated synced subscriptions don't result in synced WC Pay Subscription - ### Describe the bug
<!-- A clear and concise description of what the bug is. Please be as descriptive as possible. -->
When a merchant selects the **Never (charge the full recurring amount at sign-up)** synchronisation setting, synced subscription products aren't being synced to their anchor date in Stripe.

### To Reproduce
<!-- Describe the steps to reproduce the behavior. -->
1. With WC Subscription (the extension inactive).
2. Go to **WooCommerce > Settings > Subscriptions**
1. Enable synchronisation
2. Select `Never (charge the full recurring amount at sign-up)` **Prorate first payment** setting.
4. Create a synced subscription product (See example screenshot below).
5. Purchase that product and note the subscriptions first payment date.
6. In the Stripe dashboard, view the subscription's billing period and note the next payment date isn't synced.
<img width="1247" alt="Screen Shot 2022-08-11 at 11 09 02 am" src="https://user-images.githubusercontent.com/8490476/184048395-d97453ed-fb0d-410c-b82c-1a4e87775771.png">
### Actual behavior
<!-- A clear and concise description of what actually happens. -->
With the **Never (charge the full recurring amount at sign-up)** setting enabled we incorrectly determine that the product "isn't" synchronised. This means we don't send the `trial_end` arg to Stripe when creating the subscription.
From my short investigation I believe the problem stems from this [logic](https://github.com/automattic/woocommerce-payments/blob/4.5.1/includes/subscriptions/class-wc-payments-subscription-service.php#L712-L714).
Because there is an upfront payment (non-proration), that `self::has_delayed_payment( $subscription )` returns `false` and so no trial end date is sent.
### Screenshots
<!-- If applicable, add screenshots to help explain your problem. -->
<img width="1842" alt="Screen Shot 2022-08-11 at 11 03 02 am" src="https://user-images.githubusercontent.com/8490476/184047929-5bbeb015-da6f-48e3-bd15-5b084ac2623b.png">
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
In this situation a `trial_end` value should be sent to Stripe when creating the subscription to ensure the subscription created in Stripe is synced.
### Additional information
- Original report in Slack: p1660131760533269/1659720769.614509-slack-CGGCLBN58 | priority | non prorated synced subscriptions don t result in synced wc pay subscription describe the bug when a merchant selects the never charge the full recurring amount at sign up synchronisation setting synced subscription products aren t being synced to their anchor date in stripe to reproduce with wc subscription the extension inactive go to woocommerce settings subscriptions enable synchronisation select never charge the full recurring amount at sign up prorate first payment setting create a synced subscription product see example screenshot below purchase that product and note the subscriptions first payment date in the stripe dashboard view the subscription s billing period and note the next payment date isn t synced img width alt screen shot at am src actual behavior with the never charge the full recurring amount at sign up setting enabled we incorrectly determine that the product isn t synchronised this means we don t send the trial end arg to stripe when creating the subscription from my short investigation i believe the problem stems from this because there is an upfront payment non proration that self has delayed payment subscription returns false and so no trial end date is sent screenshots img width alt screen shot at am src expected behavior in this situation a trial end value should be sent to stripe when creating the subscription to ensure the subscription created in stripe is synced additional information original report in slack slack | 1 |
258,007 | 19,538,686,851 | IssuesEvent | 2021-12-31 14:14:19 | CalmiraLinux/cport | https://api.github.com/repos/CalmiraLinux/cport | closed | Обновить README | documentation | Добавить обновлённую информацию о работу с утилитой:
- Установка порта;
- Удаление порта;
- Просмотр информации о порте.
| 1.0 | Обновить README - Добавить обновлённую информацию о работу с утилитой:
- Установка порта;
- Удаление порта;
- Просмотр информации о порте.
| non_priority | обновить readme добавить обновлённую информацию о работу с утилитой установка порта удаление порта просмотр информации о порте | 0 |
530,419 | 15,422,730,106 | IssuesEvent | 2021-03-05 14:44:59 | jahirfiquitiva/Frames | https://api.github.com/repos/jahirfiquitiva/Frames | closed | Sec media storage crash | Priority: Low Status: Invalid Status: Not Reproducible | Whenever I try to save multiple wallpapers on apps like joywalls, wallflair, reev pro icon pack which use your wallpaper dashboard, the app sec media process and download manager crashes along with the app.
This is only happening on Samsung one ui3 3.0 and 3.1 on my s10 and s21 ultra. Apps work done on one up 2.5 and aosp rom.
I am using latest versions of all these apps and have tried clearing cache and data off crashing apps as well as cache from recovery. <!--
Any HTML comment will be stripped when the markdown is rendered, so you don't need to delete them.
-->
- [x] I have verified there are no duplicate active or recent bugs, questions, or requests
- [x] I have verified that I am using the latest version.
### Device/App info:
- Frames Version: `?`
- Android version: `?`
- Device Manufacturer: `?`
- Device Name: `?`
### Describe the bug
A clear and concise description of what the bug is.
### Reproduction Steps
1.
2.
3.
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
### Screenshots
<!-- If applicable, add screenshots or videos to help explain your problem. -->
### Code and/or Logs
<!--
Please wrap code with correct syntax highlighting. You can remove it if you think it isn't necessary.
-->
```kotlin
println("Hello, world!")
```
<!--
If you are getting an error in the LogCat, paste here the stack trace.
Please wrap logs with Gradle syntax highlighting (it makes them look better).
-->
```Gradle
java.lang.RuntimeException: This is an example Exception log
at com.package.name.HelloWorld
at com.package.name.HelloWorld$ThisIsNotARealLog
at android.app.Instrumentation.callActivityOnResume(Instrumentation.kt)
```
### Additional context
<!-- Add any other context about the problem here. -->
| 1.0 | Sec media storage crash - Whenever I try to save multiple wallpapers on apps like joywalls, wallflair, reev pro icon pack which use your wallpaper dashboard, the app sec media process and download manager crashes along with the app.
This is only happening on Samsung one ui3 3.0 and 3.1 on my s10 and s21 ultra. Apps work done on one up 2.5 and aosp rom.
I am using latest versions of all these apps and have tried clearing cache and data off crashing apps as well as cache from recovery. <!--
Any HTML comment will be stripped when the markdown is rendered, so you don't need to delete them.
-->
- [x] I have verified there are no duplicate active or recent bugs, questions, or requests
- [x] I have verified that I am using the latest version.
### Device/App info:
- Frames Version: `?`
- Android version: `?`
- Device Manufacturer: `?`
- Device Name: `?`
### Describe the bug
A clear and concise description of what the bug is.
### Reproduction Steps
1.
2.
3.
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
### Screenshots
<!-- If applicable, add screenshots or videos to help explain your problem. -->
### Code and/or Logs
<!--
Please wrap code with correct syntax highlighting. You can remove it if you think it isn't necessary.
-->
```kotlin
println("Hello, world!")
```
<!--
If you are getting an error in the LogCat, paste here the stack trace.
Please wrap logs with Gradle syntax highlighting (it makes them look better).
-->
```Gradle
java.lang.RuntimeException: This is an example Exception log
at com.package.name.HelloWorld
at com.package.name.HelloWorld$ThisIsNotARealLog
at android.app.Instrumentation.callActivityOnResume(Instrumentation.kt)
```
### Additional context
<!-- Add any other context about the problem here. -->
| priority | sec media storage crash whenever i try to save multiple wallpapers on apps like joywalls wallflair reev pro icon pack which use your wallpaper dashboard the app sec media process and download manager crashes along with the app this is only happening on samsung one and on my and ultra apps work done on one up and aosp rom i am using latest versions of all these apps and have tried clearing cache and data off crashing apps as well as cache from recovery any html comment will be stripped when the markdown is rendered so you don t need to delete them i have verified there are no duplicate active or recent bugs questions or requests i have verified that i am using the latest version device app info frames version android version device manufacturer device name describe the bug a clear and concise description of what the bug is reproduction steps expected behavior screenshots code and or logs please wrap code with correct syntax highlighting you can remove it if you think it isn t necessary kotlin println hello world if you are getting an error in the logcat paste here the stack trace please wrap logs with gradle syntax highlighting it makes them look better gradle java lang runtimeexception this is an example exception log at com package name helloworld at com package name helloworld thisisnotareallog at android app instrumentation callactivityonresume instrumentation kt additional context | 1 |
569,326 | 17,011,973,387 | IssuesEvent | 2021-07-02 06:36:01 | Uninett/Argus-frontend | https://api.github.com/repos/Uninett/Argus-frontend | opened | Filtering incidents by severity level | backend discussion enhancement priority: high size: small | Severity levels were introduced into the API as part of [Uninett/Argus#70](https://github.com/Uninett/Argus/issues/70), and issue #271 deals with showing the severity level in the frontend. The next step is to make it possible to also filter on severity level in the frontend. There is already support for this in the backend.
My suggestion is to use a dropdown to filter on max severity level. The frontend will then show issues with severity level lower than or equal to the selected value. The dropdown will probably look something like this:

| 1.0 | Filtering incidents by severity level - Severity levels were introduced into the API as part of [Uninett/Argus#70](https://github.com/Uninett/Argus/issues/70), and issue #271 deals with showing the severity level in the frontend. The next step is to make it possible to also filter on severity level in the frontend. There is already support for this in the backend.
My suggestion is to use a dropdown to filter on max severity level. The frontend will then show issues with severity level lower than or equal to the selected value. The dropdown will probably look something like this:

| priority | filtering incidents by severity level severity levels were introduced into the api as part of and issue deals with showing the severity level in the frontend the next step is to make it possible to also filter on severity level in the frontend there is already support for this in the backend my suggestion is to use a dropdown to filter on max severity level the frontend will then show issues with severity level lower than or equal to the selected value the dropdown will probably look something like this | 1 |
54,475 | 30,198,664,388 | IssuesEvent | 2023-07-05 02:06:10 | cilium/cilium | https://api.github.com/repos/cilium/cilium | closed | Tunnel mode lowers throughput by a large amount | kind/bug need-more-info sig/datapath kind/performance needs/triage kind/community-report stale | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### What happened?
When using cilium with tunnel mode enabled, throughput drops by a significant amount(30-50%). This was seen in #22898 with the attached logs. This bug is an extension of that to discuss matters related specifically to that.
### Cilium Version
1.12.3
### Kernel Version
5.4.0-42-generic
### Kubernetes Version
1.26
### Sysdump
Couldn't get it
### Relevant log output
_No response_
### Anything else?
From other issue
1) tunnel: disabled, mtu: 1450 [logs](https://gist.github.com/michaelasp/3883c8fdfd298ad192a80f6a1db13262) ~8.24gb/s
2) tunnel: vxlan [logs](https://gist.github.com/michaelasp/1081274981900abfbfefdcccd6f52a68) ~5.8 Gb/s, 30% drop
3) tunnel: geneve [logs](https://gist.github.com/michaelasp/542ba5bfb0e325be2bb16585167e7314) ~5.2 Gb/s 37% drop
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct | True | Tunnel mode lowers throughput by a large amount - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### What happened?
When using cilium with tunnel mode enabled, throughput drops by a significant amount(30-50%). This was seen in #22898 with the attached logs. This bug is an extension of that to discuss matters related specifically to that.
### Cilium Version
1.12.3
### Kernel Version
5.4.0-42-generic
### Kubernetes Version
1.26
### Sysdump
Couldn't get it
### Relevant log output
_No response_
### Anything else?
From other issue
1) tunnel: disabled, mtu: 1450 [logs](https://gist.github.com/michaelasp/3883c8fdfd298ad192a80f6a1db13262) ~8.24gb/s
2) tunnel: vxlan [logs](https://gist.github.com/michaelasp/1081274981900abfbfefdcccd6f52a68) ~5.8 Gb/s, 30% drop
3) tunnel: geneve [logs](https://gist.github.com/michaelasp/542ba5bfb0e325be2bb16585167e7314) ~5.2 Gb/s 37% drop
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct | non_priority | tunnel mode lowers throughput by a large amount is there an existing issue for this i have searched the existing issues what happened when using cilium with tunnel mode enabled throughput drops by a significant amount this was seen in with the attached logs this bug is an extension of that to discuss matters related specifically to that cilium version kernel version generic kubernetes version sysdump couldn t get it relevant log output no response anything else from other issue tunnel disabled mtu s tunnel vxlan gb s drop tunnel geneve gb s drop code of conduct i agree to follow this project s code of conduct | 0 |
647,635 | 21,132,750,417 | IssuesEvent | 2022-04-06 01:22:36 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | Make OnShowPanel and OnShowWalletOnboarding for Android only | priority/P2 QA/No release-notes/exclude feature/wallet OS/Android | OnShowPanel();
OnShowWalletOnboarding();
observer functions should be visible for Android only | 1.0 | Make OnShowPanel and OnShowWalletOnboarding for Android only - OnShowPanel();
OnShowWalletOnboarding();
observer functions should be visible for Android only | priority | make onshowpanel and onshowwalletonboarding for android only onshowpanel onshowwalletonboarding observer functions should be visible for android only | 1 |
540,084 | 15,800,575,450 | IssuesEvent | 2021-04-03 00:20:11 | JensenJ/EmbargoMC-IssueTracker | https://api.github.com/repos/JensenJ/EmbargoMC-IssueTracker | closed | [BUG] Water / Lava Deleted Source Blocks | bug low-priority | **Describe the bug**
Source blocks get deleted by custom caves from world painter
**Expected behaviour**
Consistent water/lava lakes spawning without holes in caused by caves | 1.0 | [BUG] Water / Lava Deleted Source Blocks - **Describe the bug**
Source blocks get deleted by custom caves from world painter
**Expected behaviour**
Consistent water/lava lakes spawning without holes in caused by caves | priority | water lava deleted source blocks describe the bug source blocks get deleted by custom caves from world painter expected behaviour consistent water lava lakes spawning without holes in caused by caves | 1 |
595,489 | 18,067,595,632 | IssuesEvent | 2021-09-20 21:07:17 | OpenMandrivaAssociation/test2 | https://api.github.com/repos/OpenMandrivaAssociation/test2 | closed | Adding users crashes the program (Bugzilla Bug 138) | bug high priority major | This issue was created automatically with bugzilla2github
# Bugzilla Bug 138
Date: 2013-09-14 15:05:51 +0000
From: @robxu9
To: OpenMandriva QA <<bugs@openmandriva.org>>
CC: @cris-b
Last updated: 2013-09-19 20:05:04 +0000
## Comment 822
Date: 2013-09-14 15:05:51 +0000
From: @robxu9
Theme name: rosa-elementary
Kernel version = 3.10.7-nrjQL-desktop-1omv
Distribution=OpenMandriva LX release 2013.0 Beta (Oxygen) for x86_64
CPU=QEMU Virtual CPU version 1.4.2
When I add users with userdrake, it closes unexpectedly. The user is still added, but locked.
## Comment 825
Date: 2013-09-14 15:13:58 +0000
From: @robxu9
*** Bug #139 has been marked as a duplicate of this bug. ***
## Comment 955
Date: 2013-09-18 21:00:17 +0000
From: @cris-b
looks like the recent update to userdrake-1.14 requires libuser to be updated to 0.59.
will request that we update libuser to 0.59 to sort this.
## Comment 963
Date: 2013-09-19 11:08:25 +0000
From: @cris-b
should now be resolved as libuser has been updated to 0.59
| 1.0 | Adding users crashes the program (Bugzilla Bug 138) - This issue was created automatically with bugzilla2github
# Bugzilla Bug 138
Date: 2013-09-14 15:05:51 +0000
From: @robxu9
To: OpenMandriva QA <<bugs@openmandriva.org>>
CC: @cris-b
Last updated: 2013-09-19 20:05:04 +0000
## Comment 822
Date: 2013-09-14 15:05:51 +0000
From: @robxu9
Theme name: rosa-elementary
Kernel version = 3.10.7-nrjQL-desktop-1omv
Distribution=OpenMandriva LX release 2013.0 Beta (Oxygen) for x86_64
CPU=QEMU Virtual CPU version 1.4.2
When I add users with userdrake, it closes unexpectedly. The user is still added, but locked.
## Comment 825
Date: 2013-09-14 15:13:58 +0000
From: @robxu9
*** Bug #139 has been marked as a duplicate of this bug. ***
## Comment 955
Date: 2013-09-18 21:00:17 +0000
From: @cris-b
looks like the recent update to userdrake-1.14 requires libuser to be updated to 0.59.
will request that we update libuser to 0.59 to sort this.
## Comment 963
Date: 2013-09-19 11:08:25 +0000
From: @cris-b
should now be resolved as libuser has been updated to 0.59
| priority | adding users crashes the program bugzilla bug this issue was created automatically with bugzilla bug date from to openmandriva qa lt gt cc cris b last updated comment date from theme name rosa elementary kernel version nrjql desktop distribution openmandriva lx release beta oxygen for cpu qemu virtual cpu version when i add users with userdrake it closes unexpectedly the user is still added but locked comment date from bug has been marked as a duplicate of this bug comment date from cris b looks like the recent update to userdrake requires libuser to be updated to will request that we update libuser to to sort this comment date from cris b should now be resolved as libuser has been updated to | 1 |
156,371 | 5,968,203,730 | IssuesEvent | 2017-05-30 17:36:34 | kolihub/koli | https://api.github.com/repos/kolihub/koli | opened | Better error handling when downloading releases | area/slugrunner improvement priority/P2 | The slugrunner doesn't validate any error when downloading releases from the git-server. | 1.0 | Better error handling when downloading releases - The slugrunner doesn't validate any error when downloading releases from the git-server. | priority | better error handling when downloading releases the slugrunner doesn t validate any error when downloading releases from the git server | 1 |
729,955 | 25,152,590,302 | IssuesEvent | 2022-11-10 11:07:35 | rism-digital/rism-online-issues | https://api.github.com/repos/rism-digital/rism-online-issues | closed | Add contour search for incipit | Audience: General public Priority: Moderate Status: Blocked Topic: Incipits Component: Incipit Search | In Muscat BL we used to have contour search enabled. It was the implementation of Themefinder
* http://www.themefinder.org/help/refinedcontour/
* http://www.themefinder.org/help/grosscontour/
Having one type of contour might be useful, especially since we have now highlighting. | 1.0 | Add contour search for incipit - In Muscat BL we used to have contour search enabled. It was the implementation of Themefinder
* http://www.themefinder.org/help/refinedcontour/
* http://www.themefinder.org/help/grosscontour/
Having one type of contour might be useful, especially since we have now highlighting. | priority | add contour search for incipit in muscat bl we used to have contour search enabled it was the implementation of themefinder having one type of contour might be useful especially since we have now highlighting | 1 |
289,596 | 31,986,705,316 | IssuesEvent | 2023-09-21 00:20:44 | pferronMend/Case121627 | https://api.github.com/repos/pferronMend/Case121627 | opened | rocksdbnative.6.2.2.nupkg: 1 vulnerabilities (highest severity is: 7.5) | Mend: dependency security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rocksdbnative.6.2.2.nupkg</b></p></summary>
<p></p>
<p>Path to dependency file: /examples/ExactlyOnceOldBroker/ExactlyOnceOldBroker.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (rocksdbnative.6.2.2.nupkg version) | Remediation Possible** |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2018-8292](https://www.mend.io/vulnerability-database/CVE-2018-8292) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | system.net.http.4.3.0.nupkg | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p><p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2018-8292</summary>
### Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers.
</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /examples/ExactlyOnceOldBroker/ExactlyOnceOldBroker.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- rocksdbnative.6.2.2.nupkg (Root Library)
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
<p></p>
</details> | True | rocksdbnative.6.2.2.nupkg: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rocksdbnative.6.2.2.nupkg</b></p></summary>
<p></p>
<p>Path to dependency file: /examples/ExactlyOnceOldBroker/ExactlyOnceOldBroker.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (rocksdbnative.6.2.2.nupkg version) | Remediation Possible** |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2018-8292](https://www.mend.io/vulnerability-database/CVE-2018-8292) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | system.net.http.4.3.0.nupkg | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p><p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2018-8292</summary>
### Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers.
</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /examples/ExactlyOnceOldBroker/ExactlyOnceOldBroker.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- rocksdbnative.6.2.2.nupkg (Root Library)
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
<p></p>
</details> | non_priority | rocksdbnative nupkg vulnerabilities highest severity is vulnerable library rocksdbnative nupkg path to dependency file examples exactlyonceoldbroker exactlyonceoldbroker csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg vulnerabilities cve severity cvss dependency type fixed in rocksdbnative nupkg version remediation possible high system net http nupkg transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the details section below to see if there is a version of transitive dependency where vulnerability is fixed in some cases remediation pr cannot be created automatically for a vulnerability despite the availability of remediation details cve vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that allow applications to consume web services over http and http components that can be used by both clients and servers for parsing http headers library home page a href path to dependency file examples exactlyonceoldbroker exactlyonceoldbroker csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg dependency hierarchy rocksdbnative nupkg root library netstandard library nupkg x system net http nupkg vulnerable library found in base branch master vulnerability details an information disclosure vulnerability exists in net core when authentication information is inadvertently exposed in a redirect aka net core information disclosure vulnerability this affects net core net core net core powershell core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution system net http microsoft powershell commands utility rc | 0 |
403,415 | 11,840,709,742 | IssuesEvent | 2020-03-23 19:22:58 | AlaskaAirlines/auro_docs | https://api.github.com/repos/AlaskaAirlines/auro_docs | opened | CDN Ready components | Priority: High Status: Proposal Type: Feature | # User story
As a web developer, I want to be able to consume Auro components w/o having to pre-compile code, so I can easily use resources in a non-dynamic environment.
## Outline subtasks or tasks
> Decide which specific steps need to be completed and who is responsible for each of them.
- [ ] Output pre-rendered JS version of web componet
- [ ] store WC in CDN
- [ ] append versioning to the ouput file name
- [ ] ensure stability of delivered asset
- [ ] have solution for polyfill
## Definition of “Done” (exit criteria)
> The story is generally “done” when the user can complete the outlined task, but make sure to define what that is.
An initial Auro web component has been updated with functionality to export a consumable version of the WC via a CDN.
| 1.0 | CDN Ready components - # User story
As a web developer, I want to be able to consume Auro components w/o having to pre-compile code, so I can easily use resources in a non-dynamic environment.
## Outline subtasks or tasks
> Decide which specific steps need to be completed and who is responsible for each of them.
- [ ] Output pre-rendered JS version of web componet
- [ ] store WC in CDN
- [ ] append versioning to the ouput file name
- [ ] ensure stability of delivered asset
- [ ] have solution for polyfill
## Definition of “Done” (exit criteria)
> The story is generally “done” when the user can complete the outlined task, but make sure to define what that is.
An initial Auro web component has been updated with functionality to export a consumable version of the WC via a CDN.
| priority | cdn ready components user story as a web developer i want to be able to consume auro components w o having to pre compile code so i can easily use resources in a non dynamic environment outline subtasks or tasks decide which specific steps need to be completed and who is responsible for each of them output pre rendered js version of web componet store wc in cdn append versioning to the ouput file name ensure stability of delivered asset have solution for polyfill definition of “done” exit criteria the story is generally “done” when the user can complete the outlined task but make sure to define what that is an initial auro web component has been updated with functionality to export a consumable version of the wc via a cdn | 1 |
126,977 | 26,956,700,015 | IssuesEvent | 2023-02-08 15:21:45 | arduino/arduino-ide | https://api.github.com/repos/arduino/arduino-ide | reopened | Spurious output in Serial Monitor after upload | conclusion: resolved topic: code type: imperfection topic: CLI topic: serial monitor criticality: low | ### Describe the problem
The first output printed to Serial Monitor after an upload is some "garbage" characters.

### To reproduce
#### Equipment
- Arduino board that uses the ATmega16U2 USB chip:
- Arduino Uno
- Arduino Uno Mini
- Arduino Mega
#### Steps
1. Upload a sketch to your board that prints to `Serial`:
```cpp
void setup() {
Serial.begin(9600);
Serial.println("hello world");
}
void loop() {}
```
1. Open the "**Serial Monitor**" view.
🐛 The output starts with some unexpected content:
```text
□□□□□□□□□□□hello world
```
### Expected behavior
Serial monitor output always reflects the data sent by the board.
### Arduino IDE version
2.0.0-rc5-snapshot-4de7737
### Operating system
Windows
### Operating system version
10
### Additional context
The number of `□` in the demonstration matches the number of characters that are printed by the sketch.
---
The spurious output does not occur when the output is triggered by resetting the board, so it is specific to the upload operation.
---
The spurious output still occurs even if you add a delay before the print, so it is not a matter of timing from initialization as is sometimes the cause of corrupted serial output.
---
The issue does not occur when using Arduino IDE 1.x
---
I have only been able to reproduce this issue with the Arduino boards that use an ATmega16U2 USB to serial adapter chip.
I could not reproduce it when using boards with other USB interfaces:
- Native USB (Leonardo, Nano 33 IoT)
- FTDI FT232R (Nano, Pro Mini)
- WCH CH340 (3rd party boards)
---
Originally reported at https://forum.arduino.cc/t/serial-monitor-contains-garbage-after-upload/972312
### Issue checklist
- [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=)
- [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds)
- [X] My report contains all necessary details | 1.0 | Spurious output in Serial Monitor after upload - ### Describe the problem
The first output printed to Serial Monitor after an upload is some "garbage" characters.

### To reproduce
#### Equipment
- Arduino board that uses the ATmega16U2 USB chip:
- Arduino Uno
- Arduino Uno Mini
- Arduino Mega
#### Steps
1. Upload a sketch to your board that prints to `Serial`:
```cpp
void setup() {
Serial.begin(9600);
Serial.println("hello world");
}
void loop() {}
```
1. Open the "**Serial Monitor**" view.
🐛 The output starts with some unexpected content:
```text
□□□□□□□□□□□hello world
```
### Expected behavior
Serial monitor output always reflects the data sent by the board.
### Arduino IDE version
2.0.0-rc5-snapshot-4de7737
### Operating system
Windows
### Operating system version
10
### Additional context
The number of `□` in the demonstration matches the number of characters that are printed by the sketch.
---
The spurious output does not occur when the output is triggered by resetting the board, so it is specific to the upload operation.
---
The spurious output still occurs even if you add a delay before the print, so it is not a matter of timing from initialization as is sometimes the cause of corrupted serial output.
---
The issue does not occur when using Arduino IDE 1.x
---
I have only been able to reproduce this issue with the Arduino boards that use an ATmega16U2 USB to serial adapter chip.
I could not reproduce it when using boards with other USB interfaces:
- Native USB (Leonardo, Nano 33 IoT)
- FTDI FT232R (Nano, Pro Mini)
- WCH CH340 (3rd party boards)
---
Originally reported at https://forum.arduino.cc/t/serial-monitor-contains-garbage-after-upload/972312
### Issue checklist
- [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=)
- [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds)
- [X] My report contains all necessary details | non_priority | spurious output in serial monitor after upload describe the problem the first output printed to serial monitor after an upload is some garbage characters to reproduce equipment arduino board that uses the usb chip arduino uno arduino uno mini arduino mega steps upload a sketch to your board that prints to serial cpp void setup serial begin serial println hello world void loop open the serial monitor view 🐛 the output starts with some unexpected content text □□□□□□□□□□□hello world expected behavior serial monitor output always reflects the data sent by the board arduino ide version snapshot operating system windows operating system version additional context the number of □ in the demonstration matches the number of characters that are printed by the sketch the spurious output does not occur when the output is triggered by resetting the board so it is specific to the upload operation the spurious output still occurs even if you add a delay before the print so it is not a matter of timing from initialization as is sometimes the cause of corrupted serial output the issue does not occur when using arduino ide x i have only been able to reproduce this issue with the arduino boards that use an usb to serial adapter chip i could not reproduce it when using boards with other usb interfaces native usb leonardo nano iot ftdi nano pro mini wch party boards originally reported at issue checklist i searched for previous reports in i verified the problem still occurs when using the latest my report contains all necessary details | 0 |
106,226 | 13,256,458,785 | IssuesEvent | 2020-08-20 12:43:52 | fremtind/jokul | https://api.github.com/repos/fremtind/jokul | closed | Loader: Legges til i Figma-biblioteket | 🎨 design | Komponenten eksisterer kun i kode (og i After Effects) Må også gjenspeiles i Figmabiblioteket.
Bør samtidig vurdere en visuell oppdatering. Spesielt i tilfellene der den blir brukt som loader i en knapp/pilleform
| 1.0 | Loader: Legges til i Figma-biblioteket - Komponenten eksisterer kun i kode (og i After Effects) Må også gjenspeiles i Figmabiblioteket.
Bør samtidig vurdere en visuell oppdatering. Spesielt i tilfellene der den blir brukt som loader i en knapp/pilleform
| non_priority | loader legges til i figma biblioteket komponenten eksisterer kun i kode og i after effects må også gjenspeiles i figmabiblioteket bør samtidig vurdere en visuell oppdatering spesielt i tilfellene der den blir brukt som loader i en knapp pilleform | 0 |
18,160 | 25,086,834,980 | IssuesEvent | 2022-11-08 00:58:58 | Automattic/woocommerce-subscriptions-core | https://api.github.com/repos/Automattic/woocommerce-subscriptions-core | closed | [HPOS] The billing address index field is duplicated in subscription meta | type: bug compatibility: HPOS | ## Describe the bug
<!-- A clear and concise description of what the bug is. Please be as descriptive as possible, and include screenshots to illustrate. -->
On a store with HPOS tables enabled with no sync (this may also impact stores with syncing on) the billing address index is duplicated in the `wp_wc_orders_meta` table.
eg:
<img width="699" alt="Screen Shot 2022-11-07 at 10 09 40 am" src="https://user-images.githubusercontent.com/8490476/200203146-d4092704-1eea-41b6-b61b-b2487d8f5083.png">
## To Reproduce
<!-- Describe the steps to reproduce the behavior. -->
1. Turn on HPOS and disable syncing. See screenshot below of HPOS settings below.
2. Purchase a subscription using BACs.
3. In the database look at the `wp_wc_orders_meta` meta and notice the subscription has two billing index fields.
<img width="769" alt="Screen Shot 2022-11-07 at 10 13 46 am" src="https://user-images.githubusercontent.com/8490476/200203381-c44b3a38-86f5-4ed4-a621-795eed19c2bd.png">
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
There should only be 1 billing index.
### Actual behavior
<!-- A clear and concise description of what actually happens. -->
There are 2 billing indexes and one of them is incomplete. | True | [HPOS] The billing address index field is duplicated in subscription meta - ## Describe the bug
<!-- A clear and concise description of what the bug is. Please be as descriptive as possible, and include screenshots to illustrate. -->
On a store with HPOS tables enabled with no sync (this may also impact stores with syncing on) the billing address index is duplicated in the `wp_wc_orders_meta` table.
eg:
<img width="699" alt="Screen Shot 2022-11-07 at 10 09 40 am" src="https://user-images.githubusercontent.com/8490476/200203146-d4092704-1eea-41b6-b61b-b2487d8f5083.png">
## To Reproduce
<!-- Describe the steps to reproduce the behavior. -->
1. Turn on HPOS and disable syncing. See screenshot below of HPOS settings below.
2. Purchase a subscription using BACs.
3. In the database look at the `wp_wc_orders_meta` meta and notice the subscription has two billing index fields.
<img width="769" alt="Screen Shot 2022-11-07 at 10 13 46 am" src="https://user-images.githubusercontent.com/8490476/200203381-c44b3a38-86f5-4ed4-a621-795eed19c2bd.png">
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
There should only be 1 billing index.
### Actual behavior
<!-- A clear and concise description of what actually happens. -->
There are 2 billing indexes and one of them is incomplete. | non_priority | the billing address index field is duplicated in subscription meta describe the bug on a store with hpos tables enabled with no sync this may also impact stores with syncing on the billing address index is duplicated in the wp wc orders meta table eg img width alt screen shot at am src to reproduce turn on hpos and disable syncing see screenshot below of hpos settings below purchase a subscription using bacs in the database look at the wp wc orders meta meta and notice the subscription has two billing index fields img width alt screen shot at am src expected behavior there should only be billing index actual behavior there are billing indexes and one of them is incomplete | 0 |
371,527 | 10,968,265,982 | IssuesEvent | 2019-11-28 11:14:33 | ONEARMY/community-platform | https://api.github.com/repos/ONEARMY/community-platform | opened | [bug] How-Tos not loading | Priority: Urgent❕❕❕ Type:Bug🐛 | **Description**
The How-To's are still not loading (I already cleared cache) on Windows 10, on Mozilla
**Link**
https://community.preciousplastic.com/how-to
**Screenshots**
**User contact**
mariska
Sure, I'm happy to help! | 1.0 | [bug] How-Tos not loading - **Description**
The How-To's are still not loading (I already cleared cache) on Windows 10, on Mozilla
**Link**
https://community.preciousplastic.com/how-to
**Screenshots**
**User contact**
mariska
Sure, I'm happy to help! | priority | how tos not loading description the how to s are still not loading i already cleared cache on windows on mozilla link screenshots user contact mariska sure i m happy to help | 1 |
509,077 | 14,712,381,742 | IssuesEvent | 2021-01-05 08:52:25 | wmeints/fizzylogic | https://api.github.com/repos/wmeints/fizzylogic | closed | Enhance the markdown pipeline | priority:high | This issue is my place to collect problems with the markdown pipeline in the website. Right now it's a basic pipeline. I've already spotted a few things missing:
* Figures need a caption based on the alt text for the image
* Code needs a wrapper element to correctly align | 1.0 | Enhance the markdown pipeline - This issue is my place to collect problems with the markdown pipeline in the website. Right now it's a basic pipeline. I've already spotted a few things missing:
* Figures need a caption based on the alt text for the image
* Code needs a wrapper element to correctly align | priority | enhance the markdown pipeline this issue is my place to collect problems with the markdown pipeline in the website right now it s a basic pipeline i ve already spotted a few things missing figures need a caption based on the alt text for the image code needs a wrapper element to correctly align | 1 |
91,918 | 26,519,812,842 | IssuesEvent | 2023-01-19 00:55:26 | airbytehq/airbyte | https://api.github.com/repos/airbytehq/airbyte | opened | [Connector Builder] Set default values for start and end datetimes | team/extensibility ui/connector-builder | ## Context
It is very common for users to use `{{ now_utc() }}` as the value for the `end_datetime` field in order to keep their data fully up-to-date.
Until https://github.com/airbytehq/airbyte/issues/21571 is complete (which makes this field optional), the connector builder should automatically set this as the value of the end_datetime field, as this will be the value users usually want anyway and it will also teach them that this macro exists.
Similarly, we should set a sensible default for the `start_datetime` field so that users don't need to configure that unless they want to use a specific value.
## Acceptance Criteria
- `end_datetime` is prefilled with a value of `{{ now_utc() }}` when the datetime stream slicer is selected in the builder UI
- `start_datetime` is prefilled with a value of now_utc minus 1 year when the datetime stream slicer is selected in the builder UI
Open question:
- Is now - 1 year the right time range to use as the default for start_datetime? Or would a month or a day be more sensible? | 1.0 | [Connector Builder] Set default values for start and end datetimes - ## Context
It is very common for users to use `{{ now_utc() }}` as the value for the `end_datetime` field in order to keep their data fully up-to-date.
Until https://github.com/airbytehq/airbyte/issues/21571 is complete (which makes this field optional), the connector builder should automatically set this as the value of the end_datetime field, as this will be the value users usually want anyway and it will also teach them that this macro exists.
Similarly, we should set a sensible default for the `start_datetime` field so that users don't need to configure that unless they want to use a specific value.
## Acceptance Criteria
- `end_datetime` is prefilled with a value of `{{ now_utc() }}` when the datetime stream slicer is selected in the builder UI
- `start_datetime` is prefilled with a value of now_utc minus 1 year when the datetime stream slicer is selected in the builder UI
Open question:
- Is now - 1 year the right time range to use as the default for start_datetime? Or would a month or a day be more sensible? | non_priority | set default values for start and end datetimes context it is very common for users to use now utc as the value for the end datetime field in order to keep their data fully up to date until is complete which makes this field optional the connector builder should automatically set this as the value of the end datetime field as this will be the value users usually want anyway and it will also teach them that this macro exists similarly we should set a sensible default for the start datetime field so that users don t need to configure that unless they want to use a specific value acceptance criteria end datetime is prefilled with a value of now utc when the datetime stream slicer is selected in the builder ui start datetime is prefilled with a value of now utc minus year when the datetime stream slicer is selected in the builder ui open question is now year the right time range to use as the default for start datetime or would a month or a day be more sensible | 0 |
513,963 | 14,930,038,469 | IssuesEvent | 2021-01-25 01:42:00 | SickChill/SickChill | https://api.github.com/repos/SickChill/SickChill | opened | GoFundMe | Help wanted Priority HIGH | So, my laptop is broken, and I have been unable to do anything for the project other than what I can do on the web interface on GitHub for the past several months, and I'm not sure when I might be able to fix that situation.
It was suggested that I create a GoFundMe (which I kind of detest when people do this) and someone told me it would be a good idea.
There is more info about my situation on the GoFundMe, so if you are feeling generous or you can afford to help I appreciate it. If not, I appreciate you still. Thanks.
https://www.gofundme.com/manage/laptop-for-open-source-developer-miigotu | 1.0 | GoFundMe - So, my laptop is broken, and I have been unable to do anything for the project other than what I can do on the web interface on GitHub for the past several months, and I'm not sure when I might be able to fix that situation.
It was suggested that I create a GoFundMe (which I kind of detest when people do this) and someone told me it would be a good idea.
There is more info about my situation on the GoFundMe, so if you are feeling generous or you can afford to help I appreciate it. If not, I appreciate you still. Thanks.
https://www.gofundme.com/manage/laptop-for-open-source-developer-miigotu | priority | gofundme so my laptop is broken and i have been unable to do anything for the project other than what i can do on the web interface on github for the past several months and i m not sure when i might be able to fix that situation it was suggested that i create a gofundme which i kind of detest when people do this and someone told me it would be a good idea there is more info about my situation on the gofundme so if you are feeling generous or you can afford to help i appreciate it if not i appreciate you still thanks | 1 |
135,403 | 19,570,285,907 | IssuesEvent | 2022-01-04 09:04:54 | hcengineering/anticrm | https://api.github.com/repos/hcengineering/anticrm | closed | Add additional way to expand/collapse the side menu | Design | Enable the user to expand/collapse the menu by clicking on the empty space in the side bar (marked with red color).


| 1.0 | Add additional way to expand/collapse the side menu - Enable the user to expand/collapse the menu by clicking on the empty space in the side bar (marked with red color).


| non_priority | add additional way to expand collapse the side menu enable the user to expand collapse the menu by clicking on the empty space in the side bar marked with red color | 0 |
697,744 | 23,951,634,927 | IssuesEvent | 2022-09-12 12:02:09 | feast-dev/feast | https://api.github.com/repos/feast-dev/feast | opened | AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query' using google-cloud-bigquery 2.1.0 | kind/bug priority/p2 | ## Expected Behavior
I have found a corner case trying to install feast as dependency in a project. Poetry solved `google-cloud-bigquery` to `2.1.0` when installing `feast[gcp,redis]==0.24.1` along with other dependencies. But I was not able to use feast after installing it.
## Current Behavior
I got the following exception when running the command `feast apply`:
```python
Traceback (most recent call last):
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/importer.py", line 26, in import_class
module = importlib.import_module(module_name)
File "/Users/<hidden>/.pyenv/versions/3.9.13/lib/python3.9/importlib/__init__.py", line 127, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1030, in _gcd_import
File "<frozen importlib._bootstrap>", line 1007, in _find_and_load
File "<frozen importlib._bootstrap>", line 986, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 680, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 850, in exec_module
File "<frozen importlib._bootstrap>", line 228, in _call_with_frames_removed
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/infra/offline_stores/bigquery.py", line 378, in <module>
class BigQueryRetrievalJob(RetrievalJob):
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/infra/offline_stores/bigquery.py", line 478, in BigQueryRetrievalJob
) -> Optional[bigquery.job.query.QueryJob]:
AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/bin/feast", line 8, in <module>
sys.exit(cli())
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1130, in __call__
return self.main(*args, **kwargs)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1055, in main
rv = self.invoke(ctx)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1657, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1404, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 760, in invoke
return __callback(*args, **kwargs)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/decorators.py", line 26, in new_func
return f(get_current_context(), *args, **kwargs)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/cli.py", line 517, in apply_total_command
repo_config = load_repo_config(repo, fs_yaml_file)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 499, in load_repo_config
c = RepoConfig(**raw_config)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 166, in __init__
super().__init__(**data)
File "pydantic/main.py", line 340, in pydantic.main.BaseModel.__init__
File "pydantic/main.py", line 1050, in pydantic.main.validate_model
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 345, in _validate_offline_store_config
offline_config_class = get_offline_config_from_type(offline_store_type)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 480, in get_offline_config_from_type
return import_class(module_name, config_class_name, config_class_name)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/importer.py", line 31, in import_class
raise FeastModuleImportError(module_name, class_name) from e
feast.errors.FeastModuleImportError: Could not import module 'feast.infra.offline_stores.bigquery' while attempting to load class 'BigQueryOfflineStoreConfig'
```
This exception happens because Feast's `bigquery` module imports the class `google.cloud.bigquery.job.query.QueryJob`, but that path doesn't exist in the `google-cloud-bigquery 2.1.0`.
https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/bigquery.py#L478
```python
@log_exceptions_and_usage
def _execute_query(
self, query, job_config=None, timeout: int = 1800
) -> Optional[bigquery.job.query.QueryJob]:
```
I was able to use Feast only when `google-cloud-bigquery` was upgraded to higher versions (e.g. `2.36.0`), but I think Feast should not allow the resolution of a dependency version that is not working properly.
## Steps to reproduce
```
poetry init -n
poetry add google-cloud-bigquery==2.1.0
poetry add --extras gcp,bigquery feast==0.24.1
```
After installing the requirements you can try to create some feature views from BQ or just check if that path exists in such version.
```
poetry run python
```
```python
>>> from google.cloud import bigquery
>>> bigquery.job.query.QueryJob
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query'
```
### Specifications
- Version: 0.24.1
- Platform: Python 3.9 on MacOS
- Subsystem:
## Possible Solution
This problem could possibly be solved by:
1) Increasing the minimium version allowed for this dependency
2) Changing the import statement from `bigquery.job.query.QueryJob` to `bigquery.job.QueryJob`
| 1.0 | AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query' using google-cloud-bigquery 2.1.0 - ## Expected Behavior
I have found a corner case trying to install feast as dependency in a project. Poetry solved `google-cloud-bigquery` to `2.1.0` when installing `feast[gcp,redis]==0.24.1` along with other dependencies. But I was not able to use feast after installing it.
## Current Behavior
I got the following exception when running the command `feast apply`:
```python
Traceback (most recent call last):
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/importer.py", line 26, in import_class
module = importlib.import_module(module_name)
File "/Users/<hidden>/.pyenv/versions/3.9.13/lib/python3.9/importlib/__init__.py", line 127, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1030, in _gcd_import
File "<frozen importlib._bootstrap>", line 1007, in _find_and_load
File "<frozen importlib._bootstrap>", line 986, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 680, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 850, in exec_module
File "<frozen importlib._bootstrap>", line 228, in _call_with_frames_removed
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/infra/offline_stores/bigquery.py", line 378, in <module>
class BigQueryRetrievalJob(RetrievalJob):
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/infra/offline_stores/bigquery.py", line 478, in BigQueryRetrievalJob
) -> Optional[bigquery.job.query.QueryJob]:
AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/bin/feast", line 8, in <module>
sys.exit(cli())
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1130, in __call__
return self.main(*args, **kwargs)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1055, in main
rv = self.invoke(ctx)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1657, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1404, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 760, in invoke
return __callback(*args, **kwargs)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/decorators.py", line 26, in new_func
return f(get_current_context(), *args, **kwargs)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/cli.py", line 517, in apply_total_command
repo_config = load_repo_config(repo, fs_yaml_file)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 499, in load_repo_config
c = RepoConfig(**raw_config)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 166, in __init__
super().__init__(**data)
File "pydantic/main.py", line 340, in pydantic.main.BaseModel.__init__
File "pydantic/main.py", line 1050, in pydantic.main.validate_model
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 345, in _validate_offline_store_config
offline_config_class = get_offline_config_from_type(offline_store_type)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 480, in get_offline_config_from_type
return import_class(module_name, config_class_name, config_class_name)
File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/importer.py", line 31, in import_class
raise FeastModuleImportError(module_name, class_name) from e
feast.errors.FeastModuleImportError: Could not import module 'feast.infra.offline_stores.bigquery' while attempting to load class 'BigQueryOfflineStoreConfig'
```
This exception happens because Feast's `bigquery` module imports the class `google.cloud.bigquery.job.query.QueryJob`, but that path doesn't exist in the `google-cloud-bigquery 2.1.0`.
https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/bigquery.py#L478
```python
@log_exceptions_and_usage
def _execute_query(
self, query, job_config=None, timeout: int = 1800
) -> Optional[bigquery.job.query.QueryJob]:
```
I was able to use Feast only when `google-cloud-bigquery` was upgraded to higher versions (e.g. `2.36.0`), but I think Feast should not allow the resolution of a dependency version that is not working properly.
## Steps to reproduce
```
poetry init -n
poetry add google-cloud-bigquery==2.1.0
poetry add --extras gcp,bigquery feast==0.24.1
```
After installing the requirements you can try to create some feature views from BQ or just check if that path exists in such version.
```
poetry run python
```
```python
>>> from google.cloud import bigquery
>>> bigquery.job.query.QueryJob
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query'
```
### Specifications
- Version: 0.24.1
- Platform: Python 3.9 on MacOS
- Subsystem:
## Possible Solution
This problem could possibly be solved by:
1) Increasing the minimium version allowed for this dependency
2) Changing the import statement from `bigquery.job.query.QueryJob` to `bigquery.job.QueryJob`
| priority | attributeerror module google cloud bigquery job has no attribute query using google cloud bigquery expected behavior i have found a corner case trying to install feast as dependency in a project poetry solved google cloud bigquery to when installing feast along with other dependencies but i was not able to use feast after installing it current behavior i got the following exception when running the command feast apply python traceback most recent call last file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast importer py line in import class module importlib import module module name file users pyenv versions lib importlib init py line in import module return bootstrap gcd import name package level file line in gcd import file line in find and load file line in find and load unlocked file line in load unlocked file line in exec module file line in call with frames removed file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast infra offline stores bigquery py line in class bigqueryretrievaljob retrievaljob file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast infra offline stores bigquery py line in bigqueryretrievaljob optional attributeerror module google cloud bigquery job has no attribute query the above exception was the direct cause of the following exception traceback most recent call last file users library caches pypoetry virtualenvs feast example zjueihwy bin feast line in sys exit cli file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click core py line in call return self main args kwargs file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click core py line in main rv self invoke ctx file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click core py line in invoke return ctx invoke self callback ctx params file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click core py line in invoke return callback args kwargs file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click decorators py line in new func return f get current context args kwargs file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast cli py line in apply total command repo config load repo config repo fs yaml file file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast repo config py line in load repo config c repoconfig raw config file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast repo config py line in init super init data file pydantic main py line in pydantic main basemodel init file pydantic main py line in pydantic main validate model file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast repo config py line in validate offline store config offline config class get offline config from type offline store type file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast repo config py line in get offline config from type return import class module name config class name config class name file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast importer py line in import class raise feastmoduleimporterror module name class name from e feast errors feastmoduleimporterror could not import module feast infra offline stores bigquery while attempting to load class bigqueryofflinestoreconfig this exception happens because feast s bigquery module imports the class google cloud bigquery job query queryjob but that path doesn t exist in the google cloud bigquery python log exceptions and usage def execute query self query job config none timeout int optional i was able to use feast only when google cloud bigquery was upgraded to higher versions e g but i think feast should not allow the resolution of a dependency version that is not working properly steps to reproduce poetry init n poetry add google cloud bigquery poetry add extras gcp bigquery feast after installing the requirements you can try to create some feature views from bq or just check if that path exists in such version poetry run python python from google cloud import bigquery bigquery job query queryjob traceback most recent call last file line in attributeerror module google cloud bigquery job has no attribute query specifications version platform python on macos subsystem possible solution this problem could possibly be solved by increasing the minimium version allowed for this dependency changing the import statement from bigquery job query queryjob to bigquery job queryjob | 1 |
20,535 | 2,622,852,322 | IssuesEvent | 2015-03-04 08:05:55 | max99x/pagemon-chrome-ext | https://api.github.com/repos/max99x/pagemon-chrome-ext | closed | Detecting the time of the problem(检测时间的问题 ) | auto-migrated Priority-Medium | ```
What steps will reproduce the problem?
Download something or play games. Usually will be the
case!(在下载东西的时候或者玩游戏的时候。平时也会出现这
样的情况!)
What is the expected output? What do you see instead?
Please see the QQ截图20121204201211.jpg!!
What version of the Chrome are you using? On what operating system?
版本 23.0.1266.0 canary and 360极速浏览器(内核:Branch of Chromium
21.0.1180.89).
system:windows 7 ultimate
What are the error log messages? To view them, go to Wrench -> Tools ->
Extensions -> Developer Mode -> expand Page Monitor -> "background.htm" ->
Console.
Please see the QQ截图20121204203115.jpg!!!
Please provide any additional information below.
Hello, I have a question to ask!When I use your plug-in to detect the update status of a page?I choose the interval is 10 seconds, sometimes I state into the plug-page view to the "detected last more than 10 seconds, in which case I downloaded the file did not download something also occurkinds of situations! !This is BUG it?I really do not understand! ! ! !(你好,我有一个问题想问!当我用你的插件去检测一个网页的更新状态时?我选择的间隔时间是10秒,有时候我进插件的状态页面查看到的"上次检测"大于10秒,这种情况是在我下载文件时发生的,没下载东西时也会发生这种情况!!这是BUG吗?真不明白!!!!)
```
Original issue reported on code.google.com by `qq25592...@gmail.com` on 4 Dec 2012 at 12:34
Attachments:
* [QQ截图20121204201211.jpg](https://storage.googleapis.com/google-code-attachments/pagemon-chrome-ext/issue-220/comment-0/QQ截图20121204201211.jpg)
* [QQ截图20121204203115.jpg](https://storage.googleapis.com/google-code-attachments/pagemon-chrome-ext/issue-220/comment-0/QQ截图20121204203115.jpg)
| 1.0 | Detecting the time of the problem(检测时间的问题 ) - ```
What steps will reproduce the problem?
Download something or play games. Usually will be the
case!(在下载东西的时候或者玩游戏的时候。平时也会出现这
样的情况!)
What is the expected output? What do you see instead?
Please see the QQ截图20121204201211.jpg!!
What version of the Chrome are you using? On what operating system?
版本 23.0.1266.0 canary and 360极速浏览器(内核:Branch of Chromium
21.0.1180.89).
system:windows 7 ultimate
What are the error log messages? To view them, go to Wrench -> Tools ->
Extensions -> Developer Mode -> expand Page Monitor -> "background.htm" ->
Console.
Please see the QQ截图20121204203115.jpg!!!
Please provide any additional information below.
Hello, I have a question to ask!When I use your plug-in to detect the update status of a page?I choose the interval is 10 seconds, sometimes I state into the plug-page view to the "detected last more than 10 seconds, in which case I downloaded the file did not download something also occurkinds of situations! !This is BUG it?I really do not understand! ! ! !(你好,我有一个问题想问!当我用你的插件去检测一个网页的更新状态时?我选择的间隔时间是10秒,有时候我进插件的状态页面查看到的"上次检测"大于10秒,这种情况是在我下载文件时发生的,没下载东西时也会发生这种情况!!这是BUG吗?真不明白!!!!)
```
Original issue reported on code.google.com by `qq25592...@gmail.com` on 4 Dec 2012 at 12:34
Attachments:
* [QQ截图20121204201211.jpg](https://storage.googleapis.com/google-code-attachments/pagemon-chrome-ext/issue-220/comment-0/QQ截图20121204201211.jpg)
* [QQ截图20121204203115.jpg](https://storage.googleapis.com/google-code-attachments/pagemon-chrome-ext/issue-220/comment-0/QQ截图20121204203115.jpg)
| priority | detecting the time of the problem(检测时间的问题 ) what steps will reproduce the problem download something or play games usually will be the case (在下载东西的时候或者玩游戏的时候。平时也会出现这 样的情况!) what is the expected output what do you see instead please see the jpg what version of the chrome are you using on what operating system 版本 canary and (内核:branch of chromium ) system windows ultimate what are the error log messages to view them go to wrench tools extensions developer mode expand page monitor background htm console please see the jpg!!! please provide any additional information below hello i have a question to ask when i use your plug in to detect the update status of a page i choose the interval is seconds sometimes i state into the plug page view to the detected last more than seconds in which case i downloaded the file did not download something also occurkinds of situations this is bug it i really do not understand (你好,我有一个问题想问!当我用你的插件去检测一个网页的更新状态时? ,有时候我进插件的状态页面查看到的 上次检测 这种情况是在我下载文件时发生的,没下载东西时也会发生这种情况!!这是bug吗?真不明白!!!!) original issue reported on code google com by gmail com on dec at attachments | 1 |
604,772 | 18,718,506,613 | IssuesEvent | 2021-11-03 09:04:23 | magento/magento2-phpstorm-plugin | https://api.github.com/repos/magento/magento2-phpstorm-plugin | opened | [UCT:Inspection] Importing non-existent interface | Award: special achievement Priority: P1 Upgrade Compatibility Tool | ### Description
The purpose of this inspection is to find places where there is imported interface that is removed in any of the next released versions.
**For example:**
The client has Magento V1 and wants to upgrade it to the V2 (versions are abstract to describe the idea). Some class has the TestInterface in imports. In the V2 that interface is removed. That inspection should highlight importing of this interface and describe the issue.
**The UML diagram of the example case:**

### Key points for implementation
- all PHP inspections should be grouped under the `com.magento.idea.magento2uct.inspections` package.
- all PHP inspections must be inherited from the com.jetbrains.php.lang.inspections.PhpInspection
- all inspections should have checking if it should be run with the current tool configurations
- all inspections should cover cases for running from the UCT tool run option and from the PHPStorm itself
- if covered area is not already supported it should be checked and added to the `com.magento.idea.magento2uct.inspections.UctInspectionManager`
- each new inspection must be registered in PHPStorm way in the plugin.xml file
- each new inspection must be described in the appropriate html file: `resources/inspectionDescriptions/FileName.html` (to be available in the inspections enable/disable menu)
- each new inspection must be registered for running from the UCT tool run option in the `com.magento.idea.magento2uct.packages.SupportedIssue`
Approximate time to develop: 2h ~ 3h | 1.0 | [UCT:Inspection] Importing non-existent interface - ### Description
The purpose of this inspection is to find places where there is imported interface that is removed in any of the next released versions.
**For example:**
The client has Magento V1 and wants to upgrade it to the V2 (versions are abstract to describe the idea). Some class has the TestInterface in imports. In the V2 that interface is removed. That inspection should highlight importing of this interface and describe the issue.
**The UML diagram of the example case:**

### Key points for implementation
- all PHP inspections should be grouped under the `com.magento.idea.magento2uct.inspections` package.
- all PHP inspections must be inherited from the com.jetbrains.php.lang.inspections.PhpInspection
- all inspections should have checking if it should be run with the current tool configurations
- all inspections should cover cases for running from the UCT tool run option and from the PHPStorm itself
- if covered area is not already supported it should be checked and added to the `com.magento.idea.magento2uct.inspections.UctInspectionManager`
- each new inspection must be registered in PHPStorm way in the plugin.xml file
- each new inspection must be described in the appropriate html file: `resources/inspectionDescriptions/FileName.html` (to be available in the inspections enable/disable menu)
- each new inspection must be registered for running from the UCT tool run option in the `com.magento.idea.magento2uct.packages.SupportedIssue`
Approximate time to develop: 2h ~ 3h | priority | importing non existent interface description the purpose of this inspection is to find places where there is imported interface that is removed in any of the next released versions for example the client has magento and wants to upgrade it to the versions are abstract to describe the idea some class has the testinterface in imports in the that interface is removed that inspection should highlight importing of this interface and describe the issue the uml diagram of the example case key points for implementation all php inspections should be grouped under the com magento idea inspections package all php inspections must be inherited from the com jetbrains php lang inspections phpinspection all inspections should have checking if it should be run with the current tool configurations all inspections should cover cases for running from the uct tool run option and from the phpstorm itself if covered area is not already supported it should be checked and added to the com magento idea inspections uctinspectionmanager each new inspection must be registered in phpstorm way in the plugin xml file each new inspection must be described in the appropriate html file resources inspectiondescriptions filename html to be available in the inspections enable disable menu each new inspection must be registered for running from the uct tool run option in the com magento idea packages supportedissue approximate time to develop | 1 |
160,494 | 13,791,077,852 | IssuesEvent | 2020-10-09 11:31:57 | ai4er-cdt/resources | https://api.github.com/repos/ai4er-cdt/resources | closed | Create cookiecutter template for AI4ER | documentation enhancement | [] Create cookiecutter template for AI4ER
[] add environment instatiation
[] explain
[] link to wiki | 1.0 | Create cookiecutter template for AI4ER - [] Create cookiecutter template for AI4ER
[] add environment instatiation
[] explain
[] link to wiki | non_priority | create cookiecutter template for create cookiecutter template for add environment instatiation explain link to wiki | 0 |
16,407 | 2,615,040,262 | IssuesEvent | 2015-03-01 03:34:54 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | opened | Java client for Google Maps Data API | auto-migrated Priority-Medium Type-Sample | ```
Which Google API and version (e.g. Google Calendar Data API version 2)?
Google Maps Data API
What format (e.g. JSON, Atom)?
JSON
What Authentation (e.g. OAuth, OAuth 2, ClientLogin)?
ClientLogin
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Java 6
External references, such as API reference guide?
Please provide any additional information below.
thank you!
```
Original issue reported on code.google.com by `vincenzo...@gmail.com` on 20 Apr 2012 at 12:43 | 1.0 | Java client for Google Maps Data API - ```
Which Google API and version (e.g. Google Calendar Data API version 2)?
Google Maps Data API
What format (e.g. JSON, Atom)?
JSON
What Authentation (e.g. OAuth, OAuth 2, ClientLogin)?
ClientLogin
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Java 6
External references, such as API reference guide?
Please provide any additional information below.
thank you!
```
Original issue reported on code.google.com by `vincenzo...@gmail.com` on 20 Apr 2012 at 12:43 | priority | java client for google maps data api which google api and version e g google calendar data api version google maps data api what format e g json atom json what authentation e g oauth oauth clientlogin clientlogin java environment e g java android app engine java external references such as api reference guide please provide any additional information below thank you original issue reported on code google com by vincenzo gmail com on apr at | 1 |
66,211 | 6,992,933,006 | IssuesEvent | 2017-12-15 09:20:41 | owncloud/core | https://api.github.com/repos/owncloud/core | closed | UI tests should be skippable for specific oC versions | 1 - To develop dev:acceptance-tests QA | ### Steps to reproduce
1. run UI tests on older versions of oC (e.g. 10.0.3)
2. tests that check for bugs that are fixed in master/stable10 branch but not in the tested release are failing
### Expected behaviour
it should be possible to skip tests that are known to fail in a specific version. e.g. `@skipBeforeOC10.0.4`
@phil-davis @patrickjahns @davitol @SergioBertolinSG comments?
| 1.0 | UI tests should be skippable for specific oC versions - ### Steps to reproduce
1. run UI tests on older versions of oC (e.g. 10.0.3)
2. tests that check for bugs that are fixed in master/stable10 branch but not in the tested release are failing
### Expected behaviour
it should be possible to skip tests that are known to fail in a specific version. e.g. `@skipBeforeOC10.0.4`
@phil-davis @patrickjahns @davitol @SergioBertolinSG comments?
| non_priority | ui tests should be skippable for specific oc versions steps to reproduce run ui tests on older versions of oc e g tests that check for bugs that are fixed in master branch but not in the tested release are failing expected behaviour it should be possible to skip tests that are known to fail in a specific version e g phil davis patrickjahns davitol sergiobertolinsg comments | 0 |
475,020 | 13,685,955,981 | IssuesEvent | 2020-09-30 07:59:17 | incognitochain/incognito-chain | https://api.github.com/repos/incognitochain/incognito-chain | closed | Got both output coin v1 and v2 after sending transactions | Priority: High Type: Bug | related to: #963
While running regression tests of sending output coins and init coins, got following error:
`Cannot get tx version because there are 2 versions of input coins\nCan not create tx`
| 1.0 | Got both output coin v1 and v2 after sending transactions - related to: #963
While running regression tests of sending output coins and init coins, got following error:
`Cannot get tx version because there are 2 versions of input coins\nCan not create tx`
| priority | got both output coin and after sending transactions related to while running regression tests of sending output coins and init coins got following error cannot get tx version because there are versions of input coins ncan not create tx | 1 |
459,465 | 13,193,714,850 | IssuesEvent | 2020-08-13 15:38:28 | IHTSDO/snomed-owl-toolkit | https://api.github.com/repos/IHTSDO/snomed-owl-toolkit | closed | Silenced out of memory error | bug medium-priority | Hi all, the last week I created an issue because I was not able to generate the owl ontology file for the last version of Snomed (#37). The generated owl showed an incorrect hierarchy of concepts, where most of them were outside the Snomed root concept. This was provoked because there was an out of memory error, which was silenced in the process.
I am doing experiments with the different versions of snomed along the time, and I repeated the process by assigning 4GB of memory. Then, I noticed that the owl files generated with 2GB of memory were, in general, different to the files generated with 4GB. For example, the following figure shows the snomed 2019-01-31 owl file generated with 2Gb and with 2GB:

As you can see, annotations are created incorrectly, but they could be legit so I did not detect the error.
Would it be possible to show this out of memory error, at least in the log? | 1.0 | Silenced out of memory error - Hi all, the last week I created an issue because I was not able to generate the owl ontology file for the last version of Snomed (#37). The generated owl showed an incorrect hierarchy of concepts, where most of them were outside the Snomed root concept. This was provoked because there was an out of memory error, which was silenced in the process.
I am doing experiments with the different versions of snomed along the time, and I repeated the process by assigning 4GB of memory. Then, I noticed that the owl files generated with 2GB of memory were, in general, different to the files generated with 4GB. For example, the following figure shows the snomed 2019-01-31 owl file generated with 2Gb and with 2GB:

As you can see, annotations are created incorrectly, but they could be legit so I did not detect the error.
Would it be possible to show this out of memory error, at least in the log? | priority | silenced out of memory error hi all the last week i created an issue because i was not able to generate the owl ontology file for the last version of snomed the generated owl showed an incorrect hierarchy of concepts where most of them were outside the snomed root concept this was provoked because there was an out of memory error which was silenced in the process i am doing experiments with the different versions of snomed along the time and i repeated the process by assigning of memory then i noticed that the owl files generated with of memory were in general different to the files generated with for example the following figure shows the snomed owl file generated with and with as you can see annotations are created incorrectly but they could be legit so i did not detect the error would it be possible to show this out of memory error at least in the log | 1 |
194,126 | 6,891,769,381 | IssuesEvent | 2017-11-22 18:15:08 | vmware/vic | https://api.github.com/repos/vmware/vic | closed | Image pull from artifactory fails: Failed to obtain OAuth endpoint: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403 | kind/bug priority/high team/container | After successful login to artifcatory repo, image pull fails:
```
Nov 9 2017 14:58:46.410Z INFO PullImage: reference: vic-docker-local.artifactory.eng.vmware.com/busybox:1, 127.0.0.1:2377, portlayer: "127.0.0.1:2377"
Nov 9 2017 14:58:46.410Z INFO Using UUID (564d1c2f-2a64-983b-1ec9-4185f86f38c7) for imagestore name
Nov 9 2017 14:58:46.410Z DEBUG Running with portlayer
Nov 9 2017 14:58:46.410Z DEBUG [BEGIN] [vic/lib/imagec.PingPortLayer:40] 127.0.0.1:2377
Nov 9 2017 14:58:46.411Z DEBUG [ END ] [vic/lib/imagec.PingPortLayer:40] [385.047µs] 127.0.0.1:2377
Nov 9 2017 14:58:46.411Z DEBUG [BEGIN] [vic/lib/imagec.LearnRegistryURL:75] vic-docker-local.artifactory.eng.vmware.com
Nov 9 2017 14:58:46.411Z DEBUG Trying https scheme for &imagec.Options{Reference:(*reference.taggedRef)(0xc420653e60), Registry:"vic-docker-local.artifactory.eng.vmware.com", Image:"busybox", Tag:"1", Destination:"/tmp", Host:"127.0.0.1:2377", Storename:"564d1c2f-2a64-983b-1ec9-4185f86f38c7", Username:"vic-deployer", Password:"vmware!123", Token:(*fetcher.Token)(nil), Timeout:3600000000000, Outstream:(*ioutils.WriteFlusher)(0xc42059fa40), InsecureSkipVerify:false, InsecureAllowHTTP:true, ImageManifestSchema1:(*imagec.Manifest)(nil), ImageManifestSchema2:(*schema2.DeserializedManifest)(nil), ManifestDigest:"", RegistryCAs:(*x509.CertPool)(0xc4207daf30), Standalone:false, ImageStore:""}
Nov 9 2017 14:58:46.411Z DEBUG URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/
Nov 9 2017 14:58:46.550Z DEBUG [ END ] [vic/lib/imagec.LearnRegistryURL:75] [138.852068ms] vic-docker-local.artifactory.eng.vmware.com
Nov 9 2017 14:58:46.550Z DEBUG [BEGIN] [vic/lib/imagec.LearnAuthURL:100] vic-docker-local.artifactory.eng.vmware.com/busybox:1
Nov 9 2017 14:58:46.550Z DEBUG URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1
Nov 9 2017 14:58:46.587Z DEBUG [ END ] [vic/lib/imagec.LearnAuthURL:100] [37.614669ms] vic-docker-local.artifactory.eng.vmware.com/busybox:1
Nov 9 2017 14:58:46.587Z INFO https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403, URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1
Nov 9 2017 14:58:46.587Z DEBUG [ END ] [vic/lib/apiservers/engine/backends.(*Image).PullImage:322] [226.636362ms]
Nov 9 2017 14:58:46.587Z ERROR Handler for POST /v1.21/images/create returned error: Failed to obtain OAuth endpoint: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403, URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1
```
[21-2-Artifactory-VCH-0-7754-container-logs.zip](https://github.com/vmware/vic/files/1458276/21-2-Artifactory-VCH-0-7754-container-logs.zip)
| 1.0 | Image pull from artifactory fails: Failed to obtain OAuth endpoint: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403 - After successful login to artifcatory repo, image pull fails:
```
Nov 9 2017 14:58:46.410Z INFO PullImage: reference: vic-docker-local.artifactory.eng.vmware.com/busybox:1, 127.0.0.1:2377, portlayer: "127.0.0.1:2377"
Nov 9 2017 14:58:46.410Z INFO Using UUID (564d1c2f-2a64-983b-1ec9-4185f86f38c7) for imagestore name
Nov 9 2017 14:58:46.410Z DEBUG Running with portlayer
Nov 9 2017 14:58:46.410Z DEBUG [BEGIN] [vic/lib/imagec.PingPortLayer:40] 127.0.0.1:2377
Nov 9 2017 14:58:46.411Z DEBUG [ END ] [vic/lib/imagec.PingPortLayer:40] [385.047µs] 127.0.0.1:2377
Nov 9 2017 14:58:46.411Z DEBUG [BEGIN] [vic/lib/imagec.LearnRegistryURL:75] vic-docker-local.artifactory.eng.vmware.com
Nov 9 2017 14:58:46.411Z DEBUG Trying https scheme for &imagec.Options{Reference:(*reference.taggedRef)(0xc420653e60), Registry:"vic-docker-local.artifactory.eng.vmware.com", Image:"busybox", Tag:"1", Destination:"/tmp", Host:"127.0.0.1:2377", Storename:"564d1c2f-2a64-983b-1ec9-4185f86f38c7", Username:"vic-deployer", Password:"vmware!123", Token:(*fetcher.Token)(nil), Timeout:3600000000000, Outstream:(*ioutils.WriteFlusher)(0xc42059fa40), InsecureSkipVerify:false, InsecureAllowHTTP:true, ImageManifestSchema1:(*imagec.Manifest)(nil), ImageManifestSchema2:(*schema2.DeserializedManifest)(nil), ManifestDigest:"", RegistryCAs:(*x509.CertPool)(0xc4207daf30), Standalone:false, ImageStore:""}
Nov 9 2017 14:58:46.411Z DEBUG URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/
Nov 9 2017 14:58:46.550Z DEBUG [ END ] [vic/lib/imagec.LearnRegistryURL:75] [138.852068ms] vic-docker-local.artifactory.eng.vmware.com
Nov 9 2017 14:58:46.550Z DEBUG [BEGIN] [vic/lib/imagec.LearnAuthURL:100] vic-docker-local.artifactory.eng.vmware.com/busybox:1
Nov 9 2017 14:58:46.550Z DEBUG URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1
Nov 9 2017 14:58:46.587Z DEBUG [ END ] [vic/lib/imagec.LearnAuthURL:100] [37.614669ms] vic-docker-local.artifactory.eng.vmware.com/busybox:1
Nov 9 2017 14:58:46.587Z INFO https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403, URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1
Nov 9 2017 14:58:46.587Z DEBUG [ END ] [vic/lib/apiservers/engine/backends.(*Image).PullImage:322] [226.636362ms]
Nov 9 2017 14:58:46.587Z ERROR Handler for POST /v1.21/images/create returned error: Failed to obtain OAuth endpoint: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403, URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1
```
[21-2-Artifactory-VCH-0-7754-container-logs.zip](https://github.com/vmware/vic/files/1458276/21-2-Artifactory-VCH-0-7754-container-logs.zip)
| priority | image pull from artifactory fails failed to obtain oauth endpoint returned an unexpected response unexpected http code after successful login to artifcatory repo image pull fails nov info pullimage reference vic docker local artifactory eng vmware com busybox portlayer nov info using uuid for imagestore name nov debug running with portlayer nov debug nov debug nov debug vic docker local artifactory eng vmware com nov debug trying https scheme for imagec options reference reference taggedref registry vic docker local artifactory eng vmware com image busybox tag destination tmp host storename username vic deployer password vmware token fetcher token nil timeout outstream ioutils writeflusher insecureskipverify false insecureallowhttp true imagec manifest nil deserializedmanifest nil manifestdigest registrycas certpool standalone false imagestore nov debug url nov debug vic docker local artifactory eng vmware com nov debug vic docker local artifactory eng vmware com busybox nov debug url nov debug vic docker local artifactory eng vmware com busybox nov info returned an unexpected response unexpected http code url nov debug nov error handler for post images create returned error failed to obtain oauth endpoint returned an unexpected response unexpected http code url | 1 |
128,805 | 5,076,210,360 | IssuesEvent | 2016-12-28 00:38:47 | NUKnightLab/StoryMapJS | https://api.github.com/repos/NUKnightLab/StoryMapJS | closed | New Users cannot sign in | bug priority: high | Likely introduced with changes to Google Drive. New users who have not before logged into StoryMap are receiving the error: Error getting migration list: 'google'
Currently, no workaround is known
| 1.0 | New Users cannot sign in - Likely introduced with changes to Google Drive. New users who have not before logged into StoryMap are receiving the error: Error getting migration list: 'google'
Currently, no workaround is known
| priority | new users cannot sign in likely introduced with changes to google drive new users who have not before logged into storymap are receiving the error error getting migration list google currently no workaround is known | 1 |
156,943 | 5,995,231,607 | IssuesEvent | 2017-06-03 01:25:07 | kmycode/gunucco | https://api.github.com/repos/kmycode/gunucco | closed | List-up and display private or user-only chapters on web client view page when user logining | enhancement priority-high | I forgot adding #32. | 1.0 | List-up and display private or user-only chapters on web client view page when user logining - I forgot adding #32. | priority | list up and display private or user only chapters on web client view page when user logining i forgot adding | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.