Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
433,630
12,508,038,527
IssuesEvent
2020-06-02 14:58:49
canonical-web-and-design/build.snapcraft.io
https://api.github.com/repos/canonical-web-and-design/build.snapcraft.io
closed
Should not build i386 snaps with base: core20
Priority: High
## Summary With the release of Ubuntu 20.04, snaps that specify `base: core20` should not auto-trigger builds on the i386 architecture because that will not be possible since there is not an Ubuntu 20.04 for i386, and thus there is not a core20 base snap for i386. ## Process Create a snapcraft.yaml with `base: core20` such as this one: https://github.com/anonymouse64/ubuntu-core-initramfs-snap ## Current and expected result I expected build.snapcraft.io to not try to build for i386 automatically when specifying a snap with `base: core20` in the snapcraft.yaml of a GitHub issue. Instead it triggered a i386 build, which failed on LP like this: ``` linux32: failed to execute snap: No such file or directory Install failed Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/lpbuildd/target/build_snap.py", line 259, in run self.install() File "/usr/lib/python2.7/dist-packages/lpbuildd/target/build_snap.py", line 163, in install snap_name]) File "/usr/lib/python2.7/dist-packages/lpbuildd/target/lxd.py", line 536, in run subprocess.check_call(cmd, **kwargs) File "/usr/lib/python2.7/subprocess.py", line 541, in check_call raise CalledProcessError(retcode, cmd) CalledProcessError: Command '['lxc', 'exec', 'lp-focal-i386', '--', 'linux32', 'snap', 'install', '--channel=stable', 'core20']' returned non-zero exit status 127 ``` See also https://bugs.launchpad.net/launchpad/+bug/1862258 for LP side
1.0
Should not build i386 snaps with base: core20 - ## Summary With the release of Ubuntu 20.04, snaps that specify `base: core20` should not auto-trigger builds on the i386 architecture because that will not be possible since there is not an Ubuntu 20.04 for i386, and thus there is not a core20 base snap for i386. ## Process Create a snapcraft.yaml with `base: core20` such as this one: https://github.com/anonymouse64/ubuntu-core-initramfs-snap ## Current and expected result I expected build.snapcraft.io to not try to build for i386 automatically when specifying a snap with `base: core20` in the snapcraft.yaml of a GitHub issue. Instead it triggered a i386 build, which failed on LP like this: ``` linux32: failed to execute snap: No such file or directory Install failed Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/lpbuildd/target/build_snap.py", line 259, in run self.install() File "/usr/lib/python2.7/dist-packages/lpbuildd/target/build_snap.py", line 163, in install snap_name]) File "/usr/lib/python2.7/dist-packages/lpbuildd/target/lxd.py", line 536, in run subprocess.check_call(cmd, **kwargs) File "/usr/lib/python2.7/subprocess.py", line 541, in check_call raise CalledProcessError(retcode, cmd) CalledProcessError: Command '['lxc', 'exec', 'lp-focal-i386', '--', 'linux32', 'snap', 'install', '--channel=stable', 'core20']' returned non-zero exit status 127 ``` See also https://bugs.launchpad.net/launchpad/+bug/1862258 for LP side
priority
should not build snaps with base summary with the release of ubuntu snaps that specify base should not auto trigger builds on the architecture because that will not be possible since there is not an ubuntu for and thus there is not a base snap for process create a snapcraft yaml with base such as this one current and expected result i expected build snapcraft io to not try to build for automatically when specifying a snap with base in the snapcraft yaml of a github issue instead it triggered a build which failed on lp like this failed to execute snap no such file or directory install failed traceback most recent call last file usr lib dist packages lpbuildd target build snap py line in run self install file usr lib dist packages lpbuildd target build snap py line in install snap name file usr lib dist packages lpbuildd target lxd py line in run subprocess check call cmd kwargs file usr lib subprocess py line in check call raise calledprocesserror retcode cmd calledprocesserror command returned non zero exit status see also for lp side
1
61,674
25,598,549,355
IssuesEvent
2022-12-01 18:05:05
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
closed
[Enhancement]: Add ECR Registry Permissions Resource
enhancement new-resource service/ecr
### Description I would like to be able to define the registry permissions JSON, which is necessary for cross account replication on the destination end. So https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/ecr_replication_configuration covers the origin part, but not the destination one. ### Affected Resource(s) and/or Data Source(s) * aws_ecr_replication_configuration (not per se affected, but related) ### Potential Terraform Configuration _No response_ ### References _No response_ ### Would you like to implement a fix? _No response_
1.0
[Enhancement]: Add ECR Registry Permissions Resource - ### Description I would like to be able to define the registry permissions JSON, which is necessary for cross account replication on the destination end. So https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/ecr_replication_configuration covers the origin part, but not the destination one. ### Affected Resource(s) and/or Data Source(s) * aws_ecr_replication_configuration (not per se affected, but related) ### Potential Terraform Configuration _No response_ ### References _No response_ ### Would you like to implement a fix? _No response_
non_priority
add ecr registry permissions resource description i would like to be able to define the registry permissions json which is necessary for cross account replication on the destination end so covers the origin part but not the destination one affected resource s and or data source s aws ecr replication configuration not per se affected but related potential terraform configuration no response references no response would you like to implement a fix no response
0
83,355
3,633,955,882
IssuesEvent
2016-02-11 16:21:36
rsanchez-wsu/sp16-ceg3120
https://api.github.com/repos/rsanchez-wsu/sp16-ceg3120
closed
Fix checkstyle issues with team 6 branch.
priority-high state-inprogress team-6
Code submission works on individual workstations but fails when Jenkins attempts to build it.
1.0
Fix checkstyle issues with team 6 branch. - Code submission works on individual workstations but fails when Jenkins attempts to build it.
priority
fix checkstyle issues with team branch code submission works on individual workstations but fails when jenkins attempts to build it
1
6,720
6,609,341,277
IssuesEvent
2017-09-19 14:15:23
ekylibre/ekylibre
https://api.github.com/repos/ekylibre/ekylibre
closed
Infinite map = infinite cultivable zone
Bug Cartography Security
When you create a new zone or modify an existent one, there is no limit on the size of it. The problem is the map is not a loop, it's a patern which repeat as many time as you want. If the cultivable zone is 2x earth size there is a problem, an error is send and you have a zone which is at the same place on multiple zone on the map.
True
Infinite map = infinite cultivable zone - When you create a new zone or modify an existent one, there is no limit on the size of it. The problem is the map is not a loop, it's a patern which repeat as many time as you want. If the cultivable zone is 2x earth size there is a problem, an error is send and you have a zone which is at the same place on multiple zone on the map.
non_priority
infinite map infinite cultivable zone when you create a new zone or modify an existent one there is no limit on the size of it the problem is the map is not a loop it s a patern which repeat as many time as you want if the cultivable zone is earth size there is a problem an error is send and you have a zone which is at the same place on multiple zone on the map
0
65,988
16,518,013,592
IssuesEvent
2021-05-26 11:49:44
angular/angular-cli
https://api.github.com/repos/angular/angular-cli
closed
ng e2e does not use --proxy-config for webdriver-manager update
comp: devkit/build-angular type: feature
<!-- IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING --> ### Bug Report or Feature Request (mark with an `x`) ``` - [x] bug report -> please search issues before submitting - [ ] feature request ``` ### Versions. <!-- Output from: `ng --version`. If nothing, output from: `node --version` and `npm --version`. Windows (7/8/10). Linux (incl. distribution). macOS (El Capitan? Sierra?) --> @angular/cli: 1.0.0 node: 6.10.1 os: win32 x64 @angular/common: 4.1.2 @angular/compiler: 4.1.2 @angular/core: 4.1.2 @angular/forms: 4.1.2 @angular/http: 4.1.2 @angular/platform-browser: 4.1.2 @angular/platform-browser-dynamic: 4.1.2 @angular/router: 4.1.2 @angular/cli: 1.0.0 @angular/compiler-cli: 4.1.2 ### Repro steps. <!-- Simple steps to reproduce this bug. Please include: commands run, packages added, related code changes. A link to a sample repo would help too. --> 1. Must have an out of date web driver or delete it from `node_modules/protractor/node_modules/webdriver-manager/selenium/`. For example, if using Chrome webdriver, delete all `chrome*` files in this selenium folder. 2. Must be behind a proxy, but without HTTPS_PROXY or https_proxy environment variables set. 3. Using the proxy guide described in [the cli wiki](https://github.com/angular/angular-cli/wiki/stories-using-corporate-proxy), create a `proxy.conf.js` file to pass to `ng e2e`. 4. Run `ng e2e --proxy-config proxy.conf.js`. If you are behind a proxy and https_proxy is not set, the command will fail after compiling the app. This is because by default, `ng e2e` will run `webdriver-manager update` to fetch the right browser drivers, but does not use the proxy config to do so. ### The log given by the failure. <!-- Normally this include a stack trace and some more information. --> ``` $ ng e2e --proxy-config proxy.conf.js ** NG Live Development Server is running on http://localhost:49152 ** Hash: 95e0fe13c75ed4802b8b Time: 17998ms chunk {0} 0.chunk.js, 0.chunk.js.map 491 kB {3} [rendered] chunk {1} polyfills.bundle.js, polyfills.bundle.js.map (polyfills) 237 kB {5} [initial] [rendered] chunk {2} styles.bundle.js, styles.bundle.js.map (styles) 226 kB {5} [initial] [rendered] chunk {3} main.bundle.js, main.bundle.js.map (main) 8.25 kB {4} [initial] [rendered] chunk {4} vendor.bundle.js, vendor.bundle.js.map (vendor) 3.12 MB [initial] [rendered] chunk {5} inline.bundle.js, inline.bundle.js.map (inline) 0 bytes [entry] [rendered] webpack: Compiled successfully. events.js:160 throw er; // Unhandled 'error' event ^ Error: connect ECONNREFUSED 172.217.8.16:443 at Object.exports._errnoException (util.js:1018:11) at exports._exceptionWithHostPort (util.js:1041:20) at TCPConnectWrap.afterConnect [as oncomplete] (net.js:1090:14) ``` ### Desired functionality. <!-- What would like to see implemented? What is the usecase? --> `ng e2e` should pass along proxy configuration to webdriver-manager when the `--webdriver-update` flag is set to true. For webdriver-manager, the flag to use is `--proxy`. ### Mention any other details that might be useful. <!-- Please include a link to the repo if this is related to an OSS project. --> The code that calls webdriver-manager is at Line 65 in [e2e.ts](https://github.com/angular/angular-cli/blob/master/packages/%40angular/cli/tasks/e2e.ts#L65).
1.0
ng e2e does not use --proxy-config for webdriver-manager update - <!-- IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING --> ### Bug Report or Feature Request (mark with an `x`) ``` - [x] bug report -> please search issues before submitting - [ ] feature request ``` ### Versions. <!-- Output from: `ng --version`. If nothing, output from: `node --version` and `npm --version`. Windows (7/8/10). Linux (incl. distribution). macOS (El Capitan? Sierra?) --> @angular/cli: 1.0.0 node: 6.10.1 os: win32 x64 @angular/common: 4.1.2 @angular/compiler: 4.1.2 @angular/core: 4.1.2 @angular/forms: 4.1.2 @angular/http: 4.1.2 @angular/platform-browser: 4.1.2 @angular/platform-browser-dynamic: 4.1.2 @angular/router: 4.1.2 @angular/cli: 1.0.0 @angular/compiler-cli: 4.1.2 ### Repro steps. <!-- Simple steps to reproduce this bug. Please include: commands run, packages added, related code changes. A link to a sample repo would help too. --> 1. Must have an out of date web driver or delete it from `node_modules/protractor/node_modules/webdriver-manager/selenium/`. For example, if using Chrome webdriver, delete all `chrome*` files in this selenium folder. 2. Must be behind a proxy, but without HTTPS_PROXY or https_proxy environment variables set. 3. Using the proxy guide described in [the cli wiki](https://github.com/angular/angular-cli/wiki/stories-using-corporate-proxy), create a `proxy.conf.js` file to pass to `ng e2e`. 4. Run `ng e2e --proxy-config proxy.conf.js`. If you are behind a proxy and https_proxy is not set, the command will fail after compiling the app. This is because by default, `ng e2e` will run `webdriver-manager update` to fetch the right browser drivers, but does not use the proxy config to do so. ### The log given by the failure. <!-- Normally this include a stack trace and some more information. --> ``` $ ng e2e --proxy-config proxy.conf.js ** NG Live Development Server is running on http://localhost:49152 ** Hash: 95e0fe13c75ed4802b8b Time: 17998ms chunk {0} 0.chunk.js, 0.chunk.js.map 491 kB {3} [rendered] chunk {1} polyfills.bundle.js, polyfills.bundle.js.map (polyfills) 237 kB {5} [initial] [rendered] chunk {2} styles.bundle.js, styles.bundle.js.map (styles) 226 kB {5} [initial] [rendered] chunk {3} main.bundle.js, main.bundle.js.map (main) 8.25 kB {4} [initial] [rendered] chunk {4} vendor.bundle.js, vendor.bundle.js.map (vendor) 3.12 MB [initial] [rendered] chunk {5} inline.bundle.js, inline.bundle.js.map (inline) 0 bytes [entry] [rendered] webpack: Compiled successfully. events.js:160 throw er; // Unhandled 'error' event ^ Error: connect ECONNREFUSED 172.217.8.16:443 at Object.exports._errnoException (util.js:1018:11) at exports._exceptionWithHostPort (util.js:1041:20) at TCPConnectWrap.afterConnect [as oncomplete] (net.js:1090:14) ``` ### Desired functionality. <!-- What would like to see implemented? What is the usecase? --> `ng e2e` should pass along proxy configuration to webdriver-manager when the `--webdriver-update` flag is set to true. For webdriver-manager, the flag to use is `--proxy`. ### Mention any other details that might be useful. <!-- Please include a link to the repo if this is related to an OSS project. --> The code that calls webdriver-manager is at Line 65 in [e2e.ts](https://github.com/angular/angular-cli/blob/master/packages/%40angular/cli/tasks/e2e.ts#L65).
non_priority
ng does not use proxy config for webdriver manager update if you don t fill out the following information your issue might be closed without investigating bug report or feature request mark with an x bug report please search issues before submitting feature request versions output from ng version if nothing output from node version and npm version windows linux incl distribution macos el capitan sierra angular cli node os angular common angular compiler angular core angular forms angular http angular platform browser angular platform browser dynamic angular router angular cli angular compiler cli repro steps simple steps to reproduce this bug please include commands run packages added related code changes a link to a sample repo would help too must have an out of date web driver or delete it from node modules protractor node modules webdriver manager selenium for example if using chrome webdriver delete all chrome files in this selenium folder must be behind a proxy but without https proxy or https proxy environment variables set using the proxy guide described in create a proxy conf js file to pass to ng run ng proxy config proxy conf js if you are behind a proxy and https proxy is not set the command will fail after compiling the app this is because by default ng will run webdriver manager update to fetch the right browser drivers but does not use the proxy config to do so the log given by the failure ng proxy config proxy conf js ng live development server is running on hash time chunk chunk js chunk js map kb chunk polyfills bundle js polyfills bundle js map polyfills kb chunk styles bundle js styles bundle js map styles kb chunk main bundle js main bundle js map main kb chunk vendor bundle js vendor bundle js map vendor mb chunk inline bundle js inline bundle js map inline bytes webpack compiled successfully events js throw er unhandled error event error connect econnrefused at object exports errnoexception util js at exports exceptionwithhostport util js at tcpconnectwrap afterconnect net js desired functionality what would like to see implemented what is the usecase ng should pass along proxy configuration to webdriver manager when the webdriver update flag is set to true for webdriver manager the flag to use is proxy mention any other details that might be useful the code that calls webdriver manager is at line in
0
31,044
2,731,323,356
IssuesEvent
2015-04-16 19:41:20
Theano/Theano
https://api.github.com/repos/Theano/Theano
opened
Sparse-aware addition for combining gradients
Low Priority Sparse
Reported in https://groups.google.com/d/topic/theano-users/aa-Ydpy6_0A/discussion During gradient computation, when summing the contribution of different gradient path for the same variable, `tensor.add` can get called with one sparse and one dense input, which lead to a crash. We could either: - have an Op for that, that then gets optimized into sparse/sparse or dense/dense addition, or optimized away (if one of the inputs is all zeros, for instance) - explicitly cast the sparse matrix into a dense one inside `theano.grad`, before the addition. I don't think we should fear a memory blow-up here, since we already have a dense matrix of the same dimension.
1.0
Sparse-aware addition for combining gradients - Reported in https://groups.google.com/d/topic/theano-users/aa-Ydpy6_0A/discussion During gradient computation, when summing the contribution of different gradient path for the same variable, `tensor.add` can get called with one sparse and one dense input, which lead to a crash. We could either: - have an Op for that, that then gets optimized into sparse/sparse or dense/dense addition, or optimized away (if one of the inputs is all zeros, for instance) - explicitly cast the sparse matrix into a dense one inside `theano.grad`, before the addition. I don't think we should fear a memory blow-up here, since we already have a dense matrix of the same dimension.
priority
sparse aware addition for combining gradients reported in during gradient computation when summing the contribution of different gradient path for the same variable tensor add can get called with one sparse and one dense input which lead to a crash we could either have an op for that that then gets optimized into sparse sparse or dense dense addition or optimized away if one of the inputs is all zeros for instance explicitly cast the sparse matrix into a dense one inside theano grad before the addition i don t think we should fear a memory blow up here since we already have a dense matrix of the same dimension
1
56,283
23,743,020,559
IssuesEvent
2022-08-31 13:55:01
miranda-ng/miranda-ng
https://api.github.com/repos/miranda-ng/miranda-ng
closed
VoiceService: некоторые надписи в окне вызова не переводятся
bug VoiceService
ERROR: type should be string, got "https://github.com/miranda-ng/miranda-ng/blob/master/plugins/VoiceService/src/VoiceCall.cpp#L222\r\n\r\n```\r\n\tcase VOICE_STATE_RINGING:\r\n\t\tincoming = true;\r\n\t\tSetCaption(L\"Incoming call\");\r\n\t\tm_btnAnswer.Enable(true);\r\n\t\tm_lblStatus.SetText(L\"Ringing\");\r\n\t\tSetWindowPos(GetHwnd(), HWND_TOPMOST, 0, 0, 0, 0, SWP_NOMOVE | SWP_NOSIZE | SWP_SHOWWINDOW);\r\n\t\tSetWindowPos(GetHwnd(), HWND_NOTOPMOST, 0, 0, 0, 0, SWP_NOMOVE | SWP_NOSIZE | SWP_SHOWWINDOW);\r\n\t\tbreak;\r\n\tcase VOICE_STATE_CALLING:\r\n\t\tincoming = false;\r\n\t\tSetCaption(L\"Outgoing call\");\r\n\t\tm_lblStatus.SetText(L\"Calling\");\r\n\t\tm_btnAnswer.Enable(false);\r\n\t\tbreak;\r\n\tcase VOICE_STATE_ON_HOLD:\r\n\t\tm_lblStatus.SetText(L\"Holded\");\r\n\t\tm_btnAnswer.Enable(true);\r\n\t\tm_btnAnswer.SetText(L\"Unhold\");\r\n\t\tbreak;\r\n\tcase VOICE_STATE_ENDED:\r\n\t\tm_calltimer.Stop();\r\n\t\tmir_snwprintf(text, _countof(text), L\"Call ended %s\", m_lblStatus.GetText());\r\n\t\tm_lblStatus.SetText(text);\r\n\t\tm_btnAnswer.Enable(false);\r\n\t\tm_btnDrop.SetText(L\"Close\");\r\n\t\tbreak;\r\n\tcase VOICE_STATE_BUSY:\r\n\t\tm_lblStatus.SetText(L\"Busy\");\r\n\t\tm_btnAnswer.Enable(false);\r\n\t\tm_btnDrop.SetText(L\"Close\");\r\n\t\tbreak;\r\n\tdefault:\r\n\t\tm_lblStatus.SetText(L\"Unknown state\");\r\n\t\tbreak;\r\n\t}\r\n```"
1.0
VoiceService: некоторые надписи в окне вызова не переводятся - https://github.com/miranda-ng/miranda-ng/blob/master/plugins/VoiceService/src/VoiceCall.cpp#L222 ``` case VOICE_STATE_RINGING: incoming = true; SetCaption(L"Incoming call"); m_btnAnswer.Enable(true); m_lblStatus.SetText(L"Ringing"); SetWindowPos(GetHwnd(), HWND_TOPMOST, 0, 0, 0, 0, SWP_NOMOVE | SWP_NOSIZE | SWP_SHOWWINDOW); SetWindowPos(GetHwnd(), HWND_NOTOPMOST, 0, 0, 0, 0, SWP_NOMOVE | SWP_NOSIZE | SWP_SHOWWINDOW); break; case VOICE_STATE_CALLING: incoming = false; SetCaption(L"Outgoing call"); m_lblStatus.SetText(L"Calling"); m_btnAnswer.Enable(false); break; case VOICE_STATE_ON_HOLD: m_lblStatus.SetText(L"Holded"); m_btnAnswer.Enable(true); m_btnAnswer.SetText(L"Unhold"); break; case VOICE_STATE_ENDED: m_calltimer.Stop(); mir_snwprintf(text, _countof(text), L"Call ended %s", m_lblStatus.GetText()); m_lblStatus.SetText(text); m_btnAnswer.Enable(false); m_btnDrop.SetText(L"Close"); break; case VOICE_STATE_BUSY: m_lblStatus.SetText(L"Busy"); m_btnAnswer.Enable(false); m_btnDrop.SetText(L"Close"); break; default: m_lblStatus.SetText(L"Unknown state"); break; } ```
non_priority
voiceservice некоторые надписи в окне вызова не переводятся case voice state ringing incoming true setcaption l incoming call m btnanswer enable true m lblstatus settext l ringing setwindowpos gethwnd hwnd topmost swp nomove swp nosize swp showwindow setwindowpos gethwnd hwnd notopmost swp nomove swp nosize swp showwindow break case voice state calling incoming false setcaption l outgoing call m lblstatus settext l calling m btnanswer enable false break case voice state on hold m lblstatus settext l holded m btnanswer enable true m btnanswer settext l unhold break case voice state ended m calltimer stop mir snwprintf text countof text l call ended s m lblstatus gettext m lblstatus settext text m btnanswer enable false m btndrop settext l close break case voice state busy m lblstatus settext l busy m btnanswer enable false m btndrop settext l close break default m lblstatus settext l unknown state break
0
313,548
9,564,542,184
IssuesEvent
2019-05-05 04:40:06
sevenzk/SJKTCI
https://api.github.com/repos/sevenzk/SJKTCI
closed
[Master Harga] Pembulatan 1000 ke atas atau ke bawah di Harga Setelah Pajak
Priority enhancement fixed
Ada pembulatan 1000 ke atas atau ke bawah (yang mendekati) di Harga Setelah Pajak seperti gambar di bawah: ![image](https://user-images.githubusercontent.com/44515909/54066707-6f6d2780-4267-11e9-94cf-e7aaefd315fb.png)
1.0
[Master Harga] Pembulatan 1000 ke atas atau ke bawah di Harga Setelah Pajak - Ada pembulatan 1000 ke atas atau ke bawah (yang mendekati) di Harga Setelah Pajak seperti gambar di bawah: ![image](https://user-images.githubusercontent.com/44515909/54066707-6f6d2780-4267-11e9-94cf-e7aaefd315fb.png)
priority
pembulatan ke atas atau ke bawah di harga setelah pajak ada pembulatan ke atas atau ke bawah yang mendekati di harga setelah pajak seperti gambar di bawah
1
118,389
4,744,343,449
IssuesEvent
2016-10-21 00:29:54
FeraGroup/FTCVortexScoreCounter
https://api.github.com/repos/FeraGroup/FTCVortexScoreCounter
closed
Large numbers do not fit
enhancement Low Priority
Any score higher then 99 will not fit completely in the box that displays the scores. This is only an issue while looking at the score using the smaller counter.
1.0
Large numbers do not fit - Any score higher then 99 will not fit completely in the box that displays the scores. This is only an issue while looking at the score using the smaller counter.
priority
large numbers do not fit any score higher then will not fit completely in the box that displays the scores this is only an issue while looking at the score using the smaller counter
1
580,262
17,214,353,905
IssuesEvent
2021-07-19 09:33:46
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Pin the GitHub actions we use on this repo to a full length commit SHA
priority-3-normal status:blocked type:refactor
**What would you like Renovate to be able to do?** <!-- Tell us what requirements you need solving, and be sure to mention too if this is part of any "bigger" problem you're trying to solve. --> @rarkins and @viceice now that PR #10835 is merged, we can start thinking about pinning our GitHub Actions to the current full length Git commit SHA? **Did you already have any implementation ideas?** <!-- In case you've already dug into existing options or source code and have ideas, mention them here. Try to keep implementation ideas separate from *requirements* above --> <!-- Please also mention here in case this is a feature you'd be interested in writing yourself, so you can be assigned it. --> @viceice can you explain what pattern we need to follow to make things work properly? I think you can use a comment to say what "tagged version" you're at??? As an example I've grabbed the commit to which `actions/setup-node@v2.2.0` points right now. ```yml - name: Set up Node.js ${{ env.NODE_VERSION }} uses: actions/setup-node@38d90ce44d5275ad62cc48384b3d8a58c500bb5f with: node-version: ${{ env.NODE_VERSION }} cache: yarn ``` Or maybe we can add a comment which says what tag we're "following". ```yml - name: Set up Node.js ${{ env.NODE_VERSION }} uses: actions/setup-node@38d90ce44d5275ad62cc48384b3d8a58c500bb5f # setup-node: tag=v2 with: node-version: ${{ env.NODE_VERSION }} cache: yarn ``` We'll also need to update the `default.json` over on the `renovate/.github` repository, to make use of the new feature.
1.0
Pin the GitHub actions we use on this repo to a full length commit SHA - **What would you like Renovate to be able to do?** <!-- Tell us what requirements you need solving, and be sure to mention too if this is part of any "bigger" problem you're trying to solve. --> @rarkins and @viceice now that PR #10835 is merged, we can start thinking about pinning our GitHub Actions to the current full length Git commit SHA? **Did you already have any implementation ideas?** <!-- In case you've already dug into existing options or source code and have ideas, mention them here. Try to keep implementation ideas separate from *requirements* above --> <!-- Please also mention here in case this is a feature you'd be interested in writing yourself, so you can be assigned it. --> @viceice can you explain what pattern we need to follow to make things work properly? I think you can use a comment to say what "tagged version" you're at??? As an example I've grabbed the commit to which `actions/setup-node@v2.2.0` points right now. ```yml - name: Set up Node.js ${{ env.NODE_VERSION }} uses: actions/setup-node@38d90ce44d5275ad62cc48384b3d8a58c500bb5f with: node-version: ${{ env.NODE_VERSION }} cache: yarn ``` Or maybe we can add a comment which says what tag we're "following". ```yml - name: Set up Node.js ${{ env.NODE_VERSION }} uses: actions/setup-node@38d90ce44d5275ad62cc48384b3d8a58c500bb5f # setup-node: tag=v2 with: node-version: ${{ env.NODE_VERSION }} cache: yarn ``` We'll also need to update the `default.json` over on the `renovate/.github` repository, to make use of the new feature.
priority
pin the github actions we use on this repo to a full length commit sha what would you like renovate to be able to do rarkins and viceice now that pr is merged we can start thinking about pinning our github actions to the current full length git commit sha did you already have any implementation ideas viceice can you explain what pattern we need to follow to make things work properly i think you can use a comment to say what tagged version you re at as an example i ve grabbed the commit to which actions setup node points right now yml name set up node js env node version uses actions setup node with node version env node version cache yarn or maybe we can add a comment which says what tag we re following yml name set up node js env node version uses actions setup node setup node tag with node version env node version cache yarn we ll also need to update the default json over on the renovate github repository to make use of the new feature
1
716,982
24,656,113,350
IssuesEvent
2022-10-17 23:47:14
ApplETS/Notre-Dame
https://api.github.com/repos/ApplETS/Notre-Dame
closed
Golden files not updated - CI tests failing
bug CI priority: high
**Describe the bug** Golden files have not been updated correctly in a recent PR and CI tests are failing. The files that are throwing the error are `gradesDetailsView_1.png`, `gradesDetailsView_2.png`, `gradesDetailsView_evaluation_not_completed.png` (located in `goldenFiles/`) **To Reproduce** Steps to reproduce the behavior: 1. Run CI tests on any valid PR 2. Golden files will throw errors and the tests will fail **Expected behavior** Golden files tests should not fail.
1.0
Golden files not updated - CI tests failing - **Describe the bug** Golden files have not been updated correctly in a recent PR and CI tests are failing. The files that are throwing the error are `gradesDetailsView_1.png`, `gradesDetailsView_2.png`, `gradesDetailsView_evaluation_not_completed.png` (located in `goldenFiles/`) **To Reproduce** Steps to reproduce the behavior: 1. Run CI tests on any valid PR 2. Golden files will throw errors and the tests will fail **Expected behavior** Golden files tests should not fail.
priority
golden files not updated ci tests failing describe the bug golden files have not been updated correctly in a recent pr and ci tests are failing the files that are throwing the error are gradesdetailsview png gradesdetailsview png gradesdetailsview evaluation not completed png located in goldenfiles to reproduce steps to reproduce the behavior run ci tests on any valid pr golden files will throw errors and the tests will fail expected behavior golden files tests should not fail
1
269,463
8,435,892,153
IssuesEvent
2018-10-17 14:12:52
smartdevicelink/sdl_core
https://api.github.com/repos/smartdevicelink/sdl_core
closed
Adjust code to accomodate new JsonCPP version
Bug Contributor priority 1: High
### Bug Report Adjust code to accomodate new JsonCPP version ##### Expected Behavior Need to upgrade the third-party JsonCpp library. The SDL library is currently using an old release candidate version of JsonCpp (0.6.0-rc2). This should be updated to an actually released version. ##### OS & Version Information * OS/Version: * SDL Core Version: * Testing Against:
1.0
Adjust code to accomodate new JsonCPP version - ### Bug Report Adjust code to accomodate new JsonCPP version ##### Expected Behavior Need to upgrade the third-party JsonCpp library. The SDL library is currently using an old release candidate version of JsonCpp (0.6.0-rc2). This should be updated to an actually released version. ##### OS & Version Information * OS/Version: * SDL Core Version: * Testing Against:
priority
adjust code to accomodate new jsoncpp version bug report adjust code to accomodate new jsoncpp version expected behavior need to upgrade the third party jsoncpp library the sdl library is currently using an old release candidate version of jsoncpp this should be updated to an actually released version os version information os version sdl core version testing against
1
455,448
13,127,063,960
IssuesEvent
2020-08-06 09:40:16
phovea/generator-phovea
https://api.github.com/repos/phovea/generator-phovea
opened
Update build.js after moving deploy scripts from app to product
priority: high type: bug
* Release number or git hash: 2feaac301b3ccbaffb9af5bbac124be3832218b1 * OS: Linux * Environment (local or deployed): both ### Steps to reproduce 1. build a product containing more deployment configurations than web and api ### Observed behavior * only web and api are accepted options for Dockerfiles (seee https://github.com/phovea/generator-phovea/blob/master/generators/init-product/templates/plain/build.js#L626) * additional folders like dtiles or benchmark are ignored ### Expected behavior * additional Dockerfiles are copied as well as long as they are available in templates/<name>/deploy/<name>
1.0
Update build.js after moving deploy scripts from app to product - * Release number or git hash: 2feaac301b3ccbaffb9af5bbac124be3832218b1 * OS: Linux * Environment (local or deployed): both ### Steps to reproduce 1. build a product containing more deployment configurations than web and api ### Observed behavior * only web and api are accepted options for Dockerfiles (seee https://github.com/phovea/generator-phovea/blob/master/generators/init-product/templates/plain/build.js#L626) * additional folders like dtiles or benchmark are ignored ### Expected behavior * additional Dockerfiles are copied as well as long as they are available in templates/<name>/deploy/<name>
priority
update build js after moving deploy scripts from app to product release number or git hash os linux environment local or deployed both steps to reproduce build a product containing more deployment configurations than web and api observed behavior only web and api are accepted options for dockerfiles seee additional folders like dtiles or benchmark are ignored expected behavior additional dockerfiles are copied as well as long as they are available in templates deploy
1
450,454
31,925,465,711
IssuesEvent
2023-09-19 01:10:23
ICEI-PUC-Minas-PMV-ADS/pmv-ads-2023-2-e3-proj-mov-t1-entre-time
https://api.github.com/repos/ICEI-PUC-Minas-PMV-ADS/pmv-ads-2023-2-e3-proj-mov-t1-entre-time
closed
Justificativa e Público alvo(01- Documentação de contexto)- H11a-ADS-CST
documentation
H11a-ADS-CST - Compreender os usuários e definir uma proposta de solução: definir o problema de forma clara e objetiva, apresentando os objetivos, a justificativa e a motivação da escolha.
1.0
Justificativa e Público alvo(01- Documentação de contexto)- H11a-ADS-CST - H11a-ADS-CST - Compreender os usuários e definir uma proposta de solução: definir o problema de forma clara e objetiva, apresentando os objetivos, a justificativa e a motivação da escolha.
non_priority
justificativa e público alvo documentação de contexto ads cst ads cst compreender os usuários e definir uma proposta de solução definir o problema de forma clara e objetiva apresentando os objetivos a justificativa e a motivação da escolha
0
50,728
12,549,929,659
IssuesEvent
2020-06-06 09:07:38
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
appengine.flexible.tasks.snippets_test: test_pause_queue failed
buildcop: issue priority: p1 type: bug
This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: cc68a07af4cab7b48233680996d2913fb0ba5e24 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/3ddf5a57-41f5-46d8-812c-bc99c4adc730), [Sponge](http://sponge2/3ddf5a57-41f5-46d8-812c-bc99c4adc730) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 57, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 826, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 729, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.NOT_FOUND details = "Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize." debug_error_string = "{"created":"@1591387480.150681677","description":"Error received from peer ipv4:172.217.212.95:443","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize.","grpc_status":5}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/appengine/flexible/tasks/snippets_test.py", line 94, in test_pause_queue TEST_PROJECT_ID, TEST_LOCATION, QUEUE_NAME_1) File "/workspace/appengine/flexible/tasks/snippets.py", line 218, in pause_queue response = client.pause_queue(queue_path) File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/tasks_v2/gapic/cloud_tasks_client.py", line 832, in pause_queue request, retry=retry, timeout=timeout, metadata=metadata File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 143, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 286, in retry_wrapped_func on_error=on_error, File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 184, in retry_target return target() File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/timeout.py", line 214, in func_with_timeout return func(*args, **kwargs) File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 59, in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) File "<string>", line 3, in raise_from google.api_core.exceptions.NotFound: 404 Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize.</pre></details>
1.0
appengine.flexible.tasks.snippets_test: test_pause_queue failed - This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: cc68a07af4cab7b48233680996d2913fb0ba5e24 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/3ddf5a57-41f5-46d8-812c-bc99c4adc730), [Sponge](http://sponge2/3ddf5a57-41f5-46d8-812c-bc99c4adc730) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 57, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 826, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 729, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.NOT_FOUND details = "Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize." debug_error_string = "{"created":"@1591387480.150681677","description":"Error received from peer ipv4:172.217.212.95:443","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize.","grpc_status":5}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/appengine/flexible/tasks/snippets_test.py", line 94, in test_pause_queue TEST_PROJECT_ID, TEST_LOCATION, QUEUE_NAME_1) File "/workspace/appengine/flexible/tasks/snippets.py", line 218, in pause_queue response = client.pause_queue(queue_path) File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/tasks_v2/gapic/cloud_tasks_client.py", line 832, in pause_queue request, retry=retry, timeout=timeout, metadata=metadata File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 143, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 286, in retry_wrapped_func on_error=on_error, File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 184, in retry_target return target() File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/timeout.py", line 214, in func_with_timeout return func(*args, **kwargs) File "/workspace/appengine/flexible/tasks/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 59, in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) File "<string>", line 3, in raise_from google.api_core.exceptions.NotFound: 404 Queue does not exist. If you just created the queue, wait at least a minute for the queue to initialize.</pre></details>
non_priority
appengine flexible tasks snippets test test pause queue failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace appengine flexible tasks nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace appengine flexible tasks nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace appengine flexible tasks nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode not found details queue does not exist if you just created the queue wait at least a minute for the queue to initialize debug error string created description error received from peer file src core lib surface call cc file line grpc message queue does not exist if you just created the queue wait at least a minute for the queue to initialize grpc status the above exception was the direct cause of the following exception traceback most recent call last file workspace appengine flexible tasks snippets test py line in test pause queue test project id test location queue name file workspace appengine flexible tasks snippets py line in pause queue response client pause queue queue path file workspace appengine flexible tasks nox py lib site packages google cloud tasks gapic cloud tasks client py line in pause queue request retry retry timeout timeout metadata metadata file workspace appengine flexible tasks nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace appengine flexible tasks nox py lib site packages google api core retry py line in retry wrapped func on error on error file workspace appengine flexible tasks nox py lib site packages google api core retry py line in retry target return target file workspace appengine flexible tasks nox py lib site packages google api core timeout py line in func with timeout return func args kwargs file workspace appengine flexible tasks nox py lib site packages google api core grpc helpers py line in error remapped callable six raise from exceptions from grpc error exc exc file line in raise from google api core exceptions notfound queue does not exist if you just created the queue wait at least a minute for the queue to initialize
0
786,415
27,645,769,221
IssuesEvent
2023-03-10 22:49:58
briandfoy/cpan-audit
https://api.github.com/repos/briandfoy/cpan-audit
closed
Feature: example drop-in cpan-audit.t file
Type: enhancement Priority: low
I think it would be great to offer a drop-in cpan-audit.t file for users to drop into their t/ directory, which does an audit and reports any security advisories on whatever is in cpanfile.snapshot, for example.
1.0
Feature: example drop-in cpan-audit.t file - I think it would be great to offer a drop-in cpan-audit.t file for users to drop into their t/ directory, which does an audit and reports any security advisories on whatever is in cpanfile.snapshot, for example.
priority
feature example drop in cpan audit t file i think it would be great to offer a drop in cpan audit t file for users to drop into their t directory which does an audit and reports any security advisories on whatever is in cpanfile snapshot for example
1
74,821
14,346,379,831
IssuesEvent
2020-11-29 00:09:40
Arcanorum/dungeonz
https://api.github.com/repos/Arcanorum/dungeonz
opened
Spatial audio
code
**Task description:** Need to figure out and implement a way to associate playing certain sounds with an entity on screen, and adjust the volume to be appropriate for the distance from the player. **References/notes:** Some discussion around the topic. https://phaser.discourse.group/t/sound-in-particular-place/2547/2 **Acceptance criteria:** AC: A given sound can be associated with an entity on screen. AC: If the player moves towards/away from that entity, the volume of that sound will increase/decrease based on distance. AC: If the entity moves towards/away from the player, the volume of that sound will increase/decrease based on distance.
1.0
Spatial audio - **Task description:** Need to figure out and implement a way to associate playing certain sounds with an entity on screen, and adjust the volume to be appropriate for the distance from the player. **References/notes:** Some discussion around the topic. https://phaser.discourse.group/t/sound-in-particular-place/2547/2 **Acceptance criteria:** AC: A given sound can be associated with an entity on screen. AC: If the player moves towards/away from that entity, the volume of that sound will increase/decrease based on distance. AC: If the entity moves towards/away from the player, the volume of that sound will increase/decrease based on distance.
non_priority
spatial audio task description need to figure out and implement a way to associate playing certain sounds with an entity on screen and adjust the volume to be appropriate for the distance from the player references notes some discussion around the topic acceptance criteria ac a given sound can be associated with an entity on screen ac if the player moves towards away from that entity the volume of that sound will increase decrease based on distance ac if the entity moves towards away from the player the volume of that sound will increase decrease based on distance
0
640,842
20,810,232,190
IssuesEvent
2022-03-18 01:10:45
monarch-initiative/mondo
https://api.github.com/repos/monarch-initiative/mondo
closed
MONDO:0005755 equine infectious anemia; non-human disease [Revise subclass]
Revise subclass high priority
**Mondo term (ID and Label)** MONDO:0005755 equine infectious anemia **Suggested revision and reasons** I think this belongs under MONDO:0005583 "non-human animal disease" NCI definition from UMLS CUI C0014661: "A horse disease caused by a retrovirus which is transmitted by biting flies. The acute phase symptoms include high fever, anemia, weakness, and swelling of the legs. The subacute phase symptoms include splenomegaly, anemia, and weight loss. The chronic phase symptoms include recurrent fevers and anemia. (NCI)" **Your nano-attribution (ORCID)** If you don't have an ORCID, you can sign up for one [here](https://orcid.org/)
1.0
MONDO:0005755 equine infectious anemia; non-human disease [Revise subclass] - **Mondo term (ID and Label)** MONDO:0005755 equine infectious anemia **Suggested revision and reasons** I think this belongs under MONDO:0005583 "non-human animal disease" NCI definition from UMLS CUI C0014661: "A horse disease caused by a retrovirus which is transmitted by biting flies. The acute phase symptoms include high fever, anemia, weakness, and swelling of the legs. The subacute phase symptoms include splenomegaly, anemia, and weight loss. The chronic phase symptoms include recurrent fevers and anemia. (NCI)" **Your nano-attribution (ORCID)** If you don't have an ORCID, you can sign up for one [here](https://orcid.org/)
priority
mondo equine infectious anemia non human disease mondo term id and label mondo equine infectious anemia suggested revision and reasons i think this belongs under mondo non human animal disease nci definition from umls cui a horse disease caused by a retrovirus which is transmitted by biting flies the acute phase symptoms include high fever anemia weakness and swelling of the legs the subacute phase symptoms include splenomegaly anemia and weight loss the chronic phase symptoms include recurrent fevers and anemia nci your nano attribution orcid if you don t have an orcid you can sign up for one
1
49,621
3,003,711,772
IssuesEvent
2015-07-25 05:58:51
jayway/powermock
https://api.github.com/repos/jayway/powermock
opened
mockin org.apache.http.impl.client.DefaultHttpClient class
bug imported Priority-Medium
_From [daghana...@gmail.com](https://code.google.com/u/104674216580764072044/) on July 22, 2014 06:55:51_ What steps will reproduce the problem? 1. create a constructor mock of .DefaultHttpClient using Powermock_V1.5.5 Mockito_V1.9.5 and Junit_V4.1 ektorp_V1.4.1 2.run the test call new StdHttpClient.Builder().url(dburl.toString()).caching(true) .cleanupIdleConnections(true).socketTimeout(Integer.parseInt(socketTimeout)).build() What is the expected output? What do you see instead? 1- connection established using mocked HttpClient 2- SSLContext.getInstance("TLS") throws NoSuchAlgorithmException What version of the product are you using? On what operating system? OS Ubuntu 14.04 Please provide any additional information below. Test code @RunWith(PowerMockRunner.class) @PrepareForTest(DefaultHttpClient.class) public class App_UnitTest { protected static InputStream mockUrlConnectionInputStream; protected static OutputStream mockUrlConnectionOutputStream; @InjectMocks private final App appUnderTest = new App(); @Spy private final CouchConnection couchConnection = new CouchConnection(); @Before public void setupOnce() throws Exception { MockitoAnnotations.initMocks(this); couchConnection.setDbUrl("https://test.couch.url.com"); couchConnection.setDbName("testCouchDB"); couchConnection.setPort("12211"); couchConnection.setSocketTimeout("10000"); final DefaultHttpClient mockDefaultHttpClient = PowerMockito.mock(DefaultHttpClient.class); PowerMockito.whenNew(DefaultHttpClient.class).withAnyArguments().thenReturn(mockDefaultHttpClient); try { URL.setURLStreamHandlerFactory(new MockURLStreamHandler()); } catch (final Error e) { // Do nothing, this will be called by all the extending test class but should only be called once for JVM so just // ignore } } @Test public void testifJobExists() throws IOException { final HttpClient authenticatedHttpClient = new StdHttpClient.Builder().url(dburl.toString()).caching(true) .cleanupIdleConnections(true).socketTimeout(Integer.parseInt(socketTimeout)).build(); final CouchDbInstance couchInstance = new StdCouchDbInstance(authenticatedHttpClient); couchConnector = couchInstance.createConnector(dbName, true); } public class MockURLStreamHandler extends URLStreamHandler implements URLStreamHandlerFactory { // *** URLStreamHandler @Override protected URLConnection openConnection(final URL u) throws IOException { return new MockHttpsURLConnection(u); } // *** URLStreamHandlerFactory @Override public URLStreamHandler createURLStreamHandler(final String protocol) { // this will only override the chosen protocol so that file and other protocols used by the Spring and other // classes will work as usual. // WARNING this will only mock https connections, so your URL has to start with https e.g. // https:\\something.etc.com if (protocol.equalsIgnoreCase("https")) { return this; } else { return null; } } } public class MockHttpsURLConnection extends HttpsURLConnection { protected MockHttpsURLConnection(final URL url) { super(url); } // *** HttpURLConnection @Override public InputStream getInputStream() throws IOException { return mockUrlConnectionInputStream; } @Override public void connect() throws IOException { } @Override public void disconnect() { } @Override public boolean usingProxy() { return false; } @Override public String getCipherSuite() { return null; } @Override public Certificate[] getLocalCertificates() { return null; } @Override public Certificate[] getServerCertificates() throws SSLPeerUnverifiedException { return null; } @Override public OutputStream getOutputStream() throws IOException { return mockUrlConnectionOutputStream; } } } _Original issue: http://code.google.com/p/powermock/issues/detail?id=507_
1.0
mockin org.apache.http.impl.client.DefaultHttpClient class - _From [daghana...@gmail.com](https://code.google.com/u/104674216580764072044/) on July 22, 2014 06:55:51_ What steps will reproduce the problem? 1. create a constructor mock of .DefaultHttpClient using Powermock_V1.5.5 Mockito_V1.9.5 and Junit_V4.1 ektorp_V1.4.1 2.run the test call new StdHttpClient.Builder().url(dburl.toString()).caching(true) .cleanupIdleConnections(true).socketTimeout(Integer.parseInt(socketTimeout)).build() What is the expected output? What do you see instead? 1- connection established using mocked HttpClient 2- SSLContext.getInstance("TLS") throws NoSuchAlgorithmException What version of the product are you using? On what operating system? OS Ubuntu 14.04 Please provide any additional information below. Test code @RunWith(PowerMockRunner.class) @PrepareForTest(DefaultHttpClient.class) public class App_UnitTest { protected static InputStream mockUrlConnectionInputStream; protected static OutputStream mockUrlConnectionOutputStream; @InjectMocks private final App appUnderTest = new App(); @Spy private final CouchConnection couchConnection = new CouchConnection(); @Before public void setupOnce() throws Exception { MockitoAnnotations.initMocks(this); couchConnection.setDbUrl("https://test.couch.url.com"); couchConnection.setDbName("testCouchDB"); couchConnection.setPort("12211"); couchConnection.setSocketTimeout("10000"); final DefaultHttpClient mockDefaultHttpClient = PowerMockito.mock(DefaultHttpClient.class); PowerMockito.whenNew(DefaultHttpClient.class).withAnyArguments().thenReturn(mockDefaultHttpClient); try { URL.setURLStreamHandlerFactory(new MockURLStreamHandler()); } catch (final Error e) { // Do nothing, this will be called by all the extending test class but should only be called once for JVM so just // ignore } } @Test public void testifJobExists() throws IOException { final HttpClient authenticatedHttpClient = new StdHttpClient.Builder().url(dburl.toString()).caching(true) .cleanupIdleConnections(true).socketTimeout(Integer.parseInt(socketTimeout)).build(); final CouchDbInstance couchInstance = new StdCouchDbInstance(authenticatedHttpClient); couchConnector = couchInstance.createConnector(dbName, true); } public class MockURLStreamHandler extends URLStreamHandler implements URLStreamHandlerFactory { // *** URLStreamHandler @Override protected URLConnection openConnection(final URL u) throws IOException { return new MockHttpsURLConnection(u); } // *** URLStreamHandlerFactory @Override public URLStreamHandler createURLStreamHandler(final String protocol) { // this will only override the chosen protocol so that file and other protocols used by the Spring and other // classes will work as usual. // WARNING this will only mock https connections, so your URL has to start with https e.g. // https:\\something.etc.com if (protocol.equalsIgnoreCase("https")) { return this; } else { return null; } } } public class MockHttpsURLConnection extends HttpsURLConnection { protected MockHttpsURLConnection(final URL url) { super(url); } // *** HttpURLConnection @Override public InputStream getInputStream() throws IOException { return mockUrlConnectionInputStream; } @Override public void connect() throws IOException { } @Override public void disconnect() { } @Override public boolean usingProxy() { return false; } @Override public String getCipherSuite() { return null; } @Override public Certificate[] getLocalCertificates() { return null; } @Override public Certificate[] getServerCertificates() throws SSLPeerUnverifiedException { return null; } @Override public OutputStream getOutputStream() throws IOException { return mockUrlConnectionOutputStream; } } } _Original issue: http://code.google.com/p/powermock/issues/detail?id=507_
priority
mockin org apache http impl client defaulthttpclient class from on july what steps will reproduce the problem create a constructor mock of defaulthttpclient using powermock mockito and junit ektorp run the test call new stdhttpclient builder url dburl tostring caching true cleanupidleconnections true sockettimeout integer parseint sockettimeout build what is the expected output what do you see instead connection established using mocked httpclient sslcontext getinstance tls throws nosuchalgorithmexception what version of the product are you using on what operating system os ubuntu please provide any additional information below test code runwith powermockrunner class preparefortest defaulthttpclient class public class app unittest protected static inputstream mockurlconnectioninputstream protected static outputstream mockurlconnectionoutputstream injectmocks private final app appundertest new app spy private final couchconnection couchconnection new couchconnection before public void setuponce throws exception mockitoannotations initmocks this couchconnection setdburl couchconnection setdbname testcouchdb couchconnection setport couchconnection setsockettimeout final defaulthttpclient mockdefaulthttpclient powermockito mock defaulthttpclient class powermockito whennew defaulthttpclient class withanyarguments thenreturn mockdefaulthttpclient try url seturlstreamhandlerfactory new mockurlstreamhandler catch final error e do nothing this will be called by all the extending test class but should only be called once for jvm so just ignore test public void testifjobexists throws ioexception final httpclient authenticatedhttpclient new stdhttpclient builder url dburl tostring caching true cleanupidleconnections true sockettimeout integer parseint sockettimeout build final couchdbinstance couchinstance new stdcouchdbinstance authenticatedhttpclient couchconnector couchinstance createconnector dbname true public class mockurlstreamhandler extends urlstreamhandler implements urlstreamhandlerfactory urlstreamhandler override protected urlconnection openconnection final url u throws ioexception return new mockhttpsurlconnection u urlstreamhandlerfactory override public urlstreamhandler createurlstreamhandler final string protocol this will only override the chosen protocol so that file and other protocols used by the spring and other classes will work as usual warning this will only mock https connections so your url has to start with https e g https something etc com if protocol equalsignorecase https return this else return null public class mockhttpsurlconnection extends httpsurlconnection protected mockhttpsurlconnection final url url super url httpurlconnection override public inputstream getinputstream throws ioexception return mockurlconnectioninputstream override public void connect throws ioexception override public void disconnect override public boolean usingproxy return false override public string getciphersuite return null override public certificate getlocalcertificates return null override public certificate getservercertificates throws sslpeerunverifiedexception return null override public outputstream getoutputstream throws ioexception return mockurlconnectionoutputstream original issue
1
43,168
5,529,972,604
IssuesEvent
2017-03-21 00:30:20
easydigitaldownloads/easy-digital-downloads
https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads
closed
Multiple EDD_Payments_Query's affect each other.
Bug Has PR Needs Testing Payments
If you instantiate 2 objects of EDD_Payments_Query, because the values are hooked to edd_pre_get_payments, any custom values you set up for the first object continue to be hooked for any subsequent objects. **For example:** The following code snippet will cause the payment history page in the WordPress dashboard to show no results (notice how we don't return any values or do anything with our returned data here - in theory, because of the word "get", this shouldn't be affecting anything): ``` function mess_with_edd_get_payments() { $args = array( 'end_date' => time() ); $payments = edd_get_payments( $args ); } add_action( 'admin_init', 'mess_with_edd_get_payments' ); ``` The problem is that in the EDD_Payments_Query class, the values are hooked to edd_pre_get_payments. So when the payment history page runs this in a separately instatiated object, the hook still runs which sets the "end_date" - even though we had no intention of setting the end_date for the Payment History page. **Proposed Solution:** Instead of hooking the custom values to edd_pre_get_payments, we should simply call each method from within the get_payments method in EDD_Payments_Query.
1.0
Multiple EDD_Payments_Query's affect each other. - If you instantiate 2 objects of EDD_Payments_Query, because the values are hooked to edd_pre_get_payments, any custom values you set up for the first object continue to be hooked for any subsequent objects. **For example:** The following code snippet will cause the payment history page in the WordPress dashboard to show no results (notice how we don't return any values or do anything with our returned data here - in theory, because of the word "get", this shouldn't be affecting anything): ``` function mess_with_edd_get_payments() { $args = array( 'end_date' => time() ); $payments = edd_get_payments( $args ); } add_action( 'admin_init', 'mess_with_edd_get_payments' ); ``` The problem is that in the EDD_Payments_Query class, the values are hooked to edd_pre_get_payments. So when the payment history page runs this in a separately instatiated object, the hook still runs which sets the "end_date" - even though we had no intention of setting the end_date for the Payment History page. **Proposed Solution:** Instead of hooking the custom values to edd_pre_get_payments, we should simply call each method from within the get_payments method in EDD_Payments_Query.
non_priority
multiple edd payments query s affect each other if you instantiate objects of edd payments query because the values are hooked to edd pre get payments any custom values you set up for the first object continue to be hooked for any subsequent objects for example the following code snippet will cause the payment history page in the wordpress dashboard to show no results notice how we don t return any values or do anything with our returned data here in theory because of the word get this shouldn t be affecting anything function mess with edd get payments args array end date time payments edd get payments args add action admin init mess with edd get payments the problem is that in the edd payments query class the values are hooked to edd pre get payments so when the payment history page runs this in a separately instatiated object the hook still runs which sets the end date even though we had no intention of setting the end date for the payment history page proposed solution instead of hooking the custom values to edd pre get payments we should simply call each method from within the get payments method in edd payments query
0
47,853
7,354,063,348
IssuesEvent
2018-03-09 04:24:51
Naoghuman/lib-validation
https://api.github.com/repos/Naoghuman/lib-validation
opened
[doc] Update ReadMe.md to 0.3.0.
documentation refactoring
[doc] Update ReadMe.md to 0.3.0. * New UML image for the section `Intention`. * Dependencies, Download...
1.0
[doc] Update ReadMe.md to 0.3.0. - [doc] Update ReadMe.md to 0.3.0. * New UML image for the section `Intention`. * Dependencies, Download...
non_priority
update readme md to update readme md to new uml image for the section intention dependencies download
0
242,138
7,838,626,663
IssuesEvent
2018-06-18 10:56:19
minishift/minishift-addons
https://api.github.com/repos/minishift/minishift-addons
opened
Add scenario for removal of Che addon
kind/task priority/major
Since PR #123 for removal of Che was merged there should be also a test case for the removal to cover the functionality in the future.
1.0
Add scenario for removal of Che addon - Since PR #123 for removal of Che was merged there should be also a test case for the removal to cover the functionality in the future.
priority
add scenario for removal of che addon since pr for removal of che was merged there should be also a test case for the removal to cover the functionality in the future
1
612,926
19,059,447,170
IssuesEvent
2021-11-26 04:32:14
tomusborne/generatepress
https://api.github.com/repos/tomusborne/generatepress
opened
Add missing wp_set_script_translations() functions
type: bug priority: medium
We're missing the needed `wp_set_script_translations( 'handle', 'generatepress' )` functions wherever we're adding `wp-i18n` as a dependency right now, which is preventing translations from working.
1.0
Add missing wp_set_script_translations() functions - We're missing the needed `wp_set_script_translations( 'handle', 'generatepress' )` functions wherever we're adding `wp-i18n` as a dependency right now, which is preventing translations from working.
priority
add missing wp set script translations functions we re missing the needed wp set script translations handle generatepress functions wherever we re adding wp as a dependency right now which is preventing translations from working
1
830,349
32,003,233,796
IssuesEvent
2023-09-21 13:28:57
dag-hammarskjold-library/dlx-rest
https://api.github.com/repos/dag-hammarskjold-library/dlx-rest
closed
Display and sorting in browse indexes by subfield when not in order
type: enhancement priority: high function: search sort
It looks like the sorting in the browse indexes is not by the alphabetical order of the subfields in the record, but rather by the order in which they display in the field? Here is an example: ![image](https://github.com/dag-hammarskjold-library/dlx-rest/assets/110108422/a897c3ca-b69a-4c97-af22-bd06e5eda7e6) This is what the second record in the list looks like - the subfields in 191 are not in order, but the index seems to read them as they are in the record and thus puts it in the wrong place in the browse index: ![image](https://github.com/dag-hammarskjold-library/dlx-rest/assets/110108422/df3cd60b-a171-4a4c-86c9-89582bc68f8b) It should appear here: ![image](https://github.com/dag-hammarskjold-library/dlx-rest/assets/110108422/5dc2a8a1-5f90-40ef-b0e6-d2b857da57ff)
1.0
Display and sorting in browse indexes by subfield when not in order - It looks like the sorting in the browse indexes is not by the alphabetical order of the subfields in the record, but rather by the order in which they display in the field? Here is an example: ![image](https://github.com/dag-hammarskjold-library/dlx-rest/assets/110108422/a897c3ca-b69a-4c97-af22-bd06e5eda7e6) This is what the second record in the list looks like - the subfields in 191 are not in order, but the index seems to read them as they are in the record and thus puts it in the wrong place in the browse index: ![image](https://github.com/dag-hammarskjold-library/dlx-rest/assets/110108422/df3cd60b-a171-4a4c-86c9-89582bc68f8b) It should appear here: ![image](https://github.com/dag-hammarskjold-library/dlx-rest/assets/110108422/5dc2a8a1-5f90-40ef-b0e6-d2b857da57ff)
priority
display and sorting in browse indexes by subfield when not in order it looks like the sorting in the browse indexes is not by the alphabetical order of the subfields in the record but rather by the order in which they display in the field here is an example this is what the second record in the list looks like the subfields in are not in order but the index seems to read them as they are in the record and thus puts it in the wrong place in the browse index it should appear here
1
452,720
13,058,468,105
IssuesEvent
2020-07-30 09:04:12
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
Create standard/curved shader with tinting
Category: Tech Priority: Low
Copy Curved/Avatar, could possibly just replace it with generic version. Would allow two colours, and blending between them. Could be extended to support more colours with less blending accuracy. Can be used for: * Avatar skin/clothing colour * Colouring vehicles * Colouring world objects (good use would be civics objects with nation colours)
1.0
Create standard/curved shader with tinting - Copy Curved/Avatar, could possibly just replace it with generic version. Would allow two colours, and blending between them. Could be extended to support more colours with less blending accuracy. Can be used for: * Avatar skin/clothing colour * Colouring vehicles * Colouring world objects (good use would be civics objects with nation colours)
priority
create standard curved shader with tinting copy curved avatar could possibly just replace it with generic version would allow two colours and blending between them could be extended to support more colours with less blending accuracy can be used for avatar skin clothing colour colouring vehicles colouring world objects good use would be civics objects with nation colours
1
335,417
10,153,354,408
IssuesEvent
2019-08-06 04:07:06
gitcoinco/web
https://api.github.com/repos/gitcoinco/web
opened
Team Payout -Build - Milestone I - As a funder on Gitcion - I'd like the ability to manage permissions and payouts with my team members.
8 priority: backlog sprint 16
<!-- Hello Gitcoiner! Please use the template below for issue ideas or bugs found within Gitcoin. If it is general support you need, reach out to us at gitcoin.co/slack Provide a general summary of the issue in the title above and use relevant fields below to define the problem. --> ### User Story As a funder on Gitcion - I'd like the ability to manage permissions and payouts with my team members. ### Assumptions ### What - The Gitcoin platform syncs repo data from Github related to collaborators and team members, mirroring their permissions model. TBD: what level of access will allow users to be added as "Issuers", who have full control, and "Approvers", who can accept submissions and issue payout. ### Why Funders have requested this!
1.0
Team Payout -Build - Milestone I - As a funder on Gitcion - I'd like the ability to manage permissions and payouts with my team members. - <!-- Hello Gitcoiner! Please use the template below for issue ideas or bugs found within Gitcoin. If it is general support you need, reach out to us at gitcoin.co/slack Provide a general summary of the issue in the title above and use relevant fields below to define the problem. --> ### User Story As a funder on Gitcion - I'd like the ability to manage permissions and payouts with my team members. ### Assumptions ### What - The Gitcoin platform syncs repo data from Github related to collaborators and team members, mirroring their permissions model. TBD: what level of access will allow users to be added as "Issuers", who have full control, and "Approvers", who can accept submissions and issue payout. ### Why Funders have requested this!
priority
team payout build milestone i as a funder on gitcion i d like the ability to manage permissions and payouts with my team members hello gitcoiner please use the template below for issue ideas or bugs found within gitcoin if it is general support you need reach out to us at gitcoin co slack provide a general summary of the issue in the title above and use relevant fields below to define the problem user story as a funder on gitcion i d like the ability to manage permissions and payouts with my team members assumptions what the gitcoin platform syncs repo data from github related to collaborators and team members mirroring their permissions model tbd what level of access will allow users to be added as issuers who have full control and approvers who can accept submissions and issue payout why funders have requested this
1
703,234
24,150,477,623
IssuesEvent
2022-09-21 23:47:45
googleapis/nodejs-ai-platform
https://api.github.com/repos/googleapis/nodejs-ai-platform
closed
AI platform get hyperparameter tuning job: should get the specified hyperparameter tuning job failed
type: bug priority: p1 flakybot: issue api: vertex-ai
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: e1c5cd6b5d03afb03911ba9aa685457aa359a602 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/69dc3b23-298d-461c-9279-cd7551a90fc9), [Sponge](http://sponge2/69dc3b23-298d-461c-9279-cd7551a90fc9) status: failed <details><summary>Test output</summary><br><pre>Command failed: node ./get-hyperparameter-tuning-job.js 2216298782247616512 ucaip-sample-tests us-central1 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. Error: Command failed: node ./get-hyperparameter-tuning-job.js 2216298782247616512 ucaip-sample-tests us-central1 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at checkExecSyncError (child_process.js:635:11) at Object.execSync (child_process.js:671:15) at execSync (test/get-hyperparameter-tuning-job.test.js:23:28) at Context.<anonymous> (test/get-hyperparameter-tuning-job.test.js:31:20) at processImmediate (internal/timers.js:461:21)</pre></details>
1.0
AI platform get hyperparameter tuning job: should get the specified hyperparameter tuning job failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: e1c5cd6b5d03afb03911ba9aa685457aa359a602 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/69dc3b23-298d-461c-9279-cd7551a90fc9), [Sponge](http://sponge2/69dc3b23-298d-461c-9279-cd7551a90fc9) status: failed <details><summary>Test output</summary><br><pre>Command failed: node ./get-hyperparameter-tuning-job.js 2216298782247616512 ucaip-sample-tests us-central1 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. Error: Command failed: node ./get-hyperparameter-tuning-job.js 2216298782247616512 ucaip-sample-tests us-central1 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at checkExecSyncError (child_process.js:635:11) at Object.execSync (child_process.js:671:15) at execSync (test/get-hyperparameter-tuning-job.test.js:23:28) at Context.<anonymous> (test/get-hyperparameter-tuning-job.test.js:31:20) at processImmediate (internal/timers.js:461:21)</pre></details>
priority
ai platform get hyperparameter tuning job should get the specified hyperparameter tuning job failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output command failed node get hyperparameter tuning job js ucaip sample tests us unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see error command failed node get hyperparameter tuning job js ucaip sample tests us unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at checkexecsyncerror child process js at object execsync child process js at execsync test get hyperparameter tuning job test js at context test get hyperparameter tuning job test js at processimmediate internal timers js
1
73,471
7,335,034,949
IssuesEvent
2018-03-06 01:44:21
istio/istio
https://api.github.com/repos/istio/istio
closed
Test Failure: security/tests/integration/certificateRotationTest
kind/fixit kind/test-failure
From: https://k8s-gubernator.appspot.com/build/istio-prow/pull/istio_istio/3913/istio-presubmit/6070/ ``` I0302 22:34:20.343] 2018-03-02T22:34:20.341870Z error failed to create test namespace: failed to create a role (error: failed to create role (error: roles.rbac.authorization.k8s.io "istio-ca-role" is forbidden: attempt to grant extra privileges: [PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["create"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["update"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["create"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["update"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["list"]}] user=&{istio-prow-test-job@istio-testing.iam.gserviceaccount.com [system:authenticated] map[authenticator:[GKE]]} ownerrules=[PolicyRule{Resources:["selfsubjectaccessreviews" "selfsubjectrulesreviews"], APIGroups:["authorization.k8s.io"], Verbs:["create"]} PolicyRule{NonResourceURLs:["/api" "/api/*" "/apis" "/apis/*" "/healthz" "/swagger-2.0.0.pb-v1" "/swagger.json" "/swaggerapi" "/swaggerapi/*" "/version"], Verbs:["get"]}] ruleResolutionErrors=[])) I0302 22:34:20.344] 2018-03-02T22:34:20.341877Z error test environment creation failure I0302 22:34:20.344] FAIL istio.io/istio/security/tests/integration/certificateRotationTest 2.535s E0302 22:34:20.355] Build failed I0302 22:34:20.355] process 509 exited with code 1 after 15.7m E0302 22:34:20.355] FAIL: istio-presubmit ``` <!-- Please see https://istio.io/help and if you are a user of Istio, please file issues in https://github.com/istio/issues/issues instead of here. Only confirmed, triaged and labelled issues should be filed here. Please add the correct labels and epics (and priority and milestones if you have that information) -->
1.0
Test Failure: security/tests/integration/certificateRotationTest - From: https://k8s-gubernator.appspot.com/build/istio-prow/pull/istio_istio/3913/istio-presubmit/6070/ ``` I0302 22:34:20.343] 2018-03-02T22:34:20.341870Z error failed to create test namespace: failed to create a role (error: failed to create role (error: roles.rbac.authorization.k8s.io "istio-ca-role" is forbidden: attempt to grant extra privileges: [PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["create"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["secrets"], APIGroups:["core"], Verbs:["update"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["create"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["secrets"], APIGroups:[""], Verbs:["update"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["services"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["get"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["watch"]} PolicyRule{Resources:["pods"], APIGroups:["core"], Verbs:["list"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["serviceaccounts"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["services"], APIGroups:[""], Verbs:["list"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["get"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["watch"]} PolicyRule{Resources:["pods"], APIGroups:[""], Verbs:["list"]}] user=&{istio-prow-test-job@istio-testing.iam.gserviceaccount.com [system:authenticated] map[authenticator:[GKE]]} ownerrules=[PolicyRule{Resources:["selfsubjectaccessreviews" "selfsubjectrulesreviews"], APIGroups:["authorization.k8s.io"], Verbs:["create"]} PolicyRule{NonResourceURLs:["/api" "/api/*" "/apis" "/apis/*" "/healthz" "/swagger-2.0.0.pb-v1" "/swagger.json" "/swaggerapi" "/swaggerapi/*" "/version"], Verbs:["get"]}] ruleResolutionErrors=[])) I0302 22:34:20.344] 2018-03-02T22:34:20.341877Z error test environment creation failure I0302 22:34:20.344] FAIL istio.io/istio/security/tests/integration/certificateRotationTest 2.535s E0302 22:34:20.355] Build failed I0302 22:34:20.355] process 509 exited with code 1 after 15.7m E0302 22:34:20.355] FAIL: istio-presubmit ``` <!-- Please see https://istio.io/help and if you are a user of Istio, please file issues in https://github.com/istio/issues/issues instead of here. Only confirmed, triaged and labelled issues should be filed here. Please add the correct labels and epics (and priority and milestones if you have that information) -->
non_priority
test failure security tests integration certificaterotationtest from error failed to create test namespace failed to create a role error failed to create role error roles rbac authorization io istio ca role is forbidden attempt to grant extra privileges apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs policyrule resources apigroups verbs user istio prow test job istio testing iam gserviceaccount com map ownerrules apigroups verbs policyrule nonresourceurls verbs ruleresolutionerrors error test environment creation failure fail istio io istio security tests integration certificaterotationtest build failed process exited with code after fail istio presubmit please see and if you are a user of istio please file issues in instead of here only confirmed triaged and labelled issues should be filed here please add the correct labels and epics and priority and milestones if you have that information
0
44,335
12,101,453,469
IssuesEvent
2020-04-20 15:14:09
codesmithtools/Templates
https://api.github.com/repos/codesmithtools/Templates
closed
Join Table w/ Dependent Foreign Key
Framework-NHibernate Type-Defect auto-migrated
``` Update the IsManyToMany logic and add a constraint on having no dependent foreign keys. http://community.codesmithtools.com/forums/t/10071.aspx ``` Original issue reported on code.google.com by `tdupont...@gmail.com` on 24 Aug 2009 at 4:13
1.0
Join Table w/ Dependent Foreign Key - ``` Update the IsManyToMany logic and add a constraint on having no dependent foreign keys. http://community.codesmithtools.com/forums/t/10071.aspx ``` Original issue reported on code.google.com by `tdupont...@gmail.com` on 24 Aug 2009 at 4:13
non_priority
join table w dependent foreign key update the ismanytomany logic and add a constraint on having no dependent foreign keys original issue reported on code google com by tdupont gmail com on aug at
0
29,666
2,716,767,477
IssuesEvent
2015-04-10 21:15:18
CruxFramework/crux
https://api.github.com/repos/CruxFramework/crux
closed
DataProvider clearChanges method is not working
bug imported Milestone-M14-C4 Priority-Medium TargetVersion-5.3.0
_From [trbustam...@gmail.com](https://code.google.com/u/117925048001886933493/) on September 19, 2014 14:54:16_ The clearChanges method is not undoing editions on the dataprovider. It is only cleaning the change logs. _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=537_
1.0
DataProvider clearChanges method is not working - _From [trbustam...@gmail.com](https://code.google.com/u/117925048001886933493/) on September 19, 2014 14:54:16_ The clearChanges method is not undoing editions on the dataprovider. It is only cleaning the change logs. _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=537_
priority
dataprovider clearchanges method is not working from on september the clearchanges method is not undoing editions on the dataprovider it is only cleaning the change logs original issue
1
824,309
31,149,247,347
IssuesEvent
2023-08-16 08:49:17
fossasia/open-event-frontend
https://api.github.com/repos/fossasia/open-event-frontend
opened
Speakers getting disconnected from sessions when editing social media profile
bug Priority: Urgent
When a speaker adds social media entries to their account they get disconnected from some sessions if they have several sessions.
1.0
Speakers getting disconnected from sessions when editing social media profile - When a speaker adds social media entries to their account they get disconnected from some sessions if they have several sessions.
priority
speakers getting disconnected from sessions when editing social media profile when a speaker adds social media entries to their account they get disconnected from some sessions if they have several sessions
1
657,885
21,870,262,182
IssuesEvent
2022-05-19 03:59:34
pytorch/data
https://api.github.com/repos/pytorch/data
closed
Multiprocessing with any DataPipe writing to local file
bug good first issue help wanted high priority
### 🐛 Describe the bug We need to take extra care all DataPipe that would write to file system when DataLoader2 triggered multiprocessing. If the file name on the local file system is same across multiple processes, it would be a racing condition. This is found when TorchText team is using `on_disk_cache` to cache file. DataLoader needs to know such DataPipe must be sharded with multiprocessing or enforce it into single process. As a workaround, users have to download the file to local file system to prevent writing within DataPipe. ### Versions main branch
1.0
Multiprocessing with any DataPipe writing to local file - ### 🐛 Describe the bug We need to take extra care all DataPipe that would write to file system when DataLoader2 triggered multiprocessing. If the file name on the local file system is same across multiple processes, it would be a racing condition. This is found when TorchText team is using `on_disk_cache` to cache file. DataLoader needs to know such DataPipe must be sharded with multiprocessing or enforce it into single process. As a workaround, users have to download the file to local file system to prevent writing within DataPipe. ### Versions main branch
priority
multiprocessing with any datapipe writing to local file 🐛 describe the bug we need to take extra care all datapipe that would write to file system when triggered multiprocessing if the file name on the local file system is same across multiple processes it would be a racing condition this is found when torchtext team is using on disk cache to cache file dataloader needs to know such datapipe must be sharded with multiprocessing or enforce it into single process as a workaround users have to download the file to local file system to prevent writing within datapipe versions main branch
1
50,182
3,006,232,964
IssuesEvent
2015-07-27 09:03:14
Itseez/opencv
https://api.github.com/repos/Itseez/opencv
opened
Incorrect window size on Mac OS X Lion
auto-transferred bug category: highgui-gui priority: normal
Transferred from http://code.opencv.org/issues/2189 ``` || Jan Dlabal on 2012-07-24 17:26 || Priority: Normal || Affected: None || Category: highgui-gui || Tracker: Bug || Difficulty: None || PR: None || Platform: None / None ``` Incorrect window size on Mac OS X Lion ----------- ``` See http://stackoverflow.com/questions/11635842/opencv-not-filling-entire-image. Basically this: @ cv::Mat cvSideDepthImage1(150, 150, CV_8UC1, cv::Scalar(100)); cv::imshow("side1", cvSideDepthImage1); @ Creates this 200x150px window: !http://i.stack.imgur.com/HetAA.png! When it should just show a gray 150x150 square. This is OpenCV 2.4.9 on OS X Lion (all updates installed). ``` History ------- ##### Marina Kolpakova on 2012-07-24 18:05 ``` - Category set to highgui-images ``` ##### Andrey Kamaev on 2012-08-15 13:42 ``` - Assignee set to Vadim Pisarevsky ``` ##### Andrey Kamaev on 2012-08-16 15:42 ``` - Category changed from highgui-images to highgui-gui ```
1.0
Incorrect window size on Mac OS X Lion - Transferred from http://code.opencv.org/issues/2189 ``` || Jan Dlabal on 2012-07-24 17:26 || Priority: Normal || Affected: None || Category: highgui-gui || Tracker: Bug || Difficulty: None || PR: None || Platform: None / None ``` Incorrect window size on Mac OS X Lion ----------- ``` See http://stackoverflow.com/questions/11635842/opencv-not-filling-entire-image. Basically this: @ cv::Mat cvSideDepthImage1(150, 150, CV_8UC1, cv::Scalar(100)); cv::imshow("side1", cvSideDepthImage1); @ Creates this 200x150px window: !http://i.stack.imgur.com/HetAA.png! When it should just show a gray 150x150 square. This is OpenCV 2.4.9 on OS X Lion (all updates installed). ``` History ------- ##### Marina Kolpakova on 2012-07-24 18:05 ``` - Category set to highgui-images ``` ##### Andrey Kamaev on 2012-08-15 13:42 ``` - Assignee set to Vadim Pisarevsky ``` ##### Andrey Kamaev on 2012-08-16 15:42 ``` - Category changed from highgui-images to highgui-gui ```
priority
incorrect window size on mac os x lion transferred from jan dlabal on priority normal affected none category highgui gui tracker bug difficulty none pr none platform none none incorrect window size on mac os x lion see basically this cv mat cv cv scalar cv imshow creates this window when it should just show a gray square this is opencv on os x lion all updates installed history marina kolpakova on category set to highgui images andrey kamaev on assignee set to vadim pisarevsky andrey kamaev on category changed from highgui images to highgui gui
1
269,879
8,444,066,087
IssuesEvent
2018-10-18 17:19:44
poanetwork/metamask-extension
https://api.github.com/repos/poanetwork/metamask-extension
closed
(Bug) Token info isn't displayed if switch the network to localhost and back
high priority logical bug ready for release
Steps: 1. Set Sokol network 2. Add any token 3. Switch network to localhost 4. Switch network back to Sokol Expected result: - token info should be loaded and properly displayed Actual result: - token info isn't loaded <img width="1440" alt="screen shot 2018-09-10 at 11 47 41 pm" src="https://user-images.githubusercontent.com/14067096/45345149-54408a00-b55a-11e8-8244-745f1a927606.png">
1.0
(Bug) Token info isn't displayed if switch the network to localhost and back - Steps: 1. Set Sokol network 2. Add any token 3. Switch network to localhost 4. Switch network back to Sokol Expected result: - token info should be loaded and properly displayed Actual result: - token info isn't loaded <img width="1440" alt="screen shot 2018-09-10 at 11 47 41 pm" src="https://user-images.githubusercontent.com/14067096/45345149-54408a00-b55a-11e8-8244-745f1a927606.png">
priority
bug token info isn t displayed if switch the network to localhost and back steps set sokol network add any token switch network to localhost switch network back to sokol expected result token info should be loaded and properly displayed actual result token info isn t loaded img width alt screen shot at pm src
1
114,876
17,266,880,013
IssuesEvent
2021-07-22 14:44:56
turkdevops/php-src
https://api.github.com/repos/turkdevops/php-src
closed
CVE-2019-11041 (High) detected in php-srcphp-7.1.0RC3 - autoclosed
security vulnerability
## CVE-2019-11041 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>php-srcphp-7.1.0RC3</b></p></summary> <p> <p>The PHP Interpreter</p> <p>Library home page: <a href=https://github.com/php/php-src.git>https://github.com/php/php-src.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/php-src/commit/ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2">ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2</a></p> <p>Found in base branch: <b>microseconds</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>php-src/ext/exif/exif.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When PHP EXIF extension is parsing EXIF information from an image, e.g. via exif_read_data() function, in PHP versions 7.1.x below 7.1.31, 7.2.x below 7.2.21 and 7.3.x below 7.3.8 it is possible to supply it with data what will cause it to read past the allocated buffer. This may lead to information disclosure or crash. <p>Publish Date: 2019-08-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11041>CVE-2019-11041</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11041">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11041</a></p> <p>Release Date: 2019-08-09</p> <p>Fix Resolution: 7.1.31,7.2.21,7.3.8 </p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-11041 (High) detected in php-srcphp-7.1.0RC3 - autoclosed - ## CVE-2019-11041 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>php-srcphp-7.1.0RC3</b></p></summary> <p> <p>The PHP Interpreter</p> <p>Library home page: <a href=https://github.com/php/php-src.git>https://github.com/php/php-src.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/php-src/commit/ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2">ec57f9143f2fcf2e9a8d3dfa268da689d11be5e2</a></p> <p>Found in base branch: <b>microseconds</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>php-src/ext/exif/exif.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When PHP EXIF extension is parsing EXIF information from an image, e.g. via exif_read_data() function, in PHP versions 7.1.x below 7.1.31, 7.2.x below 7.2.21 and 7.3.x below 7.3.8 it is possible to supply it with data what will cause it to read past the allocated buffer. This may lead to information disclosure or crash. <p>Publish Date: 2019-08-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11041>CVE-2019-11041</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11041">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11041</a></p> <p>Release Date: 2019-08-09</p> <p>Fix Resolution: 7.1.31,7.2.21,7.3.8 </p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in php srcphp autoclosed cve high severity vulnerability vulnerable library php srcphp the php interpreter library home page a href found in head commit a href found in base branch microseconds vulnerable source files php src ext exif exif c vulnerability details when php exif extension is parsing exif information from an image e g via exif read data function in php versions x below x below and x below it is possible to supply it with data what will cause it to read past the allocated buffer this may lead to information disclosure or crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
125,225
26,620,774,968
IssuesEvent
2023-01-24 11:03:02
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
insights: use code insight title for data export filename
team/code-insights backend
follow up on https://github.com/sourcegraph/sourcegraph/pull/46662#discussion_r1084634856 escape illegal characters
1.0
insights: use code insight title for data export filename - follow up on https://github.com/sourcegraph/sourcegraph/pull/46662#discussion_r1084634856 escape illegal characters
non_priority
insights use code insight title for data export filename follow up on escape illegal characters
0
111,739
11,741,181,964
IssuesEvent
2020-03-11 21:09:49
ISPPNightTurn/Clubby
https://api.github.com/repos/ISPPNightTurn/Clubby
closed
Diseñar presentación del 11/03/2020
desing documentation
Es necesario preparar la presentación para la clase del próximo miércoles.
1.0
Diseñar presentación del 11/03/2020 - Es necesario preparar la presentación para la clase del próximo miércoles.
non_priority
diseñar presentación del es necesario preparar la presentación para la clase del próximo miércoles
0
79,357
10,120,685,146
IssuesEvent
2019-07-31 14:12:42
kids-first/kf-api-release-coordinator
https://api.github.com/repos/kids-first/kf-api-release-coordinator
closed
Add sphinx docs site
documentation
We should update existing documentation to use a sphinx docs site like most other code bases.
1.0
Add sphinx docs site - We should update existing documentation to use a sphinx docs site like most other code bases.
non_priority
add sphinx docs site we should update existing documentation to use a sphinx docs site like most other code bases
0
490,825
14,140,593,105
IssuesEvent
2020-11-10 11:25:38
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
SHUTDOWN should not disable SHUTDOWN_HARD
Category: Core & System Community Priority: Normal Status: Accepted Type: Feature
**Description** The most annoying issue of opennebula. **To Reproduce** Regarding: - Shutdown - Power Off - Reset Try to shut down a VM. Find out it won't shut down. Now sit and wait a few minutes so you can force shutdown it. Optionally, be educated that you want to do something invalid, while in reality SW failed. Also applies to trying a hard reset if a VM didn't boot off i.e. PXE. **Expected behavior** More intrusive "power actions" are not blocked by lighter ones. I can pull the power cable on a server if it hangs during shut down. There's nothing that will first block me from doing that and suddenly, without notice it will allow it. **Details** - Affected Component: [e.g. Sunstone, Scheduler or Storage] - Hypervisor: [e.g. KVM] - Version: [e.g. 5.4.12 or development] **Additional context** Add any other context about the problem here. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
SHUTDOWN should not disable SHUTDOWN_HARD - **Description** The most annoying issue of opennebula. **To Reproduce** Regarding: - Shutdown - Power Off - Reset Try to shut down a VM. Find out it won't shut down. Now sit and wait a few minutes so you can force shutdown it. Optionally, be educated that you want to do something invalid, while in reality SW failed. Also applies to trying a hard reset if a VM didn't boot off i.e. PXE. **Expected behavior** More intrusive "power actions" are not blocked by lighter ones. I can pull the power cable on a server if it hangs during shut down. There's nothing that will first block me from doing that and suddenly, without notice it will allow it. **Details** - Affected Component: [e.g. Sunstone, Scheduler or Storage] - Hypervisor: [e.g. KVM] - Version: [e.g. 5.4.12 or development] **Additional context** Add any other context about the problem here. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
priority
shutdown should not disable shutdown hard description the most annoying issue of opennebula to reproduce regarding shutdown power off reset try to shut down a vm find out it won t shut down now sit and wait a few minutes so you can force shutdown it optionally be educated that you want to do something invalid while in reality sw failed also applies to trying a hard reset if a vm didn t boot off i e pxe expected behavior more intrusive power actions are not blocked by lighter ones i can pull the power cable on a server if it hangs during shut down there s nothing that will first block me from doing that and suddenly without notice it will allow it details affected component hypervisor version additional context add any other context about the problem here progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
1
239,848
18,285,901,013
IssuesEvent
2021-10-05 10:13:32
girlscript/winter-of-contributing
https://api.github.com/repos/girlscript/winter-of-contributing
closed
Competitive Programming : : Sliding window maximum (documentation)
documentation GWOC21 Assigned Competitive Programming
<hr> ## Description 📜 I would like to provide documentation about the Sliding window maximum problem. <hr> ## Domain of Contribution 📊 <!----Please delete options that are not relevant.And in order to tick the check box just but x inside them for example [x] like this-----> - [x] Competitive Programming <hr>
1.0
Competitive Programming : : Sliding window maximum (documentation) - <hr> ## Description 📜 I would like to provide documentation about the Sliding window maximum problem. <hr> ## Domain of Contribution 📊 <!----Please delete options that are not relevant.And in order to tick the check box just but x inside them for example [x] like this-----> - [x] Competitive Programming <hr>
non_priority
competitive programming sliding window maximum documentation description 📜 i would like to provide documentation about the sliding window maximum problem domain of contribution 📊 competitive programming
0
250,619
7,979,201,252
IssuesEvent
2018-07-17 20:52:53
neurosynth/neurosynth-web
https://api.github.com/repos/neurosynth/neurosynth-web
closed
Update code page to include this repo
enhancement priority:med
The Code page needs to add a link to and description of this repository.
1.0
Update code page to include this repo - The Code page needs to add a link to and description of this repository.
priority
update code page to include this repo the code page needs to add a link to and description of this repository
1
107,692
16,762,159,331
IssuesEvent
2021-06-14 01:03:34
ioana-nicolae/first
https://api.github.com/repos/ioana-nicolae/first
closed
WS-2018-0209 (Medium) detected in morgan-1.8.0.tgz - autoclosed
security vulnerability
## WS-2018-0209 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>morgan-1.8.0.tgz</b></p></summary> <p>HTTP request logger middleware for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.8.0.tgz">https://registry.npmjs.org/morgan/-/morgan-1.8.0.tgz</a></p> <p>Path to dependency file: first/angular.js-master/angular.js-master/yarn.lock</p> <p>Path to vulnerable library: first/angular.js-master/angular.js-master/yarn.lock</p> <p> Dependency Hierarchy: - grunt-contrib-connect-1.0.2.tgz (Root Library) - :x: **morgan-1.8.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/first/commit/6e040516f98074b37f69996b76c1c99015c1d685">6e040516f98074b37f69996b76c1c99015c1d685</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack. <p>Publish Date: 2018-11-25 <p>URL: <a href=https://github.com/Twipped/ircdkit/issues/1>WS-2018-0209</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/735">https://www.npmjs.com/advisories/735</a></p> <p>Release Date: 2019-04-08</p> <p>Fix Resolution: 1.9.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"morgan","packageVersion":"1.8.0","packageFilePaths":["/angular.js-master/angular.js-master/yarn.lock"],"isTransitiveDependency":true,"dependencyTree":"grunt-contrib-connect:1.0.2;morgan:1.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2018-0209","vulnerabilityDetails":"Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.","vulnerabilityUrl":"https://github.com/Twipped/ircdkit/issues/1","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> -->
True
WS-2018-0209 (Medium) detected in morgan-1.8.0.tgz - autoclosed - ## WS-2018-0209 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>morgan-1.8.0.tgz</b></p></summary> <p>HTTP request logger middleware for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/morgan/-/morgan-1.8.0.tgz">https://registry.npmjs.org/morgan/-/morgan-1.8.0.tgz</a></p> <p>Path to dependency file: first/angular.js-master/angular.js-master/yarn.lock</p> <p>Path to vulnerable library: first/angular.js-master/angular.js-master/yarn.lock</p> <p> Dependency Hierarchy: - grunt-contrib-connect-1.0.2.tgz (Root Library) - :x: **morgan-1.8.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/first/commit/6e040516f98074b37f69996b76c1c99015c1d685">6e040516f98074b37f69996b76c1c99015c1d685</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack. <p>Publish Date: 2018-11-25 <p>URL: <a href=https://github.com/Twipped/ircdkit/issues/1>WS-2018-0209</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/735">https://www.npmjs.com/advisories/735</a></p> <p>Release Date: 2019-04-08</p> <p>Fix Resolution: 1.9.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"morgan","packageVersion":"1.8.0","packageFilePaths":["/angular.js-master/angular.js-master/yarn.lock"],"isTransitiveDependency":true,"dependencyTree":"grunt-contrib-connect:1.0.2;morgan:1.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2018-0209","vulnerabilityDetails":"Morgan before 1.9.1 is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack.","vulnerabilityUrl":"https://github.com/Twipped/ircdkit/issues/1","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> -->
non_priority
ws medium detected in morgan tgz autoclosed ws medium severity vulnerability vulnerable library morgan tgz http request logger middleware for node js library home page a href path to dependency file first angular js master angular js master yarn lock path to vulnerable library first angular js master angular js master yarn lock dependency hierarchy grunt contrib connect tgz root library x morgan tgz vulnerable library found in head commit a href found in base branch master vulnerability details morgan before is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt contrib connect morgan isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier ws vulnerabilitydetails morgan before is vulnerable to code injection when user input is allowed into the filter or combined with a prototype pollution attack vulnerabilityurl
0
162,834
6,176,610,319
IssuesEvent
2017-07-01 15:22:03
Placeholder-Software/Dissonance
https://api.github.com/repos/Placeholder-Software/Dissonance
closed
Reconnection Error
Priority: High Status: Awaiting User Feedback Type: Bug
Hi, we've bought your asset a couple of days ago and trying out your HLAPI as well as LLAPI integrations. Unfortunately LLAPI has some issues I'm not sure how to resolve them and that seem unusual. I'm currently working with Unity 5.6.0f3. Empty Project, Dissonance Package imported as well asLLAPI integration. ## Steps 0) Open DissonanceLlapiDemo 1) Build project, x64, development build 2) run build as client/server 3) run unity editor and connect (2-3 may be flipped) 4) wait a sec, maybe talk 5) disconnect client 6) wait a sec 7) reconnect sometimes when reconnecting/disconnecting something goes wrong and the following errors appear. After that I'm not able to connect to the server anymore and have to restart the server. Server side error: ``` host id {0} has been already deleted UnityEngine.Networking.NetworkTransport:ReceiveFromHost(Int32, Int32&, Int32&, Byte[], Int32, Int32&, Byte&) Dissonance.Integrations.UNet_LLAPI.UNetServer:ReadMessages() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetServer.cs:47) Dissonance.Networking.BaseServer`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseServer.cs:78) Dissonance.Networking.BaseCommsNetwork`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:124) [Dissonance:Network] UNetServer: Error reading server socket: WrongHost UnityEngine.Debug:LogError(Object) Dissonance.Logs:SendLogMessage(String, Action`1) (at Assets/Plugins/Dissonance/Core/Log.cs:114) Dissonance.Log:WriteLog(LogLevel, String) (at Assets/Plugins/Dissonance/Core/Log.cs:167) Dissonance.Log:WriteLogFormat(LogLevel, String, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:190) Dissonance.Log:Error(String, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:340) Dissonance.Integrations.UNet_LLAPI.UNetServer:ReadMessages() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetServer.cs:50) Dissonance.Networking.BaseServer`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseServer.cs:78) Dissonance.Networking.BaseCommsNetwork`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:124) ``` Client-side error ``` [Dissonance:Network] UNetClient: Failed to cleanly disconnect from Dissonance server at 127.0.0.1:5889, Error WrongConnection UnityEngine.Debug:LogError(Object) Dissonance.Logs:SendLogMessage(String, Action`1) (at Assets/Plugins/Dissonance/Core/Log.cs:114) Dissonance.Log:WriteLog(LogLevel, String) (at Assets/Plugins/Dissonance/Core/Log.cs:167) Dissonance.Log:WriteLogFormat(LogLevel, String, String, UInt16, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:208) Dissonance.Log:Error(String, String, UInt16, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:352) Dissonance.Integrations.UNet_LLAPI.UNetClient:Disconnect() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetClient.cs:40) Dissonance.Networking.BaseCommsNetwork`3:StopClient() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:186) Dissonance.Integrations.UNet_LLAPI.Demo.LlapiManager:OnGUI() (at Assets/Dissonance/Integrations/UNet_LLAPI/Demo/LlapiManager.cs:45) ``` When restarting the server (full restart, close application, reopening the application) and do not stop the client and try to reconnect I get: ``` [Dissonance:Network] UNetClient: Exception invoking event handler: System.ArgumentNullException: Argument cannot be null. Parameter name: key at System.Collections.Generic.Dictionary`2[System.String,Dissonance.VoicePlayerState].TryGetValue (System.String key, Dissonance.VoicePlayerState& value) [0x000ba] in /Users/builduser/buildslave/mono/build/mcs/class/corlib/System.Collections.Generic/Dictionary.cs:608 at Dissonance.DissonanceComms.Net_PlayerStartedSpeaking (System.String player) [0x0000a] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\DissonanceComms.cs:417 at Dissonance.Networking.BaseCommsNetwork`3[Dissonance.Integrations.UNet_LLAPI.UNetServer,Dissonance.Integrations.UNet_LLAPI.UNetClient,System.Int32].OnPlayerStartedSpeaking (System.String obj) [0x00010] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseCommsNetwork.cs:296 at Dissonance.Networking.BaseClient`3[Dissonance.Integrations.UNet_LLAPI.UNetServer,Dissonance.Integrations.UNet_LLAPI.UNetClient,System.Int32].InvokeEvent[String] (System.String& arg, System.Action`1 handler) [0x0000f] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:316 UnityEngine.DebugLogHandler:Internal_Log(LogType, String, Object) UnityEngine.DebugLogHandler:LogFormat(LogType, Object, String, Object[]) UnityEngine.Logger:Log(LogType, Object) UnityEngine.Debug:LogError(Object) Dissonance.Logs:SendLogMessage(String, Action`1) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:114) Dissonance.Log:WriteLog(LogLevel, String) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:167) Dissonance.Log:WriteLogFormat(LogLevel, String, Exception) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:190) Dissonance.Log:Error(String, Exception) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:340) Dissonance.Networking.BaseClient`3:InvokeEvent(String&, Action`1) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:320) Dissonance.Networking.BaseClient`3:DispatchEvents() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:282) Dissonance.Networking.BaseClient`3:Update() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:174) Dissonance.Networking.BaseCommsNetwork`3:Update() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseCommsNetwork.cs:128) ``` The dicionary error may be resolved by just adding a check if state already exists for this key and if so just update/replace the state.I already tried that but it doesn't resolve the initial error. Can you reproduce the error? If not I can upload you my test project. Or if you need anything else let me know. Would be great if you could give a quick hint if this is a known issue and/or being worked on or not so that I can plan ahead. Best Marcel
1.0
Reconnection Error - Hi, we've bought your asset a couple of days ago and trying out your HLAPI as well as LLAPI integrations. Unfortunately LLAPI has some issues I'm not sure how to resolve them and that seem unusual. I'm currently working with Unity 5.6.0f3. Empty Project, Dissonance Package imported as well asLLAPI integration. ## Steps 0) Open DissonanceLlapiDemo 1) Build project, x64, development build 2) run build as client/server 3) run unity editor and connect (2-3 may be flipped) 4) wait a sec, maybe talk 5) disconnect client 6) wait a sec 7) reconnect sometimes when reconnecting/disconnecting something goes wrong and the following errors appear. After that I'm not able to connect to the server anymore and have to restart the server. Server side error: ``` host id {0} has been already deleted UnityEngine.Networking.NetworkTransport:ReceiveFromHost(Int32, Int32&, Int32&, Byte[], Int32, Int32&, Byte&) Dissonance.Integrations.UNet_LLAPI.UNetServer:ReadMessages() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetServer.cs:47) Dissonance.Networking.BaseServer`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseServer.cs:78) Dissonance.Networking.BaseCommsNetwork`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:124) [Dissonance:Network] UNetServer: Error reading server socket: WrongHost UnityEngine.Debug:LogError(Object) Dissonance.Logs:SendLogMessage(String, Action`1) (at Assets/Plugins/Dissonance/Core/Log.cs:114) Dissonance.Log:WriteLog(LogLevel, String) (at Assets/Plugins/Dissonance/Core/Log.cs:167) Dissonance.Log:WriteLogFormat(LogLevel, String, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:190) Dissonance.Log:Error(String, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:340) Dissonance.Integrations.UNet_LLAPI.UNetServer:ReadMessages() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetServer.cs:50) Dissonance.Networking.BaseServer`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseServer.cs:78) Dissonance.Networking.BaseCommsNetwork`3:Update() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:124) ``` Client-side error ``` [Dissonance:Network] UNetClient: Failed to cleanly disconnect from Dissonance server at 127.0.0.1:5889, Error WrongConnection UnityEngine.Debug:LogError(Object) Dissonance.Logs:SendLogMessage(String, Action`1) (at Assets/Plugins/Dissonance/Core/Log.cs:114) Dissonance.Log:WriteLog(LogLevel, String) (at Assets/Plugins/Dissonance/Core/Log.cs:167) Dissonance.Log:WriteLogFormat(LogLevel, String, String, UInt16, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:208) Dissonance.Log:Error(String, String, UInt16, NetworkError) (at Assets/Plugins/Dissonance/Core/Log.cs:352) Dissonance.Integrations.UNet_LLAPI.UNetClient:Disconnect() (at Assets/Dissonance/Integrations/UNet_LLAPI/UNetClient.cs:40) Dissonance.Networking.BaseCommsNetwork`3:StopClient() (at Assets/Plugins/Dissonance/Core/Networking/BaseCommsNetwork.cs:186) Dissonance.Integrations.UNet_LLAPI.Demo.LlapiManager:OnGUI() (at Assets/Dissonance/Integrations/UNet_LLAPI/Demo/LlapiManager.cs:45) ``` When restarting the server (full restart, close application, reopening the application) and do not stop the client and try to reconnect I get: ``` [Dissonance:Network] UNetClient: Exception invoking event handler: System.ArgumentNullException: Argument cannot be null. Parameter name: key at System.Collections.Generic.Dictionary`2[System.String,Dissonance.VoicePlayerState].TryGetValue (System.String key, Dissonance.VoicePlayerState& value) [0x000ba] in /Users/builduser/buildslave/mono/build/mcs/class/corlib/System.Collections.Generic/Dictionary.cs:608 at Dissonance.DissonanceComms.Net_PlayerStartedSpeaking (System.String player) [0x0000a] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\DissonanceComms.cs:417 at Dissonance.Networking.BaseCommsNetwork`3[Dissonance.Integrations.UNet_LLAPI.UNetServer,Dissonance.Integrations.UNet_LLAPI.UNetClient,System.Int32].OnPlayerStartedSpeaking (System.String obj) [0x00010] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseCommsNetwork.cs:296 at Dissonance.Networking.BaseClient`3[Dissonance.Integrations.UNet_LLAPI.UNetServer,Dissonance.Integrations.UNet_LLAPI.UNetClient,System.Int32].InvokeEvent[String] (System.String& arg, System.Action`1 handler) [0x0000f] in D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:316 UnityEngine.DebugLogHandler:Internal_Log(LogType, String, Object) UnityEngine.DebugLogHandler:LogFormat(LogType, Object, String, Object[]) UnityEngine.Logger:Log(LogType, Object) UnityEngine.Debug:LogError(Object) Dissonance.Logs:SendLogMessage(String, Action`1) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:114) Dissonance.Log:WriteLog(LogLevel, String) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:167) Dissonance.Log:WriteLogFormat(LogLevel, String, Exception) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:190) Dissonance.Log:Error(String, Exception) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Log.cs:340) Dissonance.Networking.BaseClient`3:InvokeEvent(String&, Action`1) (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:320) Dissonance.Networking.BaseClient`3:DispatchEvents() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:282) Dissonance.Networking.BaseClient`3:Update() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseClient.cs:174) Dissonance.Networking.BaseCommsNetwork`3:Update() (at D:\Tests\AudioPluginTest\Assets\Plugins\Dissonance\Core\Networking\BaseCommsNetwork.cs:128) ``` The dicionary error may be resolved by just adding a check if state already exists for this key and if so just update/replace the state.I already tried that but it doesn't resolve the initial error. Can you reproduce the error? If not I can upload you my test project. Or if you need anything else let me know. Would be great if you could give a quick hint if this is a known issue and/or being worked on or not so that I can plan ahead. Best Marcel
priority
reconnection error hi we ve bought your asset a couple of days ago and trying out your hlapi as well as llapi integrations unfortunately llapi has some issues i m not sure how to resolve them and that seem unusual i m currently working with unity empty project dissonance package imported as well asllapi integration steps open dissonancellapidemo build project development build run build as client server run unity editor and connect may be flipped wait a sec maybe talk disconnect client wait a sec reconnect sometimes when reconnecting disconnecting something goes wrong and the following errors appear after that i m not able to connect to the server anymore and have to restart the server server side error host id has been already deleted unityengine networking networktransport receivefromhost byte byte dissonance integrations unet llapi unetserver readmessages at assets dissonance integrations unet llapi unetserver cs dissonance networking baseserver update at assets plugins dissonance core networking baseserver cs dissonance networking basecommsnetwork update at assets plugins dissonance core networking basecommsnetwork cs unetserver error reading server socket wronghost unityengine debug logerror object dissonance logs sendlogmessage string action at assets plugins dissonance core log cs dissonance log writelog loglevel string at assets plugins dissonance core log cs dissonance log writelogformat loglevel string networkerror at assets plugins dissonance core log cs dissonance log error string networkerror at assets plugins dissonance core log cs dissonance integrations unet llapi unetserver readmessages at assets dissonance integrations unet llapi unetserver cs dissonance networking baseserver update at assets plugins dissonance core networking baseserver cs dissonance networking basecommsnetwork update at assets plugins dissonance core networking basecommsnetwork cs client side error unetclient failed to cleanly disconnect from dissonance server at error wrongconnection unityengine debug logerror object dissonance logs sendlogmessage string action at assets plugins dissonance core log cs dissonance log writelog loglevel string at assets plugins dissonance core log cs dissonance log writelogformat loglevel string string networkerror at assets plugins dissonance core log cs dissonance log error string string networkerror at assets plugins dissonance core log cs dissonance integrations unet llapi unetclient disconnect at assets dissonance integrations unet llapi unetclient cs dissonance networking basecommsnetwork stopclient at assets plugins dissonance core networking basecommsnetwork cs dissonance integrations unet llapi demo llapimanager ongui at assets dissonance integrations unet llapi demo llapimanager cs when restarting the server full restart close application reopening the application and do not stop the client and try to reconnect i get unetclient exception invoking event handler system argumentnullexception argument cannot be null parameter name key at system collections generic dictionary trygetvalue system string key dissonance voiceplayerstate value in users builduser buildslave mono build mcs class corlib system collections generic dictionary cs at dissonance dissonancecomms net playerstartedspeaking system string player in d tests audioplugintest assets plugins dissonance dissonancecomms cs at dissonance networking basecommsnetwork onplayerstartedspeaking system string obj in d tests audioplugintest assets plugins dissonance core networking basecommsnetwork cs at dissonance networking baseclient invokeevent system string arg system action handler in d tests audioplugintest assets plugins dissonance core networking baseclient cs unityengine debugloghandler internal log logtype string object unityengine debugloghandler logformat logtype object string object unityengine logger log logtype object unityengine debug logerror object dissonance logs sendlogmessage string action at d tests audioplugintest assets plugins dissonance core log cs dissonance log writelog loglevel string at d tests audioplugintest assets plugins dissonance core log cs dissonance log writelogformat loglevel string exception at d tests audioplugintest assets plugins dissonance core log cs dissonance log error string exception at d tests audioplugintest assets plugins dissonance core log cs dissonance networking baseclient invokeevent string action at d tests audioplugintest assets plugins dissonance core networking baseclient cs dissonance networking baseclient dispatchevents at d tests audioplugintest assets plugins dissonance core networking baseclient cs dissonance networking baseclient update at d tests audioplugintest assets plugins dissonance core networking baseclient cs dissonance networking basecommsnetwork update at d tests audioplugintest assets plugins dissonance core networking basecommsnetwork cs the dicionary error may be resolved by just adding a check if state already exists for this key and if so just update replace the state i already tried that but it doesn t resolve the initial error can you reproduce the error if not i can upload you my test project or if you need anything else let me know would be great if you could give a quick hint if this is a known issue and or being worked on or not so that i can plan ahead best marcel
1
33,935
7,302,940,520
IssuesEvent
2018-02-27 11:21:05
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
generic HazelcastException thrown, wrapping com.hazelcast.core.MemberLeftException
Team: Core Type: Defect
exception ``` com.hazelcast.core.HazelcastException: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster! at com.hazelcast.util.ExceptionUtil$1.create(ExceptionUtil.java:40) at com.hazelcast.util.ExceptionUtil.peel(ExceptionUtil.java:116) at com.hazelcast.util.ExceptionUtil.peel(ExceptionUtil.java:69) at com.hazelcast.util.ExceptionUtil.rethrow(ExceptionUtil.java:129) at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.getQueryResults(MapPublisherCreateWithValueMessageTask.java:98) at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.call(MapPublisherCreateWithValueMessageTask.java:68) at com.hazelcast.client.impl.protocol.task.AbstractCallableMessageTask.processMessage(AbstractCallableMessageTask.java:35) at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:123) at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:103) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622) at java.lang.Thread.run(Thread.java:748) at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64) at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80) at ------ submitted from ------.(Unknown Source) at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:96) at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:33) at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:155) at com.hazelcast.client.impl.querycache.subscriber.ClientInvokerWrapper.invoke(ClientInvokerWrapper.java:89) at com.hazelcast.client.impl.querycache.subscriber.ClientQueryCacheEndToEndConstructor.createPublishAccumulatorWithIncludeValue(ClientQueryCacheEndToEndConstructor.java:68) at com.hazelcast.client.impl.querycache.subscriber.ClientQueryCacheEndToEndConstructor.createPublisherAccumulator(ClientQueryCacheEndToEndConstructor.java:49) at com.hazelcast.map.impl.querycache.subscriber.AbstractQueryCacheEndToEndConstructor.createNew(AbstractQueryCacheEndToEndConstructor.java:90) at com.hazelcast.map.impl.querycache.subscriber.AbstractQueryCacheEndToEndConstructor.createNew(AbstractQueryCacheEndToEndConstructor.java:40) at com.hazelcast.map.impl.querycache.subscriber.QueryCacheEndToEndProvider.getOrCreateQueryCache(QueryCacheEndToEndProvider.java:60) at com.hazelcast.client.proxy.ClientMapProxy.createQueryCache(ClientMapProxy.java:1494) at com.hazelcast.client.proxy.ClientMapProxy.getQueryCacheInternal(ClientMapProxy.java:1487) at com.hazelcast.client.proxy.ClientMapProxy.getQueryCache(ClientMapProxy.java:1464) at hzcmd.map.cqc.CreateUseDestroy.timeStep(CreateUseDestroy.java:16) at remote.bench.marker.MetricsMarker.flatOut(MetricsMarker.java:53) at remote.bench.marker.MetricsMarker.bench(MetricsMarker.java:40) at remote.bench.BenchThread.call(BenchThread.java:38) at remote.bench.BenchThread.call(BenchThread.java:12) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:473) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622) at java.lang.Thread.run(Thread.java:748) Caused by: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster! at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$OnMemberLeftTask.onTargetLoss(InvocationMonitor.java:392) at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$OnMemberLeftTask.run0(InvocationMonitor.java:359) at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$MonitorTask.run(InvocationMonitor.java:240) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:473) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622) at java.lang.Thread.run(Thread.java:748) at ------ submitted from ------.(Unknown Source) at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolve(InvocationFuture.java:127) at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolveAndThrowIfException(InvocationFuture.java:79) at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:155) at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.getQueryResults(MapPublisherCreateWithValueMessageTask.java:96) at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.call(MapPublisherCreateWithValueMessageTask.java:68) at com.hazelcast.client.impl.protocol.task.AbstractCallableMessageTask.processMessage(AbstractCallableMessageTask.java:35) at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:123) at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:103) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622) at java.lang.Thread.run(Thread.java:748) at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64) at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80) ``` from https://hazelcast-l337.ci.cloudbees.com/view/kill/job/kill-All/33/console /disk1/jenkins/workspace/kill-All/3.10-SNAPSHOT/2018_02_21-15_05_39/stable/destroy/create-use-destroy Failed ``` fail HzClient3HZ _create-use-destroy_createUseDestroy_mapBak1HD_queryCache hzcmd.map.cqc.CreateUseDestroy threadId=0 com.hazelcast.core.HazelcastException: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster! ``` http://54.82.84.143/~jenkins/workspace/kill-All/3.10-SNAPSHOT/2018_02_21-15_05_39/stable/destroy/create-use-destroy
1.0
generic HazelcastException thrown, wrapping com.hazelcast.core.MemberLeftException - exception ``` com.hazelcast.core.HazelcastException: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster! at com.hazelcast.util.ExceptionUtil$1.create(ExceptionUtil.java:40) at com.hazelcast.util.ExceptionUtil.peel(ExceptionUtil.java:116) at com.hazelcast.util.ExceptionUtil.peel(ExceptionUtil.java:69) at com.hazelcast.util.ExceptionUtil.rethrow(ExceptionUtil.java:129) at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.getQueryResults(MapPublisherCreateWithValueMessageTask.java:98) at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.call(MapPublisherCreateWithValueMessageTask.java:68) at com.hazelcast.client.impl.protocol.task.AbstractCallableMessageTask.processMessage(AbstractCallableMessageTask.java:35) at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:123) at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:103) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622) at java.lang.Thread.run(Thread.java:748) at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64) at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80) at ------ submitted from ------.(Unknown Source) at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:96) at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolveAndThrowIfException(ClientInvocationFuture.java:33) at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:155) at com.hazelcast.client.impl.querycache.subscriber.ClientInvokerWrapper.invoke(ClientInvokerWrapper.java:89) at com.hazelcast.client.impl.querycache.subscriber.ClientQueryCacheEndToEndConstructor.createPublishAccumulatorWithIncludeValue(ClientQueryCacheEndToEndConstructor.java:68) at com.hazelcast.client.impl.querycache.subscriber.ClientQueryCacheEndToEndConstructor.createPublisherAccumulator(ClientQueryCacheEndToEndConstructor.java:49) at com.hazelcast.map.impl.querycache.subscriber.AbstractQueryCacheEndToEndConstructor.createNew(AbstractQueryCacheEndToEndConstructor.java:90) at com.hazelcast.map.impl.querycache.subscriber.AbstractQueryCacheEndToEndConstructor.createNew(AbstractQueryCacheEndToEndConstructor.java:40) at com.hazelcast.map.impl.querycache.subscriber.QueryCacheEndToEndProvider.getOrCreateQueryCache(QueryCacheEndToEndProvider.java:60) at com.hazelcast.client.proxy.ClientMapProxy.createQueryCache(ClientMapProxy.java:1494) at com.hazelcast.client.proxy.ClientMapProxy.getQueryCacheInternal(ClientMapProxy.java:1487) at com.hazelcast.client.proxy.ClientMapProxy.getQueryCache(ClientMapProxy.java:1464) at hzcmd.map.cqc.CreateUseDestroy.timeStep(CreateUseDestroy.java:16) at remote.bench.marker.MetricsMarker.flatOut(MetricsMarker.java:53) at remote.bench.marker.MetricsMarker.bench(MetricsMarker.java:40) at remote.bench.BenchThread.call(BenchThread.java:38) at remote.bench.BenchThread.call(BenchThread.java:12) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:473) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622) at java.lang.Thread.run(Thread.java:748) Caused by: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster! at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$OnMemberLeftTask.onTargetLoss(InvocationMonitor.java:392) at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$OnMemberLeftTask.run0(InvocationMonitor.java:359) at com.hazelcast.spi.impl.operationservice.impl.InvocationMonitor$MonitorTask.run(InvocationMonitor.java:240) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:473) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622) at java.lang.Thread.run(Thread.java:748) at ------ submitted from ------.(Unknown Source) at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolve(InvocationFuture.java:127) at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolveAndThrowIfException(InvocationFuture.java:79) at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:155) at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.getQueryResults(MapPublisherCreateWithValueMessageTask.java:96) at com.hazelcast.client.impl.protocol.task.map.MapPublisherCreateWithValueMessageTask.call(MapPublisherCreateWithValueMessageTask.java:68) at com.hazelcast.client.impl.protocol.task.AbstractCallableMessageTask.processMessage(AbstractCallableMessageTask.java:35) at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:123) at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:103) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1152) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622) at java.lang.Thread.run(Thread.java:748) at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64) at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80) ``` from https://hazelcast-l337.ci.cloudbees.com/view/kill/job/kill-All/33/console /disk1/jenkins/workspace/kill-All/3.10-SNAPSHOT/2018_02_21-15_05_39/stable/destroy/create-use-destroy Failed ``` fail HzClient3HZ _create-use-destroy_createUseDestroy_mapBak1HD_queryCache hzcmd.map.cqc.CreateUseDestroy threadId=0 com.hazelcast.core.HazelcastException: com.hazelcast.core.MemberLeftException: Member [10.0.0.213]:5701 - d437817c-54f7-4631-a243-ac55ef2c7694 has left cluster! ``` http://54.82.84.143/~jenkins/workspace/kill-All/3.10-SNAPSHOT/2018_02_21-15_05_39/stable/destroy/create-use-destroy
non_priority
generic hazelcastexception thrown wrapping com hazelcast core memberleftexception exception com hazelcast core hazelcastexception com hazelcast core memberleftexception member has left cluster at com hazelcast util exceptionutil create exceptionutil java at com hazelcast util exceptionutil peel exceptionutil java at com hazelcast util exceptionutil peel exceptionutil java at com hazelcast util exceptionutil rethrow exceptionutil java at com hazelcast client impl protocol task map mappublishercreatewithvaluemessagetask getqueryresults mappublishercreatewithvaluemessagetask java at com hazelcast client impl protocol task map mappublishercreatewithvaluemessagetask call mappublishercreatewithvaluemessagetask java at com hazelcast client impl protocol task abstractcallablemessagetask processmessage abstractcallablemessagetask java at com hazelcast client impl protocol task abstractmessagetask initializeandprocessmessage abstractmessagetask java at com hazelcast client impl protocol task abstractmessagetask run abstractmessagetask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast util executor hazelcastmanagedthread run hazelcastmanagedthread java at submitted from unknown source at com hazelcast client spi impl clientinvocationfuture resolveandthrowifexception clientinvocationfuture java at com hazelcast client spi impl clientinvocationfuture resolveandthrowifexception clientinvocationfuture java at com hazelcast spi impl abstractinvocationfuture get abstractinvocationfuture java at com hazelcast client impl querycache subscriber clientinvokerwrapper invoke clientinvokerwrapper java at com hazelcast client impl querycache subscriber clientquerycacheendtoendconstructor createpublishaccumulatorwithincludevalue clientquerycacheendtoendconstructor java at com hazelcast client impl querycache subscriber clientquerycacheendtoendconstructor createpublisheraccumulator clientquerycacheendtoendconstructor java at com hazelcast map impl querycache subscriber abstractquerycacheendtoendconstructor createnew abstractquerycacheendtoendconstructor java at com hazelcast map impl querycache subscriber abstractquerycacheendtoendconstructor createnew abstractquerycacheendtoendconstructor java at com hazelcast map impl querycache subscriber querycacheendtoendprovider getorcreatequerycache querycacheendtoendprovider java at com hazelcast client proxy clientmapproxy createquerycache clientmapproxy java at com hazelcast client proxy clientmapproxy getquerycacheinternal clientmapproxy java at com hazelcast client proxy clientmapproxy getquerycache clientmapproxy java at hzcmd map cqc createusedestroy timestep createusedestroy java at remote bench marker metricsmarker flatout metricsmarker java at remote bench marker metricsmarker bench metricsmarker java at remote bench benchthread call benchthread java at remote bench benchthread call benchthread java at java util concurrent futuretask run futuretask java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by com hazelcast core memberleftexception member has left cluster at com hazelcast spi impl operationservice impl invocationmonitor onmemberlefttask ontargetloss invocationmonitor java at com hazelcast spi impl operationservice impl invocationmonitor onmemberlefttask invocationmonitor java at com hazelcast spi impl operationservice impl invocationmonitor monitortask run invocationmonitor java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at submitted from unknown source at com hazelcast spi impl operationservice impl invocationfuture resolve invocationfuture java at com hazelcast spi impl operationservice impl invocationfuture resolveandthrowifexception invocationfuture java at com hazelcast spi impl abstractinvocationfuture get abstractinvocationfuture java at com hazelcast client impl protocol task map mappublishercreatewithvaluemessagetask getqueryresults mappublishercreatewithvaluemessagetask java at com hazelcast client impl protocol task map mappublishercreatewithvaluemessagetask call mappublishercreatewithvaluemessagetask java at com hazelcast client impl protocol task abstractcallablemessagetask processmessage abstractcallablemessagetask java at com hazelcast client impl protocol task abstractmessagetask initializeandprocessmessage abstractmessagetask java at com hazelcast client impl protocol task abstractmessagetask run abstractmessagetask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast util executor hazelcastmanagedthread run hazelcastmanagedthread java from jenkins workspace kill all snapshot stable destroy create use destroy failed fail create use destroy createusedestroy querycache hzcmd map cqc createusedestroy threadid com hazelcast core hazelcastexception com hazelcast core memberleftexception member has left cluster
0
240,153
26,254,327,648
IssuesEvent
2023-01-05 22:33:07
TreyM-WSS/terra-clinical
https://api.github.com/repos/TreyM-WSS/terra-clinical
opened
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz
security vulnerability
## CVE-2021-3918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - terra-toolkit-6.0.0.tgz (Root Library) - node-sass-4.14.0.tgz - request-2.88.2.tgz - http-signature-1.2.0.tgz - jsprim-1.4.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TreyM-WSS/terra-clinical/commit/0940364474cdf2812b34bb24824139d7e3f8060b">0940364474cdf2812b34bb24824139d7e3f8060b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3918>CVE-2021-3918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p> <p>Release Date: 2021-11-13</p> <p>Fix Resolution (json-schema): 0.4.0</p> <p>Direct dependency fix Resolution (terra-toolkit): 6.1.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - ## CVE-2021-3918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - terra-toolkit-6.0.0.tgz (Root Library) - node-sass-4.14.0.tgz - request-2.88.2.tgz - http-signature-1.2.0.tgz - jsprim-1.4.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TreyM-WSS/terra-clinical/commit/0940364474cdf2812b34bb24824139d7e3f8060b">0940364474cdf2812b34bb24824139d7e3f8060b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3918>CVE-2021-3918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p> <p>Release Date: 2021-11-13</p> <p>Fix Resolution (json-schema): 0.4.0</p> <p>Direct dependency fix Resolution (terra-toolkit): 6.1.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_priority
cve high detected in json schema tgz cve high severity vulnerability vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file package json path to vulnerable library node modules json schema package json dependency hierarchy terra toolkit tgz root library node sass tgz request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in head commit a href found in base branch master vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution json schema direct dependency fix resolution terra toolkit rescue worker helmet automatic remediation is available for this issue
0
14,859
9,546,931,024
IssuesEvent
2019-05-01 21:25:23
mercycorps/TolaActivity
https://api.github.com/repos/mercycorps/TolaActivity
closed
Add meaningful error messages for failed Django login on login page
Deploy Ready Verified usability
Currently, failing to enter a valid u/p combination on the Django login returns me silently to the login page. We need to add: - [ ] ~link to reset password or other help]~ _(no longer addressing in this ticket)_ - [x] error message(s)
True
Add meaningful error messages for failed Django login on login page - Currently, failing to enter a valid u/p combination on the Django login returns me silently to the login page. We need to add: - [ ] ~link to reset password or other help]~ _(no longer addressing in this ticket)_ - [x] error message(s)
non_priority
add meaningful error messages for failed django login on login page currently failing to enter a valid u p combination on the django login returns me silently to the login page we need to add link to reset password or other help no longer addressing in this ticket error message s
0
604
3,003,590,670
IssuesEvent
2015-07-25 02:32:52
mesosphere/marathon
https://api.github.com/repos/mesosphere/marathon
closed
docker: hostPath in volumes seems to be ignored
bug OKR Usability service
mesos: 0.22.1 marathon: 0.8.1 docker: 1.6.2 I define: ``` "volumes": [{ "containerPath": "/var/log/vimana", "hostPath": "/var/log/vimana" }] ``` But the mounts I get are: ``` "Volumes": { "/mnt/mesos/sandbox": "/tmp/mesos/slaves/20150522-122903-2693333002-5050-7694-S9/frameworks/20150522-085224-378208266-5050-5138-0000/executors "/var/log/vimana": "/var/lib/docker/vfs/dir/38cf9ddf3c342baec70ade21a9cce2c267f909986f06cc101f5ecb672ae99f40" }, ``` The volume is declared in the docker file, so this behavior seems to as if the containerizer just dropped my mount.
1.0
docker: hostPath in volumes seems to be ignored - mesos: 0.22.1 marathon: 0.8.1 docker: 1.6.2 I define: ``` "volumes": [{ "containerPath": "/var/log/vimana", "hostPath": "/var/log/vimana" }] ``` But the mounts I get are: ``` "Volumes": { "/mnt/mesos/sandbox": "/tmp/mesos/slaves/20150522-122903-2693333002-5050-7694-S9/frameworks/20150522-085224-378208266-5050-5138-0000/executors "/var/log/vimana": "/var/lib/docker/vfs/dir/38cf9ddf3c342baec70ade21a9cce2c267f909986f06cc101f5ecb672ae99f40" }, ``` The volume is declared in the docker file, so this behavior seems to as if the containerizer just dropped my mount.
non_priority
docker hostpath in volumes seems to be ignored mesos marathon docker i define volumes containerpath var log vimana hostpath var log vimana but the mounts i get are volumes mnt mesos sandbox tmp mesos slaves frameworks executors var log vimana var lib docker vfs dir the volume is declared in the docker file so this behavior seems to as if the containerizer just dropped my mount
0
697,530
23,942,680,719
IssuesEvent
2022-09-12 02:25:29
jrsteensen/OpenHornet
https://api.github.com/repos/jrsteensen/OpenHornet
closed
Update Native F360 Stick Model
Type: Enhancement Category: MCAD Priority: Normal
- [x] Add new hall sensors to pitch axis - [x] Add new hall sensor to roll axis - [x] Mount Electronics to base - [x] Rename everything to OH PNs - [x] Update File Properties on all subcomponents. - [ ] Design an enclosure for the controller PCB to give it some protection.
1.0
Update Native F360 Stick Model - - [x] Add new hall sensors to pitch axis - [x] Add new hall sensor to roll axis - [x] Mount Electronics to base - [x] Rename everything to OH PNs - [x] Update File Properties on all subcomponents. - [ ] Design an enclosure for the controller PCB to give it some protection.
priority
update native stick model add new hall sensors to pitch axis add new hall sensor to roll axis mount electronics to base rename everything to oh pns update file properties on all subcomponents design an enclosure for the controller pcb to give it some protection
1
243,141
7,853,763,521
IssuesEvent
2018-06-20 18:29:12
canmet-energy/btap_tasks
https://api.github.com/repos/canmet-energy/btap_tasks
closed
Merge nrcan into develop
Priority High Standards
The schedule has moved up and we need to get our code into NREL's for the next release Due Monday the 11th of June.
1.0
Merge nrcan into develop - The schedule has moved up and we need to get our code into NREL's for the next release Due Monday the 11th of June.
priority
merge nrcan into develop the schedule has moved up and we need to get our code into nrel s for the next release due monday the of june
1
263,229
28,029,746,253
IssuesEvent
2023-03-28 11:35:20
RG4421/ampere-centos-kernel
https://api.github.com/repos/RG4421/ampere-centos-kernel
reopened
CVE-2022-1974 (Medium) detected in linuxv5.2
Mend: dependency security vulnerability
## CVE-2022-1974 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A use-after-free flaw was found in the Linux kernel's NFC core functionality due to a race condition between kobject creation and delete. This vulnerability allows a local attacker with CAP_NET_ADMIN privilege to leak kernel information. <p>Publish Date: 2022-08-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1974>CVE-2022-1974</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1974">https://www.linuxkernelcves.com/cves/CVE-2022-1974</a></p> <p>Release Date: 2022-08-31</p> <p>Fix Resolution: v4.9.313,v4.14.278,v4.19.242,v5.4.193,v5.10.115,v5.15.39,v5.17.7</p> </p> </details> <p></p>
True
CVE-2022-1974 (Medium) detected in linuxv5.2 - ## CVE-2022-1974 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A use-after-free flaw was found in the Linux kernel's NFC core functionality due to a race condition between kobject creation and delete. This vulnerability allows a local attacker with CAP_NET_ADMIN privilege to leak kernel information. <p>Publish Date: 2022-08-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1974>CVE-2022-1974</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1974">https://www.linuxkernelcves.com/cves/CVE-2022-1974</a></p> <p>Release Date: 2022-08-31</p> <p>Fix Resolution: v4.9.313,v4.14.278,v4.19.242,v5.4.193,v5.10.115,v5.15.39,v5.17.7</p> </p> </details> <p></p>
non_priority
cve medium detected in cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files vulnerability details a use after free flaw was found in the linux kernel s nfc core functionality due to a race condition between kobject creation and delete this vulnerability allows a local attacker with cap net admin privilege to leak kernel information publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
178,605
6,612,727,926
IssuesEvent
2017-09-20 06:02:28
arquillian/smart-testing
https://api.github.com/repos/arquillian/smart-testing
closed
NPE is thrown is strategy is mispelled
Component: Maven Priority: High Type: Bug
##### Issue Overview NPE is thrown if you set a strategy incorrectly ##### Expected Behaviour Throw a meaningful exception instead of NPE. ##### Current Behaviour NPE
1.0
NPE is thrown is strategy is mispelled - ##### Issue Overview NPE is thrown if you set a strategy incorrectly ##### Expected Behaviour Throw a meaningful exception instead of NPE. ##### Current Behaviour NPE
priority
npe is thrown is strategy is mispelled issue overview npe is thrown if you set a strategy incorrectly expected behaviour throw a meaningful exception instead of npe current behaviour npe
1
366,200
25,572,768,437
IssuesEvent
2022-11-30 19:08:57
Westlake-AI/openmixup
https://api.github.com/repos/Westlake-AI/openmixup
opened
Release Models of Mixups and MogaNet and Update Features in V0.2.6
documentation enhancement update
Updateing new features: 1. Fix the classification heads and update implementations and config files of [AlexNet](https://dl.acm.org/doi/10.1145/3065386) and [InceptionV3](https://arxiv.org/abs/1512.00567). Uploading Benchmark Results (release): 1. Release pre-trained models and logs of mixup benchmarks on ImageNet as provided in [AutoMix](https://arxiv.org/abs/2103.13027) and update results in [Model_Zoo_sup.md](https://github.com/Westlake-AI/openmixup/tree/main/docs/en/model_zoos/Model_Zoo_sup.md). 2. Release pre-trained models and logs of [MogaNet](https://arxiv.org/abs/2211.03295). Updating documents: 1. Update documents of mixup benchmarks on ImageNet with new backbones in [Model_Zoo_sup.md](https://github.com/Westlake-AI/openmixup/tree/main/docs/en/model_zoos/Model_Zoo_sup.md). 2. Update awesome lists in [Awesome Mixups](docs/en/awesome_selfsup/MIM.md) and [Awesome MIM](docs/en/awesome_selfsup/MIM.md) and provide teaser figures of most papers as illustrations.
1.0
Release Models of Mixups and MogaNet and Update Features in V0.2.6 - Updateing new features: 1. Fix the classification heads and update implementations and config files of [AlexNet](https://dl.acm.org/doi/10.1145/3065386) and [InceptionV3](https://arxiv.org/abs/1512.00567). Uploading Benchmark Results (release): 1. Release pre-trained models and logs of mixup benchmarks on ImageNet as provided in [AutoMix](https://arxiv.org/abs/2103.13027) and update results in [Model_Zoo_sup.md](https://github.com/Westlake-AI/openmixup/tree/main/docs/en/model_zoos/Model_Zoo_sup.md). 2. Release pre-trained models and logs of [MogaNet](https://arxiv.org/abs/2211.03295). Updating documents: 1. Update documents of mixup benchmarks on ImageNet with new backbones in [Model_Zoo_sup.md](https://github.com/Westlake-AI/openmixup/tree/main/docs/en/model_zoos/Model_Zoo_sup.md). 2. Update awesome lists in [Awesome Mixups](docs/en/awesome_selfsup/MIM.md) and [Awesome MIM](docs/en/awesome_selfsup/MIM.md) and provide teaser figures of most papers as illustrations.
non_priority
release models of mixups and moganet and update features in updateing new features fix the classification heads and update implementations and config files of and uploading benchmark results release release pre trained models and logs of mixup benchmarks on imagenet as provided in and update results in release pre trained models and logs of updating documents update documents of mixup benchmarks on imagenet with new backbones in update awesome lists in docs en awesome selfsup mim md and docs en awesome selfsup mim md and provide teaser figures of most papers as illustrations
0
587,294
17,612,267,406
IssuesEvent
2021-08-18 04:08:24
goplus/gop
https://api.github.com/repos/goplus/gop
closed
repl continueMode bug
bug priority:low
![image](https://user-images.githubusercontent.com/7893714/126966507-05c60b23-b0eb-4934-85e4-b153b82ddaaf.png) Cant input multiline code. go version go version go1.14.3 darwin/amd64 os: macOS Catalina 10.5.7
1.0
repl continueMode bug - ![image](https://user-images.githubusercontent.com/7893714/126966507-05c60b23-b0eb-4934-85e4-b153b82ddaaf.png) Cant input multiline code. go version go version go1.14.3 darwin/amd64 os: macOS Catalina 10.5.7
priority
repl continuemode bug cant input multiline code go version go version darwin os macos catalina
1
61,161
17,023,621,376
IssuesEvent
2021-07-03 02:58:09
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
search from IOS is crashing
Component: nominatim Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 10.44am, Wednesday, 4th August 2010]** Trying to do a search containing a "special keyword" (POI) like: http://nominatim.openstreetmap.org/search/airport from Safari (or any other web browser) on a IOS device (iPhone/iPad) does result in: string(354) "select place_id,false as in_small,false as in_large from search_name where name_vector @> ARRAY[] and ST_DWithin(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326), 0.1) order by ST_Distance(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326)) ASC, ST_Distance(ST_SetSRID(ST_Point(9.584734,47.659290),4326), centroid) asc, search_rank ASC limit 10" object(DB_Error)#2 (8) { ["error_message_prefix"]=> string(0) "" ["mode"]=> int(1) ["level"]=> int(1024) ["code"]=> int(-2) ["message"]=> string(22) "DB Error: syntax error" ["userinfo"]=> string(541) "select place_id,false as in_small,false as in_large from search_name where name_vector @> ARRAY[] and ST_DWithin(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326), 0.1) order by ST_Distance(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326)) ASC, ST_Distance(ST_SetSRID(ST_Point(9.584734,47.659290),4326), centroid) asc, search_rank ASC limit 10 [nativecode=ERROR: syntax error at or near "]" LINE 1: ...large from search_name where name_vector @> ARRAY[] and ST_D... ... ... ...
1.0
search from IOS is crashing - **[Submitted to the original trac issue database at 10.44am, Wednesday, 4th August 2010]** Trying to do a search containing a "special keyword" (POI) like: http://nominatim.openstreetmap.org/search/airport from Safari (or any other web browser) on a IOS device (iPhone/iPad) does result in: string(354) "select place_id,false as in_small,false as in_large from search_name where name_vector @> ARRAY[] and ST_DWithin(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326), 0.1) order by ST_Distance(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326)) ASC, ST_Distance(ST_SetSRID(ST_Point(9.584734,47.659290),4326), centroid) asc, search_rank ASC limit 10" object(DB_Error)#2 (8) { ["error_message_prefix"]=> string(0) "" ["mode"]=> int(1) ["level"]=> int(1024) ["code"]=> int(-2) ["message"]=> string(22) "DB Error: syntax error" ["userinfo"]=> string(541) "select place_id,false as in_small,false as in_large from search_name where name_vector @> ARRAY[] and ST_DWithin(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326), 0.1) order by ST_Distance(centroid, ST_SetSRID(ST_Point(9.584734,47.65929),4326)) ASC, ST_Distance(ST_SetSRID(ST_Point(9.584734,47.659290),4326), centroid) asc, search_rank ASC limit 10 [nativecode=ERROR: syntax error at or near "]" LINE 1: ...large from search_name where name_vector @> ARRAY[] and ST_D... ... ... ...
non_priority
search from ios is crashing trying to do a search containing a special keyword poi like from safari or any other web browser on a ios device iphone ipad does result in string select place id false as in small false as in large from search name where name vector array and st dwithin centroid st setsrid st point order by st distance centroid st setsrid st point asc st distance st setsrid st point centroid asc search rank asc limit object db error string int int int string db error syntax error string select place id false as in small false as in large from search name where name vector array and st dwithin centroid st setsrid st point order by st distance centroid st setsrid st point asc st distance st setsrid st point centroid asc search rank asc limit line large from search name where name vector array and st d
0
50,536
13,539,630,516
IssuesEvent
2020-09-16 13:42:28
cniweb/missing-link
https://api.github.com/repos/cniweb/missing-link
opened
CVE-2020-1945 (Medium) detected in ant-1.8.2.jar
security vulnerability
## CVE-2020-1945 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ant-1.8.2.jar</b></p></summary> <p>master POM</p> <p>Path to vulnerable library: missing-link/ant-props/lib/apache-ant/apache-ant-1.8.2.jar</p> <p> Dependency Hierarchy: - :x: **ant-1.8.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/cniweb/missing-link/commit/e22e6226faae42f8975654b836d8cd2d6de9967c">e22e6226faae42f8975654b836d8cd2d6de9967c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Ant 1.1 to 1.9.14 and 1.10.0 to 1.10.7 uses the default temporary directory identified by the Java system property java.io.tmpdir for several tasks and may thus leak sensitive information. The fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process. <p>Publish Date: 2020-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1945>CVE-2020-1945</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://ant.apache.org/security.html">https://ant.apache.org/security.html</a></p> <p>Release Date: 2020-05-14</p> <p>Fix Resolution: org.apache.ant:ant-junitlauncher:1.10.8;org.apache.ant:ant:1.9.15,1.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-1945 (Medium) detected in ant-1.8.2.jar - ## CVE-2020-1945 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ant-1.8.2.jar</b></p></summary> <p>master POM</p> <p>Path to vulnerable library: missing-link/ant-props/lib/apache-ant/apache-ant-1.8.2.jar</p> <p> Dependency Hierarchy: - :x: **ant-1.8.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/cniweb/missing-link/commit/e22e6226faae42f8975654b836d8cd2d6de9967c">e22e6226faae42f8975654b836d8cd2d6de9967c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Ant 1.1 to 1.9.14 and 1.10.0 to 1.10.7 uses the default temporary directory identified by the Java system property java.io.tmpdir for several tasks and may thus leak sensitive information. The fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process. <p>Publish Date: 2020-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1945>CVE-2020-1945</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://ant.apache.org/security.html">https://ant.apache.org/security.html</a></p> <p>Release Date: 2020-05-14</p> <p>Fix Resolution: org.apache.ant:ant-junitlauncher:1.10.8;org.apache.ant:ant:1.9.15,1.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in ant jar cve medium severity vulnerability vulnerable library ant jar master pom path to vulnerable library missing link ant props lib apache ant apache ant jar dependency hierarchy x ant jar vulnerable library found in head commit a href found in base branch master vulnerability details apache ant to and to uses the default temporary directory identified by the java system property java io tmpdir for several tasks and may thus leak sensitive information the fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache ant ant junitlauncher org apache ant ant step up your open source security game with whitesource
0
230,392
7,609,805,441
IssuesEvent
2018-05-01 03:11:06
Gamebuster19901/InventoryDecrapifier
https://api.github.com/repos/Gamebuster19901/InventoryDecrapifier
closed
Items are picked up one at a time
Bug Priority - Normal ↓ Side - Client Side - Server
Items are picked up from the ground one at a time at a rate of one per tick, instead of multiple per tick.
1.0
Items are picked up one at a time - Items are picked up from the ground one at a time at a rate of one per tick, instead of multiple per tick.
priority
items are picked up one at a time items are picked up from the ground one at a time at a rate of one per tick instead of multiple per tick
1
713,347
24,525,461,879
IssuesEvent
2022-10-11 12:49:58
quadratic-funding/mpc-phase2-suite
https://api.github.com/repos/quadratic-funding/mpc-phase2-suite
closed
Get access to cloud resource
DevOps ⚙ High Priority 🔥
### Description We need to get access for GCP and AWS cloud resources in order to switch from personal billing account to EF one. We need access to Firebase and GCP Cloud Functions + Compute Engine, AWS S3 (possibly AWS EC2).
1.0
Get access to cloud resource - ### Description We need to get access for GCP and AWS cloud resources in order to switch from personal billing account to EF one. We need access to Firebase and GCP Cloud Functions + Compute Engine, AWS S3 (possibly AWS EC2).
priority
get access to cloud resource description we need to get access for gcp and aws cloud resources in order to switch from personal billing account to ef one we need access to firebase and gcp cloud functions compute engine aws possibly aws
1
48,885
13,184,766,840
IssuesEvent
2020-08-12 20:03:21
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
buggy nutau (Trac #383)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/383 , reported by olivas and owned by olivas_</summary> <p> ```json { "status": "closed", "changetime": "2014-11-22T18:26:26", "description": "it seems to me that in the nutau nugen dataset 6539 the light emission\nfrom muons (if present in the event) is missing. I attach a scatter plot\nof muon energy vs muon track length (LDir) in IceCube. This should depend\nonly on the muon energy (and vertex position), but not on the primary\ntype. However, the reconstructed track length is aleays around 100m for\nnu-tau primary. That value can be reached by the cascade at sufficiently\nhigh energy.\n\nIt looks like there are two populations of muons in the tau MC, as I would\nexpect: one at very low muon energy (below 10 GeV) originating in the\ncascade. The other component is at higher energies (>100 GeV) and is\noriginating in the 17% tau->muon decay channel. So the particle phyiscs\nseems okay in that MC, but it was probably missed to simulate the light\nfrom the muon. This makes dataset 6539 unusable. We need urgently a fixed\nnutau dataset.\n\n Cheers,\n Andreas\n\n\nThis is likely due to the way nugen uses the propagators. For NuTau it uses the tau\npropagator and would not propagate muons in the final state.", "reporter": "olivas", "cc": "", "resolution": "fixed", "_ts": "1416680786826176", "component": "combo simulation", "summary": "buggy nutau", "priority": "normal", "keywords": "", "time": "2012-03-23T19:23:00", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
buggy nutau (Trac #383) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/383 , reported by olivas and owned by olivas_</summary> <p> ```json { "status": "closed", "changetime": "2014-11-22T18:26:26", "description": "it seems to me that in the nutau nugen dataset 6539 the light emission\nfrom muons (if present in the event) is missing. I attach a scatter plot\nof muon energy vs muon track length (LDir) in IceCube. This should depend\nonly on the muon energy (and vertex position), but not on the primary\ntype. However, the reconstructed track length is aleays around 100m for\nnu-tau primary. That value can be reached by the cascade at sufficiently\nhigh energy.\n\nIt looks like there are two populations of muons in the tau MC, as I would\nexpect: one at very low muon energy (below 10 GeV) originating in the\ncascade. The other component is at higher energies (>100 GeV) and is\noriginating in the 17% tau->muon decay channel. So the particle phyiscs\nseems okay in that MC, but it was probably missed to simulate the light\nfrom the muon. This makes dataset 6539 unusable. We need urgently a fixed\nnutau dataset.\n\n Cheers,\n Andreas\n\n\nThis is likely due to the way nugen uses the propagators. For NuTau it uses the tau\npropagator and would not propagate muons in the final state.", "reporter": "olivas", "cc": "", "resolution": "fixed", "_ts": "1416680786826176", "component": "combo simulation", "summary": "buggy nutau", "priority": "normal", "keywords": "", "time": "2012-03-23T19:23:00", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
non_priority
buggy nutau trac migrated from reported by olivas and owned by olivas json status closed changetime description it seems to me that in the nutau nugen dataset the light emission nfrom muons if present in the event is missing i attach a scatter plot nof muon energy vs muon track length ldir in icecube this should depend nonly on the muon energy and vertex position but not on the primary ntype however the reconstructed track length is aleays around for nnu tau primary that value can be reached by the cascade at sufficiently nhigh energy n nit looks like there are two populations of muons in the tau mc as i would nexpect one at very low muon energy below gev originating in the ncascade the other component is at higher energies gev and is noriginating in the tau muon decay channel so the particle phyiscs nseems okay in that mc but it was probably missed to simulate the light nfrom the muon this makes dataset unusable we need urgently a fixed nnutau dataset n n cheers n andreas n n nthis is likely due to the way nugen uses the propagators for nutau it uses the tau npropagator and would not propagate muons in the final state reporter olivas cc resolution fixed ts component combo simulation summary buggy nutau priority normal keywords time milestone owner olivas type defect
0
142,936
11,500,952,658
IssuesEvent
2020-02-12 16:21:52
wazuh/wazuh-qa
https://api.github.com/repos/wazuh/wazuh-qa
opened
FIM System tests: Create common tasks to verify alerts on alerts.json and Elasticsearch
fim-system-tests
### Objective The purpose of this issue is to keep track of the progress of tasks that parse the generated file from actions tasks at https://github.com/wazuh/wazuh-qa/issues/444 and ensures that an alert has been created for every file creation/modification/deletion ### Tasks - [ ] Create tasks to compare generated files with the alerts at `alerts.json` file. (assigned @Zenidd ) - [ ] Create tasks to compare generated files with the alerts at Elasticsearch index.
1.0
FIM System tests: Create common tasks to verify alerts on alerts.json and Elasticsearch - ### Objective The purpose of this issue is to keep track of the progress of tasks that parse the generated file from actions tasks at https://github.com/wazuh/wazuh-qa/issues/444 and ensures that an alert has been created for every file creation/modification/deletion ### Tasks - [ ] Create tasks to compare generated files with the alerts at `alerts.json` file. (assigned @Zenidd ) - [ ] Create tasks to compare generated files with the alerts at Elasticsearch index.
non_priority
fim system tests create common tasks to verify alerts on alerts json and elasticsearch objective the purpose of this issue is to keep track of the progress of tasks that parse the generated file from actions tasks at and ensures that an alert has been created for every file creation modification deletion tasks create tasks to compare generated files with the alerts at alerts json file assigned zenidd create tasks to compare generated files with the alerts at elasticsearch index
0
246,106
20,822,888,405
IssuesEvent
2022-03-18 17:10:32
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
opened
Missing copy shareable link button
bug test-day
Opening the share entity dialog does not always render the copy shareable URL to clipboard button. The input is also showing a different size: https://user-images.githubusercontent.com/716185/159049927-8b9e5cf6-c7a2-4854-ba88-2e25a018aead.mp4 This is not exclusive of saved searches, although I first noticed it there. There are no errors in the browser console and I could not find any patterns when it happens but it seems to be fixed when opening the modal for the same entity a second time. ## Expected Behavior Button to copy shareable URL to clipboard should be in share dialog. ## Current Behavior Button to copy shareable URL to clipboard is not always in share dialog. ## Steps to Reproduce (for bugs) 1. See video above ## Your Environment * Graylog Version: Graylog 4.3.0-beta.2+a61a34c * Browser version: Firefox Developer Edition 99.0b4 (64-bit), Vivaldi 5.1.2567.66 (Stable channel) (x86_64)
1.0
Missing copy shareable link button - Opening the share entity dialog does not always render the copy shareable URL to clipboard button. The input is also showing a different size: https://user-images.githubusercontent.com/716185/159049927-8b9e5cf6-c7a2-4854-ba88-2e25a018aead.mp4 This is not exclusive of saved searches, although I first noticed it there. There are no errors in the browser console and I could not find any patterns when it happens but it seems to be fixed when opening the modal for the same entity a second time. ## Expected Behavior Button to copy shareable URL to clipboard should be in share dialog. ## Current Behavior Button to copy shareable URL to clipboard is not always in share dialog. ## Steps to Reproduce (for bugs) 1. See video above ## Your Environment * Graylog Version: Graylog 4.3.0-beta.2+a61a34c * Browser version: Firefox Developer Edition 99.0b4 (64-bit), Vivaldi 5.1.2567.66 (Stable channel) (x86_64)
non_priority
missing copy shareable link button opening the share entity dialog does not always render the copy shareable url to clipboard button the input is also showing a different size this is not exclusive of saved searches although i first noticed it there there are no errors in the browser console and i could not find any patterns when it happens but it seems to be fixed when opening the modal for the same entity a second time expected behavior button to copy shareable url to clipboard should be in share dialog current behavior button to copy shareable url to clipboard is not always in share dialog steps to reproduce for bugs see video above your environment graylog version graylog beta browser version firefox developer edition bit vivaldi stable channel
0
367,282
25,730,921,407
IssuesEvent
2022-12-07 20:15:24
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
closed
[Docs]:
documentation service/ec2 needs-triage
### Documentation Link https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/ami ### Description As per documentation, [public](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/ami#public) - true if the image has public launch permissions. if the public flag is used as a filter then seeing below error message. Is `public` is still valid ? if i change `public` to `is-public` then its working as expected code snippet ------------ filter { name = "public" values = [true] } error message ------------ Error: InvalidParameterValue: The filter 'public' is invalid status code: 400, request id: a7f03df8-0b25-425e-a828-21b8d7a677b1 with data.aws_ami.debian, on main.tf line 63, in data "aws_ami" "debian": 63: data "aws_ami" "debian" { ### References _No response_ ### Would you like to implement a fix? None
1.0
[Docs]: - ### Documentation Link https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/ami ### Description As per documentation, [public](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/ami#public) - true if the image has public launch permissions. if the public flag is used as a filter then seeing below error message. Is `public` is still valid ? if i change `public` to `is-public` then its working as expected code snippet ------------ filter { name = "public" values = [true] } error message ------------ Error: InvalidParameterValue: The filter 'public' is invalid status code: 400, request id: a7f03df8-0b25-425e-a828-21b8d7a677b1 with data.aws_ami.debian, on main.tf line 63, in data "aws_ami" "debian": 63: data "aws_ami" "debian" { ### References _No response_ ### Would you like to implement a fix? None
non_priority
documentation link description as per documentation true if the image has public launch permissions if the public flag is used as a filter then seeing below error message is public is still valid if i change public to is public then its working as expected code snippet filter name public values error message error invalidparametervalue the filter public is invalid status code request id with data aws ami debian on main tf line in data aws ami debian data aws ami debian references no response would you like to implement a fix none
0
304,251
9,329,465,887
IssuesEvent
2019-03-28 02:30:23
rubrikinc/use-case-aws-cloudformation-template-cloudcluster
https://api.github.com/repos/rubrikinc/use-case-aws-cloudformation-template-cloudcluster
closed
Link for sharing CF Template
exp-beginner priority-p1
The template needs to be uploaded and shared out via an AWS link like the rest of the CF templates. For example: https://s3-us-west-1.amazonaws.com/cloudformation-templates-rubrik-prod/rubrik_cloudon.template
1.0
Link for sharing CF Template - The template needs to be uploaded and shared out via an AWS link like the rest of the CF templates. For example: https://s3-us-west-1.amazonaws.com/cloudformation-templates-rubrik-prod/rubrik_cloudon.template
priority
link for sharing cf template the template needs to be uploaded and shared out via an aws link like the rest of the cf templates for example
1
357,091
10,601,825,455
IssuesEvent
2019-10-10 13:07:40
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
Create a switch/variable for data discovery with PhEDEx x Rucio
Medium Priority New Feature ReqMgr2 Rucio Transition WorkQueue
For central services. Related to: https://its.cern.ch/jira/browse/CMSRUCIO-104
1.0
Create a switch/variable for data discovery with PhEDEx x Rucio - For central services. Related to: https://its.cern.ch/jira/browse/CMSRUCIO-104
priority
create a switch variable for data discovery with phedex x rucio for central services related to
1
84,050
7,888,578,955
IssuesEvent
2018-06-27 22:45:02
linnovate/root
https://api.github.com/repos/linnovate/root
closed
translate new entity has "type your text"
2.0.4 testing week
![image](https://user-images.githubusercontent.com/1281931/41975512-97cee28e-7a23-11e8-94c8-881e557ea9ac.png) translate new entity has "type your text" to "הכנס כותרת" and design according to original design (see http://root.203.projects.linnovate.net for referece)
1.0
translate new entity has "type your text" - ![image](https://user-images.githubusercontent.com/1281931/41975512-97cee28e-7a23-11e8-94c8-881e557ea9ac.png) translate new entity has "type your text" to "הכנס כותרת" and design according to original design (see http://root.203.projects.linnovate.net for referece)
non_priority
translate new entity has type your text translate new entity has type your text to הכנס כותרת and design according to original design see for referece
0
314,163
9,593,467,818
IssuesEvent
2019-05-09 11:37:59
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.rockstargames.com - site is not usable
browser-firefox-mobile engine-gecko priority-normal
<!-- @browser: Firefox Mobile 67.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:67.0) Gecko/67.0 Firefox/67.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.rockstargames.com/GTAOnline/restricted-content/agegate/form?redirect=https%3A%2F%2Fwww.rockstargames.com%2FGTAOnline%2Fnews&options=&locale=en_us **Browser / Version**: Firefox Mobile 67.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: entering birthday to access page gets stuck in loop, can't enter page **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190506235559</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.rockstargames.com - site is not usable - <!-- @browser: Firefox Mobile 67.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:67.0) Gecko/67.0 Firefox/67.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.rockstargames.com/GTAOnline/restricted-content/agegate/form?redirect=https%3A%2F%2Fwww.rockstargames.com%2FGTAOnline%2Fnews&options=&locale=en_us **Browser / Version**: Firefox Mobile 67.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: entering birthday to access page gets stuck in loop, can't enter page **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190506235559</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description entering birthday to access page gets stuck in loop can t enter page steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta from with ❤️
1
602,643
18,492,048,092
IssuesEvent
2021-10-19 02:27:59
AY2122S1-CS2113T-T12-3/tp
https://api.github.com/repos/AY2122S1-CS2113T-T12-3/tp
closed
Add function to clear all entries being tracked
priority.Medium
So that we can have a easier time testing code. Users might also find clear all function handy if they want to start afresh.
1.0
Add function to clear all entries being tracked - So that we can have a easier time testing code. Users might also find clear all function handy if they want to start afresh.
priority
add function to clear all entries being tracked so that we can have a easier time testing code users might also find clear all function handy if they want to start afresh
1
186,103
14,394,638,184
IssuesEvent
2020-12-03 01:46:13
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
tengteng/Guava: _vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go; 3 LoC
fresh test tiny
Found a possible issue in [tengteng/Guava](https://www.github.com/tengteng/Guava) at [_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to test at line 106 may start a goroutine [Click here to see the code in its original context.](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range importerTests { runImporterTest(t, imp, initmap, &test) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: f44ca584d2dc0fe32182990065cfc4fd0e6cebe8
1.0
tengteng/Guava: _vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go; 3 LoC - Found a possible issue in [tengteng/Guava](https://www.github.com/tengteng/Guava) at [_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to test at line 106 may start a goroutine [Click here to see the code in its original context.](https://github.com/tengteng/Guava/blob/f44ca584d2dc0fe32182990065cfc4fd0e6cebe8/_vendor/src/golang.org/x/tools/go/gccgoimporter/importer_test.go#L105-L107) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range importerTests { runImporterTest(t, imp, initmap, &test) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: f44ca584d2dc0fe32182990065cfc4fd0e6cebe8
non_priority
tengteng guava vendor src golang org x tools go gccgoimporter importer test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to test at line may start a goroutine click here to show the line s of go which triggered the analyzer go for test range importertests runimportertest t imp initmap test leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
545,022
15,934,230,699
IssuesEvent
2021-04-14 08:26:14
NamanhTran/Syntax-Analyzer
https://api.github.com/repos/NamanhTran/Syntax-Analyzer
closed
Create Parsing Table Algorithm
Hard Priority
## Overview We need to implement the algorithm to read from the parse table, keep track of all production rules for each lexeme, and handle errors. ## Action Items - [ ] Implement the algorithm - [ ] Generate meaningful error message when syntax error identified. - [ ] Test the table parser algorithm ## Assumptions for implementation ## Resources/Instructions - [Youtube video on the parsing algorithm](https://www.youtube.com/watch?v=KX_RU4rVTGk&t=733s).
1.0
Create Parsing Table Algorithm - ## Overview We need to implement the algorithm to read from the parse table, keep track of all production rules for each lexeme, and handle errors. ## Action Items - [ ] Implement the algorithm - [ ] Generate meaningful error message when syntax error identified. - [ ] Test the table parser algorithm ## Assumptions for implementation ## Resources/Instructions - [Youtube video on the parsing algorithm](https://www.youtube.com/watch?v=KX_RU4rVTGk&t=733s).
priority
create parsing table algorithm overview we need to implement the algorithm to read from the parse table keep track of all production rules for each lexeme and handle errors action items implement the algorithm generate meaningful error message when syntax error identified test the table parser algorithm assumptions for implementation resources instructions
1
11,452
4,227,265,273
IssuesEvent
2016-07-03 02:40:56
ac21/sherlock
https://api.github.com/repos/ac21/sherlock
closed
Fix "Rubocop/Lint/UnusedBlockArgument" issue in lib/api/v1/defaults.rb
code_climate
Unused block argument - `e`. You can omit the argument if you don't care about it. https://codeclimate.com/github/ac21/sherlock/lib/api/v1/defaults.rb#issue_5771abba9591a1000110c0f3
1.0
Fix "Rubocop/Lint/UnusedBlockArgument" issue in lib/api/v1/defaults.rb - Unused block argument - `e`. You can omit the argument if you don't care about it. https://codeclimate.com/github/ac21/sherlock/lib/api/v1/defaults.rb#issue_5771abba9591a1000110c0f3
non_priority
fix rubocop lint unusedblockargument issue in lib api defaults rb unused block argument e you can omit the argument if you don t care about it
0
751,717
26,254,831,555
IssuesEvent
2023-01-05 23:09:13
Automattic/woocommerce-payments
https://api.github.com/repos/Automattic/woocommerce-payments
closed
Non-prorated synced subscriptions don't result in synced WC Pay Subscription
type: bug priority: low component: wcpay subscriptions category: core
### Describe the bug <!-- A clear and concise description of what the bug is. Please be as descriptive as possible. --> When a merchant selects the **Never (charge the full recurring amount at sign-up)** synchronisation setting, synced subscription products aren't being synced to their anchor date in Stripe. ![Screen Shot 2022-08-11 at 03 15 22](https://user-images.githubusercontent.com/8490476/184043939-b083afe4-9cf9-4ed2-9527-1f5e311f1bb1.png) ### To Reproduce <!-- Describe the steps to reproduce the behavior. --> 1. With WC Subscription (the extension inactive). 2. Go to **WooCommerce > Settings > Subscriptions** 1. Enable synchronisation 2. Select `Never (charge the full recurring amount at sign-up)` **Prorate first payment** setting. 4. Create a synced subscription product (See example screenshot below). 5. Purchase that product and note the subscriptions first payment date. 6. In the Stripe dashboard, view the subscription's billing period and note the next payment date isn't synced. <img width="1247" alt="Screen Shot 2022-08-11 at 11 09 02 am" src="https://user-images.githubusercontent.com/8490476/184048395-d97453ed-fb0d-410c-b82c-1a4e87775771.png"> ### Actual behavior <!-- A clear and concise description of what actually happens. --> With the **Never (charge the full recurring amount at sign-up)** setting enabled we incorrectly determine that the product "isn't" synchronised. This means we don't send the `trial_end` arg to Stripe when creating the subscription. From my short investigation I believe the problem stems from this [logic](https://github.com/automattic/woocommerce-payments/blob/4.5.1/includes/subscriptions/class-wc-payments-subscription-service.php#L712-L714). Because there is an upfront payment (non-proration), that `self::has_delayed_payment( $subscription )` returns `false` and so no trial end date is sent. ### Screenshots <!-- If applicable, add screenshots to help explain your problem. --> <img width="1842" alt="Screen Shot 2022-08-11 at 11 03 02 am" src="https://user-images.githubusercontent.com/8490476/184047929-5bbeb015-da6f-48e3-bd15-5b084ac2623b.png"> ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> In this situation a `trial_end` value should be sent to Stripe when creating the subscription to ensure the subscription created in Stripe is synced. ### Additional information - Original report in Slack: p1660131760533269/1659720769.614509-slack-CGGCLBN58
1.0
Non-prorated synced subscriptions don't result in synced WC Pay Subscription - ### Describe the bug <!-- A clear and concise description of what the bug is. Please be as descriptive as possible. --> When a merchant selects the **Never (charge the full recurring amount at sign-up)** synchronisation setting, synced subscription products aren't being synced to their anchor date in Stripe. ![Screen Shot 2022-08-11 at 03 15 22](https://user-images.githubusercontent.com/8490476/184043939-b083afe4-9cf9-4ed2-9527-1f5e311f1bb1.png) ### To Reproduce <!-- Describe the steps to reproduce the behavior. --> 1. With WC Subscription (the extension inactive). 2. Go to **WooCommerce > Settings > Subscriptions** 1. Enable synchronisation 2. Select `Never (charge the full recurring amount at sign-up)` **Prorate first payment** setting. 4. Create a synced subscription product (See example screenshot below). 5. Purchase that product and note the subscriptions first payment date. 6. In the Stripe dashboard, view the subscription's billing period and note the next payment date isn't synced. <img width="1247" alt="Screen Shot 2022-08-11 at 11 09 02 am" src="https://user-images.githubusercontent.com/8490476/184048395-d97453ed-fb0d-410c-b82c-1a4e87775771.png"> ### Actual behavior <!-- A clear and concise description of what actually happens. --> With the **Never (charge the full recurring amount at sign-up)** setting enabled we incorrectly determine that the product "isn't" synchronised. This means we don't send the `trial_end` arg to Stripe when creating the subscription. From my short investigation I believe the problem stems from this [logic](https://github.com/automattic/woocommerce-payments/blob/4.5.1/includes/subscriptions/class-wc-payments-subscription-service.php#L712-L714). Because there is an upfront payment (non-proration), that `self::has_delayed_payment( $subscription )` returns `false` and so no trial end date is sent. ### Screenshots <!-- If applicable, add screenshots to help explain your problem. --> <img width="1842" alt="Screen Shot 2022-08-11 at 11 03 02 am" src="https://user-images.githubusercontent.com/8490476/184047929-5bbeb015-da6f-48e3-bd15-5b084ac2623b.png"> ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> In this situation a `trial_end` value should be sent to Stripe when creating the subscription to ensure the subscription created in Stripe is synced. ### Additional information - Original report in Slack: p1660131760533269/1659720769.614509-slack-CGGCLBN58
priority
non prorated synced subscriptions don t result in synced wc pay subscription describe the bug when a merchant selects the never charge the full recurring amount at sign up synchronisation setting synced subscription products aren t being synced to their anchor date in stripe to reproduce with wc subscription the extension inactive go to woocommerce settings subscriptions enable synchronisation select never charge the full recurring amount at sign up prorate first payment setting create a synced subscription product see example screenshot below purchase that product and note the subscriptions first payment date in the stripe dashboard view the subscription s billing period and note the next payment date isn t synced img width alt screen shot at am src actual behavior with the never charge the full recurring amount at sign up setting enabled we incorrectly determine that the product isn t synchronised this means we don t send the trial end arg to stripe when creating the subscription from my short investigation i believe the problem stems from this because there is an upfront payment non proration that self has delayed payment subscription returns false and so no trial end date is sent screenshots img width alt screen shot at am src expected behavior in this situation a trial end value should be sent to stripe when creating the subscription to ensure the subscription created in stripe is synced additional information original report in slack slack
1
258,007
19,538,686,851
IssuesEvent
2021-12-31 14:14:19
CalmiraLinux/cport
https://api.github.com/repos/CalmiraLinux/cport
closed
Обновить README
documentation
Добавить обновлённую информацию о работу с утилитой: - Установка порта; - Удаление порта; - Просмотр информации о порте.
1.0
Обновить README - Добавить обновлённую информацию о работу с утилитой: - Установка порта; - Удаление порта; - Просмотр информации о порте.
non_priority
обновить readme добавить обновлённую информацию о работу с утилитой установка порта удаление порта просмотр информации о порте
0
530,419
15,422,730,106
IssuesEvent
2021-03-05 14:44:59
jahirfiquitiva/Frames
https://api.github.com/repos/jahirfiquitiva/Frames
closed
Sec media storage crash
Priority: Low Status: Invalid Status: Not Reproducible
Whenever I try to save multiple wallpapers on apps like joywalls, wallflair, reev pro icon pack which use your wallpaper dashboard, the app sec media process and download manager crashes along with the app. This is only happening on Samsung one ui3 3.0 and 3.1 on my s10 and s21 ultra. Apps work done on one up 2.5 and aosp rom. I am using latest versions of all these apps and have tried clearing cache and data off crashing apps as well as cache from recovery. <!-- Any HTML comment will be stripped when the markdown is rendered, so you don't need to delete them. --> - [x] I have verified there are no duplicate active or recent bugs, questions, or requests - [x] I have verified that I am using the latest version. ### Device/App info: - Frames Version: `?` - Android version: `?` - Device Manufacturer: `?` - Device Name: `?` ### Describe the bug A clear and concise description of what the bug is. ### Reproduction Steps 1. 2. 3. ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> ### Screenshots <!-- If applicable, add screenshots or videos to help explain your problem. --> ### Code and/or Logs <!-- Please wrap code with correct syntax highlighting. You can remove it if you think it isn't necessary. --> ```kotlin println("Hello, world!") ``` <!-- If you are getting an error in the LogCat, paste here the stack trace. Please wrap logs with Gradle syntax highlighting (it makes them look better). --> ```Gradle java.lang.RuntimeException: This is an example Exception log at com.package.name.HelloWorld at com.package.name.HelloWorld$ThisIsNotARealLog at android.app.Instrumentation.callActivityOnResume(Instrumentation.kt) ``` ### Additional context <!-- Add any other context about the problem here. -->
1.0
Sec media storage crash - Whenever I try to save multiple wallpapers on apps like joywalls, wallflair, reev pro icon pack which use your wallpaper dashboard, the app sec media process and download manager crashes along with the app. This is only happening on Samsung one ui3 3.0 and 3.1 on my s10 and s21 ultra. Apps work done on one up 2.5 and aosp rom. I am using latest versions of all these apps and have tried clearing cache and data off crashing apps as well as cache from recovery. <!-- Any HTML comment will be stripped when the markdown is rendered, so you don't need to delete them. --> - [x] I have verified there are no duplicate active or recent bugs, questions, or requests - [x] I have verified that I am using the latest version. ### Device/App info: - Frames Version: `?` - Android version: `?` - Device Manufacturer: `?` - Device Name: `?` ### Describe the bug A clear and concise description of what the bug is. ### Reproduction Steps 1. 2. 3. ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> ### Screenshots <!-- If applicable, add screenshots or videos to help explain your problem. --> ### Code and/or Logs <!-- Please wrap code with correct syntax highlighting. You can remove it if you think it isn't necessary. --> ```kotlin println("Hello, world!") ``` <!-- If you are getting an error in the LogCat, paste here the stack trace. Please wrap logs with Gradle syntax highlighting (it makes them look better). --> ```Gradle java.lang.RuntimeException: This is an example Exception log at com.package.name.HelloWorld at com.package.name.HelloWorld$ThisIsNotARealLog at android.app.Instrumentation.callActivityOnResume(Instrumentation.kt) ``` ### Additional context <!-- Add any other context about the problem here. -->
priority
sec media storage crash whenever i try to save multiple wallpapers on apps like joywalls wallflair reev pro icon pack which use your wallpaper dashboard the app sec media process and download manager crashes along with the app this is only happening on samsung one and on my and ultra apps work done on one up and aosp rom i am using latest versions of all these apps and have tried clearing cache and data off crashing apps as well as cache from recovery any html comment will be stripped when the markdown is rendered so you don t need to delete them i have verified there are no duplicate active or recent bugs questions or requests i have verified that i am using the latest version device app info frames version android version device manufacturer device name describe the bug a clear and concise description of what the bug is reproduction steps expected behavior screenshots code and or logs please wrap code with correct syntax highlighting you can remove it if you think it isn t necessary kotlin println hello world if you are getting an error in the logcat paste here the stack trace please wrap logs with gradle syntax highlighting it makes them look better gradle java lang runtimeexception this is an example exception log at com package name helloworld at com package name helloworld thisisnotareallog at android app instrumentation callactivityonresume instrumentation kt additional context
1
569,326
17,011,973,387
IssuesEvent
2021-07-02 06:36:01
Uninett/Argus-frontend
https://api.github.com/repos/Uninett/Argus-frontend
opened
Filtering incidents by severity level
backend discussion enhancement priority: high size: small
Severity levels were introduced into the API as part of [Uninett/Argus#70](https://github.com/Uninett/Argus/issues/70), and issue #271 deals with showing the severity level in the frontend. The next step is to make it possible to also filter on severity level in the frontend. There is already support for this in the backend. My suggestion is to use a dropdown to filter on max severity level. The frontend will then show issues with severity level lower than or equal to the selected value. The dropdown will probably look something like this: ![image](https://user-images.githubusercontent.com/59874198/124231413-8444e780-db10-11eb-8706-b467f99f10fb.png)
1.0
Filtering incidents by severity level - Severity levels were introduced into the API as part of [Uninett/Argus#70](https://github.com/Uninett/Argus/issues/70), and issue #271 deals with showing the severity level in the frontend. The next step is to make it possible to also filter on severity level in the frontend. There is already support for this in the backend. My suggestion is to use a dropdown to filter on max severity level. The frontend will then show issues with severity level lower than or equal to the selected value. The dropdown will probably look something like this: ![image](https://user-images.githubusercontent.com/59874198/124231413-8444e780-db10-11eb-8706-b467f99f10fb.png)
priority
filtering incidents by severity level severity levels were introduced into the api as part of and issue deals with showing the severity level in the frontend the next step is to make it possible to also filter on severity level in the frontend there is already support for this in the backend my suggestion is to use a dropdown to filter on max severity level the frontend will then show issues with severity level lower than or equal to the selected value the dropdown will probably look something like this
1
54,475
30,198,664,388
IssuesEvent
2023-07-05 02:06:10
cilium/cilium
https://api.github.com/repos/cilium/cilium
closed
Tunnel mode lowers throughput by a large amount
kind/bug need-more-info sig/datapath kind/performance needs/triage kind/community-report stale
### Is there an existing issue for this? - [X] I have searched the existing issues ### What happened? When using cilium with tunnel mode enabled, throughput drops by a significant amount(30-50%). This was seen in #22898 with the attached logs. This bug is an extension of that to discuss matters related specifically to that. ### Cilium Version 1.12.3 ### Kernel Version 5.4.0-42-generic ### Kubernetes Version 1.26 ### Sysdump Couldn't get it ### Relevant log output _No response_ ### Anything else? From other issue 1) tunnel: disabled, mtu: 1450 [logs](https://gist.github.com/michaelasp/3883c8fdfd298ad192a80f6a1db13262) ~8.24gb/s 2) tunnel: vxlan [logs](https://gist.github.com/michaelasp/1081274981900abfbfefdcccd6f52a68) ~5.8 Gb/s, 30% drop 3) tunnel: geneve [logs](https://gist.github.com/michaelasp/542ba5bfb0e325be2bb16585167e7314) ~5.2 Gb/s 37% drop ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
True
Tunnel mode lowers throughput by a large amount - ### Is there an existing issue for this? - [X] I have searched the existing issues ### What happened? When using cilium with tunnel mode enabled, throughput drops by a significant amount(30-50%). This was seen in #22898 with the attached logs. This bug is an extension of that to discuss matters related specifically to that. ### Cilium Version 1.12.3 ### Kernel Version 5.4.0-42-generic ### Kubernetes Version 1.26 ### Sysdump Couldn't get it ### Relevant log output _No response_ ### Anything else? From other issue 1) tunnel: disabled, mtu: 1450 [logs](https://gist.github.com/michaelasp/3883c8fdfd298ad192a80f6a1db13262) ~8.24gb/s 2) tunnel: vxlan [logs](https://gist.github.com/michaelasp/1081274981900abfbfefdcccd6f52a68) ~5.8 Gb/s, 30% drop 3) tunnel: geneve [logs](https://gist.github.com/michaelasp/542ba5bfb0e325be2bb16585167e7314) ~5.2 Gb/s 37% drop ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
non_priority
tunnel mode lowers throughput by a large amount is there an existing issue for this i have searched the existing issues what happened when using cilium with tunnel mode enabled throughput drops by a significant amount this was seen in with the attached logs this bug is an extension of that to discuss matters related specifically to that cilium version kernel version generic kubernetes version sysdump couldn t get it relevant log output no response anything else from other issue tunnel disabled mtu s tunnel vxlan gb s drop tunnel geneve gb s drop code of conduct i agree to follow this project s code of conduct
0
647,635
21,132,750,417
IssuesEvent
2022-04-06 01:22:36
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Make OnShowPanel and OnShowWalletOnboarding for Android only
priority/P2 QA/No release-notes/exclude feature/wallet OS/Android
OnShowPanel(); OnShowWalletOnboarding(); observer functions should be visible for Android only
1.0
Make OnShowPanel and OnShowWalletOnboarding for Android only - OnShowPanel(); OnShowWalletOnboarding(); observer functions should be visible for Android only
priority
make onshowpanel and onshowwalletonboarding for android only onshowpanel onshowwalletonboarding observer functions should be visible for android only
1
540,084
15,800,575,450
IssuesEvent
2021-04-03 00:20:11
JensenJ/EmbargoMC-IssueTracker
https://api.github.com/repos/JensenJ/EmbargoMC-IssueTracker
closed
[BUG] Water / Lava Deleted Source Blocks
bug low-priority
**Describe the bug** Source blocks get deleted by custom caves from world painter **Expected behaviour** Consistent water/lava lakes spawning without holes in caused by caves
1.0
[BUG] Water / Lava Deleted Source Blocks - **Describe the bug** Source blocks get deleted by custom caves from world painter **Expected behaviour** Consistent water/lava lakes spawning without holes in caused by caves
priority
water lava deleted source blocks describe the bug source blocks get deleted by custom caves from world painter expected behaviour consistent water lava lakes spawning without holes in caused by caves
1
595,489
18,067,595,632
IssuesEvent
2021-09-20 21:07:17
OpenMandrivaAssociation/test2
https://api.github.com/repos/OpenMandrivaAssociation/test2
closed
Adding users crashes the program (Bugzilla Bug 138)
bug high priority major
This issue was created automatically with bugzilla2github # Bugzilla Bug 138 Date: 2013-09-14 15:05:51 +0000 From: @robxu9 To: OpenMandriva QA &lt;<bugs@openmandriva.org>&gt; CC: @cris-b Last updated: 2013-09-19 20:05:04 +0000 ## Comment 822 Date: 2013-09-14 15:05:51 +0000 From: @robxu9 Theme name: rosa-elementary Kernel version = 3.10.7-nrjQL-desktop-1omv Distribution=OpenMandriva LX release 2013.0 Beta (Oxygen) for x86_64 CPU=QEMU Virtual CPU version 1.4.2 When I add users with userdrake, it closes unexpectedly. The user is still added, but locked. ## Comment 825 Date: 2013-09-14 15:13:58 +0000 From: @robxu9 *** Bug #139 has been marked as a duplicate of this bug. *** ## Comment 955 Date: 2013-09-18 21:00:17 +0000 From: @cris-b looks like the recent update to userdrake-1.14 requires libuser to be updated to 0.59. will request that we update libuser to 0.59 to sort this. ## Comment 963 Date: 2013-09-19 11:08:25 +0000 From: @cris-b should now be resolved as libuser has been updated to 0.59
1.0
Adding users crashes the program (Bugzilla Bug 138) - This issue was created automatically with bugzilla2github # Bugzilla Bug 138 Date: 2013-09-14 15:05:51 +0000 From: @robxu9 To: OpenMandriva QA &lt;<bugs@openmandriva.org>&gt; CC: @cris-b Last updated: 2013-09-19 20:05:04 +0000 ## Comment 822 Date: 2013-09-14 15:05:51 +0000 From: @robxu9 Theme name: rosa-elementary Kernel version = 3.10.7-nrjQL-desktop-1omv Distribution=OpenMandriva LX release 2013.0 Beta (Oxygen) for x86_64 CPU=QEMU Virtual CPU version 1.4.2 When I add users with userdrake, it closes unexpectedly. The user is still added, but locked. ## Comment 825 Date: 2013-09-14 15:13:58 +0000 From: @robxu9 *** Bug #139 has been marked as a duplicate of this bug. *** ## Comment 955 Date: 2013-09-18 21:00:17 +0000 From: @cris-b looks like the recent update to userdrake-1.14 requires libuser to be updated to 0.59. will request that we update libuser to 0.59 to sort this. ## Comment 963 Date: 2013-09-19 11:08:25 +0000 From: @cris-b should now be resolved as libuser has been updated to 0.59
priority
adding users crashes the program bugzilla bug this issue was created automatically with bugzilla bug date from to openmandriva qa lt gt cc cris b last updated comment date from theme name rosa elementary kernel version nrjql desktop distribution openmandriva lx release beta oxygen for cpu qemu virtual cpu version when i add users with userdrake it closes unexpectedly the user is still added but locked comment date from bug has been marked as a duplicate of this bug comment date from cris b looks like the recent update to userdrake requires libuser to be updated to will request that we update libuser to to sort this comment date from cris b should now be resolved as libuser has been updated to
1
156,371
5,968,203,730
IssuesEvent
2017-05-30 17:36:34
kolihub/koli
https://api.github.com/repos/kolihub/koli
opened
Better error handling when downloading releases
area/slugrunner improvement priority/P2
The slugrunner doesn't validate any error when downloading releases from the git-server.
1.0
Better error handling when downloading releases - The slugrunner doesn't validate any error when downloading releases from the git-server.
priority
better error handling when downloading releases the slugrunner doesn t validate any error when downloading releases from the git server
1
729,955
25,152,590,302
IssuesEvent
2022-11-10 11:07:35
rism-digital/rism-online-issues
https://api.github.com/repos/rism-digital/rism-online-issues
closed
Add contour search for incipit
Audience: General public Priority: Moderate Status: Blocked Topic: Incipits Component: Incipit Search
In Muscat BL we used to have contour search enabled. It was the implementation of Themefinder * http://www.themefinder.org/help/refinedcontour/ * http://www.themefinder.org/help/grosscontour/ Having one type of contour might be useful, especially since we have now highlighting.
1.0
Add contour search for incipit - In Muscat BL we used to have contour search enabled. It was the implementation of Themefinder * http://www.themefinder.org/help/refinedcontour/ * http://www.themefinder.org/help/grosscontour/ Having one type of contour might be useful, especially since we have now highlighting.
priority
add contour search for incipit in muscat bl we used to have contour search enabled it was the implementation of themefinder having one type of contour might be useful especially since we have now highlighting
1
289,596
31,986,705,316
IssuesEvent
2023-09-21 00:20:44
pferronMend/Case121627
https://api.github.com/repos/pferronMend/Case121627
opened
rocksdbnative.6.2.2.nupkg: 1 vulnerabilities (highest severity is: 7.5)
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rocksdbnative.6.2.2.nupkg</b></p></summary> <p></p> <p>Path to dependency file: /examples/ExactlyOnceOldBroker/ExactlyOnceOldBroker.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (rocksdbnative.6.2.2.nupkg version) | Remediation Possible** | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2018-8292](https://www.mend.io/vulnerability-database/CVE-2018-8292) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | system.net.http.4.3.0.nupkg | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p><p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2018-8292</summary> ### Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p> <p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers. </p> <p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p> <p>Path to dependency file: /examples/ExactlyOnceOldBroker/ExactlyOnceOldBroker.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p> <p> Dependency Hierarchy: - rocksdbnative.6.2.2.nupkg (Root Library) - netstandard.library.1.6.1.nupkg - :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0. <p>Publish Date: 2018-10-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-10-10</p> <p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p> </p> <p></p> </details>
True
rocksdbnative.6.2.2.nupkg: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rocksdbnative.6.2.2.nupkg</b></p></summary> <p></p> <p>Path to dependency file: /examples/ExactlyOnceOldBroker/ExactlyOnceOldBroker.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (rocksdbnative.6.2.2.nupkg version) | Remediation Possible** | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2018-8292](https://www.mend.io/vulnerability-database/CVE-2018-8292) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | system.net.http.4.3.0.nupkg | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p><p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2018-8292</summary> ### Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p> <p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers. </p> <p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p> <p>Path to dependency file: /examples/ExactlyOnceOldBroker/ExactlyOnceOldBroker.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p> <p> Dependency Hierarchy: - rocksdbnative.6.2.2.nupkg (Root Library) - netstandard.library.1.6.1.nupkg - :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0. <p>Publish Date: 2018-10-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-10-10</p> <p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p> </p> <p></p> </details>
non_priority
rocksdbnative nupkg vulnerabilities highest severity is vulnerable library rocksdbnative nupkg path to dependency file examples exactlyonceoldbroker exactlyonceoldbroker csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg vulnerabilities cve severity cvss dependency type fixed in rocksdbnative nupkg version remediation possible high system net http nupkg transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the details section below to see if there is a version of transitive dependency where vulnerability is fixed in some cases remediation pr cannot be created automatically for a vulnerability despite the availability of remediation details cve vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that allow applications to consume web services over http and http components that can be used by both clients and servers for parsing http headers library home page a href path to dependency file examples exactlyonceoldbroker exactlyonceoldbroker csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg dependency hierarchy rocksdbnative nupkg root library netstandard library nupkg x system net http nupkg vulnerable library found in base branch master vulnerability details an information disclosure vulnerability exists in net core when authentication information is inadvertently exposed in a redirect aka net core information disclosure vulnerability this affects net core net core net core powershell core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution system net http microsoft powershell commands utility rc
0
403,415
11,840,709,742
IssuesEvent
2020-03-23 19:22:58
AlaskaAirlines/auro_docs
https://api.github.com/repos/AlaskaAirlines/auro_docs
opened
CDN Ready components
Priority: High Status: Proposal Type: Feature
# User story As a web developer, I want to be able to consume Auro components w/o having to pre-compile code, so I can easily use resources in a non-dynamic environment. ## Outline subtasks or tasks > Decide which specific steps need to be completed and who is responsible for each of them. - [ ] Output pre-rendered JS version of web componet - [ ] store WC in CDN - [ ] append versioning to the ouput file name - [ ] ensure stability of delivered asset - [ ] have solution for polyfill ## Definition of “Done” (exit criteria) > The story is generally “done” when the user can complete the outlined task, but make sure to define what that is. An initial Auro web component has been updated with functionality to export a consumable version of the WC via a CDN.
1.0
CDN Ready components - # User story As a web developer, I want to be able to consume Auro components w/o having to pre-compile code, so I can easily use resources in a non-dynamic environment. ## Outline subtasks or tasks > Decide which specific steps need to be completed and who is responsible for each of them. - [ ] Output pre-rendered JS version of web componet - [ ] store WC in CDN - [ ] append versioning to the ouput file name - [ ] ensure stability of delivered asset - [ ] have solution for polyfill ## Definition of “Done” (exit criteria) > The story is generally “done” when the user can complete the outlined task, but make sure to define what that is. An initial Auro web component has been updated with functionality to export a consumable version of the WC via a CDN.
priority
cdn ready components user story as a web developer i want to be able to consume auro components w o having to pre compile code so i can easily use resources in a non dynamic environment outline subtasks or tasks decide which specific steps need to be completed and who is responsible for each of them output pre rendered js version of web componet store wc in cdn append versioning to the ouput file name ensure stability of delivered asset have solution for polyfill definition of “done” exit criteria the story is generally “done” when the user can complete the outlined task but make sure to define what that is an initial auro web component has been updated with functionality to export a consumable version of the wc via a cdn
1
126,977
26,956,700,015
IssuesEvent
2023-02-08 15:21:45
arduino/arduino-ide
https://api.github.com/repos/arduino/arduino-ide
reopened
Spurious output in Serial Monitor after upload
conclusion: resolved topic: code type: imperfection topic: CLI topic: serial monitor criticality: low
### Describe the problem The first output printed to Serial Monitor after an upload is some "garbage" characters. ![image](https://user-images.githubusercontent.com/8572152/159872636-d8e0c920-e628-4c91-9c91-65611fcfcaaa.png) ### To reproduce #### Equipment - Arduino board that uses the ATmega16U2 USB chip: - Arduino Uno - Arduino Uno Mini - Arduino Mega #### Steps 1. Upload a sketch to your board that prints to `Serial`: ```cpp void setup() { Serial.begin(9600); Serial.println("hello world"); } void loop() {} ``` 1. Open the "**Serial Monitor**" view. 🐛 The output starts with some unexpected content: ```text □□□□□□□□□□□hello world ``` ### Expected behavior Serial monitor output always reflects the data sent by the board. ### Arduino IDE version 2.0.0-rc5-snapshot-4de7737 ### Operating system Windows ### Operating system version 10 ### Additional context The number of `□` in the demonstration matches the number of characters that are printed by the sketch. --- The spurious output does not occur when the output is triggered by resetting the board, so it is specific to the upload operation. --- The spurious output still occurs even if you add a delay before the print, so it is not a matter of timing from initialization as is sometimes the cause of corrupted serial output. --- The issue does not occur when using Arduino IDE 1.x --- I have only been able to reproduce this issue with the Arduino boards that use an ATmega16U2 USB to serial adapter chip. I could not reproduce it when using boards with other USB interfaces: - Native USB (Leonardo, Nano 33 IoT) - FTDI FT232R (Nano, Pro Mini) - WCH CH340 (3rd party boards) --- Originally reported at https://forum.arduino.cc/t/serial-monitor-contains-garbage-after-upload/972312 ### Issue checklist - [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=) - [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds) - [X] My report contains all necessary details
1.0
Spurious output in Serial Monitor after upload - ### Describe the problem The first output printed to Serial Monitor after an upload is some "garbage" characters. ![image](https://user-images.githubusercontent.com/8572152/159872636-d8e0c920-e628-4c91-9c91-65611fcfcaaa.png) ### To reproduce #### Equipment - Arduino board that uses the ATmega16U2 USB chip: - Arduino Uno - Arduino Uno Mini - Arduino Mega #### Steps 1. Upload a sketch to your board that prints to `Serial`: ```cpp void setup() { Serial.begin(9600); Serial.println("hello world"); } void loop() {} ``` 1. Open the "**Serial Monitor**" view. 🐛 The output starts with some unexpected content: ```text □□□□□□□□□□□hello world ``` ### Expected behavior Serial monitor output always reflects the data sent by the board. ### Arduino IDE version 2.0.0-rc5-snapshot-4de7737 ### Operating system Windows ### Operating system version 10 ### Additional context The number of `□` in the demonstration matches the number of characters that are printed by the sketch. --- The spurious output does not occur when the output is triggered by resetting the board, so it is specific to the upload operation. --- The spurious output still occurs even if you add a delay before the print, so it is not a matter of timing from initialization as is sometimes the cause of corrupted serial output. --- The issue does not occur when using Arduino IDE 1.x --- I have only been able to reproduce this issue with the Arduino boards that use an ATmega16U2 USB to serial adapter chip. I could not reproduce it when using boards with other USB interfaces: - Native USB (Leonardo, Nano 33 IoT) - FTDI FT232R (Nano, Pro Mini) - WCH CH340 (3rd party boards) --- Originally reported at https://forum.arduino.cc/t/serial-monitor-contains-garbage-after-upload/972312 ### Issue checklist - [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=) - [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds) - [X] My report contains all necessary details
non_priority
spurious output in serial monitor after upload describe the problem the first output printed to serial monitor after an upload is some garbage characters to reproduce equipment arduino board that uses the usb chip arduino uno arduino uno mini arduino mega steps upload a sketch to your board that prints to serial cpp void setup serial begin serial println hello world void loop open the serial monitor view 🐛 the output starts with some unexpected content text □□□□□□□□□□□hello world expected behavior serial monitor output always reflects the data sent by the board arduino ide version snapshot operating system windows operating system version additional context the number of □ in the demonstration matches the number of characters that are printed by the sketch the spurious output does not occur when the output is triggered by resetting the board so it is specific to the upload operation the spurious output still occurs even if you add a delay before the print so it is not a matter of timing from initialization as is sometimes the cause of corrupted serial output the issue does not occur when using arduino ide x i have only been able to reproduce this issue with the arduino boards that use an usb to serial adapter chip i could not reproduce it when using boards with other usb interfaces native usb leonardo nano iot ftdi nano pro mini wch party boards originally reported at issue checklist i searched for previous reports in i verified the problem still occurs when using the latest my report contains all necessary details
0
106,226
13,256,458,785
IssuesEvent
2020-08-20 12:43:52
fremtind/jokul
https://api.github.com/repos/fremtind/jokul
closed
Loader: Legges til i Figma-biblioteket
🎨 design
Komponenten eksisterer kun i kode (og i After Effects) Må også gjenspeiles i Figmabiblioteket. Bør samtidig vurdere en visuell oppdatering. Spesielt i tilfellene der den blir brukt som loader i en knapp/pilleform
1.0
Loader: Legges til i Figma-biblioteket - Komponenten eksisterer kun i kode (og i After Effects) Må også gjenspeiles i Figmabiblioteket. Bør samtidig vurdere en visuell oppdatering. Spesielt i tilfellene der den blir brukt som loader i en knapp/pilleform
non_priority
loader legges til i figma biblioteket komponenten eksisterer kun i kode og i after effects må også gjenspeiles i figmabiblioteket bør samtidig vurdere en visuell oppdatering spesielt i tilfellene der den blir brukt som loader i en knapp pilleform
0
18,160
25,086,834,980
IssuesEvent
2022-11-08 00:58:58
Automattic/woocommerce-subscriptions-core
https://api.github.com/repos/Automattic/woocommerce-subscriptions-core
closed
[HPOS] The billing address index field is duplicated in subscription meta
type: bug compatibility: HPOS
## Describe the bug <!-- A clear and concise description of what the bug is. Please be as descriptive as possible, and include screenshots to illustrate. --> On a store with HPOS tables enabled with no sync (this may also impact stores with syncing on) the billing address index is duplicated in the `wp_wc_orders_meta` table. eg: <img width="699" alt="Screen Shot 2022-11-07 at 10 09 40 am" src="https://user-images.githubusercontent.com/8490476/200203146-d4092704-1eea-41b6-b61b-b2487d8f5083.png"> ## To Reproduce <!-- Describe the steps to reproduce the behavior. --> 1. Turn on HPOS and disable syncing. See screenshot below of HPOS settings below. 2. Purchase a subscription using BACs. 3. In the database look at the `wp_wc_orders_meta` meta and notice the subscription has two billing index fields. <img width="769" alt="Screen Shot 2022-11-07 at 10 13 46 am" src="https://user-images.githubusercontent.com/8490476/200203381-c44b3a38-86f5-4ed4-a621-795eed19c2bd.png"> ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> There should only be 1 billing index. ### Actual behavior <!-- A clear and concise description of what actually happens. --> There are 2 billing indexes and one of them is incomplete.
True
[HPOS] The billing address index field is duplicated in subscription meta - ## Describe the bug <!-- A clear and concise description of what the bug is. Please be as descriptive as possible, and include screenshots to illustrate. --> On a store with HPOS tables enabled with no sync (this may also impact stores with syncing on) the billing address index is duplicated in the `wp_wc_orders_meta` table. eg: <img width="699" alt="Screen Shot 2022-11-07 at 10 09 40 am" src="https://user-images.githubusercontent.com/8490476/200203146-d4092704-1eea-41b6-b61b-b2487d8f5083.png"> ## To Reproduce <!-- Describe the steps to reproduce the behavior. --> 1. Turn on HPOS and disable syncing. See screenshot below of HPOS settings below. 2. Purchase a subscription using BACs. 3. In the database look at the `wp_wc_orders_meta` meta and notice the subscription has two billing index fields. <img width="769" alt="Screen Shot 2022-11-07 at 10 13 46 am" src="https://user-images.githubusercontent.com/8490476/200203381-c44b3a38-86f5-4ed4-a621-795eed19c2bd.png"> ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> There should only be 1 billing index. ### Actual behavior <!-- A clear and concise description of what actually happens. --> There are 2 billing indexes and one of them is incomplete.
non_priority
the billing address index field is duplicated in subscription meta describe the bug on a store with hpos tables enabled with no sync this may also impact stores with syncing on the billing address index is duplicated in the wp wc orders meta table eg img width alt screen shot at am src to reproduce turn on hpos and disable syncing see screenshot below of hpos settings below purchase a subscription using bacs in the database look at the wp wc orders meta meta and notice the subscription has two billing index fields img width alt screen shot at am src expected behavior there should only be billing index actual behavior there are billing indexes and one of them is incomplete
0
371,527
10,968,265,982
IssuesEvent
2019-11-28 11:14:33
ONEARMY/community-platform
https://api.github.com/repos/ONEARMY/community-platform
opened
[bug] How-Tos not loading
Priority: Urgent❕❕❕ Type:Bug🐛
**Description** The How-To's are still not loading (I already cleared cache) on Windows 10, on Mozilla **Link** https://community.preciousplastic.com/how-to **Screenshots** **User contact** mariska Sure, I'm happy to help!
1.0
[bug] How-Tos not loading - **Description** The How-To's are still not loading (I already cleared cache) on Windows 10, on Mozilla **Link** https://community.preciousplastic.com/how-to **Screenshots** **User contact** mariska Sure, I'm happy to help!
priority
how tos not loading description the how to s are still not loading i already cleared cache on windows on mozilla link screenshots user contact mariska sure i m happy to help
1
509,077
14,712,381,742
IssuesEvent
2021-01-05 08:52:25
wmeints/fizzylogic
https://api.github.com/repos/wmeints/fizzylogic
closed
Enhance the markdown pipeline
priority:high
This issue is my place to collect problems with the markdown pipeline in the website. Right now it's a basic pipeline. I've already spotted a few things missing: * Figures need a caption based on the alt text for the image * Code needs a wrapper element to correctly align
1.0
Enhance the markdown pipeline - This issue is my place to collect problems with the markdown pipeline in the website. Right now it's a basic pipeline. I've already spotted a few things missing: * Figures need a caption based on the alt text for the image * Code needs a wrapper element to correctly align
priority
enhance the markdown pipeline this issue is my place to collect problems with the markdown pipeline in the website right now it s a basic pipeline i ve already spotted a few things missing figures need a caption based on the alt text for the image code needs a wrapper element to correctly align
1
91,918
26,519,812,842
IssuesEvent
2023-01-19 00:55:26
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
opened
[Connector Builder] Set default values for start and end datetimes
team/extensibility ui/connector-builder
## Context It is very common for users to use `{{ now_utc() }}` as the value for the `end_datetime` field in order to keep their data fully up-to-date. Until https://github.com/airbytehq/airbyte/issues/21571 is complete (which makes this field optional), the connector builder should automatically set this as the value of the end_datetime field, as this will be the value users usually want anyway and it will also teach them that this macro exists. Similarly, we should set a sensible default for the `start_datetime` field so that users don't need to configure that unless they want to use a specific value. ## Acceptance Criteria - `end_datetime` is prefilled with a value of `{{ now_utc() }}` when the datetime stream slicer is selected in the builder UI - `start_datetime` is prefilled with a value of now_utc minus 1 year when the datetime stream slicer is selected in the builder UI Open question: - Is now - 1 year the right time range to use as the default for start_datetime? Or would a month or a day be more sensible?
1.0
[Connector Builder] Set default values for start and end datetimes - ## Context It is very common for users to use `{{ now_utc() }}` as the value for the `end_datetime` field in order to keep their data fully up-to-date. Until https://github.com/airbytehq/airbyte/issues/21571 is complete (which makes this field optional), the connector builder should automatically set this as the value of the end_datetime field, as this will be the value users usually want anyway and it will also teach them that this macro exists. Similarly, we should set a sensible default for the `start_datetime` field so that users don't need to configure that unless they want to use a specific value. ## Acceptance Criteria - `end_datetime` is prefilled with a value of `{{ now_utc() }}` when the datetime stream slicer is selected in the builder UI - `start_datetime` is prefilled with a value of now_utc minus 1 year when the datetime stream slicer is selected in the builder UI Open question: - Is now - 1 year the right time range to use as the default for start_datetime? Or would a month or a day be more sensible?
non_priority
set default values for start and end datetimes context it is very common for users to use now utc as the value for the end datetime field in order to keep their data fully up to date until is complete which makes this field optional the connector builder should automatically set this as the value of the end datetime field as this will be the value users usually want anyway and it will also teach them that this macro exists similarly we should set a sensible default for the start datetime field so that users don t need to configure that unless they want to use a specific value acceptance criteria end datetime is prefilled with a value of now utc when the datetime stream slicer is selected in the builder ui start datetime is prefilled with a value of now utc minus year when the datetime stream slicer is selected in the builder ui open question is now year the right time range to use as the default for start datetime or would a month or a day be more sensible
0
513,963
14,930,038,469
IssuesEvent
2021-01-25 01:42:00
SickChill/SickChill
https://api.github.com/repos/SickChill/SickChill
opened
GoFundMe
Help wanted Priority HIGH
So, my laptop is broken, and I have been unable to do anything for the project other than what I can do on the web interface on GitHub for the past several months, and I'm not sure when I might be able to fix that situation. It was suggested that I create a GoFundMe (which I kind of detest when people do this) and someone told me it would be a good idea. There is more info about my situation on the GoFundMe, so if you are feeling generous or you can afford to help I appreciate it. If not, I appreciate you still. Thanks. https://www.gofundme.com/manage/laptop-for-open-source-developer-miigotu
1.0
GoFundMe - So, my laptop is broken, and I have been unable to do anything for the project other than what I can do on the web interface on GitHub for the past several months, and I'm not sure when I might be able to fix that situation. It was suggested that I create a GoFundMe (which I kind of detest when people do this) and someone told me it would be a good idea. There is more info about my situation on the GoFundMe, so if you are feeling generous or you can afford to help I appreciate it. If not, I appreciate you still. Thanks. https://www.gofundme.com/manage/laptop-for-open-source-developer-miigotu
priority
gofundme so my laptop is broken and i have been unable to do anything for the project other than what i can do on the web interface on github for the past several months and i m not sure when i might be able to fix that situation it was suggested that i create a gofundme which i kind of detest when people do this and someone told me it would be a good idea there is more info about my situation on the gofundme so if you are feeling generous or you can afford to help i appreciate it if not i appreciate you still thanks
1
135,403
19,570,285,907
IssuesEvent
2022-01-04 09:04:54
hcengineering/anticrm
https://api.github.com/repos/hcengineering/anticrm
closed
Add additional way to expand/collapse the side menu
Design
Enable the user to expand/collapse the menu by clicking on the empty space in the side bar (marked with red color). ![image](https://user-images.githubusercontent.com/22766382/147695919-9ef86edf-1535-4a48-944d-0226830b58c5.png) ![image](https://user-images.githubusercontent.com/22766382/147695944-d8739c16-aac7-450b-984d-ea4dc63f6fa2.png)
1.0
Add additional way to expand/collapse the side menu - Enable the user to expand/collapse the menu by clicking on the empty space in the side bar (marked with red color). ![image](https://user-images.githubusercontent.com/22766382/147695919-9ef86edf-1535-4a48-944d-0226830b58c5.png) ![image](https://user-images.githubusercontent.com/22766382/147695944-d8739c16-aac7-450b-984d-ea4dc63f6fa2.png)
non_priority
add additional way to expand collapse the side menu enable the user to expand collapse the menu by clicking on the empty space in the side bar marked with red color
0
697,744
23,951,634,927
IssuesEvent
2022-09-12 12:02:09
feast-dev/feast
https://api.github.com/repos/feast-dev/feast
opened
AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query' using google-cloud-bigquery 2.1.0
kind/bug priority/p2
## Expected Behavior I have found a corner case trying to install feast as dependency in a project. Poetry solved `google-cloud-bigquery` to `2.1.0` when installing `feast[gcp,redis]==0.24.1` along with other dependencies. But I was not able to use feast after installing it. ## Current Behavior I got the following exception when running the command `feast apply`: ```python Traceback (most recent call last): File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/importer.py", line 26, in import_class module = importlib.import_module(module_name) File "/Users/<hidden>/.pyenv/versions/3.9.13/lib/python3.9/importlib/__init__.py", line 127, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1030, in _gcd_import File "<frozen importlib._bootstrap>", line 1007, in _find_and_load File "<frozen importlib._bootstrap>", line 986, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 680, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 850, in exec_module File "<frozen importlib._bootstrap>", line 228, in _call_with_frames_removed File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/infra/offline_stores/bigquery.py", line 378, in <module> class BigQueryRetrievalJob(RetrievalJob): File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/infra/offline_stores/bigquery.py", line 478, in BigQueryRetrievalJob ) -> Optional[bigquery.job.query.QueryJob]: AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query' The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/bin/feast", line 8, in <module> sys.exit(cli()) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1130, in __call__ return self.main(*args, **kwargs) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1055, in main rv = self.invoke(ctx) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1657, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 760, in invoke return __callback(*args, **kwargs) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/decorators.py", line 26, in new_func return f(get_current_context(), *args, **kwargs) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/cli.py", line 517, in apply_total_command repo_config = load_repo_config(repo, fs_yaml_file) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 499, in load_repo_config c = RepoConfig(**raw_config) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 166, in __init__ super().__init__(**data) File "pydantic/main.py", line 340, in pydantic.main.BaseModel.__init__ File "pydantic/main.py", line 1050, in pydantic.main.validate_model File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 345, in _validate_offline_store_config offline_config_class = get_offline_config_from_type(offline_store_type) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 480, in get_offline_config_from_type return import_class(module_name, config_class_name, config_class_name) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/importer.py", line 31, in import_class raise FeastModuleImportError(module_name, class_name) from e feast.errors.FeastModuleImportError: Could not import module 'feast.infra.offline_stores.bigquery' while attempting to load class 'BigQueryOfflineStoreConfig' ``` This exception happens because Feast's `bigquery` module imports the class `google.cloud.bigquery.job.query.QueryJob`, but that path doesn't exist in the `google-cloud-bigquery 2.1.0`. https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/bigquery.py#L478 ```python @log_exceptions_and_usage def _execute_query( self, query, job_config=None, timeout: int = 1800 ) -> Optional[bigquery.job.query.QueryJob]: ``` I was able to use Feast only when `google-cloud-bigquery` was upgraded to higher versions (e.g. `2.36.0`), but I think Feast should not allow the resolution of a dependency version that is not working properly. ## Steps to reproduce ``` poetry init -n poetry add google-cloud-bigquery==2.1.0 poetry add --extras gcp,bigquery feast==0.24.1 ``` After installing the requirements you can try to create some feature views from BQ or just check if that path exists in such version. ``` poetry run python ``` ```python >>> from google.cloud import bigquery >>> bigquery.job.query.QueryJob Traceback (most recent call last): File "<stdin>", line 1, in <module> AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query' ``` ### Specifications - Version: 0.24.1 - Platform: Python 3.9 on MacOS - Subsystem: ## Possible Solution This problem could possibly be solved by: 1) Increasing the minimium version allowed for this dependency 2) Changing the import statement from `bigquery.job.query.QueryJob` to `bigquery.job.QueryJob`
1.0
AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query' using google-cloud-bigquery 2.1.0 - ## Expected Behavior I have found a corner case trying to install feast as dependency in a project. Poetry solved `google-cloud-bigquery` to `2.1.0` when installing `feast[gcp,redis]==0.24.1` along with other dependencies. But I was not able to use feast after installing it. ## Current Behavior I got the following exception when running the command `feast apply`: ```python Traceback (most recent call last): File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/importer.py", line 26, in import_class module = importlib.import_module(module_name) File "/Users/<hidden>/.pyenv/versions/3.9.13/lib/python3.9/importlib/__init__.py", line 127, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1030, in _gcd_import File "<frozen importlib._bootstrap>", line 1007, in _find_and_load File "<frozen importlib._bootstrap>", line 986, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 680, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 850, in exec_module File "<frozen importlib._bootstrap>", line 228, in _call_with_frames_removed File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/infra/offline_stores/bigquery.py", line 378, in <module> class BigQueryRetrievalJob(RetrievalJob): File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/infra/offline_stores/bigquery.py", line 478, in BigQueryRetrievalJob ) -> Optional[bigquery.job.query.QueryJob]: AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query' The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/bin/feast", line 8, in <module> sys.exit(cli()) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1130, in __call__ return self.main(*args, **kwargs) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1055, in main rv = self.invoke(ctx) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1657, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/core.py", line 760, in invoke return __callback(*args, **kwargs) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/click/decorators.py", line 26, in new_func return f(get_current_context(), *args, **kwargs) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/cli.py", line 517, in apply_total_command repo_config = load_repo_config(repo, fs_yaml_file) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 499, in load_repo_config c = RepoConfig(**raw_config) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 166, in __init__ super().__init__(**data) File "pydantic/main.py", line 340, in pydantic.main.BaseModel.__init__ File "pydantic/main.py", line 1050, in pydantic.main.validate_model File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 345, in _validate_offline_store_config offline_config_class = get_offline_config_from_type(offline_store_type) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/repo_config.py", line 480, in get_offline_config_from_type return import_class(module_name, config_class_name, config_class_name) File "/Users/<hidden>/Library/Caches/pypoetry/virtualenvs/feast-example-zjuEIhwY-py3.9/lib/python3.9/site-packages/feast/importer.py", line 31, in import_class raise FeastModuleImportError(module_name, class_name) from e feast.errors.FeastModuleImportError: Could not import module 'feast.infra.offline_stores.bigquery' while attempting to load class 'BigQueryOfflineStoreConfig' ``` This exception happens because Feast's `bigquery` module imports the class `google.cloud.bigquery.job.query.QueryJob`, but that path doesn't exist in the `google-cloud-bigquery 2.1.0`. https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/bigquery.py#L478 ```python @log_exceptions_and_usage def _execute_query( self, query, job_config=None, timeout: int = 1800 ) -> Optional[bigquery.job.query.QueryJob]: ``` I was able to use Feast only when `google-cloud-bigquery` was upgraded to higher versions (e.g. `2.36.0`), but I think Feast should not allow the resolution of a dependency version that is not working properly. ## Steps to reproduce ``` poetry init -n poetry add google-cloud-bigquery==2.1.0 poetry add --extras gcp,bigquery feast==0.24.1 ``` After installing the requirements you can try to create some feature views from BQ or just check if that path exists in such version. ``` poetry run python ``` ```python >>> from google.cloud import bigquery >>> bigquery.job.query.QueryJob Traceback (most recent call last): File "<stdin>", line 1, in <module> AttributeError: module 'google.cloud.bigquery.job' has no attribute 'query' ``` ### Specifications - Version: 0.24.1 - Platform: Python 3.9 on MacOS - Subsystem: ## Possible Solution This problem could possibly be solved by: 1) Increasing the minimium version allowed for this dependency 2) Changing the import statement from `bigquery.job.query.QueryJob` to `bigquery.job.QueryJob`
priority
attributeerror module google cloud bigquery job has no attribute query using google cloud bigquery expected behavior i have found a corner case trying to install feast as dependency in a project poetry solved google cloud bigquery to when installing feast along with other dependencies but i was not able to use feast after installing it current behavior i got the following exception when running the command feast apply python traceback most recent call last file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast importer py line in import class module importlib import module module name file users pyenv versions lib importlib init py line in import module return bootstrap gcd import name package level file line in gcd import file line in find and load file line in find and load unlocked file line in load unlocked file line in exec module file line in call with frames removed file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast infra offline stores bigquery py line in class bigqueryretrievaljob retrievaljob file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast infra offline stores bigquery py line in bigqueryretrievaljob optional attributeerror module google cloud bigquery job has no attribute query the above exception was the direct cause of the following exception traceback most recent call last file users library caches pypoetry virtualenvs feast example zjueihwy bin feast line in sys exit cli file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click core py line in call return self main args kwargs file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click core py line in main rv self invoke ctx file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click core py line in invoke return ctx invoke self callback ctx params file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click core py line in invoke return callback args kwargs file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages click decorators py line in new func return f get current context args kwargs file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast cli py line in apply total command repo config load repo config repo fs yaml file file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast repo config py line in load repo config c repoconfig raw config file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast repo config py line in init super init data file pydantic main py line in pydantic main basemodel init file pydantic main py line in pydantic main validate model file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast repo config py line in validate offline store config offline config class get offline config from type offline store type file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast repo config py line in get offline config from type return import class module name config class name config class name file users library caches pypoetry virtualenvs feast example zjueihwy lib site packages feast importer py line in import class raise feastmoduleimporterror module name class name from e feast errors feastmoduleimporterror could not import module feast infra offline stores bigquery while attempting to load class bigqueryofflinestoreconfig this exception happens because feast s bigquery module imports the class google cloud bigquery job query queryjob but that path doesn t exist in the google cloud bigquery python log exceptions and usage def execute query self query job config none timeout int optional i was able to use feast only when google cloud bigquery was upgraded to higher versions e g but i think feast should not allow the resolution of a dependency version that is not working properly steps to reproduce poetry init n poetry add google cloud bigquery poetry add extras gcp bigquery feast after installing the requirements you can try to create some feature views from bq or just check if that path exists in such version poetry run python python from google cloud import bigquery bigquery job query queryjob traceback most recent call last file line in attributeerror module google cloud bigquery job has no attribute query specifications version platform python on macos subsystem possible solution this problem could possibly be solved by increasing the minimium version allowed for this dependency changing the import statement from bigquery job query queryjob to bigquery job queryjob
1
20,535
2,622,852,322
IssuesEvent
2015-03-04 08:05:55
max99x/pagemon-chrome-ext
https://api.github.com/repos/max99x/pagemon-chrome-ext
closed
Detecting the time of the problem(检测时间的问题 )
auto-migrated Priority-Medium
``` What steps will reproduce the problem? Download something or play games. Usually will be the case!(在下载东西的时候或者玩游戏的时候。平时也会出现这 样的情况!) What is the expected output? What do you see instead? Please see the QQ截图20121204201211.jpg!! What version of the Chrome are you using? On what operating system? 版本 23.0.1266.0 canary and 360极速浏览器(内核:Branch of Chromium 21.0.1180.89). system:windows 7 ultimate What are the error log messages? To view them, go to Wrench -> Tools -> Extensions -> Developer Mode -> expand Page Monitor -> "background.htm" -> Console. Please see the QQ截图20121204203115.jpg!!! Please provide any additional information below. Hello, I have a question to ask!When I use your plug-in to detect the update status of a page?I choose the interval is 10 seconds, sometimes I state into the plug-page view to the "detected last more than 10 seconds, in which case I downloaded the file did not download something also occurkinds of situations! !This is BUG it?I really do not understand! ! ! !(你好,我有一个问题想问!当我用你的插件去检测一个网页的更新状态时?我选择的间隔时间是10秒,有时候我进插件的状态页面查看到的"上次检测"大于10秒,这种情况是在我下载文件时发生的,没下载东西时也会发生这种情况!!这是BUG吗?真不明白!!!!) ``` Original issue reported on code.google.com by `qq25592...@gmail.com` on 4 Dec 2012 at 12:34 Attachments: * [QQ截图20121204201211.jpg](https://storage.googleapis.com/google-code-attachments/pagemon-chrome-ext/issue-220/comment-0/QQ截图20121204201211.jpg) * [QQ截图20121204203115.jpg](https://storage.googleapis.com/google-code-attachments/pagemon-chrome-ext/issue-220/comment-0/QQ截图20121204203115.jpg)
1.0
Detecting the time of the problem(检测时间的问题 ) - ``` What steps will reproduce the problem? Download something or play games. Usually will be the case!(在下载东西的时候或者玩游戏的时候。平时也会出现这 样的情况!) What is the expected output? What do you see instead? Please see the QQ截图20121204201211.jpg!! What version of the Chrome are you using? On what operating system? 版本 23.0.1266.0 canary and 360极速浏览器(内核:Branch of Chromium 21.0.1180.89). system:windows 7 ultimate What are the error log messages? To view them, go to Wrench -> Tools -> Extensions -> Developer Mode -> expand Page Monitor -> "background.htm" -> Console. Please see the QQ截图20121204203115.jpg!!! Please provide any additional information below. Hello, I have a question to ask!When I use your plug-in to detect the update status of a page?I choose the interval is 10 seconds, sometimes I state into the plug-page view to the "detected last more than 10 seconds, in which case I downloaded the file did not download something also occurkinds of situations! !This is BUG it?I really do not understand! ! ! !(你好,我有一个问题想问!当我用你的插件去检测一个网页的更新状态时?我选择的间隔时间是10秒,有时候我进插件的状态页面查看到的"上次检测"大于10秒,这种情况是在我下载文件时发生的,没下载东西时也会发生这种情况!!这是BUG吗?真不明白!!!!) ``` Original issue reported on code.google.com by `qq25592...@gmail.com` on 4 Dec 2012 at 12:34 Attachments: * [QQ截图20121204201211.jpg](https://storage.googleapis.com/google-code-attachments/pagemon-chrome-ext/issue-220/comment-0/QQ截图20121204201211.jpg) * [QQ截图20121204203115.jpg](https://storage.googleapis.com/google-code-attachments/pagemon-chrome-ext/issue-220/comment-0/QQ截图20121204203115.jpg)
priority
detecting the time of the problem(检测时间的问题 ) what steps will reproduce the problem download something or play games usually will be the case (在下载东西的时候或者玩游戏的时候。平时也会出现这 样的情况!) what is the expected output what do you see instead please see the jpg what version of the chrome are you using on what operating system 版本 canary and (内核:branch of chromium ) system windows ultimate what are the error log messages to view them go to wrench tools extensions developer mode expand page monitor background htm console please see the jpg!!! please provide any additional information below hello i have a question to ask when i use your plug in to detect the update status of a page i choose the interval is seconds sometimes i state into the plug page view to the detected last more than seconds in which case i downloaded the file did not download something also occurkinds of situations this is bug it i really do not understand (你好,我有一个问题想问!当我用你的插件去检测一个网页的更新状态时? ,有时候我进插件的状态页面查看到的 上次检测 这种情况是在我下载文件时发生的,没下载东西时也会发生这种情况!!这是bug吗?真不明白!!!!) original issue reported on code google com by gmail com on dec at attachments
1
604,772
18,718,506,613
IssuesEvent
2021-11-03 09:04:23
magento/magento2-phpstorm-plugin
https://api.github.com/repos/magento/magento2-phpstorm-plugin
opened
[UCT:Inspection] Importing non-existent interface
Award: special achievement Priority: P1 Upgrade Compatibility Tool
### Description The purpose of this inspection is to find places where there is imported interface that is removed in any of the next released versions. **For example:** The client has Magento V1 and wants to upgrade it to the V2 (versions are abstract to describe the idea). Some class has the TestInterface in imports. In the V2 that interface is removed. That inspection should highlight importing of this interface and describe the issue. **The UML diagram of the example case:** ![importing-non-existent-interface](https://user-images.githubusercontent.com/31848341/140033054-101a2084-dc62-417a-a64e-268795df65d3.jpg) ### Key points for implementation - all PHP inspections should be grouped under the `com.magento.idea.magento2uct.inspections` package. - all PHP inspections must be inherited from the com.jetbrains.php.lang.inspections.PhpInspection - all inspections should have checking if it should be run with the current tool configurations - all inspections should cover cases for running from the UCT tool run option and from the PHPStorm itself - if covered area is not already supported it should be checked and added to the `com.magento.idea.magento2uct.inspections.UctInspectionManager` - each new inspection must be registered in PHPStorm way in the plugin.xml file - each new inspection must be described in the appropriate html file: `resources/inspectionDescriptions/FileName.html` (to be available in the inspections enable/disable menu) - each new inspection must be registered for running from the UCT tool run option in the `com.magento.idea.magento2uct.packages.SupportedIssue` Approximate time to develop: 2h ~ 3h
1.0
[UCT:Inspection] Importing non-existent interface - ### Description The purpose of this inspection is to find places where there is imported interface that is removed in any of the next released versions. **For example:** The client has Magento V1 and wants to upgrade it to the V2 (versions are abstract to describe the idea). Some class has the TestInterface in imports. In the V2 that interface is removed. That inspection should highlight importing of this interface and describe the issue. **The UML diagram of the example case:** ![importing-non-existent-interface](https://user-images.githubusercontent.com/31848341/140033054-101a2084-dc62-417a-a64e-268795df65d3.jpg) ### Key points for implementation - all PHP inspections should be grouped under the `com.magento.idea.magento2uct.inspections` package. - all PHP inspections must be inherited from the com.jetbrains.php.lang.inspections.PhpInspection - all inspections should have checking if it should be run with the current tool configurations - all inspections should cover cases for running from the UCT tool run option and from the PHPStorm itself - if covered area is not already supported it should be checked and added to the `com.magento.idea.magento2uct.inspections.UctInspectionManager` - each new inspection must be registered in PHPStorm way in the plugin.xml file - each new inspection must be described in the appropriate html file: `resources/inspectionDescriptions/FileName.html` (to be available in the inspections enable/disable menu) - each new inspection must be registered for running from the UCT tool run option in the `com.magento.idea.magento2uct.packages.SupportedIssue` Approximate time to develop: 2h ~ 3h
priority
importing non existent interface description the purpose of this inspection is to find places where there is imported interface that is removed in any of the next released versions for example the client has magento and wants to upgrade it to the versions are abstract to describe the idea some class has the testinterface in imports in the that interface is removed that inspection should highlight importing of this interface and describe the issue the uml diagram of the example case key points for implementation all php inspections should be grouped under the com magento idea inspections package all php inspections must be inherited from the com jetbrains php lang inspections phpinspection all inspections should have checking if it should be run with the current tool configurations all inspections should cover cases for running from the uct tool run option and from the phpstorm itself if covered area is not already supported it should be checked and added to the com magento idea inspections uctinspectionmanager each new inspection must be registered in phpstorm way in the plugin xml file each new inspection must be described in the appropriate html file resources inspectiondescriptions filename html to be available in the inspections enable disable menu each new inspection must be registered for running from the uct tool run option in the com magento idea packages supportedissue approximate time to develop
1
160,494
13,791,077,852
IssuesEvent
2020-10-09 11:31:57
ai4er-cdt/resources
https://api.github.com/repos/ai4er-cdt/resources
closed
Create cookiecutter template for AI4ER
documentation enhancement
[] Create cookiecutter template for AI4ER [] add environment instatiation [] explain [] link to wiki
1.0
Create cookiecutter template for AI4ER - [] Create cookiecutter template for AI4ER [] add environment instatiation [] explain [] link to wiki
non_priority
create cookiecutter template for create cookiecutter template for add environment instatiation explain link to wiki
0
16,407
2,615,040,262
IssuesEvent
2015-03-01 03:34:54
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
opened
Java client for Google Maps Data API
auto-migrated Priority-Medium Type-Sample
``` Which Google API and version (e.g. Google Calendar Data API version 2)? Google Maps Data API What format (e.g. JSON, Atom)? JSON What Authentation (e.g. OAuth, OAuth 2, ClientLogin)? ClientLogin Java environment (e.g. Java 6, Android 2.3, App Engine)? Java 6 External references, such as API reference guide? Please provide any additional information below. thank you! ``` Original issue reported on code.google.com by `vincenzo...@gmail.com` on 20 Apr 2012 at 12:43
1.0
Java client for Google Maps Data API - ``` Which Google API and version (e.g. Google Calendar Data API version 2)? Google Maps Data API What format (e.g. JSON, Atom)? JSON What Authentation (e.g. OAuth, OAuth 2, ClientLogin)? ClientLogin Java environment (e.g. Java 6, Android 2.3, App Engine)? Java 6 External references, such as API reference guide? Please provide any additional information below. thank you! ``` Original issue reported on code.google.com by `vincenzo...@gmail.com` on 20 Apr 2012 at 12:43
priority
java client for google maps data api which google api and version e g google calendar data api version google maps data api what format e g json atom json what authentation e g oauth oauth clientlogin clientlogin java environment e g java android app engine java external references such as api reference guide please provide any additional information below thank you original issue reported on code google com by vincenzo gmail com on apr at
1
66,211
6,992,933,006
IssuesEvent
2017-12-15 09:20:41
owncloud/core
https://api.github.com/repos/owncloud/core
closed
UI tests should be skippable for specific oC versions
1 - To develop dev:acceptance-tests QA
### Steps to reproduce 1. run UI tests on older versions of oC (e.g. 10.0.3) 2. tests that check for bugs that are fixed in master/stable10 branch but not in the tested release are failing ### Expected behaviour it should be possible to skip tests that are known to fail in a specific version. e.g. `@skipBeforeOC10.0.4` @phil-davis @patrickjahns @davitol @SergioBertolinSG comments?
1.0
UI tests should be skippable for specific oC versions - ### Steps to reproduce 1. run UI tests on older versions of oC (e.g. 10.0.3) 2. tests that check for bugs that are fixed in master/stable10 branch but not in the tested release are failing ### Expected behaviour it should be possible to skip tests that are known to fail in a specific version. e.g. `@skipBeforeOC10.0.4` @phil-davis @patrickjahns @davitol @SergioBertolinSG comments?
non_priority
ui tests should be skippable for specific oc versions steps to reproduce run ui tests on older versions of oc e g tests that check for bugs that are fixed in master branch but not in the tested release are failing expected behaviour it should be possible to skip tests that are known to fail in a specific version e g phil davis patrickjahns davitol sergiobertolinsg comments
0
475,020
13,685,955,981
IssuesEvent
2020-09-30 07:59:17
incognitochain/incognito-chain
https://api.github.com/repos/incognitochain/incognito-chain
closed
Got both output coin v1 and v2 after sending transactions
Priority: High Type: Bug
related to: #963 While running regression tests of sending output coins and init coins, got following error: `Cannot get tx version because there are 2 versions of input coins\nCan not create tx`
1.0
Got both output coin v1 and v2 after sending transactions - related to: #963 While running regression tests of sending output coins and init coins, got following error: `Cannot get tx version because there are 2 versions of input coins\nCan not create tx`
priority
got both output coin and after sending transactions related to while running regression tests of sending output coins and init coins got following error cannot get tx version because there are versions of input coins ncan not create tx
1
459,465
13,193,714,850
IssuesEvent
2020-08-13 15:38:28
IHTSDO/snomed-owl-toolkit
https://api.github.com/repos/IHTSDO/snomed-owl-toolkit
closed
Silenced out of memory error
bug medium-priority
Hi all, the last week I created an issue because I was not able to generate the owl ontology file for the last version of Snomed (#37). The generated owl showed an incorrect hierarchy of concepts, where most of them were outside the Snomed root concept. This was provoked because there was an out of memory error, which was silenced in the process. I am doing experiments with the different versions of snomed along the time, and I repeated the process by assigning 4GB of memory. Then, I noticed that the owl files generated with 2GB of memory were, in general, different to the files generated with 4GB. For example, the following figure shows the snomed 2019-01-31 owl file generated with 2Gb and with 2GB: ![imagen](https://user-images.githubusercontent.com/11376639/74641157-21e48480-5171-11ea-9de6-e37180820307.png) As you can see, annotations are created incorrectly, but they could be legit so I did not detect the error. Would it be possible to show this out of memory error, at least in the log?
1.0
Silenced out of memory error - Hi all, the last week I created an issue because I was not able to generate the owl ontology file for the last version of Snomed (#37). The generated owl showed an incorrect hierarchy of concepts, where most of them were outside the Snomed root concept. This was provoked because there was an out of memory error, which was silenced in the process. I am doing experiments with the different versions of snomed along the time, and I repeated the process by assigning 4GB of memory. Then, I noticed that the owl files generated with 2GB of memory were, in general, different to the files generated with 4GB. For example, the following figure shows the snomed 2019-01-31 owl file generated with 2Gb and with 2GB: ![imagen](https://user-images.githubusercontent.com/11376639/74641157-21e48480-5171-11ea-9de6-e37180820307.png) As you can see, annotations are created incorrectly, but they could be legit so I did not detect the error. Would it be possible to show this out of memory error, at least in the log?
priority
silenced out of memory error hi all the last week i created an issue because i was not able to generate the owl ontology file for the last version of snomed the generated owl showed an incorrect hierarchy of concepts where most of them were outside the snomed root concept this was provoked because there was an out of memory error which was silenced in the process i am doing experiments with the different versions of snomed along the time and i repeated the process by assigning of memory then i noticed that the owl files generated with of memory were in general different to the files generated with for example the following figure shows the snomed owl file generated with and with as you can see annotations are created incorrectly but they could be legit so i did not detect the error would it be possible to show this out of memory error at least in the log
1
194,126
6,891,769,381
IssuesEvent
2017-11-22 18:15:08
vmware/vic
https://api.github.com/repos/vmware/vic
closed
Image pull from artifactory fails: Failed to obtain OAuth endpoint: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403
kind/bug priority/high team/container
After successful login to artifcatory repo, image pull fails: ``` Nov 9 2017 14:58:46.410Z INFO PullImage: reference: vic-docker-local.artifactory.eng.vmware.com/busybox:1, 127.0.0.1:2377, portlayer: "127.0.0.1:2377" Nov 9 2017 14:58:46.410Z INFO Using UUID (564d1c2f-2a64-983b-1ec9-4185f86f38c7) for imagestore name Nov 9 2017 14:58:46.410Z DEBUG Running with portlayer Nov 9 2017 14:58:46.410Z DEBUG [BEGIN] [vic/lib/imagec.PingPortLayer:40] 127.0.0.1:2377 Nov 9 2017 14:58:46.411Z DEBUG [ END ] [vic/lib/imagec.PingPortLayer:40] [385.047µs] 127.0.0.1:2377 Nov 9 2017 14:58:46.411Z DEBUG [BEGIN] [vic/lib/imagec.LearnRegistryURL:75] vic-docker-local.artifactory.eng.vmware.com Nov 9 2017 14:58:46.411Z DEBUG Trying https scheme for &imagec.Options{Reference:(*reference.taggedRef)(0xc420653e60), Registry:"vic-docker-local.artifactory.eng.vmware.com", Image:"busybox", Tag:"1", Destination:"/tmp", Host:"127.0.0.1:2377", Storename:"564d1c2f-2a64-983b-1ec9-4185f86f38c7", Username:"vic-deployer", Password:"vmware!123", Token:(*fetcher.Token)(nil), Timeout:3600000000000, Outstream:(*ioutils.WriteFlusher)(0xc42059fa40), InsecureSkipVerify:false, InsecureAllowHTTP:true, ImageManifestSchema1:(*imagec.Manifest)(nil), ImageManifestSchema2:(*schema2.DeserializedManifest)(nil), ManifestDigest:"", RegistryCAs:(*x509.CertPool)(0xc4207daf30), Standalone:false, ImageStore:""} Nov 9 2017 14:58:46.411Z DEBUG URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/ Nov 9 2017 14:58:46.550Z DEBUG [ END ] [vic/lib/imagec.LearnRegistryURL:75] [138.852068ms] vic-docker-local.artifactory.eng.vmware.com Nov 9 2017 14:58:46.550Z DEBUG [BEGIN] [vic/lib/imagec.LearnAuthURL:100] vic-docker-local.artifactory.eng.vmware.com/busybox:1 Nov 9 2017 14:58:46.550Z DEBUG URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 Nov 9 2017 14:58:46.587Z DEBUG [ END ] [vic/lib/imagec.LearnAuthURL:100] [37.614669ms] vic-docker-local.artifactory.eng.vmware.com/busybox:1 Nov 9 2017 14:58:46.587Z INFO https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403, URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 Nov 9 2017 14:58:46.587Z DEBUG [ END ] [vic/lib/apiservers/engine/backends.(*Image).PullImage:322] [226.636362ms] Nov 9 2017 14:58:46.587Z ERROR Handler for POST /v1.21/images/create returned error: Failed to obtain OAuth endpoint: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403, URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 ``` [21-2-Artifactory-VCH-0-7754-container-logs.zip](https://github.com/vmware/vic/files/1458276/21-2-Artifactory-VCH-0-7754-container-logs.zip)
1.0
Image pull from artifactory fails: Failed to obtain OAuth endpoint: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403 - After successful login to artifcatory repo, image pull fails: ``` Nov 9 2017 14:58:46.410Z INFO PullImage: reference: vic-docker-local.artifactory.eng.vmware.com/busybox:1, 127.0.0.1:2377, portlayer: "127.0.0.1:2377" Nov 9 2017 14:58:46.410Z INFO Using UUID (564d1c2f-2a64-983b-1ec9-4185f86f38c7) for imagestore name Nov 9 2017 14:58:46.410Z DEBUG Running with portlayer Nov 9 2017 14:58:46.410Z DEBUG [BEGIN] [vic/lib/imagec.PingPortLayer:40] 127.0.0.1:2377 Nov 9 2017 14:58:46.411Z DEBUG [ END ] [vic/lib/imagec.PingPortLayer:40] [385.047µs] 127.0.0.1:2377 Nov 9 2017 14:58:46.411Z DEBUG [BEGIN] [vic/lib/imagec.LearnRegistryURL:75] vic-docker-local.artifactory.eng.vmware.com Nov 9 2017 14:58:46.411Z DEBUG Trying https scheme for &imagec.Options{Reference:(*reference.taggedRef)(0xc420653e60), Registry:"vic-docker-local.artifactory.eng.vmware.com", Image:"busybox", Tag:"1", Destination:"/tmp", Host:"127.0.0.1:2377", Storename:"564d1c2f-2a64-983b-1ec9-4185f86f38c7", Username:"vic-deployer", Password:"vmware!123", Token:(*fetcher.Token)(nil), Timeout:3600000000000, Outstream:(*ioutils.WriteFlusher)(0xc42059fa40), InsecureSkipVerify:false, InsecureAllowHTTP:true, ImageManifestSchema1:(*imagec.Manifest)(nil), ImageManifestSchema2:(*schema2.DeserializedManifest)(nil), ManifestDigest:"", RegistryCAs:(*x509.CertPool)(0xc4207daf30), Standalone:false, ImageStore:""} Nov 9 2017 14:58:46.411Z DEBUG URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/ Nov 9 2017 14:58:46.550Z DEBUG [ END ] [vic/lib/imagec.LearnRegistryURL:75] [138.852068ms] vic-docker-local.artifactory.eng.vmware.com Nov 9 2017 14:58:46.550Z DEBUG [BEGIN] [vic/lib/imagec.LearnAuthURL:100] vic-docker-local.artifactory.eng.vmware.com/busybox:1 Nov 9 2017 14:58:46.550Z DEBUG URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 Nov 9 2017 14:58:46.587Z DEBUG [ END ] [vic/lib/imagec.LearnAuthURL:100] [37.614669ms] vic-docker-local.artifactory.eng.vmware.com/busybox:1 Nov 9 2017 14:58:46.587Z INFO https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403, URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 Nov 9 2017 14:58:46.587Z DEBUG [ END ] [vic/lib/apiservers/engine/backends.(*Image).PullImage:322] [226.636362ms] Nov 9 2017 14:58:46.587Z ERROR Handler for POST /v1.21/images/create returned error: Failed to obtain OAuth endpoint: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 returned an unexpected response: Unexpected http code: 403, URL: https://vic-docker-local.artifactory.eng.vmware.com/v2/busybox/manifests/1 ``` [21-2-Artifactory-VCH-0-7754-container-logs.zip](https://github.com/vmware/vic/files/1458276/21-2-Artifactory-VCH-0-7754-container-logs.zip)
priority
image pull from artifactory fails failed to obtain oauth endpoint returned an unexpected response unexpected http code after successful login to artifcatory repo image pull fails nov info pullimage reference vic docker local artifactory eng vmware com busybox portlayer nov info using uuid for imagestore name nov debug running with portlayer nov debug nov debug nov debug vic docker local artifactory eng vmware com nov debug trying https scheme for imagec options reference reference taggedref registry vic docker local artifactory eng vmware com image busybox tag destination tmp host storename username vic deployer password vmware token fetcher token nil timeout outstream ioutils writeflusher insecureskipverify false insecureallowhttp true imagec manifest nil deserializedmanifest nil manifestdigest registrycas certpool standalone false imagestore nov debug url nov debug vic docker local artifactory eng vmware com nov debug vic docker local artifactory eng vmware com busybox nov debug url nov debug vic docker local artifactory eng vmware com busybox nov info returned an unexpected response unexpected http code url nov debug nov error handler for post images create returned error failed to obtain oauth endpoint returned an unexpected response unexpected http code url
1
128,805
5,076,210,360
IssuesEvent
2016-12-28 00:38:47
NUKnightLab/StoryMapJS
https://api.github.com/repos/NUKnightLab/StoryMapJS
closed
New Users cannot sign in
bug priority: high
Likely introduced with changes to Google Drive. New users who have not before logged into StoryMap are receiving the error: Error getting migration list: 'google' Currently, no workaround is known
1.0
New Users cannot sign in - Likely introduced with changes to Google Drive. New users who have not before logged into StoryMap are receiving the error: Error getting migration list: 'google' Currently, no workaround is known
priority
new users cannot sign in likely introduced with changes to google drive new users who have not before logged into storymap are receiving the error error getting migration list google currently no workaround is known
1
156,943
5,995,231,607
IssuesEvent
2017-06-03 01:25:07
kmycode/gunucco
https://api.github.com/repos/kmycode/gunucco
closed
List-up and display private or user-only chapters on web client view page when user logining
enhancement priority-high
I forgot adding #32.
1.0
List-up and display private or user-only chapters on web client view page when user logining - I forgot adding #32.
priority
list up and display private or user only chapters on web client view page when user logining i forgot adding
1