Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
144,430
| 11,616,182,295
|
IssuesEvent
|
2020-02-26 15:19:50
|
godaddy-wordpress/coblocks
|
https://api.github.com/repos/godaddy-wordpress/coblocks
|
closed
|
ISBAT ensure the integrity of the Gallery Masonry block through automated tests
|
[Type] Tests
|
Automated tests need to be added to the `gallery-masonry` block which follow the groundwork from #835.
#### Tests Required:
- [ ] The save function
- [x] Block transforms #1200
- [x] Block deprecation #967
#### AC
- Any attribute that modifies the serialized block output needs to have tests.
- Any existing transforms to or from a block needs to have tests (if applicable).
- Any existing block deprecations need to have tests (if applicable).
|
1.0
|
ISBAT ensure the integrity of the Gallery Masonry block through automated tests - Automated tests need to be added to the `gallery-masonry` block which follow the groundwork from #835.
#### Tests Required:
- [ ] The save function
- [x] Block transforms #1200
- [x] Block deprecation #967
#### AC
- Any attribute that modifies the serialized block output needs to have tests.
- Any existing transforms to or from a block needs to have tests (if applicable).
- Any existing block deprecations need to have tests (if applicable).
|
non_defect
|
isbat ensure the integrity of the gallery masonry block through automated tests automated tests need to be added to the gallery masonry block which follow the groundwork from tests required the save function block transforms block deprecation ac any attribute that modifies the serialized block output needs to have tests any existing transforms to or from a block needs to have tests if applicable any existing block deprecations need to have tests if applicable
| 0
|
418,492
| 12,198,594,795
|
IssuesEvent
|
2020-04-29 23:15:44
|
CESARBR/knot-babeltower
|
https://api.github.com/repos/CESARBR/knot-babeltower
|
closed
|
Add auth device
|
enhancement priority: medium
|
As a platform user, I want to authenticate my device on the fog.
- [ ] Add event listener to authenticate a device
- [ ] Authenticate the devive on the things service
- [ ] Update documentation
|
1.0
|
Add auth device - As a platform user, I want to authenticate my device on the fog.
- [ ] Add event listener to authenticate a device
- [ ] Authenticate the devive on the things service
- [ ] Update documentation
|
non_defect
|
add auth device as a platform user i want to authenticate my device on the fog add event listener to authenticate a device authenticate the devive on the things service update documentation
| 0
|
700,502
| 24,062,877,490
|
IssuesEvent
|
2022-09-17 03:58:02
|
carpentries/amy
|
https://api.github.com/repos/carpentries/amy
|
closed
|
Require dates for role
|
priority: essential
|
For all community roles, dates should be required.
If inactive, both the start and end date should be required. ~~Additionally, the current date can not be in between the start and end dates.~~
If active, the start date should be required. The end date should display as `present` rather than as `???`.
EDIT @pbanaszkiewicz to reflect that the end date can be in the future.
|
1.0
|
Require dates for role - For all community roles, dates should be required.
If inactive, both the start and end date should be required. ~~Additionally, the current date can not be in between the start and end dates.~~
If active, the start date should be required. The end date should display as `present` rather than as `???`.
EDIT @pbanaszkiewicz to reflect that the end date can be in the future.
|
non_defect
|
require dates for role for all community roles dates should be required if inactive both the start and end date should be required additionally the current date can not be in between the start and end dates if active the start date should be required the end date should display as present rather than as edit pbanaszkiewicz to reflect that the end date can be in the future
| 0
|
70,671
| 23,281,992,292
|
IssuesEvent
|
2022-08-05 13:01:00
|
vector-im/element-call
|
https://api.github.com/repos/vector-im/element-call
|
opened
|
Scrolling the bottom participants videos in spotlight also scrolls the main "spotlight" video
|
T-Defect
|
### Steps to reproduce
1. Where are you starting? What can you see?
Group call with more than 4 ppl
2. What do you click?
I click on the spotlight view and then scroll the list of videos horizontally on the botton.
### Outcome
#### What did you expect?
I expected the big highlighted video to stay where it is and all the bottom videos should move with my finger.
#### What happened instead?
The list scrolls (my video does not scroll which is covered in a different issue: https://github.com/vector-im/element-call/issues/398) but the big video of the talking person at the top does scroll as well.
### Operating system
Android
### Browser information
Chrome
### URL for webapp
element-call-sfu-simon.netlify.app
### Will you send logs?
No
|
1.0
|
Scrolling the bottom participants videos in spotlight also scrolls the main "spotlight" video - ### Steps to reproduce
1. Where are you starting? What can you see?
Group call with more than 4 ppl
2. What do you click?
I click on the spotlight view and then scroll the list of videos horizontally on the botton.
### Outcome
#### What did you expect?
I expected the big highlighted video to stay where it is and all the bottom videos should move with my finger.
#### What happened instead?
The list scrolls (my video does not scroll which is covered in a different issue: https://github.com/vector-im/element-call/issues/398) but the big video of the talking person at the top does scroll as well.
### Operating system
Android
### Browser information
Chrome
### URL for webapp
element-call-sfu-simon.netlify.app
### Will you send logs?
No
|
defect
|
scrolling the bottom participants videos in spotlight also scrolls the main spotlight video steps to reproduce where are you starting what can you see group call with more than ppl what do you click i click on the spotlight view and then scroll the list of videos horizontally on the botton outcome what did you expect i expected the big highlighted video to stay where it is and all the bottom videos should move with my finger what happened instead the list scrolls my video does not scroll which is covered in a different issue but the big video of the talking person at the top does scroll as well operating system android browser information chrome url for webapp element call sfu simon netlify app will you send logs no
| 1
|
48,509
| 13,106,233,127
|
IssuesEvent
|
2020-08-04 13:32:36
|
hazelcast/hazelcast-jet
|
https://api.github.com/repos/hazelcast/hazelcast-jet
|
closed
|
Job with long stage name and metrics after job completion doesn't terminate
|
defect metrics
|
Reproducer:
```java
JetInstance jet = Jet.newJetInstance();
Pipeline p = Pipeline.create();
p.readFrom(TestSources.items(1))
.writeTo(Sinks.noop()).setName("verylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongname");
Job job = jet.newJob(p, new JobConfig().setStoreMetricsAfterJobCompletion(true));
job.join();
```
|
1.0
|
Job with long stage name and metrics after job completion doesn't terminate - Reproducer:
```java
JetInstance jet = Jet.newJetInstance();
Pipeline p = Pipeline.create();
p.readFrom(TestSources.items(1))
.writeTo(Sinks.noop()).setName("verylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongname");
Job job = jet.newJob(p, new JobConfig().setStoreMetricsAfterJobCompletion(true));
job.join();
```
|
defect
|
job with long stage name and metrics after job completion doesn t terminate reproducer java jetinstance jet jet newjetinstance pipeline p pipeline create p readfrom testsources items writeto sinks noop setname verylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongnameverylongname job job jet newjob p new jobconfig setstoremetricsafterjobcompletion true job join
| 1
|
10,805
| 2,622,190,855
|
IssuesEvent
|
2015-03-04 00:22:59
|
byzhang/cudpp
|
https://api.github.com/repos/byzhang/cudpp
|
closed
|
cudppSort error for a large array
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. tar -xzvf sort_test.tar.gz
2. cd sort_test
3. make
4. ./testsort 1000000
What is the expected output? What do you see instead?
expected :
before sort
radix sort : 0.00833379 s 1000000 elements
what I see :
before sort
radix sort : 0.00833379 s 1000000 elements
sort error 4 720476 541723
What version of the product are you using? On what operating system?
Using device 0: Quadroplex 2200 S4
Quadroplex 2200 S4; global mem: 4294705152B; compute v1.3; clock: 1296000
kHz
cudpp 1.1.1
CUDA SDK 2.3
on linux 2.6 kernel
$uname -a
Linux tesla 2.6.18-128.1.1.el5 #1 SMP Tue Feb 10 11:36:29 EST 2009 x86_64
x86_64 x86_64 GNU/Linux
$cat /proc/driver/nvidia/version
NVRM version: NVIDIA UNIX x86_64 Kernel Module 190.53 Wed Dec 9
15:29:46 PST 2009
GCC version: gcc version 4.1.2 20080704 (Red Hat 4.1.2-44)
Please provide any additional information below.
This is a simple test to use cudppSort.
For a small array, it passes the test, but for a large array, it fails.
It also fails in cudpp_testrig as follows.
$./cudpp_testrig -sort -n=1000000
Using device 0: Quadroplex 2200 S4
Quadroplex 2200 S4; global mem: 4294705152B; compute v1.3; clock: 1296000
kHz
Running a sort of 1000000 unsigned int key-value pairs
Unordered key[3]:746051 > key[4]:16173
Incorrectly sorted value[0] (583160) 1153083146 != 460036
GPU test FAILED
Average execution time: 8.024296 ms
1 tests failed
If this is a driver version mismatch, please let me know which driver
version is needed. Thank you.
```
Original issue reported on code.google.com by `Eunjin...@gmail.com` on 30 Mar 2010 at 12:06
* Merged into: #51
Attachments:
* [sort_test.tar.gz](https://storage.googleapis.com/google-code-attachments/cudpp/issue-50/comment-0/sort_test.tar.gz)
|
1.0
|
cudppSort error for a large array - ```
What steps will reproduce the problem?
1. tar -xzvf sort_test.tar.gz
2. cd sort_test
3. make
4. ./testsort 1000000
What is the expected output? What do you see instead?
expected :
before sort
radix sort : 0.00833379 s 1000000 elements
what I see :
before sort
radix sort : 0.00833379 s 1000000 elements
sort error 4 720476 541723
What version of the product are you using? On what operating system?
Using device 0: Quadroplex 2200 S4
Quadroplex 2200 S4; global mem: 4294705152B; compute v1.3; clock: 1296000
kHz
cudpp 1.1.1
CUDA SDK 2.3
on linux 2.6 kernel
$uname -a
Linux tesla 2.6.18-128.1.1.el5 #1 SMP Tue Feb 10 11:36:29 EST 2009 x86_64
x86_64 x86_64 GNU/Linux
$cat /proc/driver/nvidia/version
NVRM version: NVIDIA UNIX x86_64 Kernel Module 190.53 Wed Dec 9
15:29:46 PST 2009
GCC version: gcc version 4.1.2 20080704 (Red Hat 4.1.2-44)
Please provide any additional information below.
This is a simple test to use cudppSort.
For a small array, it passes the test, but for a large array, it fails.
It also fails in cudpp_testrig as follows.
$./cudpp_testrig -sort -n=1000000
Using device 0: Quadroplex 2200 S4
Quadroplex 2200 S4; global mem: 4294705152B; compute v1.3; clock: 1296000
kHz
Running a sort of 1000000 unsigned int key-value pairs
Unordered key[3]:746051 > key[4]:16173
Incorrectly sorted value[0] (583160) 1153083146 != 460036
GPU test FAILED
Average execution time: 8.024296 ms
1 tests failed
If this is a driver version mismatch, please let me know which driver
version is needed. Thank you.
```
Original issue reported on code.google.com by `Eunjin...@gmail.com` on 30 Mar 2010 at 12:06
* Merged into: #51
Attachments:
* [sort_test.tar.gz](https://storage.googleapis.com/google-code-attachments/cudpp/issue-50/comment-0/sort_test.tar.gz)
|
defect
|
cudppsort error for a large array what steps will reproduce the problem tar xzvf sort test tar gz cd sort test make testsort what is the expected output what do you see instead expected before sort radix sort s elements what i see before sort radix sort s elements sort error what version of the product are you using on what operating system using device quadroplex quadroplex global mem compute clock khz cudpp cuda sdk on linux kernel uname a linux tesla smp tue feb est gnu linux cat proc driver nvidia version nvrm version nvidia unix kernel module wed dec pst gcc version gcc version red hat please provide any additional information below this is a simple test to use cudppsort for a small array it passes the test but for a large array it fails it also fails in cudpp testrig as follows cudpp testrig sort n using device quadroplex quadroplex global mem compute clock khz running a sort of unsigned int key value pairs unordered key key incorrectly sorted value gpu test failed average execution time ms tests failed if this is a driver version mismatch please let me know which driver version is needed thank you original issue reported on code google com by eunjin gmail com on mar at merged into attachments
| 1
|
632,467
| 20,198,050,887
|
IssuesEvent
|
2022-02-11 12:34:29
|
therealbluepandabear/PyxlMoose
|
https://api.github.com/repos/therealbluepandabear/PyxlMoose
|
closed
|
[Improvement] Add clear labeling to the 'FindAndReplaceFragment'
|
low priority improvement
|
#### Improvement description
Add clear labelling to the 'FindAndReplaceFragment' showing which color will be found, and which color it will be replaced with.
#### Why is this improvement important to add?
Because some users are getting confused as to what the two blocks represent at the bottom of the 'FindAndReplaceFragment'.
|
1.0
|
[Improvement] Add clear labeling to the 'FindAndReplaceFragment' - #### Improvement description
Add clear labelling to the 'FindAndReplaceFragment' showing which color will be found, and which color it will be replaced with.
#### Why is this improvement important to add?
Because some users are getting confused as to what the two blocks represent at the bottom of the 'FindAndReplaceFragment'.
|
non_defect
|
add clear labeling to the findandreplacefragment improvement description add clear labelling to the findandreplacefragment showing which color will be found and which color it will be replaced with why is this improvement important to add because some users are getting confused as to what the two blocks represent at the bottom of the findandreplacefragment
| 0
|
19,089
| 11,139,298,272
|
IssuesEvent
|
2019-12-21 03:46:28
|
ritsec/cluster-duck
|
https://api.github.com/repos/ritsec/cluster-duck
|
closed
|
Deploy OpenStack
|
club management new-service
|
Deploy OpenStack
==============
OpenStack needs to be deployed on our local cluster using OpenStack-Ansible.
Depends on
-----
This issue depends on the following issues:
- [x] #4
- [x] #6
- [x] #7
Tasks
-----
All of the following tasks must be complete before this issue can be closed. Be sure to reference this issue in the relevant issues/PRs in other repositories.
- [x] Run OpenStack-Ansible deployment
- [x] Back up secret files to Vault
- [x] Update the RITSEC OpenStack-Ansible repository with any changes
- [x] Explain how to run a deployment of OpenStack-Ansible in documentation
- [ ] Add messages to host pages
|
1.0
|
Deploy OpenStack - Deploy OpenStack
==============
OpenStack needs to be deployed on our local cluster using OpenStack-Ansible.
Depends on
-----
This issue depends on the following issues:
- [x] #4
- [x] #6
- [x] #7
Tasks
-----
All of the following tasks must be complete before this issue can be closed. Be sure to reference this issue in the relevant issues/PRs in other repositories.
- [x] Run OpenStack-Ansible deployment
- [x] Back up secret files to Vault
- [x] Update the RITSEC OpenStack-Ansible repository with any changes
- [x] Explain how to run a deployment of OpenStack-Ansible in documentation
- [ ] Add messages to host pages
|
non_defect
|
deploy openstack deploy openstack openstack needs to be deployed on our local cluster using openstack ansible depends on this issue depends on the following issues tasks all of the following tasks must be complete before this issue can be closed be sure to reference this issue in the relevant issues prs in other repositories run openstack ansible deployment back up secret files to vault update the ritsec openstack ansible repository with any changes explain how to run a deployment of openstack ansible in documentation add messages to host pages
| 0
|
48,033
| 13,067,405,814
|
IssuesEvent
|
2020-07-31 00:21:03
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
[steamshovel] PyArtist.addCleanupAction doesn't work on bound functions (Trac #1687)
|
Migrated from Trac combo core defect
|
Using PyArtists.addCleanupAction with a bound function to the same object doesn't actually result in anything (e.g. writing to a log) happening. A bound function on a different object results in segfaults.
''Example 1: Same object''
```text
class MyArtists(PyArtist):
def __init__(self):
...
self.addCleanupAction(self.doAction)
def doAction(self):
[action]
```
Expected behaviour: [action] is performed when the artists is destroyed.
Actual behaviour: [action] is not performed.
''Example 2: Other object''
```text
class OtherClass:
def doAction(self):
[action]
global_other = OtherClass()
class MyArtists(PyArtist):
def __init__(self):
...
self.addCleanupAction(global_other.doAction)
```
Expected behaviour: [action] is performed when the artists is destroyed.
Actual behaviour: steamshovel segfaults without performing [action].
(Can't run steamshovel on the device I'm submitting this ticket on, but can provide more details later if need be)
Migrated from https://code.icecube.wisc.edu/ticket/1687
```json
{
"status": "closed",
"changetime": "2016-05-19T15:57:35",
"description": "Using PyArtists.addCleanupAction with a bound function to the same object doesn't actually result in anything (e.g. writing to a log) happening. A bound function on a different object results in segfaults.\n\n''Example 1: Same object''\n{{{\nclass MyArtists(PyArtist):\n def __init__(self):\n ...\n self.addCleanupAction(self.doAction)\n\n def doAction(self):\n [action]\n}}}\n\nExpected behaviour: [action] is performed when the artists is destroyed.\nActual behaviour: [action] is not performed.\n\n''Example 2: Other object''\n{{{\nclass OtherClass:\n def doAction(self):\n [action]\n\nglobal_other = OtherClass()\n\nclass MyArtists(PyArtist):\n def __init__(self):\n ...\n self.addCleanupAction(global_other.doAction)\n}}}\n\nExpected behaviour: [action] is performed when the artists is destroyed.\nActual behaviour: steamshovel segfaults without performing [action].\n\n(Can't run steamshovel on the device I'm submitting this ticket on, but can provide more details later if need be)",
"reporter": "sander.vanheule",
"cc": "",
"resolution": "fixed",
"_ts": "1463673455749367",
"component": "combo core",
"summary": "[steamshovel] PyArtist.addCleanupAction doesn't work on bound functions",
"priority": "major",
"keywords": "steamshovel pybindings",
"time": "2016-05-03T18:26:25",
"milestone": "Long-Term Future",
"owner": "hdembinski",
"type": "defect"
}
```
|
1.0
|
[steamshovel] PyArtist.addCleanupAction doesn't work on bound functions (Trac #1687) - Using PyArtists.addCleanupAction with a bound function to the same object doesn't actually result in anything (e.g. writing to a log) happening. A bound function on a different object results in segfaults.
''Example 1: Same object''
```text
class MyArtists(PyArtist):
def __init__(self):
...
self.addCleanupAction(self.doAction)
def doAction(self):
[action]
```
Expected behaviour: [action] is performed when the artists is destroyed.
Actual behaviour: [action] is not performed.
''Example 2: Other object''
```text
class OtherClass:
def doAction(self):
[action]
global_other = OtherClass()
class MyArtists(PyArtist):
def __init__(self):
...
self.addCleanupAction(global_other.doAction)
```
Expected behaviour: [action] is performed when the artists is destroyed.
Actual behaviour: steamshovel segfaults without performing [action].
(Can't run steamshovel on the device I'm submitting this ticket on, but can provide more details later if need be)
Migrated from https://code.icecube.wisc.edu/ticket/1687
```json
{
"status": "closed",
"changetime": "2016-05-19T15:57:35",
"description": "Using PyArtists.addCleanupAction with a bound function to the same object doesn't actually result in anything (e.g. writing to a log) happening. A bound function on a different object results in segfaults.\n\n''Example 1: Same object''\n{{{\nclass MyArtists(PyArtist):\n def __init__(self):\n ...\n self.addCleanupAction(self.doAction)\n\n def doAction(self):\n [action]\n}}}\n\nExpected behaviour: [action] is performed when the artists is destroyed.\nActual behaviour: [action] is not performed.\n\n''Example 2: Other object''\n{{{\nclass OtherClass:\n def doAction(self):\n [action]\n\nglobal_other = OtherClass()\n\nclass MyArtists(PyArtist):\n def __init__(self):\n ...\n self.addCleanupAction(global_other.doAction)\n}}}\n\nExpected behaviour: [action] is performed when the artists is destroyed.\nActual behaviour: steamshovel segfaults without performing [action].\n\n(Can't run steamshovel on the device I'm submitting this ticket on, but can provide more details later if need be)",
"reporter": "sander.vanheule",
"cc": "",
"resolution": "fixed",
"_ts": "1463673455749367",
"component": "combo core",
"summary": "[steamshovel] PyArtist.addCleanupAction doesn't work on bound functions",
"priority": "major",
"keywords": "steamshovel pybindings",
"time": "2016-05-03T18:26:25",
"milestone": "Long-Term Future",
"owner": "hdembinski",
"type": "defect"
}
```
|
defect
|
pyartist addcleanupaction doesn t work on bound functions trac using pyartists addcleanupaction with a bound function to the same object doesn t actually result in anything e g writing to a log happening a bound function on a different object results in segfaults example same object text class myartists pyartist def init self self addcleanupaction self doaction def doaction self expected behaviour is performed when the artists is destroyed actual behaviour is not performed example other object text class otherclass def doaction self global other otherclass class myartists pyartist def init self self addcleanupaction global other doaction expected behaviour is performed when the artists is destroyed actual behaviour steamshovel segfaults without performing can t run steamshovel on the device i m submitting this ticket on but can provide more details later if need be migrated from json status closed changetime description using pyartists addcleanupaction with a bound function to the same object doesn t actually result in anything e g writing to a log happening a bound function on a different object results in segfaults n n example same object n nclass myartists pyartist n def init self n n self addcleanupaction self doaction n n def doaction self n n n nexpected behaviour is performed when the artists is destroyed nactual behaviour is not performed n n example other object n nclass otherclass n def doaction self n n nglobal other otherclass n nclass myartists pyartist n def init self n n self addcleanupaction global other doaction n n nexpected behaviour is performed when the artists is destroyed nactual behaviour steamshovel segfaults without performing n n can t run steamshovel on the device i m submitting this ticket on but can provide more details later if need be reporter sander vanheule cc resolution fixed ts component combo core summary pyartist addcleanupaction doesn t work on bound functions priority major keywords steamshovel pybindings time milestone long term future owner hdembinski type defect
| 1
|
82,196
| 7,833,863,013
|
IssuesEvent
|
2018-06-16 04:42:33
|
scalatra/scalatra
|
https://api.github.com/repos/scalatra/scalatra
|
closed
|
Multi project builds, run `resourceBasePath` different to test one
|
test
|
Hi,
I have a very simple [hello world app](https://github.com/plippe/scalatra-hello-resource). The main purpose of the app was to highlight how to read resource files. Those resources are stored in `src/main/webapp/WEB-INF`, and accessed with `getServletContext.getResource`. This worked perfectly until I tried with a multi project build.
`resourceBasePath` points to `PROJECT/src/main/webapp/WEB-INF` in run, but `src/main/webapp/WEB-INF` in test.
I was able to fix my tests with `servletContextHandler.setResourceBase(s"PROJECT/${resourceBasePath}")`, but I was wondering if this was the expected behaviour.
Kind regards
|
1.0
|
Multi project builds, run `resourceBasePath` different to test one - Hi,
I have a very simple [hello world app](https://github.com/plippe/scalatra-hello-resource). The main purpose of the app was to highlight how to read resource files. Those resources are stored in `src/main/webapp/WEB-INF`, and accessed with `getServletContext.getResource`. This worked perfectly until I tried with a multi project build.
`resourceBasePath` points to `PROJECT/src/main/webapp/WEB-INF` in run, but `src/main/webapp/WEB-INF` in test.
I was able to fix my tests with `servletContextHandler.setResourceBase(s"PROJECT/${resourceBasePath}")`, but I was wondering if this was the expected behaviour.
Kind regards
|
non_defect
|
multi project builds run resourcebasepath different to test one hi i have a very simple the main purpose of the app was to highlight how to read resource files those resources are stored in src main webapp web inf and accessed with getservletcontext getresource this worked perfectly until i tried with a multi project build resourcebasepath points to project src main webapp web inf in run but src main webapp web inf in test i was able to fix my tests with servletcontexthandler setresourcebase s project resourcebasepath but i was wondering if this was the expected behaviour kind regards
| 0
|
50,871
| 13,187,921,739
|
IssuesEvent
|
2020-08-13 05:02:19
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
[recclasses] Non-offline dependencies (Trac #1567)
|
Migrated from Trac combo reconstruction defect
|
Currently recclasses depends on portia (r142454) and ophelia(r142462). This is breakage that needs to be fixed. One of the main reasons for the creation of the project was to provide a project with minimal dependencies that people could include into any meta-project. This project therefore can't have any dependencies outside of offline-software.
Can we remove those dependencies and get whatever fixes into portia and ophelia that are needed so that their tests pass on all platforms?
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1567">https://code.icecube.wisc.edu/ticket/1567</a>, reported by olivas and owned by hdembinski</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "Currently recclasses depends on portia (r142454) and ophelia(r142462). This is breakage that needs to be fixed. One of the main reasons for the creation of the project was to provide a project with minimal dependencies that people could include into any meta-project. This project therefore can't have any dependencies outside of offline-software.\n\nCan we remove those dependencies and get whatever fixes into portia and ophelia that are needed so that their tests pass on all platforms?\n\n",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "[recclasses] Non-offline dependencies",
"priority": "blocker",
"keywords": "",
"time": "2016-02-25T16:34:10",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[recclasses] Non-offline dependencies (Trac #1567) - Currently recclasses depends on portia (r142454) and ophelia(r142462). This is breakage that needs to be fixed. One of the main reasons for the creation of the project was to provide a project with minimal dependencies that people could include into any meta-project. This project therefore can't have any dependencies outside of offline-software.
Can we remove those dependencies and get whatever fixes into portia and ophelia that are needed so that their tests pass on all platforms?
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1567">https://code.icecube.wisc.edu/ticket/1567</a>, reported by olivas and owned by hdembinski</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "Currently recclasses depends on portia (r142454) and ophelia(r142462). This is breakage that needs to be fixed. One of the main reasons for the creation of the project was to provide a project with minimal dependencies that people could include into any meta-project. This project therefore can't have any dependencies outside of offline-software.\n\nCan we remove those dependencies and get whatever fixes into portia and ophelia that are needed so that their tests pass on all platforms?\n\n",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "[recclasses] Non-offline dependencies",
"priority": "blocker",
"keywords": "",
"time": "2016-02-25T16:34:10",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
</p>
</details>
|
defect
|
non offline dependencies trac currently recclasses depends on portia and ophelia this is breakage that needs to be fixed one of the main reasons for the creation of the project was to provide a project with minimal dependencies that people could include into any meta project this project therefore can t have any dependencies outside of offline software can we remove those dependencies and get whatever fixes into portia and ophelia that are needed so that their tests pass on all platforms migrated from json status closed changetime description currently recclasses depends on portia and ophelia this is breakage that needs to be fixed one of the main reasons for the creation of the project was to provide a project with minimal dependencies that people could include into any meta project this project therefore can t have any dependencies outside of offline software n ncan we remove those dependencies and get whatever fixes into portia and ophelia that are needed so that their tests pass on all platforms n n reporter olivas cc resolution fixed ts component combo reconstruction summary non offline dependencies priority blocker keywords time milestone owner hdembinski type defect
| 1
|
237,091
| 26,078,788,409
|
IssuesEvent
|
2022-12-25 01:13:30
|
kapseliboi/Node-Data
|
https://api.github.com/repos/kapseliboi/Node-Data
|
opened
|
CVE-2022-23541 (Medium) detected in jsonwebtoken-5.7.0.tgz
|
security vulnerability
|
## CVE-2022-23541 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonwebtoken-5.7.0.tgz</b></p></summary>
<p>JSON Web Token implementation (symmetric and asymmetric)</p>
<p>Library home page: <a href="https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-5.7.0.tgz">https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-5.7.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/jsonwebtoken/package.json</p>
<p>
Dependency Hierarchy:
- :x: **jsonwebtoken-5.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/Node-Data/commit/289c77565fc637d4c0e4bf4a9a1e81df96cd190a">289c77565fc637d4c0e4bf4a9a1e81df96cd190a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jsonwebtoken is an implementation of JSON Web Tokens. Versions `<= 8.5.1` of `jsonwebtoken` library can be misconfigured so that passing a poorly implemented key retrieval function referring to the `secretOrPublicKey` argument from the readme link will result in incorrect verification of tokens. There is a possibility of using a different algorithm and key combination in verification, other than the one that was used to sign the tokens. Specifically, tokens signed with an asymmetric public key could be verified with a symmetric HS256 algorithm. This can lead to successful validation of forged tokens. If your application is supporting usage of both symmetric key and asymmetric key in jwt.verify() implementation with the same key retrieval function. This issue has been patched, please update to version 9.0.0.
<p>Publish Date: 2022-12-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23541>CVE-2022-23541</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/auth0/node-jsonwebtoken/security/advisories/GHSA-hjrf-2m68-5959">https://github.com/auth0/node-jsonwebtoken/security/advisories/GHSA-hjrf-2m68-5959</a></p>
<p>Release Date: 2022-12-22</p>
<p>Fix Resolution: jsonwebtoken - 9.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-23541 (Medium) detected in jsonwebtoken-5.7.0.tgz - ## CVE-2022-23541 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonwebtoken-5.7.0.tgz</b></p></summary>
<p>JSON Web Token implementation (symmetric and asymmetric)</p>
<p>Library home page: <a href="https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-5.7.0.tgz">https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-5.7.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/jsonwebtoken/package.json</p>
<p>
Dependency Hierarchy:
- :x: **jsonwebtoken-5.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/Node-Data/commit/289c77565fc637d4c0e4bf4a9a1e81df96cd190a">289c77565fc637d4c0e4bf4a9a1e81df96cd190a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jsonwebtoken is an implementation of JSON Web Tokens. Versions `<= 8.5.1` of `jsonwebtoken` library can be misconfigured so that passing a poorly implemented key retrieval function referring to the `secretOrPublicKey` argument from the readme link will result in incorrect verification of tokens. There is a possibility of using a different algorithm and key combination in verification, other than the one that was used to sign the tokens. Specifically, tokens signed with an asymmetric public key could be verified with a symmetric HS256 algorithm. This can lead to successful validation of forged tokens. If your application is supporting usage of both symmetric key and asymmetric key in jwt.verify() implementation with the same key retrieval function. This issue has been patched, please update to version 9.0.0.
<p>Publish Date: 2022-12-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23541>CVE-2022-23541</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/auth0/node-jsonwebtoken/security/advisories/GHSA-hjrf-2m68-5959">https://github.com/auth0/node-jsonwebtoken/security/advisories/GHSA-hjrf-2m68-5959</a></p>
<p>Release Date: 2022-12-22</p>
<p>Fix Resolution: jsonwebtoken - 9.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in jsonwebtoken tgz cve medium severity vulnerability vulnerable library jsonwebtoken tgz json web token implementation symmetric and asymmetric library home page a href path to dependency file package json path to vulnerable library node modules jsonwebtoken package json dependency hierarchy x jsonwebtoken tgz vulnerable library found in head commit a href found in base branch master vulnerability details jsonwebtoken is an implementation of json web tokens versions of jsonwebtoken library can be misconfigured so that passing a poorly implemented key retrieval function referring to the secretorpublickey argument from the readme link will result in incorrect verification of tokens there is a possibility of using a different algorithm and key combination in verification other than the one that was used to sign the tokens specifically tokens signed with an asymmetric public key could be verified with a symmetric algorithm this can lead to successful validation of forged tokens if your application is supporting usage of both symmetric key and asymmetric key in jwt verify implementation with the same key retrieval function this issue has been patched please update to version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jsonwebtoken step up your open source security game with mend
| 0
|
16,027
| 2,870,252,322
|
IssuesEvent
|
2015-06-07 00:37:36
|
pdelia/away3d
|
https://api.github.com/repos/pdelia/away3d
|
opened
|
In away3d.loaders.utils.GeometryLibrary.as, _geometryArrayDirty should default to true
|
auto-migrated Priority-Medium Type-Defect
|
#81 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:39Z
```
If try to load a Collada model, can get error if call getGeometry() before
any addGeometry() calls, because _geometryArray has never been updated,
because _geometryArrayDirty defaults to false.
Fix:
private var _geometryArrayDirty:Boolean = true
```
Original issue reported on code.google.com by `dtgriscom@gmail.com` on 23 Nov 2009 at 4:50
|
1.0
|
In away3d.loaders.utils.GeometryLibrary.as, _geometryArrayDirty should default to true - #81 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:39Z
```
If try to load a Collada model, can get error if call getGeometry() before
any addGeometry() calls, because _geometryArray has never been updated,
because _geometryArrayDirty defaults to false.
Fix:
private var _geometryArrayDirty:Boolean = true
```
Original issue reported on code.google.com by `dtgriscom@gmail.com` on 23 Nov 2009 at 4:50
|
defect
|
in loaders utils geometrylibrary as geometryarraydirty should default to true issue by googlecodeexporter created on if try to load a collada model can get error if call getgeometry before any addgeometry calls because geometryarray has never been updated because geometryarraydirty defaults to false fix private var geometryarraydirty boolean true original issue reported on code google com by dtgriscom gmail com on nov at
| 1
|
81,752
| 31,494,577,659
|
IssuesEvent
|
2023-08-31 00:30:49
|
idaholab/malamute
|
https://api.github.com/repos/idaholab/malamute
|
closed
|
Apptainer failure: Cincotti example testing producing warnings that fail as errors.
|
c: infrastructure p: normal t: defect
|
## Bug Description
The Cincotti example tests are failing during apptainer builds (See #93 CIVET results) because warnings are not allowed according to the MALAMUTE `testroot`. Oddly, this hasn't impacted other CI testing thus far, but likely should have caused more widespread failures.
## Steps to Reproduce
Run testing without allowing for warnings.
## Impact
Stops MALAMUTE from having an apptainer and should be resolved.
|
1.0
|
Apptainer failure: Cincotti example testing producing warnings that fail as errors. - ## Bug Description
The Cincotti example tests are failing during apptainer builds (See #93 CIVET results) because warnings are not allowed according to the MALAMUTE `testroot`. Oddly, this hasn't impacted other CI testing thus far, but likely should have caused more widespread failures.
## Steps to Reproduce
Run testing without allowing for warnings.
## Impact
Stops MALAMUTE from having an apptainer and should be resolved.
|
defect
|
apptainer failure cincotti example testing producing warnings that fail as errors bug description the cincotti example tests are failing during apptainer builds see civet results because warnings are not allowed according to the malamute testroot oddly this hasn t impacted other ci testing thus far but likely should have caused more widespread failures steps to reproduce run testing without allowing for warnings impact stops malamute from having an apptainer and should be resolved
| 1
|
304,437
| 23,065,830,420
|
IssuesEvent
|
2022-07-25 13:51:32
|
Gamify-IT/issues
|
https://api.github.com/repos/Gamify-IT/issues
|
closed
|
Meeting with Uwe regarding REST API design
|
documentation
|
Uwe wants to tell us how to design REST APIs
## DoD
- [x] Meeting was planned
- [x] Meeting took place
- [x] Everyone attending it knows how to design REST APIs
- [x] (Optional) A recording of the meeting can be referenced in the docs
|
1.0
|
Meeting with Uwe regarding REST API design - Uwe wants to tell us how to design REST APIs
## DoD
- [x] Meeting was planned
- [x] Meeting took place
- [x] Everyone attending it knows how to design REST APIs
- [x] (Optional) A recording of the meeting can be referenced in the docs
|
non_defect
|
meeting with uwe regarding rest api design uwe wants to tell us how to design rest apis dod meeting was planned meeting took place everyone attending it knows how to design rest apis optional a recording of the meeting can be referenced in the docs
| 0
|
22,346
| 4,790,615,560
|
IssuesEvent
|
2016-10-31 09:22:47
|
kss-node/kss-node
|
https://api.github.com/repos/kss-node/kss-node
|
reopened
|
LESS variables
|
documentation
|
I'm trying to configure my project to use KSSNode. So far I have been able to get it to generate a style guide, but my current issue is I am trying to figure out how to display LESS variables. I have tried finding the solution, but I had no luck. How would I display a LESS variable `@amber: rgb(255,193,7);` like this within my styleguide?
|
1.0
|
LESS variables - I'm trying to configure my project to use KSSNode. So far I have been able to get it to generate a style guide, but my current issue is I am trying to figure out how to display LESS variables. I have tried finding the solution, but I had no luck. How would I display a LESS variable `@amber: rgb(255,193,7);` like this within my styleguide?
|
non_defect
|
less variables i m trying to configure my project to use kssnode so far i have been able to get it to generate a style guide but my current issue is i am trying to figure out how to display less variables i have tried finding the solution but i had no luck how would i display a less variable amber rgb like this within my styleguide
| 0
|
12,600
| 2,711,986,200
|
IssuesEvent
|
2015-04-09 10:37:27
|
codenameone/CodenameOne
|
https://api.github.com/repos/codenameone/CodenameOne
|
closed
|
Changing SpanLabel style via code
|
Priority-Medium Type-Defect
|
Original [issue 1102](https://code.google.com/p/codenameone/issues/detail?id=1102) created by codenameone on 2014-04-08T12:33:37.000Z:
Its not possible to change SpanLabel style via code (for instance when wanting to set Font via code or Foreground color)
Since it apply only to the container and not the actual text beneath it.
probably the getStyle() method should return the a style that implement that each of its setters apply to both the styles of the icon and textarea styles.
<b>What steps will reproduce the problem?</b>
1. Add SpanLabel to a form
2. Try to change its text Font via code
<b>What is the expected output? What do you see instead?</b>
The text font should be change to the new Font, it does not changed.
<b>What version of the product are you using? On what operating system?</b>
Latest source code.
|
1.0
|
Changing SpanLabel style via code - Original [issue 1102](https://code.google.com/p/codenameone/issues/detail?id=1102) created by codenameone on 2014-04-08T12:33:37.000Z:
Its not possible to change SpanLabel style via code (for instance when wanting to set Font via code or Foreground color)
Since it apply only to the container and not the actual text beneath it.
probably the getStyle() method should return the a style that implement that each of its setters apply to both the styles of the icon and textarea styles.
<b>What steps will reproduce the problem?</b>
1. Add SpanLabel to a form
2. Try to change its text Font via code
<b>What is the expected output? What do you see instead?</b>
The text font should be change to the new Font, it does not changed.
<b>What version of the product are you using? On what operating system?</b>
Latest source code.
|
defect
|
changing spanlabel style via code original created by codenameone on its not possible to change spanlabel style via code for instance when wanting to set font via code or foreground color since it apply only to the container and not the actual text beneath it probably the getstyle method should return the a style that implement that each of its setters apply to both the styles of the icon and textarea styles what steps will reproduce the problem add spanlabel to a form try to change its text font via code what is the expected output what do you see instead the text font should be change to the new font it does not changed what version of the product are you using on what operating system latest source code
| 1
|
40,689
| 10,128,216,327
|
IssuesEvent
|
2019-08-01 12:15:13
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
opened
|
Prevent parse error
|
defect
|
This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* CakePHP Version: latest master on local dev
### What you did
Development on a controller and view with some helpers.
### What happened
Got `var_export does not handle circular references` in `src/Core/ObjectRegistry.php` line 147 `$msg .= var_export($existingConfig, true);`

### What you expected to happen
Trying to catch/prevent this parse error from happening during development as this one is really hard to understand and track down, as the stracktrace is completely missing on that one.
|
1.0
|
Prevent parse error - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* CakePHP Version: latest master on local dev
### What you did
Development on a controller and view with some helpers.
### What happened
Got `var_export does not handle circular references` in `src/Core/ObjectRegistry.php` line 147 `$msg .= var_export($existingConfig, true);`

### What you expected to happen
Trying to catch/prevent this parse error from happening during development as this one is really hard to understand and track down, as the stracktrace is completely missing on that one.
|
defect
|
prevent parse error this is a multiple allowed bug enhancement cakephp version latest master on local dev what you did development on a controller and view with some helpers what happened got var export does not handle circular references in src core objectregistry php line msg var export existingconfig true what you expected to happen trying to catch prevent this parse error from happening during development as this one is really hard to understand and track down as the stracktrace is completely missing on that one
| 1
|
47,221
| 13,056,061,533
|
IssuesEvent
|
2020-07-30 03:32:20
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
failure in make tarball (Trac #155)
|
Migrated from Trac cmake defect
|
Adding a debug statement to install_shlib.pl, we get the following ( as an example ) :
system(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . "/" . libCore.so.5, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);
system(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . "/" . libCore.so.5.17, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);
system(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . "/" . /afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/var/db/dports/software/root_5.17.06/5.17.06_2/afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/root-v5.17.06/lib/libCore.so.5.17, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);
Couldn't copy /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib//afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/var/db/dports/software/root_5.17.06/5.17.06_2/afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/root-v5.17.06/lib/libCore.so.5.17 to simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools: No such file or directory at /afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl line 75.
This happens for all libraries because all symlinks ahve absolute paths.
Migrated from https://code.icecube.wisc.edu/ticket/155
```json
{
"status": "closed",
"changetime": "2009-09-30T14:43:59",
"description": "Adding a debug statement to install_shlib.pl, we get the following ( as an example ) :\n\n system(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . \"/\" . libCore.so.5, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);\nsystem(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . \"/\" . libCore.so.5.17, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);\nsystem(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . \"/\" . /afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/var/db/dports/software/root_5.17.06/5.17.06_2/afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/root-v5.17.06/lib/libCore.so.5.17, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);\nCouldn't copy /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib//afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/var/db/dports/software/root_5.17.06/5.17.06_2/afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/root-v5.17.06/lib/libCore.so.5.17 to simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools: No such file or directory at /afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl line 75.\n\n This happens for all libraries because all symlinks ahve absolute paths. ",
"reporter": "pratik",
"cc": "rfranke",
"resolution": "fixed",
"_ts": "1254321839000000",
"component": "cmake",
"summary": "failure in make tarball",
"priority": "normal",
"keywords": "",
"time": "2008-11-19T02:51:38",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
|
1.0
|
failure in make tarball (Trac #155) - Adding a debug statement to install_shlib.pl, we get the following ( as an example ) :
system(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . "/" . libCore.so.5, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);
system(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . "/" . libCore.so.5.17, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);
system(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . "/" . /afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/var/db/dports/software/root_5.17.06/5.17.06_2/afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/root-v5.17.06/lib/libCore.so.5.17, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);
Couldn't copy /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib//afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/var/db/dports/software/root_5.17.06/5.17.06_2/afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/root-v5.17.06/lib/libCore.so.5.17 to simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools: No such file or directory at /afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl line 75.
This happens for all libraries because all symlinks ahve absolute paths.
Migrated from https://code.icecube.wisc.edu/ticket/155
```json
{
"status": "closed",
"changetime": "2009-09-30T14:43:59",
"description": "Adding a debug statement to install_shlib.pl, we get the following ( as an example ) :\n\n system(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . \"/\" . libCore.so.5, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);\nsystem(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . \"/\" . libCore.so.5.17, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);\nsystem(/afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl, /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib . \"/\" . /afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/var/db/dports/software/root_5.17.06/5.17.06_2/afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/root-v5.17.06/lib/libCore.so.5.17, simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools);\nCouldn't copy /afs/ifh.de/group/amanda/icecube/ports/RHEL_5.0_amd64/gcc-4.1.2/I3_PORTS/root-v5.17.06/lib//afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/var/db/dports/software/root_5.17.06/5.17.06_2/afs/ifh.de/group/amanda/icecube/ports/amd64_rhel50/gcc-4.1.1/I3_PORTS/root-v5.17.06/lib/libCore.so.5.17 to simulation.releases.V02-02-05.r0.Linux-x86_64.gcc-4.1.2/lib/tools: No such file or directory at /afs/ifh.de/user/i/iceprod/simulation/V02-02-05/src/cmake/install_shlib.pl line 75.\n\n This happens for all libraries because all symlinks ahve absolute paths. ",
"reporter": "pratik",
"cc": "rfranke",
"resolution": "fixed",
"_ts": "1254321839000000",
"component": "cmake",
"summary": "failure in make tarball",
"priority": "normal",
"keywords": "",
"time": "2008-11-19T02:51:38",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
|
defect
|
failure in make tarball trac adding a debug statement to install shlib pl we get the following as an example system afs ifh de user i iceprod simulation src cmake install shlib pl afs ifh de group amanda icecube ports rhel gcc ports root lib libcore so simulation releases linux gcc lib tools system afs ifh de user i iceprod simulation src cmake install shlib pl afs ifh de group amanda icecube ports rhel gcc ports root lib libcore so simulation releases linux gcc lib tools system afs ifh de user i iceprod simulation src cmake install shlib pl afs ifh de group amanda icecube ports rhel gcc ports root lib afs ifh de group amanda icecube ports gcc ports var db dports software root afs ifh de group amanda icecube ports gcc ports root lib libcore so simulation releases linux gcc lib tools couldn t copy afs ifh de group amanda icecube ports rhel gcc ports root lib afs ifh de group amanda icecube ports gcc ports var db dports software root afs ifh de group amanda icecube ports gcc ports root lib libcore so to simulation releases linux gcc lib tools no such file or directory at afs ifh de user i iceprod simulation src cmake install shlib pl line this happens for all libraries because all symlinks ahve absolute paths migrated from json status closed changetime description adding a debug statement to install shlib pl we get the following as an example n n system afs ifh de user i iceprod simulation src cmake install shlib pl afs ifh de group amanda icecube ports rhel gcc ports root lib libcore so simulation releases linux gcc lib tools nsystem afs ifh de user i iceprod simulation src cmake install shlib pl afs ifh de group amanda icecube ports rhel gcc ports root lib libcore so simulation releases linux gcc lib tools nsystem afs ifh de user i iceprod simulation src cmake install shlib pl afs ifh de group amanda icecube ports rhel gcc ports root lib afs ifh de group amanda icecube ports gcc ports var db dports software root afs ifh de group amanda icecube ports gcc ports root lib libcore so simulation releases linux gcc lib tools ncouldn t copy afs ifh de group amanda icecube ports rhel gcc ports root lib afs ifh de group amanda icecube ports gcc ports var db dports software root afs ifh de group amanda icecube ports gcc ports root lib libcore so to simulation releases linux gcc lib tools no such file or directory at afs ifh de user i iceprod simulation src cmake install shlib pl line n n this happens for all libraries because all symlinks ahve absolute paths reporter pratik cc rfranke resolution fixed ts component cmake summary failure in make tarball priority normal keywords time milestone owner troy type defect
| 1
|
11,476
| 2,652,258,638
|
IssuesEvent
|
2015-03-16 16:23:48
|
JoseExposito/touchegg
|
https://api.github.com/repos/JoseExposito/touchegg
|
closed
|
Touchegg on ubuntu 13.10 with Mate
|
auto-migrated Type-Defect
|
```
I have a Clevo P150SM with a synaptics trackpad and I wish to have multi-touch
gestures for expo, scale, show desktop, change view port, etc...
I am using Ubuntu 13.10 with mate as a desktop environment and have synclient.
I searched the web and found touchegg and I installed it from the repos and it
did not work, then I tried to install it by the source and it did not work.
Touchegg runs on terminal and displays the information of the gestures but then
if I make a gesture it does not do anything.
An example I created a gesture for "all" to make a mouse click (button 1) with
4 fingers I rebooted the PC logged in but it did nothing.
(I have the touchegg in the startup applications).
Does touchegg work in ubuntu 13.10?
Is there something I am missing?
```
Original issue reported on code.google.com by `edsous...@gmail.com` on 25 Mar 2014 at 10:37
|
1.0
|
Touchegg on ubuntu 13.10 with Mate - ```
I have a Clevo P150SM with a synaptics trackpad and I wish to have multi-touch
gestures for expo, scale, show desktop, change view port, etc...
I am using Ubuntu 13.10 with mate as a desktop environment and have synclient.
I searched the web and found touchegg and I installed it from the repos and it
did not work, then I tried to install it by the source and it did not work.
Touchegg runs on terminal and displays the information of the gestures but then
if I make a gesture it does not do anything.
An example I created a gesture for "all" to make a mouse click (button 1) with
4 fingers I rebooted the PC logged in but it did nothing.
(I have the touchegg in the startup applications).
Does touchegg work in ubuntu 13.10?
Is there something I am missing?
```
Original issue reported on code.google.com by `edsous...@gmail.com` on 25 Mar 2014 at 10:37
|
defect
|
touchegg on ubuntu with mate i have a clevo with a synaptics trackpad and i wish to have multi touch gestures for expo scale show desktop change view port etc i am using ubuntu with mate as a desktop environment and have synclient i searched the web and found touchegg and i installed it from the repos and it did not work then i tried to install it by the source and it did not work touchegg runs on terminal and displays the information of the gestures but then if i make a gesture it does not do anything an example i created a gesture for all to make a mouse click button with fingers i rebooted the pc logged in but it did nothing i have the touchegg in the startup applications does touchegg work in ubuntu is there something i am missing original issue reported on code google com by edsous gmail com on mar at
| 1
|
47,792
| 13,066,231,447
|
IssuesEvent
|
2020-07-30 21:15:50
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
spline-reco documentaion improvements (Trac #1184)
|
Migrated from Trac combo reconstruction defect
|
docs don't list maintainer and no link to doxygen documentaion
release notes don't contain trunk section
the only example script dosn't use test data and instead relies on user input for input filenames
Migrated from https://code.icecube.wisc.edu/ticket/1184
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "docs don't list maintainer and no link to doxygen documentaion\nrelease notes don't contain trunk section\nthe only example script dosn't use test data and instead relies on user input for input filenames",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "spline-reco documentaion improvements",
"priority": "blocker",
"keywords": "",
"time": "2015-08-19T13:14:01",
"milestone": "",
"owner": "mvoge",
"type": "defect"
}
```
|
1.0
|
spline-reco documentaion improvements (Trac #1184) - docs don't list maintainer and no link to doxygen documentaion
release notes don't contain trunk section
the only example script dosn't use test data and instead relies on user input for input filenames
Migrated from https://code.icecube.wisc.edu/ticket/1184
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "docs don't list maintainer and no link to doxygen documentaion\nrelease notes don't contain trunk section\nthe only example script dosn't use test data and instead relies on user input for input filenames",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "spline-reco documentaion improvements",
"priority": "blocker",
"keywords": "",
"time": "2015-08-19T13:14:01",
"milestone": "",
"owner": "mvoge",
"type": "defect"
}
```
|
defect
|
spline reco documentaion improvements trac docs don t list maintainer and no link to doxygen documentaion release notes don t contain trunk section the only example script dosn t use test data and instead relies on user input for input filenames migrated from json status closed changetime description docs don t list maintainer and no link to doxygen documentaion nrelease notes don t contain trunk section nthe only example script dosn t use test data and instead relies on user input for input filenames reporter kjmeagher cc resolution fixed ts component combo reconstruction summary spline reco documentaion improvements priority blocker keywords time milestone owner mvoge type defect
| 1
|
237,087
| 19,592,847,316
|
IssuesEvent
|
2022-01-05 14:46:15
|
theislab/scvelo
|
https://api.github.com/repos/theislab/scvelo
|
closed
|
Unit test `neighbors.py`
|
enhancement testing
|
<!-- What kind of feature would you like to request? -->
## Description
For a more robust and reliable codebase, the code in `neighbors.py` needs to be unit tested. This is one step to achieving #226.
|
1.0
|
Unit test `neighbors.py` - <!-- What kind of feature would you like to request? -->
## Description
For a more robust and reliable codebase, the code in `neighbors.py` needs to be unit tested. This is one step to achieving #226.
|
non_defect
|
unit test neighbors py description for a more robust and reliable codebase the code in neighbors py needs to be unit tested this is one step to achieving
| 0
|
16,661
| 2,925,158,117
|
IssuesEvent
|
2015-06-26 02:11:38
|
FreeRADIUS/freeradius-server
|
https://api.github.com/repos/FreeRADIUS/freeradius-server
|
closed
|
Problems with radmin
|
defect v3.0.x v3.1.x
|
Hi,
I noticed that the radmin doesn't works in the v3.0.x/HEAD
1) The debug condition does nothing.
radmin> debug condition '(User-Name != "jorge")'
2) same in the show.
radmin> show debug condition
radmin>
|
1.0
|
Problems with radmin - Hi,
I noticed that the radmin doesn't works in the v3.0.x/HEAD
1) The debug condition does nothing.
radmin> debug condition '(User-Name != "jorge")'
2) same in the show.
radmin> show debug condition
radmin>
|
defect
|
problems with radmin hi i noticed that the radmin doesn t works in the x head the debug condition does nothing radmin debug condition user name jorge same in the show radmin show debug condition radmin
| 1
|
41,432
| 6,905,517,747
|
IssuesEvent
|
2017-11-27 07:32:36
|
godotengine/godot
|
https://api.github.com/repos/godotengine/godot
|
closed
|
Update 'Rect3' to AABB in the documentation
|
documentation
|
When i try to `godot -s tcp.gd` this script :
```python
#tcp.gd
extends SceneTree
var connection = null
var peerstream = null
var test = null
func _init():
print("Start client TCP")
# Connect
connection = StreamPeerTCP.new()
connection.connect_to_host("127.0.0.1", 8080)
peerstream = PacketPeerStream.new()
peerstream.set_stream_peer(connection)
# rect3 bug
test = Vector3(12.0,12.0,12.0)
peerstream.put_var(Rect3(test,test))
while(true):
if connection.is_connected_to_host() and connection.get_available_bytes() >0:
test = connection.get_var()
print(typeof(test))
print(test)
quit()
```
I get
```
SCRIPT ERROR: _init: Invalid call. Nonexistent function 'Rect3' in base 'SceneTree (tcp.gd)'.
At: res://tcp.gd:18
```
All other types works already
|
1.0
|
Update 'Rect3' to AABB in the documentation - When i try to `godot -s tcp.gd` this script :
```python
#tcp.gd
extends SceneTree
var connection = null
var peerstream = null
var test = null
func _init():
print("Start client TCP")
# Connect
connection = StreamPeerTCP.new()
connection.connect_to_host("127.0.0.1", 8080)
peerstream = PacketPeerStream.new()
peerstream.set_stream_peer(connection)
# rect3 bug
test = Vector3(12.0,12.0,12.0)
peerstream.put_var(Rect3(test,test))
while(true):
if connection.is_connected_to_host() and connection.get_available_bytes() >0:
test = connection.get_var()
print(typeof(test))
print(test)
quit()
```
I get
```
SCRIPT ERROR: _init: Invalid call. Nonexistent function 'Rect3' in base 'SceneTree (tcp.gd)'.
At: res://tcp.gd:18
```
All other types works already
|
non_defect
|
update to aabb in the documentation when i try to godot s tcp gd this script python tcp gd extends scenetree var connection null var peerstream null var test null func init print start client tcp connect connection streampeertcp new connection connect to host peerstream packetpeerstream new peerstream set stream peer connection bug test peerstream put var test test while true if connection is connected to host and connection get available bytes test connection get var print typeof test print test quit i get script error init invalid call nonexistent function in base scenetree tcp gd at res tcp gd all other types works already
| 0
|
103,581
| 16,602,927,509
|
IssuesEvent
|
2021-06-01 22:16:53
|
gms-ws-sandbox/nibrs
|
https://api.github.com/repos/gms-ws-sandbox/nibrs
|
opened
|
CVE-2019-10072 (High) detected in multiple libraries
|
security vulnerability
|
## CVE-2019-10072 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tomcat-embed-core-9.0.19.jar</b>, <b>tomcat-embed-core-8.5.34.jar</b>, <b>tomcat-embed-core-8.5.20.jar</b></p></summary>
<p>
<details><summary><b>tomcat-embed-core-9.0.19.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-summary-report-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/9.0.19/tomcat-embed-core-9.0.19.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.1.5.RELEASE.jar
- :x: **tomcat-embed-core-9.0.19.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-embed-core-8.5.34.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: nibrs/web/nibrs-web/pom.xml</p>
<p>Path to vulnerable library: nibrs/tools/nibrs-route/target/nibrs-route-1.0.0/WEB-INF/lib/tomcat-embed-core-8.5.34.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.34/tomcat-embed-core-8.5.34.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.34/tomcat-embed-core-8.5.34.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.34/tomcat-embed-core-8.5.34.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.34/tomcat-embed-core-8.5.34.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.34/tomcat-embed-core-8.5.34.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.0.5.RELEASE.jar
- :x: **tomcat-embed-core-8.5.34.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-embed-core-8.5.20.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/tomcat-embed-core-8.5.20.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.20/tomcat-embed-core-8.5.20.jar</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-embed-core-8.5.20.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs/commit/9fb1c19bd26c2113d1961640de126a33eacdc946">9fb1c19bd26c2113d1961640de126a33eacdc946</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The fix for CVE-2019-0199 was incomplete and did not address HTTP/2 connection window exhaustion on write in Apache Tomcat versions 9.0.0.M1 to 9.0.19 and 8.5.0 to 8.5.40 . By not sending WINDOW_UPDATE messages for the connection window (stream 0) clients were able to cause server-side threads to block eventually leading to thread exhaustion and a DoS.
<p>Publish Date: 2019-06-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10072>CVE-2019-10072</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://tomcat.apache.org/security-8.html#Fixed_in_Apache_Tomcat_8.5.41">http://tomcat.apache.org/security-8.html#Fixed_in_Apache_Tomcat_8.5.41</a></p>
<p>Release Date: 2019-06-21</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.20,8.5.41,org.apache.tomcat:tomcat-coyote:9.0.20,8.5.41</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"9.0.19","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.1.5.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:9.0.19","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.20,8.5.41,org.apache.tomcat:tomcat-coyote:9.0.20,8.5.41"},{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.34","packageFilePaths":["/web/nibrs-web/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-staging-data/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.0.5.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.0.5.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.5.34","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.20,8.5.41,org.apache.tomcat:tomcat-coyote:9.0.20,8.5.41"},{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.20","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.tomcat.embed:tomcat-embed-core:8.5.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.20,8.5.41,org.apache.tomcat:tomcat-coyote:9.0.20,8.5.41"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-10072","vulnerabilityDetails":"The fix for CVE-2019-0199 was incomplete and did not address HTTP/2 connection window exhaustion on write in Apache Tomcat versions 9.0.0.M1 to 9.0.19 and 8.5.0 to 8.5.40 . By not sending WINDOW_UPDATE messages for the connection window (stream 0) clients were able to cause server-side threads to block eventually leading to thread exhaustion and a DoS.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10072","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-10072 (High) detected in multiple libraries - ## CVE-2019-10072 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tomcat-embed-core-9.0.19.jar</b>, <b>tomcat-embed-core-8.5.34.jar</b>, <b>tomcat-embed-core-8.5.20.jar</b></p></summary>
<p>
<details><summary><b>tomcat-embed-core-9.0.19.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-summary-report-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/9.0.19/tomcat-embed-core-9.0.19.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.1.5.RELEASE.jar
- :x: **tomcat-embed-core-9.0.19.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-embed-core-8.5.34.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: nibrs/web/nibrs-web/pom.xml</p>
<p>Path to vulnerable library: nibrs/tools/nibrs-route/target/nibrs-route-1.0.0/WEB-INF/lib/tomcat-embed-core-8.5.34.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.34/tomcat-embed-core-8.5.34.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.34/tomcat-embed-core-8.5.34.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.34/tomcat-embed-core-8.5.34.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.34/tomcat-embed-core-8.5.34.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.34/tomcat-embed-core-8.5.34.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.0.5.RELEASE.jar
- :x: **tomcat-embed-core-8.5.34.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-embed-core-8.5.20.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/tomcat-embed-core-8.5.20.jar,/home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.20/tomcat-embed-core-8.5.20.jar</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-embed-core-8.5.20.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs/commit/9fb1c19bd26c2113d1961640de126a33eacdc946">9fb1c19bd26c2113d1961640de126a33eacdc946</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The fix for CVE-2019-0199 was incomplete and did not address HTTP/2 connection window exhaustion on write in Apache Tomcat versions 9.0.0.M1 to 9.0.19 and 8.5.0 to 8.5.40 . By not sending WINDOW_UPDATE messages for the connection window (stream 0) clients were able to cause server-side threads to block eventually leading to thread exhaustion and a DoS.
<p>Publish Date: 2019-06-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10072>CVE-2019-10072</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://tomcat.apache.org/security-8.html#Fixed_in_Apache_Tomcat_8.5.41">http://tomcat.apache.org/security-8.html#Fixed_in_Apache_Tomcat_8.5.41</a></p>
<p>Release Date: 2019-06-21</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.20,8.5.41,org.apache.tomcat:tomcat-coyote:9.0.20,8.5.41</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"9.0.19","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.1.5.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:9.0.19","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.20,8.5.41,org.apache.tomcat:tomcat-coyote:9.0.20,8.5.41"},{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.34","packageFilePaths":["/web/nibrs-web/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-staging-data/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.0.5.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.0.5.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.5.34","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.20,8.5.41,org.apache.tomcat:tomcat-coyote:9.0.20,8.5.41"},{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.20","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.tomcat.embed:tomcat-embed-core:8.5.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.20,8.5.41,org.apache.tomcat:tomcat-coyote:9.0.20,8.5.41"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-10072","vulnerabilityDetails":"The fix for CVE-2019-0199 was incomplete and did not address HTTP/2 connection window exhaustion on write in Apache Tomcat versions 9.0.0.M1 to 9.0.19 and 8.5.0 to 8.5.40 . By not sending WINDOW_UPDATE messages for the connection window (stream 0) clients were able to cause server-side threads to block eventually leading to thread exhaustion and a DoS.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10072","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries tomcat embed core jar tomcat embed core jar tomcat embed core jar tomcat embed core jar core tomcat implementation library home page a href path to dependency file nibrs tools nibrs summary report common pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file nibrs web nibrs web pom xml path to vulnerable library nibrs tools nibrs route target nibrs route web inf lib tomcat embed core jar home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file nibrs tools nibrs fbi service pom xml path to vulnerable library nibrs tools nibrs fbi service target nibrs fbi service web inf lib tomcat embed core jar home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy x tomcat embed core jar vulnerable library found in head commit a href found in base branch master vulnerability details the fix for cve was incomplete and did not address http connection window exhaustion on write in apache tomcat versions to and to by not sending window update messages for the connection window stream clients were able to cause server side threads to block eventually leading to thread exhaustion and a dos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter tomcat release org apache tomcat embed tomcat embed core isminimumfixversionavailable true minimumfixversion org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote packagetype java groupid org apache tomcat embed packagename tomcat embed core packageversion packagefilepaths istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter tomcat release org apache tomcat embed tomcat embed core isminimumfixversionavailable true minimumfixversion org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote packagetype java groupid org apache tomcat embed packagename tomcat embed core packageversion packagefilepaths istransitivedependency false dependencytree org apache tomcat embed tomcat embed core isminimumfixversionavailable true minimumfixversion org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote basebranches vulnerabilityidentifier cve vulnerabilitydetails the fix for cve was incomplete and did not address http connection window exhaustion on write in apache tomcat versions to and to by not sending window update messages for the connection window stream clients were able to cause server side threads to block eventually leading to thread exhaustion and a dos vulnerabilityurl
| 0
|
59,115
| 17,015,831,330
|
IssuesEvent
|
2021-07-02 11:54:38
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
opened
|
Tirex-backend-manager spins with 100% CPU if all backends die
|
Component: tirex Priority: minor Type: defect
|
**[Submitted to the original trac issue database at 3.43pm, Saturday, 4th February 2012]**
If all backend processes die (with non restartable errors) and no more processes are running without tirex-backend-manager having received a sighup, then tirex-backend-manager sits there spinning with 100% CPU usage.
If all backends exited with non-restartable errors, tirex-backend-manager should rather terminate as well than use up 100% CPU.
|
1.0
|
Tirex-backend-manager spins with 100% CPU if all backends die - **[Submitted to the original trac issue database at 3.43pm, Saturday, 4th February 2012]**
If all backend processes die (with non restartable errors) and no more processes are running without tirex-backend-manager having received a sighup, then tirex-backend-manager sits there spinning with 100% CPU usage.
If all backends exited with non-restartable errors, tirex-backend-manager should rather terminate as well than use up 100% CPU.
|
defect
|
tirex backend manager spins with cpu if all backends die if all backend processes die with non restartable errors and no more processes are running without tirex backend manager having received a sighup then tirex backend manager sits there spinning with cpu usage if all backends exited with non restartable errors tirex backend manager should rather terminate as well than use up cpu
| 1
|
349,035
| 31,768,849,975
|
IssuesEvent
|
2023-09-12 10:26:22
|
dieter-project/WithPT-BE
|
https://api.github.com/repos/dieter-project/WithPT-BE
|
opened
|
feat(gym) : 트레이너-체육관 로직 추가
|
✨ Feature ✅ Test
|
## 이슈 내용
- 트레이너 회원 가입 시 체육관 등록
- 트레이너 신규 회원 등록 시 체육관 추가 가능
- 수업관리 메에서 트레이너 소속 체육관 및 회원 수 조회
|
1.0
|
feat(gym) : 트레이너-체육관 로직 추가 - ## 이슈 내용
- 트레이너 회원 가입 시 체육관 등록
- 트레이너 신규 회원 등록 시 체육관 추가 가능
- 수업관리 메에서 트레이너 소속 체육관 및 회원 수 조회
|
non_defect
|
feat gym 트레이너 체육관 로직 추가 이슈 내용 트레이너 회원 가입 시 체육관 등록 트레이너 신규 회원 등록 시 체육관 추가 가능 수업관리 메에서 트레이너 소속 체육관 및 회원 수 조회
| 0
|
26,230
| 4,631,353,834
|
IssuesEvent
|
2016-09-28 15:15:11
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
opened
|
Parse/TryParse not support hex string
|
defect
|
### Expected
v has value
### Actual
Exception
http://forums.bridge.net/forum/community/help/2828-tryparse-not-support-hex-string
### Steps To Reproduce
```csharp
public class App
{
public static void Main()
{
var v = uint.Parse("0xffff",16);
}
}
```
|
1.0
|
Parse/TryParse not support hex string - ### Expected
v has value
### Actual
Exception
http://forums.bridge.net/forum/community/help/2828-tryparse-not-support-hex-string
### Steps To Reproduce
```csharp
public class App
{
public static void Main()
{
var v = uint.Parse("0xffff",16);
}
}
```
|
defect
|
parse tryparse not support hex string expected v has value actual exception steps to reproduce csharp public class app public static void main var v uint parse
| 1
|
810,312
| 30,235,923,647
|
IssuesEvent
|
2023-07-06 10:12:03
|
pendulum-chain/pendulum
|
https://api.github.com/repos/pendulum-chain/pendulum
|
closed
|
Update runtimes to `polkadot-v0.9.40`
|
priority:medium
|
We should update our runtime dependencies to version `polkadot-v0.9.38` so that we get rid of some dependency conflicts.
For example, we were running into problems because our spacewalk pallets use `jsonrpsee` with version >0.16.0 but other parachain pallets that are on `polkadot-v0.9.37` still use `jsonrpsee` version 0.15.x (e.g. Zenlink) which then leads to conflicts when importing RPC modules from their pallets.
|
1.0
|
Update runtimes to `polkadot-v0.9.40` - We should update our runtime dependencies to version `polkadot-v0.9.38` so that we get rid of some dependency conflicts.
For example, we were running into problems because our spacewalk pallets use `jsonrpsee` with version >0.16.0 but other parachain pallets that are on `polkadot-v0.9.37` still use `jsonrpsee` version 0.15.x (e.g. Zenlink) which then leads to conflicts when importing RPC modules from their pallets.
|
non_defect
|
update runtimes to polkadot we should update our runtime dependencies to version polkadot so that we get rid of some dependency conflicts for example we were running into problems because our spacewalk pallets use jsonrpsee with version but other parachain pallets that are on polkadot still use jsonrpsee version x e g zenlink which then leads to conflicts when importing rpc modules from their pallets
| 0
|
172,465
| 27,285,733,284
|
IssuesEvent
|
2023-02-23 13:22:40
|
vendure-ecommerce/vendure
|
https://api.github.com/repos/vendure-ecommerce/vendure
|
closed
|
Admin and Customer accounts with the same email
|
@vendure/core design 📐 v2
|
Currently you can create a customer for `alice@example.com`, and an administrator for `alice@example.com`. This will create 2 different user entities and 2 different authentication methods. Then, if `alice@example.com` tries to log in as an admin, she can't because the `BaseAuthResolver` finds her customer authentication method and it can't find an admin user linked to it.
Vendure should either:
1. Not allow administrators to use the same email as customers (effectively tie the email identifier with a user entity); OR
2. Allow two (admin and customer) accounts on the same email.
I think that 2 is preferable, since with multi-tenant setups you might not want to ask your shop owners to use different emails for administration and shopping. Ideally you should have 2 separate sessions depending on whether you're logged on the admin-api or shop-api. This would require adding 2 session keys to the cookie, since both APIs are on the same domain.
|
1.0
|
Admin and Customer accounts with the same email - Currently you can create a customer for `alice@example.com`, and an administrator for `alice@example.com`. This will create 2 different user entities and 2 different authentication methods. Then, if `alice@example.com` tries to log in as an admin, she can't because the `BaseAuthResolver` finds her customer authentication method and it can't find an admin user linked to it.
Vendure should either:
1. Not allow administrators to use the same email as customers (effectively tie the email identifier with a user entity); OR
2. Allow two (admin and customer) accounts on the same email.
I think that 2 is preferable, since with multi-tenant setups you might not want to ask your shop owners to use different emails for administration and shopping. Ideally you should have 2 separate sessions depending on whether you're logged on the admin-api or shop-api. This would require adding 2 session keys to the cookie, since both APIs are on the same domain.
|
non_defect
|
admin and customer accounts with the same email currently you can create a customer for alice example com and an administrator for alice example com this will create different user entities and different authentication methods then if alice example com tries to log in as an admin she can t because the baseauthresolver finds her customer authentication method and it can t find an admin user linked to it vendure should either not allow administrators to use the same email as customers effectively tie the email identifier with a user entity or allow two admin and customer accounts on the same email i think that is preferable since with multi tenant setups you might not want to ask your shop owners to use different emails for administration and shopping ideally you should have separate sessions depending on whether you re logged on the admin api or shop api this would require adding session keys to the cookie since both apis are on the same domain
| 0
|
65,313
| 19,345,858,639
|
IssuesEvent
|
2021-12-15 10:43:20
|
vector-im/element-ios
|
https://api.github.com/repos/vector-im/element-ios
|
closed
|
Voice call sound is audible bidirectionally only on the third connection attempt
|
T-Defect A-VoIP S-Major
|
### Steps to reproduce
1. I open the latest Element version on iPadOS 14.8.1 (or older versions)
2. During text chats with contacts, I press on the voice call icon to talk
3. On first attempt either I can hear them, but they cannot hear me, or vice versa
4. On second attempt either I can hear them, but they cannot hear me, or vice versa
5. On third attempt we can hear each other
### Outcome
#### I expected that the voice calls would connect with bidirectional audio on the first attempt
#### Instead, it takes three attempts
### Your phone model
iPad Pro (9.7-inch) MLQ52B/A A1674
### Operating system version
iPadOS 14.8.1
### Application version
Element 1.6.8 Matrix SDK 0.20.10 OLM 3.2.5
### Homeserver
https://matrix.org
### Will you send logs?
Yes
|
1.0
|
Voice call sound is audible bidirectionally only on the third connection attempt - ### Steps to reproduce
1. I open the latest Element version on iPadOS 14.8.1 (or older versions)
2. During text chats with contacts, I press on the voice call icon to talk
3. On first attempt either I can hear them, but they cannot hear me, or vice versa
4. On second attempt either I can hear them, but they cannot hear me, or vice versa
5. On third attempt we can hear each other
### Outcome
#### I expected that the voice calls would connect with bidirectional audio on the first attempt
#### Instead, it takes three attempts
### Your phone model
iPad Pro (9.7-inch) MLQ52B/A A1674
### Operating system version
iPadOS 14.8.1
### Application version
Element 1.6.8 Matrix SDK 0.20.10 OLM 3.2.5
### Homeserver
https://matrix.org
### Will you send logs?
Yes
|
defect
|
voice call sound is audible bidirectionally only on the third connection attempt steps to reproduce i open the latest element version on ipados or older versions during text chats with contacts i press on the voice call icon to talk on first attempt either i can hear them but they cannot hear me or vice versa on second attempt either i can hear them but they cannot hear me or vice versa on third attempt we can hear each other outcome i expected that the voice calls would connect with bidirectional audio on the first attempt instead it takes three attempts your phone model ipad pro inch a operating system version ipados application version element matrix sdk olm homeserver will you send logs yes
| 1
|
27,406
| 5,003,178,616
|
IssuesEvent
|
2016-12-11 19:45:03
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
opened
|
GetGenericArguments returns null for generic type definition
|
defect
|
### Expected
```js
Array
```
### Actual
```js
null
```
### Steps To Reproduce
[Deck](http://deck.net/6d4b2e89d25d3b84dd6a74c03ed9b609)
```cs
public class Base<T, U> { }
public class Derived<V> : Base<V, V> { }
public class Program
{
public static void Main()
{
Type derivedType = typeof(Derived<>);
Console.WriteLine(derivedType.GetGenericArguments());
}
}
```
|
1.0
|
GetGenericArguments returns null for generic type definition - ### Expected
```js
Array
```
### Actual
```js
null
```
### Steps To Reproduce
[Deck](http://deck.net/6d4b2e89d25d3b84dd6a74c03ed9b609)
```cs
public class Base<T, U> { }
public class Derived<V> : Base<V, V> { }
public class Program
{
public static void Main()
{
Type derivedType = typeof(Derived<>);
Console.WriteLine(derivedType.GetGenericArguments());
}
}
```
|
defect
|
getgenericarguments returns null for generic type definition expected js array actual js null steps to reproduce cs public class base public class derived base public class program public static void main type derivedtype typeof derived console writeline derivedtype getgenericarguments
| 1
|
68,783
| 21,896,607,456
|
IssuesEvent
|
2022-05-20 09:14:38
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Input area is not aligned with event body's first letter in a thread on IRC/modern layout
|
T-Defect S-Tolerable A-Message-Editing O-Frequent A-Threads
|
### Steps to reproduce
1. Enable modern layout
2. Open a thread
3. Send a message
4. Edit the message
### Outcome
#### What did you expect?
The input area should be aligned with display name and other event tiles' content.

#### What happened instead?
The input area is not aligned.

### Operating system
Debian
### Browser information
Firefox ESR 99
### URL for webapp
localhost
### Application version
9889aa0de29b35e610eb51d96dae568f3debd8f0
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Input area is not aligned with event body's first letter in a thread on IRC/modern layout - ### Steps to reproduce
1. Enable modern layout
2. Open a thread
3. Send a message
4. Edit the message
### Outcome
#### What did you expect?
The input area should be aligned with display name and other event tiles' content.

#### What happened instead?
The input area is not aligned.

### Operating system
Debian
### Browser information
Firefox ESR 99
### URL for webapp
localhost
### Application version
9889aa0de29b35e610eb51d96dae568f3debd8f0
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
input area is not aligned with event body s first letter in a thread on irc modern layout steps to reproduce enable modern layout open a thread send a message edit the message outcome what did you expect the input area should be aligned with display name and other event tiles content what happened instead the input area is not aligned operating system debian browser information firefox esr url for webapp localhost application version homeserver no response will you send logs no
| 1
|
60,458
| 17,023,429,541
|
IssuesEvent
|
2021-07-03 01:59:32
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
OSMA: render landuse=* (eg. =brownfield) at layer=-5
|
Component: osmarender Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 5.41pm, Thursday, 25th June 2009]**
Currently landuse= draws over the top of tunnels. Ideally it should render lower down.
Eg. Stratford Olympic park:
http://www.openstreetmap.org/?lat=51.36959&lon=-0.079&zoom=15&layers=0B00FTF
|
1.0
|
OSMA: render landuse=* (eg. =brownfield) at layer=-5 - **[Submitted to the original trac issue database at 5.41pm, Thursday, 25th June 2009]**
Currently landuse= draws over the top of tunnels. Ideally it should render lower down.
Eg. Stratford Olympic park:
http://www.openstreetmap.org/?lat=51.36959&lon=-0.079&zoom=15&layers=0B00FTF
|
defect
|
osma render landuse eg brownfield at layer currently landuse draws over the top of tunnels ideally it should render lower down eg stratford olympic park
| 1
|
94,985
| 10,863,169,966
|
IssuesEvent
|
2019-11-14 14:40:18
|
georchestra/mapstore2-georchestra
|
https://api.github.com/repos/georchestra/mapstore2-georchestra
|
opened
|
Review Documentation
|
documentation
|
- [ ] Remove sample Cadastrapp page
- è ] Remove indicies and tables
- [ ] Add logo (as in this [sample](https://docs.ckan.org/en/2.8/) )
|
1.0
|
Review Documentation - - [ ] Remove sample Cadastrapp page
- è ] Remove indicies and tables
- [ ] Add logo (as in this [sample](https://docs.ckan.org/en/2.8/) )
|
non_defect
|
review documentation remove sample cadastrapp page è remove indicies and tables add logo as in this
| 0
|
48,593
| 13,157,184,595
|
IssuesEvent
|
2020-08-10 12:17:22
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Providing the primary key and using INSERT ... RETURNING returns null for emulating databases
|
C: Functionality E: All Editions P: Medium R: Worksforme T: Defect
|
When using the `INSERT ... RETURNING` syntax and providing the primary key, I've observed different behavior in case the feature is natively supported and if the feature is emulated.
E.g. doing
```sql
INSERT INTO foo(some_id) VALUES(1) RETURNING some_id
```
returns `1` for Postgres, but `null` for SQLite which does internally
```sql
INSERT INTO foo(some_id) VALUES(1);
select last_insert_rowid(); //return 0
select foo.some_id from foo where foo.some_id = 0; //null
```
This affects jOOQ 3.10.6.
|
1.0
|
Providing the primary key and using INSERT ... RETURNING returns null for emulating databases - When using the `INSERT ... RETURNING` syntax and providing the primary key, I've observed different behavior in case the feature is natively supported and if the feature is emulated.
E.g. doing
```sql
INSERT INTO foo(some_id) VALUES(1) RETURNING some_id
```
returns `1` for Postgres, but `null` for SQLite which does internally
```sql
INSERT INTO foo(some_id) VALUES(1);
select last_insert_rowid(); //return 0
select foo.some_id from foo where foo.some_id = 0; //null
```
This affects jOOQ 3.10.6.
|
defect
|
providing the primary key and using insert returning returns null for emulating databases when using the insert returning syntax and providing the primary key i ve observed different behavior in case the feature is natively supported and if the feature is emulated e g doing sql insert into foo some id values returning some id returns for postgres but null for sqlite which does internally sql insert into foo some id values select last insert rowid return select foo some id from foo where foo some id null this affects jooq
| 1
|
203,674
| 23,168,990,961
|
IssuesEvent
|
2022-07-30 12:04:49
|
MatBenfield/news
|
https://api.github.com/repos/MatBenfield/news
|
closed
|
[SecurityWeek] GitHub Improves npm Account Security as Incidents Rise
|
SecurityWeek Stale
|
**Microsoft-owned GitHub this week announced new npm security improvements, amid an increase in incidents involving malicious npm packages.**
[read more](https://www.securityweek.com/github-improves-npm-account-security-incidents-rise)
<https://www.securityweek.com/github-improves-npm-account-security-incidents-rise>
|
True
|
[SecurityWeek] GitHub Improves npm Account Security as Incidents Rise -
**Microsoft-owned GitHub this week announced new npm security improvements, amid an increase in incidents involving malicious npm packages.**
[read more](https://www.securityweek.com/github-improves-npm-account-security-incidents-rise)
<https://www.securityweek.com/github-improves-npm-account-security-incidents-rise>
|
non_defect
|
github improves npm account security as incidents rise microsoft owned github this week announced new npm security improvements amid an increase in incidents involving malicious npm packages
| 0
|
2,864
| 2,607,963,631
|
IssuesEvent
|
2015-02-26 00:41:20
|
chrsmithdemos/leveldb
|
https://api.github.com/repos/chrsmithdemos/leveldb
|
closed
|
Add GNU/kFreeBSD support
|
auto-migrated Priority-Medium Type-Defect
|
```
Hi!
The attached patch will allow leveldb to compile on kFreeBSD platforms.
Regards,
```
-----
Original issue reported on code.google.com by `quadris...@gmail.com` on 5 Sep 2011 at 7:56
Attachments:
* [1002-kfreebsd.patch](https://storage.googleapis.com/google-code-attachments/leveldb/issue-38/comment-0/1002-kfreebsd.patch)
|
1.0
|
Add GNU/kFreeBSD support - ```
Hi!
The attached patch will allow leveldb to compile on kFreeBSD platforms.
Regards,
```
-----
Original issue reported on code.google.com by `quadris...@gmail.com` on 5 Sep 2011 at 7:56
Attachments:
* [1002-kfreebsd.patch](https://storage.googleapis.com/google-code-attachments/leveldb/issue-38/comment-0/1002-kfreebsd.patch)
|
defect
|
add gnu kfreebsd support hi the attached patch will allow leveldb to compile on kfreebsd platforms regards original issue reported on code google com by quadris gmail com on sep at attachments
| 1
|
33,471
| 7,130,688,213
|
IssuesEvent
|
2018-01-22 08:01:36
|
opencaching/opencaching-pl
|
https://api.github.com/repos/opencaching/opencaching-pl
|
closed
|
new cache page missing borders
|
Component_CacheEdit Priority_Low Type_Defect
|
New cache page is missing some borders in some sections.

|
1.0
|
new cache page missing borders - New cache page is missing some borders in some sections.

|
defect
|
new cache page missing borders new cache page is missing some borders in some sections
| 1
|
169,126
| 20,828,052,065
|
IssuesEvent
|
2022-03-19 01:26:07
|
Seagate/cortx-utils
|
https://api.github.com/repos/Seagate/cortx-utils
|
opened
|
CVE-2022-24302 (Medium) detected in paramiko-2.7.1-py2.py3-none-any.whl
|
security vulnerability
|
## CVE-2022-24302 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-2.7.1-py2.py3-none-any.whl</b></p></summary>
<p>SSH2 protocol library</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /py-utils</p>
<p>Path to vulnerable library: /py-utils,/py-utils/python_requirements.txt,/py-utils/test,/py-utils/src/utils/setup/openldap/openldap,/py-utils/src/setup/utils,/py-utils/src/utils/setup/kafka/kafka</p>
<p>
Dependency Hierarchy:
- :x: **paramiko-2.7.1-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302>CVE-2022-24302</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.paramiko.org/changelog.html">https://www.paramiko.org/changelog.html</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: paramiko - 2.10.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"paramiko","packageVersion":"2.7.1","packageFilePaths":["/py-utils"],"isTransitiveDependency":false,"dependencyTree":"paramiko:2.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"paramiko - 2.10.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-24302","vulnerabilityDetails":"In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2022-24302 (Medium) detected in paramiko-2.7.1-py2.py3-none-any.whl - ## CVE-2022-24302 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-2.7.1-py2.py3-none-any.whl</b></p></summary>
<p>SSH2 protocol library</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /py-utils</p>
<p>Path to vulnerable library: /py-utils,/py-utils/python_requirements.txt,/py-utils/test,/py-utils/src/utils/setup/openldap/openldap,/py-utils/src/setup/utils,/py-utils/src/utils/setup/kafka/kafka</p>
<p>
Dependency Hierarchy:
- :x: **paramiko-2.7.1-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302>CVE-2022-24302</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.paramiko.org/changelog.html">https://www.paramiko.org/changelog.html</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: paramiko - 2.10.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"paramiko","packageVersion":"2.7.1","packageFilePaths":["/py-utils"],"isTransitiveDependency":false,"dependencyTree":"paramiko:2.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"paramiko - 2.10.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-24302","vulnerabilityDetails":"In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve medium detected in paramiko none any whl cve medium severity vulnerability vulnerable library paramiko none any whl protocol library library home page a href path to dependency file py utils path to vulnerable library py utils py utils python requirements txt py utils test py utils src utils setup openldap openldap py utils src setup utils py utils src utils setup kafka kafka dependency hierarchy x paramiko none any whl vulnerable library found in base branch main vulnerability details in paramiko before a race condition between creation and chmod in the write private key file function could allow unauthorized information disclosure publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution paramiko rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree paramiko isminimumfixversionavailable true minimumfixversion paramiko isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in paramiko before a race condition between creation and chmod in the write private key file function could allow unauthorized information disclosure vulnerabilityurl
| 0
|
1,770
| 3,941,820,804
|
IssuesEvent
|
2016-04-27 09:23:22
|
tripikad/trip2
|
https://api.github.com/repos/tripikad/trip2
|
opened
|
Log execptions if needed
|
external services
|
First test whenever Exceptions are logged to ```syslog```. If not, do this:
In https://github.com/tripikad/trip2/blob/master/app/Exceptions/Handler.php#L27 add
```php
Log::error($e)
```
See also https://github.com/foxxmd/laravel-loggly#usage and #629
|
1.0
|
Log execptions if needed - First test whenever Exceptions are logged to ```syslog```. If not, do this:
In https://github.com/tripikad/trip2/blob/master/app/Exceptions/Handler.php#L27 add
```php
Log::error($e)
```
See also https://github.com/foxxmd/laravel-loggly#usage and #629
|
non_defect
|
log execptions if needed first test whenever exceptions are logged to syslog if not do this in add php log error e see also and
| 0
|
55,453
| 7,987,939,232
|
IssuesEvent
|
2018-07-19 09:24:39
|
hyn/multi-tenant
|
https://api.github.com/repos/hyn/multi-tenant
|
closed
|
Documentation on website needs previous/next button
|
documentation enhancement
|
Currently you have to go to the index to be able to view a documentation page. It would be nice if you can go to the previous or next documentation page by adding two links on each page (or one for the first/last page).
|
1.0
|
Documentation on website needs previous/next button - Currently you have to go to the index to be able to view a documentation page. It would be nice if you can go to the previous or next documentation page by adding two links on each page (or one for the first/last page).
|
non_defect
|
documentation on website needs previous next button currently you have to go to the index to be able to view a documentation page it would be nice if you can go to the previous or next documentation page by adding two links on each page or one for the first last page
| 0
|
153,620
| 13,520,124,209
|
IssuesEvent
|
2020-09-15 03:52:46
|
simonw/datasette
|
https://api.github.com/repos/simonw/datasette
|
closed
|
Remove _request_ip example from canned queries documentation
|
bug documentation
|
`_request_ip` isn't valid, so it shouldn't be in the example: https://github.com/simonw/datasette/blob/cb515a9d75430adaf5e545a840bbc111648e8bfd/docs/sql_queries.rst#L320-L322
|
1.0
|
Remove _request_ip example from canned queries documentation - `_request_ip` isn't valid, so it shouldn't be in the example: https://github.com/simonw/datasette/blob/cb515a9d75430adaf5e545a840bbc111648e8bfd/docs/sql_queries.rst#L320-L322
|
non_defect
|
remove request ip example from canned queries documentation request ip isn t valid so it shouldn t be in the example
| 0
|
38,009
| 8,633,003,679
|
IssuesEvent
|
2018-11-22 12:33:54
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
SlideMenu/TieredMenu: When a submenu is clicked the entire panel is hidden
|
defect
|
## 1) Environment
- PrimeFaces version: 6.3-SNAPSHOT
- Does it work on the newest released PrimeFaces version? Yes Version? 6.2
- Does it work on the newest sources in GitHub? Not
- Application server + version: WildFly 13
- Affected browsers: Any
## 2) Expected behavior
When we click the submenu this action shouldn't hide the entire panel.
## 3) Actual behavior
When we click the submenu this action hide the entire panel.
## 4) Steps to reproduce
1. Run the showcase and go to: /showcase/ui/menu/slideMenu.xhtml or /showcase/ui/menu/tieredMenu.xhtml
2. Click the button "Show"
3. Click the first submenu, the entire panel is hidden. That behavior is not the right one.
|
1.0
|
SlideMenu/TieredMenu: When a submenu is clicked the entire panel is hidden - ## 1) Environment
- PrimeFaces version: 6.3-SNAPSHOT
- Does it work on the newest released PrimeFaces version? Yes Version? 6.2
- Does it work on the newest sources in GitHub? Not
- Application server + version: WildFly 13
- Affected browsers: Any
## 2) Expected behavior
When we click the submenu this action shouldn't hide the entire panel.
## 3) Actual behavior
When we click the submenu this action hide the entire panel.
## 4) Steps to reproduce
1. Run the showcase and go to: /showcase/ui/menu/slideMenu.xhtml or /showcase/ui/menu/tieredMenu.xhtml
2. Click the button "Show"
3. Click the first submenu, the entire panel is hidden. That behavior is not the right one.
|
defect
|
slidemenu tieredmenu when a submenu is clicked the entire panel is hidden environment primefaces version snapshot does it work on the newest released primefaces version yes version does it work on the newest sources in github not application server version wildfly affected browsers any expected behavior when we click the submenu this action shouldn t hide the entire panel actual behavior when we click the submenu this action hide the entire panel steps to reproduce run the showcase and go to showcase ui menu slidemenu xhtml or showcase ui menu tieredmenu xhtml click the button show click the first submenu the entire panel is hidden that behavior is not the right one
| 1
|
642,983
| 20,919,953,334
|
IssuesEvent
|
2022-03-24 16:28:46
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Custom Column aliasing can cause incorrect results or wrong alias on nested queries
|
Type:Bug Priority:P2 .Correctness Querying/Nested Queries .Backend .Regression Querying/Notebook/Custom Column
|
**Describe the bug**
Custom Columns with names similar to existing table column names causes failed queries or incorrect results depending on database.
On BigQuery it does not fail, but shows wrong results. Perhaps other databases too.
Almost a repeat of #14255, but for nested queries.
Regression since 0.42.0 for some database, while BigQuery would fail previously with ambiguous errors.
We might have a frontend issue too, since it's possible to self-reference a Custom Column, which causes "No results". #21135
**To Reproduce**
1. Question > **BigQuery** Sample > Products
2. Custom Column `[Price] + 2` as "Price" (yes, the name of the Custom Column is the same as the underlying column name)

3. Save the question and use it as the base in a new question (or alternatively convert it to a Model, would cause the same problem)

<details><summary>Generated SQL for BigQuery</summary>
```
SELECT
`source`.`id` AS `id`,
`source`.`ean` AS `ean`,
`source`.`title` AS `title`,
`source`.`category` AS `category`,
`source`.`vendor` AS `vendor`,
`source`.`price` AS `price`,
`source`.`rating` AS `rating`,
`source`.`created_at` AS `created_at`,
`source`.`Price` AS `Price_2`
FROM
(
SELECT
`v6_sample_dataset.products`.`id` AS `id`,
`v6_sample_dataset.products`.`ean` AS `ean`,
`v6_sample_dataset.products`.`title` AS `title`,
`v6_sample_dataset.products`.`category` AS `category`,
`v6_sample_dataset.products`.`vendor` AS `vendor`,
`v6_sample_dataset.products`.`price` AS `price`,
`v6_sample_dataset.products`.`rating` AS `rating`,
`v6_sample_dataset.products`.`created_at` AS `created_at`,
(
`v6_sample_dataset.products`.`price` + 2
)
AS `Price_2`
FROM
`v6_sample_dataset.products`
) `source`
LIMIT 1048575
```
</details>
Doing similar repro with H2 would fail on step 3 with `Column "source.Price" not found;`.
If the Custom Column was called "PRICE" (all uppercase), then it works as expected. Guess it might just be a H2 thing.
<details><summary>Generated SQL for H2</summary>
```
SELECT
"source"."ID" AS "ID",
"source"."TITLE" AS "TITLE",
"source"."CATEGORY" AS "CATEGORY",
"source"."VENDOR" AS "VENDOR",
"source"."PRICE" AS "PRICE",
"source"."RATING" AS "RATING",
"source"."CREATED_AT" AS "CREATED_AT",
"source"."Price" AS "Price_2"
FROM
(
SELECT
"PUBLIC"."PRODUCTS"."ID" AS "ID",
"PUBLIC"."PRODUCTS"."TITLE" AS "TITLE",
"PUBLIC"."PRODUCTS"."CATEGORY" AS "CATEGORY",
"PUBLIC"."PRODUCTS"."VENDOR" AS "VENDOR",
"PUBLIC"."PRODUCTS"."PRICE" AS "PRICE",
"PUBLIC"."PRODUCTS"."RATING" AS "RATING",
"PUBLIC"."PRODUCTS"."CREATED_AT" AS "CREATED_AT",
(
"PUBLIC"."PRODUCTS"."PRICE" + 2
)
AS "Price_2"
FROM
"PUBLIC"."PRODUCTS"
)
"source"
LIMIT 1048575
```
</details>
<details><summary>Full stacktrace for H2</summary>
```
2022-03-21 16:01:30,597 INFO api.dataset :: Source query for this query is Card 867
2022-03-21 16:01:30,613 INFO middleware.fetch-source-query :: Fetched source query from Card 867:
{:source-table 10, :expressions {:Price [:+ [:field 104 nil] 2]}}
2022-03-21 16:01:30,620 INFO middleware.fetch-source-query :: Fetched source query from Card 867:
{:source-table 10, :expressions {:Price [:+ [:field 104 nil] 2]}}
2022-03-21 16:01:31,725 ERROR middleware.catch-exceptions :: Error processing query: null
{:database_id 4,
:started_at #t "2022-03-21T16:01:30.602959+01:00[Europe/Copenhagen]",
:via
[{:status :failed,
:class clojure.lang.ExceptionInfo,
:error
"Error executing query: Column \"source.Price\" not found; SQL statement:\n-- Metabase:: userID: 1 queryType: MBQL queryHash: 2994018ec812a2df87ffe8ffdf14e51cc433ebbae710fff85243178a6c0c0fa2\nSELECT \"source\".\"ID\" AS \"ID\", \"source\".\"TITLE\" AS \"TITLE\", \"source\".\"CATEGORY\" AS \"CATEGORY\", \"source\".\"VENDOR\" AS \"VENDOR\", \"source\".\"PRICE\" AS \"PRICE\", \"source\".\"RATING\" AS \"RATING\", \"source\".\"CREATED_AT\" AS \"CREATED_AT\", \"source\".\"Price\" AS \"Price_2\" FROM (SELECT \"PUBLIC\".\"PRODUCTS\".\"ID\" AS \"ID\", \"PUBLIC\".\"PRODUCTS\".\"TITLE\" AS \"TITLE\", \"PUBLIC\".\"PRODUCTS\".\"CATEGORY\" AS \"CATEGORY\", \"PUBLIC\".\"PRODUCTS\".\"VENDOR\" AS \"VENDOR\", \"PUBLIC\".\"PRODUCTS\".\"PRICE\" AS \"PRICE\", \"PUBLIC\".\"PRODUCTS\".\"RATING\" AS \"RATING\", \"PUBLIC\".\"PRODUCTS\".\"CREATED_AT\" AS \"CREATED_AT\", (\"PUBLIC\".\"PRODUCTS\".\"PRICE\" + 2) AS \"Price_2\" FROM \"PUBLIC\".\"PRODUCTS\") \"source\" LIMIT 2000 [42122-197]",
:stacktrace
["--> driver.sql_jdbc.execute$execute_reducible_query$fn__53301.invoke(execute.clj:504)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:501)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:487)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:496)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:487)"
"driver.sql_jdbc$fn__81999.invokeStatic(sql_jdbc.clj:54)"
"driver.sql_jdbc$fn__81999.invoke(sql_jdbc.clj:52)"
"driver.h2$fn__79149.invokeStatic(h2.clj:90)"
"driver.h2$fn__79149.invoke(h2.clj:87)"
"query_processor.context$executef.invokeStatic(context.clj:59)"
"query_processor.context$executef.invoke(context.clj:48)"
"query_processor.context.default$default_runf.invokeStatic(default.clj:68)"
"query_processor.context.default$default_runf.invoke(default.clj:66)"
"query_processor.context$runf.invokeStatic(context.clj:45)"
"query_processor.context$runf.invoke(context.clj:39)"
"query_processor.reducible$pivot.invokeStatic(reducible.clj:34)"
"query_processor.reducible$pivot.invoke(reducible.clj:31)"
"query_processor.middleware.mbql_to_native$mbql__GT_native$fn__49655.invoke(mbql_to_native.clj:25)"
"query_processor.middleware.check_features$check_features$fn__50401.invoke(check_features.clj:42)"
"query_processor.middleware.limit$limit$fn__47991.invoke(limit.clj:37)"
"query_processor.middleware.cache$maybe_return_cached_results$fn__50784.invoke(cache.clj:204)"
"query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__51848.invoke(optimize_temporal_filters.clj:204)"
"query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__51892.invoke(validate_temporal_bucketing.clj:50)"
"query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__49712.invoke(auto_parse_filter_values.clj:43)"
"query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__39780.invoke(wrap_value_literals.clj:161)"
"query_processor.middleware.annotate$add_column_info$fn__44546.invoke(annotate.clj:659)"
"query_processor.middleware.permissions$check_query_permissions$fn__46288.invoke(permissions.clj:108)"
"query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__51006.invoke(pre_alias_aggregations.clj:40)"
"query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46684.invoke(cumulative_aggregations.clj:60)"
"query_processor.middleware.visualization_settings$update_viz_settings$fn__46622.invoke(visualization_settings.clj:63)"
"query_processor.middleware.escape_join_aliases$escape_join_aliases_middleware$fn__48032.invoke(escape_join_aliases.clj:64)"
"query_processor.middleware.fix_bad_references$fix_bad_references_middleware$fn__50971.invoke(fix_bad_references.clj:91)"
"query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__47569.invoke(resolve_joined_fields.clj:111)"
"query_processor.middleware.resolve_joins$resolve_joins$fn__51618.invoke(resolve_joins.clj:178)"
"query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__51160.invoke(add_implicit_joins.clj:246)"
"query_processor.middleware.large_int_id$convert_id_to_string$fn__47588.invoke(large_int_id.clj:59)"
"query_processor.middleware.format_rows$format_rows$fn__51212.invoke(format_rows.clj:74)"
"query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__46956.invoke(add_default_temporal_unit.clj:23)"
"query_processor.middleware.desugar$desugar$fn__46595.invoke(desugar.clj:21)"
"query_processor.middleware.binning$update_binning_strategy$fn__39509.invoke(binning.clj:229)"
"query_processor.middleware.resolve_fields$resolve_fields$fn__45930.invoke(resolve_fields.clj:34)"
"query_processor.middleware.add_dimension_projections$add_remapping$fn__50338.invoke(add_dimension_projections.clj:487)"
"query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__50632.invoke(add_implicit_clauses.clj:164)"
"query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__47976.invoke(upgrade_field_literals.clj:117)"
"query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__47333.invoke(add_source_metadata.clj:125)"
"query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__50883.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)"
"query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__48970.invoke(auto_bucket_datetimes.clj:147)"
"query_processor.middleware.resolve_source_table$resolve_source_tables$fn__45911.invoke(resolve_source_table.clj:45)"
"query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)"
"query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)"
"query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)"
"query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)"
"query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)"
"driver$do_with_driver.invokeStatic(driver.clj:60)"
"driver$do_with_driver.invoke(driver.clj:56)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)"
"query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:282)"
"query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)"
"query_processor.store$do_with_store.invokeStatic(store.clj:44)"
"query_processor.store$do_with_store.invoke(store.clj:38)"
"query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)"
"query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)"
"query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)"
"query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)"
"query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)"
"query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)"
"query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)"
"query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)"
"query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)"
"query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)"
"query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)"
"query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)"
"query_processor$process_userland_query.invokeStatic(query_processor.clj:247)"
"query_processor$process_userland_query.doInvoke(query_processor.clj:243)"
"query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)"
"query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)"
"query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)"
"query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)"
"api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)"
"query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)"
"query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)"
"async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)"
"async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)"
"async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"],
:error_type :invalid-query,
:ex-data
{:sql
"-- Metabase:: userID: 1 queryType: MBQL queryHash: 2994018ec812a2df87ffe8ffdf14e51cc433ebbae710fff85243178a6c0c0fa2\nSELECT \"source\".\"ID\" AS \"ID\", \"source\".\"TITLE\" AS \"TITLE\", \"source\".\"CATEGORY\" AS \"CATEGORY\", \"source\".\"VENDOR\" AS \"VENDOR\", \"source\".\"PRICE\" AS \"PRICE\", \"source\".\"RATING\" AS \"RATING\", \"source\".\"CREATED_AT\" AS \"CREATED_AT\", \"source\".\"Price\" AS \"Price_2\" FROM (SELECT \"PUBLIC\".\"PRODUCTS\".\"ID\" AS \"ID\", \"PUBLIC\".\"PRODUCTS\".\"TITLE\" AS \"TITLE\", \"PUBLIC\".\"PRODUCTS\".\"CATEGORY\" AS \"CATEGORY\", \"PUBLIC\".\"PRODUCTS\".\"VENDOR\" AS \"VENDOR\", \"PUBLIC\".\"PRODUCTS\".\"PRICE\" AS \"PRICE\", \"PUBLIC\".\"PRODUCTS\".\"RATING\" AS \"RATING\", \"PUBLIC\".\"PRODUCTS\".\"CREATED_AT\" AS \"CREATED_AT\", (\"PUBLIC\".\"PRODUCTS\".\"PRICE\" + 2) AS \"Price_2\" FROM \"PUBLIC\".\"PRODUCTS\") \"source\" LIMIT 2000",
:params nil,
:type :invalid-query}}],
:state "42S22",
:error_type :invalid-query,
:json_query
{:database 4,
:query {:source-table "card__867"},
:type "query",
:parameters [],
:middleware {:js-int-to-string? true, :add-default-userland-constraints? true}},
:native
{:query
"SELECT \"source\".\"ID\" AS \"ID\", \"source\".\"TITLE\" AS \"TITLE\", \"source\".\"CATEGORY\" AS \"CATEGORY\", \"source\".\"VENDOR\" AS \"VENDOR\", \"source\".\"PRICE\" AS \"PRICE\", \"source\".\"RATING\" AS \"RATING\", \"source\".\"CREATED_AT\" AS \"CREATED_AT\", \"source\".\"Price\" AS \"Price_2\" FROM (SELECT \"PUBLIC\".\"PRODUCTS\".\"ID\" AS \"ID\", \"PUBLIC\".\"PRODUCTS\".\"TITLE\" AS \"TITLE\", \"PUBLIC\".\"PRODUCTS\".\"CATEGORY\" AS \"CATEGORY\", \"PUBLIC\".\"PRODUCTS\".\"VENDOR\" AS \"VENDOR\", \"PUBLIC\".\"PRODUCTS\".\"PRICE\" AS \"PRICE\", \"PUBLIC\".\"PRODUCTS\".\"RATING\" AS \"RATING\", \"PUBLIC\".\"PRODUCTS\".\"CREATED_AT\" AS \"CREATED_AT\", (\"PUBLIC\".\"PRODUCTS\".\"PRICE\" + 2) AS \"Price_2\" FROM \"PUBLIC\".\"PRODUCTS\") \"source\" LIMIT 2000",
:params nil},
:status :failed,
:class org.h2.jdbc.JdbcSQLException,
:stacktrace
["org.h2.message.DbException.getJdbcSQLException(DbException.java:357)"
"org.h2.message.DbException.get(DbException.java:179)"
"org.h2.message.DbException.get(DbException.java:155)"
"org.h2.expression.ExpressionColumn.optimize(ExpressionColumn.java:150)"
"org.h2.expression.Alias.optimize(Alias.java:51)"
"org.h2.command.dml.Select.prepare(Select.java:858)"
"org.h2.command.Parser.prepareCommand(Parser.java:283)"
"org.h2.engine.Session.prepareLocal(Session.java:611)"
"org.h2.engine.Session.prepareCommand(Session.java:549)"
"org.h2.jdbc.JdbcConnection.prepareCommand(JdbcConnection.java:1247)"
"org.h2.jdbc.JdbcStatement.executeInternal(JdbcStatement.java:217)"
"org.h2.jdbc.JdbcStatement.execute(JdbcStatement.java:205)"
"com.mchange.v2.c3p0.impl.NewProxyStatement.execute(NewProxyStatement.java:75)"
"--> driver.sql_jdbc.execute$fn__53221.invokeStatic(execute.clj:368)"
"driver.sql_jdbc.execute$fn__53221.invoke(execute.clj:366)"
"driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invokeStatic(execute.clj:376)"
"driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invoke(execute.clj:373)"
"driver.sql_jdbc.execute$execute_reducible_query$fn__53301.invoke(execute.clj:502)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:501)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:487)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:496)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:487)"
"driver.sql_jdbc$fn__81999.invokeStatic(sql_jdbc.clj:54)"
"driver.sql_jdbc$fn__81999.invoke(sql_jdbc.clj:52)"
"driver.h2$fn__79149.invokeStatic(h2.clj:90)"
"driver.h2$fn__79149.invoke(h2.clj:87)"
"query_processor.context$executef.invokeStatic(context.clj:59)"
"query_processor.context$executef.invoke(context.clj:48)"
"query_processor.context.default$default_runf.invokeStatic(default.clj:68)"
"query_processor.context.default$default_runf.invoke(default.clj:66)"
"query_processor.context$runf.invokeStatic(context.clj:45)"
"query_processor.context$runf.invoke(context.clj:39)"
"query_processor.reducible$pivot.invokeStatic(reducible.clj:34)"
"query_processor.reducible$pivot.invoke(reducible.clj:31)"
"query_processor.middleware.mbql_to_native$mbql__GT_native$fn__49655.invoke(mbql_to_native.clj:25)"
"query_processor.middleware.check_features$check_features$fn__50401.invoke(check_features.clj:42)"
"query_processor.middleware.limit$limit$fn__47991.invoke(limit.clj:37)"
"query_processor.middleware.cache$maybe_return_cached_results$fn__50784.invoke(cache.clj:204)"
"query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__51848.invoke(optimize_temporal_filters.clj:204)"
"query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__51892.invoke(validate_temporal_bucketing.clj:50)"
"query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__49712.invoke(auto_parse_filter_values.clj:43)"
"query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__39780.invoke(wrap_value_literals.clj:161)"
"query_processor.middleware.annotate$add_column_info$fn__44546.invoke(annotate.clj:659)"
"query_processor.middleware.permissions$check_query_permissions$fn__46288.invoke(permissions.clj:108)"
"query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__51006.invoke(pre_alias_aggregations.clj:40)"
"query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46684.invoke(cumulative_aggregations.clj:60)"
"query_processor.middleware.visualization_settings$update_viz_settings$fn__46622.invoke(visualization_settings.clj:63)"
"query_processor.middleware.escape_join_aliases$escape_join_aliases_middleware$fn__48032.invoke(escape_join_aliases.clj:64)"
"query_processor.middleware.fix_bad_references$fix_bad_references_middleware$fn__50971.invoke(fix_bad_references.clj:91)"
"query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__47569.invoke(resolve_joined_fields.clj:111)"
"query_processor.middleware.resolve_joins$resolve_joins$fn__51618.invoke(resolve_joins.clj:178)"
"query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__51160.invoke(add_implicit_joins.clj:246)"
"query_processor.middleware.large_int_id$convert_id_to_string$fn__47588.invoke(large_int_id.clj:59)"
"query_processor.middleware.format_rows$format_rows$fn__51212.invoke(format_rows.clj:74)"
"query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__46956.invoke(add_default_temporal_unit.clj:23)"
"query_processor.middleware.desugar$desugar$fn__46595.invoke(desugar.clj:21)"
"query_processor.middleware.binning$update_binning_strategy$fn__39509.invoke(binning.clj:229)"
"query_processor.middleware.resolve_fields$resolve_fields$fn__45930.invoke(resolve_fields.clj:34)"
"query_processor.middleware.add_dimension_projections$add_remapping$fn__50338.invoke(add_dimension_projections.clj:487)"
"query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__50632.invoke(add_implicit_clauses.clj:164)"
"query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__47976.invoke(upgrade_field_literals.clj:117)"
"query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__47333.invoke(add_source_metadata.clj:125)"
"query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__50883.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)"
"query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__48970.invoke(auto_bucket_datetimes.clj:147)"
"query_processor.middleware.resolve_source_table$resolve_source_tables$fn__45911.invoke(resolve_source_table.clj:45)"
"query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)"
"query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)"
"query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)"
"query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)"
"query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)"
"driver$do_with_driver.invokeStatic(driver.clj:60)"
"driver$do_with_driver.invoke(driver.clj:56)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)"
"query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:282)"
"query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)"
"query_processor.store$do_with_store.invokeStatic(store.clj:44)"
"query_processor.store$do_with_store.invoke(store.clj:38)"
"query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)"
"query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)"
"query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)"
"query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)"
"query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)"
"query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)"
"query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)"
"query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)"
"query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)"
"query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)"
"query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)"
"query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)"
"query_processor$process_userland_query.invokeStatic(query_processor.clj:247)"
"query_processor$process_userland_query.doInvoke(query_processor.clj:243)"
"query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)"
"query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)"
"query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)"
"query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)"
"api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)"
"query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)"
"query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)"
"async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)"
"async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)"
"async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"],
:card_id 867,
:context :ad-hoc,
:error
"Column \"source.Price\" not found; SQL statement:\n-- Metabase:: userID: 1 queryType: MBQL queryHash: 2994018ec812a2df87ffe8ffdf14e51cc433ebbae710fff85243178a6c0c0fa2\nSELECT \"source\".\"ID\" AS \"ID\", \"source\".\"TITLE\" AS \"TITLE\", \"source\".\"CATEGORY\" AS \"CATEGORY\", \"source\".\"VENDOR\" AS \"VENDOR\", \"source\".\"PRICE\" AS \"PRICE\", \"source\".\"RATING\" AS \"RATING\", \"source\".\"CREATED_AT\" AS \"CREATED_AT\", \"source\".\"Price\" AS \"Price_2\" FROM (SELECT \"PUBLIC\".\"PRODUCTS\".\"ID\" AS \"ID\", \"PUBLIC\".\"PRODUCTS\".\"TITLE\" AS \"TITLE\", \"PUBLIC\".\"PRODUCTS\".\"CATEGORY\" AS \"CATEGORY\", \"PUBLIC\".\"PRODUCTS\".\"VENDOR\" AS \"VENDOR\", \"PUBLIC\".\"PRODUCTS\".\"PRICE\" AS \"PRICE\", \"PUBLIC\".\"PRODUCTS\".\"RATING\" AS \"RATING\", \"PUBLIC\".\"PRODUCTS\".\"CREATED_AT\" AS \"CREATED_AT\", (\"PUBLIC\".\"PRODUCTS\".\"PRICE\" + 2) AS \"Price_2\" FROM \"PUBLIC\".\"PRODUCTS\") \"source\" LIMIT 2000 [42122-197]",
:row_count 0,
:running_time 0,
:preprocessed
{:database 4,
:query
{:source-card-id 867,
:source-metadata
[{:description
"The numerical product number. Only used internally. All external communication should use the title or EAN.",
:semantic_type :type/PK,
:coercion_strategy nil,
:name "ID",
:settings nil,
:field_ref [:field 105 nil],
:effective_type :type/BigInteger,
:id 105,
:display_name "ID",
:fingerprint nil,
:base_type :type/BigInteger}
{:description "The name of the product as it should be displayed to customers.",
:semantic_type :type/Title,
:coercion_strategy nil,
:name "TITLE",
:settings nil,
:field_ref [:field 106 nil],
:effective_type :type/Text,
:id 106,
:display_name "Title",
:fingerprint
{:global {:distinct-count 199, :nil% 0.0},
:type
{:type/Text
{:percent-json 0.0, :percent-url 0.0, :percent-email 0.0, :percent-state 0.0, :average-length 21.495}}},
:base_type :type/Text}
{:description "The type of product, valid values include: Doohicky, Gadget, Gizmo and Widget",
:semantic_type :type/Category,
:coercion_strategy nil,
:name "CATEGORY",
:settings nil,
:field_ref [:field 107 nil],
:effective_type :type/Text,
:id 107,
:display_name "Category",
:fingerprint
{:global {:distinct-count 4, :nil% 0.0},
:type
{:type/Text
{:percent-json 0.0, :percent-url 0.0, :percent-email 0.0, :percent-state 0.0, :average-length 6.375}}},
:base_type :type/Text}
{:description "The source of the product.",
:semantic_type :type/Company,
:coercion_strategy nil,
:name "VENDOR",
:settings nil,
:field_ref [:field 109 nil],
:effective_type :type/Text,
:id 109,
:display_name "Vendor",
:fingerprint
{:global {:distinct-count 200, :nil% 0.0},
:type
{:type/Text
{:percent-json 0.0, :percent-url 0.0, :percent-email 0.0, :percent-state 0.0, :average-length 20.6}}},
:base_type :type/Text}
{:description
"The list price of the product. Note that this is not always the price the product sold for due to discounts, promotions, etc.",
:semantic_type :type/Price,
:coercion_strategy nil,
:name "PRICE",
:settings {},
:field_ref [:field 104 nil],
:effective_type :type/Float,
:id 104,
:display_name "Price",
:fingerprint
{:global {:distinct-count 170, :nil% 0.0},
:type
{:type/Number
{:min 15.691943673970439,
:q1 37.25154462926434,
:q3 75.45898071609447,
:max 98.81933684368194,
:sd 21.711481557852057,
:avg 55.74639966792074}}},
:base_type :type/Float}
{:description "The average rating users have given the product. This ranges from 1 - 5",
:semantic_type :type/Score,
:coercion_strategy nil,
:name "RATING",
:settings nil,
:field_ref [:field 103 nil],
:effective_type :type/Float,
:id 103,
:display_name "Rating",
:fingerprint
{:global {:distinct-count 23, :nil% 0.0},
:type
{:type/Number
{:min 0.0, :q1 3.5120465053408525, :q3 4.216124969497314, :max 5.0, :sd 1.3605488657451452, :avg 3.4715}}},
:base_type :type/Float}
{:description "The date the product was added to our catalog.",
:semantic_type :type/CreationTimestamp,
:coercion_strategy nil,
:unit :default,
:name "CREATED_AT",
:settings {:date_separator ".", :time_style "k:mm"},
:field_ref [:field 108 {:temporal-unit :default}],
:effective_type :type/DateTime,
:id 108,
:display_name "Created At",
:fingerprint
{:global {:distinct-count 200, :nil% 0.0},
:type {:type/DateTime {:earliest "2016-04-26T19:29:55.147Z", :latest "2019-04-15T13:34:19.931Z"}}},
:base_type :type/DateTime}
{:display_name "Price",
:field_ref [:expression "Price"],
:name "Price",
:base_type :type/Float,
:effective_type :type/Float,
:semantic_type nil,
:fingerprint
{:global {:distinct-count 170, :nil% 0.0},
:type
{:type/Number
{:min 17.69194367397044,
:q1 39.25154462926434,
:q3 77.45898071609447,
:max 100.81933684368194,
:sd 21.711481557852057,
:avg 57.74639966792074}}}}],
:fields
[[:field 105 nil]
[:field 106 nil]
[:field 107 nil]
[:field 109 nil]
[:field 104 nil]
[:field 103 nil]
[:field 108 {:temporal-unit :default}]
[:field "Price" {:base-type :type/Float}]],
:source-query
{:source-table 10,
:expressions {:Price [:+ [:field 104 nil] 2]},
:fields
[[:field 105 nil]
[:field 106 nil]
[:field 107 nil]
[:field 109 nil]
[:field 104 nil]
[:field 103 nil]
[:field 108 {:temporal-unit :default}]
[:expression "Price"]]},
:limit 2000},
:type :query,
:middleware {:js-int-to-string? true, :add-default-userland-constraints? true},
:info
{:executed-by 1,
:context :ad-hoc,
:card-id 867,
:nested? true,
:query-hash
[41, -108, 1, -114, -56, 18, -94, -33, -121, -1, -24, -1, -33, 20, -27, 28, -60, 51, -21, -70, -25, 16, -1, -8, 82,
67, 23, -118, 108, 12, 15, -94]},
:constraints {:max-results 10000, :max-results-bare-rows 2000}},
:data {:rows [], :cols []}}
2022-03-21 16:01:31,731 DEBUG middleware.log :: POST /api/dataset 202 [ASYNC: completed] 1.1 s (22 DB calls) App DB connections: 0/13 Jetty threads: 3/50 (4 idle, 0 queued) (96 total active threads) Queries in flight: 0 (0 queued); h2 DB 4 connections: 0/1 (0 threads blocked)
```
</details>
**Information about your Metabase Installation:**
Tested 0.41.6 thru 0.42.2 - regression since 0.42.0
|
1.0
|
Custom Column aliasing can cause incorrect results or wrong alias on nested queries - **Describe the bug**
Custom Columns with names similar to existing table column names causes failed queries or incorrect results depending on database.
On BigQuery it does not fail, but shows wrong results. Perhaps other databases too.
Almost a repeat of #14255, but for nested queries.
Regression since 0.42.0 for some database, while BigQuery would fail previously with ambiguous errors.
We might have a frontend issue too, since it's possible to self-reference a Custom Column, which causes "No results". #21135
**To Reproduce**
1. Question > **BigQuery** Sample > Products
2. Custom Column `[Price] + 2` as "Price" (yes, the name of the Custom Column is the same as the underlying column name)

3. Save the question and use it as the base in a new question (or alternatively convert it to a Model, would cause the same problem)

<details><summary>Generated SQL for BigQuery</summary>
```
SELECT
`source`.`id` AS `id`,
`source`.`ean` AS `ean`,
`source`.`title` AS `title`,
`source`.`category` AS `category`,
`source`.`vendor` AS `vendor`,
`source`.`price` AS `price`,
`source`.`rating` AS `rating`,
`source`.`created_at` AS `created_at`,
`source`.`Price` AS `Price_2`
FROM
(
SELECT
`v6_sample_dataset.products`.`id` AS `id`,
`v6_sample_dataset.products`.`ean` AS `ean`,
`v6_sample_dataset.products`.`title` AS `title`,
`v6_sample_dataset.products`.`category` AS `category`,
`v6_sample_dataset.products`.`vendor` AS `vendor`,
`v6_sample_dataset.products`.`price` AS `price`,
`v6_sample_dataset.products`.`rating` AS `rating`,
`v6_sample_dataset.products`.`created_at` AS `created_at`,
(
`v6_sample_dataset.products`.`price` + 2
)
AS `Price_2`
FROM
`v6_sample_dataset.products`
) `source`
LIMIT 1048575
```
</details>
Doing similar repro with H2 would fail on step 3 with `Column "source.Price" not found;`.
If the Custom Column was called "PRICE" (all uppercase), then it works as expected. Guess it might just be a H2 thing.
<details><summary>Generated SQL for H2</summary>
```
SELECT
"source"."ID" AS "ID",
"source"."TITLE" AS "TITLE",
"source"."CATEGORY" AS "CATEGORY",
"source"."VENDOR" AS "VENDOR",
"source"."PRICE" AS "PRICE",
"source"."RATING" AS "RATING",
"source"."CREATED_AT" AS "CREATED_AT",
"source"."Price" AS "Price_2"
FROM
(
SELECT
"PUBLIC"."PRODUCTS"."ID" AS "ID",
"PUBLIC"."PRODUCTS"."TITLE" AS "TITLE",
"PUBLIC"."PRODUCTS"."CATEGORY" AS "CATEGORY",
"PUBLIC"."PRODUCTS"."VENDOR" AS "VENDOR",
"PUBLIC"."PRODUCTS"."PRICE" AS "PRICE",
"PUBLIC"."PRODUCTS"."RATING" AS "RATING",
"PUBLIC"."PRODUCTS"."CREATED_AT" AS "CREATED_AT",
(
"PUBLIC"."PRODUCTS"."PRICE" + 2
)
AS "Price_2"
FROM
"PUBLIC"."PRODUCTS"
)
"source"
LIMIT 1048575
```
</details>
<details><summary>Full stacktrace for H2</summary>
```
2022-03-21 16:01:30,597 INFO api.dataset :: Source query for this query is Card 867
2022-03-21 16:01:30,613 INFO middleware.fetch-source-query :: Fetched source query from Card 867:
{:source-table 10, :expressions {:Price [:+ [:field 104 nil] 2]}}
2022-03-21 16:01:30,620 INFO middleware.fetch-source-query :: Fetched source query from Card 867:
{:source-table 10, :expressions {:Price [:+ [:field 104 nil] 2]}}
2022-03-21 16:01:31,725 ERROR middleware.catch-exceptions :: Error processing query: null
{:database_id 4,
:started_at #t "2022-03-21T16:01:30.602959+01:00[Europe/Copenhagen]",
:via
[{:status :failed,
:class clojure.lang.ExceptionInfo,
:error
"Error executing query: Column \"source.Price\" not found; SQL statement:\n-- Metabase:: userID: 1 queryType: MBQL queryHash: 2994018ec812a2df87ffe8ffdf14e51cc433ebbae710fff85243178a6c0c0fa2\nSELECT \"source\".\"ID\" AS \"ID\", \"source\".\"TITLE\" AS \"TITLE\", \"source\".\"CATEGORY\" AS \"CATEGORY\", \"source\".\"VENDOR\" AS \"VENDOR\", \"source\".\"PRICE\" AS \"PRICE\", \"source\".\"RATING\" AS \"RATING\", \"source\".\"CREATED_AT\" AS \"CREATED_AT\", \"source\".\"Price\" AS \"Price_2\" FROM (SELECT \"PUBLIC\".\"PRODUCTS\".\"ID\" AS \"ID\", \"PUBLIC\".\"PRODUCTS\".\"TITLE\" AS \"TITLE\", \"PUBLIC\".\"PRODUCTS\".\"CATEGORY\" AS \"CATEGORY\", \"PUBLIC\".\"PRODUCTS\".\"VENDOR\" AS \"VENDOR\", \"PUBLIC\".\"PRODUCTS\".\"PRICE\" AS \"PRICE\", \"PUBLIC\".\"PRODUCTS\".\"RATING\" AS \"RATING\", \"PUBLIC\".\"PRODUCTS\".\"CREATED_AT\" AS \"CREATED_AT\", (\"PUBLIC\".\"PRODUCTS\".\"PRICE\" + 2) AS \"Price_2\" FROM \"PUBLIC\".\"PRODUCTS\") \"source\" LIMIT 2000 [42122-197]",
:stacktrace
["--> driver.sql_jdbc.execute$execute_reducible_query$fn__53301.invoke(execute.clj:504)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:501)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:487)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:496)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:487)"
"driver.sql_jdbc$fn__81999.invokeStatic(sql_jdbc.clj:54)"
"driver.sql_jdbc$fn__81999.invoke(sql_jdbc.clj:52)"
"driver.h2$fn__79149.invokeStatic(h2.clj:90)"
"driver.h2$fn__79149.invoke(h2.clj:87)"
"query_processor.context$executef.invokeStatic(context.clj:59)"
"query_processor.context$executef.invoke(context.clj:48)"
"query_processor.context.default$default_runf.invokeStatic(default.clj:68)"
"query_processor.context.default$default_runf.invoke(default.clj:66)"
"query_processor.context$runf.invokeStatic(context.clj:45)"
"query_processor.context$runf.invoke(context.clj:39)"
"query_processor.reducible$pivot.invokeStatic(reducible.clj:34)"
"query_processor.reducible$pivot.invoke(reducible.clj:31)"
"query_processor.middleware.mbql_to_native$mbql__GT_native$fn__49655.invoke(mbql_to_native.clj:25)"
"query_processor.middleware.check_features$check_features$fn__50401.invoke(check_features.clj:42)"
"query_processor.middleware.limit$limit$fn__47991.invoke(limit.clj:37)"
"query_processor.middleware.cache$maybe_return_cached_results$fn__50784.invoke(cache.clj:204)"
"query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__51848.invoke(optimize_temporal_filters.clj:204)"
"query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__51892.invoke(validate_temporal_bucketing.clj:50)"
"query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__49712.invoke(auto_parse_filter_values.clj:43)"
"query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__39780.invoke(wrap_value_literals.clj:161)"
"query_processor.middleware.annotate$add_column_info$fn__44546.invoke(annotate.clj:659)"
"query_processor.middleware.permissions$check_query_permissions$fn__46288.invoke(permissions.clj:108)"
"query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__51006.invoke(pre_alias_aggregations.clj:40)"
"query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46684.invoke(cumulative_aggregations.clj:60)"
"query_processor.middleware.visualization_settings$update_viz_settings$fn__46622.invoke(visualization_settings.clj:63)"
"query_processor.middleware.escape_join_aliases$escape_join_aliases_middleware$fn__48032.invoke(escape_join_aliases.clj:64)"
"query_processor.middleware.fix_bad_references$fix_bad_references_middleware$fn__50971.invoke(fix_bad_references.clj:91)"
"query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__47569.invoke(resolve_joined_fields.clj:111)"
"query_processor.middleware.resolve_joins$resolve_joins$fn__51618.invoke(resolve_joins.clj:178)"
"query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__51160.invoke(add_implicit_joins.clj:246)"
"query_processor.middleware.large_int_id$convert_id_to_string$fn__47588.invoke(large_int_id.clj:59)"
"query_processor.middleware.format_rows$format_rows$fn__51212.invoke(format_rows.clj:74)"
"query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__46956.invoke(add_default_temporal_unit.clj:23)"
"query_processor.middleware.desugar$desugar$fn__46595.invoke(desugar.clj:21)"
"query_processor.middleware.binning$update_binning_strategy$fn__39509.invoke(binning.clj:229)"
"query_processor.middleware.resolve_fields$resolve_fields$fn__45930.invoke(resolve_fields.clj:34)"
"query_processor.middleware.add_dimension_projections$add_remapping$fn__50338.invoke(add_dimension_projections.clj:487)"
"query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__50632.invoke(add_implicit_clauses.clj:164)"
"query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__47976.invoke(upgrade_field_literals.clj:117)"
"query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__47333.invoke(add_source_metadata.clj:125)"
"query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__50883.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)"
"query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__48970.invoke(auto_bucket_datetimes.clj:147)"
"query_processor.middleware.resolve_source_table$resolve_source_tables$fn__45911.invoke(resolve_source_table.clj:45)"
"query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)"
"query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)"
"query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)"
"query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)"
"query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)"
"driver$do_with_driver.invokeStatic(driver.clj:60)"
"driver$do_with_driver.invoke(driver.clj:56)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)"
"query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:282)"
"query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)"
"query_processor.store$do_with_store.invokeStatic(store.clj:44)"
"query_processor.store$do_with_store.invoke(store.clj:38)"
"query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)"
"query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)"
"query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)"
"query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)"
"query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)"
"query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)"
"query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)"
"query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)"
"query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)"
"query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)"
"query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)"
"query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)"
"query_processor$process_userland_query.invokeStatic(query_processor.clj:247)"
"query_processor$process_userland_query.doInvoke(query_processor.clj:243)"
"query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)"
"query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)"
"query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)"
"query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)"
"api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)"
"query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)"
"query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)"
"async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)"
"async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)"
"async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"],
:error_type :invalid-query,
:ex-data
{:sql
"-- Metabase:: userID: 1 queryType: MBQL queryHash: 2994018ec812a2df87ffe8ffdf14e51cc433ebbae710fff85243178a6c0c0fa2\nSELECT \"source\".\"ID\" AS \"ID\", \"source\".\"TITLE\" AS \"TITLE\", \"source\".\"CATEGORY\" AS \"CATEGORY\", \"source\".\"VENDOR\" AS \"VENDOR\", \"source\".\"PRICE\" AS \"PRICE\", \"source\".\"RATING\" AS \"RATING\", \"source\".\"CREATED_AT\" AS \"CREATED_AT\", \"source\".\"Price\" AS \"Price_2\" FROM (SELECT \"PUBLIC\".\"PRODUCTS\".\"ID\" AS \"ID\", \"PUBLIC\".\"PRODUCTS\".\"TITLE\" AS \"TITLE\", \"PUBLIC\".\"PRODUCTS\".\"CATEGORY\" AS \"CATEGORY\", \"PUBLIC\".\"PRODUCTS\".\"VENDOR\" AS \"VENDOR\", \"PUBLIC\".\"PRODUCTS\".\"PRICE\" AS \"PRICE\", \"PUBLIC\".\"PRODUCTS\".\"RATING\" AS \"RATING\", \"PUBLIC\".\"PRODUCTS\".\"CREATED_AT\" AS \"CREATED_AT\", (\"PUBLIC\".\"PRODUCTS\".\"PRICE\" + 2) AS \"Price_2\" FROM \"PUBLIC\".\"PRODUCTS\") \"source\" LIMIT 2000",
:params nil,
:type :invalid-query}}],
:state "42S22",
:error_type :invalid-query,
:json_query
{:database 4,
:query {:source-table "card__867"},
:type "query",
:parameters [],
:middleware {:js-int-to-string? true, :add-default-userland-constraints? true}},
:native
{:query
"SELECT \"source\".\"ID\" AS \"ID\", \"source\".\"TITLE\" AS \"TITLE\", \"source\".\"CATEGORY\" AS \"CATEGORY\", \"source\".\"VENDOR\" AS \"VENDOR\", \"source\".\"PRICE\" AS \"PRICE\", \"source\".\"RATING\" AS \"RATING\", \"source\".\"CREATED_AT\" AS \"CREATED_AT\", \"source\".\"Price\" AS \"Price_2\" FROM (SELECT \"PUBLIC\".\"PRODUCTS\".\"ID\" AS \"ID\", \"PUBLIC\".\"PRODUCTS\".\"TITLE\" AS \"TITLE\", \"PUBLIC\".\"PRODUCTS\".\"CATEGORY\" AS \"CATEGORY\", \"PUBLIC\".\"PRODUCTS\".\"VENDOR\" AS \"VENDOR\", \"PUBLIC\".\"PRODUCTS\".\"PRICE\" AS \"PRICE\", \"PUBLIC\".\"PRODUCTS\".\"RATING\" AS \"RATING\", \"PUBLIC\".\"PRODUCTS\".\"CREATED_AT\" AS \"CREATED_AT\", (\"PUBLIC\".\"PRODUCTS\".\"PRICE\" + 2) AS \"Price_2\" FROM \"PUBLIC\".\"PRODUCTS\") \"source\" LIMIT 2000",
:params nil},
:status :failed,
:class org.h2.jdbc.JdbcSQLException,
:stacktrace
["org.h2.message.DbException.getJdbcSQLException(DbException.java:357)"
"org.h2.message.DbException.get(DbException.java:179)"
"org.h2.message.DbException.get(DbException.java:155)"
"org.h2.expression.ExpressionColumn.optimize(ExpressionColumn.java:150)"
"org.h2.expression.Alias.optimize(Alias.java:51)"
"org.h2.command.dml.Select.prepare(Select.java:858)"
"org.h2.command.Parser.prepareCommand(Parser.java:283)"
"org.h2.engine.Session.prepareLocal(Session.java:611)"
"org.h2.engine.Session.prepareCommand(Session.java:549)"
"org.h2.jdbc.JdbcConnection.prepareCommand(JdbcConnection.java:1247)"
"org.h2.jdbc.JdbcStatement.executeInternal(JdbcStatement.java:217)"
"org.h2.jdbc.JdbcStatement.execute(JdbcStatement.java:205)"
"com.mchange.v2.c3p0.impl.NewProxyStatement.execute(NewProxyStatement.java:75)"
"--> driver.sql_jdbc.execute$fn__53221.invokeStatic(execute.clj:368)"
"driver.sql_jdbc.execute$fn__53221.invoke(execute.clj:366)"
"driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invokeStatic(execute.clj:376)"
"driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invoke(execute.clj:373)"
"driver.sql_jdbc.execute$execute_reducible_query$fn__53301.invoke(execute.clj:502)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:501)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:487)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:496)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:487)"
"driver.sql_jdbc$fn__81999.invokeStatic(sql_jdbc.clj:54)"
"driver.sql_jdbc$fn__81999.invoke(sql_jdbc.clj:52)"
"driver.h2$fn__79149.invokeStatic(h2.clj:90)"
"driver.h2$fn__79149.invoke(h2.clj:87)"
"query_processor.context$executef.invokeStatic(context.clj:59)"
"query_processor.context$executef.invoke(context.clj:48)"
"query_processor.context.default$default_runf.invokeStatic(default.clj:68)"
"query_processor.context.default$default_runf.invoke(default.clj:66)"
"query_processor.context$runf.invokeStatic(context.clj:45)"
"query_processor.context$runf.invoke(context.clj:39)"
"query_processor.reducible$pivot.invokeStatic(reducible.clj:34)"
"query_processor.reducible$pivot.invoke(reducible.clj:31)"
"query_processor.middleware.mbql_to_native$mbql__GT_native$fn__49655.invoke(mbql_to_native.clj:25)"
"query_processor.middleware.check_features$check_features$fn__50401.invoke(check_features.clj:42)"
"query_processor.middleware.limit$limit$fn__47991.invoke(limit.clj:37)"
"query_processor.middleware.cache$maybe_return_cached_results$fn__50784.invoke(cache.clj:204)"
"query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__51848.invoke(optimize_temporal_filters.clj:204)"
"query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__51892.invoke(validate_temporal_bucketing.clj:50)"
"query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__49712.invoke(auto_parse_filter_values.clj:43)"
"query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__39780.invoke(wrap_value_literals.clj:161)"
"query_processor.middleware.annotate$add_column_info$fn__44546.invoke(annotate.clj:659)"
"query_processor.middleware.permissions$check_query_permissions$fn__46288.invoke(permissions.clj:108)"
"query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__51006.invoke(pre_alias_aggregations.clj:40)"
"query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46684.invoke(cumulative_aggregations.clj:60)"
"query_processor.middleware.visualization_settings$update_viz_settings$fn__46622.invoke(visualization_settings.clj:63)"
"query_processor.middleware.escape_join_aliases$escape_join_aliases_middleware$fn__48032.invoke(escape_join_aliases.clj:64)"
"query_processor.middleware.fix_bad_references$fix_bad_references_middleware$fn__50971.invoke(fix_bad_references.clj:91)"
"query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__47569.invoke(resolve_joined_fields.clj:111)"
"query_processor.middleware.resolve_joins$resolve_joins$fn__51618.invoke(resolve_joins.clj:178)"
"query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__51160.invoke(add_implicit_joins.clj:246)"
"query_processor.middleware.large_int_id$convert_id_to_string$fn__47588.invoke(large_int_id.clj:59)"
"query_processor.middleware.format_rows$format_rows$fn__51212.invoke(format_rows.clj:74)"
"query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__46956.invoke(add_default_temporal_unit.clj:23)"
"query_processor.middleware.desugar$desugar$fn__46595.invoke(desugar.clj:21)"
"query_processor.middleware.binning$update_binning_strategy$fn__39509.invoke(binning.clj:229)"
"query_processor.middleware.resolve_fields$resolve_fields$fn__45930.invoke(resolve_fields.clj:34)"
"query_processor.middleware.add_dimension_projections$add_remapping$fn__50338.invoke(add_dimension_projections.clj:487)"
"query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__50632.invoke(add_implicit_clauses.clj:164)"
"query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__47976.invoke(upgrade_field_literals.clj:117)"
"query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__47333.invoke(add_source_metadata.clj:125)"
"query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__50883.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)"
"query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__48970.invoke(auto_bucket_datetimes.clj:147)"
"query_processor.middleware.resolve_source_table$resolve_source_tables$fn__45911.invoke(resolve_source_table.clj:45)"
"query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)"
"query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)"
"query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)"
"query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)"
"query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)"
"driver$do_with_driver.invokeStatic(driver.clj:60)"
"driver$do_with_driver.invoke(driver.clj:56)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)"
"query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:282)"
"query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)"
"query_processor.store$do_with_store.invokeStatic(store.clj:44)"
"query_processor.store$do_with_store.invoke(store.clj:38)"
"query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)"
"query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)"
"query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)"
"query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)"
"query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)"
"query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)"
"query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)"
"query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)"
"query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)"
"query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)"
"query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)"
"query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)"
"query_processor$process_userland_query.invokeStatic(query_processor.clj:247)"
"query_processor$process_userland_query.doInvoke(query_processor.clj:243)"
"query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)"
"query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)"
"query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)"
"query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)"
"api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)"
"query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)"
"query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)"
"async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)"
"async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)"
"async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"],
:card_id 867,
:context :ad-hoc,
:error
"Column \"source.Price\" not found; SQL statement:\n-- Metabase:: userID: 1 queryType: MBQL queryHash: 2994018ec812a2df87ffe8ffdf14e51cc433ebbae710fff85243178a6c0c0fa2\nSELECT \"source\".\"ID\" AS \"ID\", \"source\".\"TITLE\" AS \"TITLE\", \"source\".\"CATEGORY\" AS \"CATEGORY\", \"source\".\"VENDOR\" AS \"VENDOR\", \"source\".\"PRICE\" AS \"PRICE\", \"source\".\"RATING\" AS \"RATING\", \"source\".\"CREATED_AT\" AS \"CREATED_AT\", \"source\".\"Price\" AS \"Price_2\" FROM (SELECT \"PUBLIC\".\"PRODUCTS\".\"ID\" AS \"ID\", \"PUBLIC\".\"PRODUCTS\".\"TITLE\" AS \"TITLE\", \"PUBLIC\".\"PRODUCTS\".\"CATEGORY\" AS \"CATEGORY\", \"PUBLIC\".\"PRODUCTS\".\"VENDOR\" AS \"VENDOR\", \"PUBLIC\".\"PRODUCTS\".\"PRICE\" AS \"PRICE\", \"PUBLIC\".\"PRODUCTS\".\"RATING\" AS \"RATING\", \"PUBLIC\".\"PRODUCTS\".\"CREATED_AT\" AS \"CREATED_AT\", (\"PUBLIC\".\"PRODUCTS\".\"PRICE\" + 2) AS \"Price_2\" FROM \"PUBLIC\".\"PRODUCTS\") \"source\" LIMIT 2000 [42122-197]",
:row_count 0,
:running_time 0,
:preprocessed
{:database 4,
:query
{:source-card-id 867,
:source-metadata
[{:description
"The numerical product number. Only used internally. All external communication should use the title or EAN.",
:semantic_type :type/PK,
:coercion_strategy nil,
:name "ID",
:settings nil,
:field_ref [:field 105 nil],
:effective_type :type/BigInteger,
:id 105,
:display_name "ID",
:fingerprint nil,
:base_type :type/BigInteger}
{:description "The name of the product as it should be displayed to customers.",
:semantic_type :type/Title,
:coercion_strategy nil,
:name "TITLE",
:settings nil,
:field_ref [:field 106 nil],
:effective_type :type/Text,
:id 106,
:display_name "Title",
:fingerprint
{:global {:distinct-count 199, :nil% 0.0},
:type
{:type/Text
{:percent-json 0.0, :percent-url 0.0, :percent-email 0.0, :percent-state 0.0, :average-length 21.495}}},
:base_type :type/Text}
{:description "The type of product, valid values include: Doohicky, Gadget, Gizmo and Widget",
:semantic_type :type/Category,
:coercion_strategy nil,
:name "CATEGORY",
:settings nil,
:field_ref [:field 107 nil],
:effective_type :type/Text,
:id 107,
:display_name "Category",
:fingerprint
{:global {:distinct-count 4, :nil% 0.0},
:type
{:type/Text
{:percent-json 0.0, :percent-url 0.0, :percent-email 0.0, :percent-state 0.0, :average-length 6.375}}},
:base_type :type/Text}
{:description "The source of the product.",
:semantic_type :type/Company,
:coercion_strategy nil,
:name "VENDOR",
:settings nil,
:field_ref [:field 109 nil],
:effective_type :type/Text,
:id 109,
:display_name "Vendor",
:fingerprint
{:global {:distinct-count 200, :nil% 0.0},
:type
{:type/Text
{:percent-json 0.0, :percent-url 0.0, :percent-email 0.0, :percent-state 0.0, :average-length 20.6}}},
:base_type :type/Text}
{:description
"The list price of the product. Note that this is not always the price the product sold for due to discounts, promotions, etc.",
:semantic_type :type/Price,
:coercion_strategy nil,
:name "PRICE",
:settings {},
:field_ref [:field 104 nil],
:effective_type :type/Float,
:id 104,
:display_name "Price",
:fingerprint
{:global {:distinct-count 170, :nil% 0.0},
:type
{:type/Number
{:min 15.691943673970439,
:q1 37.25154462926434,
:q3 75.45898071609447,
:max 98.81933684368194,
:sd 21.711481557852057,
:avg 55.74639966792074}}},
:base_type :type/Float}
{:description "The average rating users have given the product. This ranges from 1 - 5",
:semantic_type :type/Score,
:coercion_strategy nil,
:name "RATING",
:settings nil,
:field_ref [:field 103 nil],
:effective_type :type/Float,
:id 103,
:display_name "Rating",
:fingerprint
{:global {:distinct-count 23, :nil% 0.0},
:type
{:type/Number
{:min 0.0, :q1 3.5120465053408525, :q3 4.216124969497314, :max 5.0, :sd 1.3605488657451452, :avg 3.4715}}},
:base_type :type/Float}
{:description "The date the product was added to our catalog.",
:semantic_type :type/CreationTimestamp,
:coercion_strategy nil,
:unit :default,
:name "CREATED_AT",
:settings {:date_separator ".", :time_style "k:mm"},
:field_ref [:field 108 {:temporal-unit :default}],
:effective_type :type/DateTime,
:id 108,
:display_name "Created At",
:fingerprint
{:global {:distinct-count 200, :nil% 0.0},
:type {:type/DateTime {:earliest "2016-04-26T19:29:55.147Z", :latest "2019-04-15T13:34:19.931Z"}}},
:base_type :type/DateTime}
{:display_name "Price",
:field_ref [:expression "Price"],
:name "Price",
:base_type :type/Float,
:effective_type :type/Float,
:semantic_type nil,
:fingerprint
{:global {:distinct-count 170, :nil% 0.0},
:type
{:type/Number
{:min 17.69194367397044,
:q1 39.25154462926434,
:q3 77.45898071609447,
:max 100.81933684368194,
:sd 21.711481557852057,
:avg 57.74639966792074}}}}],
:fields
[[:field 105 nil]
[:field 106 nil]
[:field 107 nil]
[:field 109 nil]
[:field 104 nil]
[:field 103 nil]
[:field 108 {:temporal-unit :default}]
[:field "Price" {:base-type :type/Float}]],
:source-query
{:source-table 10,
:expressions {:Price [:+ [:field 104 nil] 2]},
:fields
[[:field 105 nil]
[:field 106 nil]
[:field 107 nil]
[:field 109 nil]
[:field 104 nil]
[:field 103 nil]
[:field 108 {:temporal-unit :default}]
[:expression "Price"]]},
:limit 2000},
:type :query,
:middleware {:js-int-to-string? true, :add-default-userland-constraints? true},
:info
{:executed-by 1,
:context :ad-hoc,
:card-id 867,
:nested? true,
:query-hash
[41, -108, 1, -114, -56, 18, -94, -33, -121, -1, -24, -1, -33, 20, -27, 28, -60, 51, -21, -70, -25, 16, -1, -8, 82,
67, 23, -118, 108, 12, 15, -94]},
:constraints {:max-results 10000, :max-results-bare-rows 2000}},
:data {:rows [], :cols []}}
2022-03-21 16:01:31,731 DEBUG middleware.log :: POST /api/dataset 202 [ASYNC: completed] 1.1 s (22 DB calls) App DB connections: 0/13 Jetty threads: 3/50 (4 idle, 0 queued) (96 total active threads) Queries in flight: 0 (0 queued); h2 DB 4 connections: 0/1 (0 threads blocked)
```
</details>
**Information about your Metabase Installation:**
Tested 0.41.6 thru 0.42.2 - regression since 0.42.0
|
non_defect
|
custom column aliasing can cause incorrect results or wrong alias on nested queries describe the bug custom columns with names similar to existing table column names causes failed queries or incorrect results depending on database on bigquery it does not fail but shows wrong results perhaps other databases too almost a repeat of but for nested queries regression since for some database while bigquery would fail previously with ambiguous errors we might have a frontend issue too since it s possible to self reference a custom column which causes no results to reproduce question bigquery sample products custom column as price yes the name of the custom column is the same as the underlying column name save the question and use it as the base in a new question or alternatively convert it to a model would cause the same problem generated sql for bigquery select source id as id source ean as ean source title as title source category as category source vendor as vendor source price as price source rating as rating source created at as created at source price as price from select sample dataset products id as id sample dataset products ean as ean sample dataset products title as title sample dataset products category as category sample dataset products vendor as vendor sample dataset products price as price sample dataset products rating as rating sample dataset products created at as created at sample dataset products price as price from sample dataset products source limit doing similar repro with would fail on step with column source price not found if the custom column was called price all uppercase then it works as expected guess it might just be a thing generated sql for select source id as id source title as title source category as category source vendor as vendor source price as price source rating as rating source created at as created at source price as price from select public products id as id public products title as title public products category as category public products vendor as vendor public products price as price public products rating as rating public products created at as created at public products price as price from public products source limit full stacktrace for info api dataset source query for this query is card info middleware fetch source query fetched source query from card source table expressions price info middleware fetch source query fetched source query from card source table expressions price error middleware catch exceptions error processing query null database id started at t via status failed class clojure lang exceptioninfo error error executing query column source price not found sql statement n metabase userid querytype mbql queryhash nselect source id as id source title as title source category as category source vendor as vendor source price as price source rating as rating source created at as created at source price as price from select public products id as id public products title as title public products category as category public products vendor as vendor public products price as price public products rating as rating public products created at as created at public products price as price from public products source limit stacktrace driver sql jdbc execute execute reducible query fn invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc fn invokestatic sql jdbc clj driver sql jdbc fn invoke sql jdbc clj driver fn invokestatic clj driver fn invoke clj query processor context executef invokestatic context clj query processor context executef invoke context clj query processor context default default runf invokestatic default clj query processor context default default runf invoke default clj query processor context runf invokestatic context clj query processor context runf invoke context clj query processor reducible pivot invokestatic reducible clj query processor reducible pivot invoke reducible clj query processor middleware mbql to native mbql gt native fn invoke mbql to native clj query processor middleware check features check features fn invoke check features clj query processor middleware limit limit fn invoke limit clj query processor middleware cache maybe return cached results fn invoke cache clj query processor middleware optimize temporal filters optimize temporal filters fn invoke optimize temporal filters clj query processor middleware validate temporal bucketing validate temporal bucketing fn invoke validate temporal bucketing clj query processor middleware auto parse filter values auto parse filter values fn invoke auto parse filter values clj query processor middleware wrap value literals wrap value literals fn invoke wrap value literals clj query processor middleware annotate add column info fn invoke annotate clj query processor middleware permissions check query permissions fn invoke permissions clj query processor middleware pre alias aggregations pre alias aggregations fn invoke pre alias aggregations clj query processor middleware cumulative aggregations handle cumulative aggregations fn invoke cumulative aggregations clj query processor middleware visualization settings update viz settings fn invoke visualization settings clj query processor middleware escape join aliases escape join aliases middleware fn invoke escape join aliases clj query processor middleware fix bad references fix bad references middleware fn invoke fix bad references clj query processor middleware resolve joined fields resolve joined fields fn invoke resolve joined fields clj query processor middleware resolve joins resolve joins fn invoke resolve joins clj query processor middleware add implicit joins add implicit joins fn invoke add implicit joins clj query processor middleware large int id convert id to string fn invoke large int id clj query processor middleware format rows format rows fn invoke format rows clj query processor middleware add default temporal unit add default temporal unit fn invoke add default temporal unit clj query processor middleware desugar desugar fn invoke desugar clj query processor middleware binning update binning strategy fn invoke binning clj query processor middleware resolve fields resolve fields fn invoke resolve fields clj query processor middleware add dimension projections add remapping fn invoke add dimension projections clj query processor middleware add implicit clauses add implicit clauses fn invoke add implicit clauses clj query processor middleware upgrade field literals upgrade field literals fn invoke upgrade field literals clj query processor middleware add source metadata add source metadata for source queries fn invoke add source metadata clj query processor middleware reconcile breakout and order by bucketing reconcile breakout and order by bucketing fn invoke reconcile breakout and order by bucketing clj query processor middleware auto bucket datetimes auto bucket datetimes fn invoke auto bucket datetimes clj query processor middleware resolve source table resolve source tables fn invoke resolve source table clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor middleware constraints add default userland constraints fn invoke constraints clj query processor middleware process userland query process userland query fn invoke process userland query clj query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor process userland query invokestatic query processor clj query processor process userland query doinvoke query processor clj query processor fn process query and save execution bang fn invoke query processor clj query processor fn process query and save execution bang invoke query processor clj query processor fn process query and save with max results constraints bang fn invoke query processor clj query processor fn process query and save with max results constraints bang invoke query processor clj api dataset run query async fn invoke dataset clj query processor streaming streaming response star fn fn invoke streaming clj query processor streaming streaming response star fn invoke streaming clj async streaming response do f star invokestatic streaming response clj async streaming response do f star invoke streaming response clj async streaming response do f async task invoke streaming response clj error type invalid query ex data sql metabase userid querytype mbql queryhash nselect source id as id source title as title source category as category source vendor as vendor source price as price source rating as rating source created at as created at source price as price from select public products id as id public products title as title public products category as category public products vendor as vendor public products price as price public products rating as rating public products created at as created at public products price as price from public products source limit params nil type invalid query state error type invalid query json query database query source table card type query parameters middleware js int to string true add default userland constraints true native query select source id as id source title as title source category as category source vendor as vendor source price as price source rating as rating source created at as created at source price as price from select public products id as id public products title as title public products category as category public products vendor as vendor public products price as price public products rating as rating public products created at as created at public products price as price from public products source limit params nil status failed class org jdbc jdbcsqlexception stacktrace org message dbexception getjdbcsqlexception dbexception java org message dbexception get dbexception java org message dbexception get dbexception java org expression expressioncolumn optimize expressioncolumn java org expression alias optimize alias java org command dml select prepare select java org command parser preparecommand parser java org engine session preparelocal session java org engine session preparecommand session java org jdbc jdbcconnection preparecommand jdbcconnection java org jdbc jdbcstatement executeinternal jdbcstatement java org jdbc jdbcstatement execute jdbcstatement java com mchange impl newproxystatement execute newproxystatement java driver sql jdbc execute fn invokestatic execute clj driver sql jdbc execute fn invoke execute clj driver sql jdbc execute execute statement or prepared statement bang invokestatic execute clj driver sql jdbc execute execute statement or prepared statement bang invoke execute clj driver sql jdbc execute execute reducible query fn invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc fn invokestatic sql jdbc clj driver sql jdbc fn invoke sql jdbc clj driver fn invokestatic clj driver fn invoke clj query processor context executef invokestatic context clj query processor context executef invoke context clj query processor context default default runf invokestatic default clj query processor context default default runf invoke default clj query processor context runf invokestatic context clj query processor context runf invoke context clj query processor reducible pivot invokestatic reducible clj query processor reducible pivot invoke reducible clj query processor middleware mbql to native mbql gt native fn invoke mbql to native clj query processor middleware check features check features fn invoke check features clj query processor middleware limit limit fn invoke limit clj query processor middleware cache maybe return cached results fn invoke cache clj query processor middleware optimize temporal filters optimize temporal filters fn invoke optimize temporal filters clj query processor middleware validate temporal bucketing validate temporal bucketing fn invoke validate temporal bucketing clj query processor middleware auto parse filter values auto parse filter values fn invoke auto parse filter values clj query processor middleware wrap value literals wrap value literals fn invoke wrap value literals clj query processor middleware annotate add column info fn invoke annotate clj query processor middleware permissions check query permissions fn invoke permissions clj query processor middleware pre alias aggregations pre alias aggregations fn invoke pre alias aggregations clj query processor middleware cumulative aggregations handle cumulative aggregations fn invoke cumulative aggregations clj query processor middleware visualization settings update viz settings fn invoke visualization settings clj query processor middleware escape join aliases escape join aliases middleware fn invoke escape join aliases clj query processor middleware fix bad references fix bad references middleware fn invoke fix bad references clj query processor middleware resolve joined fields resolve joined fields fn invoke resolve joined fields clj query processor middleware resolve joins resolve joins fn invoke resolve joins clj query processor middleware add implicit joins add implicit joins fn invoke add implicit joins clj query processor middleware large int id convert id to string fn invoke large int id clj query processor middleware format rows format rows fn invoke format rows clj query processor middleware add default temporal unit add default temporal unit fn invoke add default temporal unit clj query processor middleware desugar desugar fn invoke desugar clj query processor middleware binning update binning strategy fn invoke binning clj query processor middleware resolve fields resolve fields fn invoke resolve fields clj query processor middleware add dimension projections add remapping fn invoke add dimension projections clj query processor middleware add implicit clauses add implicit clauses fn invoke add implicit clauses clj query processor middleware upgrade field literals upgrade field literals fn invoke upgrade field literals clj query processor middleware add source metadata add source metadata for source queries fn invoke add source metadata clj query processor middleware reconcile breakout and order by bucketing reconcile breakout and order by bucketing fn invoke reconcile breakout and order by bucketing clj query processor middleware auto bucket datetimes auto bucket datetimes fn invoke auto bucket datetimes clj query processor middleware resolve source table resolve source tables fn invoke resolve source table clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor middleware constraints add default userland constraints fn invoke constraints clj query processor middleware process userland query process userland query fn invoke process userland query clj query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor process userland query invokestatic query processor clj query processor process userland query doinvoke query processor clj query processor fn process query and save execution bang fn invoke query processor clj query processor fn process query and save execution bang invoke query processor clj query processor fn process query and save with max results constraints bang fn invoke query processor clj query processor fn process query and save with max results constraints bang invoke query processor clj api dataset run query async fn invoke dataset clj query processor streaming streaming response star fn fn invoke streaming clj query processor streaming streaming response star fn invoke streaming clj async streaming response do f star invokestatic streaming response clj async streaming response do f star invoke streaming response clj async streaming response do f async task invoke streaming response clj card id context ad hoc error column source price not found sql statement n metabase userid querytype mbql queryhash nselect source id as id source title as title source category as category source vendor as vendor source price as price source rating as rating source created at as created at source price as price from select public products id as id public products title as title public products category as category public products vendor as vendor public products price as price public products rating as rating public products created at as created at public products price as price from public products source limit row count running time preprocessed database query source card id source metadata description the numerical product number only used internally all external communication should use the title or ean semantic type type pk coercion strategy nil name id settings nil field ref effective type type biginteger id display name id fingerprint nil base type type biginteger description the name of the product as it should be displayed to customers semantic type type title coercion strategy nil name title settings nil field ref effective type type text id display name title fingerprint global distinct count nil type type text percent json percent url percent email percent state average length base type type text description the type of product valid values include doohicky gadget gizmo and widget semantic type type category coercion strategy nil name category settings nil field ref effective type type text id display name category fingerprint global distinct count nil type type text percent json percent url percent email percent state average length base type type text description the source of the product semantic type type company coercion strategy nil name vendor settings nil field ref effective type type text id display name vendor fingerprint global distinct count nil type type text percent json percent url percent email percent state average length base type type text description the list price of the product note that this is not always the price the product sold for due to discounts promotions etc semantic type type price coercion strategy nil name price settings field ref effective type type float id display name price fingerprint global distinct count nil type type number min max sd avg base type type float description the average rating users have given the product this ranges from semantic type type score coercion strategy nil name rating settings nil field ref effective type type float id display name rating fingerprint global distinct count nil type type number min max sd avg base type type float description the date the product was added to our catalog semantic type type creationtimestamp coercion strategy nil unit default name created at settings date separator time style k mm field ref effective type type datetime id display name created at fingerprint global distinct count nil type type datetime earliest latest base type type datetime display name price field ref name price base type type float effective type type float semantic type nil fingerprint global distinct count nil type type number min max sd avg fields source query source table expressions price fields limit type query middleware js int to string true add default userland constraints true info executed by context ad hoc card id nested true query hash constraints max results max results bare rows data rows cols debug middleware log post api dataset s db calls app db connections jetty threads idle queued total active threads queries in flight queued db connections threads blocked information about your metabase installation tested thru regression since
| 0
|
24,391
| 4,076,810,274
|
IssuesEvent
|
2016-05-30 03:10:06
|
xcat2/xcat-core
|
https://api.github.com/repos/xcat2/xcat-core
|
opened
|
[FVT]Please update x_ubuntu_cmd.bundle file to make the "Full_installation_flat_docker" at the beginning
|
component:test
|
@junxiawang
In order to make the docker installation is the first case of ubuntu x86, please update x_ubuntu_cmd.bundle file to make the "Full_installation_flat_docker" at the beginning of the file.
|
1.0
|
[FVT]Please update x_ubuntu_cmd.bundle file to make the "Full_installation_flat_docker" at the beginning - @junxiawang
In order to make the docker installation is the first case of ubuntu x86, please update x_ubuntu_cmd.bundle file to make the "Full_installation_flat_docker" at the beginning of the file.
|
non_defect
|
please update x ubuntu cmd bundle file to make the full installation flat docker at the beginning junxiawang in order to make the docker installation is the first case of ubuntu please update x ubuntu cmd bundle file to make the full installation flat docker at the beginning of the file
| 0
|
5,841
| 2,610,216,472
|
IssuesEvent
|
2015-02-26 19:08:59
|
chrsmith/somefinders
|
https://api.github.com/repos/chrsmith/somefinders
|
opened
|
сигнализация tiger evolution инструкция
|
auto-migrated Priority-Medium Type-Defect
|
```
'''Август Маслов'''
День добрый никак не могу найти
.сигнализация tiger evolution инструкция. как то
выкладывали уже
'''Вилли Русаков'''
Качай тут http://bit.ly/16pKem3
'''Гермоген Пономарёв'''
Просит ввести номер мобилы!Не опасно ли это?
'''Валериан Соболев'''
Не это не влияет на баланс
'''Всемил Шилов'''
Не это не влияет на баланс
Информация о файле: сигнализация tiger evolution
инструкция
Загружен: В этом месяце
Скачан раз: 655
Рейтинг: 1175
Средняя скорость скачивания: 1275
Похожих файлов: 35
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 7:47
|
1.0
|
сигнализация tiger evolution инструкция - ```
'''Август Маслов'''
День добрый никак не могу найти
.сигнализация tiger evolution инструкция. как то
выкладывали уже
'''Вилли Русаков'''
Качай тут http://bit.ly/16pKem3
'''Гермоген Пономарёв'''
Просит ввести номер мобилы!Не опасно ли это?
'''Валериан Соболев'''
Не это не влияет на баланс
'''Всемил Шилов'''
Не это не влияет на баланс
Информация о файле: сигнализация tiger evolution
инструкция
Загружен: В этом месяце
Скачан раз: 655
Рейтинг: 1175
Средняя скорость скачивания: 1275
Похожих файлов: 35
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 7:47
|
defect
|
сигнализация tiger evolution инструкция август маслов день добрый никак не могу найти сигнализация tiger evolution инструкция как то выкладывали уже вилли русаков качай тут гермоген пономарёв просит ввести номер мобилы не опасно ли это валериан соболев не это не влияет на баланс всемил шилов не это не влияет на баланс информация о файле сигнализация tiger evolution инструкция загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
| 1
|
178,634
| 21,509,443,728
|
IssuesEvent
|
2022-04-28 01:41:52
|
bsbtd/Teste
|
https://api.github.com/repos/bsbtd/Teste
|
closed
|
CVE-2020-17510 (High) detected in shiro-web-1.5.0.jar - autoclosed
|
security vulnerability
|
## CVE-2020-17510 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shiro-web-1.5.0.jar</b></p></summary>
<p>Apache Shiro is a powerful and flexible open-source security framework that cleanly handles
authentication, authorization, enterprise session management, single sign-on and cryptography services.</p>
<p>Path to dependency file: /OpenLegislation/pom.xml</p>
<p>Path to vulnerable library: /.m2/repository/org/apache/shiro/shiro-web/1.5.0/shiro-web-1.5.0.jar,/16064326_QOFTMV/downloadResource_GWKSRP/20220416070814/shiro-web-1.5.0.jar</p>
<p>
Dependency Hierarchy:
- :x: **shiro-web-1.5.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Shiro before 1.7.0, when using Apache Shiro with Spring, a specially crafted HTTP request may cause an authentication bypass.
<p>Publish Date: 2020-11-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-17510>CVE-2020-17510</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/rc2cff2538b683d480426393eecf1ce8dd80e052fbef49303b4f47171%40%3Cdev.shiro.apache.org%3E">https://lists.apache.org/thread.html/rc2cff2538b683d480426393eecf1ce8dd80e052fbef49303b4f47171%40%3Cdev.shiro.apache.org%3E</a></p>
<p>Release Date: 2020-11-05</p>
<p>Fix Resolution: org.apache.shiro:shiro-web:1.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-17510 (High) detected in shiro-web-1.5.0.jar - autoclosed - ## CVE-2020-17510 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shiro-web-1.5.0.jar</b></p></summary>
<p>Apache Shiro is a powerful and flexible open-source security framework that cleanly handles
authentication, authorization, enterprise session management, single sign-on and cryptography services.</p>
<p>Path to dependency file: /OpenLegislation/pom.xml</p>
<p>Path to vulnerable library: /.m2/repository/org/apache/shiro/shiro-web/1.5.0/shiro-web-1.5.0.jar,/16064326_QOFTMV/downloadResource_GWKSRP/20220416070814/shiro-web-1.5.0.jar</p>
<p>
Dependency Hierarchy:
- :x: **shiro-web-1.5.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Shiro before 1.7.0, when using Apache Shiro with Spring, a specially crafted HTTP request may cause an authentication bypass.
<p>Publish Date: 2020-11-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-17510>CVE-2020-17510</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/rc2cff2538b683d480426393eecf1ce8dd80e052fbef49303b4f47171%40%3Cdev.shiro.apache.org%3E">https://lists.apache.org/thread.html/rc2cff2538b683d480426393eecf1ce8dd80e052fbef49303b4f47171%40%3Cdev.shiro.apache.org%3E</a></p>
<p>Release Date: 2020-11-05</p>
<p>Fix Resolution: org.apache.shiro:shiro-web:1.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in shiro web jar autoclosed cve high severity vulnerability vulnerable library shiro web jar apache shiro is a powerful and flexible open source security framework that cleanly handles authentication authorization enterprise session management single sign on and cryptography services path to dependency file openlegislation pom xml path to vulnerable library repository org apache shiro shiro web shiro web jar qoftmv downloadresource gwksrp shiro web jar dependency hierarchy x shiro web jar vulnerable library found in head commit a href vulnerability details apache shiro before when using apache shiro with spring a specially crafted http request may cause an authentication bypass publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache shiro shiro web step up your open source security game with whitesource
| 0
|
275,872
| 20,960,147,620
|
IssuesEvent
|
2022-03-27 17:22:18
|
project-serum/anchor
|
https://api.github.com/repos/project-serum/anchor
|
closed
|
lang: add docs for `emit!` macro
|
documentation lang
|
currently there's almost nothing there.
Should include an example
https://docs.rs/anchor-lang/latest/anchor_lang/macro.emit.html
|
1.0
|
lang: add docs for `emit!` macro - currently there's almost nothing there.
Should include an example
https://docs.rs/anchor-lang/latest/anchor_lang/macro.emit.html
|
non_defect
|
lang add docs for emit macro currently there s almost nothing there should include an example
| 0
|
58,804
| 16,769,691,640
|
IssuesEvent
|
2021-06-14 13:27:53
|
line/armeria
|
https://api.github.com/repos/line/armeria
|
closed
|
Which event loop should late subscriber signals be run on
|
defect
|
Not sure if this is intentional behavior, so wanted to ask.
Currently, it seems like for a late subscriber, the current subscription's executor is used to signal events.
From the code level, it seemed like `AbstractStreamMessage` makes a deliberate effort to ensure the currently subscribed executor is always used.
https://github.com/line/armeria/blob/4256dec753213d091d98fe160fefe8385787e0eb/core/src/main/java/com/linecorp/armeria/common/stream/AbstractStreamMessage.java#L70
Instead, I'm wondering if it is more consistent behavior to run the signal on the late subscriber's event loop.
**Current**
If `subscriber1` subscribes with `eventLoop1`, and `subscriber2` subscribers with `eventLoop2`, `subscriber2` would receive signals via `eventLoop1`
**Proposal**
If `subscriber1` subscribes with `eventLoop1`, and `subscriber2` subscribers with `eventLoop2`, `subscriber2` would receive signals via `eventLoop2`
---
For instance, if a user `abort`s a stream and then `subscribe`s, the signal will always be run on the executing thread.
(the following test fails due to this behavior)
```
@Test
void testLateSubscribe() {
final DefaultStreamMessage<String> streamMessage = new DefaultStreamMessage<>();
streamMessage.abort();
final AtomicBoolean completed = new AtomicBoolean();
final EventLoop executor = eventLoop.get();
streamMessage.subscribe(new Subscriber<String>() {
@Override
public void onSubscribe(Subscription s) {
assertThat(executor.inEventLoop()).isTrue();
s.request(1);
}
@Override
public void onNext(String s) {}
@Override
public void onError(Throwable t) {
completed.set(true);
}
@Override
public void onComplete() {
}
}, executor);
await().untilTrue(completed);
}
```
This is me asking without previous context, so let me know if this isn't a good idea
NOTE: For more context on why I'm asking this, see `subscribe` at #3604. An aborted message would end up signaling late subscribers using the internal event loop. This would result in inconsistent behavior across `StreamMessage`s.
|
1.0
|
Which event loop should late subscriber signals be run on - Not sure if this is intentional behavior, so wanted to ask.
Currently, it seems like for a late subscriber, the current subscription's executor is used to signal events.
From the code level, it seemed like `AbstractStreamMessage` makes a deliberate effort to ensure the currently subscribed executor is always used.
https://github.com/line/armeria/blob/4256dec753213d091d98fe160fefe8385787e0eb/core/src/main/java/com/linecorp/armeria/common/stream/AbstractStreamMessage.java#L70
Instead, I'm wondering if it is more consistent behavior to run the signal on the late subscriber's event loop.
**Current**
If `subscriber1` subscribes with `eventLoop1`, and `subscriber2` subscribers with `eventLoop2`, `subscriber2` would receive signals via `eventLoop1`
**Proposal**
If `subscriber1` subscribes with `eventLoop1`, and `subscriber2` subscribers with `eventLoop2`, `subscriber2` would receive signals via `eventLoop2`
---
For instance, if a user `abort`s a stream and then `subscribe`s, the signal will always be run on the executing thread.
(the following test fails due to this behavior)
```
@Test
void testLateSubscribe() {
final DefaultStreamMessage<String> streamMessage = new DefaultStreamMessage<>();
streamMessage.abort();
final AtomicBoolean completed = new AtomicBoolean();
final EventLoop executor = eventLoop.get();
streamMessage.subscribe(new Subscriber<String>() {
@Override
public void onSubscribe(Subscription s) {
assertThat(executor.inEventLoop()).isTrue();
s.request(1);
}
@Override
public void onNext(String s) {}
@Override
public void onError(Throwable t) {
completed.set(true);
}
@Override
public void onComplete() {
}
}, executor);
await().untilTrue(completed);
}
```
This is me asking without previous context, so let me know if this isn't a good idea
NOTE: For more context on why I'm asking this, see `subscribe` at #3604. An aborted message would end up signaling late subscribers using the internal event loop. This would result in inconsistent behavior across `StreamMessage`s.
|
defect
|
which event loop should late subscriber signals be run on not sure if this is intentional behavior so wanted to ask currently it seems like for a late subscriber the current subscription s executor is used to signal events from the code level it seemed like abstractstreammessage makes a deliberate effort to ensure the currently subscribed executor is always used instead i m wondering if it is more consistent behavior to run the signal on the late subscriber s event loop current if subscribes with and subscribers with would receive signals via proposal if subscribes with and subscribers with would receive signals via for instance if a user abort s a stream and then subscribe s the signal will always be run on the executing thread the following test fails due to this behavior test void testlatesubscribe final defaultstreammessage streammessage new defaultstreammessage streammessage abort final atomicboolean completed new atomicboolean final eventloop executor eventloop get streammessage subscribe new subscriber override public void onsubscribe subscription s assertthat executor ineventloop istrue s request override public void onnext string s override public void onerror throwable t completed set true override public void oncomplete executor await untiltrue completed this is me asking without previous context so let me know if this isn t a good idea note for more context on why i m asking this see subscribe at an aborted message would end up signaling late subscribers using the internal event loop this would result in inconsistent behavior across streammessage s
| 1
|
32,366
| 6,767,389,113
|
IssuesEvent
|
2017-10-26 03:00:48
|
Shopkeepers/Shopkeepers
|
https://api.github.com/repos/Shopkeepers/Shopkeepers
|
closed
|
An internal error has occurred.
|
Defect invalid migrated
|
**Migrated from:** https://dev.bukkit.org/projects/shopkeepers/issues/71
**Originally posted by DayneOram (Dec 30, 2012):**
What steps will reproduce the problem?
1. Difficult to say, it seems to be a problem with us or our server. I simply installed it and tried to use it.
2.
3.What is the expected output? What do you see instead?
After typing /shopkeeper (like I did when I had a previous version of ShopKeepers on a previous map), I expect a shopkeeper to appear and remain still. After holding shift and right clicking the shopkeeper, I should get the extra menu as I used to.
Instead, a standard villager with a default trade setup appears, and he runs to the nearest door like a regular villager would.It also tells me "An internal error has occurred while attempting to perform this command."What version of the product are you using?1.9, as installed by Multicraft. The jar file appears to be there, as does the Shopkeepers folder. I tried removing these and then reinstalling it yet again, but the same problem occurs. Server is on 1.4.6.Do you have an error log of what happened?
30.12 16:08:36 [Server] INFO ... 22 more
30.12 16:08:36 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source)
30.12 16:08:36 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source)
30.12 16:08:36 [Server] INFO at org.bukkit.plugin.java.PluginClassLoader.findClass(PluginClassLoader.java:53)
30.12 16:08:36 [Server] INFO at org.bukkit.plugin.java.PluginClassLoader.findClass0(PluginClassLoader.java:70)
30.12 16:08:36 [Server] INFO Caused by: java.lang.ClassNotFoundException: net.minecraft.server.PathfinderGoal
30.12 16:08:36 [Server] INFO ... 15 more
30.12 16:08:36 [Server] INFO at org.bukkit.command.PluginCommand.execute(PluginCommand.java:44)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.ShopkeepersPlugin.onCommand(ShopkeepersPlugin.java:302)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.ShopkeepersPlugin.createNewAdminShopkeeper(ShopkeepersPlugin.java:332)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.Shopkeeper.spawn(Shopkeeper.java:87)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.shopobjects.VillagerShop.spawn(VillagerShop.java:36)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.shopobjects.LivingEntityShop.spawn(LivingEntityShop.java:64)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.shopobjects.VillagerShop.overwriteAI(VillagerShop.java:73)
30.12 16:08:36 [Server] INFO Caused by: java.lang.NoClassDefFoundError: net/minecraft/server/PathfinderGoal
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.ThreadServerApplication.run(SourceFile:849)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.MinecraftServer.run(MinecraftServer.java:427)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.MinecraftServer.q(MinecraftServer.java:494)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.DedicatedServer.r(DedicatedServer.java:224)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.MinecraftServer.r(MinecraftServer.java:598)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.DedicatedServerConnection.b(SourceFile:30)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.ServerConnection.b(SourceFile:39)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.PlayerConnection.d(PlayerConnection.java:112)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.NetworkManager.b(NetworkManager.java:290)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.Packet3Chat.handle(Packet3Chat.java:44)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.PlayerConnection.a(PlayerConnection.java:853)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.PlayerConnection.chat(PlayerConnection.java:898)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.PlayerConnection.handleCommand(PlayerConnection.java:980)
30.12 16:08:36 [Server] INFO at org.bukkit.craftbukkit.v1_4_6.CraftServer.dispatchCommand(CraftServer.java:510)
30.12 16:08:36 [Server] INFO at org.bukkit.command.SimpleCommandMap.dispatch(SimpleCommandMap.java:186)
30.12 16:08:36 [Server] INFO at org.bukkit.command.PluginCommand.execute(PluginCommand.java:46)
30.12 16:08:36 [Server] INFO org.bukkit.command.CommandException: Unhandled exception executing command 'shopkeeper' in plugin Shopkeepers v1.9
30.12 16:08:36 [Server] SEVERE nullPlease provide any additional information below.
|
1.0
|
An internal error has occurred. - **Migrated from:** https://dev.bukkit.org/projects/shopkeepers/issues/71
**Originally posted by DayneOram (Dec 30, 2012):**
What steps will reproduce the problem?
1. Difficult to say, it seems to be a problem with us or our server. I simply installed it and tried to use it.
2.
3.What is the expected output? What do you see instead?
After typing /shopkeeper (like I did when I had a previous version of ShopKeepers on a previous map), I expect a shopkeeper to appear and remain still. After holding shift and right clicking the shopkeeper, I should get the extra menu as I used to.
Instead, a standard villager with a default trade setup appears, and he runs to the nearest door like a regular villager would.It also tells me "An internal error has occurred while attempting to perform this command."What version of the product are you using?1.9, as installed by Multicraft. The jar file appears to be there, as does the Shopkeepers folder. I tried removing these and then reinstalling it yet again, but the same problem occurs. Server is on 1.4.6.Do you have an error log of what happened?
30.12 16:08:36 [Server] INFO ... 22 more
30.12 16:08:36 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source)
30.12 16:08:36 [Server] INFO at java.lang.ClassLoader.loadClass(Unknown Source)
30.12 16:08:36 [Server] INFO at org.bukkit.plugin.java.PluginClassLoader.findClass(PluginClassLoader.java:53)
30.12 16:08:36 [Server] INFO at org.bukkit.plugin.java.PluginClassLoader.findClass0(PluginClassLoader.java:70)
30.12 16:08:36 [Server] INFO Caused by: java.lang.ClassNotFoundException: net.minecraft.server.PathfinderGoal
30.12 16:08:36 [Server] INFO ... 15 more
30.12 16:08:36 [Server] INFO at org.bukkit.command.PluginCommand.execute(PluginCommand.java:44)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.ShopkeepersPlugin.onCommand(ShopkeepersPlugin.java:302)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.ShopkeepersPlugin.createNewAdminShopkeeper(ShopkeepersPlugin.java:332)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.Shopkeeper.spawn(Shopkeeper.java:87)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.shopobjects.VillagerShop.spawn(VillagerShop.java:36)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.shopobjects.LivingEntityShop.spawn(LivingEntityShop.java:64)
30.12 16:08:36 [Server] INFO at com.nisovin.shopkeepers.shopobjects.VillagerShop.overwriteAI(VillagerShop.java:73)
30.12 16:08:36 [Server] INFO Caused by: java.lang.NoClassDefFoundError: net/minecraft/server/PathfinderGoal
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.ThreadServerApplication.run(SourceFile:849)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.MinecraftServer.run(MinecraftServer.java:427)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.MinecraftServer.q(MinecraftServer.java:494)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.DedicatedServer.r(DedicatedServer.java:224)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.MinecraftServer.r(MinecraftServer.java:598)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.DedicatedServerConnection.b(SourceFile:30)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.ServerConnection.b(SourceFile:39)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.PlayerConnection.d(PlayerConnection.java:112)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.NetworkManager.b(NetworkManager.java:290)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.Packet3Chat.handle(Packet3Chat.java:44)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.PlayerConnection.a(PlayerConnection.java:853)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.PlayerConnection.chat(PlayerConnection.java:898)
30.12 16:08:36 [Server] INFO at net.minecraft.server.v1_4_6.PlayerConnection.handleCommand(PlayerConnection.java:980)
30.12 16:08:36 [Server] INFO at org.bukkit.craftbukkit.v1_4_6.CraftServer.dispatchCommand(CraftServer.java:510)
30.12 16:08:36 [Server] INFO at org.bukkit.command.SimpleCommandMap.dispatch(SimpleCommandMap.java:186)
30.12 16:08:36 [Server] INFO at org.bukkit.command.PluginCommand.execute(PluginCommand.java:46)
30.12 16:08:36 [Server] INFO org.bukkit.command.CommandException: Unhandled exception executing command 'shopkeeper' in plugin Shopkeepers v1.9
30.12 16:08:36 [Server] SEVERE nullPlease provide any additional information below.
|
defect
|
an internal error has occurred migrated from originally posted by dayneoram dec what steps will reproduce the problem difficult to say it seems to be a problem with us or our server i simply installed it and tried to use it what is the expected output what do you see instead after typing shopkeeper like i did when i had a previous version of shopkeepers on a previous map i expect a shopkeeper to appear and remain still after holding shift and right clicking the shopkeeper i should get the extra menu as i used to instead a standard villager with a default trade setup appears and he runs to the nearest door like a regular villager would it also tells me an internal error has occurred while attempting to perform this command what version of the product are you using as installed by multicraft the jar file appears to be there as does the shopkeepers folder i tried removing these and then reinstalling it yet again but the same problem occurs server is on do you have an error log of what happened info more info at java lang classloader loadclass unknown source info at java lang classloader loadclass unknown source info at org bukkit plugin java pluginclassloader findclass pluginclassloader java info at org bukkit plugin java pluginclassloader pluginclassloader java info caused by java lang classnotfoundexception net minecraft server pathfindergoal info more info at org bukkit command plugincommand execute plugincommand java info at com nisovin shopkeepers shopkeepersplugin oncommand shopkeepersplugin java info at com nisovin shopkeepers shopkeepersplugin createnewadminshopkeeper shopkeepersplugin java info at com nisovin shopkeepers shopkeeper spawn shopkeeper java info at com nisovin shopkeepers shopobjects villagershop spawn villagershop java info at com nisovin shopkeepers shopobjects livingentityshop spawn livingentityshop java info at com nisovin shopkeepers shopobjects villagershop overwriteai villagershop java info caused by java lang noclassdeffounderror net minecraft server pathfindergoal info at net minecraft server threadserverapplication run sourcefile info at net minecraft server minecraftserver run minecraftserver java info at net minecraft server minecraftserver q minecraftserver java info at net minecraft server dedicatedserver r dedicatedserver java info at net minecraft server minecraftserver r minecraftserver java info at net minecraft server dedicatedserverconnection b sourcefile info at net minecraft server serverconnection b sourcefile info at net minecraft server playerconnection d playerconnection java info at net minecraft server networkmanager b networkmanager java info at net minecraft server handle java info at net minecraft server playerconnection a playerconnection java info at net minecraft server playerconnection chat playerconnection java info at net minecraft server playerconnection handlecommand playerconnection java info at org bukkit craftbukkit craftserver dispatchcommand craftserver java info at org bukkit command simplecommandmap dispatch simplecommandmap java info at org bukkit command plugincommand execute plugincommand java info org bukkit command commandexception unhandled exception executing command shopkeeper in plugin shopkeepers severe nullplease provide any additional information below
| 1
|
73,118
| 24,467,715,013
|
IssuesEvent
|
2022-10-07 16:30:44
|
BOINC/boinc
|
https://api.github.com/repos/BOINC/boinc
|
closed
|
[Simple View] Ubuntu 19.10 Empty Computing Preferences screen
|
C: Manager P: Major R: duplicate T: Defect E: to be determined C: Manager - Simple View Validate
|
I've installed BOINC from the Software Package Manager and it runs / computes fine. The GUI theme however is pretty broken. The preferences page is 100% blank.

|
1.0
|
[Simple View] Ubuntu 19.10 Empty Computing Preferences screen - I've installed BOINC from the Software Package Manager and it runs / computes fine. The GUI theme however is pretty broken. The preferences page is 100% blank.

|
defect
|
ubuntu empty computing preferences screen i ve installed boinc from the software package manager and it runs computes fine the gui theme however is pretty broken the preferences page is blank
| 1
|
433,410
| 30,326,631,920
|
IssuesEvent
|
2023-07-11 00:56:08
|
aws/aws-sdk-net
|
https://api.github.com/repos/aws/aws-sdk-net
|
reopened
|
No instructions on POST presigned url generation?
|
response-requested documentation module/sdk-core
|
### Describe the issue
I am attempting to add contitions such as max payload size using [POST policies](https://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-HTTPPOSTConstructPolicy.html) when generating presigned urls to allow the uploading of files to an S3 bucket. While the .NET AWS SDK client does have the option to create presigned urls for PUT requests, it does not for POST requests. I was wondering if the .NET SDK had methods/mechansim like the Python's [generate_presigned_post](https://boto3.amazonaws.com/v1/documentation/api/latest/reference/services/s3/client/generate_presigned_post.html) for generating POST presigned URLs?
### Links
https://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-HTTPPOSTConstructPolicy.html
https://boto3.amazonaws.com/v1/documentation/api/latest/reference/services/s3/client/generate_presigned_post.html
Other links
https://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-post-example.html
https://docs.aws.amazon.com/AmazonS3/latest/API/sig-v4-header-based-auth.html
|
1.0
|
No instructions on POST presigned url generation? - ### Describe the issue
I am attempting to add contitions such as max payload size using [POST policies](https://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-HTTPPOSTConstructPolicy.html) when generating presigned urls to allow the uploading of files to an S3 bucket. While the .NET AWS SDK client does have the option to create presigned urls for PUT requests, it does not for POST requests. I was wondering if the .NET SDK had methods/mechansim like the Python's [generate_presigned_post](https://boto3.amazonaws.com/v1/documentation/api/latest/reference/services/s3/client/generate_presigned_post.html) for generating POST presigned URLs?
### Links
https://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-HTTPPOSTConstructPolicy.html
https://boto3.amazonaws.com/v1/documentation/api/latest/reference/services/s3/client/generate_presigned_post.html
Other links
https://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-post-example.html
https://docs.aws.amazon.com/AmazonS3/latest/API/sig-v4-header-based-auth.html
|
non_defect
|
no instructions on post presigned url generation describe the issue i am attempting to add contitions such as max payload size using when generating presigned urls to allow the uploading of files to an bucket while the net aws sdk client does have the option to create presigned urls for put requests it does not for post requests i was wondering if the net sdk had methods mechansim like the python s for generating post presigned urls links other links
| 0
|
287,937
| 31,856,517,809
|
IssuesEvent
|
2023-09-15 07:52:25
|
Trinadh465/linux-4.1.15_CVE-2023-26607
|
https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-26607
|
opened
|
CVE-2022-4379 (High) detected in linuxlinux-4.6
|
Mend: dependency security vulnerability
|
## CVE-2022-4379 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-26607/commit/6fca0e3f2f14e1e851258fd815766531370084b0">6fca0e3f2f14e1e851258fd815766531370084b0</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A use-after-free vulnerability was found in __nfs42_ssc_open() in fs/nfs/nfs4file.c in the Linux kernel. This flaw allows an attacker to conduct a remote denial
<p>Publish Date: 2023-01-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-4379>CVE-2022-4379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-4379">https://www.linuxkernelcves.com/cves/CVE-2022-4379</a></p>
<p>Release Date: 2023-01-10</p>
<p>Fix Resolution: v6.1.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-4379 (High) detected in linuxlinux-4.6 - ## CVE-2022-4379 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-26607/commit/6fca0e3f2f14e1e851258fd815766531370084b0">6fca0e3f2f14e1e851258fd815766531370084b0</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A use-after-free vulnerability was found in __nfs42_ssc_open() in fs/nfs/nfs4file.c in the Linux kernel. This flaw allows an attacker to conduct a remote denial
<p>Publish Date: 2023-01-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-4379>CVE-2022-4379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-4379">https://www.linuxkernelcves.com/cves/CVE-2022-4379</a></p>
<p>Release Date: 2023-01-10</p>
<p>Fix Resolution: v6.1.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files vulnerability details a use after free vulnerability was found in ssc open in fs nfs c in the linux kernel this flaw allows an attacker to conduct a remote denial publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
55,538
| 14,534,649,842
|
IssuesEvent
|
2020-12-15 03:32:44
|
naev/naev
|
https://api.github.com/repos/naev/naev
|
closed
|
Game crashes if you switch from a ship carrying cargo to a ship with stowed cargo.
|
Priority-Critical Type-Defect
|
Repro: load this little fellow, go to the equipment screen, and double-click the pirate Kestrel.
[True pirate.ns.gz](https://github.com/naev/naev/files/5650772/True.pirate.ns.gz)
The save data says
```
<ship name="P-KISTAL" model="Pirate Kestrel>
...
<commodities>
<commodity quantity="35">Food</commodity>
<commodity quantity="23">Diamond</commodity>
</commodities>
...
</ship>
```
If that's a legitimate situation, the pilot_cargoMove code needs to handle it correctly.
(If it's not, that code still needs to change.)
It tries to make space for new entries in the dest->commodities array, but clobbers the originals rather than appending.
But also, can the source and destination ships have overlaps in the commodity types? Do we have to dedup? Should we simply call a helper method to add each cargo to the destination?
|
1.0
|
Game crashes if you switch from a ship carrying cargo to a ship with stowed cargo. - Repro: load this little fellow, go to the equipment screen, and double-click the pirate Kestrel.
[True pirate.ns.gz](https://github.com/naev/naev/files/5650772/True.pirate.ns.gz)
The save data says
```
<ship name="P-KISTAL" model="Pirate Kestrel>
...
<commodities>
<commodity quantity="35">Food</commodity>
<commodity quantity="23">Diamond</commodity>
</commodities>
...
</ship>
```
If that's a legitimate situation, the pilot_cargoMove code needs to handle it correctly.
(If it's not, that code still needs to change.)
It tries to make space for new entries in the dest->commodities array, but clobbers the originals rather than appending.
But also, can the source and destination ships have overlaps in the commodity types? Do we have to dedup? Should we simply call a helper method to add each cargo to the destination?
|
defect
|
game crashes if you switch from a ship carrying cargo to a ship with stowed cargo repro load this little fellow go to the equipment screen and double click the pirate kestrel the save data says food diamond if that s a legitimate situation the pilot cargomove code needs to handle it correctly if it s not that code still needs to change it tries to make space for new entries in the dest commodities array but clobbers the originals rather than appending but also can the source and destination ships have overlaps in the commodity types do we have to dedup should we simply call a helper method to add each cargo to the destination
| 1
|
49,808
| 13,187,275,191
|
IssuesEvent
|
2020-08-13 02:53:56
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
[CascadeVariables] review branch by mjurkovic (Trac #2118)
|
Incomplete Migration Migrated from Trac analysis defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2118">https://code.icecube.wisc.edu/ticket/2118</a>, reported by kjmeagher and owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-12T22:16:55",
"description": "http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/sandbox/mjurkovic/projects/CascadeVariables/trunk",
"reporter": "kjmeagher",
"cc": "",
"resolution": "insufficient resources",
"_ts": "1550009815229985",
"component": "analysis",
"summary": "[CascadeVariables] review branch by mjurkovic",
"priority": "normal",
"keywords": "",
"time": "2017-11-30T09:01:47",
"milestone": "Vernal Equinox 2019",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[CascadeVariables] review branch by mjurkovic (Trac #2118) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2118">https://code.icecube.wisc.edu/ticket/2118</a>, reported by kjmeagher and owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-12T22:16:55",
"description": "http://code.icecube.wisc.edu/projects/icecube/browser/IceCube/sandbox/mjurkovic/projects/CascadeVariables/trunk",
"reporter": "kjmeagher",
"cc": "",
"resolution": "insufficient resources",
"_ts": "1550009815229985",
"component": "analysis",
"summary": "[CascadeVariables] review branch by mjurkovic",
"priority": "normal",
"keywords": "",
"time": "2017-11-30T09:01:47",
"milestone": "Vernal Equinox 2019",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
|
defect
|
review branch by mjurkovic trac migrated from json status closed changetime description reporter kjmeagher cc resolution insufficient resources ts component analysis summary review branch by mjurkovic priority normal keywords time milestone vernal equinox owner kjmeagher type defect
| 1
|
15,260
| 10,271,319,611
|
IssuesEvent
|
2019-08-23 13:52:05
|
kyma-project/kyma
|
https://api.github.com/repos/kyma-project/kyma
|
closed
|
Move API controller integration tests from component to tests
|
area/service-mesh enhancement stale
|
There are 2 api-controller tests that require running Kyma cluster to pass:
* components/api-controller/pkg/controller/networking/v1/networkingv1_integration_test.go
* components/api-controller/pkg/controller/authentication/v2/authenticationv2_integration_test.go
They are not run on CI because Kyma is not available during unit testing. It would be better to have them in tests/api-controller-acceptance instead, so they are run every time.
|
1.0
|
Move API controller integration tests from component to tests - There are 2 api-controller tests that require running Kyma cluster to pass:
* components/api-controller/pkg/controller/networking/v1/networkingv1_integration_test.go
* components/api-controller/pkg/controller/authentication/v2/authenticationv2_integration_test.go
They are not run on CI because Kyma is not available during unit testing. It would be better to have them in tests/api-controller-acceptance instead, so they are run every time.
|
non_defect
|
move api controller integration tests from component to tests there are api controller tests that require running kyma cluster to pass components api controller pkg controller networking integration test go components api controller pkg controller authentication integration test go they are not run on ci because kyma is not available during unit testing it would be better to have them in tests api controller acceptance instead so they are run every time
| 0
|
36,917
| 2,813,567,515
|
IssuesEvent
|
2015-05-18 15:20:07
|
CruxFramework/crux
|
https://api.github.com/repos/CruxFramework/crux
|
closed
|
Split CSS files from Smart Faces
|
enhancement imported invalid Module-CruxSmartFaces Priority-Medium
|
_From [claudio....@cruxframework.org](https://code.google.com/u/102254381191677355567/) on August 29, 2014 10:38:07_
Today we have 3 main stylesheets for Smart Faces. Large, Small, and Common. But this approach still not the better one. By splitting the CSS into one file for each component we made it really easier to maintain due to the organization, not to mention that it will be easier to edit and create new themes further ahead.
We can take advantage of the GWT CSS processor (future GSS) to compile it all together automatically.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=505_
|
1.0
|
Split CSS files from Smart Faces - _From [claudio....@cruxframework.org](https://code.google.com/u/102254381191677355567/) on August 29, 2014 10:38:07_
Today we have 3 main stylesheets for Smart Faces. Large, Small, and Common. But this approach still not the better one. By splitting the CSS into one file for each component we made it really easier to maintain due to the organization, not to mention that it will be easier to edit and create new themes further ahead.
We can take advantage of the GWT CSS processor (future GSS) to compile it all together automatically.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=505_
|
non_defect
|
split css files from smart faces from on august today we have main stylesheets for smart faces large small and common but this approach still not the better one by splitting the css into one file for each component we made it really easier to maintain due to the organization not to mention that it will be easier to edit and create new themes further ahead we can take advantage of the gwt css processor future gss to compile it all together automatically original issue
| 0
|
3,418
| 2,610,062,302
|
IssuesEvent
|
2015-02-26 18:18:19
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
黄岩治疗男性不育哪家好
|
auto-migrated Priority-Medium Type-Defect
|
```
黄岩治疗男性不育哪家好【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:35
|
1.0
|
黄岩治疗男性不育哪家好 - ```
黄岩治疗男性不育哪家好【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:35
|
defect
|
黄岩治疗男性不育哪家好 黄岩治疗男性不育哪家好【台州五洲生殖医院】 询热线 微信号tzwzszyy 医院地址 台州 (枫南大转盘旁)乘车线路 、 � �� 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
156,456
| 5,969,666,845
|
IssuesEvent
|
2017-05-30 20:48:28
|
osuosl/iam
|
https://api.github.com/repos/osuosl/iam
|
closed
|
Change client and project views to summary style
|
High Priority Ready for Review
|
Based on feedback from @ramereth:
Change the current client view to be more like the recently created Summary table - list all the nodes and databases along with their measurements, and leave the resource details (cluster, type, etc) to the individual resource view. This view will be essentially the same as the Project view, except that a client view may contain data from multiple projects.
Create a new view, 'billing' which will display a condensed summary table that has only the totals for all VPCU, RAM, DB size, etc. In this view, ganeti node measurements will be multiplied by 2 if the disk template is drbd. This view will also contain a date range selector for adjusting the date range and quick links to select month to date, last calendar month, and last quarter.
|
1.0
|
Change client and project views to summary style - Based on feedback from @ramereth:
Change the current client view to be more like the recently created Summary table - list all the nodes and databases along with their measurements, and leave the resource details (cluster, type, etc) to the individual resource view. This view will be essentially the same as the Project view, except that a client view may contain data from multiple projects.
Create a new view, 'billing' which will display a condensed summary table that has only the totals for all VPCU, RAM, DB size, etc. In this view, ganeti node measurements will be multiplied by 2 if the disk template is drbd. This view will also contain a date range selector for adjusting the date range and quick links to select month to date, last calendar month, and last quarter.
|
non_defect
|
change client and project views to summary style based on feedback from ramereth change the current client view to be more like the recently created summary table list all the nodes and databases along with their measurements and leave the resource details cluster type etc to the individual resource view this view will be essentially the same as the project view except that a client view may contain data from multiple projects create a new view billing which will display a condensed summary table that has only the totals for all vpcu ram db size etc in this view ganeti node measurements will be multiplied by if the disk template is drbd this view will also contain a date range selector for adjusting the date range and quick links to select month to date last calendar month and last quarter
| 0
|
105,562
| 16,652,829,192
|
IssuesEvent
|
2021-06-05 01:31:48
|
cfscode/react-photoswipe
|
https://api.github.com/repos/cfscode/react-photoswipe
|
opened
|
CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js
|
security vulnerability
|
## CVE-2012-6708 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: react-photoswipe/node_modules/vm-browserify/example/run/index.html</p>
<p>Path to vulnerable library: react-photoswipe/node_modules/vm-browserify/example/run/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js - ## CVE-2012-6708 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: react-photoswipe/node_modules/vm-browserify/example/run/index.html</p>
<p>Path to vulnerable library: react-photoswipe/node_modules/vm-browserify/example/run/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file react photoswipe node modules vm browserify example run index html path to vulnerable library react photoswipe node modules vm browserify example run index html dependency hierarchy x jquery min js vulnerable library found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
62,957
| 17,270,161,819
|
IssuesEvent
|
2021-07-22 18:40:12
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Crash of version 1.7.33 on hirsute on amd64
|
T-Defect
|
<!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->
Same symptoms as https://github.com/vector-im/element-web/issues/18173 but apparently different root cause
(stack trace at: https://gist.github.com/monochromec/14ce9a604a368d0ee017730923ab9265)
<!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
#### Description
Random crashes on Ubuntu 21.04
#### Steps to reproduce
- Start Element
- Observe crash while app becomes unresponsive
<!-- Please send us logs for your bug report. They're very important for bugs
which are hard to reproduce. To do this, create this issue then go to your
account settings and click 'Submit Debug Logs' from the Help & About tab -->
Logs being sent: yes
(Cf. above)
<!-- Include screenshots if possible: you can drag and drop images below. -->
#### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: web (in-browser) or desktop?
Desktop 1.7.33
- Ubuntu 21.04
1.7.33 apparently really missed the boat on several platforms. More than happy to provide additional information - please let me know what else is required...
|
1.0
|
Crash of version 1.7.33 on hirsute on amd64 - <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->
Same symptoms as https://github.com/vector-im/element-web/issues/18173 but apparently different root cause
(stack trace at: https://gist.github.com/monochromec/14ce9a604a368d0ee017730923ab9265)
<!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
#### Description
Random crashes on Ubuntu 21.04
#### Steps to reproduce
- Start Element
- Observe crash while app becomes unresponsive
<!-- Please send us logs for your bug report. They're very important for bugs
which are hard to reproduce. To do this, create this issue then go to your
account settings and click 'Submit Debug Logs' from the Help & About tab -->
Logs being sent: yes
(Cf. above)
<!-- Include screenshots if possible: you can drag and drop images below. -->
#### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: web (in-browser) or desktop?
Desktop 1.7.33
- Ubuntu 21.04
1.7.33 apparently really missed the boat on several platforms. More than happy to provide additional information - please let me know what else is required...
|
defect
|
crash of version on hirsute on same symptoms as but apparently different root cause stack trace at this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description random crashes on ubuntu steps to reproduce start element observe crash while app becomes unresponsive please send us logs for your bug report they re very important for bugs which are hard to reproduce to do this create this issue then go to your account settings and click submit debug logs from the help about tab logs being sent yes cf above version information platform web in browser or desktop desktop ubuntu apparently really missed the boat on several platforms more than happy to provide additional information please let me know what else is required
| 1
|
194,059
| 14,668,656,145
|
IssuesEvent
|
2020-12-29 21:53:49
|
calba5141114/studentpass
|
https://api.github.com/repos/calba5141114/studentpass
|
opened
|
Implement Tests for studentpass
|
help wanted testing
|
Testing Individual Classes, Functions and REST routes is very important for insuring that
studentpass is a highly reliable and stable API.
|
1.0
|
Implement Tests for studentpass - Testing Individual Classes, Functions and REST routes is very important for insuring that
studentpass is a highly reliable and stable API.
|
non_defect
|
implement tests for studentpass testing individual classes functions and rest routes is very important for insuring that studentpass is a highly reliable and stable api
| 0
|
26,270
| 4,647,121,124
|
IssuesEvent
|
2016-10-01 09:16:09
|
KronoZed/urtdsc-old
|
https://api.github.com/repos/KronoZed/urtdsc-old
|
closed
|
Нужно почистить код
|
auto-migrated Component-Logic Priority-Medium Type-Defect
|
```
Нужно почистить код.
Бросаются в глаза demorealdate и screenrealdate (из func.py)
с одинаковым поведением, и непонятно где
используемые.
```
Original issue reported on code.google.com by `endenis@gmail.com` on 11 Sep 2011 at 9:42
|
1.0
|
Нужно почистить код - ```
Нужно почистить код.
Бросаются в глаза demorealdate и screenrealdate (из func.py)
с одинаковым поведением, и непонятно где
используемые.
```
Original issue reported on code.google.com by `endenis@gmail.com` on 11 Sep 2011 at 9:42
|
defect
|
нужно почистить код нужно почистить код бросаются в глаза demorealdate и screenrealdate из func py с одинаковым поведением и непонятно где используемые original issue reported on code google com by endenis gmail com on sep at
| 1
|
41,942
| 10,722,199,627
|
IssuesEvent
|
2019-10-27 10:06:46
|
vesoft-inc/nebula
|
https://api.github.com/repos/vesoft-inc/nebula
|
closed
|
Chinese Guideline shows error when try to insert
|
defect-p3
|
**Describe the bug(__must be provided__)**
Follow the guideline, when try to execute this
> INSERT VERTEX student(name, age, gender) VALUES 200:("Monica", 16, "female");
shows error:
> [ERROR (-8)]: No schema found for `student'
**Your Environments(__must be provided__)**
* OS: `18.7.0 Darwin Kernel Version 18.7.0: Thu Jun 20 18:42:21 PDT 2019; root:xnu-4903.270.47~4/RELEASE_X86_64 x86_64`
* Compliler: Configured with: --prefix=/Applications/Xcode.app/Contents/Developer/usr --with-gxx-include-dir=/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.14.sdk/usr/include/c++/4.2.1
Apple LLVM version 10.0.1 (clang-1001.0.46.4)
Target: x86_64-apple-darwin18.7.0
Thread model: posix
InstalledDir: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin
* CPU: Intel(R) Core(TM) i7-4980HQ CPU @ 2.80GHz
**How To Reproduce(__must be provided__)**
Steps to reproduce the behavior:
1. Just follow the chinese guideline step by step
2. try to execute command 'INSERT VERTEX student(name, age, gender) VALUES 200:("Monica", 16, "female");'
3. shows error above
**Expected behavior**
Should execute success or shows error meta is not ok
**Additional context**
first launch
When you execute some other command,it wiil be ok, not sure if it caused by meta problme
|
1.0
|
Chinese Guideline shows error when try to insert - **Describe the bug(__must be provided__)**
Follow the guideline, when try to execute this
> INSERT VERTEX student(name, age, gender) VALUES 200:("Monica", 16, "female");
shows error:
> [ERROR (-8)]: No schema found for `student'
**Your Environments(__must be provided__)**
* OS: `18.7.0 Darwin Kernel Version 18.7.0: Thu Jun 20 18:42:21 PDT 2019; root:xnu-4903.270.47~4/RELEASE_X86_64 x86_64`
* Compliler: Configured with: --prefix=/Applications/Xcode.app/Contents/Developer/usr --with-gxx-include-dir=/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.14.sdk/usr/include/c++/4.2.1
Apple LLVM version 10.0.1 (clang-1001.0.46.4)
Target: x86_64-apple-darwin18.7.0
Thread model: posix
InstalledDir: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin
* CPU: Intel(R) Core(TM) i7-4980HQ CPU @ 2.80GHz
**How To Reproduce(__must be provided__)**
Steps to reproduce the behavior:
1. Just follow the chinese guideline step by step
2. try to execute command 'INSERT VERTEX student(name, age, gender) VALUES 200:("Monica", 16, "female");'
3. shows error above
**Expected behavior**
Should execute success or shows error meta is not ok
**Additional context**
first launch
When you execute some other command,it wiil be ok, not sure if it caused by meta problme
|
defect
|
chinese guideline shows error when try to insert describe the bug must be provided follow the guideline when try to execute this insert vertex student name age gender values monica female shows error no schema found for student your environments must be provided os darwin kernel version thu jun pdt root xnu release compliler configured with prefix applications xcode app contents developer usr with gxx include dir applications xcode app contents developer platforms macosx platform developer sdks sdk usr include c apple llvm version clang target apple thread model posix installeddir applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin cpu intel r core tm cpu how to reproduce must be provided steps to reproduce the behavior just follow the chinese guideline step by step try to execute command insert vertex student name age gender values monica female shows error above expected behavior should execute success or shows error meta is not ok additional context first launch when you execute some other command,it wiil be ok, not sure if it caused by meta problme
| 1
|
365,742
| 10,791,174,245
|
IssuesEvent
|
2019-11-05 16:10:40
|
AY1920S1-CS2113T-W12-3/main
|
https://api.github.com/repos/AY1920S1-CS2113T-W12-3/main
|
closed
|
As a hall resident, I can cancel booking of the facility
|
priority.High type.Story
|
So that I can free up the room for others
|
1.0
|
As a hall resident, I can cancel booking of the facility - So that I can free up the room for others
|
non_defect
|
as a hall resident i can cancel booking of the facility so that i can free up the room for others
| 0
|
4,825
| 5,314,258,134
|
IssuesEvent
|
2017-02-13 14:37:55
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
2nd factor authentication turns off 1st factor authentication
|
1.severity: blocker 1.severity: security 6.topic: nixos
|
When I did "security.pam.services.sshd.oathAuth = true", the resulting pam file was:
auth sufficient pam_unix.so likeauth try_first_pass
auth sufficient ${pkgs.oathToolkit}/lib/security/pam_oath.so window=5 usersfile=/etc/users.oath digits=6
which allows me to type a wrong password and make up for it by a correct OTP. I haven't reviewed the resulting PAM file carefully after making that change, and I was taken by surprise by that behavior. I would have expected a more conservative "conjunctive" semantic here, as in "OTP" is required in addition to everything else.
Justification:
While the other authentication mechanisms surrounding the "auth" directives in
https://github.com/NixOS/nixpkgs/blob/master/nixos/modules/security/pam.nix#L262
could be argued to have password like qualities and rightfully replace pam_unix.so (pam_ldap, pam_otpw), I would argue that 2nd factor auth methods like xOTP don't have those qualities (2^6^8 for a 8 char password vs 10^6~=2^16 for xOTP). Otherwise, they wouldn't be called 2nd factor :)
Implementation:
Move 2nd factor mechanisms to the top of the auth section and mark them required. Unfortunately that makes all authenticators prompt for 2nd factor first, which is a bit odd. I haven't yet thought about a more complex restructuring of pam.nix.
While at it, we could split up the "pam.text" section into four overridable ones for account/auth/session/password, so that people, that don't agree with my conservative stance and want 2nd factor only, can change their configs easily.
- System: 16.03.616.4ac4b35 (Emu)
- Nix version: nix-env (Nix) 1.11.2
- Nixpkgs version: 16.03.616.4ac4b35
|
True
|
2nd factor authentication turns off 1st factor authentication - When I did "security.pam.services.sshd.oathAuth = true", the resulting pam file was:
auth sufficient pam_unix.so likeauth try_first_pass
auth sufficient ${pkgs.oathToolkit}/lib/security/pam_oath.so window=5 usersfile=/etc/users.oath digits=6
which allows me to type a wrong password and make up for it by a correct OTP. I haven't reviewed the resulting PAM file carefully after making that change, and I was taken by surprise by that behavior. I would have expected a more conservative "conjunctive" semantic here, as in "OTP" is required in addition to everything else.
Justification:
While the other authentication mechanisms surrounding the "auth" directives in
https://github.com/NixOS/nixpkgs/blob/master/nixos/modules/security/pam.nix#L262
could be argued to have password like qualities and rightfully replace pam_unix.so (pam_ldap, pam_otpw), I would argue that 2nd factor auth methods like xOTP don't have those qualities (2^6^8 for a 8 char password vs 10^6~=2^16 for xOTP). Otherwise, they wouldn't be called 2nd factor :)
Implementation:
Move 2nd factor mechanisms to the top of the auth section and mark them required. Unfortunately that makes all authenticators prompt for 2nd factor first, which is a bit odd. I haven't yet thought about a more complex restructuring of pam.nix.
While at it, we could split up the "pam.text" section into four overridable ones for account/auth/session/password, so that people, that don't agree with my conservative stance and want 2nd factor only, can change their configs easily.
- System: 16.03.616.4ac4b35 (Emu)
- Nix version: nix-env (Nix) 1.11.2
- Nixpkgs version: 16.03.616.4ac4b35
|
non_defect
|
factor authentication turns off factor authentication when i did security pam services sshd oathauth true the resulting pam file was auth sufficient pam unix so likeauth try first pass auth sufficient pkgs oathtoolkit lib security pam oath so window usersfile etc users oath digits which allows me to type a wrong password and make up for it by a correct otp i haven t reviewed the resulting pam file carefully after making that change and i was taken by surprise by that behavior i would have expected a more conservative conjunctive semantic here as in otp is required in addition to everything else justification while the other authentication mechanisms surrounding the auth directives in could be argued to have password like qualities and rightfully replace pam unix so pam ldap pam otpw i would argue that factor auth methods like xotp don t have those qualities for a char password vs for xotp otherwise they wouldn t be called factor implementation move factor mechanisms to the top of the auth section and mark them required unfortunately that makes all authenticators prompt for factor first which is a bit odd i haven t yet thought about a more complex restructuring of pam nix while at it we could split up the pam text section into four overridable ones for account auth session password so that people that don t agree with my conservative stance and want factor only can change their configs easily system emu nix version nix env nix nixpkgs version
| 0
|
239,717
| 26,232,057,227
|
IssuesEvent
|
2023-01-05 01:43:00
|
kapseliboi/DoIt
|
https://api.github.com/repos/kapseliboi/DoIt
|
opened
|
CVE-2021-44906 (High) detected in minimist-1.2.5.tgz
|
security vulnerability
|
## CVE-2021-44906 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-1.2.5.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- eslint-plugin-import-2.25.4.tgz (Root Library)
- tsconfig-paths-3.12.0.tgz
- :x: **minimist-1.2.5.tgz** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95).
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-44906>CVE-2021-44906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution (minimist): 1.2.6</p>
<p>Direct dependency fix Resolution (eslint-plugin-import): 2.26.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-44906 (High) detected in minimist-1.2.5.tgz - ## CVE-2021-44906 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-1.2.5.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- eslint-plugin-import-2.25.4.tgz (Root Library)
- tsconfig-paths-3.12.0.tgz
- :x: **minimist-1.2.5.tgz** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95).
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-44906>CVE-2021-44906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution (minimist): 1.2.6</p>
<p>Direct dependency fix Resolution (eslint-plugin-import): 2.26.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in minimist tgz cve high severity vulnerability vulnerable library minimist tgz parse argument options library home page a href path to dependency file package json path to vulnerable library node modules minimist package json dependency hierarchy eslint plugin import tgz root library tsconfig paths tgz x minimist tgz vulnerable library found in base branch main vulnerability details minimist is vulnerable to prototype pollution via file index js function setkey lines publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimist direct dependency fix resolution eslint plugin import step up your open source security game with mend
| 0
|
135,346
| 5,246,947,368
|
IssuesEvent
|
2017-02-01 11:17:21
|
pmem/issues
|
https://api.github.com/repos/pmem/issues
|
closed
|
unit tests: obj_list_recovery/TEST0, TEST1, TEST2 (all/pmem/debug/memcheck) fails
|
Exposure: Low OS: Linux Priority: 4 low Type: Bug
|
Found on revision: 27ac7ed4ef539adf64cd712aefe578ec61ec6c70
> obj_list_recovery/TEST0: SETUP (all/pmem/debug/memcheck)
> obj_list_recovery/TEST0: START: obj_list
> obj_list_recovery/TEST0: START: obj_list
> obj_list_recovery/TEST0 failed with Valgrind. See memcheck0.log. First 20 lines below.
> obj_list_recovery/TEST0 memcheck0.log ==5595==
> obj_list_recovery/TEST0 memcheck0.log ==5595== HEAP SUMMARY:
> obj_list_recovery/TEST0 memcheck0.log ==5595== in use at exit: 112 bytes in 1 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595== total heap usage: 47 allocs, 46 frees, 91,956 bytes allocated
> obj_list_recovery/TEST0 memcheck0.log ==5595==
> obj_list_recovery/TEST0 memcheck0.log ==5595== 112 bytes in 1 blocks are definitely lost in loss record 1 of 1
> obj_list_recovery/TEST0 memcheck0.log ==5595== at 0x4C2DB90: malloc (vg_replace_malloc.c:296)
> obj_list_recovery/TEST0 memcheck0.log ==5595== by 0x41E0CE: redo_log_config_new (redo.c:69)
> obj_list_recovery/TEST0 memcheck0.log ==5595== by 0x404E3C: __wrap_pmemobj_open (obj_list_mocks.c:187)
> obj_list_recovery/TEST0 memcheck0.log ==5595== by 0x4044C0: main (obj_list.c:405)
> obj_list_recovery/TEST0 memcheck0.log ==5595==
> obj_list_recovery/TEST0 memcheck0.log ==5595== LEAK SUMMARY:
> obj_list_recovery/TEST0 memcheck0.log ==5595== definitely lost: 112 bytes in 1 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595== indirectly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595== possibly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595== still reachable: 0 bytes in 0 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595== suppressed: 0 bytes in 0 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595==
> obj_list_recovery/TEST0 memcheck0.log ==5595== For counts of detected and suppressed errors, rerun with: -v
> obj_list_recovery/TEST0 memcheck0.log ==5595== ERROR SUMMARY: 1 errors from 1 contexts (suppressed: 0 from 0)
> RUNTESTS: stopping: obj_list_recovery/TEST0 failed, TEST=all FS=pmem BUILD=debug
> ../Makefile.inc:320: recipe for target 'TEST0' failed
> make[3]: *** [TEST0] Error 1
> obj_list_recovery/TEST1: SETUP (all/pmem/debug/memcheck)
> obj_list_recovery/TEST1: START: obj_list
> obj_list_recovery/TEST1: START: obj_list
> obj_list_recovery/TEST1 failed with Valgrind. See memcheck1.log. First 20 lines below.
> obj_list_recovery/TEST1 memcheck1.log ==5675==
> obj_list_recovery/TEST1 memcheck1.log ==5675== HEAP SUMMARY:
> obj_list_recovery/TEST1 memcheck1.log ==5675== in use at exit: 112 bytes in 1 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675== total heap usage: 47 allocs, 46 frees, 91,956 bytes allocated
> obj_list_recovery/TEST1 memcheck1.log ==5675==
> obj_list_recovery/TEST1 memcheck1.log ==5675== 112 bytes in 1 blocks are definitely lost in loss record 1 of 1
> obj_list_recovery/TEST1 memcheck1.log ==5675== at 0x4C2DB90: malloc (vg_replace_malloc.c:296)
> obj_list_recovery/TEST1 memcheck1.log ==5675== by 0x41E0CE: redo_log_config_new (redo.c:69)
> obj_list_recovery/TEST1 memcheck1.log ==5675== by 0x404E3C: __wrap_pmemobj_open (obj_list_mocks.c:187)
> obj_list_recovery/TEST1 memcheck1.log ==5675== by 0x4044C0: main (obj_list.c:405)
> obj_list_recovery/TEST1 memcheck1.log ==5675==
> obj_list_recovery/TEST1 memcheck1.log ==5675== LEAK SUMMARY:
> obj_list_recovery/TEST1 memcheck1.log ==5675== definitely lost: 112 bytes in 1 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675== indirectly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675== possibly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675== still reachable: 0 bytes in 0 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675== suppressed: 0 bytes in 0 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675==
> obj_list_recovery/TEST1 memcheck1.log ==5675== For counts of detected and suppressed errors, rerun with: -v
> obj_list_recovery/TEST1 memcheck1.log ==5675== ERROR SUMMARY: 1 errors from 1 contexts (suppressed: 0 from 0)
> RUNTESTS: stopping: obj_list_recovery/TEST1 failed, TEST=all FS=pmem BUILD=debug
> ../Makefile.inc:320: recipe for target 'TEST1' failed
> make[3]: *** [TEST1] Error 1
> obj_list_recovery/TEST2: SETUP (all/pmem/debug/memcheck)
> obj_list_recovery/TEST2: START: obj_list
> obj_list_recovery/TEST2: START: obj_list
> obj_list_recovery/TEST2 failed with Valgrind. See memcheck2.log. First 20 lines below.
> obj_list_recovery/TEST2 memcheck2.log ==5755==
> obj_list_recovery/TEST2 memcheck2.log ==5755== HEAP SUMMARY:
> obj_list_recovery/TEST2 memcheck2.log ==5755== in use at exit: 112 bytes in 1 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755== total heap usage: 47 allocs, 46 frees, 91,956 bytes allocated
> obj_list_recovery/TEST2 memcheck2.log ==5755==
> obj_list_recovery/TEST2 memcheck2.log ==5755== 112 bytes in 1 blocks are definitely lost in loss record 1 of 1
> obj_list_recovery/TEST2 memcheck2.log ==5755== at 0x4C2DB90: malloc (vg_replace_malloc.c:296)
> obj_list_recovery/TEST2 memcheck2.log ==5755== by 0x41E0CE: redo_log_config_new (redo.c:69)
> obj_list_recovery/TEST2 memcheck2.log ==5755== by 0x404E3C: __wrap_pmemobj_open (obj_list_mocks.c:187)
> obj_list_recovery/TEST2 memcheck2.log ==5755== by 0x4044C0: main (obj_list.c:405)
> obj_list_recovery/TEST2 memcheck2.log ==5755==
> obj_list_recovery/TEST2 memcheck2.log ==5755== LEAK SUMMARY:
> obj_list_recovery/TEST2 memcheck2.log ==5755== definitely lost: 112 bytes in 1 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755== indirectly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755== possibly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755== still reachable: 0 bytes in 0 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755== suppressed: 0 bytes in 0 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755==
> obj_list_recovery/TEST2 memcheck2.log ==5755== For counts of detected and suppressed errors, rerun with: -v
> obj_list_recovery/TEST2 memcheck2.log ==5755== ERROR SUMMARY: 1 errors from 1 contexts (suppressed: 0 from 0)
> RUNTESTS: stopping: obj_list_recovery/TEST2 failed, TEST=all FS=pmem BUILD=debug
|
1.0
|
unit tests: obj_list_recovery/TEST0, TEST1, TEST2 (all/pmem/debug/memcheck) fails - Found on revision: 27ac7ed4ef539adf64cd712aefe578ec61ec6c70
> obj_list_recovery/TEST0: SETUP (all/pmem/debug/memcheck)
> obj_list_recovery/TEST0: START: obj_list
> obj_list_recovery/TEST0: START: obj_list
> obj_list_recovery/TEST0 failed with Valgrind. See memcheck0.log. First 20 lines below.
> obj_list_recovery/TEST0 memcheck0.log ==5595==
> obj_list_recovery/TEST0 memcheck0.log ==5595== HEAP SUMMARY:
> obj_list_recovery/TEST0 memcheck0.log ==5595== in use at exit: 112 bytes in 1 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595== total heap usage: 47 allocs, 46 frees, 91,956 bytes allocated
> obj_list_recovery/TEST0 memcheck0.log ==5595==
> obj_list_recovery/TEST0 memcheck0.log ==5595== 112 bytes in 1 blocks are definitely lost in loss record 1 of 1
> obj_list_recovery/TEST0 memcheck0.log ==5595== at 0x4C2DB90: malloc (vg_replace_malloc.c:296)
> obj_list_recovery/TEST0 memcheck0.log ==5595== by 0x41E0CE: redo_log_config_new (redo.c:69)
> obj_list_recovery/TEST0 memcheck0.log ==5595== by 0x404E3C: __wrap_pmemobj_open (obj_list_mocks.c:187)
> obj_list_recovery/TEST0 memcheck0.log ==5595== by 0x4044C0: main (obj_list.c:405)
> obj_list_recovery/TEST0 memcheck0.log ==5595==
> obj_list_recovery/TEST0 memcheck0.log ==5595== LEAK SUMMARY:
> obj_list_recovery/TEST0 memcheck0.log ==5595== definitely lost: 112 bytes in 1 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595== indirectly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595== possibly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595== still reachable: 0 bytes in 0 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595== suppressed: 0 bytes in 0 blocks
> obj_list_recovery/TEST0 memcheck0.log ==5595==
> obj_list_recovery/TEST0 memcheck0.log ==5595== For counts of detected and suppressed errors, rerun with: -v
> obj_list_recovery/TEST0 memcheck0.log ==5595== ERROR SUMMARY: 1 errors from 1 contexts (suppressed: 0 from 0)
> RUNTESTS: stopping: obj_list_recovery/TEST0 failed, TEST=all FS=pmem BUILD=debug
> ../Makefile.inc:320: recipe for target 'TEST0' failed
> make[3]: *** [TEST0] Error 1
> obj_list_recovery/TEST1: SETUP (all/pmem/debug/memcheck)
> obj_list_recovery/TEST1: START: obj_list
> obj_list_recovery/TEST1: START: obj_list
> obj_list_recovery/TEST1 failed with Valgrind. See memcheck1.log. First 20 lines below.
> obj_list_recovery/TEST1 memcheck1.log ==5675==
> obj_list_recovery/TEST1 memcheck1.log ==5675== HEAP SUMMARY:
> obj_list_recovery/TEST1 memcheck1.log ==5675== in use at exit: 112 bytes in 1 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675== total heap usage: 47 allocs, 46 frees, 91,956 bytes allocated
> obj_list_recovery/TEST1 memcheck1.log ==5675==
> obj_list_recovery/TEST1 memcheck1.log ==5675== 112 bytes in 1 blocks are definitely lost in loss record 1 of 1
> obj_list_recovery/TEST1 memcheck1.log ==5675== at 0x4C2DB90: malloc (vg_replace_malloc.c:296)
> obj_list_recovery/TEST1 memcheck1.log ==5675== by 0x41E0CE: redo_log_config_new (redo.c:69)
> obj_list_recovery/TEST1 memcheck1.log ==5675== by 0x404E3C: __wrap_pmemobj_open (obj_list_mocks.c:187)
> obj_list_recovery/TEST1 memcheck1.log ==5675== by 0x4044C0: main (obj_list.c:405)
> obj_list_recovery/TEST1 memcheck1.log ==5675==
> obj_list_recovery/TEST1 memcheck1.log ==5675== LEAK SUMMARY:
> obj_list_recovery/TEST1 memcheck1.log ==5675== definitely lost: 112 bytes in 1 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675== indirectly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675== possibly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675== still reachable: 0 bytes in 0 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675== suppressed: 0 bytes in 0 blocks
> obj_list_recovery/TEST1 memcheck1.log ==5675==
> obj_list_recovery/TEST1 memcheck1.log ==5675== For counts of detected and suppressed errors, rerun with: -v
> obj_list_recovery/TEST1 memcheck1.log ==5675== ERROR SUMMARY: 1 errors from 1 contexts (suppressed: 0 from 0)
> RUNTESTS: stopping: obj_list_recovery/TEST1 failed, TEST=all FS=pmem BUILD=debug
> ../Makefile.inc:320: recipe for target 'TEST1' failed
> make[3]: *** [TEST1] Error 1
> obj_list_recovery/TEST2: SETUP (all/pmem/debug/memcheck)
> obj_list_recovery/TEST2: START: obj_list
> obj_list_recovery/TEST2: START: obj_list
> obj_list_recovery/TEST2 failed with Valgrind. See memcheck2.log. First 20 lines below.
> obj_list_recovery/TEST2 memcheck2.log ==5755==
> obj_list_recovery/TEST2 memcheck2.log ==5755== HEAP SUMMARY:
> obj_list_recovery/TEST2 memcheck2.log ==5755== in use at exit: 112 bytes in 1 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755== total heap usage: 47 allocs, 46 frees, 91,956 bytes allocated
> obj_list_recovery/TEST2 memcheck2.log ==5755==
> obj_list_recovery/TEST2 memcheck2.log ==5755== 112 bytes in 1 blocks are definitely lost in loss record 1 of 1
> obj_list_recovery/TEST2 memcheck2.log ==5755== at 0x4C2DB90: malloc (vg_replace_malloc.c:296)
> obj_list_recovery/TEST2 memcheck2.log ==5755== by 0x41E0CE: redo_log_config_new (redo.c:69)
> obj_list_recovery/TEST2 memcheck2.log ==5755== by 0x404E3C: __wrap_pmemobj_open (obj_list_mocks.c:187)
> obj_list_recovery/TEST2 memcheck2.log ==5755== by 0x4044C0: main (obj_list.c:405)
> obj_list_recovery/TEST2 memcheck2.log ==5755==
> obj_list_recovery/TEST2 memcheck2.log ==5755== LEAK SUMMARY:
> obj_list_recovery/TEST2 memcheck2.log ==5755== definitely lost: 112 bytes in 1 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755== indirectly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755== possibly lost: 0 bytes in 0 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755== still reachable: 0 bytes in 0 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755== suppressed: 0 bytes in 0 blocks
> obj_list_recovery/TEST2 memcheck2.log ==5755==
> obj_list_recovery/TEST2 memcheck2.log ==5755== For counts of detected and suppressed errors, rerun with: -v
> obj_list_recovery/TEST2 memcheck2.log ==5755== ERROR SUMMARY: 1 errors from 1 contexts (suppressed: 0 from 0)
> RUNTESTS: stopping: obj_list_recovery/TEST2 failed, TEST=all FS=pmem BUILD=debug
|
non_defect
|
unit tests obj list recovery all pmem debug memcheck fails found on revision obj list recovery setup all pmem debug memcheck obj list recovery start obj list obj list recovery start obj list obj list recovery failed with valgrind see log first lines below obj list recovery log obj list recovery log heap summary obj list recovery log in use at exit bytes in blocks obj list recovery log total heap usage allocs frees bytes allocated obj list recovery log obj list recovery log bytes in blocks are definitely lost in loss record of obj list recovery log at malloc vg replace malloc c obj list recovery log by redo log config new redo c obj list recovery log by wrap pmemobj open obj list mocks c obj list recovery log by main obj list c obj list recovery log obj list recovery log leak summary obj list recovery log definitely lost bytes in blocks obj list recovery log indirectly lost bytes in blocks obj list recovery log possibly lost bytes in blocks obj list recovery log still reachable bytes in blocks obj list recovery log suppressed bytes in blocks obj list recovery log obj list recovery log for counts of detected and suppressed errors rerun with v obj list recovery log error summary errors from contexts suppressed from runtests stopping obj list recovery failed test all fs pmem build debug makefile inc recipe for target failed make error obj list recovery setup all pmem debug memcheck obj list recovery start obj list obj list recovery start obj list obj list recovery failed with valgrind see log first lines below obj list recovery log obj list recovery log heap summary obj list recovery log in use at exit bytes in blocks obj list recovery log total heap usage allocs frees bytes allocated obj list recovery log obj list recovery log bytes in blocks are definitely lost in loss record of obj list recovery log at malloc vg replace malloc c obj list recovery log by redo log config new redo c obj list recovery log by wrap pmemobj open obj list mocks c obj list recovery log by main obj list c obj list recovery log obj list recovery log leak summary obj list recovery log definitely lost bytes in blocks obj list recovery log indirectly lost bytes in blocks obj list recovery log possibly lost bytes in blocks obj list recovery log still reachable bytes in blocks obj list recovery log suppressed bytes in blocks obj list recovery log obj list recovery log for counts of detected and suppressed errors rerun with v obj list recovery log error summary errors from contexts suppressed from runtests stopping obj list recovery failed test all fs pmem build debug makefile inc recipe for target failed make error obj list recovery setup all pmem debug memcheck obj list recovery start obj list obj list recovery start obj list obj list recovery failed with valgrind see log first lines below obj list recovery log obj list recovery log heap summary obj list recovery log in use at exit bytes in blocks obj list recovery log total heap usage allocs frees bytes allocated obj list recovery log obj list recovery log bytes in blocks are definitely lost in loss record of obj list recovery log at malloc vg replace malloc c obj list recovery log by redo log config new redo c obj list recovery log by wrap pmemobj open obj list mocks c obj list recovery log by main obj list c obj list recovery log obj list recovery log leak summary obj list recovery log definitely lost bytes in blocks obj list recovery log indirectly lost bytes in blocks obj list recovery log possibly lost bytes in blocks obj list recovery log still reachable bytes in blocks obj list recovery log suppressed bytes in blocks obj list recovery log obj list recovery log for counts of detected and suppressed errors rerun with v obj list recovery log error summary errors from contexts suppressed from runtests stopping obj list recovery failed test all fs pmem build debug
| 0
|
24,203
| 3,924,306,331
|
IssuesEvent
|
2016-04-22 14:47:44
|
opencaching/opencaching-pl
|
https://api.github.com/repos/opencaching/opencaching-pl
|
closed
|
PHP notices (main website)
|
Priority_Medium Server_Administration Type_Defect x_Maintainability
|
```
Undefined variable: cryptedhints in viewcache.php on line 1454
Undefined index: deleted in viewcache.php on line 1659
Use of undefined constant name - assumed 'name' in lib/common.inc.php(616) : eval()'d code on line 302
Use of undefined constant name - assumed 'code' in lib/common.inc.php(616) : eval()'d code on line 302
Undefined variable: showspoiler in viewlogs.php on line 448
Undefined variable: delByCOG in viewlogs.php on line 279
Undefined variable: enable_google_translation in viewcache.php on line 1995
Undefined variable: enable_google_translation in viewcache.php on line 2094
Undefined variable: showspoiler in gallery_cache.php on line 121
Undefined variable: tmplog_username in gallery_cache.php on line 122
Undefined variable: record in gallery_cache.php on line 122
Undefined variable: clogs in gallery_cache.php on line 139
Undefined variable: logs in gallery_cache.php on line 188
Undefined variable: r_log in newcaches.php on line 86
Use of undefined constant Stats_t5_01 - assumed 'Stats_t5_01' in lib/t5.php on line 12
Use of undefined constant Stats_t5_02 - assumed 'Stats_t5_02' in lib/t5.php on line 12
Use of undefined constant Stats_t5_03 - assumed 'Stats_t5_03' in lib/t5.php on line 12
Use of undefined constant Stats_t5_04 - assumed 'Stats_t5_04' in lib/t5.php on line 12
Use of undefined constant Stats_t5_05 - assumed 'Stats_t5_05' in lib/t5.php on line 12
Use of undefined constant Stats_t5_06 - assumed 'Stats_t5_06' in lib/t5.php on line 12
Use of undefined constant Stats_t5_07 - assumed 'Stats_t5_07' in lib/t5.php on line 16
Use of undefined constant Stats_t5_08 - assumed 'Stats_t5_08' in lib/t5.php on line 16
Use of undefined constant Stats_t5_09 - assumed 'Stats_t5_09' in lib/t5.php on line 16
```
-------------------------------
These show up in recent logs. As there is an ongoing effort to clean up PHP messages that end up in error.log, I am listing these here.
Thank you.
|
1.0
|
PHP notices (main website) - ```
Undefined variable: cryptedhints in viewcache.php on line 1454
Undefined index: deleted in viewcache.php on line 1659
Use of undefined constant name - assumed 'name' in lib/common.inc.php(616) : eval()'d code on line 302
Use of undefined constant name - assumed 'code' in lib/common.inc.php(616) : eval()'d code on line 302
Undefined variable: showspoiler in viewlogs.php on line 448
Undefined variable: delByCOG in viewlogs.php on line 279
Undefined variable: enable_google_translation in viewcache.php on line 1995
Undefined variable: enable_google_translation in viewcache.php on line 2094
Undefined variable: showspoiler in gallery_cache.php on line 121
Undefined variable: tmplog_username in gallery_cache.php on line 122
Undefined variable: record in gallery_cache.php on line 122
Undefined variable: clogs in gallery_cache.php on line 139
Undefined variable: logs in gallery_cache.php on line 188
Undefined variable: r_log in newcaches.php on line 86
Use of undefined constant Stats_t5_01 - assumed 'Stats_t5_01' in lib/t5.php on line 12
Use of undefined constant Stats_t5_02 - assumed 'Stats_t5_02' in lib/t5.php on line 12
Use of undefined constant Stats_t5_03 - assumed 'Stats_t5_03' in lib/t5.php on line 12
Use of undefined constant Stats_t5_04 - assumed 'Stats_t5_04' in lib/t5.php on line 12
Use of undefined constant Stats_t5_05 - assumed 'Stats_t5_05' in lib/t5.php on line 12
Use of undefined constant Stats_t5_06 - assumed 'Stats_t5_06' in lib/t5.php on line 12
Use of undefined constant Stats_t5_07 - assumed 'Stats_t5_07' in lib/t5.php on line 16
Use of undefined constant Stats_t5_08 - assumed 'Stats_t5_08' in lib/t5.php on line 16
Use of undefined constant Stats_t5_09 - assumed 'Stats_t5_09' in lib/t5.php on line 16
```
-------------------------------
These show up in recent logs. As there is an ongoing effort to clean up PHP messages that end up in error.log, I am listing these here.
Thank you.
|
defect
|
php notices main website undefined variable cryptedhints in viewcache php on line undefined index deleted in viewcache php on line use of undefined constant name assumed name in lib common inc php eval d code on line use of undefined constant name assumed code in lib common inc php eval d code on line undefined variable showspoiler in viewlogs php on line undefined variable delbycog in viewlogs php on line undefined variable enable google translation in viewcache php on line undefined variable enable google translation in viewcache php on line undefined variable showspoiler in gallery cache php on line undefined variable tmplog username in gallery cache php on line undefined variable record in gallery cache php on line undefined variable clogs in gallery cache php on line undefined variable logs in gallery cache php on line undefined variable r log in newcaches php on line use of undefined constant stats assumed stats in lib php on line use of undefined constant stats assumed stats in lib php on line use of undefined constant stats assumed stats in lib php on line use of undefined constant stats assumed stats in lib php on line use of undefined constant stats assumed stats in lib php on line use of undefined constant stats assumed stats in lib php on line use of undefined constant stats assumed stats in lib php on line use of undefined constant stats assumed stats in lib php on line use of undefined constant stats assumed stats in lib php on line these show up in recent logs as there is an ongoing effort to clean up php messages that end up in error log i am listing these here thank you
| 1
|
15,212
| 2,850,318,310
|
IssuesEvent
|
2015-05-31 13:34:36
|
damonkohler/sl4a
|
https://api.github.com/repos/damonkohler/sl4a
|
opened
|
sl4a not working
|
auto-migrated Priority-Medium Type-Defect
|
_From @GoogleCodeExporter on May 31, 2015 11:31_
```
What device(s) are you experiencing the problem on?
eeepc
What firmware version are you running on the device?
android x86 4.04 eeepc
What steps will reproduce the problem?
1.installing sl4a
2.installing Python for android
3.running the hello_world.py script
What is the expected output? What do you see instead?
in step of showin hello world, sl4a crashes
What version of the product are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `bennor.i...@gmail.com` on 29 Oct 2012 at 4:16
_Copied from original issue: damonkohler/android-scripting#669_
|
1.0
|
sl4a not working - _From @GoogleCodeExporter on May 31, 2015 11:31_
```
What device(s) are you experiencing the problem on?
eeepc
What firmware version are you running on the device?
android x86 4.04 eeepc
What steps will reproduce the problem?
1.installing sl4a
2.installing Python for android
3.running the hello_world.py script
What is the expected output? What do you see instead?
in step of showin hello world, sl4a crashes
What version of the product are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `bennor.i...@gmail.com` on 29 Oct 2012 at 4:16
_Copied from original issue: damonkohler/android-scripting#669_
|
defect
|
not working from googlecodeexporter on may what device s are you experiencing the problem on eeepc what firmware version are you running on the device android eeepc what steps will reproduce the problem installing installing python for android running the hello world py script what is the expected output what do you see instead in step of showin hello world crashes what version of the product are you using on what operating system please provide any additional information below original issue reported on code google com by bennor i gmail com on oct at copied from original issue damonkohler android scripting
| 1
|
58,203
| 8,233,774,387
|
IssuesEvent
|
2018-09-08 05:33:59
|
gatsbyjs/gatsby
|
https://api.github.com/repos/gatsbyjs/gatsby
|
closed
|
TO DO after v2 is done: separate search for v1 and v2 docs
|
type: documentation
|
## Summary
## Relevant information
@m-allanson and I think that Algolia search is collecting all search data from v1 and v2 docs into one stream. Seems like down the road, as v2 progresses, it might be handy to separate that stream into two, so we can get data specific to each site and improve the docs based on the data.
Wondering: are there any reasons to keep v1 and v2 search data in one stream?
|
1.0
|
TO DO after v2 is done: separate search for v1 and v2 docs - ## Summary
## Relevant information
@m-allanson and I think that Algolia search is collecting all search data from v1 and v2 docs into one stream. Seems like down the road, as v2 progresses, it might be handy to separate that stream into two, so we can get data specific to each site and improve the docs based on the data.
Wondering: are there any reasons to keep v1 and v2 search data in one stream?
|
non_defect
|
to do after is done separate search for and docs summary relevant information m allanson and i think that algolia search is collecting all search data from and docs into one stream seems like down the road as progresses it might be handy to separate that stream into two so we can get data specific to each site and improve the docs based on the data wondering are there any reasons to keep and search data in one stream
| 0
|
23,047
| 3,755,593,102
|
IssuesEvent
|
2016-03-12 19:27:48
|
RomanGolovanov/aMetro
|
https://api.github.com/repos/RomanGolovanov/aMetro
|
closed
|
Уведомление завершения обработки (Загрузки, Импорта, Удаления)
|
auto-migrated Component-UI Priority-Low Type-Defect
|
```
Отображать всплывающее уведомление об
успешном выполнении обработки карты как в
детализации карты (если делали вручную
поштучно), так и в списке (если делали
массово).
```
Original issue reported on code.google.com by `G.Glaur...@gmail.com` on 27 Jul 2010 at 11:30
|
1.0
|
Уведомление завершения обработки (Загрузки, Импорта, Удаления) - ```
Отображать всплывающее уведомление об
успешном выполнении обработки карты как в
детализации карты (если делали вручную
поштучно), так и в списке (если делали
массово).
```
Original issue reported on code.google.com by `G.Glaur...@gmail.com` on 27 Jul 2010 at 11:30
|
defect
|
уведомление завершения обработки загрузки импорта удаления отображать всплывающее уведомление об успешном выполнении обработки карты как в детализации карты если делали вручную поштучно так и в списке если делали массово original issue reported on code google com by g glaur gmail com on jul at
| 1
|
332,815
| 10,111,283,024
|
IssuesEvent
|
2019-07-30 12:24:47
|
linkerd/linkerd2
|
https://api.github.com/repos/linkerd/linkerd2
|
closed
|
`top` output has no column headers
|
area/cli good first issue help wanted priority/P1
|
The output of `linkerd top` has no column headers, making it unclear what is actually being reported. (2.4.0)
<img width="1340" alt="image" src="https://user-images.githubusercontent.com/4026/61723195-8d23c080-ad20-11e9-9339-accc3707c7ca.png">
|
1.0
|
`top` output has no column headers - The output of `linkerd top` has no column headers, making it unclear what is actually being reported. (2.4.0)
<img width="1340" alt="image" src="https://user-images.githubusercontent.com/4026/61723195-8d23c080-ad20-11e9-9339-accc3707c7ca.png">
|
non_defect
|
top output has no column headers the output of linkerd top has no column headers making it unclear what is actually being reported img width alt image src
| 0
|
3,590
| 6,500,770,156
|
IssuesEvent
|
2017-08-23 06:44:38
|
shoes/shoes4
|
https://api.github.com/repos/shoes/shoes4
|
closed
|
Focus method for edit_line not available
|
compatibility enhancement
|
It appears that the focus method for the input_box class is not implemented.
The following code will recreate the issue:
```ruby
class Focus < Shoes
url '/', :index
url '/text_input', :text_input
def index
para link("enter data", click: "/text_input")
end
def text_input
@input = edit_line width: 400, top: 20, left: 20
@input.focus()
end
end
```
After clicking on "enter data" the following error will be reported:
```
ERROR: undefined method `focus' for #<Shoes::Swt::EditLine:0xff6077>
```
(If `edit_line` is called in the `:index` method, the input will still receive focus automatically despite the failure to focus, so it isn't immediately clear there is an issue.)
Untested solution: Add the following to `shoes-swt/lib/shoes/swt/input_box.rb`
```ruby
def focus
@real.set_focus()
end
```
|
True
|
Focus method for edit_line not available - It appears that the focus method for the input_box class is not implemented.
The following code will recreate the issue:
```ruby
class Focus < Shoes
url '/', :index
url '/text_input', :text_input
def index
para link("enter data", click: "/text_input")
end
def text_input
@input = edit_line width: 400, top: 20, left: 20
@input.focus()
end
end
```
After clicking on "enter data" the following error will be reported:
```
ERROR: undefined method `focus' for #<Shoes::Swt::EditLine:0xff6077>
```
(If `edit_line` is called in the `:index` method, the input will still receive focus automatically despite the failure to focus, so it isn't immediately clear there is an issue.)
Untested solution: Add the following to `shoes-swt/lib/shoes/swt/input_box.rb`
```ruby
def focus
@real.set_focus()
end
```
|
non_defect
|
focus method for edit line not available it appears that the focus method for the input box class is not implemented the following code will recreate the issue ruby class focus shoes url index url text input text input def index para link enter data click text input end def text input input edit line width top left input focus end end after clicking on enter data the following error will be reported error undefined method focus for if edit line is called in the index method the input will still receive focus automatically despite the failure to focus so it isn t immediately clear there is an issue untested solution add the following to shoes swt lib shoes swt input box rb ruby def focus real set focus end
| 0
|
212,357
| 16,444,839,605
|
IssuesEvent
|
2021-05-20 18:16:50
|
BabylonJS/Babylon.js
|
https://api.github.com/repos/BabylonJS/Babylon.js
|
closed
|
Create Unit Tests for Input
|
Testing
|
We need to create unit tests to verify that any changes to the DeviceInputSystem/DeviceSourceManager won't break things like gizmos
|
1.0
|
Create Unit Tests for Input - We need to create unit tests to verify that any changes to the DeviceInputSystem/DeviceSourceManager won't break things like gizmos
|
non_defect
|
create unit tests for input we need to create unit tests to verify that any changes to the deviceinputsystem devicesourcemanager won t break things like gizmos
| 0
|
67,143
| 27,737,560,195
|
IssuesEvent
|
2023-03-15 12:18:27
|
dotnet/fsharp
|
https://api.github.com/repos/dotnet/fsharp
|
opened
|
Finding references of additional constructors in VS
|
Feature Request Area-LangService-FindAllReferences Needs-Triage
|
Consider the following example:
```fsharp
type MyClass(x) =
new() = MyClass(0)
let myInstance = MyClass()
```
If we issue _Find All References_ on the additional constructor (`new` on line 2) it would be nice if we get back all the places this constructor is used (`MyClass()` on line 4). At the moment we don't get any results.
|
1.0
|
Finding references of additional constructors in VS - Consider the following example:
```fsharp
type MyClass(x) =
new() = MyClass(0)
let myInstance = MyClass()
```
If we issue _Find All References_ on the additional constructor (`new` on line 2) it would be nice if we get back all the places this constructor is used (`MyClass()` on line 4). At the moment we don't get any results.
|
non_defect
|
finding references of additional constructors in vs consider the following example fsharp type myclass x new myclass let myinstance myclass if we issue find all references on the additional constructor new on line it would be nice if we get back all the places this constructor is used myclass on line at the moment we don t get any results
| 0
|
32,888
| 13,941,605,853
|
IssuesEvent
|
2020-10-22 19:40:29
|
Azure/azure-sdk-for-js
|
https://api.github.com/repos/Azure/azure-sdk-for-js
|
closed
|
[Service Bus] Swap the order of messages & enqueuedTime in scheduled_messages method
|
Client Service Bus good first issue help wanted
|
The `scheduled_messages()` method in the Service Bus package takes the time first and the messages later. It is more intuitive to swap this order because the message(s) is the main entity here. This also makes us consistent with the Service Bus packages in other languages
PRs are welcome to add this feature.
Code Pointers:
- [Set up environment](https://github.com/Azure/azure-sdk-for-js/blob/master/CONTRIBUTING.md#setting-up-your-environment) and run `rush build`
- Open the `sdk/servicebus/service-bus` folder in your editor/IDE, find the `scheduled_messages()` method and swap the order
|
1.0
|
[Service Bus] Swap the order of messages & enqueuedTime in scheduled_messages method - The `scheduled_messages()` method in the Service Bus package takes the time first and the messages later. It is more intuitive to swap this order because the message(s) is the main entity here. This also makes us consistent with the Service Bus packages in other languages
PRs are welcome to add this feature.
Code Pointers:
- [Set up environment](https://github.com/Azure/azure-sdk-for-js/blob/master/CONTRIBUTING.md#setting-up-your-environment) and run `rush build`
- Open the `sdk/servicebus/service-bus` folder in your editor/IDE, find the `scheduled_messages()` method and swap the order
|
non_defect
|
swap the order of messages enqueuedtime in scheduled messages method the scheduled messages method in the service bus package takes the time first and the messages later it is more intuitive to swap this order because the message s is the main entity here this also makes us consistent with the service bus packages in other languages prs are welcome to add this feature code pointers and run rush build open the sdk servicebus service bus folder in your editor ide find the scheduled messages method and swap the order
| 0
|
19,827
| 3,264,912,071
|
IssuesEvent
|
2015-10-22 14:11:00
|
contao/core-bundle
|
https://api.github.com/repos/contao/core-bundle
|
closed
|
Correctly escape back slashes in single quoted strings
|
defect
|
We are often doing the following:
```php
private $mapper = [
'Contao\\CoreBundle\\Exception\\AccessDeniedException' => 'AccessDeniedHttpException',
'Contao\\CoreBundle\\Exception\\ForwardPageNotFoundException' => 'InternalServerErrorHttpException',
'Contao\\CoreBundle\\Exception\\IncompleteInstallationException' => 'InternalServerErrorHttpException',
];
```
But in fact, the double back slashes are wrong here. According to [the PHP manual](http://php.net/manual/en/language.types.string.php#language.types.string.syntax.single), the back slash only needs to be escaped if it is followed by a single quote:
```php
$a = 'No need to escape the \ here.';
$b = 'The following character needs to be escaped here: \\';
```
We should fix this in the non-legacy code.
|
1.0
|
Correctly escape back slashes in single quoted strings - We are often doing the following:
```php
private $mapper = [
'Contao\\CoreBundle\\Exception\\AccessDeniedException' => 'AccessDeniedHttpException',
'Contao\\CoreBundle\\Exception\\ForwardPageNotFoundException' => 'InternalServerErrorHttpException',
'Contao\\CoreBundle\\Exception\\IncompleteInstallationException' => 'InternalServerErrorHttpException',
];
```
But in fact, the double back slashes are wrong here. According to [the PHP manual](http://php.net/manual/en/language.types.string.php#language.types.string.syntax.single), the back slash only needs to be escaped if it is followed by a single quote:
```php
$a = 'No need to escape the \ here.';
$b = 'The following character needs to be escaped here: \\';
```
We should fix this in the non-legacy code.
|
defect
|
correctly escape back slashes in single quoted strings we are often doing the following php private mapper contao corebundle exception accessdeniedexception accessdeniedhttpexception contao corebundle exception forwardpagenotfoundexception internalservererrorhttpexception contao corebundle exception incompleteinstallationexception internalservererrorhttpexception but in fact the double back slashes are wrong here according to the back slash only needs to be escaped if it is followed by a single quote php a no need to escape the here b the following character needs to be escaped here we should fix this in the non legacy code
| 1
|
89,342
| 11,216,352,118
|
IssuesEvent
|
2020-01-07 06:02:02
|
ZeffonWu/algo
|
https://api.github.com/repos/ZeffonWu/algo
|
opened
|
341.扁平化嵌套列表迭代器 | Zeffon's blog | 算法博客
|
Gitalk https://algo.zeffon.design/posts/b4c84e9e.html
|
https://algo.zeffon.design/posts/b4c84e9e.html
题目要求给定一个嵌套的整型列表。设计一个迭代器,使其能够遍历这个整型列表中的所有整数。列表中的项或者为一个整数,或者是另一个列表。 题目示例 示例 1: 123输入: [[1,1],2,[1,1]]输出: [1,1,2,1,1]解释: 通过重复调用 next 直到 hasNext 返回false,next 返回的元素的顺序应该是: [1,1,2,1,1]。 示例
|
1.0
|
341.扁平化嵌套列表迭代器 | Zeffon's blog | 算法博客 - https://algo.zeffon.design/posts/b4c84e9e.html
题目要求给定一个嵌套的整型列表。设计一个迭代器,使其能够遍历这个整型列表中的所有整数。列表中的项或者为一个整数,或者是另一个列表。 题目示例 示例 1: 123输入: [[1,1],2,[1,1]]输出: [1,1,2,1,1]解释: 通过重复调用 next 直到 hasNext 返回false,next 返回的元素的顺序应该是: [1,1,2,1,1]。 示例
|
non_defect
|
扁平化嵌套列表迭代器 zeffon s blog 算法博客 题目要求给定一个嵌套的整型列表。设计一个迭代器,使其能够遍历这个整型列表中的所有整数。列表中的项或者为一个整数,或者是另一个列表。 题目示例 示例 输出 解释 通过重复调用 nbsp next 直到 nbsp hasnext 返回false,next nbsp 返回的元素的顺序应该是 。 示例
| 0
|
75,813
| 9,330,995,804
|
IssuesEvent
|
2019-03-28 08:41:42
|
canonical-websites/build.snapcraft.io
|
https://api.github.com/repos/canonical-websites/build.snapcraft.io
|
closed
|
"Hi, you" should link to repos page (/user/you)
|
Design: Required Priority: Low
|
Could the "Hi, you" link to /users/you (the repos page) on pages other than /user/you?
I did some quick user testing with @thomir last night and he missed the "let's take a look at your repos" link on the front page.
|
1.0
|
"Hi, you" should link to repos page (/user/you) - Could the "Hi, you" link to /users/you (the repos page) on pages other than /user/you?
I did some quick user testing with @thomir last night and he missed the "let's take a look at your repos" link on the front page.
|
non_defect
|
hi you should link to repos page user you could the hi you link to users you the repos page on pages other than user you i did some quick user testing with thomir last night and he missed the let s take a look at your repos link on the front page
| 0
|
104,461
| 8,972,675,419
|
IssuesEvent
|
2019-01-29 18:55:46
|
codercom/bugs
|
https://api.github.com/repos/codercom/bugs
|
opened
|
Python debugger receives wrong environment when building debugging commands
|
bug testing
|
Jan 29, 2019 at 12:48:05 PM
**Describe the bug**
A user on discord has reported that the Python extension/debugger appears to intercept the OS the browser is running on when building the command to run the debugger. As a result, if a user is on windows, the command is built using `cmd` which will fail as a result of the container running Ubuntu.
**Environment**
Production
**To Reproduce**
Steps to reproduce the behavior:
1. Open the IDE (on windows)
2. Install flask
3. Create a hello world file
4. Click the debugger in the activity bar
5. Click start
6. See that the command that is built is incorrect
**Expected behavior**
The command should be based on the Ubuntu container it is running on
**Screenshots**

**Desktop:**
Windows (I assume)
Chrome
|
1.0
|
Python debugger receives wrong environment when building debugging commands - Jan 29, 2019 at 12:48:05 PM
**Describe the bug**
A user on discord has reported that the Python extension/debugger appears to intercept the OS the browser is running on when building the command to run the debugger. As a result, if a user is on windows, the command is built using `cmd` which will fail as a result of the container running Ubuntu.
**Environment**
Production
**To Reproduce**
Steps to reproduce the behavior:
1. Open the IDE (on windows)
2. Install flask
3. Create a hello world file
4. Click the debugger in the activity bar
5. Click start
6. See that the command that is built is incorrect
**Expected behavior**
The command should be based on the Ubuntu container it is running on
**Screenshots**

**Desktop:**
Windows (I assume)
Chrome
|
non_defect
|
python debugger receives wrong environment when building debugging commands jan at pm describe the bug a user on discord has reported that the python extension debugger appears to intercept the os the browser is running on when building the command to run the debugger as a result if a user is on windows the command is built using cmd which will fail as a result of the container running ubuntu environment production to reproduce steps to reproduce the behavior open the ide on windows install flask create a hello world file click the debugger in the activity bar click start see that the command that is built is incorrect expected behavior the command should be based on the ubuntu container it is running on screenshots desktop windows i assume chrome
| 0
|
396,681
| 27,131,026,978
|
IssuesEvent
|
2023-02-16 09:48:05
|
wagtail/wagtail
|
https://api.github.com/repos/wagtail/wagtail
|
closed
|
The left side of the Algolia InstantSearch results panel is cut off when viewing in portrait mode on mobile phone
|
Documentation good first issue
|
The left side of the Algolia InstantSearch results panel is cut off when viewing in portrait mode on mobile phone (see attached).
Mobile browser is Chrome. Is there a way to modify the CSS in the theme to adjust for this? Thanks!

|
1.0
|
The left side of the Algolia InstantSearch results panel is cut off when viewing in portrait mode on mobile phone - The left side of the Algolia InstantSearch results panel is cut off when viewing in portrait mode on mobile phone (see attached).
Mobile browser is Chrome. Is there a way to modify the CSS in the theme to adjust for this? Thanks!

|
non_defect
|
the left side of the algolia instantsearch results panel is cut off when viewing in portrait mode on mobile phone the left side of the algolia instantsearch results panel is cut off when viewing in portrait mode on mobile phone see attached mobile browser is chrome is there a way to modify the css in the theme to adjust for this thanks
| 0
|
408,528
| 27,694,535,796
|
IssuesEvent
|
2023-03-14 00:22:15
|
futantan/OpenGpt
|
https://api.github.com/repos/futantan/OpenGpt
|
closed
|
📝 文档改善
|
documentation
|
# 文档改善建议 📝
欢迎在此分享您对文档的改善建议,我们期待听到您的想法和建议。
## 您的建议是什么? 🤔
请简要描述您的文档改善建议,包括您的目标和想法。
如果您的建议是解决某个特定问题的,请尽可能提供更多的上下文和细节。
## 您的建议有哪些优势? 🌟
请简要描述您的建议的优势和特点,比如:
- 是否可以提高文档的可读性和易用性?
- 是否可以使文档更加详细和准确?
- 是否可以让文档更好地反映项目的实际情况?
感谢您的分享和支持!🙏
|
1.0
|
📝 文档改善 - # 文档改善建议 📝
欢迎在此分享您对文档的改善建议,我们期待听到您的想法和建议。
## 您的建议是什么? 🤔
请简要描述您的文档改善建议,包括您的目标和想法。
如果您的建议是解决某个特定问题的,请尽可能提供更多的上下文和细节。
## 您的建议有哪些优势? 🌟
请简要描述您的建议的优势和特点,比如:
- 是否可以提高文档的可读性和易用性?
- 是否可以使文档更加详细和准确?
- 是否可以让文档更好地反映项目的实际情况?
感谢您的分享和支持!🙏
|
non_defect
|
📝 文档改善 文档改善建议 📝 欢迎在此分享您对文档的改善建议,我们期待听到您的想法和建议。 您的建议是什么? 🤔 请简要描述您的文档改善建议,包括您的目标和想法。 如果您的建议是解决某个特定问题的,请尽可能提供更多的上下文和细节。 您的建议有哪些优势? 🌟 请简要描述您的建议的优势和特点,比如: 是否可以提高文档的可读性和易用性? 是否可以使文档更加详细和准确? 是否可以让文档更好地反映项目的实际情况? 感谢您的分享和支持!🙏
| 0
|
311,183
| 26,774,275,658
|
IssuesEvent
|
2023-01-31 16:04:54
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix math.test_tensorflow_reduce_prod
|
TensorFlow Frontend Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4011265503/jobs/6888662116" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4054925606/jobs/6977418144" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4054925606/jobs/6977418144" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4054925606/jobs/6977402738" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>Not found</summary>
Not found
</details>
|
1.0
|
Fix math.test_tensorflow_reduce_prod - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4011265503/jobs/6888662116" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4054925606/jobs/6977418144" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4054925606/jobs/6977418144" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4054925606/jobs/6977402738" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>Not found</summary>
Not found
</details>
|
non_defect
|
fix math test tensorflow reduce prod tensorflow img src torch img src numpy img src jax img src not found not found
| 0
|
67,589
| 27,957,748,992
|
IssuesEvent
|
2023-03-24 13:37:23
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
[Bug] 311 service request assets are not associating to a location
|
Workgroup: AMD Type: Bug Report Impact: 3-Minor Service: Dev Product: AMD Data Tracker
|
<!-- Email -->
<!-- lance.ballard@austintexas.gov -->
> What application are you using?
Data Tracker
> Describe the problem.
Assets are not being automatically associated when brought over from 311 even when the location is exactly on top of the asset.
> Website Address
https://atd.knack.com/amd#mmc/
Internet Browser: Chrome
> Requested By
Lance B.
Request ID: DTS22-106010
|
1.0
|
[Bug] 311 service request assets are not associating to a location - <!-- Email -->
<!-- lance.ballard@austintexas.gov -->
> What application are you using?
Data Tracker
> Describe the problem.
Assets are not being automatically associated when brought over from 311 even when the location is exactly on top of the asset.
> Website Address
https://atd.knack.com/amd#mmc/
Internet Browser: Chrome
> Requested By
Lance B.
Request ID: DTS22-106010
|
non_defect
|
service request assets are not associating to a location what application are you using data tracker describe the problem assets are not being automatically associated when brought over from even when the location is exactly on top of the asset website address internet browser chrome requested by lance b request id
| 0
|
78,038
| 27,290,126,513
|
IssuesEvent
|
2023-02-23 16:03:08
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Admin request in room where you are common user
|
T-Defect
|
### Steps to reproduce
1. Visit any room where you are a common user.
2. Open user list in room.
3. Open console.
4. View user information.
5. Element tries GET https://matrix-client.matrix.org/_synapse/admin/v1/users/%40nightybrowser%3Amatrix.org/admin 403
fetch @ fetch.ts:62
requestOtherUrl @ fetch.ts:264
request @ fetch.ts:205
authedRequest @ fetch.ts:156
isSynapseAdministrator @ client.ts:6927
(anonymous) @ UserInfo.tsx:964
Bl @ react-dom.production.min.js:262
t.unstable_runWithPriority @ scheduler.production.min.js:18
$i @ react-dom.production.min.js:122
Ml @ react-dom.production.min.js:261
(anonymous) @ react-dom.production.min.js:261
L @ scheduler.production.min.js:16
E.port1.onmessage @ scheduler.production.min.js:12
### Outcome
#### What did you expect?
Element works like for a common user
#### What happened instead?
Element tries to work like for admin
```
fetch.ts:62
GET https://matrix-client.matrix.org/_synapse/admin/v1/users/%40nightybrowser%3Amatrix.org/admin 403
fetch @ fetch.ts:62
requestOtherUrl @ fetch.ts:264
request @ fetch.ts:205
authedRequest @ fetch.ts:156
isSynapseAdministrator @ client.ts:6927
(anonymous) @ UserInfo.tsx:964
Bl @ react-dom.production.min.js:262
t.unstable_runWithPriority @ scheduler.production.min.js:18
$i @ react-dom.production.min.js:122
Ml @ react-dom.production.min.js:261
(anonymous) @ react-dom.production.min.js:261
L @ scheduler.production.min.js:16
E.port1.onmessage @ scheduler.production.min.js:12
```
### Operating system
Windows
### Application version
1.11.23
### How did you install the app?
In-app update
### Homeserver
matrix.org
### Will you send logs?
No
|
1.0
|
Admin request in room where you are common user - ### Steps to reproduce
1. Visit any room where you are a common user.
2. Open user list in room.
3. Open console.
4. View user information.
5. Element tries GET https://matrix-client.matrix.org/_synapse/admin/v1/users/%40nightybrowser%3Amatrix.org/admin 403
fetch @ fetch.ts:62
requestOtherUrl @ fetch.ts:264
request @ fetch.ts:205
authedRequest @ fetch.ts:156
isSynapseAdministrator @ client.ts:6927
(anonymous) @ UserInfo.tsx:964
Bl @ react-dom.production.min.js:262
t.unstable_runWithPriority @ scheduler.production.min.js:18
$i @ react-dom.production.min.js:122
Ml @ react-dom.production.min.js:261
(anonymous) @ react-dom.production.min.js:261
L @ scheduler.production.min.js:16
E.port1.onmessage @ scheduler.production.min.js:12
### Outcome
#### What did you expect?
Element works like for a common user
#### What happened instead?
Element tries to work like for admin
```
fetch.ts:62
GET https://matrix-client.matrix.org/_synapse/admin/v1/users/%40nightybrowser%3Amatrix.org/admin 403
fetch @ fetch.ts:62
requestOtherUrl @ fetch.ts:264
request @ fetch.ts:205
authedRequest @ fetch.ts:156
isSynapseAdministrator @ client.ts:6927
(anonymous) @ UserInfo.tsx:964
Bl @ react-dom.production.min.js:262
t.unstable_runWithPriority @ scheduler.production.min.js:18
$i @ react-dom.production.min.js:122
Ml @ react-dom.production.min.js:261
(anonymous) @ react-dom.production.min.js:261
L @ scheduler.production.min.js:16
E.port1.onmessage @ scheduler.production.min.js:12
```
### Operating system
Windows
### Application version
1.11.23
### How did you install the app?
In-app update
### Homeserver
matrix.org
### Will you send logs?
No
|
defect
|
admin request in room where you are common user steps to reproduce visit any room where you are a common user open user list in room open console view user information element tries get fetch fetch ts requestotherurl fetch ts request fetch ts authedrequest fetch ts issynapseadministrator client ts anonymous userinfo tsx bl react dom production min js t unstable runwithpriority scheduler production min js i react dom production min js ml react dom production min js anonymous react dom production min js l scheduler production min js e onmessage scheduler production min js outcome what did you expect element works like for a common user what happened instead element tries to work like for admin fetch ts get fetch fetch ts requestotherurl fetch ts request fetch ts authedrequest fetch ts issynapseadministrator client ts anonymous userinfo tsx bl react dom production min js t unstable runwithpriority scheduler production min js i react dom production min js ml react dom production min js anonymous react dom production min js l scheduler production min js e onmessage scheduler production min js operating system windows application version how did you install the app in app update homeserver matrix org will you send logs no
| 1
|
12,334
| 2,692,109,353
|
IssuesEvent
|
2015-04-01 03:55:19
|
JesusFreke/smali
|
https://api.github.com/repos/JesusFreke/smali
|
closed
|
[CLOSED] Java 5 JVMs not suppored
|
auto-migrated Priority-Medium Type-Defect
|
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Wed Apr 1 03:09:04 2015_
_Originally opened as https://github.com/JesusFreke/smali_imported/issues/1_
----
```
What steps will reproduce the problem?
1. Try running smali-0.9.1.jar on Java 5
What is the expected output? What do you see instead?
It's supposed to work, but I get missing class errors (java.util.ArrayQueue
is one)
What version of the product are you using? On what operating system?
0.9.1, on Mac OS X. On Max OS X, Java 5 is the default so this is basically
blocking out most Mac users.
Please provide any additional information below.
I think this should be fixed as most of these problems can be fixed (swap
LinkedList for ArrayQueue for example) and it stops especially Mac users
from using the software (you can work around it, but this is hard).
```
Original issue reported on code.google.com by `isaacwal...@gmail.com` on 24 Jun 2009 at 9:37
|
1.0
|
[CLOSED] Java 5 JVMs not suppored - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Wed Apr 1 03:09:04 2015_
_Originally opened as https://github.com/JesusFreke/smali_imported/issues/1_
----
```
What steps will reproduce the problem?
1. Try running smali-0.9.1.jar on Java 5
What is the expected output? What do you see instead?
It's supposed to work, but I get missing class errors (java.util.ArrayQueue
is one)
What version of the product are you using? On what operating system?
0.9.1, on Mac OS X. On Max OS X, Java 5 is the default so this is basically
blocking out most Mac users.
Please provide any additional information below.
I think this should be fixed as most of these problems can be fixed (swap
LinkedList for ArrayQueue for example) and it stops especially Mac users
from using the software (you can work around it, but this is hard).
```
Original issue reported on code.google.com by `isaacwal...@gmail.com` on 24 Jun 2009 at 9:37
|
defect
|
java jvms not suppored issue by wed apr originally opened as what steps will reproduce the problem try running smali jar on java what is the expected output what do you see instead it s supposed to work but i get missing class errors java util arrayqueue is one what version of the product are you using on what operating system on mac os x on max os x java is the default so this is basically blocking out most mac users please provide any additional information below i think this should be fixed as most of these problems can be fixed swap linkedlist for arrayqueue for example and it stops especially mac users from using the software you can work around it but this is hard original issue reported on code google com by isaacwal gmail com on jun at
| 1
|
45,553
| 24,105,647,384
|
IssuesEvent
|
2022-09-20 07:12:49
|
The-FireHub-Project-old/FireHub_Core
|
https://api.github.com/repos/The-FireHub-Project-old/FireHub_Core
|
opened
|
[Performance] : bvfnb
|
status: Waiting Triage type: Performance
|
### For which platform you have performance issue?
FireHub Core
### Describe the solution and feature you'd have performance issues with
fcngg
### Describe alternatives you've considered
b
### Additional context
_No response_
|
True
|
[Performance] : bvfnb - ### For which platform you have performance issue?
FireHub Core
### Describe the solution and feature you'd have performance issues with
fcngg
### Describe alternatives you've considered
b
### Additional context
_No response_
|
non_defect
|
bvfnb for which platform you have performance issue firehub core describe the solution and feature you d have performance issues with fcngg describe alternatives you ve considered b additional context no response
| 0
|
63,247
| 17,484,441,590
|
IssuesEvent
|
2021-08-09 09:07:45
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
FileUpload: Chunked file upload throws ClassCastException if resume and commons-upload
|
defect
|
**Describe the defect**
Chunked FileUpload in advanced mode throws ClassCastException if `UPLOADER=commons` and `FileUpload Resume Servlet` is configured.
```
org.primefaces.webapp.FileUploadChunksServlet sendError
SEVERE: class org.eclipse.jetty.server.Request cannot be cast to class org.primefaces.webapp.MultipartRequest (org.eclipse.jetty.server.Request is in unnamed module of loader org.codehaus.plexus.classworlds.realm.ClassRealm @60a7e509; org.primefaces.webapp.MultipartRequest is in unnamed module of loader org.eclipse.jetty.webapp.WebAppClassLoader @76973a6)
java.lang.ClassCastException: class org.eclipse.jetty.server.Request cannot be cast to class org.primefaces.webapp.MultipartRequest (org.eclipse.jetty.server.Request is in unnamed module of loader org.codehaus.plexus.classworlds.realm.ClassRealm @60a7e509; org.primefaces.webapp.MultipartRequest is in unnamed module of loader org.eclipse.jetty.webapp.WebAppClassLoader @76973a6)
at org.primefaces.component.fileupload.CommonsFileUploadDecoder.getUploadDirectory(CommonsFileUploadDecoder.java:36)
at org.primefaces.util.FileUploadUtils.getChunkDir(FileUploadUtils.java:397)
at org.primefaces.util.FileUploadUtils.listChunks(FileUploadUtils.java:376)
at org.primefaces.component.fileupload.AbstractFileUploadDecoder.decodeUploadedBytes(AbstractFileUploadDecoder.java:132)
at org.primefaces.webapp.FileUploadChunksServlet.doGet(FileUploadChunksServlet.java:47)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:687)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
at org.eclipse.jetty.servlet.ServletHolder$NotAsync.service(ServletHolder.java:1450)
```
Browser console shows
```
GET http://localhost:8080/primefaces-test/file/resume/?X-File-Id=file.ext,1625515344526,application/x-zip-compressed,7464948
500 Server Error
{
"servlet":"FileUpload Resume Servlet",
"message":"Server Error",
"url":"/primefaces-test/file/resume/",
"status":"500"
}
```
**Reproducer**
[PF-7364-resume.zip](https://github.com/primefaces/primefaces/files/6918554/PF-7364-resume.zip)
**Environment:**
- PF Version: _10.0_
- JSF + version: _Mojarra and MyFaces_
- Affected browsers: _ALL_
**To Reproduce**
Steps to reproduce the behavior:
1. Start server `mvn jetty:run`
2. Go to 'localhost:8080/primefaces-test/'
3. Click 'choose'
4. Select any non empty file
5. See errors in server log and browser console
**Expected behavior**
ClassCastException should not be thrown.
**Example XHTML**
```html
<h:form id="frmTest">
<p:fileUpload id="fileupload" mode="advanced" maxChunkSize="65536" .../>
</h:form>
```
**Example web.xml**
```
...
<context-param>
<param-name>primefaces.UPLOADER</param-name>
<param-value>commons</param-value>
</context-param>
...
<filter>
<filter-name>PrimeFaces FileUpload Filter</filter-name>
<filter-class>org.primefaces.webapp.filter.FileUploadFilter</filter-class>
</filter>
<filter-mapping>
<filter-name>PrimeFaces FileUpload Filter</filter-name>
<servlet-name>Faces Servlet</servlet-name>
</filter-mapping>
...
<servlet>
<servlet-name>FileUpload Resume Servlet</servlet-name>
<servlet-class>org.primefaces.webapp.FileUploadChunksServlet</servlet-class>
</servlet>
<servlet-mapping>
<servlet-name>FileUpload Resume Servlet</servlet-name>
<url-pattern>/file/resume/</url-pattern>
</servlet-mapping>
```
**Example Bean**
see Reproducer
|
1.0
|
FileUpload: Chunked file upload throws ClassCastException if resume and commons-upload - **Describe the defect**
Chunked FileUpload in advanced mode throws ClassCastException if `UPLOADER=commons` and `FileUpload Resume Servlet` is configured.
```
org.primefaces.webapp.FileUploadChunksServlet sendError
SEVERE: class org.eclipse.jetty.server.Request cannot be cast to class org.primefaces.webapp.MultipartRequest (org.eclipse.jetty.server.Request is in unnamed module of loader org.codehaus.plexus.classworlds.realm.ClassRealm @60a7e509; org.primefaces.webapp.MultipartRequest is in unnamed module of loader org.eclipse.jetty.webapp.WebAppClassLoader @76973a6)
java.lang.ClassCastException: class org.eclipse.jetty.server.Request cannot be cast to class org.primefaces.webapp.MultipartRequest (org.eclipse.jetty.server.Request is in unnamed module of loader org.codehaus.plexus.classworlds.realm.ClassRealm @60a7e509; org.primefaces.webapp.MultipartRequest is in unnamed module of loader org.eclipse.jetty.webapp.WebAppClassLoader @76973a6)
at org.primefaces.component.fileupload.CommonsFileUploadDecoder.getUploadDirectory(CommonsFileUploadDecoder.java:36)
at org.primefaces.util.FileUploadUtils.getChunkDir(FileUploadUtils.java:397)
at org.primefaces.util.FileUploadUtils.listChunks(FileUploadUtils.java:376)
at org.primefaces.component.fileupload.AbstractFileUploadDecoder.decodeUploadedBytes(AbstractFileUploadDecoder.java:132)
at org.primefaces.webapp.FileUploadChunksServlet.doGet(FileUploadChunksServlet.java:47)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:687)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
at org.eclipse.jetty.servlet.ServletHolder$NotAsync.service(ServletHolder.java:1450)
```
Browser console shows
```
GET http://localhost:8080/primefaces-test/file/resume/?X-File-Id=file.ext,1625515344526,application/x-zip-compressed,7464948
500 Server Error
{
"servlet":"FileUpload Resume Servlet",
"message":"Server Error",
"url":"/primefaces-test/file/resume/",
"status":"500"
}
```
**Reproducer**
[PF-7364-resume.zip](https://github.com/primefaces/primefaces/files/6918554/PF-7364-resume.zip)
**Environment:**
- PF Version: _10.0_
- JSF + version: _Mojarra and MyFaces_
- Affected browsers: _ALL_
**To Reproduce**
Steps to reproduce the behavior:
1. Start server `mvn jetty:run`
2. Go to 'localhost:8080/primefaces-test/'
3. Click 'choose'
4. Select any non empty file
5. See errors in server log and browser console
**Expected behavior**
ClassCastException should not be thrown.
**Example XHTML**
```html
<h:form id="frmTest">
<p:fileUpload id="fileupload" mode="advanced" maxChunkSize="65536" .../>
</h:form>
```
**Example web.xml**
```
...
<context-param>
<param-name>primefaces.UPLOADER</param-name>
<param-value>commons</param-value>
</context-param>
...
<filter>
<filter-name>PrimeFaces FileUpload Filter</filter-name>
<filter-class>org.primefaces.webapp.filter.FileUploadFilter</filter-class>
</filter>
<filter-mapping>
<filter-name>PrimeFaces FileUpload Filter</filter-name>
<servlet-name>Faces Servlet</servlet-name>
</filter-mapping>
...
<servlet>
<servlet-name>FileUpload Resume Servlet</servlet-name>
<servlet-class>org.primefaces.webapp.FileUploadChunksServlet</servlet-class>
</servlet>
<servlet-mapping>
<servlet-name>FileUpload Resume Servlet</servlet-name>
<url-pattern>/file/resume/</url-pattern>
</servlet-mapping>
```
**Example Bean**
see Reproducer
|
defect
|
fileupload chunked file upload throws classcastexception if resume and commons upload describe the defect chunked fileupload in advanced mode throws classcastexception if uploader commons and fileupload resume servlet is configured org primefaces webapp fileuploadchunksservlet senderror severe class org eclipse jetty server request cannot be cast to class org primefaces webapp multipartrequest org eclipse jetty server request is in unnamed module of loader org codehaus plexus classworlds realm classrealm org primefaces webapp multipartrequest is in unnamed module of loader org eclipse jetty webapp webappclassloader java lang classcastexception class org eclipse jetty server request cannot be cast to class org primefaces webapp multipartrequest org eclipse jetty server request is in unnamed module of loader org codehaus plexus classworlds realm classrealm org primefaces webapp multipartrequest is in unnamed module of loader org eclipse jetty webapp webappclassloader at org primefaces component fileupload commonsfileuploaddecoder getuploaddirectory commonsfileuploaddecoder java at org primefaces util fileuploadutils getchunkdir fileuploadutils java at org primefaces util fileuploadutils listchunks fileuploadutils java at org primefaces component fileupload abstractfileuploaddecoder decodeuploadedbytes abstractfileuploaddecoder java at org primefaces webapp fileuploadchunksservlet doget fileuploadchunksservlet java at javax servlet http httpservlet service httpservlet java at javax servlet http httpservlet service httpservlet java at org eclipse jetty servlet servletholder notasync service servletholder java browser console shows get server error servlet fileupload resume servlet message server error url primefaces test file resume status reproducer environment pf version jsf version mojarra and myfaces affected browsers all to reproduce steps to reproduce the behavior start server mvn jetty run go to localhost primefaces test click choose select any non empty file see errors in server log and browser console expected behavior classcastexception should not be thrown example xhtml html example web xml primefaces uploader commons primefaces fileupload filter org primefaces webapp filter fileuploadfilter primefaces fileupload filter faces servlet fileupload resume servlet org primefaces webapp fileuploadchunksservlet fileupload resume servlet file resume example bean see reproducer
| 1
|
212,410
| 23,893,914,517
|
IssuesEvent
|
2022-09-08 13:31:40
|
RG4421/ampere-centos-kernel
|
https://api.github.com/repos/RG4421/ampere-centos-kernel
|
closed
|
WS-2021-0441 (Medium) detected in linuxv5.2 - autoclosed
|
security vulnerability
|
## WS-2021-0441 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/usbhid/hid-core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/usbhid/hid-core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Linux/Kernel in versions v2.6.11-tree to v4.4.285, v4.5-rc1 to v4.9.284, v4.10-rc1 to v4.14.248, v4.15-rc1 to v4.19.208, v5.0-rc1 t0 v5.4.150, v5.5-rc1 to v5.10.70, v5.10-rc1 to v5.14.
Is vulnerable to free raw_report buffers in usbhid_stop in drivers/hid/usbhid/hid-core.c
<p>Publish Date: 2021-11-29
<p>URL: <a href=https://github.com/gregkh/linux/commit/2b704864c92dcec2b295f276fcfbfb81d9831f81>WS-2021-0441</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1001712">https://osv.dev/vulnerability/UVI-2021-1001712</a></p>
<p>Release Date: 2021-11-29</p>
<p>Fix Resolution: Linux/Kernel - v4.4.286, v4.9.285, v4.14.249, v4.19.209, v5.4.151, v5.10.71, v5.14.10</p>
</p>
</details>
<p></p>
|
True
|
WS-2021-0441 (Medium) detected in linuxv5.2 - autoclosed - ## WS-2021-0441 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/usbhid/hid-core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/usbhid/hid-core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Linux/Kernel in versions v2.6.11-tree to v4.4.285, v4.5-rc1 to v4.9.284, v4.10-rc1 to v4.14.248, v4.15-rc1 to v4.19.208, v5.0-rc1 t0 v5.4.150, v5.5-rc1 to v5.10.70, v5.10-rc1 to v5.14.
Is vulnerable to free raw_report buffers in usbhid_stop in drivers/hid/usbhid/hid-core.c
<p>Publish Date: 2021-11-29
<p>URL: <a href=https://github.com/gregkh/linux/commit/2b704864c92dcec2b295f276fcfbfb81d9831f81>WS-2021-0441</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1001712">https://osv.dev/vulnerability/UVI-2021-1001712</a></p>
<p>Release Date: 2021-11-29</p>
<p>Fix Resolution: Linux/Kernel - v4.4.286, v4.9.285, v4.14.249, v4.19.209, v5.4.151, v5.10.71, v5.14.10</p>
</p>
</details>
<p></p>
|
non_defect
|
ws medium detected in autoclosed ws medium severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files drivers hid usbhid hid core c drivers hid usbhid hid core c vulnerability details in linux kernel in versions tree to to to to to to is vulnerable to free raw report buffers in usbhid stop in drivers hid usbhid hid core c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux kernel
| 0
|
480,714
| 13,865,744,947
|
IssuesEvent
|
2020-10-16 05:14:02
|
TheTypingMatch/events-bot
|
https://api.github.com/repos/TheTypingMatch/events-bot
|
reopened
|
Participant status count - tournament info
|
feature low priority no-issue-activity
|
When displaying tournament info, show how many people are eliminated, in the losers bracket, and in the winners bracket.
|
1.0
|
Participant status count - tournament info - When displaying tournament info, show how many people are eliminated, in the losers bracket, and in the winners bracket.
|
non_defect
|
participant status count tournament info when displaying tournament info show how many people are eliminated in the losers bracket and in the winners bracket
| 0
|
64,637
| 18,774,929,670
|
IssuesEvent
|
2021-11-07 14:00:52
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Bubbles: Context menu button of deleted message hidden behind avatar
|
T-Defect
|
### Steps to reproduce
1. Delete own message.
2. Point mouse cursor to message placeholder.
3. See partially hidden button.
### Outcome
#### What did you expect?
Button shall be on located next to avatar same as for normal messages.
#### What happened instead?

### Operating system
Windows 10
### Browser information
Firefox 95b3
### URL for webapp
own
### Application version
Element version: 1.9.4-rc.1 Olm version: 3.2.3
### Homeserver
own
### Will you send logs?
No
|
1.0
|
Bubbles: Context menu button of deleted message hidden behind avatar - ### Steps to reproduce
1. Delete own message.
2. Point mouse cursor to message placeholder.
3. See partially hidden button.
### Outcome
#### What did you expect?
Button shall be on located next to avatar same as for normal messages.
#### What happened instead?

### Operating system
Windows 10
### Browser information
Firefox 95b3
### URL for webapp
own
### Application version
Element version: 1.9.4-rc.1 Olm version: 3.2.3
### Homeserver
own
### Will you send logs?
No
|
defect
|
bubbles context menu button of deleted message hidden behind avatar steps to reproduce delete own message point mouse cursor to message placeholder see partially hidden button outcome what did you expect button shall be on located next to avatar same as for normal messages what happened instead operating system windows browser information firefox url for webapp own application version element version rc olm version homeserver own will you send logs no
| 1
|
6,781
| 15,255,757,453
|
IssuesEvent
|
2021-02-20 17:21:53
|
livgust/covid-vaccine-scrapers
|
https://api.github.com/repos/livgust/covid-vaccine-scrapers
|
closed
|
Separating scrapers
|
architecture
|
Without getting into too many details, in my past scraping work we ran each site separately. I pushed a couple ways we could go about doing that below. Would be interested in any feedback from the group.
See options here: https://github.com/livgust/covid-vaccine-scrapers/compare/dynamic-scrapers
Usage would be:
ENV `SCRAPERS="PriceChopper Hannaford"`
or
`node main.js PriceChopper Hannaford`
|
1.0
|
Separating scrapers - Without getting into too many details, in my past scraping work we ran each site separately. I pushed a couple ways we could go about doing that below. Would be interested in any feedback from the group.
See options here: https://github.com/livgust/covid-vaccine-scrapers/compare/dynamic-scrapers
Usage would be:
ENV `SCRAPERS="PriceChopper Hannaford"`
or
`node main.js PriceChopper Hannaford`
|
non_defect
|
separating scrapers without getting into too many details in my past scraping work we ran each site separately i pushed a couple ways we could go about doing that below would be interested in any feedback from the group see options here usage would be env scrapers pricechopper hannaford or node main js pricechopper hannaford
| 0
|
243,914
| 7,868,274,553
|
IssuesEvent
|
2018-06-23 19:28:24
|
gnunn1/tilix
|
https://api.github.com/repos/gnunn1/tilix
|
closed
|
Quake mode: always appear on primary screen
|
Low Priority enhancement
|
Hello,
It would be great to have an option "Always on primary screen" for the quake mode - the same way Guake does it

The issue is that when I'm moving and plugging/unplugging external monitors (that I use as primary display), the display number changes and sometime tilix will show up on the laptop screen instead of the external monitor
This is a minor quality of life change - thanks for that awesome terminal!
|
1.0
|
Quake mode: always appear on primary screen - Hello,
It would be great to have an option "Always on primary screen" for the quake mode - the same way Guake does it

The issue is that when I'm moving and plugging/unplugging external monitors (that I use as primary display), the display number changes and sometime tilix will show up on the laptop screen instead of the external monitor
This is a minor quality of life change - thanks for that awesome terminal!
|
non_defect
|
quake mode always appear on primary screen hello it would be great to have an option always on primary screen for the quake mode the same way guake does it the issue is that when i m moving and plugging unplugging external monitors that i use as primary display the display number changes and sometime tilix will show up on the laptop screen instead of the external monitor this is a minor quality of life change thanks for that awesome terminal
| 0
|
336,298
| 24,491,896,521
|
IssuesEvent
|
2022-10-10 03:34:33
|
risingwavelabs/risingwave-docs
|
https://api.github.com/repos/risingwavelabs/risingwave-docs
|
closed
|
Update CREATE INDEX
|
documentation
|
### Related code PR
https://github.com/singularity-data/risingwave/pull/4340/files
### Which part(s) of the docs might be affected or should be updated? And how?
Update https://github.com/singularity-data/risingwave-docs/blob/main/temp/sql-create-index.md (syntax and examples).
CREATE INDEX is not shown to users because:
<img width="444" alt="image" src="https://user-images.githubusercontent.com/100549427/184094586-560e3fbc-7ee5-485c-82b8-31ddfea85602.png">
So keep the file in the temp folder.
Skip Release Notes.
### Reference
https://www.postgresql.org/docs/14/sql-createindex.html
|
1.0
|
Update CREATE INDEX - ### Related code PR
https://github.com/singularity-data/risingwave/pull/4340/files
### Which part(s) of the docs might be affected or should be updated? And how?
Update https://github.com/singularity-data/risingwave-docs/blob/main/temp/sql-create-index.md (syntax and examples).
CREATE INDEX is not shown to users because:
<img width="444" alt="image" src="https://user-images.githubusercontent.com/100549427/184094586-560e3fbc-7ee5-485c-82b8-31ddfea85602.png">
So keep the file in the temp folder.
Skip Release Notes.
### Reference
https://www.postgresql.org/docs/14/sql-createindex.html
|
non_defect
|
update create index related code pr which part s of the docs might be affected or should be updated and how update syntax and examples create index is not shown to users because img width alt image src so keep the file in the temp folder skip release notes reference
| 0
|
41,674
| 10,563,627,485
|
IssuesEvent
|
2019-10-04 21:32:35
|
mpawlow/github-milestone-generator
|
https://api.github.com/repos/mpawlow/github-milestone-generator
|
closed
|
Exception encountered trying to close milestones: TypeError: expected string or bytes-like object
|
priority-high severity-major type-defect
|
## Exception
```
$ python index.py -c -r dap/dap-planning
[2019-10-03 18:50:59,222] [INFO ] [__main__] [-- GITHUB MILESTONE GENERATOR ------------------------------------------------
[2019-10-03 18:50:59,222] [INFO ] [__main__] Parsed Command-line Arguments:
GitHub API Domain: github.ibm.com.
GitHub Repository: dap/dap-planning.
New Milestone Name: None.
New Milestone Due Date: None.
Close Milestones: True.
Logging Level: INFO.
Traceback (most recent call last):
File "index.py", line 744, in <module>
_main()
File "index.py", line 693, in _main
date=args.milestone_due_date)
File "index.py", line 244, in _validate_milestone_due_date
status = _validate_iso_8601_date(date)
File "index.py", line 259, in _validate_iso_8601_date
match = re.fullmatch(ISO_8601_DATE_REGEX, date)
File "C:\Users\MikePawlowski\AppData\Local\Programs\Python\Python37\lib\re.py", line 178, in fullmatch
return _compile(pattern, flags).fullmatch(string)
TypeError: expected string or bytes-like object
(github-milestone-generator)
```
|
1.0
|
Exception encountered trying to close milestones: TypeError: expected string or bytes-like object - ## Exception
```
$ python index.py -c -r dap/dap-planning
[2019-10-03 18:50:59,222] [INFO ] [__main__] [-- GITHUB MILESTONE GENERATOR ------------------------------------------------
[2019-10-03 18:50:59,222] [INFO ] [__main__] Parsed Command-line Arguments:
GitHub API Domain: github.ibm.com.
GitHub Repository: dap/dap-planning.
New Milestone Name: None.
New Milestone Due Date: None.
Close Milestones: True.
Logging Level: INFO.
Traceback (most recent call last):
File "index.py", line 744, in <module>
_main()
File "index.py", line 693, in _main
date=args.milestone_due_date)
File "index.py", line 244, in _validate_milestone_due_date
status = _validate_iso_8601_date(date)
File "index.py", line 259, in _validate_iso_8601_date
match = re.fullmatch(ISO_8601_DATE_REGEX, date)
File "C:\Users\MikePawlowski\AppData\Local\Programs\Python\Python37\lib\re.py", line 178, in fullmatch
return _compile(pattern, flags).fullmatch(string)
TypeError: expected string or bytes-like object
(github-milestone-generator)
```
|
defect
|
exception encountered trying to close milestones typeerror expected string or bytes like object exception python index py c r dap dap planning github milestone generator parsed command line arguments github api domain github ibm com github repository dap dap planning new milestone name none new milestone due date none close milestones true logging level info traceback most recent call last file index py line in main file index py line in main date args milestone due date file index py line in validate milestone due date status validate iso date date file index py line in validate iso date match re fullmatch iso date regex date file c users mikepawlowski appdata local programs python lib re py line in fullmatch return compile pattern flags fullmatch string typeerror expected string or bytes like object github milestone generator
| 1
|
76,399
| 26,409,841,702
|
IssuesEvent
|
2023-01-13 11:14:08
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
Slider: javax.faces.FacesException: No converter available for: class java.lang.String
|
:lady_beetle: defect
|
### Describe the bug
```
javax.faces.FacesException: No converter available for: class java.lang.String
at org.primefaces.component.slider.Slider.getValueAsStringOfAttachedInput(Slider.java:192) ~[primefaces-12.0.0.jar:12.0.0]
```
The method should return a string and already got one.
```
diff --git a/primefaces/src/main/java/org/primefaces/component/slider/Slider.java b/primefaces/src/main/java/org/primefaces/component/slider/Slider.java
index 2e8a538d8..7b483e2d2 100644
--- a/primefaces/src/main/java/org/primefaces/component/slider/Slider.java
+++ b/primefaces/src/main/java/org/primefaces/component/slider/Slider.java
@@ -184,6 +184,9 @@ public class Slider extends SliderBase {
Object value = input.getValue();
if (value != null) {
+ if (value instanceof String) {
+ return (String) value;
+ }
Converter converter = ComponentUtils.getConverter(context, input);
if (converter != null) {
return converter.getAsString(context, this, value);
```
### Reproducer
_No response_
### Expected behavior
_No response_
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.x
### Java version
11
### Browser(s)
_No response_
|
1.0
|
Slider: javax.faces.FacesException: No converter available for: class java.lang.String - ### Describe the bug
```
javax.faces.FacesException: No converter available for: class java.lang.String
at org.primefaces.component.slider.Slider.getValueAsStringOfAttachedInput(Slider.java:192) ~[primefaces-12.0.0.jar:12.0.0]
```
The method should return a string and already got one.
```
diff --git a/primefaces/src/main/java/org/primefaces/component/slider/Slider.java b/primefaces/src/main/java/org/primefaces/component/slider/Slider.java
index 2e8a538d8..7b483e2d2 100644
--- a/primefaces/src/main/java/org/primefaces/component/slider/Slider.java
+++ b/primefaces/src/main/java/org/primefaces/component/slider/Slider.java
@@ -184,6 +184,9 @@ public class Slider extends SliderBase {
Object value = input.getValue();
if (value != null) {
+ if (value instanceof String) {
+ return (String) value;
+ }
Converter converter = ComponentUtils.getConverter(context, input);
if (converter != null) {
return converter.getAsString(context, this, value);
```
### Reproducer
_No response_
### Expected behavior
_No response_
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.x
### Java version
11
### Browser(s)
_No response_
|
defect
|
slider javax faces facesexception no converter available for class java lang string describe the bug javax faces facesexception no converter available for class java lang string at org primefaces component slider slider getvalueasstringofattachedinput slider java the method should return a string and already got one diff git a primefaces src main java org primefaces component slider slider java b primefaces src main java org primefaces component slider slider java index a primefaces src main java org primefaces component slider slider java b primefaces src main java org primefaces component slider slider java public class slider extends sliderbase object value input getvalue if value null if value instanceof string return string value converter converter componentutils getconverter context input if converter null return converter getasstring context this value reproducer no response expected behavior no response primefaces edition community primefaces version theme no response jsf implementation mojarra jsf version x java version browser s no response
| 1
|
68,193
| 21,556,307,876
|
IssuesEvent
|
2022-04-30 13:30:36
|
ofalk/libdnet
|
https://api.github.com/repos/ofalk/libdnet
|
closed
|
OSError: Result too large
|
Priority-Medium Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
>>> import dnet
>>> dnet.eth("eth0")
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "dnet.pyx", line 112, in dnet.eth.__init__
OSError: Result too large
>>> dnet.eth("eth7")
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "dnet.pyx", line 112, in dnet.eth.__init__
OSError: Result too large
>>> dnet.eth(None)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "dnet.pyx", line 110, in dnet.eth.__init__
TypeError: expected string or Unicode object, NoneType found
What is the expected output? What do you see instead?
there is no way to specify that we want the type=6 (ethernet) interface. The
dnet.eth() method and the underlying eth_open() C function only take
the plain interface name as arguments.
What version of the product are you using? On what operating system?
Python 2.6.3
dnet 1.12 win32
Windows 7 / Vista
Please provide any additional information below.
>>> import dnet
>>> for iface in dnet.intf(): print iface
...
{'flags': 32, 'type': 1, 'name': 'eth0', 'mtu': 0L}
...
{'flags': 33, 'link_addr': 0a:aa:20:52:41:53, 'type': 6, 'name': 'eth0', 'mtu':
1500L}
...
{'addr': 192.168.1.99/24, 'mtu': 1500L, 'flags': 33, 'link_addr':
00:1f:3c:34:1c:0c, 'type': 71, 'name': 'eth0'}
...
```
Original issue reported on code.google.com by `fastian2...@gmail.com` on 31 Jan 2010 at 12:14
|
1.0
|
OSError: Result too large - ```
What steps will reproduce the problem?
>>> import dnet
>>> dnet.eth("eth0")
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "dnet.pyx", line 112, in dnet.eth.__init__
OSError: Result too large
>>> dnet.eth("eth7")
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "dnet.pyx", line 112, in dnet.eth.__init__
OSError: Result too large
>>> dnet.eth(None)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "dnet.pyx", line 110, in dnet.eth.__init__
TypeError: expected string or Unicode object, NoneType found
What is the expected output? What do you see instead?
there is no way to specify that we want the type=6 (ethernet) interface. The
dnet.eth() method and the underlying eth_open() C function only take
the plain interface name as arguments.
What version of the product are you using? On what operating system?
Python 2.6.3
dnet 1.12 win32
Windows 7 / Vista
Please provide any additional information below.
>>> import dnet
>>> for iface in dnet.intf(): print iface
...
{'flags': 32, 'type': 1, 'name': 'eth0', 'mtu': 0L}
...
{'flags': 33, 'link_addr': 0a:aa:20:52:41:53, 'type': 6, 'name': 'eth0', 'mtu':
1500L}
...
{'addr': 192.168.1.99/24, 'mtu': 1500L, 'flags': 33, 'link_addr':
00:1f:3c:34:1c:0c, 'type': 71, 'name': 'eth0'}
...
```
Original issue reported on code.google.com by `fastian2...@gmail.com` on 31 Jan 2010 at 12:14
|
defect
|
oserror result too large what steps will reproduce the problem import dnet dnet eth traceback most recent call last file line in file dnet pyx line in dnet eth init oserror result too large dnet eth traceback most recent call last file line in file dnet pyx line in dnet eth init oserror result too large dnet eth none traceback most recent call last file line in file dnet pyx line in dnet eth init typeerror expected string or unicode object nonetype found what is the expected output what do you see instead there is no way to specify that we want the type ethernet interface the dnet eth method and the underlying eth open c function only take the plain interface name as arguments what version of the product are you using on what operating system python dnet windows vista please provide any additional information below import dnet for iface in dnet intf print iface flags type name mtu flags link addr aa type name mtu addr mtu flags link addr type name original issue reported on code google com by gmail com on jan at
| 1
|
79,607
| 28,439,917,831
|
IssuesEvent
|
2023-04-15 19:35:17
|
thomasleplus/tinkerit
|
https://api.github.com/repos/thomasleplus/tinkerit
|
closed
|
Dmx Delay
|
Priority-Medium auto-migrated Type-Defect
|
```
What steps will reproduce the problem?
1. load the SerialToDMX sketch
2. Create a Pd Patch which sends Serial data through the object [comport 2
9600]
3. Try to send 16 serial messages every 70 ms in order to control 4 RGB lamps
(see attached patch)
What is the expected output?
changing the light synchronously with the variation of the values in the Pd
patch.
What do you see instead?
the lights change with a noticeable delay.
Please see attached Pd patch.
When I stop the data flow the light shouldn't change anymore.
What version of the product are you using? On what operating system?
OS: OSX 10.10.4
Pd-extended
thinker kit DMX master shield COM/T040060
Light source 4 x LA86LED
(http://www.amazon.co.uk/Lightahead®-DMX512-Colorful-Lighting-Projector/dp/B00U
VYK366)
Please provide any additional information below.
```
Original issue reported on code.google.com by `didonato...@gmail.com` on 6 Aug 2015 at 3:43
Attachments:
- [issue.pd](https://storage.googleapis.com/google-code-attachments/tinkerit/issue-26/comment-0/issue.pd)
|
1.0
|
Dmx Delay - ```
What steps will reproduce the problem?
1. load the SerialToDMX sketch
2. Create a Pd Patch which sends Serial data through the object [comport 2
9600]
3. Try to send 16 serial messages every 70 ms in order to control 4 RGB lamps
(see attached patch)
What is the expected output?
changing the light synchronously with the variation of the values in the Pd
patch.
What do you see instead?
the lights change with a noticeable delay.
Please see attached Pd patch.
When I stop the data flow the light shouldn't change anymore.
What version of the product are you using? On what operating system?
OS: OSX 10.10.4
Pd-extended
thinker kit DMX master shield COM/T040060
Light source 4 x LA86LED
(http://www.amazon.co.uk/Lightahead®-DMX512-Colorful-Lighting-Projector/dp/B00U
VYK366)
Please provide any additional information below.
```
Original issue reported on code.google.com by `didonato...@gmail.com` on 6 Aug 2015 at 3:43
Attachments:
- [issue.pd](https://storage.googleapis.com/google-code-attachments/tinkerit/issue-26/comment-0/issue.pd)
|
defect
|
dmx delay what steps will reproduce the problem load the serialtodmx sketch create a pd patch which sends serial data through the object comport try to send serial messages every ms in order to control rgb lamps see attached patch what is the expected output changing the light synchronously with the variation of the values in the pd patch what do you see instead the lights change with a noticeable delay please see attached pd patch when i stop the data flow the light shouldn t change anymore what version of the product are you using on what operating system os osx pd extended thinker kit dmx master shield com light source x please provide any additional information below original issue reported on code google com by didonato gmail com on aug at attachments
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.