Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
473,566
| 13,644,379,796
|
IssuesEvent
|
2020-09-25 18:47:31
|
microsoft/PowerToys
|
https://api.github.com/repos/microsoft/PowerToys
|
closed
|
PowerToys Run not working on first boot
|
Issue-Bug Priority-0
|
- PowerToys version: 0.20.1
- Running PowerToys as Admin: Yes
- Windows build number: 19041.388
When pc boot up the PowerToys Run does not work ( Pressing Alt + Space does nothing ). If i terminate the app and restart it then it works as intended. Issue occurred when i updated to 0.20.1 version. Previous version was fine.
|
1.0
|
PowerToys Run not working on first boot - - PowerToys version: 0.20.1
- Running PowerToys as Admin: Yes
- Windows build number: 19041.388
When pc boot up the PowerToys Run does not work ( Pressing Alt + Space does nothing ). If i terminate the app and restart it then it works as intended. Issue occurred when i updated to 0.20.1 version. Previous version was fine.
|
non_process
|
powertoys run not working on first boot powertoys version running powertoys as admin yes windows build number when pc boot up the powertoys run does not work pressing alt space does nothing if i terminate the app and restart it then it works as intended issue occurred when i updated to version previous version was fine
| 0
|
178,815
| 14,680,430,847
|
IssuesEvent
|
2020-12-31 10:03:06
|
RedHatInsights/insights-data-schemas
|
https://api.github.com/repos/RedHatInsights/insights-data-schemas
|
opened
|
Create a documentation for all implemented predicates.
|
documentation
|
Create a documentation for all implemented predicates.
|
1.0
|
Create a documentation for all implemented predicates. - Create a documentation for all implemented predicates.
|
non_process
|
create a documentation for all implemented predicates create a documentation for all implemented predicates
| 0
|
18,735
| 11,045,858,832
|
IssuesEvent
|
2019-12-09 15:50:37
|
cityofaustin/atd-mobility-project-database
|
https://api.github.com/repos/cityofaustin/atd-mobility-project-database
|
closed
|
Document feature-based roadmap
|
Epic Project: Mobility Project Database Service: Apps Workgroup: ATSD
|
As we speak with various users, we'll be collecting a list of potential MPD features. We'll use this down the line to derive user stories, prioritize, etc.
[Documented here](https://docs.google.com/spreadsheets/d/1gNCz63qFpztiuL_JK4aPNnicjCglmATOW1dbpU0wxWk/edit#gid=179231394).
|
1.0
|
Document feature-based roadmap - As we speak with various users, we'll be collecting a list of potential MPD features. We'll use this down the line to derive user stories, prioritize, etc.
[Documented here](https://docs.google.com/spreadsheets/d/1gNCz63qFpztiuL_JK4aPNnicjCglmATOW1dbpU0wxWk/edit#gid=179231394).
|
non_process
|
document feature based roadmap as we speak with various users we ll be collecting a list of potential mpd features we ll use this down the line to derive user stories prioritize etc
| 0
|
22,738
| 32,056,193,507
|
IssuesEvent
|
2023-09-24 05:19:29
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Need event metric like span metric
|
enhancement Stale processor/spanmetrics closed as inactive
|
### Is your feature request related to a problem? Please describe.
I have use span metric for some time. And I found in some other platform , it provide event metric. I think it's a good idea.
### Describe the solution you'd like
Add event metric in span metric processor. Maybe just calls_total is enough, and add some event attribute to the metric.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_
|
1.0
|
Need event metric like span metric - ### Is your feature request related to a problem? Please describe.
I have use span metric for some time. And I found in some other platform , it provide event metric. I think it's a good idea.
### Describe the solution you'd like
Add event metric in span metric processor. Maybe just calls_total is enough, and add some event attribute to the metric.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_
|
process
|
need event metric like span metric is your feature request related to a problem please describe i have use span metric for some time and i found in some other platform it provide event metric i think it s a good idea describe the solution you d like add event metric in span metric processor maybe just calls total is enough and add some event attribute to the metric describe alternatives you ve considered no response additional context no response
| 1
|
239,051
| 18,258,207,362
|
IssuesEvent
|
2021-10-03 11:59:23
|
GowthamGoush/Amazing_Sites
|
https://api.github.com/repos/GowthamGoush/Amazing_Sites
|
closed
|
replit
|
documentation hacktoberfest
|
Replit allows users to write code and build apps and websites using a browser. The site also has various collaborative features, including capability for real-time, multiuser editing with a live chat feed. It supports over 50 programming and markup languages, including Java, Python, and HTML, enabling users to build apps and websites. The site is integrated with GitHub, a code-hosting platform, providing a way to import and run projects from GitHub.
link : https://replit.com/
|
1.0
|
replit - Replit allows users to write code and build apps and websites using a browser. The site also has various collaborative features, including capability for real-time, multiuser editing with a live chat feed. It supports over 50 programming and markup languages, including Java, Python, and HTML, enabling users to build apps and websites. The site is integrated with GitHub, a code-hosting platform, providing a way to import and run projects from GitHub.
link : https://replit.com/
|
non_process
|
replit replit allows users to write code and build apps and websites using a browser the site also has various collaborative features including capability for real time multiuser editing with a live chat feed it supports over programming and markup languages including java python and html enabling users to build apps and websites the site is integrated with github a code hosting platform providing a way to import and run projects from github link
| 0
|
7,055
| 10,211,824,558
|
IssuesEvent
|
2019-08-14 17:54:36
|
googleapis/google-cloud-java
|
https://api.github.com/repos/googleapis/google-cloud-java
|
closed
|
Beta release of Cloud Dataproc
|
api: dataproc type: process
|
Package name: **google-cloud-dataproc**
Current release: **alpha**
Proposed release: **beta**
## Instructions
Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue.
## Required
- [x] Server API is beta or GA
- [x] Service API is public
- [x] Client surface is mostly stable (no known issues that could significantly change the surface)
- [x] All manual types and methods have comment documentation
- [x] Package name is idiomatic for the platform
- [x] At least one integration/smoke test is defined and passing
- [x] Central GitHub README lists and points to the per-API README
- [x] Per-API README links to product page on cloud.google.com
- [x] Manual code has been reviewed for API stability by repo owner
## Optional
- [ ] Most common / important scenarios have descriptive samples
- [ ] Public manual methods have at least one usage sample each (excluding overloads)
- [ ] Per-API README includes a full description of the API
- [ ] Per-API README contains at least one “getting started” sample using the most common API scenario
- [x] Manual code has been reviewed by API producer
- [x] Manual code has been reviewed by a DPE responsible for samples
- [x] 'Client LIbraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
|
1.0
|
Beta release of Cloud Dataproc - Package name: **google-cloud-dataproc**
Current release: **alpha**
Proposed release: **beta**
## Instructions
Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue.
## Required
- [x] Server API is beta or GA
- [x] Service API is public
- [x] Client surface is mostly stable (no known issues that could significantly change the surface)
- [x] All manual types and methods have comment documentation
- [x] Package name is idiomatic for the platform
- [x] At least one integration/smoke test is defined and passing
- [x] Central GitHub README lists and points to the per-API README
- [x] Per-API README links to product page on cloud.google.com
- [x] Manual code has been reviewed for API stability by repo owner
## Optional
- [ ] Most common / important scenarios have descriptive samples
- [ ] Public manual methods have at least one usage sample each (excluding overloads)
- [ ] Per-API README includes a full description of the API
- [ ] Per-API README contains at least one “getting started” sample using the most common API scenario
- [x] Manual code has been reviewed by API producer
- [x] Manual code has been reviewed by a DPE responsible for samples
- [x] 'Client LIbraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
|
process
|
beta release of cloud dataproc package name google cloud dataproc current release alpha proposed release beta instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required server api is beta or ga service api is public client surface is mostly stable no known issues that could significantly change the surface all manual types and methods have comment documentation package name is idiomatic for the platform at least one integration smoke test is defined and passing central github readme lists and points to the per api readme per api readme links to product page on cloud google com manual code has been reviewed for api stability by repo owner optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
| 1
|
334,347
| 24,414,742,690
|
IssuesEvent
|
2022-10-05 15:00:49
|
Archimonde666/Challenge-drone-EA
|
https://api.github.com/repos/Archimonde666/Challenge-drone-EA
|
closed
|
Simulator : Challenge 7 trap
|
documentation invalid Tips and advises
|
The checkpoint numbers are messed up in the challenge 7 (10 -> 12 -> 11 -> 13)
|
1.0
|
Simulator : Challenge 7 trap - The checkpoint numbers are messed up in the challenge 7 (10 -> 12 -> 11 -> 13)
|
non_process
|
simulator challenge trap the checkpoint numbers are messed up in the challenge
| 0
|
3,301
| 6,395,712,305
|
IssuesEvent
|
2017-08-04 13:55:36
|
tvorogme/gotosite
|
https://api.github.com/repos/tvorogme/gotosite
|
closed
|
[Backend] Add socials
|
backend in process wontfix
|
Socials:
1. Github
- [x] In profile add / del
- [x] Login
2. Gitlab
- [x] In profile add / del
- [x] Login
3. Vk
- [x] In profile add / del
- [x] Login
|
1.0
|
[Backend] Add socials - Socials:
1. Github
- [x] In profile add / del
- [x] Login
2. Gitlab
- [x] In profile add / del
- [x] Login
3. Vk
- [x] In profile add / del
- [x] Login
|
process
|
add socials socials github in profile add del login gitlab in profile add del login vk in profile add del login
| 1
|
14,797
| 18,074,493,927
|
IssuesEvent
|
2021-09-21 08:20:54
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] My account tab > Text change is required
|
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
Change the from **`My Account`** to **`My account`**

|
3.0
|
[PM] My account tab > Text change is required - Change the from **`My Account`** to **`My account`**

|
process
|
my account tab text change is required change the from my account to my account
| 1
|
57,853
| 3,084,070,473
|
IssuesEvent
|
2015-08-24 13:13:18
|
pavel-pimenov/flylinkdc-r5xx
|
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
|
closed
|
[zzxy] улучшение Настройки=>Дополнительно=>"Держать файл с изменениями SQLite в памяти. ..."
|
bug imported Priority-Medium wontfix
|
_From [zzzxzzzy...@gmail.com](https://code.google.com/u/111612712877897236331/) on August 01, 2013 17:30:32_
(502-betta92)
+ Настройки=>Дополнительно=>"Держать файл с изменениями SQLite в памяти. ...", а чтобы не повреждало интервально сейвить, раз в 2+ минут, при отс.обр.к диску и при нал.изменений
PS: а, правильнее - "интервально сейвить" запросы изменений, только.
Актуально когда сразу ОЧЕНЬ много обновлений базы, чтобы реже дёргать винт.
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1147_
|
1.0
|
[zzxy] улучшение Настройки=>Дополнительно=>"Держать файл с изменениями SQLite в памяти. ..." - _From [zzzxzzzy...@gmail.com](https://code.google.com/u/111612712877897236331/) on August 01, 2013 17:30:32_
(502-betta92)
+ Настройки=>Дополнительно=>"Держать файл с изменениями SQLite в памяти. ...", а чтобы не повреждало интервально сейвить, раз в 2+ минут, при отс.обр.к диску и при нал.изменений
PS: а, правильнее - "интервально сейвить" запросы изменений, только.
Актуально когда сразу ОЧЕНЬ много обновлений базы, чтобы реже дёргать винт.
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1147_
|
non_process
|
улучшение настройки дополнительно держать файл с изменениями sqlite в памяти from on august настройки дополнительно держать файл с изменениями sqlite в памяти а чтобы не повреждало интервально сейвить раз в минут при отс обр к диску и при нал изменений ps а правильнее интервально сейвить запросы изменений только актуально когда сразу очень много обновлений базы чтобы реже дёргать винт original issue
| 0
|
1,540
| 4,151,158,996
|
IssuesEvent
|
2016-06-15 19:41:24
|
kerubistan/kerub
|
https://api.github.com/repos/kerubistan/kerub
|
opened
|
add support for freebsd native iscsi target
|
component:data processing enhancement priority: normal
|
At the moment kerub only supports tgtd (available in most distros) but FreeBSD does not have a tgt package. Freebsd 10+ has a 'native' iscsi server.
https://www.freebsd.org/doc/handbook/network-iscsi.html
|
1.0
|
add support for freebsd native iscsi target - At the moment kerub only supports tgtd (available in most distros) but FreeBSD does not have a tgt package. Freebsd 10+ has a 'native' iscsi server.
https://www.freebsd.org/doc/handbook/network-iscsi.html
|
process
|
add support for freebsd native iscsi target at the moment kerub only supports tgtd available in most distros but freebsd does not have a tgt package freebsd has a native iscsi server
| 1
|
9,981
| 13,024,370,970
|
IssuesEvent
|
2020-07-27 11:43:36
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[FEATURE][processing] Allow drag and drop of model files onto QGIS window
to execute them
|
3.0 Automatic new feature Graphical modeler Processing
|
Original commit: https://github.com/qgis/QGIS/commit/6300464275914ab8bfef6dd5eb2617bc9c9ff26e by nyalldawson
Unfortunately this naughty coder did not write a description... :-(
|
1.0
|
[FEATURE][processing] Allow drag and drop of model files onto QGIS window
to execute them - Original commit: https://github.com/qgis/QGIS/commit/6300464275914ab8bfef6dd5eb2617bc9c9ff26e by nyalldawson
Unfortunately this naughty coder did not write a description... :-(
|
process
|
allow drag and drop of model files onto qgis window to execute them original commit by nyalldawson unfortunately this naughty coder did not write a description
| 1
|
17,562
| 23,376,405,795
|
IssuesEvent
|
2022-08-11 03:55:25
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Orthogonalize Processing Forces Close To Desktop
|
Processing Bug Crash/Data Corruption
|
### What is the bug or the crash?
When trying to `run` the Orthogonalize processing, the QGIS forces close without any warning or any report about crash :(
### Steps to reproduce the issue
1. I have vector building with total of 860 features manually digitized and some are slightly not squared.
2. Go to `Processing Toolbox`, search `Orthogonalize`, and choose the building layer as Input Layer and using default parameter.
3. QGIS Force Close
### Versions
3.26.0 Buenos Aires
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
When copying as Python command and running it, the console will look like this:
```
processing.run("native:orthogonalize", {'INPUT':'I:/RDTR/SIG/Temp/ALI 2/Bangunan.shp','ANGLE_TOLERANCE':15,'MAX_ITERATIONS':1000,'OUTPUT':'TEMPORARY_OUTPUT'})
Traceback (most recent call last):
File "C:\PROGRA~1\QGIS\apps\Python39\lib\code.py", line 90, in runcode
exec(code, self.locals)
File "<input>", line 1, in <module>
File "C:\PROGRA~1/QGIS/apps/qgis/./python/plugins\processing\tools\general.py", line 108, in run
return Processing.runAlgorithm(algOrName, parameters, onFinish, feedback, context)
File "C:\PROGRA~1/QGIS/apps/qgis/./python/plugins\processing\core\Processing.py", line 187, in runAlgorithm
ret, results = execute(alg, parameters, context, feedback, catch_exceptions=False)
File "C:\PROGRA~1/QGIS/apps/qgis/./python/plugins\processing\gui\AlgorithmExecutor.py", line 72, in execute
results, ok = alg.run(parameters, context, feedback, {}, False)
Exception: unknown
```
|
1.0
|
Orthogonalize Processing Forces Close To Desktop - ### What is the bug or the crash?
When trying to `run` the Orthogonalize processing, the QGIS forces close without any warning or any report about crash :(
### Steps to reproduce the issue
1. I have vector building with total of 860 features manually digitized and some are slightly not squared.
2. Go to `Processing Toolbox`, search `Orthogonalize`, and choose the building layer as Input Layer and using default parameter.
3. QGIS Force Close
### Versions
3.26.0 Buenos Aires
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
When copying as Python command and running it, the console will look like this:
```
processing.run("native:orthogonalize", {'INPUT':'I:/RDTR/SIG/Temp/ALI 2/Bangunan.shp','ANGLE_TOLERANCE':15,'MAX_ITERATIONS':1000,'OUTPUT':'TEMPORARY_OUTPUT'})
Traceback (most recent call last):
File "C:\PROGRA~1\QGIS\apps\Python39\lib\code.py", line 90, in runcode
exec(code, self.locals)
File "<input>", line 1, in <module>
File "C:\PROGRA~1/QGIS/apps/qgis/./python/plugins\processing\tools\general.py", line 108, in run
return Processing.runAlgorithm(algOrName, parameters, onFinish, feedback, context)
File "C:\PROGRA~1/QGIS/apps/qgis/./python/plugins\processing\core\Processing.py", line 187, in runAlgorithm
ret, results = execute(alg, parameters, context, feedback, catch_exceptions=False)
File "C:\PROGRA~1/QGIS/apps/qgis/./python/plugins\processing\gui\AlgorithmExecutor.py", line 72, in execute
results, ok = alg.run(parameters, context, feedback, {}, False)
Exception: unknown
```
|
process
|
orthogonalize processing forces close to desktop what is the bug or the crash when trying to run the orthogonalize processing the qgis forces close without any warning or any report about crash steps to reproduce the issue i have vector building with total of features manually digitized and some are slightly not squared go to processing toolbox search orthogonalize and choose the building layer as input layer and using default parameter qgis force close versions buenos aires supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context when copying as python command and running it the console will look like this processing run native orthogonalize input i rdtr sig temp ali bangunan shp angle tolerance max iterations output temporary output traceback most recent call last file c progra qgis apps lib code py line in runcode exec code self locals file line in file c progra qgis apps qgis python plugins processing tools general py line in run return processing runalgorithm algorname parameters onfinish feedback context file c progra qgis apps qgis python plugins processing core processing py line in runalgorithm ret results execute alg parameters context feedback catch exceptions false file c progra qgis apps qgis python plugins processing gui algorithmexecutor py line in execute results ok alg run parameters context feedback false exception unknown
| 1
|
492,256
| 14,199,270,134
|
IssuesEvent
|
2020-11-16 01:50:22
|
ita-social-projects/horondi_client_fe
|
https://api.github.com/repos/ita-social-projects/horondi_client_fe
|
closed
|
[Home] Categories slider from bags/accessories/backpacks displayed on home page
|
UI bug priority: high severity: minor
|
**Environment:** Windows X, Google Chrome: Version 86.0.4240.75 (Official Build) (64-bit)
**Reproducible:** always.
**Build found**: 23/Oct/20 | DB: 23
**Preconditions:**
Run HORONDI localhost
**Steps to reproduce:**
1. Go to HORONDI homepage (http://localhost:3000/)
2. Click on 'Bags' or 'Accessories' or 'Backpacks'
3. Click on HORONDI label
**Actual result:**
Categories slider is displayed on homepage (see slider-err.png)
**Expected result:**
Homepage displayed without categories slider
**Workaround:**
Refresh homepage.
User story and test case links
User story #2 As a user I want to see home page
Test case: ad hoc test

|
1.0
|
[Home] Categories slider from bags/accessories/backpacks displayed on home page - **Environment:** Windows X, Google Chrome: Version 86.0.4240.75 (Official Build) (64-bit)
**Reproducible:** always.
**Build found**: 23/Oct/20 | DB: 23
**Preconditions:**
Run HORONDI localhost
**Steps to reproduce:**
1. Go to HORONDI homepage (http://localhost:3000/)
2. Click on 'Bags' or 'Accessories' or 'Backpacks'
3. Click on HORONDI label
**Actual result:**
Categories slider is displayed on homepage (see slider-err.png)
**Expected result:**
Homepage displayed without categories slider
**Workaround:**
Refresh homepage.
User story and test case links
User story #2 As a user I want to see home page
Test case: ad hoc test

|
non_process
|
categories slider from bags accessories backpacks displayed on home page environment windows x google chrome version official build bit reproducible always build found oct db preconditions run horondi localhost steps to reproduce go to horondi homepage click on bags or accessories or backpacks click on horondi label actual result categories slider is displayed on homepage see slider err png expected result homepage displayed without categories slider workaround refresh homepage user story and test case links user story as a user i want to see home page test case ad hoc test
| 0
|
112,408
| 11,768,350,386
|
IssuesEvent
|
2020-03-15 09:53:28
|
SE701Group3/git-brunching
|
https://api.github.com/repos/SE701Group3/git-brunching
|
opened
|
Test case template
|
documentation testing
|
**User Story:**
As a.. I want.. so I can..
**Acceptance Criteria:**
* A/C
**Additional Context:**
|
1.0
|
Test case template - **User Story:**
As a.. I want.. so I can..
**Acceptance Criteria:**
* A/C
**Additional Context:**
|
non_process
|
test case template user story as a i want so i can acceptance criteria a c additional context
| 0
|
430,901
| 12,467,865,625
|
IssuesEvent
|
2020-05-28 17:49:26
|
theintern/intern
|
https://api.github.com/repos/theintern/intern
|
opened
|
Allow easy use of GitHub Action's CHROMEWEBDRIVER environment variable
|
effort-medium enhancement priority-medium repo-digdug
|
GitHub Actions virtual environments (linux and windows) expose a `CHROMEWEBDRIVER` environment variable that points to the local chromedriver.exe. Intern should be able to use this information to allow use of the existing webdriver vs downloading a new one.
This could be accomplished in a couple of ways:
1. Update SeleniumTunnel to understand CHROMEWEBDRIVER (and FIREFOXWEBDRIVER, etc) env vars
2. Update the config processing code to understand env vars (which would be more generally useful). This would also require updating dig dug to allow executables to be specified for drivers.
|
1.0
|
Allow easy use of GitHub Action's CHROMEWEBDRIVER environment variable - GitHub Actions virtual environments (linux and windows) expose a `CHROMEWEBDRIVER` environment variable that points to the local chromedriver.exe. Intern should be able to use this information to allow use of the existing webdriver vs downloading a new one.
This could be accomplished in a couple of ways:
1. Update SeleniumTunnel to understand CHROMEWEBDRIVER (and FIREFOXWEBDRIVER, etc) env vars
2. Update the config processing code to understand env vars (which would be more generally useful). This would also require updating dig dug to allow executables to be specified for drivers.
|
non_process
|
allow easy use of github action s chromewebdriver environment variable github actions virtual environments linux and windows expose a chromewebdriver environment variable that points to the local chromedriver exe intern should be able to use this information to allow use of the existing webdriver vs downloading a new one this could be accomplished in a couple of ways update seleniumtunnel to understand chromewebdriver and firefoxwebdriver etc env vars update the config processing code to understand env vars which would be more generally useful this would also require updating dig dug to allow executables to be specified for drivers
| 0
|
20,810
| 27,569,743,444
|
IssuesEvent
|
2023-03-08 08:16:39
|
cse442-at-ub/project_s23-iweatherify
|
https://api.github.com/repos/cse442-at-ub/project_s23-iweatherify
|
reopened
|
Integrate Weather API into the Homepage
|
Processing Task Sprint 2
|
**Tests**
*Test 1*
1) Visit the Website
2) See the Current Weather for Buffalo
3) Confirm it's the current weather by typing "current weather in buffalo" in google
[Screenshot 2023-03-08 at 3.13.27 AM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/b4cd664b-ea3e-4365-b604-6dc715f0fe49)
*Test 2*
1) In the input search bar type "Llanfairpwllgwyngyllgogerychwyrndrobwllllantysiliogogogoch"
2) Confirm that this div is scrollable
2) Confirm that the correct current Weather API Data is shown by searching "Llanfairpwllgwyngyllgogerychwyrndrobwllllantysiliogogogoch" weather in Google.

*Test 3*
1) In the input search bar type "Manhattan"
2) Confirm that the Weather API Data is Shown!

|
1.0
|
Integrate Weather API into the Homepage - **Tests**
*Test 1*
1) Visit the Website
2) See the Current Weather for Buffalo
3) Confirm it's the current weather by typing "current weather in buffalo" in google
[Screenshot 2023-03-08 at 3.13.27 AM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/b4cd664b-ea3e-4365-b604-6dc715f0fe49)
*Test 2*
1) In the input search bar type "Llanfairpwllgwyngyllgogerychwyrndrobwllllantysiliogogogoch"
2) Confirm that this div is scrollable
2) Confirm that the correct current Weather API Data is shown by searching "Llanfairpwllgwyngyllgogerychwyrndrobwllllantysiliogogogoch" weather in Google.

*Test 3*
1) In the input search bar type "Manhattan"
2) Confirm that the Weather API Data is Shown!

|
process
|
integrate weather api into the homepage tests test visit the website see the current weather for buffalo confirm it s the current weather by typing current weather in buffalo in google test in the input search bar type llanfairpwllgwyngyllgogerychwyrndrobwllllantysiliogogogoch confirm that this div is scrollable confirm that the correct current weather api data is shown by searching llanfairpwllgwyngyllgogerychwyrndrobwllllantysiliogogogoch weather in google test in the input search bar type manhattan confirm that the weather api data is shown
| 1
|
19,477
| 25,788,374,504
|
IssuesEvent
|
2022-12-09 23:27:33
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
closed
|
Replace Maven with Gradle in CI
|
enhancement process
|
### Problem
As a developer, I'd like to consolidate build tools to reduce dependabot PRs, reduce newcomer confusion, and take advantage of the benefits that Gradle provides over Maven.
### Solution
* [x] Update helm acceptance tests to use Gradle
* [x] Update Gradle to generate tgz for grpc, importer and rest
* [x] Move windows and Rosetta workflows to nightly
* [x] Remove separate module workflows in favor of consolidated Gradle matrix workflow
* [x] Replace use of Maven in CI with Gradle
* [x] Change Dockerfile to use build/ instead of target/ and update docker Gradle
* [x] Update Rosetta Dockerfile
* [x] Run Sonar in Gradle
* [x] Remove Maven from Dependabot
* [x] Update docs to reference Gradle
* [x] Remove pom.xml
* [ ] Create an infra ticket to have them update their scripts
### Alternatives
_No response_
|
1.0
|
Replace Maven with Gradle in CI - ### Problem
As a developer, I'd like to consolidate build tools to reduce dependabot PRs, reduce newcomer confusion, and take advantage of the benefits that Gradle provides over Maven.
### Solution
* [x] Update helm acceptance tests to use Gradle
* [x] Update Gradle to generate tgz for grpc, importer and rest
* [x] Move windows and Rosetta workflows to nightly
* [x] Remove separate module workflows in favor of consolidated Gradle matrix workflow
* [x] Replace use of Maven in CI with Gradle
* [x] Change Dockerfile to use build/ instead of target/ and update docker Gradle
* [x] Update Rosetta Dockerfile
* [x] Run Sonar in Gradle
* [x] Remove Maven from Dependabot
* [x] Update docs to reference Gradle
* [x] Remove pom.xml
* [ ] Create an infra ticket to have them update their scripts
### Alternatives
_No response_
|
process
|
replace maven with gradle in ci problem as a developer i d like to consolidate build tools to reduce dependabot prs reduce newcomer confusion and take advantage of the benefits that gradle provides over maven solution update helm acceptance tests to use gradle update gradle to generate tgz for grpc importer and rest move windows and rosetta workflows to nightly remove separate module workflows in favor of consolidated gradle matrix workflow replace use of maven in ci with gradle change dockerfile to use build instead of target and update docker gradle update rosetta dockerfile run sonar in gradle remove maven from dependabot update docs to reference gradle remove pom xml create an infra ticket to have them update their scripts alternatives no response
| 1
|
22,666
| 31,896,008,191
|
IssuesEvent
|
2023-09-18 01:48:55
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - dynamicProperties (fix typo `tragusLengthInMeters`
|
Term - change Class - Record-level non-normative Process - complete
|
## Term change
* Submitter: @peterdesmet
* Efficacy Justification (why is this change necessary?):
> **fixing typo** `tragusLengthInMeters` is very likely a typo and should be `targusLengthInMeters` (a common measurement for birds)
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): n/a
* Stability Justification (what concerns are there that this might affect existing implementations?): n/a
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: n/a
Current Term definition: https://dwc.tdwg.org/list/#dwc_dynamicProperties
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes):
* Term label (English, not normative):
* Organized in Class (e.g., Occurrence, Event, Location, Taxon):
* Definition of the term (normative):
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative):
> {"heightInMeters":1.5}, {"~tragusLengthInMeters~ **targusLengthInMeters**":0.014, "weightInGrams":120}, {"natureOfID":"expert identification", "identificationEvidence":"cytochrome B sequence"}, {"relativeHumidity":28, "airTemperatureInCelsius":22, "sampleSizeInKilograms":10}, {"aspectHeading":277, "slopeInDegrees":6}, {"iucnStatus":"vulnerable", "taxonDistribution":"Neuquén, Argentina"}
* Refines (identifier of the broader term this term refines; normative):
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative):
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative):
|
1.0
|
Change term - dynamicProperties (fix typo `tragusLengthInMeters` - ## Term change
* Submitter: @peterdesmet
* Efficacy Justification (why is this change necessary?):
> **fixing typo** `tragusLengthInMeters` is very likely a typo and should be `targusLengthInMeters` (a common measurement for birds)
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): n/a
* Stability Justification (what concerns are there that this might affect existing implementations?): n/a
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: n/a
Current Term definition: https://dwc.tdwg.org/list/#dwc_dynamicProperties
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes):
* Term label (English, not normative):
* Organized in Class (e.g., Occurrence, Event, Location, Taxon):
* Definition of the term (normative):
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative):
> {"heightInMeters":1.5}, {"~tragusLengthInMeters~ **targusLengthInMeters**":0.014, "weightInGrams":120}, {"natureOfID":"expert identification", "identificationEvidence":"cytochrome B sequence"}, {"relativeHumidity":28, "airTemperatureInCelsius":22, "sampleSizeInKilograms":10}, {"aspectHeading":277, "slopeInDegrees":6}, {"iucnStatus":"vulnerable", "taxonDistribution":"Neuquén, Argentina"}
* Refines (identifier of the broader term this term refines; normative):
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative):
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative):
|
process
|
change term dynamicproperties fix typo traguslengthinmeters term change submitter peterdesmet efficacy justification why is this change necessary fixing typo traguslengthinmeters is very likely a typo and should be targuslengthinmeters a common measurement for birds demand justification if the change is semantic in nature name at least two organizations that independently need this term n a stability justification what concerns are there that this might affect existing implementations n a implications for dwciri namespace does this change affect a dwciri term version n a current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes term label english not normative organized in class e g occurrence event location taxon definition of the term normative usage comments recommendations regarding content etc not normative examples not normative heightinmeters traguslengthinmeters targuslengthinmeters weightingrams natureofid expert identification identificationevidence cytochrome b sequence relativehumidity airtemperatureincelsius samplesizeinkilograms aspectheading slopeindegrees iucnstatus vulnerable taxondistribution neuquén argentina refines identifier of the broader term this term refines normative replaces identifier of the existing term that would be deprecated and replaced by this term normative abcd xpath of the equivalent term in abcd or efg not normative
| 1
|
3,450
| 6,541,833,869
|
IssuesEvent
|
2017-09-01 22:11:13
|
zero-os/0-Disk
|
https://api.github.com/repos/zero-os/0-Disk
|
closed
|
Use protobuf instead of capnp for tlog
|
process_wontfix type_feature
|
Based on the benchmarking done for https://github.com/zero-os/0-stor/issues/199 it seems very attractive to use protobuf instead of capnp for tlog.
|
1.0
|
Use protobuf instead of capnp for tlog - Based on the benchmarking done for https://github.com/zero-os/0-stor/issues/199 it seems very attractive to use protobuf instead of capnp for tlog.
|
process
|
use protobuf instead of capnp for tlog based on the benchmarking done for it seems very attractive to use protobuf instead of capnp for tlog
| 1
|
720,422
| 24,792,164,768
|
IssuesEvent
|
2022-10-24 14:30:26
|
AY2223S1-CS2103T-T17-4/tp
|
https://api.github.com/repos/AY2223S1-CS2103T-T17-4/tp
|
closed
|
AppointmentList Class
|
type.Enhancement priority.Medium
|
Person class could have an AppointmentList class rather than List containing appointment objects so that we can implement methods for them
|
1.0
|
AppointmentList Class - Person class could have an AppointmentList class rather than List containing appointment objects so that we can implement methods for them
|
non_process
|
appointmentlist class person class could have an appointmentlist class rather than list containing appointment objects so that we can implement methods for them
| 0
|
6,956
| 10,113,960,824
|
IssuesEvent
|
2019-07-30 18:00:44
|
material-components/material-components-ios
|
https://api.github.com/repos/material-components/material-components-ios
|
closed
|
[schemes/Color] Umbrella bug for tracking internal client migrations to new color scheme APIs
|
[schemes] type:Process
|
This is an internal issue. If you are a Googler, please visit [b/112284592](http://b/112284592) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/112284592](http://b/112284592)
|
1.0
|
[schemes/Color] Umbrella bug for tracking internal client migrations to new color scheme APIs - This is an internal issue. If you are a Googler, please visit [b/112284592](http://b/112284592) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/112284592](http://b/112284592)
|
process
|
umbrella bug for tracking internal client migrations to new color scheme apis this is an internal issue if you are a googler please visit for more details internal data associated internal bug
| 1
|
135,814
| 18,722,119,338
|
IssuesEvent
|
2021-11-03 12:59:10
|
KDWSS/dd-trace-java
|
https://api.github.com/repos/KDWSS/dd-trace-java
|
opened
|
CVE-2020-36186 (High) detected in multiple libraries
|
security vulnerability
|
## CVE-2020-36186 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.10.jar</b>, <b>jackson-databind-2.6.6.jar</b>, <b>jackson-databind-2.9.8.jar</b>, <b>jackson-databind-2.7.1.jar</b>, <b>jackson-databind-2.8.4.jar</b>, <b>jackson-databind-2.3.3.jar</b>, <b>jackson-databind-2.8.11.jar</b>, <b>jackson-databind-2.9.1.jar</b>, <b>jackson-databind-2.6.5.jar</b>, <b>jackson-databind-2.7.9.1.jar</b>, <b>jackson-databind-2.8.11.1.jar</b>, <b>jackson-databind-2.9.4.jar</b>, <b>jackson-databind-2.8.11.2.jar</b>, <b>jackson-databind-2.5.3.jar</b>, <b>jackson-databind-2.8.11.3.jar</b>, <b>jackson-databind-2.3.2.jar</b>, <b>jackson-databind-2.9.9.3.jar</b>, <b>jackson-databind-2.8.5.jar</b>, <b>jackson-databind-2.7.4.jar</b>, <b>jackson-databind-2.9.10.1.jar</b>, <b>jackson-databind-2.8.9.jar</b>, <b>jackson-databind-2.5.4.jar</b>, <b>jackson-databind-2.8.7.jar</b>, <b>jackson-databind-2.7.8.jar</b>, <b>jackson-databind-2.6.4.jar</b>, <b>jackson-databind-2.9.10.6.jar</b>, <b>jackson-databind-2.8.11.4.jar</b>, <b>jackson-databind-2.9.9.jar</b>, <b>jackson-databind-2.7.9.3.jar</b>, <b>jackson-databind-2.8.3.jar</b>, <b>jackson-databind-2.9.7.jar</b>, <b>jackson-databind-2.7.5.jar</b>, <b>jackson-databind-2.9.0.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.10.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/jax-rs-annotations-2/jax-rs-annotations-2.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar</p>
<p>
Dependency Hierarchy:
- dropwizard-testing-0.8.0.jar (Root Library)
- dropwizard-core-0.8.0.jar
- dropwizard-configuration-0.8.0.jar
- jackson-dataformat-yaml-2.5.1.jar
- :x: **jackson-databind-2.9.10.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/aws-java-sqs-1.0/aws-java-sqs-1.0.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.6/5108dde6049374ba980b360e1ecff49847baba4a/jackson-databind-2.6.6.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.6/5108dde6049374ba980b360e1ecff49847baba4a/jackson-databind-2.6.6.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-kinesis-1.11.106.jar (Root Library)
- aws-java-sdk-core-1.11.106.jar
- :x: **jackson-databind-2.6.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/elasticsearch/transport-5.3/transport-5.3.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.8/11283f21cc480aa86c4df7a0a3243ec508372ed2/jackson-databind-2.9.8.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.8/11283f21cc480aa86c4df7a0a3243ec508372ed2/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- spring-data-elasticsearch-3.0.14.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/play-2.4/play-2.4.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.1/14d88822bca655de7aa6ed3e4c498d115505710a/jackson-databind-2.7.1.jar</p>
<p>
Dependency Hierarchy:
- play-java_2.11-2.5.0.jar (Root Library)
- play_2.11-2.5.0.jar
- jackson-datatype-jdk8-2.7.1.jar
- :x: **jackson-databind-2.7.1.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/spring-rabbit/spring-rabbit.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.4/1c36c81e79cacdf48116afba8495e3393d267ba1/jackson-databind-2.8.4.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.4/1c36c81e79cacdf48116afba8495e3393d267ba1/jackson-databind-2.8.4.jar</p>
<p>
Dependency Hierarchy:
- spring-rabbit-2.0.0.RELEASE.jar (Root Library)
- http-client-1.3.0.RELEASE.jar
- :x: **jackson-databind-2.8.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.3.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/jax-rs-annotations-1/jax-rs-annotations-1.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.3.3/63b77400b5f1cf83a81823562c48d3120ef5518e/jackson-databind-2.3.3.jar</p>
<p>
Dependency Hierarchy:
- dropwizard-testing-0.7.1.jar (Root Library)
- dropwizard-core-0.7.1.jar
- dropwizard-jackson-0.7.1.jar
- :x: **jackson-databind-2.3.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/akka-http-10.0/akka-http-10.0.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11/569a9f220273024523799dba9dd358121b0ee09/jackson-databind-2.8.11.jar</p>
<p>
Dependency Hierarchy:
- lagom-javadsl-testkit_2.11-1.4.0.jar (Root Library)
- lagom-persistence-core_2.11-1.4.0.jar
- play_2.11-2.6.11.jar
- jjwt-0.7.0.jar
- :x: **jackson-databind-2.8.11.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/elasticsearch/transport-5.3/transport-5.3.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.1/716da1830a2043f18882fc036ec26eb32cbe5aff/jackson-databind-2.9.1.jar</p>
<p>
Dependency Hierarchy:
- spring-data-elasticsearch-3.0.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.9.1.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/elasticsearch/transport-2/transport-2.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar</p>
<p>
Dependency Hierarchy:
- spring-data-elasticsearch-2.0.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.6.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.9.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/appsec/weblog/weblog-spring-app/weblog-spring-app.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9.1/85343e40e4f68d4a25226d53736646abaf0ae039/jackson-databind-2.7.9.1.jar,/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9.1/85343e40e4f68d4a25226d53736646abaf0ae039/jackson-databind-2.7.9.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.9.1.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/benchmark-integration/play-perftest/play-perftest.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.1/341edc63fdd8b44e17b2c36abbc9b451d8fd05a5/jackson-databind-2.8.11.1.jar</p>
<p>
Dependency Hierarchy:
- play_2.12-2.6.20.jar (Root Library)
- jjwt-0.7.0.jar
- :x: **jackson-databind-2.8.11.1.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/spring-webflux-5/spring-webflux-5.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.4/498bbc3b94f566982c7f7c6d4d303fce365529be/jackson-databind-2.9.4.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-webflux-2.0.0.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.0.0.RELEASE.jar
- :x: **jackson-databind-2.9.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/spring-webmvc-3.1/spring-webmvc-3.1.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.2/2c5051e8e84d2c16316b758ebf746f9e90bef5a4/jackson-databind-2.8.11.2.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.2/2c5051e8e84d2c16316b758ebf746f9e90bef5a4/jackson-databind-2.8.11.2.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.17.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.11.2.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.5.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/aws-java-sdk-1.11.0/aws-java-sdk-1.11.0.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.3/c37875ff66127d93e5f672708cb2dcc14c8232ab/jackson-databind-2.5.3.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.3/c37875ff66127d93e5f672708cb2dcc14c8232ab/jackson-databind-2.5.3.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.3/c37875ff66127d93e5f672708cb2dcc14c8232ab/jackson-databind-2.5.3.jar</p>
<p>
Dependency Hierarchy:
- play_2.11-2.4.0.jar (Root Library)
- :x: **jackson-databind-2.5.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-smoke-tests/springboot-grpc/springboot-grpc.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.3/844df5aba5a1a56e00905b165b12bb34116ee858/jackson-databind-2.8.11.3.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.3/844df5aba5a1a56e00905b165b12bb34116ee858/jackson-databind-2.8.11.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.18.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.3.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/dropwizard/dropwizard-views/dropwizard-views.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.3.2/c75edc740a6d8cb1cef6fa82fa594e0bce561916/jackson-databind-2.3.2.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.3.2/c75edc740a6d8cb1cef6fa82fa594e0bce561916/jackson-databind-2.3.2.jar</p>
<p>
Dependency Hierarchy:
- play-java-ws_2.11-2.3.10.jar (Root Library)
- play_2.11-2.3.10.jar
- :x: **jackson-databind-2.3.2.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.9.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-trace-core/dd-trace-core.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar</p>
<p>
Dependency Hierarchy:
- jackson-dataformat-msgpack-0.8.20.jar (Root Library)
- :x: **jackson-databind-2.9.9.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/kafka-streams-0.11/kafka-streams-0.11.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.5/b3035f37e674c04dafe36a660c3815cc59f764e2/jackson-databind-2.8.5.jar</p>
<p>
Dependency Hierarchy:
- kafka-streams-0.11.0.0.jar (Root Library)
- connect-json-0.11.0.0.jar
- :x: **jackson-databind-2.8.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/vertx-web-3.4/vertx-web-3.4.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.4/1e9c6f3659644aeac84872c3b62d8e363bf4c96d/jackson-databind-2.7.4.jar</p>
<p>
Dependency Hierarchy:
- vertx-web-3.4.0.jar (Root Library)
- vertx-core-3.4.0.jar
- :x: **jackson-databind-2.7.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.10.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/finatra-2.9/finatra-2.9.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.9.10.1.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10.1/cadb5d93d16151946468cd1a2e96f41bc7d4e198/jackson-databind-2.9.10.1.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10.1/cadb5d93d16151946468cd1a2e96f41bc7d4e198/jackson-databind-2.9.10.1.jar,/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10.1/cadb5d93d16151946468cd1a2e96f41bc7d4e198/jackson-databind-2.9.10.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.10.1.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/play-2.6/play-2.6.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.9/4dfca3975be3c1a98eacb829e70f02e9a71bc159/jackson-databind-2.8.9.jar</p>
<p>
Dependency Hierarchy:
- play_2.11-2.6.0.jar (Root Library)
- jjwt-0.7.0.jar
- :x: **jackson-databind-2.8.9.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.5.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-smoke-tests/play-2.4/play-2.4.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.4/5dfa42af84584b4a862ea488da84bbbebbb06c35/jackson-databind-2.5.4.jar</p>
<p>
Dependency Hierarchy:
- play_2.11-2.4.11.jar (Root Library)
- jackson-datatype-jsr310-2.5.4.jar
- :x: **jackson-databind-2.5.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.7.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/twilio/twilio.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.7/6c3257ef458ac58a8da69a6dca3d2a15286d88c8/jackson-databind-2.8.7.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.7/6c3257ef458ac58a8da69a6dca3d2a15286d88c8/jackson-databind-2.8.7.jar</p>
<p>
Dependency Hierarchy:
- ratpack-core-1.5.0.jar (Root Library)
- :x: **jackson-databind-2.8.7.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-smoke-tests/play-2.5/play-2.5.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.8/9bc551426f1e19b4e2d87bb4bb2e19f8ecf8d578/jackson-databind-2.7.8.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.8/9bc551426f1e19b4e2d87bb4bb2e19f8ecf8d578/jackson-databind-2.7.8.jar</p>
<p>
Dependency Hierarchy:
- play_2.11-2.5.19.jar (Root Library)
- :x: **jackson-databind-2.7.8.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/couchbase-2.0/couchbase-2.0.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.4/f2abadd10891512268b16a1a1a6f81890f3e2976/jackson-databind-2.6.4.jar</p>
<p>
Dependency Hierarchy:
- spring-data-couchbase-2.0.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.6.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.10.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/couchbase-2.6/couchbase-2.6.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10.6/fbe40c0535b836082be7e3f8cac79275b9c8ff4a/jackson-databind-2.9.10.6.jar</p>
<p>
Dependency Hierarchy:
- encryption-2.0.1.jar (Root Library)
- :x: **jackson-databind-2.9.10.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-smoke-tests/play-2.6/play-2.6.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.4/596d6923ff4cf7ea72ded3ac32903b9c618ce9f1/jackson-databind-2.8.11.4.jar</p>
<p>
Dependency Hierarchy:
- play_2.12-2.6.25.jar (Root Library)
- jjwt-0.7.0.jar
- :x: **jackson-databind-2.8.11.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/mule-4/mule-4.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9/d6eb9817d9c7289a91f043ac5ee02a6b3cc86238/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- mule-module-launcher-4.2.2.jar (Root Library)
- mule-module-deployment-4.2.2.jar
- mule-module-extensions-xml-support-4.2.2.jar
- mule-metadata-model-catalog-1.2.2.jar
- mule-metadata-model-raml-1.2.2.jar
- raml-parser-2-1.0.40.jar
- json-schema-validator-2.2.10.jar
- json-schema-core-1.2.10.jar
- jackson-coreutils-1.9.jar
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.9.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/datastax-cassandra-3/datastax-cassandra-3.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9.3/fc6d8373d2f5a012473c764c3556704be6da15e/jackson-databind-2.7.9.3.jar</p>
<p>
Dependency Hierarchy:
- cassandra-driver-core-3.11.0.jar (Root Library)
- :x: **jackson-databind-2.7.9.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-smoke-tests/log-injection/log-injection.gradle</p>
<p>Path to vulnerable library: /caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.3/cea3788c72271d45676ce32c0665991674b24cc5/jackson-databind-2.8.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.7.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/couchbase-2.6/couchbase-2.6.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.7/e6faad47abd3179666e89068485a1b88a195ceb7/jackson-databind-2.9.7.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.7/e6faad47abd3179666e89068485a1b88a195ceb7/jackson-databind-2.9.7.jar</p>
<p>
Dependency Hierarchy:
- encryption-1.0.0.jar (Root Library)
- :x: **jackson-databind-2.9.7.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/spring-cloud-zuul-2/spring-cloud-zuul-2.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.5/ca7084021d9f213003eafe2583d1783d3d6a3685/jackson-databind-2.7.5.jar</p>
<p>
Dependency Hierarchy:
- zuul-core-1.3.1.jar (Root Library)
- archaius-core-0.7.6.jar
- :x: **jackson-databind-2.7.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/vertx-rx-3.5/vertx-rx-3.5.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.0/14fb5f088cc0b0dc90a73ba745bcade4961a3ee3/jackson-databind-2.9.0.jar</p>
<p>
Dependency Hierarchy:
- vertx-rx-java2-3.5.0.jar (Root Library)
- vertx-core-3.5.0.jar
- :x: **jackson-databind-2.9.0.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/KDWSS/dd-trace-java/commit/2819174635979a19573ec0ce8e3e2b63a3848079">2819174635979a19573ec0ce8e3e2b63a3848079</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.PerUserPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36186>CVE-2020-36186</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2997">https://github.com/FasterXML/jackson-databind/issues/2997</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10","packageFilePaths":["/dd-java-agent/instrumentation/jax-rs-annotations-2/jax-rs-annotations-2.gradle","/dd-java-agent/instrumentation/mule-4/mule-4.gradle","/dd-java-agent/instrumentation/dropwizard/dropwizard.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.dropwizard:dropwizard-testing:0.8.0;io.dropwizard:dropwizard-core:0.8.0;io.dropwizard:dropwizard-configuration:0.8.0;com.fasterxml.jackson.dataformat:jackson-dataformat-yaml:2.5.1;com.fasterxml.jackson.core:jackson-databind:2.9.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.6","packageFilePaths":["/dd-java-agent/instrumentation/aws-java-sqs-1.0/aws-java-sqs-1.0.gradle","/dd-java-agent/instrumentation/aws-java-sdk-1.11.0/aws-java-sdk-1.11.0.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.amazonaws:aws-java-sdk-kinesis:1.11.106;com.amazonaws:aws-java-sdk-core:1.11.106;com.fasterxml.jackson.core:jackson-databind:2.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/dd-java-agent/instrumentation/elasticsearch/transport-5.3/transport-5.3.gradle","/dd-java-agent/instrumentation/spring-webflux-5/spring-webflux-5.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.data:spring-data-elasticsearch:3.0.14.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.1","packageFilePaths":["/dd-java-agent/instrumentation/play-2.4/play-2.4.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play-java_2.11:2.5.0;com.typesafe.play:play_2.11:2.5.0;com.fasterxml.jackson.datatype:jackson-datatype-jdk8:2.7.1;com.fasterxml.jackson.core:jackson-databind:2.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.4","packageFilePaths":["/dd-java-agent/instrumentation/spring-rabbit/spring-rabbit.gradle","/dd-java-agent/instrumentation/finatra-2.9/finatra-2.9.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.amqp:spring-rabbit:2.0.0.RELEASE;com.rabbitmq:http-client:1.3.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.3.3","packageFilePaths":["/dd-java-agent/instrumentation/jax-rs-annotations-1/jax-rs-annotations-1.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.dropwizard:dropwizard-testing:0.7.1;io.dropwizard:dropwizard-core:0.7.1;io.dropwizard:dropwizard-jackson:0.7.1;com.fasterxml.jackson.core:jackson-databind:2.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11","packageFilePaths":["/dd-java-agent/instrumentation/akka-http-10.0/akka-http-10.0.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.lightbend.lagom:lagom-javadsl-testkit_2.11:1.4.0;com.lightbend.lagom:lagom-persistence-core_2.11:1.4.0;com.typesafe.play:play_2.11:2.6.11;io.jsonwebtoken:jjwt:0.7.0;com.fasterxml.jackson.core:jackson-databind:2.8.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.1","packageFilePaths":["/dd-java-agent/instrumentation/elasticsearch/transport-5.3/transport-5.3.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.data:spring-data-elasticsearch:3.0.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.5","packageFilePaths":["/dd-java-agent/instrumentation/elasticsearch/transport-2/transport-2.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.data:spring-data-elasticsearch:2.0.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.6.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.9.1","packageFilePaths":["/dd-java-agent/appsec/weblog/weblog-spring-app/weblog-spring-app.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.9.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.1","packageFilePaths":["/dd-java-agent/benchmark-integration/play-perftest/play-perftest.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.12:2.6.20;io.jsonwebtoken:jjwt:0.7.0;com.fasterxml.jackson.core:jackson-databind:2.8.11.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.4","packageFilePaths":["/dd-java-agent/instrumentation/spring-webflux-5/spring-webflux-5.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-webflux:2.0.0.RELEASE;org.springframework.boot:spring-boot-starter-json:2.0.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.2","packageFilePaths":["/dd-java-agent/instrumentation/spring-webmvc-3.1/spring-webmvc-3.1.gradle","/dd-java-agent/instrumentation/elasticsearch/transport-2/transport-2.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.5.17.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.11.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.5.3","packageFilePaths":["/dd-java-agent/instrumentation/aws-java-sdk-1.11.0/aws-java-sdk-1.11.0.gradle","/dd-java-agent/instrumentation/aws-java-sqs-1.0/aws-java-sqs-1.0.gradle","/dd-java-agent/instrumentation/play-2.4/play-2.4.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.11:2.4.0;com.fasterxml.jackson.core:jackson-databind:2.5.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.3","packageFilePaths":["/dd-smoke-tests/springboot-grpc/springboot-grpc.gradle","/dd-smoke-tests/springboot/springboot.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.5.18.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.11.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.3.2","packageFilePaths":["/dd-java-agent/instrumentation/dropwizard/dropwizard-views/dropwizard-views.gradle","/dd-java-agent/instrumentation/play-2.3/play-2.3.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play-java-ws_2.11:2.3.10;com.typesafe.play:play_2.11:2.3.10;com.fasterxml.jackson.core:jackson-databind:2.3.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9.3","packageFilePaths":["/dd-trace-core/dd-trace-core.gradle","/communication/communication.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.msgpack:jackson-dataformat-msgpack:0.8.20;com.fasterxml.jackson.core:jackson-databind:2.9.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.5","packageFilePaths":["/dd-java-agent/instrumentation/kafka-streams-0.11/kafka-streams-0.11.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.apache.kafka:kafka-streams:0.11.0.0;org.apache.kafka:connect-json:0.11.0.0;com.fasterxml.jackson.core:jackson-databind:2.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.4","packageFilePaths":["/dd-java-agent/instrumentation/vertx-web-3.4/vertx-web-3.4.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.vertx:vertx-web:3.4.0;io.vertx:vertx-core:3.4.0;com.fasterxml.jackson.core:jackson-databind:2.7.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.1","packageFilePaths":["/dd-java-agent/instrumentation/finatra-2.9/finatra-2.9.gradle","/dd-java-agent/instrumentation/play-2.6/play-2.6.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.9","packageFilePaths":["/dd-java-agent/instrumentation/play-2.6/play-2.6.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.11:2.6.0;io.jsonwebtoken:jjwt:0.7.0;com.fasterxml.jackson.core:jackson-databind:2.8.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.5.4","packageFilePaths":["/dd-smoke-tests/play-2.4/play-2.4.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.11:2.4.11;com.fasterxml.jackson.datatype:jackson-datatype-jsr310:2.5.4;com.fasterxml.jackson.core:jackson-databind:2.5.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.7","packageFilePaths":["/dd-java-agent/instrumentation/twilio/twilio.gradle","/dd-java-agent/instrumentation/ratpack-1.5/ratpack-1.5.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.ratpack:ratpack-core:1.5.0;com.fasterxml.jackson.core:jackson-databind:2.8.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.8","packageFilePaths":["/dd-smoke-tests/play-2.5/play-2.5.gradle","/dd-java-agent/instrumentation/play-2.4/play-2.4.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.11:2.5.19;com.fasterxml.jackson.core:jackson-databind:2.7.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.4","packageFilePaths":["/dd-java-agent/instrumentation/couchbase-2.0/couchbase-2.0.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.data:spring-data-couchbase:2.0.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.6.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.6","packageFilePaths":["/dd-java-agent/instrumentation/couchbase-2.6/couchbase-2.6.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.couchbase.client:encryption:2.0.1;com.fasterxml.jackson.core:jackson-databind:2.9.10.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.4","packageFilePaths":["/dd-smoke-tests/play-2.6/play-2.6.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.12:2.6.25;io.jsonwebtoken:jjwt:0.7.0;com.fasterxml.jackson.core:jackson-databind:2.8.11.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9","packageFilePaths":["/dd-java-agent/instrumentation/mule-4/mule-4.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.mule.runtime:mule-module-launcher:4.2.2;org.mule.runtime:mule-module-deployment:4.2.2;org.mule.runtime:mule-module-extensions-xml-support:4.2.2;org.mule.runtime:mule-metadata-model-catalog:1.2.2;org.mule.runtime:mule-metadata-model-raml:1.2.2;org.raml:raml-parser-2:1.0.40;com.github.java-json-tools:json-schema-validator:2.2.10;com.github.java-json-tools:json-schema-core:1.2.10;com.github.java-json-tools:jackson-coreutils:1.9;com.fasterxml.jackson.core:jackson-databind:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.9.3","packageFilePaths":["/dd-java-agent/instrumentation/datastax-cassandra-3/datastax-cassandra-3.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.datastax.cassandra:cassandra-driver-core:3.11.0;com.fasterxml.jackson.core:jackson-databind:2.7.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.3","packageFilePaths":["/dd-smoke-tests/log-injection/log-injection.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.7","packageFilePaths":["/dd-java-agent/instrumentation/couchbase-2.6/couchbase-2.6.gradle","/dd-java-agent/instrumentation/aws-java-sdk-2.2/aws-java-sdk-2.2.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.couchbase.client:encryption:1.0.0;com.fasterxml.jackson.core:jackson-databind:2.9.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.5","packageFilePaths":["/dd-java-agent/instrumentation/spring-cloud-zuul-2/spring-cloud-zuul-2.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.netflix.zuul:zuul-core:1.3.1;com.netflix.archaius:archaius-core:0.7.6;com.fasterxml.jackson.core:jackson-databind:2.7.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.0","packageFilePaths":["/dd-java-agent/instrumentation/vertx-rx-3.5/vertx-rx-3.5.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.vertx:vertx-rx-java2:3.5.0;io.vertx:vertx-core:3.5.0;com.fasterxml.jackson.core:jackson-databind:2.9.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36186","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.PerUserPoolDataSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36186","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-36186 (High) detected in multiple libraries - ## CVE-2020-36186 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.10.jar</b>, <b>jackson-databind-2.6.6.jar</b>, <b>jackson-databind-2.9.8.jar</b>, <b>jackson-databind-2.7.1.jar</b>, <b>jackson-databind-2.8.4.jar</b>, <b>jackson-databind-2.3.3.jar</b>, <b>jackson-databind-2.8.11.jar</b>, <b>jackson-databind-2.9.1.jar</b>, <b>jackson-databind-2.6.5.jar</b>, <b>jackson-databind-2.7.9.1.jar</b>, <b>jackson-databind-2.8.11.1.jar</b>, <b>jackson-databind-2.9.4.jar</b>, <b>jackson-databind-2.8.11.2.jar</b>, <b>jackson-databind-2.5.3.jar</b>, <b>jackson-databind-2.8.11.3.jar</b>, <b>jackson-databind-2.3.2.jar</b>, <b>jackson-databind-2.9.9.3.jar</b>, <b>jackson-databind-2.8.5.jar</b>, <b>jackson-databind-2.7.4.jar</b>, <b>jackson-databind-2.9.10.1.jar</b>, <b>jackson-databind-2.8.9.jar</b>, <b>jackson-databind-2.5.4.jar</b>, <b>jackson-databind-2.8.7.jar</b>, <b>jackson-databind-2.7.8.jar</b>, <b>jackson-databind-2.6.4.jar</b>, <b>jackson-databind-2.9.10.6.jar</b>, <b>jackson-databind-2.8.11.4.jar</b>, <b>jackson-databind-2.9.9.jar</b>, <b>jackson-databind-2.7.9.3.jar</b>, <b>jackson-databind-2.8.3.jar</b>, <b>jackson-databind-2.9.7.jar</b>, <b>jackson-databind-2.7.5.jar</b>, <b>jackson-databind-2.9.0.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.10.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/jax-rs-annotations-2/jax-rs-annotations-2.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10/e201bb70b7469ba18dd58ed8268aa44e702fa2f0/jackson-databind-2.9.10.jar</p>
<p>
Dependency Hierarchy:
- dropwizard-testing-0.8.0.jar (Root Library)
- dropwizard-core-0.8.0.jar
- dropwizard-configuration-0.8.0.jar
- jackson-dataformat-yaml-2.5.1.jar
- :x: **jackson-databind-2.9.10.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/aws-java-sqs-1.0/aws-java-sqs-1.0.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.6/5108dde6049374ba980b360e1ecff49847baba4a/jackson-databind-2.6.6.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.6/5108dde6049374ba980b360e1ecff49847baba4a/jackson-databind-2.6.6.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-kinesis-1.11.106.jar (Root Library)
- aws-java-sdk-core-1.11.106.jar
- :x: **jackson-databind-2.6.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/elasticsearch/transport-5.3/transport-5.3.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.8/11283f21cc480aa86c4df7a0a3243ec508372ed2/jackson-databind-2.9.8.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.8/11283f21cc480aa86c4df7a0a3243ec508372ed2/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- spring-data-elasticsearch-3.0.14.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/play-2.4/play-2.4.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.1/14d88822bca655de7aa6ed3e4c498d115505710a/jackson-databind-2.7.1.jar</p>
<p>
Dependency Hierarchy:
- play-java_2.11-2.5.0.jar (Root Library)
- play_2.11-2.5.0.jar
- jackson-datatype-jdk8-2.7.1.jar
- :x: **jackson-databind-2.7.1.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/spring-rabbit/spring-rabbit.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.4/1c36c81e79cacdf48116afba8495e3393d267ba1/jackson-databind-2.8.4.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.4/1c36c81e79cacdf48116afba8495e3393d267ba1/jackson-databind-2.8.4.jar</p>
<p>
Dependency Hierarchy:
- spring-rabbit-2.0.0.RELEASE.jar (Root Library)
- http-client-1.3.0.RELEASE.jar
- :x: **jackson-databind-2.8.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.3.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/jax-rs-annotations-1/jax-rs-annotations-1.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.3.3/63b77400b5f1cf83a81823562c48d3120ef5518e/jackson-databind-2.3.3.jar</p>
<p>
Dependency Hierarchy:
- dropwizard-testing-0.7.1.jar (Root Library)
- dropwizard-core-0.7.1.jar
- dropwizard-jackson-0.7.1.jar
- :x: **jackson-databind-2.3.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/akka-http-10.0/akka-http-10.0.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11/569a9f220273024523799dba9dd358121b0ee09/jackson-databind-2.8.11.jar</p>
<p>
Dependency Hierarchy:
- lagom-javadsl-testkit_2.11-1.4.0.jar (Root Library)
- lagom-persistence-core_2.11-1.4.0.jar
- play_2.11-2.6.11.jar
- jjwt-0.7.0.jar
- :x: **jackson-databind-2.8.11.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/elasticsearch/transport-5.3/transport-5.3.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.1/716da1830a2043f18882fc036ec26eb32cbe5aff/jackson-databind-2.9.1.jar</p>
<p>
Dependency Hierarchy:
- spring-data-elasticsearch-3.0.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.9.1.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/elasticsearch/transport-2/transport-2.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar</p>
<p>
Dependency Hierarchy:
- spring-data-elasticsearch-2.0.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.6.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.9.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/appsec/weblog/weblog-spring-app/weblog-spring-app.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9.1/85343e40e4f68d4a25226d53736646abaf0ae039/jackson-databind-2.7.9.1.jar,/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9.1/85343e40e4f68d4a25226d53736646abaf0ae039/jackson-databind-2.7.9.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.9.1.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/benchmark-integration/play-perftest/play-perftest.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.1/341edc63fdd8b44e17b2c36abbc9b451d8fd05a5/jackson-databind-2.8.11.1.jar</p>
<p>
Dependency Hierarchy:
- play_2.12-2.6.20.jar (Root Library)
- jjwt-0.7.0.jar
- :x: **jackson-databind-2.8.11.1.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/spring-webflux-5/spring-webflux-5.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.4/498bbc3b94f566982c7f7c6d4d303fce365529be/jackson-databind-2.9.4.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-webflux-2.0.0.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.0.0.RELEASE.jar
- :x: **jackson-databind-2.9.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/spring-webmvc-3.1/spring-webmvc-3.1.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.2/2c5051e8e84d2c16316b758ebf746f9e90bef5a4/jackson-databind-2.8.11.2.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.2/2c5051e8e84d2c16316b758ebf746f9e90bef5a4/jackson-databind-2.8.11.2.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.17.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.11.2.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.5.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/aws-java-sdk-1.11.0/aws-java-sdk-1.11.0.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.3/c37875ff66127d93e5f672708cb2dcc14c8232ab/jackson-databind-2.5.3.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.3/c37875ff66127d93e5f672708cb2dcc14c8232ab/jackson-databind-2.5.3.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.3/c37875ff66127d93e5f672708cb2dcc14c8232ab/jackson-databind-2.5.3.jar</p>
<p>
Dependency Hierarchy:
- play_2.11-2.4.0.jar (Root Library)
- :x: **jackson-databind-2.5.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-smoke-tests/springboot-grpc/springboot-grpc.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.3/844df5aba5a1a56e00905b165b12bb34116ee858/jackson-databind-2.8.11.3.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.3/844df5aba5a1a56e00905b165b12bb34116ee858/jackson-databind-2.8.11.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.18.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.3.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/dropwizard/dropwizard-views/dropwizard-views.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.3.2/c75edc740a6d8cb1cef6fa82fa594e0bce561916/jackson-databind-2.3.2.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.3.2/c75edc740a6d8cb1cef6fa82fa594e0bce561916/jackson-databind-2.3.2.jar</p>
<p>
Dependency Hierarchy:
- play-java-ws_2.11-2.3.10.jar (Root Library)
- play_2.11-2.3.10.jar
- :x: **jackson-databind-2.3.2.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.9.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-trace-core/dd-trace-core.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar</p>
<p>
Dependency Hierarchy:
- jackson-dataformat-msgpack-0.8.20.jar (Root Library)
- :x: **jackson-databind-2.9.9.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/kafka-streams-0.11/kafka-streams-0.11.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.5/b3035f37e674c04dafe36a660c3815cc59f764e2/jackson-databind-2.8.5.jar</p>
<p>
Dependency Hierarchy:
- kafka-streams-0.11.0.0.jar (Root Library)
- connect-json-0.11.0.0.jar
- :x: **jackson-databind-2.8.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/vertx-web-3.4/vertx-web-3.4.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.4/1e9c6f3659644aeac84872c3b62d8e363bf4c96d/jackson-databind-2.7.4.jar</p>
<p>
Dependency Hierarchy:
- vertx-web-3.4.0.jar (Root Library)
- vertx-core-3.4.0.jar
- :x: **jackson-databind-2.7.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.10.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/finatra-2.9/finatra-2.9.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.9.10.1.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10.1/cadb5d93d16151946468cd1a2e96f41bc7d4e198/jackson-databind-2.9.10.1.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10.1/cadb5d93d16151946468cd1a2e96f41bc7d4e198/jackson-databind-2.9.10.1.jar,/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10.1/cadb5d93d16151946468cd1a2e96f41bc7d4e198/jackson-databind-2.9.10.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.10.1.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/play-2.6/play-2.6.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.9/4dfca3975be3c1a98eacb829e70f02e9a71bc159/jackson-databind-2.8.9.jar</p>
<p>
Dependency Hierarchy:
- play_2.11-2.6.0.jar (Root Library)
- jjwt-0.7.0.jar
- :x: **jackson-databind-2.8.9.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.5.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-smoke-tests/play-2.4/play-2.4.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.4/5dfa42af84584b4a862ea488da84bbbebbb06c35/jackson-databind-2.5.4.jar</p>
<p>
Dependency Hierarchy:
- play_2.11-2.4.11.jar (Root Library)
- jackson-datatype-jsr310-2.5.4.jar
- :x: **jackson-databind-2.5.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.7.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/twilio/twilio.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.7/6c3257ef458ac58a8da69a6dca3d2a15286d88c8/jackson-databind-2.8.7.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.7/6c3257ef458ac58a8da69a6dca3d2a15286d88c8/jackson-databind-2.8.7.jar</p>
<p>
Dependency Hierarchy:
- ratpack-core-1.5.0.jar (Root Library)
- :x: **jackson-databind-2.8.7.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-smoke-tests/play-2.5/play-2.5.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.8/9bc551426f1e19b4e2d87bb4bb2e19f8ecf8d578/jackson-databind-2.7.8.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.8/9bc551426f1e19b4e2d87bb4bb2e19f8ecf8d578/jackson-databind-2.7.8.jar</p>
<p>
Dependency Hierarchy:
- play_2.11-2.5.19.jar (Root Library)
- :x: **jackson-databind-2.7.8.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/couchbase-2.0/couchbase-2.0.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.4/f2abadd10891512268b16a1a1a6f81890f3e2976/jackson-databind-2.6.4.jar</p>
<p>
Dependency Hierarchy:
- spring-data-couchbase-2.0.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.6.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.10.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/couchbase-2.6/couchbase-2.6.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.10.6/fbe40c0535b836082be7e3f8cac79275b9c8ff4a/jackson-databind-2.9.10.6.jar</p>
<p>
Dependency Hierarchy:
- encryption-2.0.1.jar (Root Library)
- :x: **jackson-databind-2.9.10.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-smoke-tests/play-2.6/play-2.6.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.4/596d6923ff4cf7ea72ded3ac32903b9c618ce9f1/jackson-databind-2.8.11.4.jar</p>
<p>
Dependency Hierarchy:
- play_2.12-2.6.25.jar (Root Library)
- jjwt-0.7.0.jar
- :x: **jackson-databind-2.8.11.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/mule-4/mule-4.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9/d6eb9817d9c7289a91f043ac5ee02a6b3cc86238/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- mule-module-launcher-4.2.2.jar (Root Library)
- mule-module-deployment-4.2.2.jar
- mule-module-extensions-xml-support-4.2.2.jar
- mule-metadata-model-catalog-1.2.2.jar
- mule-metadata-model-raml-1.2.2.jar
- raml-parser-2-1.0.40.jar
- json-schema-validator-2.2.10.jar
- json-schema-core-1.2.10.jar
- jackson-coreutils-1.9.jar
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.9.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/datastax-cassandra-3/datastax-cassandra-3.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9.3/fc6d8373d2f5a012473c764c3556704be6da15e/jackson-databind-2.7.9.3.jar</p>
<p>
Dependency Hierarchy:
- cassandra-driver-core-3.11.0.jar (Root Library)
- :x: **jackson-databind-2.7.9.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-smoke-tests/log-injection/log-injection.gradle</p>
<p>Path to vulnerable library: /caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.3/cea3788c72271d45676ce32c0665991674b24cc5/jackson-databind-2.8.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.7.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/couchbase-2.6/couchbase-2.6.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.7/e6faad47abd3179666e89068485a1b88a195ceb7/jackson-databind-2.9.7.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.7/e6faad47abd3179666e89068485a1b88a195ceb7/jackson-databind-2.9.7.jar</p>
<p>
Dependency Hierarchy:
- encryption-1.0.0.jar (Root Library)
- :x: **jackson-databind-2.9.7.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.7.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/spring-cloud-zuul-2/spring-cloud-zuul-2.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.5/ca7084021d9f213003eafe2583d1783d3d6a3685/jackson-databind-2.7.5.jar</p>
<p>
Dependency Hierarchy:
- zuul-core-1.3.1.jar (Root Library)
- archaius-core-0.7.6.jar
- :x: **jackson-databind-2.7.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: dd-trace-java/dd-java-agent/instrumentation/vertx-rx-3.5/vertx-rx-3.5.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.0/14fb5f088cc0b0dc90a73ba745bcade4961a3ee3/jackson-databind-2.9.0.jar</p>
<p>
Dependency Hierarchy:
- vertx-rx-java2-3.5.0.jar (Root Library)
- vertx-core-3.5.0.jar
- :x: **jackson-databind-2.9.0.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/KDWSS/dd-trace-java/commit/2819174635979a19573ec0ce8e3e2b63a3848079">2819174635979a19573ec0ce8e3e2b63a3848079</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.PerUserPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36186>CVE-2020-36186</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2997">https://github.com/FasterXML/jackson-databind/issues/2997</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10","packageFilePaths":["/dd-java-agent/instrumentation/jax-rs-annotations-2/jax-rs-annotations-2.gradle","/dd-java-agent/instrumentation/mule-4/mule-4.gradle","/dd-java-agent/instrumentation/dropwizard/dropwizard.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.dropwizard:dropwizard-testing:0.8.0;io.dropwizard:dropwizard-core:0.8.0;io.dropwizard:dropwizard-configuration:0.8.0;com.fasterxml.jackson.dataformat:jackson-dataformat-yaml:2.5.1;com.fasterxml.jackson.core:jackson-databind:2.9.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.6","packageFilePaths":["/dd-java-agent/instrumentation/aws-java-sqs-1.0/aws-java-sqs-1.0.gradle","/dd-java-agent/instrumentation/aws-java-sdk-1.11.0/aws-java-sdk-1.11.0.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.amazonaws:aws-java-sdk-kinesis:1.11.106;com.amazonaws:aws-java-sdk-core:1.11.106;com.fasterxml.jackson.core:jackson-databind:2.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/dd-java-agent/instrumentation/elasticsearch/transport-5.3/transport-5.3.gradle","/dd-java-agent/instrumentation/spring-webflux-5/spring-webflux-5.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.data:spring-data-elasticsearch:3.0.14.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.1","packageFilePaths":["/dd-java-agent/instrumentation/play-2.4/play-2.4.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play-java_2.11:2.5.0;com.typesafe.play:play_2.11:2.5.0;com.fasterxml.jackson.datatype:jackson-datatype-jdk8:2.7.1;com.fasterxml.jackson.core:jackson-databind:2.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.4","packageFilePaths":["/dd-java-agent/instrumentation/spring-rabbit/spring-rabbit.gradle","/dd-java-agent/instrumentation/finatra-2.9/finatra-2.9.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.amqp:spring-rabbit:2.0.0.RELEASE;com.rabbitmq:http-client:1.3.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.3.3","packageFilePaths":["/dd-java-agent/instrumentation/jax-rs-annotations-1/jax-rs-annotations-1.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.dropwizard:dropwizard-testing:0.7.1;io.dropwizard:dropwizard-core:0.7.1;io.dropwizard:dropwizard-jackson:0.7.1;com.fasterxml.jackson.core:jackson-databind:2.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11","packageFilePaths":["/dd-java-agent/instrumentation/akka-http-10.0/akka-http-10.0.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.lightbend.lagom:lagom-javadsl-testkit_2.11:1.4.0;com.lightbend.lagom:lagom-persistence-core_2.11:1.4.0;com.typesafe.play:play_2.11:2.6.11;io.jsonwebtoken:jjwt:0.7.0;com.fasterxml.jackson.core:jackson-databind:2.8.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.1","packageFilePaths":["/dd-java-agent/instrumentation/elasticsearch/transport-5.3/transport-5.3.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.data:spring-data-elasticsearch:3.0.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.5","packageFilePaths":["/dd-java-agent/instrumentation/elasticsearch/transport-2/transport-2.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.data:spring-data-elasticsearch:2.0.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.6.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.9.1","packageFilePaths":["/dd-java-agent/appsec/weblog/weblog-spring-app/weblog-spring-app.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.9.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.1","packageFilePaths":["/dd-java-agent/benchmark-integration/play-perftest/play-perftest.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.12:2.6.20;io.jsonwebtoken:jjwt:0.7.0;com.fasterxml.jackson.core:jackson-databind:2.8.11.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.4","packageFilePaths":["/dd-java-agent/instrumentation/spring-webflux-5/spring-webflux-5.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-webflux:2.0.0.RELEASE;org.springframework.boot:spring-boot-starter-json:2.0.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.2","packageFilePaths":["/dd-java-agent/instrumentation/spring-webmvc-3.1/spring-webmvc-3.1.gradle","/dd-java-agent/instrumentation/elasticsearch/transport-2/transport-2.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.5.17.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.11.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.5.3","packageFilePaths":["/dd-java-agent/instrumentation/aws-java-sdk-1.11.0/aws-java-sdk-1.11.0.gradle","/dd-java-agent/instrumentation/aws-java-sqs-1.0/aws-java-sqs-1.0.gradle","/dd-java-agent/instrumentation/play-2.4/play-2.4.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.11:2.4.0;com.fasterxml.jackson.core:jackson-databind:2.5.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.3","packageFilePaths":["/dd-smoke-tests/springboot-grpc/springboot-grpc.gradle","/dd-smoke-tests/springboot/springboot.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.5.18.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.11.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.3.2","packageFilePaths":["/dd-java-agent/instrumentation/dropwizard/dropwizard-views/dropwizard-views.gradle","/dd-java-agent/instrumentation/play-2.3/play-2.3.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play-java-ws_2.11:2.3.10;com.typesafe.play:play_2.11:2.3.10;com.fasterxml.jackson.core:jackson-databind:2.3.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9.3","packageFilePaths":["/dd-trace-core/dd-trace-core.gradle","/communication/communication.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.msgpack:jackson-dataformat-msgpack:0.8.20;com.fasterxml.jackson.core:jackson-databind:2.9.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.5","packageFilePaths":["/dd-java-agent/instrumentation/kafka-streams-0.11/kafka-streams-0.11.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.apache.kafka:kafka-streams:0.11.0.0;org.apache.kafka:connect-json:0.11.0.0;com.fasterxml.jackson.core:jackson-databind:2.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.4","packageFilePaths":["/dd-java-agent/instrumentation/vertx-web-3.4/vertx-web-3.4.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.vertx:vertx-web:3.4.0;io.vertx:vertx-core:3.4.0;com.fasterxml.jackson.core:jackson-databind:2.7.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.1","packageFilePaths":["/dd-java-agent/instrumentation/finatra-2.9/finatra-2.9.gradle","/dd-java-agent/instrumentation/play-2.6/play-2.6.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.9","packageFilePaths":["/dd-java-agent/instrumentation/play-2.6/play-2.6.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.11:2.6.0;io.jsonwebtoken:jjwt:0.7.0;com.fasterxml.jackson.core:jackson-databind:2.8.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.5.4","packageFilePaths":["/dd-smoke-tests/play-2.4/play-2.4.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.11:2.4.11;com.fasterxml.jackson.datatype:jackson-datatype-jsr310:2.5.4;com.fasterxml.jackson.core:jackson-databind:2.5.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.7","packageFilePaths":["/dd-java-agent/instrumentation/twilio/twilio.gradle","/dd-java-agent/instrumentation/ratpack-1.5/ratpack-1.5.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.ratpack:ratpack-core:1.5.0;com.fasterxml.jackson.core:jackson-databind:2.8.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.8","packageFilePaths":["/dd-smoke-tests/play-2.5/play-2.5.gradle","/dd-java-agent/instrumentation/play-2.4/play-2.4.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.11:2.5.19;com.fasterxml.jackson.core:jackson-databind:2.7.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.4","packageFilePaths":["/dd-java-agent/instrumentation/couchbase-2.0/couchbase-2.0.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.data:spring-data-couchbase:2.0.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.6.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.6","packageFilePaths":["/dd-java-agent/instrumentation/couchbase-2.6/couchbase-2.6.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.couchbase.client:encryption:2.0.1;com.fasterxml.jackson.core:jackson-databind:2.9.10.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.4","packageFilePaths":["/dd-smoke-tests/play-2.6/play-2.6.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.typesafe.play:play_2.12:2.6.25;io.jsonwebtoken:jjwt:0.7.0;com.fasterxml.jackson.core:jackson-databind:2.8.11.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9","packageFilePaths":["/dd-java-agent/instrumentation/mule-4/mule-4.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.mule.runtime:mule-module-launcher:4.2.2;org.mule.runtime:mule-module-deployment:4.2.2;org.mule.runtime:mule-module-extensions-xml-support:4.2.2;org.mule.runtime:mule-metadata-model-catalog:1.2.2;org.mule.runtime:mule-metadata-model-raml:1.2.2;org.raml:raml-parser-2:1.0.40;com.github.java-json-tools:json-schema-validator:2.2.10;com.github.java-json-tools:json-schema-core:1.2.10;com.github.java-json-tools:jackson-coreutils:1.9;com.fasterxml.jackson.core:jackson-databind:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.9.3","packageFilePaths":["/dd-java-agent/instrumentation/datastax-cassandra-3/datastax-cassandra-3.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.datastax.cassandra:cassandra-driver-core:3.11.0;com.fasterxml.jackson.core:jackson-databind:2.7.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.3","packageFilePaths":["/dd-smoke-tests/log-injection/log-injection.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.7","packageFilePaths":["/dd-java-agent/instrumentation/couchbase-2.6/couchbase-2.6.gradle","/dd-java-agent/instrumentation/aws-java-sdk-2.2/aws-java-sdk-2.2.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.couchbase.client:encryption:1.0.0;com.fasterxml.jackson.core:jackson-databind:2.9.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.5","packageFilePaths":["/dd-java-agent/instrumentation/spring-cloud-zuul-2/spring-cloud-zuul-2.gradle"],"isTransitiveDependency":true,"dependencyTree":"com.netflix.zuul:zuul-core:1.3.1;com.netflix.archaius:archaius-core:0.7.6;com.fasterxml.jackson.core:jackson-databind:2.7.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.0","packageFilePaths":["/dd-java-agent/instrumentation/vertx-rx-3.5/vertx-rx-3.5.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.vertx:vertx-rx-java2:3.5.0;io.vertx:vertx-core:3.5.0;com.fasterxml.jackson.core:jackson-databind:2.9.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36186","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.PerUserPoolDataSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36186","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation jax rs annotations jax rs annotations gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy dropwizard testing jar root library dropwizard core jar dropwizard configuration jar jackson dataformat yaml jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation aws java sqs aws java sqs gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy aws java sdk kinesis jar root library aws java sdk core jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation elasticsearch transport transport gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring data elasticsearch release jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation play play gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy play java jar root library play jar jackson datatype jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation spring rabbit spring rabbit gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring rabbit release jar root library http client release jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to dependency file dd trace java dd java agent instrumentation jax rs annotations jax rs annotations gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy dropwizard testing jar root library dropwizard core jar dropwizard jackson jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation akka http akka http gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy lagom javadsl testkit jar root library lagom persistence core jar play jar jjwt jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation elasticsearch transport transport gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring data elasticsearch release jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation elasticsearch transport transport gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring data elasticsearch release jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent appsec weblog weblog spring app weblog spring app gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent benchmark integration play perftest play perftest gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy play jar root library jjwt jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation spring webflux spring webflux gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter webflux release jar root library spring boot starter json release jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation spring webmvc spring webmvc gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation aws java sdk aws java sdk gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy play jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd smoke tests springboot grpc springboot grpc gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to dependency file dd trace java dd java agent instrumentation dropwizard dropwizard views dropwizard views gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy play java ws jar root library play jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd trace core dd trace core gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy jackson dataformat msgpack jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation kafka streams kafka streams gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy kafka streams jar root library connect json jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation vertx web vertx web gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy vertx web jar root library vertx core jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation finatra finatra gradle path to vulnerable library home wss scanner cache com fasterxml jackson core jackson databind bundles jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation play play gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy play jar root library jjwt jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd smoke tests play play gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy play jar root library jackson datatype jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation twilio twilio gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy ratpack core jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd smoke tests play play gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy play jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation couchbase couchbase gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring data couchbase release jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation couchbase couchbase gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy encryption jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd smoke tests play play gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy play jar root library jjwt jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation mule mule gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy mule module launcher jar root library mule module deployment jar mule module extensions xml support jar mule metadata model catalog jar mule metadata model raml jar raml parser jar json schema validator jar json schema core jar jackson coreutils jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation datastax cassandra datastax cassandra gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy cassandra driver core jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd smoke tests log injection log injection gradle path to vulnerable library caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation couchbase couchbase gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy encryption jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation spring cloud zuul spring cloud zuul gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy zuul core jar root library archaius core jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file dd trace java dd java agent instrumentation vertx rx vertx rx gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy vertx rx jar root library vertx core jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree io dropwizard dropwizard testing io dropwizard dropwizard core io dropwizard dropwizard configuration com fasterxml jackson dataformat jackson dataformat yaml com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com amazonaws aws java sdk kinesis com amazonaws aws java sdk core com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework data spring data elasticsearch release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com typesafe play play java com typesafe play play com fasterxml jackson datatype jackson datatype com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework amqp spring rabbit release com rabbitmq http client release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree io dropwizard dropwizard testing io dropwizard dropwizard core io dropwizard dropwizard jackson com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com lightbend lagom lagom javadsl testkit com lightbend lagom lagom persistence core com typesafe play play io jsonwebtoken jjwt com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework data spring data elasticsearch release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework data spring data elasticsearch release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com typesafe play play io jsonwebtoken jjwt com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework boot spring boot starter webflux release org springframework boot spring boot starter json release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework boot spring boot starter web release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com typesafe play play com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework boot spring boot starter web release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com typesafe play play java ws com typesafe play play com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org msgpack jackson dataformat msgpack com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org apache kafka kafka streams org apache kafka connect json com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree io vertx vertx web io vertx vertx core com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com typesafe play play io jsonwebtoken jjwt com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com typesafe play play com fasterxml jackson datatype jackson datatype com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree io ratpack ratpack core com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com typesafe play play com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework data spring data couchbase release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com couchbase client encryption com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com typesafe play play io jsonwebtoken jjwt com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org mule runtime mule module launcher org mule runtime mule module deployment org mule runtime mule module extensions xml support org mule runtime mule metadata model catalog org mule runtime mule metadata model raml org raml raml parser com github java json tools json schema validator com github java json tools json schema core com github java json tools jackson coreutils com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com datastax cassandra cassandra driver core com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com couchbase client encryption com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree com netflix zuul zuul core com netflix archaius archaius core com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree io vertx vertx rx io vertx vertx core com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp datasources peruserpooldatasource vulnerabilityurl
| 0
|
520,184
| 15,081,429,427
|
IssuesEvent
|
2021-02-05 13:11:42
|
canonical-web-and-design/multipass.run
|
https://api.github.com/repos/canonical-web-and-design/multipass.run
|
closed
|
Add Ubuntu Pro tips Cheat sheet section/form
|
Priority: Medium
|
[copy doc](https://docs.google.com/document/d/1uI5hsuhe5VUpFSSiQas6FVB2F7RhlbNXcr5Pvr_Lm2c/edit#)
Add a form for users to download cheat sheet same as https://ubuntu.com/download/server#
HTML below for section of landing page
```<section class="p-strip--suru-accent">
<div class="row u-equal-height u-vertically-center">
<div class="col-6 u-align--center u-hide--small">
<img src="https://assets.ubuntu.com/v1/39a8dac8-Ubuntu_Server_CLI_pro_tips_2020-04.jpg" style="width: 350px;" alt="Ubuntu Server CLI pro tips"/>
</div>
<div class="col-6">
<p>The ”Ubuntu Server CLI cheat sheet“ is your fast path to learning the Linux command line - from basic file management to deploying Kubernetes and OpenStack.</p>
<a class="p-button--positive js-invoke-modal" href="#">Download Cheat Sheet</a>
</div>
</div>
</section>
|
1.0
|
Add Ubuntu Pro tips Cheat sheet section/form - [copy doc](https://docs.google.com/document/d/1uI5hsuhe5VUpFSSiQas6FVB2F7RhlbNXcr5Pvr_Lm2c/edit#)
Add a form for users to download cheat sheet same as https://ubuntu.com/download/server#
HTML below for section of landing page
```<section class="p-strip--suru-accent">
<div class="row u-equal-height u-vertically-center">
<div class="col-6 u-align--center u-hide--small">
<img src="https://assets.ubuntu.com/v1/39a8dac8-Ubuntu_Server_CLI_pro_tips_2020-04.jpg" style="width: 350px;" alt="Ubuntu Server CLI pro tips"/>
</div>
<div class="col-6">
<p>The ”Ubuntu Server CLI cheat sheet“ is your fast path to learning the Linux command line - from basic file management to deploying Kubernetes and OpenStack.</p>
<a class="p-button--positive js-invoke-modal" href="#">Download Cheat Sheet</a>
</div>
</div>
</section>
|
non_process
|
add ubuntu pro tips cheat sheet section form add a form for users to download cheat sheet same as html below for section of landing page the rdquo ubuntu server cli cheat sheet ldquo is your fast path to learning the linux command line from basic file management to deploying kubernetes and openstack download cheat sheet
| 0
|
9,944
| 12,976,196,690
|
IssuesEvent
|
2020-07-21 18:20:20
|
obinnaokechukwu/internship-2020
|
https://api.github.com/repos/obinnaokechukwu/internship-2020
|
opened
|
Perform final code review of changes before making pull request to OBS
|
process
|
Perform final code review of changes before making pull request to OBS.
|
1.0
|
Perform final code review of changes before making pull request to OBS - Perform final code review of changes before making pull request to OBS.
|
process
|
perform final code review of changes before making pull request to obs perform final code review of changes before making pull request to obs
| 1
|
12,158
| 14,741,472,193
|
IssuesEvent
|
2021-01-07 10:40:24
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
NCSM Advanced PICC Specialist Payment not showing
|
anc-process anp-important ant-bug ant-enhancement
|
In GitLab by @kdjstudios on Jan 22, 2019, 13:23
**Submitted by:** "Martin Villegas" <martin.villegas@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-01-22-81991
**Server:** Internal
**Client/Site:** NCSM
**Account:** Advanced PICC Specialist
**Issue:**
We posted a payment of $10,000 on 01/15/19 but the check bounced back due to NSF. I had added a stage fee for the same amount but Richard (CFO) told me I could enter a payment with a negative sign in front of it so that the amount would be added back. I did it and the amount was added back to the balance but I can’t see it under payment status or billing history. Was this the right thing to do or we should have waited for the stage fee to be posted?
Balance before adding the -$10,000 was 9344.18
Balance after adding the -$10,000 is 19344.18
|
1.0
|
NCSM Advanced PICC Specialist Payment not showing - In GitLab by @kdjstudios on Jan 22, 2019, 13:23
**Submitted by:** "Martin Villegas" <martin.villegas@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-01-22-81991
**Server:** Internal
**Client/Site:** NCSM
**Account:** Advanced PICC Specialist
**Issue:**
We posted a payment of $10,000 on 01/15/19 but the check bounced back due to NSF. I had added a stage fee for the same amount but Richard (CFO) told me I could enter a payment with a negative sign in front of it so that the amount would be added back. I did it and the amount was added back to the balance but I can’t see it under payment status or billing history. Was this the right thing to do or we should have waited for the stage fee to be posted?
Balance before adding the -$10,000 was 9344.18
Balance after adding the -$10,000 is 19344.18
|
process
|
ncsm advanced picc specialist payment not showing in gitlab by kdjstudios on jan submitted by martin villegas helpdesk server internal client site ncsm account advanced picc specialist issue we posted a payment of on but the check bounced back due to nsf i had added a stage fee for the same amount but richard cfo told me i could enter a payment with a negative sign in front of it so that the amount would be added back i did it and the amount was added back to the balance but i can’t see it under payment status or billing history was this the right thing to do or we should have waited for the stage fee to be posted balance before adding the was balance after adding the is
| 1
|
218,128
| 16,750,117,569
|
IssuesEvent
|
2021-06-11 21:28:54
|
ctc-oss/fapolicy-analyzer
|
https://api.github.com/repos/ctc-oss/fapolicy-analyzer
|
opened
|
Starting (and keeping) a changelog
|
documentation
|
We need to start keeping a changelog. This is a good reference
https://keepachangelog.com/en/1.0.0/
I like pattern [Argo](https://github.com/argoproj/argo-workflows) uses to generate the release notes linking back to the changelog.
The RPM spec also has a section for changelog where this could be included. Just need to ensure the formats match.
Will add some doc to the Wiki as well for this to provide some basic guidelines.
|
1.0
|
Starting (and keeping) a changelog - We need to start keeping a changelog. This is a good reference
https://keepachangelog.com/en/1.0.0/
I like pattern [Argo](https://github.com/argoproj/argo-workflows) uses to generate the release notes linking back to the changelog.
The RPM spec also has a section for changelog where this could be included. Just need to ensure the formats match.
Will add some doc to the Wiki as well for this to provide some basic guidelines.
|
non_process
|
starting and keeping a changelog we need to start keeping a changelog this is a good reference i like pattern uses to generate the release notes linking back to the changelog the rpm spec also has a section for changelog where this could be included just need to ensure the formats match will add some doc to the wiki as well for this to provide some basic guidelines
| 0
|
424,761
| 29,175,699,831
|
IssuesEvent
|
2023-05-19 07:43:54
|
imAsparky/django-cookiecutter
|
https://api.github.com/repos/imAsparky/django-cookiecutter
|
closed
|
[DOCS]: Fix broken docs build.
|
documentation
|
**Describe alternatives you've considered**
Would you please share your clear and concise description of any alternative/similar documentation you've considered.
Version dependency issues using Sphinx 7, Go back to latest version of Sphinx 6.
**Additional context**
Would you please share any other context or screenshots about the documentation request here.
|
1.0
|
[DOCS]: Fix broken docs build. - **Describe alternatives you've considered**
Would you please share your clear and concise description of any alternative/similar documentation you've considered.
Version dependency issues using Sphinx 7, Go back to latest version of Sphinx 6.
**Additional context**
Would you please share any other context or screenshots about the documentation request here.
|
non_process
|
fix broken docs build describe alternatives you ve considered would you please share your clear and concise description of any alternative similar documentation you ve considered version dependency issues using sphinx go back to latest version of sphinx additional context would you please share any other context or screenshots about the documentation request here
| 0
|
106,212
| 23,194,734,514
|
IssuesEvent
|
2022-08-01 15:23:27
|
nilearn/nilearn
|
https://api.github.com/repos/nilearn/nilearn
|
opened
|
Check for inconsistent naming of functions according to PEP8 convention
|
code quality Maintenance
|
As pointed out in https://github.com/nilearn/nilearn/pull/3300#discussion_r929071958 I think it is worth checking for inconsistencies in function naming in nilearn. It might also be a good idea to discuss any ambiguities regarding function naming.
|
1.0
|
Check for inconsistent naming of functions according to PEP8 convention - As pointed out in https://github.com/nilearn/nilearn/pull/3300#discussion_r929071958 I think it is worth checking for inconsistencies in function naming in nilearn. It might also be a good idea to discuss any ambiguities regarding function naming.
|
non_process
|
check for inconsistent naming of functions according to convention as pointed out in i think it is worth checking for inconsistencies in function naming in nilearn it might also be a good idea to discuss any ambiguities regarding function naming
| 0
|
20,741
| 27,440,637,663
|
IssuesEvent
|
2023-03-02 10:44:48
|
pystatgen/sgkit
|
https://api.github.com/repos/pystatgen/sgkit
|
closed
|
Sphinx build is failing with extension error (pydata_sphinx_theme)
|
process + tools upstream
|
E.g. from https://github.com/pystatgen/sgkit/actions/runs/4293787906/jobs/7481927063:
```
Running Sphinx v4.2.0
python exec: /opt/hostedtoolcache/Python/3.[9](https://github.com/pystatgen/sgkit/actions/runs/4293787906/jobs/7481927063#step:6:10).16/x64/bin/python
sys.path: ['/home/runner/work/sgkit/sgkit', '/opt/hostedtoolcache/Python/3.9.16/x64/bin', '/home/runner/work/sgkit/sgkit', '/opt/hostedtoolcache/Python/3.9.16/x64/lib/python39.zip', '/opt/hostedtoolcache/Python/3.9.16/x64/lib/python3.9', '/opt/hostedtoolcache/Python/3.9.16/x64/lib/python3.9/lib-dynload', '/opt/hostedtoolcache/Python/3.9.16/x64/lib/python3.9/site-packages']
making output directory... done
[autosummary] generating autosummary for: about.rst, api.rst, changelog.rst, contributing.rst, examples/gwas_tutorial.ipynb, examples/index.rst, getting_started.rst, how_do_i.rst, index.rst, news.rst, news/introducing_sgkit.md, user_guide.rst, vcf.rst
[autosummary] generating autosummary for: /home/runner/work/sgkit/sgkit/docs/generated/sgkit.Fst.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.Garud_H.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.Tajimas_D.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.Weir_Goudet_beta.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.call_allele_frequencies.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.cohort_allele_frequencies.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.convert_probability_to_call.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.count_call_alleles.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.count_cohort_alleles.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.count_variant_alleles.rst, ..., /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variables.variant_position_spec.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variables.variant_score_spec.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variables.window_contig_spec.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variables.window_start_spec.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variables.window_stop_spec.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variant_stats.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.window_by_genome.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.window_by_interval.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.window_by_position.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.window_by_variant.rst
loading intersphinx inventory from https://docs.dask.org/en/stable/objects.inv...
loading intersphinx inventory from https://xarray.pydata.org/en/stable/objects.inv...
loading intersphinx inventory from https://zarr.readthedocs.io/en/stable/objects.inv...
loading intersphinx inventory from https://numpy.org/doc/stable/objects.inv...
loading intersphinx inventory from https://docs.python.org/3/objects.inv...
loading intersphinx inventory from https://scikit-learn.org/stable/objects.inv...
intersphinx inventory has moved: https://xarray.pydata.org/en/stable/objects.inv -> https://docs.xarray.dev/en/stable/objects.inv
myst v0.18.1: MdParserConfig(commonmark_only=False, gfm_only=False, enable_extensions=[], disable_syntax=[], all_links_external=False, url_schemes=('http', 'https', 'mailto', 'ftp'), ref_domains=None, highlight_code_blocks=True, number_code_blocks=[], title_to_header=False, heading_anchors=None, heading_slug_func=None, footnote_transition=True, words_per_minute=200, sub_delimiters=('{', '}'), linkify_fuzzy_links=True, dmath_allow_labels=True, dmath_allow_space=True, dmath_allow_digits=True, dmath_double_inline=False, update_mathjax=True, mathjax_classes='tex2jax_process|mathjax_process|math|output_area')
myst-nb v0.17.1: NbParserConfig(custom_formats={}, metadata_key='mystnb', cell_metadata_key='mystnb', kernel_rgx_aliases={}, execution_mode='auto', execution_cache_path='', execution_excludepatterns=(), execution_timeout=30, execution_in_temp=False, execution_allow_errors=False, execution_raise_on_error=False, execution_show_tb=False, merge_streams=False, render_plugin='default', remove_code_source=False, remove_code_outputs=False, code_prompt_show='Show code cell {type}', code_prompt_hide='Hide code cell {type}', number_source_lines=False, output_stderr='show', render_text_lexer='myst-ansi', render_error_lexer='ipythontb', render_image_options={}, render_figure_options={}, render_markdown_format='commonmark', output_folder='build', append_css=True, metadata_to_fm=False)
Using jupyter-cache at: /home/runner/work/sgkit/sgkit/docs/_build/.jupyter_cache
Extension error (pydata_sphinx_theme):
Handler <function update_config at 0x7f89d55e5[16](https://github.com/pystatgen/sgkit/actions/runs/4293787906/jobs/7481927063#step:6:17)0> for event 'builder-inited' threw an exception (exception: 'icon_links')
make: *** [Makefile:[27](https://github.com/pystatgen/sgkit/actions/runs/4293787906/jobs/7481927063#step:6:28): html] Error 2
Error: Process completed with exit code 2.
```
Looks like it's due to https://github.com/pydata/pydata-sphinx-theme/releases/tag/v0.13.0
|
1.0
|
Sphinx build is failing with extension error (pydata_sphinx_theme) - E.g. from https://github.com/pystatgen/sgkit/actions/runs/4293787906/jobs/7481927063:
```
Running Sphinx v4.2.0
python exec: /opt/hostedtoolcache/Python/3.[9](https://github.com/pystatgen/sgkit/actions/runs/4293787906/jobs/7481927063#step:6:10).16/x64/bin/python
sys.path: ['/home/runner/work/sgkit/sgkit', '/opt/hostedtoolcache/Python/3.9.16/x64/bin', '/home/runner/work/sgkit/sgkit', '/opt/hostedtoolcache/Python/3.9.16/x64/lib/python39.zip', '/opt/hostedtoolcache/Python/3.9.16/x64/lib/python3.9', '/opt/hostedtoolcache/Python/3.9.16/x64/lib/python3.9/lib-dynload', '/opt/hostedtoolcache/Python/3.9.16/x64/lib/python3.9/site-packages']
making output directory... done
[autosummary] generating autosummary for: about.rst, api.rst, changelog.rst, contributing.rst, examples/gwas_tutorial.ipynb, examples/index.rst, getting_started.rst, how_do_i.rst, index.rst, news.rst, news/introducing_sgkit.md, user_guide.rst, vcf.rst
[autosummary] generating autosummary for: /home/runner/work/sgkit/sgkit/docs/generated/sgkit.Fst.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.Garud_H.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.Tajimas_D.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.Weir_Goudet_beta.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.call_allele_frequencies.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.cohort_allele_frequencies.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.convert_probability_to_call.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.count_call_alleles.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.count_cohort_alleles.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.count_variant_alleles.rst, ..., /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variables.variant_position_spec.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variables.variant_score_spec.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variables.window_contig_spec.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variables.window_start_spec.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variables.window_stop_spec.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.variant_stats.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.window_by_genome.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.window_by_interval.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.window_by_position.rst, /home/runner/work/sgkit/sgkit/docs/generated/sgkit.window_by_variant.rst
loading intersphinx inventory from https://docs.dask.org/en/stable/objects.inv...
loading intersphinx inventory from https://xarray.pydata.org/en/stable/objects.inv...
loading intersphinx inventory from https://zarr.readthedocs.io/en/stable/objects.inv...
loading intersphinx inventory from https://numpy.org/doc/stable/objects.inv...
loading intersphinx inventory from https://docs.python.org/3/objects.inv...
loading intersphinx inventory from https://scikit-learn.org/stable/objects.inv...
intersphinx inventory has moved: https://xarray.pydata.org/en/stable/objects.inv -> https://docs.xarray.dev/en/stable/objects.inv
myst v0.18.1: MdParserConfig(commonmark_only=False, gfm_only=False, enable_extensions=[], disable_syntax=[], all_links_external=False, url_schemes=('http', 'https', 'mailto', 'ftp'), ref_domains=None, highlight_code_blocks=True, number_code_blocks=[], title_to_header=False, heading_anchors=None, heading_slug_func=None, footnote_transition=True, words_per_minute=200, sub_delimiters=('{', '}'), linkify_fuzzy_links=True, dmath_allow_labels=True, dmath_allow_space=True, dmath_allow_digits=True, dmath_double_inline=False, update_mathjax=True, mathjax_classes='tex2jax_process|mathjax_process|math|output_area')
myst-nb v0.17.1: NbParserConfig(custom_formats={}, metadata_key='mystnb', cell_metadata_key='mystnb', kernel_rgx_aliases={}, execution_mode='auto', execution_cache_path='', execution_excludepatterns=(), execution_timeout=30, execution_in_temp=False, execution_allow_errors=False, execution_raise_on_error=False, execution_show_tb=False, merge_streams=False, render_plugin='default', remove_code_source=False, remove_code_outputs=False, code_prompt_show='Show code cell {type}', code_prompt_hide='Hide code cell {type}', number_source_lines=False, output_stderr='show', render_text_lexer='myst-ansi', render_error_lexer='ipythontb', render_image_options={}, render_figure_options={}, render_markdown_format='commonmark', output_folder='build', append_css=True, metadata_to_fm=False)
Using jupyter-cache at: /home/runner/work/sgkit/sgkit/docs/_build/.jupyter_cache
Extension error (pydata_sphinx_theme):
Handler <function update_config at 0x7f89d55e5[16](https://github.com/pystatgen/sgkit/actions/runs/4293787906/jobs/7481927063#step:6:17)0> for event 'builder-inited' threw an exception (exception: 'icon_links')
make: *** [Makefile:[27](https://github.com/pystatgen/sgkit/actions/runs/4293787906/jobs/7481927063#step:6:28): html] Error 2
Error: Process completed with exit code 2.
```
Looks like it's due to https://github.com/pydata/pydata-sphinx-theme/releases/tag/v0.13.0
|
process
|
sphinx build is failing with extension error pydata sphinx theme e g from running sphinx python exec opt hostedtoolcache python sys path making output directory done generating autosummary for about rst api rst changelog rst contributing rst examples gwas tutorial ipynb examples index rst getting started rst how do i rst index rst news rst news introducing sgkit md user guide rst vcf rst generating autosummary for home runner work sgkit sgkit docs generated sgkit fst rst home runner work sgkit sgkit docs generated sgkit garud h rst home runner work sgkit sgkit docs generated sgkit tajimas d rst home runner work sgkit sgkit docs generated sgkit weir goudet beta rst home runner work sgkit sgkit docs generated sgkit call allele frequencies rst home runner work sgkit sgkit docs generated sgkit cohort allele frequencies rst home runner work sgkit sgkit docs generated sgkit convert probability to call rst home runner work sgkit sgkit docs generated sgkit count call alleles rst home runner work sgkit sgkit docs generated sgkit count cohort alleles rst home runner work sgkit sgkit docs generated sgkit count variant alleles rst home runner work sgkit sgkit docs generated sgkit variables variant position spec rst home runner work sgkit sgkit docs generated sgkit variables variant score spec rst home runner work sgkit sgkit docs generated sgkit variables window contig spec rst home runner work sgkit sgkit docs generated sgkit variables window start spec rst home runner work sgkit sgkit docs generated sgkit variables window stop spec rst home runner work sgkit sgkit docs generated sgkit variant stats rst home runner work sgkit sgkit docs generated sgkit window by genome rst home runner work sgkit sgkit docs generated sgkit window by interval rst home runner work sgkit sgkit docs generated sgkit window by position rst home runner work sgkit sgkit docs generated sgkit window by variant rst loading intersphinx inventory from loading intersphinx inventory from loading intersphinx inventory from loading intersphinx inventory from loading intersphinx inventory from loading intersphinx inventory from intersphinx inventory has moved myst mdparserconfig commonmark only false gfm only false enable extensions disable syntax all links external false url schemes http https mailto ftp ref domains none highlight code blocks true number code blocks title to header false heading anchors none heading slug func none footnote transition true words per minute sub delimiters linkify fuzzy links true dmath allow labels true dmath allow space true dmath allow digits true dmath double inline false update mathjax true mathjax classes process mathjax process math output area myst nb nbparserconfig custom formats metadata key mystnb cell metadata key mystnb kernel rgx aliases execution mode auto execution cache path execution excludepatterns execution timeout execution in temp false execution allow errors false execution raise on error false execution show tb false merge streams false render plugin default remove code source false remove code outputs false code prompt show show code cell type code prompt hide hide code cell type number source lines false output stderr show render text lexer myst ansi render error lexer ipythontb render image options render figure options render markdown format commonmark output folder build append css true metadata to fm false using jupyter cache at home runner work sgkit sgkit docs build jupyter cache extension error pydata sphinx theme handler function update config at for event builder inited threw an exception exception icon links make html error error process completed with exit code looks like it s due to
| 1
|
10,576
| 13,386,268,850
|
IssuesEvent
|
2020-09-02 14:31:16
|
googleapis/python-bigquery-datatransfer
|
https://api.github.com/repos/googleapis/python-bigquery-datatransfer
|
opened
|
Transition the library to the new microgenerator
|
type: process
|
Microgenerator is ready and we can use it to regenerate the code here.
Doing that is a breaking change (e.g. drops Python 2.7, 3.5), meaning that it needs to be released in a new **major** version.
|
1.0
|
Transition the library to the new microgenerator - Microgenerator is ready and we can use it to regenerate the code here.
Doing that is a breaking change (e.g. drops Python 2.7, 3.5), meaning that it needs to be released in a new **major** version.
|
process
|
transition the library to the new microgenerator microgenerator is ready and we can use it to regenerate the code here doing that is a breaking change e g drops python meaning that it needs to be released in a new major version
| 1
|
4,488
| 6,996,090,513
|
IssuesEvent
|
2017-12-15 22:20:24
|
amcnulty/Whoop-It-Up
|
https://api.github.com/repos/amcnulty/Whoop-It-Up
|
closed
|
Insert new user into database
|
Requirement
|
Once a user is signed up their information needs to be stored into the database.
|
1.0
|
Insert new user into database - Once a user is signed up their information needs to be stored into the database.
|
non_process
|
insert new user into database once a user is signed up their information needs to be stored into the database
| 0
|
185,656
| 6,726,328,158
|
IssuesEvent
|
2017-10-17 09:29:31
|
metasfresh/metasfresh
|
https://api.github.com/repos/metasfresh/metasfresh
|
closed
|
Set correct identifiers in the table M_ShipmentSchedule (v2)
|
branch:master priority:high
|
### Is this a bug or feature request?
FR
Follow-up of https://github.com/metasfresh/metasfresh/issues/2700 where we set the table/record as identifiers.
Needed for https://github.com/metasfresh/metasfresh/issues/472
### What is the current behavior?
Shipment schedule string representation shows the table/record Id.
#### Which are the steps to reproduce?
* check subscription progress window.
* check the "Pick" process params (see https://github.com/metasfresh/metasfresh/issues/472)
### What is the expected or desired behavior?
Following fields ONLY shall be part of the identifier:
* bpartner
* bp location
* date promised
* order document no
* product
* QtyToDeliver
|
1.0
|
Set correct identifiers in the table M_ShipmentSchedule (v2) - ### Is this a bug or feature request?
FR
Follow-up of https://github.com/metasfresh/metasfresh/issues/2700 where we set the table/record as identifiers.
Needed for https://github.com/metasfresh/metasfresh/issues/472
### What is the current behavior?
Shipment schedule string representation shows the table/record Id.
#### Which are the steps to reproduce?
* check subscription progress window.
* check the "Pick" process params (see https://github.com/metasfresh/metasfresh/issues/472)
### What is the expected or desired behavior?
Following fields ONLY shall be part of the identifier:
* bpartner
* bp location
* date promised
* order document no
* product
* QtyToDeliver
|
non_process
|
set correct identifiers in the table m shipmentschedule is this a bug or feature request fr follow up of where we set the table record as identifiers needed for what is the current behavior shipment schedule string representation shows the table record id which are the steps to reproduce check subscription progress window check the pick process params see what is the expected or desired behavior following fields only shall be part of the identifier bpartner bp location date promised order document no product qtytodeliver
| 0
|
195
| 2,600,253,533
|
IssuesEvent
|
2015-02-23 15:17:13
|
cfpb/hmda-viz-prototype
|
https://api.github.com/repos/cfpb/hmda-viz-prototype
|
opened
|
Create log files
|
enhancement Processing
|
Log which reports are created for which state/msa
Log which json template files are created (front end use files)
|
1.0
|
Create log files - Log which reports are created for which state/msa
Log which json template files are created (front end use files)
|
process
|
create log files log which reports are created for which state msa log which json template files are created front end use files
| 1
|
162,799
| 25,598,173,685
|
IssuesEvent
|
2022-12-01 17:47:20
|
carbon-design-system/carbon
|
https://api.github.com/repos/carbon-design-system/carbon
|
closed
|
Fluid input: finalize form spec
|
component: form role: design :pencil2:
|
The form component for Fluid input needs to be spec'ed and finalized.
Does it include a header, footer, etc.
[stub]
|
1.0
|
Fluid input: finalize form spec - The form component for Fluid input needs to be spec'ed and finalized.
Does it include a header, footer, etc.
[stub]
|
non_process
|
fluid input finalize form spec the form component for fluid input needs to be spec ed and finalized does it include a header footer etc
| 0
|
11,839
| 14,656,687,864
|
IssuesEvent
|
2020-12-28 13:58:56
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
Account lock and Forgot password email for PM
|
Auth server Bug P1 Process: Release 2 Process: Tested QA Process: Tested dev
|
(Rephrasing this issue), the Auth Server is currently sending the same emails to the Participant Manager and Mobile App for the Forgot Password and Account Lock functionality. This needs to be corrected to send application-specific emails.
|
3.0
|
Account lock and Forgot password email for PM - (Rephrasing this issue), the Auth Server is currently sending the same emails to the Participant Manager and Mobile App for the Forgot Password and Account Lock functionality. This needs to be corrected to send application-specific emails.
|
process
|
account lock and forgot password email for pm rephrasing this issue the auth server is currently sending the same emails to the participant manager and mobile app for the forgot password and account lock functionality this needs to be corrected to send application specific emails
| 1
|
22,305
| 7,154,313,148
|
IssuesEvent
|
2018-01-26 07:52:44
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
[NEW] gcc 7 and -Werror=implicit-fallthrough
|
area/build
|
### What version of gRPC and what language are you using?
1.8.2
### What operating system (Linux, Windows, …) and version?
Ubuntu 17.10
### What runtime / compiler are you using (e.g. python version or version of gcc)
gcc 7.2.0
### What did you do?
```
$ git clone -b $(curl -L https://grpc.io/release) https://github.com/grpc/grpc
$ cd grpc
$ git submodule update --init
$ make
```
### What did you expect to see?
Successful compilation of source
### What did you see instead?
```
[C] Compiling third_party/boringssl/crypto/bio/file.c
third_party/boringssl/crypto/bio/file.c: In function ‘file_ctrl’:
third_party/boringssl/crypto/bio/file.c:186:11: error: this statement may fall through [-Werror=implicit-fallthrough=]
num = 0;
~~~~^~~
third_party/boringssl/crypto/bio/file.c:187:5: note: here
case BIO_C_FILE_SEEK:
^~~~
cc1: all warnings being treated as errors
Makefile:2569: recipe for target '/home/oleg/Downloads/grpc/objs/opt/third_party/boringssl/crypto/bio/file.o' failed
make: *** [/home/oleg/Downloads/grpc/objs/opt/third_party/boringssl/crypto/bio/file.o] Error 1
```
### Anything else we should know about your project / environment?
Related to #13640 and to closed #12536 and #12022
### Workaround
`make CFLAGS='-Wno-implicit-fallthrough'`
|
1.0
|
[NEW] gcc 7 and -Werror=implicit-fallthrough - ### What version of gRPC and what language are you using?
1.8.2
### What operating system (Linux, Windows, …) and version?
Ubuntu 17.10
### What runtime / compiler are you using (e.g. python version or version of gcc)
gcc 7.2.0
### What did you do?
```
$ git clone -b $(curl -L https://grpc.io/release) https://github.com/grpc/grpc
$ cd grpc
$ git submodule update --init
$ make
```
### What did you expect to see?
Successful compilation of source
### What did you see instead?
```
[C] Compiling third_party/boringssl/crypto/bio/file.c
third_party/boringssl/crypto/bio/file.c: In function ‘file_ctrl’:
third_party/boringssl/crypto/bio/file.c:186:11: error: this statement may fall through [-Werror=implicit-fallthrough=]
num = 0;
~~~~^~~
third_party/boringssl/crypto/bio/file.c:187:5: note: here
case BIO_C_FILE_SEEK:
^~~~
cc1: all warnings being treated as errors
Makefile:2569: recipe for target '/home/oleg/Downloads/grpc/objs/opt/third_party/boringssl/crypto/bio/file.o' failed
make: *** [/home/oleg/Downloads/grpc/objs/opt/third_party/boringssl/crypto/bio/file.o] Error 1
```
### Anything else we should know about your project / environment?
Related to #13640 and to closed #12536 and #12022
### Workaround
`make CFLAGS='-Wno-implicit-fallthrough'`
|
non_process
|
gcc and werror implicit fallthrough what version of grpc and what language are you using what operating system linux windows … and version ubuntu what runtime compiler are you using e g python version or version of gcc gcc what did you do git clone b curl l cd grpc git submodule update init make what did you expect to see successful compilation of source what did you see instead compiling third party boringssl crypto bio file c third party boringssl crypto bio file c in function ‘file ctrl’ third party boringssl crypto bio file c error this statement may fall through num third party boringssl crypto bio file c note here case bio c file seek all warnings being treated as errors makefile recipe for target home oleg downloads grpc objs opt third party boringssl crypto bio file o failed make error anything else we should know about your project environment related to and to closed and workaround make cflags wno implicit fallthrough
| 0
|
11,259
| 14,040,239,137
|
IssuesEvent
|
2020-11-01 01:12:35
|
PPHubApp/PPHub-Feedback
|
https://api.github.com/repos/PPHubApp/PPHub-Feedback
|
closed
|
v2.4.10 - 浅色模式下,小组件里的git提交图展示空...
|
Bug 🐛 Processing 👨🏻💻🚧
|
浅色模式下,小组件里的git提交图展示空白。深色模式下正常。
运行环境: iPhone 6s Plus - iOS14.1 - v2.4.10(189)
|
1.0
|
v2.4.10 - 浅色模式下,小组件里的git提交图展示空... - 浅色模式下,小组件里的git提交图展示空白。深色模式下正常。
运行环境: iPhone 6s Plus - iOS14.1 - v2.4.10(189)
|
process
|
浅色模式下,小组件里的git提交图展示空 浅色模式下,小组件里的git提交图展示空白。深色模式下正常。 运行环境 iphone plus
| 1
|
1,904
| 4,728,745,728
|
IssuesEvent
|
2016-10-18 16:45:55
|
sysown/proxysql
|
https://api.github.com/repos/sysown/proxysql
|
closed
|
feature request: new gloval_variable digest_use_lowercase
|
QUERY PROCESSOR
|
Normally hashes are case sensitive, but I'm not sure if it is supposed to be related to digest_text, since mysql doesn't care if statements are lower or upper case.
```
mysql> select digest,digest_text from stats_mysql_query_digest where digest_text
like "select id from sample_table limit ?";
+--------------------+--------------------------------------+
| digest | digest_text |
+--------------------+--------------------------------------+
| 0xE1612CA769ADE07E | SELECT id from sample_table limit ? |
| 0x575A2061B1BCC5F0 | select id from sample_table limit ? |
+--------------------+--------------------------------------+
```
How about a global variable like `digest_use_lowercase` which is default `0`, but if set to `1` always convert digest_text to lowercase before creating the digest, to don't break compatibility with existing installations using the hash?
Otherwise rules on digest won't match if a user uses uppercase, while the rule was based on lowercase - so it's somehow related on #717 :-)
|
1.0
|
feature request: new gloval_variable digest_use_lowercase - Normally hashes are case sensitive, but I'm not sure if it is supposed to be related to digest_text, since mysql doesn't care if statements are lower or upper case.
```
mysql> select digest,digest_text from stats_mysql_query_digest where digest_text
like "select id from sample_table limit ?";
+--------------------+--------------------------------------+
| digest | digest_text |
+--------------------+--------------------------------------+
| 0xE1612CA769ADE07E | SELECT id from sample_table limit ? |
| 0x575A2061B1BCC5F0 | select id from sample_table limit ? |
+--------------------+--------------------------------------+
```
How about a global variable like `digest_use_lowercase` which is default `0`, but if set to `1` always convert digest_text to lowercase before creating the digest, to don't break compatibility with existing installations using the hash?
Otherwise rules on digest won't match if a user uses uppercase, while the rule was based on lowercase - so it's somehow related on #717 :-)
|
process
|
feature request new gloval variable digest use lowercase normally hashes are case sensitive but i m not sure if it is supposed to be related to digest text since mysql doesn t care if statements are lower or upper case mysql select digest digest text from stats mysql query digest where digest text like select id from sample table limit digest digest text select id from sample table limit select id from sample table limit how about a global variable like digest use lowercase which is default but if set to always convert digest text to lowercase before creating the digest to don t break compatibility with existing installations using the hash otherwise rules on digest won t match if a user uses uppercase while the rule was based on lowercase so it s somehow related on
| 1
|
18,707
| 24,600,155,710
|
IssuesEvent
|
2022-10-14 11:48:03
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Consent API] Consent API Disable > Cloud Storage > Data sharing image is not getting displayed in the cloud storage in the following scenario
|
Bug P0 Process: Fixed Process: Tested QA Process: Tested dev
|
**Pre-Condition:** Study should be created with the data sharing permission
**Steps:**
1. Install the mobile app
2. Sign in / Sign up
3. Enroll to the study by providing data sharing permission
4. Withdraw from that study
5. Again enroll to the same study without providing data sharing permission
6. Now, Go to Cloud storage
7. Search the studyID in the consent bucket
8. Click on the participant record and Verify
**AR:** Data sharing image is not getting displayed in the cloud storage in the following scenario
**ER:** Data sharing image should get displayed in the cloud storage in the following scenario
**Note:** Issue needs to be fixed even when without providing data sharing permission in step 3 and by providing data sharing permission in step 5 scenario
|
3.0
|
[Consent API] Consent API Disable > Cloud Storage > Data sharing image is not getting displayed in the cloud storage in the following scenario - **Pre-Condition:** Study should be created with the data sharing permission
**Steps:**
1. Install the mobile app
2. Sign in / Sign up
3. Enroll to the study by providing data sharing permission
4. Withdraw from that study
5. Again enroll to the same study without providing data sharing permission
6. Now, Go to Cloud storage
7. Search the studyID in the consent bucket
8. Click on the participant record and Verify
**AR:** Data sharing image is not getting displayed in the cloud storage in the following scenario
**ER:** Data sharing image should get displayed in the cloud storage in the following scenario
**Note:** Issue needs to be fixed even when without providing data sharing permission in step 3 and by providing data sharing permission in step 5 scenario
|
process
|
consent api disable cloud storage data sharing image is not getting displayed in the cloud storage in the following scenario pre condition study should be created with the data sharing permission steps install the mobile app sign in sign up enroll to the study by providing data sharing permission withdraw from that study again enroll to the same study without providing data sharing permission now go to cloud storage search the studyid in the consent bucket click on the participant record and verify ar data sharing image is not getting displayed in the cloud storage in the following scenario er data sharing image should get displayed in the cloud storage in the following scenario note issue needs to be fixed even when without providing data sharing permission in step and by providing data sharing permission in step scenario
| 1
|
37,310
| 6,589,277,877
|
IssuesEvent
|
2017-09-14 08:18:34
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
Using towncrier for release notes?
|
Documentation needs-decision
|
How about using https://pypi.python.org/pypi/towncrier for release notes?
In this approach, we add `doc/release/changes/<submodule>` folders, into which each PR that implements a feature adds a new file of format `<pr-number>.<suffix>` where `<suffix>` is `bugfix`, `trivial`, `feature`, ... etc.
At release time, these files are concatenated (with a jinja template) to form the release notes, and removed.
The good thing is that this eliminates merge conflicts in release notes, and perhaps reduces the work for the release manager when the burden of writing the content is pushed mainly to PR authors.
The approach is used e.g. by pytest.
|
1.0
|
Using towncrier for release notes? - How about using https://pypi.python.org/pypi/towncrier for release notes?
In this approach, we add `doc/release/changes/<submodule>` folders, into which each PR that implements a feature adds a new file of format `<pr-number>.<suffix>` where `<suffix>` is `bugfix`, `trivial`, `feature`, ... etc.
At release time, these files are concatenated (with a jinja template) to form the release notes, and removed.
The good thing is that this eliminates merge conflicts in release notes, and perhaps reduces the work for the release manager when the burden of writing the content is pushed mainly to PR authors.
The approach is used e.g. by pytest.
|
non_process
|
using towncrier for release notes how about using for release notes in this approach we add doc release changes folders into which each pr that implements a feature adds a new file of format where is bugfix trivial feature etc at release time these files are concatenated with a jinja template to form the release notes and removed the good thing is that this eliminates merge conflicts in release notes and perhaps reduces the work for the release manager when the burden of writing the content is pushed mainly to pr authors the approach is used e g by pytest
| 0
|
17,198
| 22,773,863,567
|
IssuesEvent
|
2022-07-08 12:43:19
|
camunda/zeebe
|
https://api.github.com/repos/camunda/zeebe
|
closed
|
Stack Overflow Error in DMN Parser
|
kind/bug area/performance severity/high area/ux area/reliability team/process-automation
|
**Describe the bug**
Observed in production logs: https://console.cloud.google.com/errors/detail/CIHn_-jg7uShyQE;service=zeebe;time=P7D?project=camunda-cloud-240911
in cluster: 4ae19f04-89da-40e4-9021-2b77029acf1a
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
**Log/Stacktrace**
<details><summary>Full Stacktrace</summary>
<p>
```
java.lang.StackOverflowError: null
at org.camunda.bpm.model.xml.impl.util.DomUtil.filterNodeList(DomUtil.java:140) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at org.camunda.bpm.model.xml.impl.util.DomUtil.filterNodeListByType(DomUtil.java:184) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at org.camunda.bpm.model.xml.impl.instance.DomElementImpl.getChildElementsByType(DomElementImpl.java:136) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at org.camunda.bpm.model.xml.impl.type.child.ChildElementCollectionImpl.getView(ChildElementCollectionImpl.java:82) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at org.camunda.bpm.model.xml.impl.type.child.ChildElementCollectionImpl.access$000(ChildElementCollectionImpl.java:38) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at org.camunda.bpm.model.xml.impl.type.child.ChildElementCollectionImpl$1.iterator(ChildElementCollectionImpl.java:164) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at scala.collection.convert.JavaCollectionWrappers$JCollectionWrapper.iterator(JavaCollectionWrappers.scala:72) ~[scala-library-2.13.8.jar:?]
at scala.collection.StrictOptimizedIterableOps.map(StrictOptimizedIterableOps.scala:98) ~[scala-library-2.13.8.jar:?]
at scala.collection.StrictOptimizedIterableOps.map$(StrictOptimizedIterableOps.scala:87) ~[scala-library-2.13.8.jar:?]
at scala.collection.convert.JavaCollectionWrappers$JCollectionWrapper.map(JavaCollectionWrappers.scala:68) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.parseDecision(DmnParser.scala:118) ~[dmn-engine-1.7.1.jar:1.7.1]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$3(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.mutable.HashMap.getOrElseUpdate(HashMap.scala:454) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$2(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.StrictOptimizedIterableOps.map(StrictOptimizedIterableOps.scala:100) ~[scala-library-2.13.8.jar:?]
at scala.collection.StrictOptimizedIterableOps.map$(StrictOptimizedIterableOps.scala:87) ~[scala-library-2.13.8.jar:?]
at scala.collection.mutable.ArrayBuffer.map(ArrayBuffer.scala:43) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.parseDecision(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$3(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.mutable.HashMap.getOrElseUpdate(HashMap.scala:454) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$2(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.StrictOptimizedIterableOps.map(StrictOptimizedIterableOps.scala:100) ~[scala-library-2.13.8.jar:?]
at scala.collection.StrictOptimizedIterableOps.map$(StrictOptimizedIterableOps.scala:87) ~[scala-library-2.13.8.jar:?]
at scala.collection.mutable.ArrayBuffer.map(ArrayBuffer.scala:43) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.parseDecision(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$3(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.mutable.HashMap.getOrElseUpdate(HashMap.scala:454) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$2(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.StrictOptimizedIterableOps.map(StrictOptimizedIterableOps.scala:100) ~[scala-library-2.13.8.jar:?]
at scala.collection.StrictOptimizedIterableOps.map$(StrictOptimizedIterableOps.scala:87) ~[scala-library-2.13.8.jar:?]
...
```
</p>
</details>
**Environment:**
- OS: Camunda Cloud
- Zeebe Version: 8.0.2
- Configuration: <!-- [e.g. exporters etc.] -->
|
1.0
|
Stack Overflow Error in DMN Parser - **Describe the bug**
Observed in production logs: https://console.cloud.google.com/errors/detail/CIHn_-jg7uShyQE;service=zeebe;time=P7D?project=camunda-cloud-240911
in cluster: 4ae19f04-89da-40e4-9021-2b77029acf1a
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
**Log/Stacktrace**
<details><summary>Full Stacktrace</summary>
<p>
```
java.lang.StackOverflowError: null
at org.camunda.bpm.model.xml.impl.util.DomUtil.filterNodeList(DomUtil.java:140) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at org.camunda.bpm.model.xml.impl.util.DomUtil.filterNodeListByType(DomUtil.java:184) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at org.camunda.bpm.model.xml.impl.instance.DomElementImpl.getChildElementsByType(DomElementImpl.java:136) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at org.camunda.bpm.model.xml.impl.type.child.ChildElementCollectionImpl.getView(ChildElementCollectionImpl.java:82) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at org.camunda.bpm.model.xml.impl.type.child.ChildElementCollectionImpl.access$000(ChildElementCollectionImpl.java:38) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at org.camunda.bpm.model.xml.impl.type.child.ChildElementCollectionImpl$1.iterator(ChildElementCollectionImpl.java:164) ~[camunda-xml-model-7.16.0.jar:7.16.0]
at scala.collection.convert.JavaCollectionWrappers$JCollectionWrapper.iterator(JavaCollectionWrappers.scala:72) ~[scala-library-2.13.8.jar:?]
at scala.collection.StrictOptimizedIterableOps.map(StrictOptimizedIterableOps.scala:98) ~[scala-library-2.13.8.jar:?]
at scala.collection.StrictOptimizedIterableOps.map$(StrictOptimizedIterableOps.scala:87) ~[scala-library-2.13.8.jar:?]
at scala.collection.convert.JavaCollectionWrappers$JCollectionWrapper.map(JavaCollectionWrappers.scala:68) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.parseDecision(DmnParser.scala:118) ~[dmn-engine-1.7.1.jar:1.7.1]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$3(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.mutable.HashMap.getOrElseUpdate(HashMap.scala:454) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$2(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.StrictOptimizedIterableOps.map(StrictOptimizedIterableOps.scala:100) ~[scala-library-2.13.8.jar:?]
at scala.collection.StrictOptimizedIterableOps.map$(StrictOptimizedIterableOps.scala:87) ~[scala-library-2.13.8.jar:?]
at scala.collection.mutable.ArrayBuffer.map(ArrayBuffer.scala:43) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.parseDecision(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$3(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.mutable.HashMap.getOrElseUpdate(HashMap.scala:454) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$2(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.StrictOptimizedIterableOps.map(StrictOptimizedIterableOps.scala:100) ~[scala-library-2.13.8.jar:?]
at scala.collection.StrictOptimizedIterableOps.map$(StrictOptimizedIterableOps.scala:87) ~[scala-library-2.13.8.jar:?]
at scala.collection.mutable.ArrayBuffer.map(ArrayBuffer.scala:43) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.parseDecision(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$3(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.mutable.HashMap.getOrElseUpdate(HashMap.scala:454) ~[scala-library-2.13.8.jar:?]
at org.camunda.dmn.parser.DmnParser.$anonfun$parseDecision$2(DmnParser.scala:120) ~[dmn-engine-1.7.1.jar:1.7.1]
at scala.collection.StrictOptimizedIterableOps.map(StrictOptimizedIterableOps.scala:100) ~[scala-library-2.13.8.jar:?]
at scala.collection.StrictOptimizedIterableOps.map$(StrictOptimizedIterableOps.scala:87) ~[scala-library-2.13.8.jar:?]
...
```
</p>
</details>
**Environment:**
- OS: Camunda Cloud
- Zeebe Version: 8.0.2
- Configuration: <!-- [e.g. exporters etc.] -->
|
process
|
stack overflow error in dmn parser describe the bug observed in production logs in cluster expected behavior log stacktrace full stacktrace java lang stackoverflowerror null at org camunda bpm model xml impl util domutil filternodelist domutil java at org camunda bpm model xml impl util domutil filternodelistbytype domutil java at org camunda bpm model xml impl instance domelementimpl getchildelementsbytype domelementimpl java at org camunda bpm model xml impl type child childelementcollectionimpl getview childelementcollectionimpl java at org camunda bpm model xml impl type child childelementcollectionimpl access childelementcollectionimpl java at org camunda bpm model xml impl type child childelementcollectionimpl iterator childelementcollectionimpl java at scala collection convert javacollectionwrappers jcollectionwrapper iterator javacollectionwrappers scala at scala collection strictoptimizediterableops map strictoptimizediterableops scala at scala collection strictoptimizediterableops map strictoptimizediterableops scala at scala collection convert javacollectionwrappers jcollectionwrapper map javacollectionwrappers scala at org camunda dmn parser dmnparser parsedecision dmnparser scala at org camunda dmn parser dmnparser anonfun parsedecision dmnparser scala at scala collection mutable hashmap getorelseupdate hashmap scala at org camunda dmn parser dmnparser anonfun parsedecision dmnparser scala at scala collection strictoptimizediterableops map strictoptimizediterableops scala at scala collection strictoptimizediterableops map strictoptimizediterableops scala at scala collection mutable arraybuffer map arraybuffer scala at org camunda dmn parser dmnparser parsedecision dmnparser scala at org camunda dmn parser dmnparser anonfun parsedecision dmnparser scala at scala collection mutable hashmap getorelseupdate hashmap scala at org camunda dmn parser dmnparser anonfun parsedecision dmnparser scala at scala collection strictoptimizediterableops map strictoptimizediterableops scala at scala collection strictoptimizediterableops map strictoptimizediterableops scala at scala collection mutable arraybuffer map arraybuffer scala at org camunda dmn parser dmnparser parsedecision dmnparser scala at org camunda dmn parser dmnparser anonfun parsedecision dmnparser scala at scala collection mutable hashmap getorelseupdate hashmap scala at org camunda dmn parser dmnparser anonfun parsedecision dmnparser scala at scala collection strictoptimizediterableops map strictoptimizediterableops scala at scala collection strictoptimizediterableops map strictoptimizediterableops scala environment os camunda cloud zeebe version configuration
| 1
|
2,143
| 4,995,932,229
|
IssuesEvent
|
2016-12-09 11:59:30
|
Alfresco/alfresco-ng2-components
|
https://api.github.com/repos/Alfresco/alfresco-ng2-components
|
opened
|
When clicking on an app after login an error is showed in console.
|
bug comp: activiti-processList
|
<!--
PLEASE FILL OUT THE FOLLOWING INFORMATION, THIS WILL HELP US TO RESOLVE YOUR PROBLEM FASTER.
REMEMBER FOR SUPPORT REQUESTS YOU CAN ALSO ASK ON OUR GITTER CHAT:
Please ask before on our gitter channel https://gitter.im/Alfresco/alfresco-ng2-components
-->
**Type of issue:** (check with "[x]")
```
- [ ] New feature request
- [x] Bug
- [ ] Support request
```
**Current behavior:**
<!-- Describe the current behavior. -->
**Expected behavior:**
<!-- Describe the expected behavior. -->
**Steps to reproduce the issue:**
<!-- Describe the steps to reproduce the issue. -->
1. Login into the demo shell with an app already deployed on activiti
2.Click on Activiti
3.Click on the deployed app
**Component name and version:**
<!-- Example: ng2-alfresco-login. Check before if this issue is still present in the most recent version -->
**Browser and version:**
<!-- [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] -->
**Node version (for build issues):**
<!-- To check the version: node --version -->
**New feature request:**
<!-- Describe the feature, motivation and the concrete use case (only in case of new feature request) -->
|
1.0
|
When clicking on an app after login an error is showed in console. - <!--
PLEASE FILL OUT THE FOLLOWING INFORMATION, THIS WILL HELP US TO RESOLVE YOUR PROBLEM FASTER.
REMEMBER FOR SUPPORT REQUESTS YOU CAN ALSO ASK ON OUR GITTER CHAT:
Please ask before on our gitter channel https://gitter.im/Alfresco/alfresco-ng2-components
-->
**Type of issue:** (check with "[x]")
```
- [ ] New feature request
- [x] Bug
- [ ] Support request
```
**Current behavior:**
<!-- Describe the current behavior. -->
**Expected behavior:**
<!-- Describe the expected behavior. -->
**Steps to reproduce the issue:**
<!-- Describe the steps to reproduce the issue. -->
1. Login into the demo shell with an app already deployed on activiti
2.Click on Activiti
3.Click on the deployed app
**Component name and version:**
<!-- Example: ng2-alfresco-login. Check before if this issue is still present in the most recent version -->
**Browser and version:**
<!-- [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] -->
**Node version (for build issues):**
<!-- To check the version: node --version -->
**New feature request:**
<!-- Describe the feature, motivation and the concrete use case (only in case of new feature request) -->
|
process
|
when clicking on an app after login an error is showed in console please fill out the following information this will help us to resolve your problem faster remember for support requests you can also ask on our gitter chat please ask before on our gitter channel type of issue check with new feature request bug support request current behavior expected behavior steps to reproduce the issue login into the demo shell with an app already deployed on activiti click on activiti click on the deployed app component name and version browser and version node version for build issues new feature request
| 1
|
9,715
| 12,710,091,174
|
IssuesEvent
|
2020-06-23 13:23:48
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
opened
|
Custom Column after aggregation creates wrong query and fails
|
Priority:P2 Querying/Notebook Querying/Processor Type:Bug
|
**Describe the bug**
When adding a Custom Column with fields from aggregated data, then the query generated is mixing the original columns and the aggregated columns, which fails query with column not found.
**To Reproduce**
Steps to reproduce the behavior:
1. Custom question > Sample Dataset > Orders
2. Summarize "Sum of Subtotal" and "Sum of Total" by "CreatedAt:Year"
3. Custom Column `[Sum of Subtotal] + [Sum of Total]` as "MegaTotal"

4. Fails with error:
```
Column "source.SUBTOTAL" not found; SQL statement: CREATE FORCE VIEW PUBLIC._8 AS SELECT ("source"."sum" + "source"."sum_2") AS "MegaTotal", "source".SUBTOTAL AS SUBTOTAL, "source".TOTAL AS TOTAL, "source".CREATED_AT AS CREATED_AT, "source"."sum" AS "sum", "source"."sum_2" AS "sum_2", "source".CREATED_AT AS CREATED_AT_2 FROM ( SELECT PARSEDATETIME(YEAR(PUBLIC.ORDERS.CREATED_AT), 'yyyy') AS CREATED_AT, SUM(PUBLIC.ORDERS.SUBTOTAL) AS "sum", SUM(PUBLIC.ORDERS.TOTAL) AS "sum_2" FROM PUBLIC.ORDERS GROUP BY PARSEDATETIME(YEAR(PUBLIC.ORDERS.CREATED_AT), 'yyyy') ORDER BY 1 ) "source" [42122-197]
```
5. "View the SQL":
```SQL
SELECT "source"."CREATED_AT" AS "CREATED_AT",
"source"."sum" AS "sum",
"source"."sum_2" AS "sum_2",
"source"."MegaTotal" AS "MegaTotal"
FROM (SELECT ( "source"."sum" + "source"."sum_2" ) AS "MegaTotal",
"source"."SUBTOTAL" AS "SUBTOTAL",
"source"."TOTAL" AS "TOTAL",
"source"."CREATED_AT" AS "CREATED_AT",
"source"."sum" AS "sum",
"source"."sum_2" AS "sum_2",
"source"."CREATED_AT" AS "CREATED_AT_2"
FROM (SELECT parsedatetime(year("PUBLIC"."ORDERS"."CREATED_AT"), 'yyyy') AS "CREATED_AT",
sum("PUBLIC"."ORDERS"."SUBTOTAL") AS "sum",
sum("PUBLIC"."ORDERS"."TOTAL") AS "sum_2"
FROM "PUBLIC"."ORDERS"
GROUP BY parsedatetime(year("PUBLIC"."ORDERS"."CREATED_AT"), 'yyyy')
ORDER BY parsedatetime(year("PUBLIC"."ORDERS"."CREATED_AT"), 'yyyy') ASC) "source") "source"
LIMIT 1048576
```
**Information about your Metabase Installation:**
Metabase 0.35.4 and `master` on various backends and datasources
**Additional context**
I have a feeling that this might be a duplicate of #12507
https://discourse.metabase.com/t/error-on-use-summarized-column-on-custom-column/10623
|
1.0
|
Custom Column after aggregation creates wrong query and fails - **Describe the bug**
When adding a Custom Column with fields from aggregated data, then the query generated is mixing the original columns and the aggregated columns, which fails query with column not found.
**To Reproduce**
Steps to reproduce the behavior:
1. Custom question > Sample Dataset > Orders
2. Summarize "Sum of Subtotal" and "Sum of Total" by "CreatedAt:Year"
3. Custom Column `[Sum of Subtotal] + [Sum of Total]` as "MegaTotal"

4. Fails with error:
```
Column "source.SUBTOTAL" not found; SQL statement: CREATE FORCE VIEW PUBLIC._8 AS SELECT ("source"."sum" + "source"."sum_2") AS "MegaTotal", "source".SUBTOTAL AS SUBTOTAL, "source".TOTAL AS TOTAL, "source".CREATED_AT AS CREATED_AT, "source"."sum" AS "sum", "source"."sum_2" AS "sum_2", "source".CREATED_AT AS CREATED_AT_2 FROM ( SELECT PARSEDATETIME(YEAR(PUBLIC.ORDERS.CREATED_AT), 'yyyy') AS CREATED_AT, SUM(PUBLIC.ORDERS.SUBTOTAL) AS "sum", SUM(PUBLIC.ORDERS.TOTAL) AS "sum_2" FROM PUBLIC.ORDERS GROUP BY PARSEDATETIME(YEAR(PUBLIC.ORDERS.CREATED_AT), 'yyyy') ORDER BY 1 ) "source" [42122-197]
```
5. "View the SQL":
```SQL
SELECT "source"."CREATED_AT" AS "CREATED_AT",
"source"."sum" AS "sum",
"source"."sum_2" AS "sum_2",
"source"."MegaTotal" AS "MegaTotal"
FROM (SELECT ( "source"."sum" + "source"."sum_2" ) AS "MegaTotal",
"source"."SUBTOTAL" AS "SUBTOTAL",
"source"."TOTAL" AS "TOTAL",
"source"."CREATED_AT" AS "CREATED_AT",
"source"."sum" AS "sum",
"source"."sum_2" AS "sum_2",
"source"."CREATED_AT" AS "CREATED_AT_2"
FROM (SELECT parsedatetime(year("PUBLIC"."ORDERS"."CREATED_AT"), 'yyyy') AS "CREATED_AT",
sum("PUBLIC"."ORDERS"."SUBTOTAL") AS "sum",
sum("PUBLIC"."ORDERS"."TOTAL") AS "sum_2"
FROM "PUBLIC"."ORDERS"
GROUP BY parsedatetime(year("PUBLIC"."ORDERS"."CREATED_AT"), 'yyyy')
ORDER BY parsedatetime(year("PUBLIC"."ORDERS"."CREATED_AT"), 'yyyy') ASC) "source") "source"
LIMIT 1048576
```
**Information about your Metabase Installation:**
Metabase 0.35.4 and `master` on various backends and datasources
**Additional context**
I have a feeling that this might be a duplicate of #12507
https://discourse.metabase.com/t/error-on-use-summarized-column-on-custom-column/10623
|
process
|
custom column after aggregation creates wrong query and fails describe the bug when adding a custom column with fields from aggregated data then the query generated is mixing the original columns and the aggregated columns which fails query with column not found to reproduce steps to reproduce the behavior custom question sample dataset orders summarize sum of subtotal and sum of total by createdat year custom column as megatotal fails with error column source subtotal not found sql statement create force view public as select source sum source sum as megatotal source subtotal as subtotal source total as total source created at as created at source sum as sum source sum as sum source created at as created at from select parsedatetime year public orders created at yyyy as created at sum public orders subtotal as sum sum public orders total as sum from public orders group by parsedatetime year public orders created at yyyy order by source view the sql sql select source created at as created at source sum as sum source sum as sum source megatotal as megatotal from select source sum source sum as megatotal source subtotal as subtotal source total as total source created at as created at source sum as sum source sum as sum source created at as created at from select parsedatetime year public orders created at yyyy as created at sum public orders subtotal as sum sum public orders total as sum from public orders group by parsedatetime year public orders created at yyyy order by parsedatetime year public orders created at yyyy asc source source limit information about your metabase installation metabase and master on various backends and datasources additional context i have a feeling that this might be a duplicate of
| 1
|
9,748
| 12,735,867,947
|
IssuesEvent
|
2020-06-25 15:58:16
|
pelias/pelias
|
https://api.github.com/repos/pelias/pelias
|
closed
|
Searching for Zoo in SF does not return the San Fransisco Zoo
|
POIs feedback processed
|
Pelias Reference for: mapzen/eraser-map#141
Logged by @nvkelso
- Device name: Samsung Galaxy S6 edge
- Android Version: 5.1.1
- App build number: master-481
- What did you expected to happen? I searched for zoo and expected to find the San Francisco Zoo to nav to as the top result.
- What happened instead? I didn't see the SF Zoo in the first 10 results, so had to fully qualify it as san francisco zoo to get that as the top result. We should be seeding search with the current map view so results are more specific.
- Steps to reproduce: Search for zoo with current location lat/lng somewhere in SF.
- Attach a screenshot See below.
- Attach device logs eh.
If something happened while you were searching, share with us:
- Where were you? 273 29th St (San Francisco, California, USA)
Searching for `zoo` near the SF Zoo should return that result first:

Instead I had to search for `san francisco zoo`:

|
1.0
|
Searching for Zoo in SF does not return the San Fransisco Zoo - Pelias Reference for: mapzen/eraser-map#141
Logged by @nvkelso
- Device name: Samsung Galaxy S6 edge
- Android Version: 5.1.1
- App build number: master-481
- What did you expected to happen? I searched for zoo and expected to find the San Francisco Zoo to nav to as the top result.
- What happened instead? I didn't see the SF Zoo in the first 10 results, so had to fully qualify it as san francisco zoo to get that as the top result. We should be seeding search with the current map view so results are more specific.
- Steps to reproduce: Search for zoo with current location lat/lng somewhere in SF.
- Attach a screenshot See below.
- Attach device logs eh.
If something happened while you were searching, share with us:
- Where were you? 273 29th St (San Francisco, California, USA)
Searching for `zoo` near the SF Zoo should return that result first:

Instead I had to search for `san francisco zoo`:

|
process
|
searching for zoo in sf does not return the san fransisco zoo pelias reference for mapzen eraser map logged by nvkelso device name samsung galaxy edge android version app build number master what did you expected to happen i searched for zoo and expected to find the san francisco zoo to nav to as the top result what happened instead i didn t see the sf zoo in the first results so had to fully qualify it as san francisco zoo to get that as the top result we should be seeding search with the current map view so results are more specific steps to reproduce search for zoo with current location lat lng somewhere in sf attach a screenshot see below attach device logs eh if something happened while you were searching share with us where were you st san francisco california usa searching for zoo near the sf zoo should return that result first instead i had to search for san francisco zoo
| 1
|
9,511
| 12,497,568,566
|
IssuesEvent
|
2020-06-01 16:43:09
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Determine triggering pipeline
|
Pri1 devops-cicd-process/tech devops/prod product-question
|
Hi team. Since a pipeline can be triggered by multiple other pipelines, is there any mechanism that exists to identify which exact pipeline was the trigger? My purpose is to control what happens in the triggered pipeline based on which pipeline triggered it.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 86285f72-9e28-da97-59bb-c29eb60f627d
* Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54
* Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml#feedback)
* Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @ashkir
* Microsoft Alias: **ashkir**
|
1.0
|
Determine triggering pipeline - Hi team. Since a pipeline can be triggered by multiple other pipelines, is there any mechanism that exists to identify which exact pipeline was the trigger? My purpose is to control what happens in the triggered pipeline based on which pipeline triggered it.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 86285f72-9e28-da97-59bb-c29eb60f627d
* Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54
* Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml#feedback)
* Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @ashkir
* Microsoft Alias: **ashkir**
|
process
|
determine triggering pipeline hi team since a pipeline can be triggered by multiple other pipelines is there any mechanism that exists to identify which exact pipeline was the trigger my purpose is to control what happens in the triggered pipeline based on which pipeline triggered it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login ashkir microsoft alias ashkir
| 1
|
20,805
| 27,565,864,875
|
IssuesEvent
|
2023-03-08 03:40:58
|
bitfocus/companion-module-requests
|
https://api.github.com/repos/bitfocus/companion-module-requests
|
closed
|
Youtube Studio
|
NOT YET PROCESSED
|
- [ ] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested**
The name of the device, hardware, or software you would like to control:
Youtube Studio
What you would like to be able to make it do from Companion:
Looking mainly to be able to start and stop livestreams in YT Studio from a companion button. I am sure there are plenty more functions and applications available, but for now that is the main function I am searching for.
Direct links or attachments to the ethernet control protocol or API:
https://developers.google.com/youtube/v3
https://github.com/adasq/youtube-studio
https://pypi.org/project/ytstudio/
|
1.0
|
Youtube Studio - - [ ] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested**
The name of the device, hardware, or software you would like to control:
Youtube Studio
What you would like to be able to make it do from Companion:
Looking mainly to be able to start and stop livestreams in YT Studio from a companion button. I am sure there are plenty more functions and applications available, but for now that is the main function I am searching for.
Direct links or attachments to the ethernet control protocol or API:
https://developers.google.com/youtube/v3
https://github.com/adasq/youtube-studio
https://pypi.org/project/ytstudio/
|
process
|
youtube studio i have researched the list of existing companion modules and requests and have determined this has not yet been requested the name of the device hardware or software you would like to control youtube studio what you would like to be able to make it do from companion looking mainly to be able to start and stop livestreams in yt studio from a companion button i am sure there are plenty more functions and applications available but for now that is the main function i am searching for direct links or attachments to the ethernet control protocol or api
| 1
|
16,654
| 21,722,716,683
|
IssuesEvent
|
2022-05-11 03:06:28
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
"Split vector Layer" error with GPKG inputs
|
Feedback stale Processing Bug
|
[aaa.gpkg.zip](https://github.com/qgis/QGIS/files/5735108/aaa.gpkg.zip)
Do the following:
- Create new GPKG and Layer
- Add some Points
- Add two new Field: fid_2 and SplitID
- Fill in your data for each row:
fid_2 has to be unique
SplitID repeat number 1 and 2 for all your points.
- Save layer
- Right click --> Export as... Layer
- In Layer Options, **set FID = fid_2**
- Save layer as tmp2 or something.
- Open that layer and edit all fid values to one number (eg all to "1")
Try to use: "Split vector Layer" with "Unique Field ID" = SplitID
Error:
First feature for every new split ID will be written. All others not:
"Feature could not be written to <Path to File>"
As soon as you delete the "fid" field, the tool will not give the error and work as exspected.
Long story short:
When giving the FID a different name then "fid" and leave the original "fid" field with non unique data in, an error will occur.
(Sidenode: Very funny in conjunction with tools like "Points along geometry", which are not updating the original fid field)
Version: 3.10.3
[aaa.gpkg.zip](https://github.com/qgis/QGIS/files/5735110/aaa.gpkg.zip)
|
1.0
|
"Split vector Layer" error with GPKG inputs -
[aaa.gpkg.zip](https://github.com/qgis/QGIS/files/5735108/aaa.gpkg.zip)
Do the following:
- Create new GPKG and Layer
- Add some Points
- Add two new Field: fid_2 and SplitID
- Fill in your data for each row:
fid_2 has to be unique
SplitID repeat number 1 and 2 for all your points.
- Save layer
- Right click --> Export as... Layer
- In Layer Options, **set FID = fid_2**
- Save layer as tmp2 or something.
- Open that layer and edit all fid values to one number (eg all to "1")
Try to use: "Split vector Layer" with "Unique Field ID" = SplitID
Error:
First feature for every new split ID will be written. All others not:
"Feature could not be written to <Path to File>"
As soon as you delete the "fid" field, the tool will not give the error and work as exspected.
Long story short:
When giving the FID a different name then "fid" and leave the original "fid" field with non unique data in, an error will occur.
(Sidenode: Very funny in conjunction with tools like "Points along geometry", which are not updating the original fid field)
Version: 3.10.3
[aaa.gpkg.zip](https://github.com/qgis/QGIS/files/5735110/aaa.gpkg.zip)
|
process
|
split vector layer error with gpkg inputs do the following create new gpkg and layer add some points add two new field fid and splitid fill in your data for each row fid has to be unique splitid repeat number and for all your points save layer right click export as layer in layer options set fid fid save layer as or something open that layer and edit all fid values to one number eg all to try to use split vector layer with unique field id splitid error first feature for every new split id will be written all others not feature could not be written to as soon as you delete the fid field the tool will not give the error and work as exspected long story short when giving the fid a different name then fid and leave the original fid field with non unique data in an error will occur sidenode very funny in conjunction with tools like points along geometry which are not updating the original fid field version
| 1
|
4,463
| 7,330,773,498
|
IssuesEvent
|
2018-03-05 11:04:41
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
closed
|
Script processor damages iterator properties in for..in loop
|
AREA: client SYSTEM: resource processing TYPE: bug
|
Reproducible in playground.
Hammerhead turns
```
for (obj.iterator in obj.enum)
div.innerHTML += obj.iterator + '<br/>';
```
to
```
for ( var __set$temp in obj.enum)
__set$(div,"innerHTML",__get$(div,"innerHTML")+(obj.iterator+'<br/>'))iv.innerHTML += obj.iterator + '<br/>' obj.iterator=__set$temp;iv.innerHTML += obj.iterator + '<br/>';
```
Example page:
https://andreybelym.github.io/test-pages/mailbox-2018-03-05-1.html
|
1.0
|
Script processor damages iterator properties in for..in loop - Reproducible in playground.
Hammerhead turns
```
for (obj.iterator in obj.enum)
div.innerHTML += obj.iterator + '<br/>';
```
to
```
for ( var __set$temp in obj.enum)
__set$(div,"innerHTML",__get$(div,"innerHTML")+(obj.iterator+'<br/>'))iv.innerHTML += obj.iterator + '<br/>' obj.iterator=__set$temp;iv.innerHTML += obj.iterator + '<br/>';
```
Example page:
https://andreybelym.github.io/test-pages/mailbox-2018-03-05-1.html
|
process
|
script processor damages iterator properties in for in loop reproducible in playground hammerhead turns for obj iterator in obj enum div innerhtml obj iterator to for var set temp in obj enum set div innerhtml get div innerhtml obj iterator iv innerhtml obj iterator obj iterator set temp iv innerhtml obj iterator example page
| 1
|
3,736
| 6,733,146,656
|
IssuesEvent
|
2017-10-18 13:59:22
|
york-region-tpss/stp
|
https://api.github.com/repos/york-region-tpss/stp
|
closed
|
Extra Work Dashboard - Display the Total Number and Remaining Unpaid Items
|
enhancement process workflow
|
Calculate the unpaid items based on the paid history.
Display the remaining unpaid items and only allow users to pay within the remaining unpaid limit.
|
1.0
|
Extra Work Dashboard - Display the Total Number and Remaining Unpaid Items - Calculate the unpaid items based on the paid history.
Display the remaining unpaid items and only allow users to pay within the remaining unpaid limit.
|
process
|
extra work dashboard display the total number and remaining unpaid items calculate the unpaid items based on the paid history display the remaining unpaid items and only allow users to pay within the remaining unpaid limit
| 1
|
559,728
| 16,574,983,176
|
IssuesEvent
|
2021-05-31 02:18:58
|
GoogleContainerTools/skaffold
|
https://api.github.com/repos/GoogleContainerTools/skaffold
|
closed
|
Detect .csproj files during init and leverage the Buildpack builder
|
area/init kind/friction planning/Q2-21 priority/p2 source/partnerships
|
<!--
Issues without logs and details are more complicated to fix.
Please help us by filling the template below!
-->
### Expected behavior
.net projects work with init when they dont have a Dockerfile
### Actual behavior
The init flow fails as it does not detect a proper builder
### Information
- Skaffold version: v1.23.0
- Operating system: MacOS
- Installed via: Cloud Code
- Contents of skaffold.yaml: None
### Steps to reproduce the behavior
1. Install .net Core https://dotnet.microsoft.com/download
2. Create a directory, cd to it and run `dotnet new mvc`
3. Try `skaffold init`
```
❯ skaffold init -vdebug
INFO[0000] Skaffold &{Version:v1.23.0 ConfigVersion:skaffold/v2beta15 GitVersion: GitCommit:e8f3c652112c338e75e03497bc8ab09b9081142d BuildDate:2021-04-28T00:55:12Z GoVersion:go1.14.14 Compiler:gc Platform:darwin/amd64}
DEBU[0000] Disable housekeeping messages for command explicitly
DEBU[0000] apiVersion not present in yaml, continuing
DEBU[0000] apiVersion not present in yaml, continuing
DEBU[0000] apiVersion not present in yaml, continuing
DEBU[0000] apiVersion not present in yaml, continuing
DEBU[0000] apiVersion not present in yaml, continuing
one or more valid builder configuration (Dockerfile or Jib configuration) must be present to build images with skaffold; please provide at least one build config and try again or run `skaffold init --skip-build`
```
|
1.0
|
Detect .csproj files during init and leverage the Buildpack builder - <!--
Issues without logs and details are more complicated to fix.
Please help us by filling the template below!
-->
### Expected behavior
.net projects work with init when they dont have a Dockerfile
### Actual behavior
The init flow fails as it does not detect a proper builder
### Information
- Skaffold version: v1.23.0
- Operating system: MacOS
- Installed via: Cloud Code
- Contents of skaffold.yaml: None
### Steps to reproduce the behavior
1. Install .net Core https://dotnet.microsoft.com/download
2. Create a directory, cd to it and run `dotnet new mvc`
3. Try `skaffold init`
```
❯ skaffold init -vdebug
INFO[0000] Skaffold &{Version:v1.23.0 ConfigVersion:skaffold/v2beta15 GitVersion: GitCommit:e8f3c652112c338e75e03497bc8ab09b9081142d BuildDate:2021-04-28T00:55:12Z GoVersion:go1.14.14 Compiler:gc Platform:darwin/amd64}
DEBU[0000] Disable housekeeping messages for command explicitly
DEBU[0000] apiVersion not present in yaml, continuing
DEBU[0000] apiVersion not present in yaml, continuing
DEBU[0000] apiVersion not present in yaml, continuing
DEBU[0000] apiVersion not present in yaml, continuing
DEBU[0000] apiVersion not present in yaml, continuing
one or more valid builder configuration (Dockerfile or Jib configuration) must be present to build images with skaffold; please provide at least one build config and try again or run `skaffold init --skip-build`
```
|
non_process
|
detect csproj files during init and leverage the buildpack builder issues without logs and details are more complicated to fix please help us by filling the template below expected behavior net projects work with init when they dont have a dockerfile actual behavior the init flow fails as it does not detect a proper builder information skaffold version operating system macos installed via cloud code contents of skaffold yaml none steps to reproduce the behavior install net core create a directory cd to it and run dotnet new mvc try skaffold init ❯ skaffold init vdebug info skaffold version configversion skaffold gitversion gitcommit builddate goversion compiler gc platform darwin debu disable housekeeping messages for command explicitly debu apiversion not present in yaml continuing debu apiversion not present in yaml continuing debu apiversion not present in yaml continuing debu apiversion not present in yaml continuing debu apiversion not present in yaml continuing one or more valid builder configuration dockerfile or jib configuration must be present to build images with skaffold please provide at least one build config and try again or run skaffold init skip build
| 0
|
6,657
| 9,780,732,703
|
IssuesEvent
|
2019-06-07 17:45:23
|
googleapis/nodejs-pubsub
|
https://api.github.com/repos/googleapis/nodejs-pubsub
|
closed
|
lint/docs tests appear to be breaking
|
priority: p1 type: process
|
See:
https://github.com/googleapis/nodejs-pubsub/pull/641
https://github.com/googleapis/nodejs-pubsub/pull/639
It would be good to get these PRs green, so that we can star working through the backlog of open issues.
|
1.0
|
lint/docs tests appear to be breaking - See:
https://github.com/googleapis/nodejs-pubsub/pull/641
https://github.com/googleapis/nodejs-pubsub/pull/639
It would be good to get these PRs green, so that we can star working through the backlog of open issues.
|
process
|
lint docs tests appear to be breaking see it would be good to get these prs green so that we can star working through the backlog of open issues
| 1
|
18,507
| 24,551,353,595
|
IssuesEvent
|
2022-10-12 12:51:19
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Participants details screen > Consent history > Text case issue
|
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
Participants details screen > Consent history > Text case issue,
**AR:** Text in 'Data sharing permissions' is getting displayed as Not Applicable and Not Provided
**ER:** Text in 'Data sharing permissions' should get displayed as Not applicable and Not provided
**1.**

**2.**

|
3.0
|
[PM] Participants details screen > Consent history > Text case issue - Participants details screen > Consent history > Text case issue,
**AR:** Text in 'Data sharing permissions' is getting displayed as Not Applicable and Not Provided
**ER:** Text in 'Data sharing permissions' should get displayed as Not applicable and Not provided
**1.**

**2.**

|
process
|
participants details screen consent history text case issue participants details screen consent history text case issue ar text in data sharing permissions is getting displayed as not applicable and not provided er text in data sharing permissions should get displayed as not applicable and not provided
| 1
|
514,529
| 14,940,561,917
|
IssuesEvent
|
2021-01-25 18:27:31
|
oilshell/oil
|
https://api.github.com/repos/oilshell/oil
|
closed
|
simple word eval globbing bug
|
high-priority pending-release
|
```
$ bin/osh -O simple_word_eval -c 'shopt -p nullglob; for x in asdl/*.cc; do echo $x; done'
shopt -u nullglob
asdl/*.cc
$ bash -c 'shopt -p nullglob; for x in asdl/*.cc; do echo $x; done'
shopt -u nullglob
asdl/gc_test.cc
asdl/gen_cpp_test.cc
asdl/runtime.cc
```
|
1.0
|
simple word eval globbing bug - ```
$ bin/osh -O simple_word_eval -c 'shopt -p nullglob; for x in asdl/*.cc; do echo $x; done'
shopt -u nullglob
asdl/*.cc
$ bash -c 'shopt -p nullglob; for x in asdl/*.cc; do echo $x; done'
shopt -u nullglob
asdl/gc_test.cc
asdl/gen_cpp_test.cc
asdl/runtime.cc
```
|
non_process
|
simple word eval globbing bug bin osh o simple word eval c shopt p nullglob for x in asdl cc do echo x done shopt u nullglob asdl cc bash c shopt p nullglob for x in asdl cc do echo x done shopt u nullglob asdl gc test cc asdl gen cpp test cc asdl runtime cc
| 0
|
107,192
| 9,203,770,533
|
IssuesEvent
|
2019-03-08 04:05:29
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
opened
|
handling of webrtc fingerprinting changed
|
QA/Test-Plan-Specified QA/Yes regression
|
Found when testing https://github.com/brave/brave-browser/issues/3626
Handling of webrtc blocking changed from 0.60.x. Not sure if this is expected.
webrtc blocking is now more aggressive.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile
2. Open https://diafygi.github.io/webrtc-ips/
## Actual result:
<!--Please add screenshots if needed-->
From 0.61.49

## Expected result:
From 0.60.48

## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easily reproduced
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 0.61.49 Chromium: 73.0.3683.67 (Official Build) (64-bit)
-- | --
Revision | a83fd4f3207ae83412d329a9ca1239dd1e068345-refs/branch-heads/3683@{#760}
OS | Windows 7 Service Pack 1 Build 7601.24312
cc @brave/legacy_qa
|
1.0
|
handling of webrtc fingerprinting changed - Found when testing https://github.com/brave/brave-browser/issues/3626
Handling of webrtc blocking changed from 0.60.x. Not sure if this is expected.
webrtc blocking is now more aggressive.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile
2. Open https://diafygi.github.io/webrtc-ips/
## Actual result:
<!--Please add screenshots if needed-->
From 0.61.49

## Expected result:
From 0.60.48

## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easily reproduced
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 0.61.49 Chromium: 73.0.3683.67 (Official Build) (64-bit)
-- | --
Revision | a83fd4f3207ae83412d329a9ca1239dd1e068345-refs/branch-heads/3683@{#760}
OS | Windows 7 Service Pack 1 Build 7601.24312
cc @brave/legacy_qa
|
non_process
|
handling of webrtc fingerprinting changed found when testing handling of webrtc blocking changed from x not sure if this is expected webrtc blocking is now more aggressive steps to reproduce clean profile open actual result from expected result from reproduces how often easily reproduced brave version brave version info brave chromium official build bit revision refs branch heads os windows service pack build cc brave legacy qa
| 0
|
49,387
| 12,337,312,691
|
IssuesEvent
|
2020-05-14 14:49:25
|
grafana/grafana
|
https://api.github.com/repos/grafana/grafana
|
closed
|
Signing: manifest should allow missing files
|
type/build-packaging
|
Our current manifest check verifies that everything in the manifest exists, and fails if a file does not exist. In order to better support platform specific bundles, we should allow removing files from the bundle, but still having the signature work. For example:
```
{
"plugin": "grafana-googlesheets-datasource",
"version": "1.0.0-dev",
"files": {
"LICENSE": "7df059597099bb7dcf25d2a9aedfaf4465f72d8d",
"README.md": "08ec6d704b6115bef57710f6d7e866c050cb50ee",
"gfx_sheets_darwin_amd64": "1b8ae92c6e80e502bb0bf2d0ae9d7223805993ab",
"gfx_sheets_linux_amd64": "f39e0cc7344d3186b1052e6d356eecaf54d75b49",
"gfx_sheets_windows_amd64.exe": "c8825dfec512c1c235244f7998ee95182f9968de",
"module.js": "aaec6f51a995b7b843b843cd14041925274d960d",
"module.js.LICENSE.txt": "7f822fe9341af8f82ad1b0c69aba957822a377cf",
"module.js.map": "c5a524f5c4237f6ed6a016d43cd46938efeadb45",
"plugin.json": "55556b845e91935cc48fae3aa67baf0f22694c3f"
},
"time": 1586817677115,
"keyId": "7e4d0c6a708866e7"
}
```
Should not need the three binary copies (each 18MB)
This should be a simple change to:
https://github.com/grafana/grafana/blob/2661054fe8009704c55646c551e3bca7ed6ea6fe/pkg/plugins/manifest.go#L105
|
1.0
|
Signing: manifest should allow missing files - Our current manifest check verifies that everything in the manifest exists, and fails if a file does not exist. In order to better support platform specific bundles, we should allow removing files from the bundle, but still having the signature work. For example:
```
{
"plugin": "grafana-googlesheets-datasource",
"version": "1.0.0-dev",
"files": {
"LICENSE": "7df059597099bb7dcf25d2a9aedfaf4465f72d8d",
"README.md": "08ec6d704b6115bef57710f6d7e866c050cb50ee",
"gfx_sheets_darwin_amd64": "1b8ae92c6e80e502bb0bf2d0ae9d7223805993ab",
"gfx_sheets_linux_amd64": "f39e0cc7344d3186b1052e6d356eecaf54d75b49",
"gfx_sheets_windows_amd64.exe": "c8825dfec512c1c235244f7998ee95182f9968de",
"module.js": "aaec6f51a995b7b843b843cd14041925274d960d",
"module.js.LICENSE.txt": "7f822fe9341af8f82ad1b0c69aba957822a377cf",
"module.js.map": "c5a524f5c4237f6ed6a016d43cd46938efeadb45",
"plugin.json": "55556b845e91935cc48fae3aa67baf0f22694c3f"
},
"time": 1586817677115,
"keyId": "7e4d0c6a708866e7"
}
```
Should not need the three binary copies (each 18MB)
This should be a simple change to:
https://github.com/grafana/grafana/blob/2661054fe8009704c55646c551e3bca7ed6ea6fe/pkg/plugins/manifest.go#L105
|
non_process
|
signing manifest should allow missing files our current manifest check verifies that everything in the manifest exists and fails if a file does not exist in order to better support platform specific bundles we should allow removing files from the bundle but still having the signature work for example plugin grafana googlesheets datasource version dev files license readme md gfx sheets darwin gfx sheets linux gfx sheets windows exe module js module js license txt module js map plugin json time keyid should not need the three binary copies each this should be a simple change to
| 0
|
649,683
| 21,317,556,222
|
IssuesEvent
|
2022-04-16 14:52:10
|
DimensionDev/Maskbook
|
https://api.github.com/repos/DimensionDev/Maskbook
|
closed
|
[Bug] 'Extra Specific Contacts' should be an independent option to choose
|
Priority: P3 (Normal) Will Not Fix Type: Bug
|
## Bug Info
'Extra Specific Contacts' is not an independent option to choose.
### How to reproduce?
1.
2.
3.
### What happened?
### What should happen?
## Environment
### Mask Version: 1.32.3
- [x] (Browser) Installed from the extension store (version: )
- [ ] (App) Installed from the app store (version: )
- [ ] master branch
- [ ] Build of version/commit hash:
(Please write down build command/flags if there is anything special)
### System
<!-- Correct the version if it is not the case -->
- [ ] Windows 10
- [x] Mac OS X Catalina
- [ ] Linux: <!-- What distro? -->
### Platform/Browser
<!-- Correct the version if it is not the case -->
- [x] Latest Chrome (stable channel)
- [ ] Latest Firefox (stable channel)

https://user-images.githubusercontent.com/61523201/119637813-a0d26d80-be48-11eb-9fc2-1aeb22292fe7.mov
- [ ] Android 11
- [ ] iOS 14
|
1.0
|
[Bug] 'Extra Specific Contacts' should be an independent option to choose - ## Bug Info
'Extra Specific Contacts' is not an independent option to choose.
### How to reproduce?
1.
2.
3.
### What happened?
### What should happen?
## Environment
### Mask Version: 1.32.3
- [x] (Browser) Installed from the extension store (version: )
- [ ] (App) Installed from the app store (version: )
- [ ] master branch
- [ ] Build of version/commit hash:
(Please write down build command/flags if there is anything special)
### System
<!-- Correct the version if it is not the case -->
- [ ] Windows 10
- [x] Mac OS X Catalina
- [ ] Linux: <!-- What distro? -->
### Platform/Browser
<!-- Correct the version if it is not the case -->
- [x] Latest Chrome (stable channel)
- [ ] Latest Firefox (stable channel)

https://user-images.githubusercontent.com/61523201/119637813-a0d26d80-be48-11eb-9fc2-1aeb22292fe7.mov
- [ ] Android 11
- [ ] iOS 14
|
non_process
|
extra specific contacts should be an independent option to choose bug info extra specific contacts is not an independent option to choose how to reproduce what happened what should happen environment mask version browser installed from the extension store version app installed from the app store version master branch build of version commit hash please write down build command flags if there is anything special system windows mac os x catalina linux platform browser latest chrome stable channel latest firefox stable channel android ios
| 0
|
2,713
| 5,579,872,577
|
IssuesEvent
|
2017-03-28 15:27:00
|
openvstorage/framework-cinder-plugin
|
https://api.github.com/repos/openvstorage/framework-cinder-plugin
|
closed
|
Integrate Cinder with the Edge
|
process_wontfix type_feature
|
_From @wimpers on February 15, 2017 9:13_
In order to better support OpenStack, we want to integrate OpenStack Cinder with the OVS Edge.
OpenStack Libert only. Other version to be tackled in https://github.com/openvstorage/framework-cinder-plugin/issues/16.
Our of scope is getting CI back up and running.
_Copied from original issue: openvstorage/framework#1450_
|
1.0
|
Integrate Cinder with the Edge - _From @wimpers on February 15, 2017 9:13_
In order to better support OpenStack, we want to integrate OpenStack Cinder with the OVS Edge.
OpenStack Libert only. Other version to be tackled in https://github.com/openvstorage/framework-cinder-plugin/issues/16.
Our of scope is getting CI back up and running.
_Copied from original issue: openvstorage/framework#1450_
|
process
|
integrate cinder with the edge from wimpers on february in order to better support openstack we want to integrate openstack cinder with the ovs edge openstack libert only other version to be tackled in our of scope is getting ci back up and running copied from original issue openvstorage framework
| 1
|
125,571
| 17,836,373,751
|
IssuesEvent
|
2021-09-03 02:05:20
|
alpersonalwebsite/react-redux-example-redux-thunk
|
https://api.github.com/repos/alpersonalwebsite/react-redux-example-redux-thunk
|
opened
|
CVE-2021-23436 (Medium) detected in immer-1.10.0.tgz
|
security vulnerability
|
## CVE-2021-23436 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>immer-1.10.0.tgz</b></p></summary>
<p>Create your next immutable state by mutating the current one</p>
<p>Library home page: <a href="https://registry.npmjs.org/immer/-/immer-1.10.0.tgz">https://registry.npmjs.org/immer/-/immer-1.10.0.tgz</a></p>
<p>Path to dependency file: react-redux-example-redux-thunk/package.json</p>
<p>Path to vulnerable library: react-redux-example-redux-thunk/node_modules/immer/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- react-dev-utils-9.1.0.tgz
- :x: **immer-1.10.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package immer before 9.0.6. A type confusion vulnerability can lead to a bypass of CVE-2020-28477 when the user-provided keys used in the path parameter are arrays. In particular, this bypass is possible because the condition (p === "__proto__" || p === "constructor") in applyPatches_ returns false if p is ['__proto__'] (or ['constructor']). The === operator (strict equality operator) returns false if the operands have different type.
<p>Publish Date: 2021-09-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23436>CVE-2021-23436</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436</a></p>
<p>Release Date: 2021-09-01</p>
<p>Fix Resolution: immer - 9.0.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23436 (Medium) detected in immer-1.10.0.tgz - ## CVE-2021-23436 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>immer-1.10.0.tgz</b></p></summary>
<p>Create your next immutable state by mutating the current one</p>
<p>Library home page: <a href="https://registry.npmjs.org/immer/-/immer-1.10.0.tgz">https://registry.npmjs.org/immer/-/immer-1.10.0.tgz</a></p>
<p>Path to dependency file: react-redux-example-redux-thunk/package.json</p>
<p>Path to vulnerable library: react-redux-example-redux-thunk/node_modules/immer/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- react-dev-utils-9.1.0.tgz
- :x: **immer-1.10.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package immer before 9.0.6. A type confusion vulnerability can lead to a bypass of CVE-2020-28477 when the user-provided keys used in the path parameter are arrays. In particular, this bypass is possible because the condition (p === "__proto__" || p === "constructor") in applyPatches_ returns false if p is ['__proto__'] (or ['constructor']). The === operator (strict equality operator) returns false if the operands have different type.
<p>Publish Date: 2021-09-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23436>CVE-2021-23436</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436</a></p>
<p>Release Date: 2021-09-01</p>
<p>Fix Resolution: immer - 9.0.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in immer tgz cve medium severity vulnerability vulnerable library immer tgz create your next immutable state by mutating the current one library home page a href path to dependency file react redux example redux thunk package json path to vulnerable library react redux example redux thunk node modules immer package json dependency hierarchy react scripts tgz root library react dev utils tgz x immer tgz vulnerable library found in base branch master vulnerability details this affects the package immer before a type confusion vulnerability can lead to a bypass of cve when the user provided keys used in the path parameter are arrays in particular this bypass is possible because the condition p proto p constructor in applypatches returns false if p is or the operator strict equality operator returns false if the operands have different type publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution immer step up your open source security game with whitesource
| 0
|
10,057
| 13,044,161,746
|
IssuesEvent
|
2020-07-29 03:47:25
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `TimestampDiff` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `TimestampDiff` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `TimestampDiff` from TiDB -
## Description
Port the scalar function `TimestampDiff` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function timestampdiff from tidb description port the scalar function timestampdiff from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
51,012
| 6,484,600,251
|
IssuesEvent
|
2017-08-19 01:26:32
|
python-trio/trio
|
https://api.github.com/repos/python-trio/trio
|
closed
|
Design: how to handle async generator gc callbacks
|
design discussion
|
My inclination is to only register if it lets us print some useful message like "stop depending on the gc callbacks"? Not sure what this would look like exactly.
|
1.0
|
Design: how to handle async generator gc callbacks - My inclination is to only register if it lets us print some useful message like "stop depending on the gc callbacks"? Not sure what this would look like exactly.
|
non_process
|
design how to handle async generator gc callbacks my inclination is to only register if it lets us print some useful message like stop depending on the gc callbacks not sure what this would look like exactly
| 0
|
20,025
| 26,504,083,344
|
IssuesEvent
|
2023-01-18 12:34:44
|
MPMG-DCC-UFMG/C01
|
https://api.github.com/repos/MPMG-DCC-UFMG/C01
|
closed
|
Integração do form parser no mecanismo de processamento dinâmico
|
[1] Requisito [0] Desenvolvimento [2] Média Prioridade [3] Processamento Dinâmico
|
Para o mecanismo dos passos conseguir atomizar bastante os passos em mini-coletas, é necessário que ele saiba de antemão por onde seus "for"s estão percorrendo, assim ele pode executar cada possibilidade de valor para o iterador do for simultaneamente. Também temos que na maioria das vezes, os fors vão ser utilizados para percorrer opções de selects. Porém, nós só sabemos essas opções durante a execução dos passos, quando a função opcoe_em descobre eles. O mecanismo de form parser vem para preprocessar os passos, mudando os lugares onde tem o uso da função opcoes_em pela lista de opcoes daquele select, porém antes da execução, claro se isso for possível. Dessa forma, como sabemos por onde alguns "for"s, vão estar iterando, podemos rodas cada iteração desse for simultaneamente.
|
1.0
|
Integração do form parser no mecanismo de processamento dinâmico - Para o mecanismo dos passos conseguir atomizar bastante os passos em mini-coletas, é necessário que ele saiba de antemão por onde seus "for"s estão percorrendo, assim ele pode executar cada possibilidade de valor para o iterador do for simultaneamente. Também temos que na maioria das vezes, os fors vão ser utilizados para percorrer opções de selects. Porém, nós só sabemos essas opções durante a execução dos passos, quando a função opcoe_em descobre eles. O mecanismo de form parser vem para preprocessar os passos, mudando os lugares onde tem o uso da função opcoes_em pela lista de opcoes daquele select, porém antes da execução, claro se isso for possível. Dessa forma, como sabemos por onde alguns "for"s, vão estar iterando, podemos rodas cada iteração desse for simultaneamente.
|
process
|
integração do form parser no mecanismo de processamento dinâmico para o mecanismo dos passos conseguir atomizar bastante os passos em mini coletas é necessário que ele saiba de antemão por onde seus for s estão percorrendo assim ele pode executar cada possibilidade de valor para o iterador do for simultaneamente também temos que na maioria das vezes os fors vão ser utilizados para percorrer opções de selects porém nós só sabemos essas opções durante a execução dos passos quando a função opcoe em descobre eles o mecanismo de form parser vem para preprocessar os passos mudando os lugares onde tem o uso da função opcoes em pela lista de opcoes daquele select porém antes da execução claro se isso for possível dessa forma como sabemos por onde alguns for s vão estar iterando podemos rodas cada iteração desse for simultaneamente
| 1
|
6,534
| 9,634,116,191
|
IssuesEvent
|
2019-05-15 20:24:07
|
peteroas/Apprenticeship-Curriculum
|
https://api.github.com/repos/peteroas/Apprenticeship-Curriculum
|
closed
|
Agile at Revelry
|
process
|
Read this:
http://revelry.co/lean-agile/
Note: this document is descriptive, not prescriptive - we're always experimenting with ways to refine the way we work, so not every project will reflect this 100%.
Questions:
- What is a sprint?
- What is a user story? What are the parts of a user story?
- What are Acceptance Criteria? What are some examples?
- What is the function of a business rule?
- Why do we add sketches or screenshots?
- What is planning poker? Why do we do planning poker? When do we do planning
poker?
- How do we start new projects?
- Who is the Product Owner? What is their job?
- How do we "work" issues?
- What is QA? How is it different than UAT?
- What is a standup? What does an engineer do during the standup? What are some
things that _don't_ happen in a standup?
- What do we do at the end of a sprint?
- How would you handle an issue or project that starts to go bad?
- What is Waffle?
- Who is Lintron and what is his job?
- What other tools do we use?
|
1.0
|
Agile at Revelry -
Read this:
http://revelry.co/lean-agile/
Note: this document is descriptive, not prescriptive - we're always experimenting with ways to refine the way we work, so not every project will reflect this 100%.
Questions:
- What is a sprint?
- What is a user story? What are the parts of a user story?
- What are Acceptance Criteria? What are some examples?
- What is the function of a business rule?
- Why do we add sketches or screenshots?
- What is planning poker? Why do we do planning poker? When do we do planning
poker?
- How do we start new projects?
- Who is the Product Owner? What is their job?
- How do we "work" issues?
- What is QA? How is it different than UAT?
- What is a standup? What does an engineer do during the standup? What are some
things that _don't_ happen in a standup?
- What do we do at the end of a sprint?
- How would you handle an issue or project that starts to go bad?
- What is Waffle?
- Who is Lintron and what is his job?
- What other tools do we use?
|
process
|
agile at revelry read this note this document is descriptive not prescriptive we re always experimenting with ways to refine the way we work so not every project will reflect this questions what is a sprint what is a user story what are the parts of a user story what are acceptance criteria what are some examples what is the function of a business rule why do we add sketches or screenshots what is planning poker why do we do planning poker when do we do planning poker how do we start new projects who is the product owner what is their job how do we work issues what is qa how is it different than uat what is a standup what does an engineer do during the standup what are some things that don t happen in a standup what do we do at the end of a sprint how would you handle an issue or project that starts to go bad what is waffle who is lintron and what is his job what other tools do we use
| 1
|
3,196
| 2,743,617,076
|
IssuesEvent
|
2015-04-21 22:54:38
|
elastic/curator
|
https://api.github.com/repos/elastic/curator
|
closed
|
[DOC] Curator fails with 'missing snapshot' intermittently (solved)
|
cannot_replicate Documentation
|
I run a job daily which uses curator (in that order) to:
```
1. Takes snapshot all-indices
curator --timeout $CONNECT_TIMEOUT --host $ELASTICSEARCHSERVER snapshot --repository $REPONAME --request_timeout $SNAPSHOT_TIMEOUT indices --all-indices
2. Deletes snapshot older than 45 days
curator --timeout $CONNECT_TIMEOUT --host $ELASTICSEARCHSERVER delete snapshots --older-than $SNAPSHOTRETENTION --time-unit days --repository $REPONAME
3. Closes indices after 15 days
curator --timeout $CONNECT_TIMEOUT --host $ELASTICSEARCHSERVER close indices --older-than $CLOSERETENTION --time-unit days --timestring '%Y.%m.%d' --prefix "logstash-"
4. Deletes indices after 30 days (prefix "logstash")
curator --timeout $CONNECT_TIMEOUT --host $ELASTICSEARCHSERVER delete indices --older-than $LOGRETENTION --time-unit days --timestring '%Y.%m.%d' --prefix "logstash-"
(Snapshots on S3)
```
From time to time I see (2) fail with missing "SnapshotMissingException" but it does exist (as I see metadata-curator-[timestamp] and snapshot-curator-[timestamp])and it does delete it. My script terminates as the return code is not zero.
Here's the output log:
```
Now deleting snapshots ..
13:18:34 2015-03-30 13:18:34,228 INFO Job starting...
13:20:01 Traceback (most recent call last):
13:20:01 File "/usr/bin/curator", line 9, in <module>
13:20:01 load_entry_point('elasticsearch-curator==3.0.2', 'console_scripts', 'curator')()
13:20:01 File "/usr/lib/python2.6/site-packages/curator/curator.py", line 5, in main
13:20:01 cli( obj={ "filters": [] } )
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 610, in __call__
13:20:01 return self.main(*args, **kwargs)
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 590, in main
13:20:01 rv = self.invoke(ctx)
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 936, in invoke
13:20:01 return _process_result(sub_ctx.command.invoke(sub_ctx))
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 936, in invoke
13:20:01 return _process_result(sub_ctx.command.invoke(sub_ctx))
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 782, in invoke
13:20:01 return ctx.invoke(self.callback, **ctx.params)
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 416, in invoke
13:20:01 return callback(*args, **kwargs)
13:20:01 File "/usr/lib/python2.6/site-packages/curator/cli/snapshot_selection.py", line 91, in snapshots
13:20:01 retval = delete_snapshot(client, snapshot=snap, repository=repository)
13:20:01 File "/usr/lib/python2.6/site-packages/curator/api/snapshot.py", line 87, in delete_snapshot
13:20:01 client.snapshot.delete(repository=repository, snapshot=snapshot)
13:20:01 File "/usr/lib/python2.6/site-packages/elasticsearch/client/utils.py", line 68, in _wrapped
13:20:01 return func(*args, params=params, **kwargs)
13:20:01 File "/usr/lib/python2.6/site-packages/elasticsearch/client/snapshot.py", line 40, in delete
13:20:01 _make_path('_snapshot', repository, snapshot), params=params)
13:20:01 File "/usr/lib/python2.6/site-packages/elasticsearch/transport.py", line 301, in perform_request
13:20:01 status, headers, data = connection.perform_request(method, url, params, body, ignore=ignore, timeout=timeout)
13:20:01 File "/usr/lib/python2.6/site-packages/elasticsearch/connection/http_urllib3.py", line 82, in perform_request
13:20:01 self._raise_error(response.status, raw_data)
13:20:01 File "/usr/lib/python2.6/site-packages/elasticsearch/connection/base.py", line 102, in _raise_error
13:20:01 raise HTTP_EXCEPTIONS.get(status_code, TransportError)(status_code, error_message, additional_info)
13:20:01 elasticsearch.exceptions.NotFoundError: TransportError(404, u'RemoteTransportException[[Lemuel Dorcas][inet[/10.11.102.53:9300]][cluster/snapshot/delete]]; nested: SnapshotMissingException[[my_s3_repository:curator-20150213001500] is missing]; nested: FileNotFoundException[The specified key does not exist. (Service: Amazon S3; Status Code: 404; Error Code: NoSuchKey; Request ID: 5531B70DBC6BD25B)]; ')
13:20:01 Unable to delete snapshots ..
13:20:01 Result: 1
```
|
1.0
|
[DOC] Curator fails with 'missing snapshot' intermittently (solved) - I run a job daily which uses curator (in that order) to:
```
1. Takes snapshot all-indices
curator --timeout $CONNECT_TIMEOUT --host $ELASTICSEARCHSERVER snapshot --repository $REPONAME --request_timeout $SNAPSHOT_TIMEOUT indices --all-indices
2. Deletes snapshot older than 45 days
curator --timeout $CONNECT_TIMEOUT --host $ELASTICSEARCHSERVER delete snapshots --older-than $SNAPSHOTRETENTION --time-unit days --repository $REPONAME
3. Closes indices after 15 days
curator --timeout $CONNECT_TIMEOUT --host $ELASTICSEARCHSERVER close indices --older-than $CLOSERETENTION --time-unit days --timestring '%Y.%m.%d' --prefix "logstash-"
4. Deletes indices after 30 days (prefix "logstash")
curator --timeout $CONNECT_TIMEOUT --host $ELASTICSEARCHSERVER delete indices --older-than $LOGRETENTION --time-unit days --timestring '%Y.%m.%d' --prefix "logstash-"
(Snapshots on S3)
```
From time to time I see (2) fail with missing "SnapshotMissingException" but it does exist (as I see metadata-curator-[timestamp] and snapshot-curator-[timestamp])and it does delete it. My script terminates as the return code is not zero.
Here's the output log:
```
Now deleting snapshots ..
13:18:34 2015-03-30 13:18:34,228 INFO Job starting...
13:20:01 Traceback (most recent call last):
13:20:01 File "/usr/bin/curator", line 9, in <module>
13:20:01 load_entry_point('elasticsearch-curator==3.0.2', 'console_scripts', 'curator')()
13:20:01 File "/usr/lib/python2.6/site-packages/curator/curator.py", line 5, in main
13:20:01 cli( obj={ "filters": [] } )
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 610, in __call__
13:20:01 return self.main(*args, **kwargs)
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 590, in main
13:20:01 rv = self.invoke(ctx)
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 936, in invoke
13:20:01 return _process_result(sub_ctx.command.invoke(sub_ctx))
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 936, in invoke
13:20:01 return _process_result(sub_ctx.command.invoke(sub_ctx))
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 782, in invoke
13:20:01 return ctx.invoke(self.callback, **ctx.params)
13:20:01 File "/usr/lib/python2.6/site-packages/click/core.py", line 416, in invoke
13:20:01 return callback(*args, **kwargs)
13:20:01 File "/usr/lib/python2.6/site-packages/curator/cli/snapshot_selection.py", line 91, in snapshots
13:20:01 retval = delete_snapshot(client, snapshot=snap, repository=repository)
13:20:01 File "/usr/lib/python2.6/site-packages/curator/api/snapshot.py", line 87, in delete_snapshot
13:20:01 client.snapshot.delete(repository=repository, snapshot=snapshot)
13:20:01 File "/usr/lib/python2.6/site-packages/elasticsearch/client/utils.py", line 68, in _wrapped
13:20:01 return func(*args, params=params, **kwargs)
13:20:01 File "/usr/lib/python2.6/site-packages/elasticsearch/client/snapshot.py", line 40, in delete
13:20:01 _make_path('_snapshot', repository, snapshot), params=params)
13:20:01 File "/usr/lib/python2.6/site-packages/elasticsearch/transport.py", line 301, in perform_request
13:20:01 status, headers, data = connection.perform_request(method, url, params, body, ignore=ignore, timeout=timeout)
13:20:01 File "/usr/lib/python2.6/site-packages/elasticsearch/connection/http_urllib3.py", line 82, in perform_request
13:20:01 self._raise_error(response.status, raw_data)
13:20:01 File "/usr/lib/python2.6/site-packages/elasticsearch/connection/base.py", line 102, in _raise_error
13:20:01 raise HTTP_EXCEPTIONS.get(status_code, TransportError)(status_code, error_message, additional_info)
13:20:01 elasticsearch.exceptions.NotFoundError: TransportError(404, u'RemoteTransportException[[Lemuel Dorcas][inet[/10.11.102.53:9300]][cluster/snapshot/delete]]; nested: SnapshotMissingException[[my_s3_repository:curator-20150213001500] is missing]; nested: FileNotFoundException[The specified key does not exist. (Service: Amazon S3; Status Code: 404; Error Code: NoSuchKey; Request ID: 5531B70DBC6BD25B)]; ')
13:20:01 Unable to delete snapshots ..
13:20:01 Result: 1
```
|
non_process
|
curator fails with missing snapshot intermittently solved i run a job daily which uses curator in that order to takes snapshot all indices curator timeout connect timeout host elasticsearchserver snapshot repository reponame request timeout snapshot timeout indices all indices deletes snapshot older than days curator timeout connect timeout host elasticsearchserver delete snapshots older than snapshotretention time unit days repository reponame closes indices after days curator timeout connect timeout host elasticsearchserver close indices older than closeretention time unit days timestring y m d prefix logstash deletes indices after days prefix logstash curator timeout connect timeout host elasticsearchserver delete indices older than logretention time unit days timestring y m d prefix logstash snapshots on from time to time i see fail with missing snapshotmissingexception but it does exist as i see metadata curator and snapshot curator and it does delete it my script terminates as the return code is not zero here s the output log now deleting snapshots info job starting traceback most recent call last file usr bin curator line in load entry point elasticsearch curator console scripts curator file usr lib site packages curator curator py line in main cli obj filters file usr lib site packages click core py line in call return self main args kwargs file usr lib site packages click core py line in main rv self invoke ctx file usr lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file usr lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file usr lib site packages click core py line in invoke return ctx invoke self callback ctx params file usr lib site packages click core py line in invoke return callback args kwargs file usr lib site packages curator cli snapshot selection py line in snapshots retval delete snapshot client snapshot snap repository repository file usr lib site packages curator api snapshot py line in delete snapshot client snapshot delete repository repository snapshot snapshot file usr lib site packages elasticsearch client utils py line in wrapped return func args params params kwargs file usr lib site packages elasticsearch client snapshot py line in delete make path snapshot repository snapshot params params file usr lib site packages elasticsearch transport py line in perform request status headers data connection perform request method url params body ignore ignore timeout timeout file usr lib site packages elasticsearch connection http py line in perform request self raise error response status raw data file usr lib site packages elasticsearch connection base py line in raise error raise http exceptions get status code transporterror status code error message additional info elasticsearch exceptions notfounderror transporterror u remotetransportexception nested snapshotmissingexception is missing nested filenotfoundexception unable to delete snapshots result
| 0
|
21,658
| 30,107,268,421
|
IssuesEvent
|
2023-06-30 03:14:10
|
diffgram/diffgram
|
https://api.github.com/repos/diffgram/diffgram
|
closed
|
Process Media Queue Pressure Refactor
|
refactor process_media
|
We have `wait_until_queue_pressure_is_lower` and we also have `check_if_add_items_to_queue()` with `add_deferred_items_time`.
What doesn't really make sense is that it should never get to FRAME_QUEUE.qsize() >= (limit defined in wait_until_queue_pressure_is_lower).
If it does, it could indicate that processing is stuck.
But, at that point it's awkward to log useful information.
next steps here is probably to align `check_if_add_items_to_queue()` to assumptions around usage of `wait_until_queue_pressure_is_lower()` and generally refactor.
For example:
these blocks could easily be their own function
```
try:
update_input = Update_Input(session = session).automatic_retry()
except Exception as exception:
logger.error(f"Couldn't find Update_Input {str(exception)}")
return 30
```
```
input = session.query(Input).with_for_update(skip_locked = True).filter(
Input.processing_deferred == True,
Input.archived == False,
Input.status != 'success'
).first()
```
|
1.0
|
Process Media Queue Pressure Refactor - We have `wait_until_queue_pressure_is_lower` and we also have `check_if_add_items_to_queue()` with `add_deferred_items_time`.
What doesn't really make sense is that it should never get to FRAME_QUEUE.qsize() >= (limit defined in wait_until_queue_pressure_is_lower).
If it does, it could indicate that processing is stuck.
But, at that point it's awkward to log useful information.
next steps here is probably to align `check_if_add_items_to_queue()` to assumptions around usage of `wait_until_queue_pressure_is_lower()` and generally refactor.
For example:
these blocks could easily be their own function
```
try:
update_input = Update_Input(session = session).automatic_retry()
except Exception as exception:
logger.error(f"Couldn't find Update_Input {str(exception)}")
return 30
```
```
input = session.query(Input).with_for_update(skip_locked = True).filter(
Input.processing_deferred == True,
Input.archived == False,
Input.status != 'success'
).first()
```
|
process
|
process media queue pressure refactor we have wait until queue pressure is lower and we also have check if add items to queue with add deferred items time what doesn t really make sense is that it should never get to frame queue qsize limit defined in wait until queue pressure is lower if it does it could indicate that processing is stuck but at that point it s awkward to log useful information next steps here is probably to align check if add items to queue to assumptions around usage of wait until queue pressure is lower and generally refactor for example these blocks could easily be their own function try update input update input session session automatic retry except exception as exception logger error f couldn t find update input str exception return input session query input with for update skip locked true filter input processing deferred true input archived false input status success first
| 1
|
251,855
| 27,218,108,461
|
IssuesEvent
|
2023-02-21 01:06:17
|
rvvergara/react-boilerplate
|
https://api.github.com/repos/rvvergara/react-boilerplate
|
opened
|
CVE-2022-4904 (Medium) detected in nodev10.14.0
|
security vulnerability
|
## CVE-2022-4904 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nodev10.14.0</b></p></summary>
<p>
<p>Node.js JavaScript runtime :sparkles::turtle::rocket::sparkles:</p>
<p>Library home page: <a href=https://github.com/nodejs/node.git>https://github.com/nodejs/node.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/node_modules/grpc/deps/grpc/third_party/cares/cares/ares_init.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
CVE-2022-4904 c-ares: buffer overflow in config_sortlist() due to missing string length check
<p>Publish Date: 2023-02-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-4904>CVE-2022-4904</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2023-02-13</p>
<p>Fix Resolution: cares-1_19_0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-4904 (Medium) detected in nodev10.14.0 - ## CVE-2022-4904 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nodev10.14.0</b></p></summary>
<p>
<p>Node.js JavaScript runtime :sparkles::turtle::rocket::sparkles:</p>
<p>Library home page: <a href=https://github.com/nodejs/node.git>https://github.com/nodejs/node.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/node_modules/grpc/deps/grpc/third_party/cares/cares/ares_init.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
CVE-2022-4904 c-ares: buffer overflow in config_sortlist() due to missing string length check
<p>Publish Date: 2023-02-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-4904>CVE-2022-4904</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2023-02-13</p>
<p>Fix Resolution: cares-1_19_0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in cve medium severity vulnerability vulnerable library node js javascript runtime sparkles turtle rocket sparkles library home page a href found in base branch master vulnerable source files node modules grpc deps grpc third party cares cares ares init c vulnerability details cve c ares buffer overflow in config sortlist due to missing string length check publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution cares step up your open source security game with mend
| 0
|
24,855
| 7,573,435,781
|
IssuesEvent
|
2018-04-23 17:46:12
|
openshiftio/openshift.io
|
https://api.github.com/repos/openshiftio/openshift.io
|
opened
|
Importing existing codebase with new launcher creates 2 build configs
|
area/app-wizard area/pipelines team/build-cd team/launcher type/bug
|
When using the new launcher experience to create an application by importing an existing codebase, I ended up with 2 new build configs. This seems to be due to the application name (app-test-1) differing from the Maven artifact ID (mynewvertx) in the repository. I believe the former is created by the launcher and the latter is created by Jenkins.
Steps to reproduce:
1. Using new launcher experience, select "Import an existing codebase".
2. Either choose a GitHub repository with Maven artifact ID different from the application name, or change the application name in the wizard to something else.

3. Wait for pipeline to start, observe 2 build configs (one named after application, one named after Maven artifact ID)

4. Application is built through pipeline twice in build config named after artifact ID:

5. Application is finally built through pipeline in build config named after the application:

Some consequences of this are that only the build config created by the launcher (named after the application) has the appropriate space label attached to it. Thus, this is the only pipeline that is reported to the user in the OSIO pipelines page.
|
1.0
|
Importing existing codebase with new launcher creates 2 build configs - When using the new launcher experience to create an application by importing an existing codebase, I ended up with 2 new build configs. This seems to be due to the application name (app-test-1) differing from the Maven artifact ID (mynewvertx) in the repository. I believe the former is created by the launcher and the latter is created by Jenkins.
Steps to reproduce:
1. Using new launcher experience, select "Import an existing codebase".
2. Either choose a GitHub repository with Maven artifact ID different from the application name, or change the application name in the wizard to something else.

3. Wait for pipeline to start, observe 2 build configs (one named after application, one named after Maven artifact ID)

4. Application is built through pipeline twice in build config named after artifact ID:

5. Application is finally built through pipeline in build config named after the application:

Some consequences of this are that only the build config created by the launcher (named after the application) has the appropriate space label attached to it. Thus, this is the only pipeline that is reported to the user in the OSIO pipelines page.
|
non_process
|
importing existing codebase with new launcher creates build configs when using the new launcher experience to create an application by importing an existing codebase i ended up with new build configs this seems to be due to the application name app test differing from the maven artifact id mynewvertx in the repository i believe the former is created by the launcher and the latter is created by jenkins steps to reproduce using new launcher experience select import an existing codebase either choose a github repository with maven artifact id different from the application name or change the application name in the wizard to something else wait for pipeline to start observe build configs one named after application one named after maven artifact id application is built through pipeline twice in build config named after artifact id application is finally built through pipeline in build config named after the application some consequences of this are that only the build config created by the launcher named after the application has the appropriate space label attached to it thus this is the only pipeline that is reported to the user in the osio pipelines page
| 0
|
22,196
| 30,751,836,367
|
IssuesEvent
|
2023-07-28 20:04:08
|
streamnative/pulsar-spark
|
https://api.github.com/repos/streamnative/pulsar-spark
|
closed
|
[FEATURE] support pyspark
|
type/feature compute/data-processing
|
**Is your feature request related to a problem? Please describe.**
It seems that for now only java version is supported for spark. I'm using pyspark, and when I try 2 use standard pulsar client to send data to pulsar, it returns error saying `PicklingError: Could not serialize object: RuntimeError: Pickling of "pulsar.Producer" instances is not enabled`
**Describe the solution you'd like**
also support pyspark
|
1.0
|
[FEATURE] support pyspark - **Is your feature request related to a problem? Please describe.**
It seems that for now only java version is supported for spark. I'm using pyspark, and when I try 2 use standard pulsar client to send data to pulsar, it returns error saying `PicklingError: Could not serialize object: RuntimeError: Pickling of "pulsar.Producer" instances is not enabled`
**Describe the solution you'd like**
also support pyspark
|
process
|
support pyspark is your feature request related to a problem please describe it seems that for now only java version is supported for spark i m using pyspark and when i try use standard pulsar client to send data to pulsar it returns error saying picklingerror could not serialize object runtimeerror pickling of pulsar producer instances is not enabled describe the solution you d like also support pyspark
| 1
|
119
| 3,835,665,963
|
IssuesEvent
|
2016-04-01 15:07:35
|
DevExpress/testcafe
|
https://api.github.com/repos/DevExpress/testcafe
|
opened
|
Active iframe window doesn't consider itself as active if it was reloaded
|
AREA: client SYSTEM: automations TYPE: bug
|
When an iframe is reloaded by changing location or submitting a from, it's `contentWindow` will be reinitialized, but it will be the same window object from top page view (`iframe.contentWindow` before reload === `iframe.contentWindow` after reload). So after iframe reloading it will be still considered as active by top window, but doesn't be considered as active from inside. It leads to infinite activation loop when trying to perform a click in this iframe.
|
1.0
|
Active iframe window doesn't consider itself as active if it was reloaded - When an iframe is reloaded by changing location or submitting a from, it's `contentWindow` will be reinitialized, but it will be the same window object from top page view (`iframe.contentWindow` before reload === `iframe.contentWindow` after reload). So after iframe reloading it will be still considered as active by top window, but doesn't be considered as active from inside. It leads to infinite activation loop when trying to perform a click in this iframe.
|
non_process
|
active iframe window doesn t consider itself as active if it was reloaded when an iframe is reloaded by changing location or submitting a from it s contentwindow will be reinitialized but it will be the same window object from top page view iframe contentwindow before reload iframe contentwindow after reload so after iframe reloading it will be still considered as active by top window but doesn t be considered as active from inside it leads to infinite activation loop when trying to perform a click in this iframe
| 0
|
375,781
| 11,134,575,521
|
IssuesEvent
|
2019-12-20 12:10:43
|
jenkins-x/jx
|
https://api.github.com/repos/jenkins-x/jx
|
closed
|
users.jenkins.io "<user>" already exists
|
area/git good first issue in progress kind/bug priority/important-longterm
|
### Summary
Creating a preview can fail with:
```
users.jenkins.io "foo" already exists
```
This happens when there is already a User CRD with the name `foo` and `GitUserResolver.Resolve` fails tries to create a new User CRD with the same name.
|
1.0
|
users.jenkins.io "<user>" already exists - ### Summary
Creating a preview can fail with:
```
users.jenkins.io "foo" already exists
```
This happens when there is already a User CRD with the name `foo` and `GitUserResolver.Resolve` fails tries to create a new User CRD with the same name.
|
non_process
|
users jenkins io already exists summary creating a preview can fail with users jenkins io foo already exists this happens when there is already a user crd with the name foo and gituserresolver resolve fails tries to create a new user crd with the same name
| 0
|
8,198
| 11,394,979,897
|
IssuesEvent
|
2020-01-30 10:27:15
|
prisma/prisma2
|
https://api.github.com/repos/prisma/prisma2
|
opened
|
Init flow should default to using environment variables for connection URLs
|
process/candidate topic: cli-init
|
`prisma2 init` should be adjusted in the following way:
- Additionally generate a `.env` file with the connection url comments moved
```
# Other examples for connection strings are:
# SQLite: url = "sqlite:./dev.db"
# MySQL: url = "mysql://johndoe:johndoe@localhost:3306/mydb"
# You can also use environment variables to specify the connection string: https://pris.ly/prisma-schema#using-environment-variables
DATABASE_URL="postgresql://schickling@localhost:5432/schickling?schema=public"
```
- Use `DATABASE_URL` in the `schema.prisma`
```prisma
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
```
|
1.0
|
Init flow should default to using environment variables for connection URLs - `prisma2 init` should be adjusted in the following way:
- Additionally generate a `.env` file with the connection url comments moved
```
# Other examples for connection strings are:
# SQLite: url = "sqlite:./dev.db"
# MySQL: url = "mysql://johndoe:johndoe@localhost:3306/mydb"
# You can also use environment variables to specify the connection string: https://pris.ly/prisma-schema#using-environment-variables
DATABASE_URL="postgresql://schickling@localhost:5432/schickling?schema=public"
```
- Use `DATABASE_URL` in the `schema.prisma`
```prisma
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
```
|
process
|
init flow should default to using environment variables for connection urls init should be adjusted in the following way additionally generate a env file with the connection url comments moved other examples for connection strings are sqlite url sqlite dev db mysql url mysql johndoe johndoe localhost mydb you can also use environment variables to specify the connection string database url postgresql schickling localhost schickling schema public use database url in the schema prisma prisma datasource db provider postgresql url env database url
| 1
|
186,668
| 6,741,810,805
|
IssuesEvent
|
2017-10-20 03:27:15
|
opencurrents/opencurrents
|
https://api.github.com/repos/opencurrents/opencurrents
|
opened
|
Create Event: When assigning another "volunteer admin" to event there is no email being sent.
|
priority high priority medium
|
Let's discuss if we want to keep this feature. If so we would need to add conditional logic that it is only possible if the person is affiliated with the Org.
|
2.0
|
Create Event: When assigning another "volunteer admin" to event there is no email being sent. - Let's discuss if we want to keep this feature. If so we would need to add conditional logic that it is only possible if the person is affiliated with the Org.
|
non_process
|
create event when assigning another volunteer admin to event there is no email being sent let s discuss if we want to keep this feature if so we would need to add conditional logic that it is only possible if the person is affiliated with the org
| 0
|
673,376
| 22,960,171,919
|
IssuesEvent
|
2022-07-19 14:46:44
|
project-chip/zap
|
https://api.github.com/repos/project-chip/zap
|
closed
|
Need to support fabric-scoped commands
|
matter high priority
|
We need to be able to mark commands as fabric-scoped in the XML.
The relevant pieces are:
1. Have `isFabricScoped="true"` in the XML on the relevant commands.
2. Add a field to the COMMAND table to store the value (similar to MUST_USE_TIMED_INVOKE).
3. Fix XML import to set the COMMAND field based on the XML.
4. Allow some way to do things with fabric-scoped commands via helpers. Either expose the state in the data `zcl_commands` fetches, or have a custom helper like `zcl_commands_that_need_timed_invoke`; no really strong opinions here, though the `zcl_commands` approach seems a bit more flexible.
|
1.0
|
Need to support fabric-scoped commands - We need to be able to mark commands as fabric-scoped in the XML.
The relevant pieces are:
1. Have `isFabricScoped="true"` in the XML on the relevant commands.
2. Add a field to the COMMAND table to store the value (similar to MUST_USE_TIMED_INVOKE).
3. Fix XML import to set the COMMAND field based on the XML.
4. Allow some way to do things with fabric-scoped commands via helpers. Either expose the state in the data `zcl_commands` fetches, or have a custom helper like `zcl_commands_that_need_timed_invoke`; no really strong opinions here, though the `zcl_commands` approach seems a bit more flexible.
|
non_process
|
need to support fabric scoped commands we need to be able to mark commands as fabric scoped in the xml the relevant pieces are have isfabricscoped true in the xml on the relevant commands add a field to the command table to store the value similar to must use timed invoke fix xml import to set the command field based on the xml allow some way to do things with fabric scoped commands via helpers either expose the state in the data zcl commands fetches or have a custom helper like zcl commands that need timed invoke no really strong opinions here though the zcl commands approach seems a bit more flexible
| 0
|
10,416
| 13,208,915,978
|
IssuesEvent
|
2020-08-15 08:03:51
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
broken point ("cursor") movement when (eval)ing inside `ess-r-transcript-mode`
|
bug process:eval
|
In previous versions of ESS, navigating *and* evaluation R code was phantastically simple, thanks to our `ess-r-transcript-mode` which is turned on automagically, e.g., for files ending in `*.Rout` (including those `*.Rout.save` ones, known from base R and package testing).
Let's take one R's own files, download it locally to open in emacs (om a decent OS, in the shell/terminal:)
` wget https://svn.r-project.org/R/trunk/tests/reg-tests-2.Rout.save
`
If you open in it in an emacs with a correctly working ESS, I use ESS 18.10.2, also starting R, e.g. by
`
emacs -f R <your_path_to>/reg-tests-2.Rout.save
`
and go to the first prompt, e.g., by `C-s Reg` you can press `[Enter]` repeatedly to evaluate things to the `*R*` buffer. At the beginning this also works nicely with current ESS (from github, MELPA, ..).
Things start differing between correctly working and current ESS dramatically, once you get to line 29 (of `reg-tests-2.Rout.save`): In current ESS, on `[Enter]` there, point (aka "the cursor") just goes one line further (line 30), which is a "+ ...." continuation line of a multi-line evaluation, I'm appending a small screen shot (using `linum-mode` to show line numbers).

whereas in ESS 18.10.2 (and *all* earlier versions of ESS), it nicely jumps down to line 69 which is exactly where the next prompt is ... so you can continue the evaluations by continuing with `[Enter]`.
Probably much related to the similarly "new" bugous behavior in issues #956 and #973 .
All of these are probably related to prompt detection, and all of these were working correctly in all released versions of ESS and also for quite some time after releasing ESS 18.10.2 (2018-11-10).
|
1.0
|
broken point ("cursor") movement when (eval)ing inside `ess-r-transcript-mode` - In previous versions of ESS, navigating *and* evaluation R code was phantastically simple, thanks to our `ess-r-transcript-mode` which is turned on automagically, e.g., for files ending in `*.Rout` (including those `*.Rout.save` ones, known from base R and package testing).
Let's take one R's own files, download it locally to open in emacs (om a decent OS, in the shell/terminal:)
` wget https://svn.r-project.org/R/trunk/tests/reg-tests-2.Rout.save
`
If you open in it in an emacs with a correctly working ESS, I use ESS 18.10.2, also starting R, e.g. by
`
emacs -f R <your_path_to>/reg-tests-2.Rout.save
`
and go to the first prompt, e.g., by `C-s Reg` you can press `[Enter]` repeatedly to evaluate things to the `*R*` buffer. At the beginning this also works nicely with current ESS (from github, MELPA, ..).
Things start differing between correctly working and current ESS dramatically, once you get to line 29 (of `reg-tests-2.Rout.save`): In current ESS, on `[Enter]` there, point (aka "the cursor") just goes one line further (line 30), which is a "+ ...." continuation line of a multi-line evaluation, I'm appending a small screen shot (using `linum-mode` to show line numbers).

whereas in ESS 18.10.2 (and *all* earlier versions of ESS), it nicely jumps down to line 69 which is exactly where the next prompt is ... so you can continue the evaluations by continuing with `[Enter]`.
Probably much related to the similarly "new" bugous behavior in issues #956 and #973 .
All of these are probably related to prompt detection, and all of these were working correctly in all released versions of ESS and also for quite some time after releasing ESS 18.10.2 (2018-11-10).
|
process
|
broken point cursor movement when eval ing inside ess r transcript mode in previous versions of ess navigating and evaluation r code was phantastically simple thanks to our ess r transcript mode which is turned on automagically e g for files ending in rout including those rout save ones known from base r and package testing let s take one r s own files download it locally to open in emacs om a decent os in the shell terminal wget if you open in it in an emacs with a correctly working ess i use ess also starting r e g by emacs f r reg tests rout save and go to the first prompt e g by c s reg you can press repeatedly to evaluate things to the r buffer at the beginning this also works nicely with current ess from github melpa things start differing between correctly working and current ess dramatically once you get to line of reg tests rout save in current ess on there point aka the cursor just goes one line further line which is a continuation line of a multi line evaluation i m appending a small screen shot using linum mode to show line numbers whereas in ess and all earlier versions of ess it nicely jumps down to line which is exactly where the next prompt is so you can continue the evaluations by continuing with probably much related to the similarly new bugous behavior in issues and all of these are probably related to prompt detection and all of these were working correctly in all released versions of ess and also for quite some time after releasing ess
| 1
|
292,830
| 22,037,327,413
|
IssuesEvent
|
2022-05-28 20:01:51
|
Interlisp/medley
|
https://api.github.com/repos/Interlisp/medley
|
closed
|
Opening a shell in Interlisp which runs in a container
|
bug documentation config
|
I'm not sure what (CHAT 'SHELL) should do when running in a docker container.
(CHAT "hostname") should run "ssh" without having to set up a tunnel to the server -- should we set up the ssh tunnel?
Also -- using docker with X rather than VNC -- is that important?
@fghalasz @stumbo what do you think?
### Discussed in https://github.com/Interlisp/medley/discussions/688
<div type='discussions-op-text'>
<sup>Originally posted by **reflektoin** February 20, 2022</sup>
**EDIT: This isn't really an issue for me at the moment. I have started to use the online version**
Hi, Thank you for this project! By following the [Docker guide here](https://github.com/Interlisp/medley/wiki/Running-with-Docker), I was able to get Interlisp running easily.
But I got stuck when I tried to open a terminal in Interlisp with command (IL:CHAT 'SHELL). I got a message
`[Connection closed by remote host]`
My current setup is that I'm running Interlisp Docker container on a Ubuntu server. From my own computer I create ssh-tunnel to the server and use either SSVNC or xtightvnc to connect to Interlisp.
Does someone have tips how to proceed debugging this?
And thanks again for making this!</div>
|
1.0
|
Opening a shell in Interlisp which runs in a container - I'm not sure what (CHAT 'SHELL) should do when running in a docker container.
(CHAT "hostname") should run "ssh" without having to set up a tunnel to the server -- should we set up the ssh tunnel?
Also -- using docker with X rather than VNC -- is that important?
@fghalasz @stumbo what do you think?
### Discussed in https://github.com/Interlisp/medley/discussions/688
<div type='discussions-op-text'>
<sup>Originally posted by **reflektoin** February 20, 2022</sup>
**EDIT: This isn't really an issue for me at the moment. I have started to use the online version**
Hi, Thank you for this project! By following the [Docker guide here](https://github.com/Interlisp/medley/wiki/Running-with-Docker), I was able to get Interlisp running easily.
But I got stuck when I tried to open a terminal in Interlisp with command (IL:CHAT 'SHELL). I got a message
`[Connection closed by remote host]`
My current setup is that I'm running Interlisp Docker container on a Ubuntu server. From my own computer I create ssh-tunnel to the server and use either SSVNC or xtightvnc to connect to Interlisp.
Does someone have tips how to proceed debugging this?
And thanks again for making this!</div>
|
non_process
|
opening a shell in interlisp which runs in a container i m not sure what chat shell should do when running in a docker container chat hostname should run ssh without having to set up a tunnel to the server should we set up the ssh tunnel also using docker with x rather than vnc is that important fghalasz stumbo what do you think discussed in originally posted by reflektoin february edit this isn t really an issue for me at the moment i have started to use the online version hi thank you for this project by following the i was able to get interlisp running easily but i got stuck when i tried to open a terminal in interlisp with command il chat shell i got a message my current setup is that i m running interlisp docker container on a ubuntu server from my own computer i create ssh tunnel to the server and use either ssvnc or xtightvnc to connect to interlisp does someone have tips how to proceed debugging this and thanks again for making this
| 0
|
4,560
| 7,389,881,078
|
IssuesEvent
|
2018-03-16 10:17:39
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
.NET Core support?
|
app-service cxp in-process product-question triaged
|
It looks like this is only available for the .NET Framework. Are there are any resources available for .NET Core?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7c1407af-81a5-52e7-a800-79d57704a876
* Version Independent ID: 5ec30406-731f-6198-99a1-9d4ee5c67475
* Content: [Develop and deploy WebJobs using Visual Studio - Azure | Microsoft Docs](https://docs.microsoft.com/en-us/azure/app-service/websites-dotnet-deploy-webjobs)
* Content Source: [articles/app-service/websites-dotnet-deploy-webjobs.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/websites-dotnet-deploy-webjobs.md)
* Service: **app-service**
* GitHub Login: @ggailey777
* Microsoft Alias: **glenga;david.ebbo;suwatch;pbatum;naren.soni**
|
1.0
|
.NET Core support? - It looks like this is only available for the .NET Framework. Are there are any resources available for .NET Core?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7c1407af-81a5-52e7-a800-79d57704a876
* Version Independent ID: 5ec30406-731f-6198-99a1-9d4ee5c67475
* Content: [Develop and deploy WebJobs using Visual Studio - Azure | Microsoft Docs](https://docs.microsoft.com/en-us/azure/app-service/websites-dotnet-deploy-webjobs)
* Content Source: [articles/app-service/websites-dotnet-deploy-webjobs.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/websites-dotnet-deploy-webjobs.md)
* Service: **app-service**
* GitHub Login: @ggailey777
* Microsoft Alias: **glenga;david.ebbo;suwatch;pbatum;naren.soni**
|
process
|
net core support it looks like this is only available for the net framework are there are any resources available for net core document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service github login microsoft alias glenga david ebbo suwatch pbatum naren soni
| 1
|
17,766
| 23,698,546,429
|
IssuesEvent
|
2022-08-29 16:43:21
|
googleapis/cloud-trace-nodejs
|
https://api.github.com/repos/googleapis/cloud-trace-nodejs
|
closed
|
make test requirements closer to other Node.js repos
|
type: process api: cloudtrace
|
we have been finding that GitHub actions work quite well for most other Node.js repos, currently cloud-trace-nodejs is one of our few repos using CircleCI. Ideally I would like us to consolidate on using GitHub actions and kokoro, as it will make managing a large number of repositories easier.
Let me know if there's anything I can do to help.
|
1.0
|
make test requirements closer to other Node.js repos - we have been finding that GitHub actions work quite well for most other Node.js repos, currently cloud-trace-nodejs is one of our few repos using CircleCI. Ideally I would like us to consolidate on using GitHub actions and kokoro, as it will make managing a large number of repositories easier.
Let me know if there's anything I can do to help.
|
process
|
make test requirements closer to other node js repos we have been finding that github actions work quite well for most other node js repos currently cloud trace nodejs is one of our few repos using circleci ideally i would like us to consolidate on using github actions and kokoro as it will make managing a large number of repositories easier let me know if there s anything i can do to help
| 1
|
22,473
| 31,388,832,670
|
IssuesEvent
|
2023-08-26 04:34:08
|
shopstarco/shopstatus
|
https://api.github.com/repos/shopstarco/shopstatus
|
closed
|
🛑 Shopstar Image Processor is down
|
status shopstar-image-processor
|
In [`ac69f10`](https://github.com/shopstarco/shopstatus/commit/ac69f10b2ef8e71f6fbd0baaf4505934ff5ef27d
), Shopstar Image Processor (https://imgpro.shopstar.co.za/) was **down**:
- HTTP code: 0
- Response time: 0 ms
|
1.0
|
🛑 Shopstar Image Processor is down - In [`ac69f10`](https://github.com/shopstarco/shopstatus/commit/ac69f10b2ef8e71f6fbd0baaf4505934ff5ef27d
), Shopstar Image Processor (https://imgpro.shopstar.co.za/) was **down**:
- HTTP code: 0
- Response time: 0 ms
|
process
|
🛑 shopstar image processor is down in shopstar image processor was down http code response time ms
| 1
|
328,706
| 9,999,080,791
|
IssuesEvent
|
2019-07-12 09:45:10
|
Sinapse-Energia/IoT-Hub-Home
|
https://api.github.com/repos/Sinapse-Energia/IoT-Hub-Home
|
opened
|
[CL][BUG] The LED is not working as expected
|
Client Priority: high Size: 2 Status: new Type: bug good first issue
|
In the version 19.28.00 we have a very strange behaviour with the LED.
We have two different behaviours:
1. Green, Red, Green...
2. Red fix
All the devices are connected and publishing.
The LED should work as expected
|
1.0
|
[CL][BUG] The LED is not working as expected - In the version 19.28.00 we have a very strange behaviour with the LED.
We have two different behaviours:
1. Green, Red, Green...
2. Red fix
All the devices are connected and publishing.
The LED should work as expected
|
non_process
|
the led is not working as expected in the version we have a very strange behaviour with the led we have two different behaviours green red green red fix all the devices are connected and publishing the led should work as expected
| 0
|
63,874
| 26,540,875,182
|
IssuesEvent
|
2023-01-19 19:09:31
|
teambtcmap/btcmap-data
|
https://api.github.com/repos/teambtcmap/btcmap-data
|
closed
|
JheDev
|
help wanted good first issue location-submission service Indonesia
|
Merchant name: JheDev
Country: Indonesia
Address: Kertamulya, Kecamatan Pedes, Kabupaten Karawang
Lat: -6.088377363099079
Long: 107.34325299278923
OSM: https://www.openstreetmap.org/edit#map=21/-6.088377363099079/107.34325299278923
Category: DevOps Room
Payment methods: onchain,lightning,nfc
Website: https://jhedev.cf
Phone: 085693316296
Opening hours:
Twitter merchant:
Twitter submitter:
Notes:
Data Source: Customer
Details (if applicable):
Contact:
Status: Todo
Created at: 2023-01-19T14:45:45.195Z
If you are a new contributor please read our Tagging Instructions [here](https://github.com/teambtcmap/btcmap-data/wiki/Tagging-Instructions).
|
1.0
|
JheDev - Merchant name: JheDev
Country: Indonesia
Address: Kertamulya, Kecamatan Pedes, Kabupaten Karawang
Lat: -6.088377363099079
Long: 107.34325299278923
OSM: https://www.openstreetmap.org/edit#map=21/-6.088377363099079/107.34325299278923
Category: DevOps Room
Payment methods: onchain,lightning,nfc
Website: https://jhedev.cf
Phone: 085693316296
Opening hours:
Twitter merchant:
Twitter submitter:
Notes:
Data Source: Customer
Details (if applicable):
Contact:
Status: Todo
Created at: 2023-01-19T14:45:45.195Z
If you are a new contributor please read our Tagging Instructions [here](https://github.com/teambtcmap/btcmap-data/wiki/Tagging-Instructions).
|
non_process
|
jhedev merchant name jhedev country indonesia address kertamulya kecamatan pedes kabupaten karawang lat long osm category devops room payment methods onchain lightning nfc website phone opening hours twitter merchant twitter submitter notes data source customer details if applicable contact status todo created at if you are a new contributor please read our tagging instructions
| 0
|
22,321
| 30,884,545,855
|
IssuesEvent
|
2023-08-03 20:29:39
|
IMAP-Science-Operations-Center/sds-data-manager
|
https://api.github.com/repos/IMAP-Science-Operations-Center/sds-data-manager
|
closed
|
Initial Implementation of Trigger jobs based on data availability
|
enhancement Parent Req: SDC Processing Monitoring
|
### Summary
This work is for SIT-2 which is to setup minimal processing pipeline using step function and lambda. Post SIT-2, it be evaluated again through formal process through SPIKE.
### Requirements
The more specific the better.
- Setup processing pipeline using step function
- Use lambda to check database and do decom work
### Related tickets
_No response_
|
1.0
|
Initial Implementation of Trigger jobs based on data availability - ### Summary
This work is for SIT-2 which is to setup minimal processing pipeline using step function and lambda. Post SIT-2, it be evaluated again through formal process through SPIKE.
### Requirements
The more specific the better.
- Setup processing pipeline using step function
- Use lambda to check database and do decom work
### Related tickets
_No response_
|
process
|
initial implementation of trigger jobs based on data availability summary this work is for sit which is to setup minimal processing pipeline using step function and lambda post sit it be evaluated again through formal process through spike requirements the more specific the better setup processing pipeline using step function use lambda to check database and do decom work related tickets no response
| 1
|
227,224
| 18,054,146,976
|
IssuesEvent
|
2021-09-20 05:06:04
|
logicmoo/logicmoo_workspace
|
https://api.github.com/repos/logicmoo/logicmoo_workspace
|
opened
|
logicmoo.pfc.test.sanity_base.HYBRID_01A JUnit
|
Test_9999 logicmoo.pfc.test.sanity_base unit_test HYBRID_01A Passing
|
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s swipl -x /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-clif hybrid_01a.pfc)
% ISSUE: https://github.com/logicmoo/logicmoo_workspace/issues/
% EDIT: https://github.com/logicmoo/logicmoo_workspace/edit/master/packs_sys/pfc/t/sanity_base/hybrid_01a.pfc
% JENKINS: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/HYBRID_01A/logicmoo_pfc_test_sanity_base_HYBRID_01A_JUnit/
% ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AHYBRID_01A
```
%~ init_phase(after_load)
%~ init_phase(restore_state)
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/hybrid_01a.pfc'),
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
:- if((pfc_test_feature(localMt,X=1),X==1)).
:- endif.
:- if((pfc_test_feature(mt,X=1),X==1)).
:- endif.
:- kb_shared(a/0).
:- \+ a.
% ISSUE: https://github.com/logicmoo/logicmoo_workspace/issues/
% EDIT: https://github.com/logicmoo/logicmoo_workspace/edit/master/packs_sys/pfc/t/sanity_base/hybrid_01a.pfc
% JENKINS: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/HYBRID_01A/logicmoo_pfc_test_sanity_base_HYBRID_01A_JUnit/
% ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AHYBRID_01A
%~ mpred_test("Test_0001_Line_0000__naf_A",baseKB:(\+a))
/*~
%~ mpred_test("Test_0001_Line_0000__naf_A",baseKB:(\+a))
passed=info(why_was_true(baseKB:(\+a)))
no_proof_for(\+a).
no_proof_for(\+a).
no_proof_for(\+a).
name = 'logicmoo.pfc.test.sanity_base.HYBRID_01A-Test_0001_Line_0000__naf_A'.
JUNIT_CLASSNAME = 'logicmoo.pfc.test.sanity_base.HYBRID_01A'.
JUNIT_CMD = 'timeout --foreground --preserve-status -s SIGKILL -k 10s 10s swipl -x /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-clif hybrid_01a.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-pfc-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.HYBRID_01A-Test_0001_Line_0000__naf_A-junit.xml
~*/
%~ unused(no_junit_results)
Test_0001_Line_0000__naf_A result = passed.
%~ test_completed_exit(64)
```
totalTime=1.000
SUCCESS: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k hybrid_01a.pfc (returned 64) Add_LABELS='' Rem_LABELS='Skipped,Errors,Warnings,Overtime,Skipped,Skipped'
|
3.0
|
logicmoo.pfc.test.sanity_base.HYBRID_01A JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s swipl -x /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-clif hybrid_01a.pfc)
% ISSUE: https://github.com/logicmoo/logicmoo_workspace/issues/
% EDIT: https://github.com/logicmoo/logicmoo_workspace/edit/master/packs_sys/pfc/t/sanity_base/hybrid_01a.pfc
% JENKINS: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/HYBRID_01A/logicmoo_pfc_test_sanity_base_HYBRID_01A_JUnit/
% ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AHYBRID_01A
```
%~ init_phase(after_load)
%~ init_phase(restore_state)
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/hybrid_01a.pfc'),
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
:- if((pfc_test_feature(localMt,X=1),X==1)).
:- endif.
:- if((pfc_test_feature(mt,X=1),X==1)).
:- endif.
:- kb_shared(a/0).
:- \+ a.
% ISSUE: https://github.com/logicmoo/logicmoo_workspace/issues/
% EDIT: https://github.com/logicmoo/logicmoo_workspace/edit/master/packs_sys/pfc/t/sanity_base/hybrid_01a.pfc
% JENKINS: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/HYBRID_01A/logicmoo_pfc_test_sanity_base_HYBRID_01A_JUnit/
% ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AHYBRID_01A
%~ mpred_test("Test_0001_Line_0000__naf_A",baseKB:(\+a))
/*~
%~ mpred_test("Test_0001_Line_0000__naf_A",baseKB:(\+a))
passed=info(why_was_true(baseKB:(\+a)))
no_proof_for(\+a).
no_proof_for(\+a).
no_proof_for(\+a).
name = 'logicmoo.pfc.test.sanity_base.HYBRID_01A-Test_0001_Line_0000__naf_A'.
JUNIT_CLASSNAME = 'logicmoo.pfc.test.sanity_base.HYBRID_01A'.
JUNIT_CMD = 'timeout --foreground --preserve-status -s SIGKILL -k 10s 10s swipl -x /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-clif hybrid_01a.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-pfc-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.HYBRID_01A-Test_0001_Line_0000__naf_A-junit.xml
~*/
%~ unused(no_junit_results)
Test_0001_Line_0000__naf_A result = passed.
%~ test_completed_exit(64)
```
totalTime=1.000
SUCCESS: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k hybrid_01a.pfc (returned 64) Add_LABELS='' Rem_LABELS='Skipped,Errors,Warnings,Overtime,Skipped,Skipped'
|
non_process
|
logicmoo pfc test sanity base hybrid junit cd var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base timeout foreground preserve status s sigkill k swipl x var lib jenkins workspace logicmoo workspace bin lmoo clif hybrid pfc issue edit jenkins issue search init phase after load init phase restore state running var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base hybrid pfc this test might need use module library logicmoo plarkc if pfc test feature localmt x x endif if pfc test feature mt x x endif kb shared a a issue edit jenkins issue search mpred test test line naf a basekb a mpred test test line naf a basekb a passed info why was true basekb a no proof for a no proof for a no proof for a name logicmoo pfc test sanity base hybrid test line naf a junit classname logicmoo pfc test sanity base hybrid junit cmd timeout foreground preserve status s sigkill k swipl x var lib jenkins workspace logicmoo workspace bin lmoo clif hybrid pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo pfc test sanity base units logicmoo pfc test sanity base hybrid test line naf a junit xml unused no junit results test line naf a result passed test completed exit totaltime success var lib jenkins workspace logicmoo workspace bin lmoo junit minor k hybrid pfc returned add labels rem labels skipped errors warnings overtime skipped skipped
| 0
|
263,726
| 8,301,091,960
|
IssuesEvent
|
2018-09-21 10:10:48
|
MagiCircles/BanGDream
|
https://api.github.com/repos/MagiCircles/BanGDream
|
opened
|
Asset Title Text Overflowing
|
bug low priority
|

It doesn't happen often (this is the only example I've seen) but the title text can overflow into the other items text area.
This is most likely because of the length of the word, so we probably would need to apply/improve some algorithm that decides what the most "suitable" place would be to cut short too long of sentences.
|
1.0
|
Asset Title Text Overflowing - 
It doesn't happen often (this is the only example I've seen) but the title text can overflow into the other items text area.
This is most likely because of the length of the word, so we probably would need to apply/improve some algorithm that decides what the most "suitable" place would be to cut short too long of sentences.
|
non_process
|
asset title text overflowing it doesn t happen often this is the only example i ve seen but the title text can overflow into the other items text area this is most likely because of the length of the word so we probably would need to apply improve some algorithm that decides what the most suitable place would be to cut short too long of sentences
| 0
|
15,422
| 19,608,207,751
|
IssuesEvent
|
2022-01-06 12:15:57
|
shirou/gopsutil
|
https://api.github.com/repos/shirou/gopsutil
|
closed
|
v3.21.12: compilation error on Windows arm64
|
os:windows package:process
|
**Describe the bug**
The `process` module does not compile on Windows arm64 in v3.21.12, it works in v.3.21.11
**To Reproduce**
```go
package main
import (
"fmt"
"github.com/shirou/gopsutil/v3/process"
)
func main() {
procs, err := process.Processes()
if err != nil {
panic(err)
}
fmt.Printf("processes: %+v\n", procs)
}
```
**Expected behavior**
this code should compile
**Environment (please complete the following information):**
- [X] Windows: the version used in GitHub actions, take a look [here](https://github.com/drakkan/sftpgo/runs/4701720408?check_suite_focus=true)
- [X] Linux:
```
cat /etc/os-release
NAME="Arch Linux"
PRETTY_NAME="Arch Linux"
ID=arch
BUILD_ID=rolling
ANSI_COLOR="38;2;23;147;209"
HOME_URL="https://archlinux.org/"
DOCUMENTATION_URL="https://wiki.archlinux.org/"
SUPPORT_URL="https://bbs.archlinux.org/"
BUG_REPORT_URL="https://bugs.archlinux.org/"
LOGO=archlinux-logo
uname -a
Linux p1 5.15.12-arch1-1 #1 SMP PREEMPT Wed, 29 Dec 2021 12:04:56 +0000 x86_64 GNU/Linux
```
**Additional context**
the GitHub actions cross compile from windows amd64 to windows arm64, take a look [here](https://github.com/drakkan/sftpgo/blob/main/.github/workflows/development.yml#L61)
on my Linux laptop I can reproduce trying to compile the code above
```
env | grep GO
GOARCH=arm64
CGO_ENABLED=0
GOOS=windows
go build
# github.com/shirou/gopsutil/v3/process
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:412:11: undefined: readProcessMemory
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:425:11: undefined: readProcessMemory
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:896:32: undefined: PROCESS_MEMORY_COUNTERS
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:897:10: undefined: PROCESS_MEMORY_COUNTERS
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:910:50: undefined: PROCESS_MEMORY_COUNTERS
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:950:21: undefined: queryPebAddress
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:955:9: undefined: readProcessMemory
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:961:8: undefined: readProcessMemory
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:969:21: undefined: queryPebAddress
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:974:9: undefined: readProcessMemory
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:974:9: too many errors
```
|
1.0
|
v3.21.12: compilation error on Windows arm64 - **Describe the bug**
The `process` module does not compile on Windows arm64 in v3.21.12, it works in v.3.21.11
**To Reproduce**
```go
package main
import (
"fmt"
"github.com/shirou/gopsutil/v3/process"
)
func main() {
procs, err := process.Processes()
if err != nil {
panic(err)
}
fmt.Printf("processes: %+v\n", procs)
}
```
**Expected behavior**
this code should compile
**Environment (please complete the following information):**
- [X] Windows: the version used in GitHub actions, take a look [here](https://github.com/drakkan/sftpgo/runs/4701720408?check_suite_focus=true)
- [X] Linux:
```
cat /etc/os-release
NAME="Arch Linux"
PRETTY_NAME="Arch Linux"
ID=arch
BUILD_ID=rolling
ANSI_COLOR="38;2;23;147;209"
HOME_URL="https://archlinux.org/"
DOCUMENTATION_URL="https://wiki.archlinux.org/"
SUPPORT_URL="https://bbs.archlinux.org/"
BUG_REPORT_URL="https://bugs.archlinux.org/"
LOGO=archlinux-logo
uname -a
Linux p1 5.15.12-arch1-1 #1 SMP PREEMPT Wed, 29 Dec 2021 12:04:56 +0000 x86_64 GNU/Linux
```
**Additional context**
the GitHub actions cross compile from windows amd64 to windows arm64, take a look [here](https://github.com/drakkan/sftpgo/blob/main/.github/workflows/development.yml#L61)
on my Linux laptop I can reproduce trying to compile the code above
```
env | grep GO
GOARCH=arm64
CGO_ENABLED=0
GOOS=windows
go build
# github.com/shirou/gopsutil/v3/process
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:412:11: undefined: readProcessMemory
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:425:11: undefined: readProcessMemory
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:896:32: undefined: PROCESS_MEMORY_COUNTERS
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:897:10: undefined: PROCESS_MEMORY_COUNTERS
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:910:50: undefined: PROCESS_MEMORY_COUNTERS
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:950:21: undefined: queryPebAddress
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:955:9: undefined: readProcessMemory
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:961:8: undefined: readProcessMemory
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:969:21: undefined: queryPebAddress
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:974:9: undefined: readProcessMemory
/home/nicola/go/pkg/mod/github.com/shirou/gopsutil/v3@v3.21.12/process/process_windows.go:974:9: too many errors
```
|
process
|
compilation error on windows describe the bug the process module does not compile on windows in it works in v to reproduce go package main import fmt github com shirou gopsutil process func main procs err process processes if err nil panic err fmt printf processes v n procs expected behavior this code should compile environment please complete the following information windows the version used in github actions take a look linux cat etc os release name arch linux pretty name arch linux id arch build id rolling ansi color home url documentation url support url bug report url logo archlinux logo uname a linux smp preempt wed dec gnu linux additional context the github actions cross compile from windows to windows take a look on my linux laptop i can reproduce trying to compile the code above env grep go goarch cgo enabled goos windows go build github com shirou gopsutil process home nicola go pkg mod github com shirou gopsutil process process windows go undefined readprocessmemory home nicola go pkg mod github com shirou gopsutil process process windows go undefined readprocessmemory home nicola go pkg mod github com shirou gopsutil process process windows go undefined process memory counters home nicola go pkg mod github com shirou gopsutil process process windows go undefined process memory counters home nicola go pkg mod github com shirou gopsutil process process windows go undefined process memory counters home nicola go pkg mod github com shirou gopsutil process process windows go undefined querypebaddress home nicola go pkg mod github com shirou gopsutil process process windows go undefined readprocessmemory home nicola go pkg mod github com shirou gopsutil process process windows go undefined readprocessmemory home nicola go pkg mod github com shirou gopsutil process process windows go undefined querypebaddress home nicola go pkg mod github com shirou gopsutil process process windows go undefined readprocessmemory home nicola go pkg mod github com shirou gopsutil process process windows go too many errors
| 1
|
12,426
| 14,926,093,566
|
IssuesEvent
|
2021-01-24 09:53:28
|
scikit-learn/scikit-learn
|
https://api.github.com/repos/scikit-learn/scikit-learn
|
closed
|
Add Spline Transformer
|
New Feature module:preprocessing
|
#### Describe the workflow you want to enable
I propose to add a `SplineTransformer` to `preprocessing`. This is similiar to `PolynomialFeatures`, but gives more flexibility (and numerical stability) for linear models to deal with continuous numerical features.
```python
from sklearn.linear_model import LogisticRegression
from sklearn.preprocessing import SplineTransformer
from sklearn.pipeline import make_pipeline
# get data X, y
...
model = make_pipeline(SplineTransformer(degree=3, n_knots=20,
positioning='quantile'),
LogisticRegression())
model.fit(X, y)
```
#### Describe your proposed solution
Add `SplineTransformer` and internally use scipy for splines. Start with
- [1-dimensional b-splines](PolynomialFeatures)
- equidistant knots
- quantile based knots
#### Additional context
[Patsy](https://patsy.readthedocs.io/en/latest/spline-regression.html) has an implementation of those that matches the R versions.
#### References
[Eilers, Marx "Flexible Smoothing with B-splines and Penalties"](https://doi.org/10.1214/ss/1038425655) passes the [scikit-learn inclusion criteria](https://scikit-learn.org/stable/faq.html?highlight=inclusion#what-are-the-inclusion-criteria-for-new-algorithms) by some margin :smirk:
|
1.0
|
Add Spline Transformer - #### Describe the workflow you want to enable
I propose to add a `SplineTransformer` to `preprocessing`. This is similiar to `PolynomialFeatures`, but gives more flexibility (and numerical stability) for linear models to deal with continuous numerical features.
```python
from sklearn.linear_model import LogisticRegression
from sklearn.preprocessing import SplineTransformer
from sklearn.pipeline import make_pipeline
# get data X, y
...
model = make_pipeline(SplineTransformer(degree=3, n_knots=20,
positioning='quantile'),
LogisticRegression())
model.fit(X, y)
```
#### Describe your proposed solution
Add `SplineTransformer` and internally use scipy for splines. Start with
- [1-dimensional b-splines](PolynomialFeatures)
- equidistant knots
- quantile based knots
#### Additional context
[Patsy](https://patsy.readthedocs.io/en/latest/spline-regression.html) has an implementation of those that matches the R versions.
#### References
[Eilers, Marx "Flexible Smoothing with B-splines and Penalties"](https://doi.org/10.1214/ss/1038425655) passes the [scikit-learn inclusion criteria](https://scikit-learn.org/stable/faq.html?highlight=inclusion#what-are-the-inclusion-criteria-for-new-algorithms) by some margin :smirk:
|
process
|
add spline transformer describe the workflow you want to enable i propose to add a splinetransformer to preprocessing this is similiar to polynomialfeatures but gives more flexibility and numerical stability for linear models to deal with continuous numerical features python from sklearn linear model import logisticregression from sklearn preprocessing import splinetransformer from sklearn pipeline import make pipeline get data x y model make pipeline splinetransformer degree n knots positioning quantile logisticregression model fit x y describe your proposed solution add splinetransformer and internally use scipy for splines start with polynomialfeatures equidistant knots quantile based knots additional context has an implementation of those that matches the r versions references passes the by some margin smirk
| 1
|
73,760
| 9,696,130,710
|
IssuesEvent
|
2019-05-25 04:11:56
|
fossasia/susi_server
|
https://api.github.com/repos/fossasia/susi_server
|
closed
|
Implement documentation site using travis and sphinx
|
documentation
|
It seems to work out well with the site generated for the loklak documentation.
Please have a look at the script (https://github.com/loklak/loklak_server/blob/development/.utility/push-docs-to-gh-pages.sh) and docs site and duplicate the way it is done here.
|
1.0
|
Implement documentation site using travis and sphinx - It seems to work out well with the site generated for the loklak documentation.
Please have a look at the script (https://github.com/loklak/loklak_server/blob/development/.utility/push-docs-to-gh-pages.sh) and docs site and duplicate the way it is done here.
|
non_process
|
implement documentation site using travis and sphinx it seems to work out well with the site generated for the loklak documentation please have a look at the script and docs site and duplicate the way it is done here
| 0
|
101,059
| 16,490,756,329
|
IssuesEvent
|
2021-05-25 03:13:14
|
valdisiljuconoks/episerverless
|
https://api.github.com/repos/valdisiljuconoks/episerverless
|
opened
|
CVE-2018-1285 (High) detected in log4net.2.0.8.nupkg
|
security vulnerability
|
## CVE-2018-1285 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4net.2.0.8.nupkg</b></p></summary>
<p>The Apache log4net library is a tool to help the programmer output log statements to a variety of ou...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/log4net.2.0.8.nupkg">https://api.nuget.org/packages/log4net.2.0.8.nupkg</a></p>
<p>Path to dependency file: /episerverless/src/Web/packages.config</p>
<p>Path to vulnerable library: episerverless/src/Web/packages.config,episerverless/packages/log4net.2.0.8/log4net.2.0.8.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **log4net.2.0.8.nupkg** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache log4net versions before 2.0.10 do not disable XML external entities when parsing log4net configuration files. This allows for XXE-based attacks in applications that accept attacker-controlled log4net configuration files.
<p>Publish Date: 2020-05-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1285>CVE-2018-1285</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/logging-log4net/tree/rel/2.0.10">https://github.com/apache/logging-log4net/tree/rel/2.0.10</a></p>
<p>Release Date: 2020-05-11</p>
<p>Fix Resolution: log4net - 2.0.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-1285 (High) detected in log4net.2.0.8.nupkg - ## CVE-2018-1285 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4net.2.0.8.nupkg</b></p></summary>
<p>The Apache log4net library is a tool to help the programmer output log statements to a variety of ou...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/log4net.2.0.8.nupkg">https://api.nuget.org/packages/log4net.2.0.8.nupkg</a></p>
<p>Path to dependency file: /episerverless/src/Web/packages.config</p>
<p>Path to vulnerable library: episerverless/src/Web/packages.config,episerverless/packages/log4net.2.0.8/log4net.2.0.8.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **log4net.2.0.8.nupkg** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache log4net versions before 2.0.10 do not disable XML external entities when parsing log4net configuration files. This allows for XXE-based attacks in applications that accept attacker-controlled log4net configuration files.
<p>Publish Date: 2020-05-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1285>CVE-2018-1285</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/logging-log4net/tree/rel/2.0.10">https://github.com/apache/logging-log4net/tree/rel/2.0.10</a></p>
<p>Release Date: 2020-05-11</p>
<p>Fix Resolution: log4net - 2.0.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in nupkg cve high severity vulnerability vulnerable library nupkg the apache library is a tool to help the programmer output log statements to a variety of ou library home page a href path to dependency file episerverless src web packages config path to vulnerable library episerverless src web packages config episerverless packages nupkg dependency hierarchy x nupkg vulnerable library vulnerability details apache versions before do not disable xml external entities when parsing configuration files this allows for xxe based attacks in applications that accept attacker controlled configuration files publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
19,807
| 26,193,890,002
|
IssuesEvent
|
2023-01-03 11:38:52
|
firebase/firebase-cpp-sdk
|
https://api.github.com/repos/firebase/firebase-cpp-sdk
|
closed
|
[C++] Nightly Integration Testing Report for Firestore
|
type: process nightly-testing
|
<hidden value="integration-test-status-comment"></hidden>
### [build against repo] Integration test with FLAKINESS (succeeded after retry)
Requested by @sunmou99 on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7
Last updated: Mon Jan 2 03:45 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3821739062)**
| Failures | Configs |
|----------|---------|
| firestore | [TEST] [FLAKINESS] [Android] [1/3 os: windows] [2/4 android_device: emulator_ftl_target emulator_ftl_latest]<details><summary>(1 failed tests)</summary> CRASH/TIMEOUT</details>[TEST] [FLAKINESS] [Android] [2/3 os: ubuntu macos] [1/4 android_device: android_target]<details><summary>(1 failed tests)</summary> CRASH/TIMEOUT</details> |
Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7
Last updated: Mon Jan 2 08:59 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3822423583)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against tip] Integration test succeeded!
Requested by @sunmou99 on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7
Last updated: Mon Jan 2 08:59 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3822051488)**
|
1.0
|
[C++] Nightly Integration Testing Report for Firestore -
<hidden value="integration-test-status-comment"></hidden>
### [build against repo] Integration test with FLAKINESS (succeeded after retry)
Requested by @sunmou99 on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7
Last updated: Mon Jan 2 03:45 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3821739062)**
| Failures | Configs |
|----------|---------|
| firestore | [TEST] [FLAKINESS] [Android] [1/3 os: windows] [2/4 android_device: emulator_ftl_target emulator_ftl_latest]<details><summary>(1 failed tests)</summary> CRASH/TIMEOUT</details>[TEST] [FLAKINESS] [Android] [2/3 os: ubuntu macos] [1/4 android_device: android_target]<details><summary>(1 failed tests)</summary> CRASH/TIMEOUT</details> |
Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7
Last updated: Mon Jan 2 08:59 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3822423583)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against tip] Integration test succeeded!
Requested by @sunmou99 on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7
Last updated: Mon Jan 2 08:59 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3822051488)**
|
process
|
nightly integration testing report for firestore integration test with flakiness succeeded after retry requested by on commit last updated mon jan pst failures configs firestore failed tests nbsp nbsp crash timeout failed tests nbsp nbsp crash timeout add flaky tests to ✅ nbsp integration test succeeded requested by firebase workflow trigger on commit last updated mon jan pst ✅ nbsp integration test succeeded requested by on commit last updated mon jan pst
| 1
|
11,792
| 14,619,868,773
|
IssuesEvent
|
2020-12-22 18:38:32
|
ewen-lbh/portfolio
|
https://api.github.com/repos/ewen-lbh/portfolio
|
closed
|
Implement `page background` and `title style`
|
processing styling
|
For now, I only use those in the 'Neptune' description.md to match [the prototype's matching page](https://github.com/ewen-lbh/portfolio/blob/master/PROTOTYPE.pdf) as such:
```yaml
title style: outlined # i guess available options are outlined and filled with more to come and "filled" the default?
page background: starry-sky.png # this is not on the prototype but I'd be nice to be able to set a fullwidth background for some works
```
## Formal specification
### `title style`
Go struct: `Work.Metadata.TitleStyle`
Allowed values:
- `"outlined"`
- `"filled"`
Default value:
`"filled"`
### `page backgroud`
Go struct: `Work.Metadata.PageBackground`
Allowed values:
- Any valid image (including Graphics Interchange Format images) file path relative to `description.md`'s folder (`.portfoliodb`)
- `""` _(the empty string)_
Default value:
- `""` _(the empty string)_
|
1.0
|
Implement `page background` and `title style` - For now, I only use those in the 'Neptune' description.md to match [the prototype's matching page](https://github.com/ewen-lbh/portfolio/blob/master/PROTOTYPE.pdf) as such:
```yaml
title style: outlined # i guess available options are outlined and filled with more to come and "filled" the default?
page background: starry-sky.png # this is not on the prototype but I'd be nice to be able to set a fullwidth background for some works
```
## Formal specification
### `title style`
Go struct: `Work.Metadata.TitleStyle`
Allowed values:
- `"outlined"`
- `"filled"`
Default value:
`"filled"`
### `page backgroud`
Go struct: `Work.Metadata.PageBackground`
Allowed values:
- Any valid image (including Graphics Interchange Format images) file path relative to `description.md`'s folder (`.portfoliodb`)
- `""` _(the empty string)_
Default value:
- `""` _(the empty string)_
|
process
|
implement page background and title style for now i only use those in the neptune description md to match as such yaml title style outlined i guess available options are outlined and filled with more to come and filled the default page background starry sky png this is not on the prototype but i d be nice to be able to set a fullwidth background for some works formal specification title style go struct work metadata titlestyle allowed values outlined filled default value filled page backgroud go struct work metadata pagebackground allowed values any valid image including graphics interchange format images file path relative to description md s folder portfoliodb the empty string default value the empty string
| 1
|
405,965
| 11,885,246,112
|
IssuesEvent
|
2020-03-27 19:12:53
|
google/knative-gcp
|
https://api.github.com/repos/google/knative-gcp
|
closed
|
Adding a status condition for workload identity
|
area/security kind/cleanup kind/feature-request priority/2
|
**Problem**
We need a status condition for objects for workload identity. When workload identity wasn't properly set, it will be marked.
If the user is not using workload identity (using secret), the status type WorkloadIdentityReady should be unknown. This status will be not included in condset, so that the unknown status will not affect the whole condition to be ready.
This condition should be set in channel and all sources.
**Exit Criteria**
Adding a status condition for workload identity
|
1.0
|
Adding a status condition for workload identity - **Problem**
We need a status condition for objects for workload identity. When workload identity wasn't properly set, it will be marked.
If the user is not using workload identity (using secret), the status type WorkloadIdentityReady should be unknown. This status will be not included in condset, so that the unknown status will not affect the whole condition to be ready.
This condition should be set in channel and all sources.
**Exit Criteria**
Adding a status condition for workload identity
|
non_process
|
adding a status condition for workload identity problem we need a status condition for objects for workload identity when workload identity wasn t properly set it will be marked if the user is not using workload identity using secret the status type workloadidentityready should be unknown this status will be not included in condset so that the unknown status will not affect the whole condition to be ready this condition should be set in channel and all sources exit criteria adding a status condition for workload identity
| 0
|
40,874
| 6,875,291,691
|
IssuesEvent
|
2017-11-19 12:14:18
|
junit-team/junit5
|
https://api.github.com/repos/junit-team/junit5
|
closed
|
Update Asciidoctor PDF backend of the user-guide
|
status: blocked theme: documentation
|
## Overview
The PDF backend of is disabled at the moment, as does not work with Java 9.
See `documentation/documentation.gradle:83` for backend configuration.
See https://github.com/jruby/jruby/issues/4805 for the underlying issue.
## Deliverables
- [ ] Enable PDF backend when https://github.com/jruby/jruby/issues/4805 is solved and JRuby **9.1.14** is released.
|
1.0
|
Update Asciidoctor PDF backend of the user-guide - ## Overview
The PDF backend of is disabled at the moment, as does not work with Java 9.
See `documentation/documentation.gradle:83` for backend configuration.
See https://github.com/jruby/jruby/issues/4805 for the underlying issue.
## Deliverables
- [ ] Enable PDF backend when https://github.com/jruby/jruby/issues/4805 is solved and JRuby **9.1.14** is released.
|
non_process
|
update asciidoctor pdf backend of the user guide overview the pdf backend of is disabled at the moment as does not work with java see documentation documentation gradle for backend configuration see for the underlying issue deliverables enable pdf backend when is solved and jruby is released
| 0
|
422,977
| 28,490,775,345
|
IssuesEvent
|
2023-04-18 11:04:12
|
carpentries/amy
|
https://api.github.com/repos/carpentries/amy
|
closed
|
Add how to send real emails to the README
|
component: documentation
|
Testing emails end to end was required for some of the features I'd been working on recently. This is a reminder to myself to remember to document how I got everything working.
|
1.0
|
Add how to send real emails to the README - Testing emails end to end was required for some of the features I'd been working on recently. This is a reminder to myself to remember to document how I got everything working.
|
non_process
|
add how to send real emails to the readme testing emails end to end was required for some of the features i d been working on recently this is a reminder to myself to remember to document how i got everything working
| 0
|
2,181
| 5,031,682,168
|
IssuesEvent
|
2016-12-16 08:20:21
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
opened
|
Raise an event when a new script was executed
|
AREA: client SYSTEM: resource processing TYPE: enhancement
|
by appending to the document, via `eval` etc..
It's necessary for https://github.com/DevExpress/testcafe/issues/1072
|
1.0
|
Raise an event when a new script was executed - by appending to the document, via `eval` etc..
It's necessary for https://github.com/DevExpress/testcafe/issues/1072
|
process
|
raise an event when a new script was executed by appending to the document via eval etc it s necessary for
| 1
|
18,058
| 3,664,524,064
|
IssuesEvent
|
2016-02-19 12:08:54
|
handsontable/handsontable
|
https://api.github.com/repos/handsontable/handsontable
|
closed
|
Inserting rows/cols doesn't shift cell meta data
|
Bug Core: handsontable Priority: high Released Tested
|
Removing rows and columns works just fine, but inserting new ones makes meta data lose track of their cells - the meta data isn't shifted along with the cell values.
Try the following JSFiddle: http://jsfiddle.net/zq5fzegx/1/
1. Notice that the "#" cell is green and centered (because of the cell's `className` meta data).
2. Now remove the "Kia" column and the "2008" row.
* The "#" is still green and centered as one would expect.
3. Insert a new row and a new column where the old ones were removed in step 2 (insert row above "2009" and insert column to the left of "Nissan").
* Now, the number 11 is green and centered instead.
|
1.0
|
Inserting rows/cols doesn't shift cell meta data - Removing rows and columns works just fine, but inserting new ones makes meta data lose track of their cells - the meta data isn't shifted along with the cell values.
Try the following JSFiddle: http://jsfiddle.net/zq5fzegx/1/
1. Notice that the "#" cell is green and centered (because of the cell's `className` meta data).
2. Now remove the "Kia" column and the "2008" row.
* The "#" is still green and centered as one would expect.
3. Insert a new row and a new column where the old ones were removed in step 2 (insert row above "2009" and insert column to the left of "Nissan").
* Now, the number 11 is green and centered instead.
|
non_process
|
inserting rows cols doesn t shift cell meta data removing rows and columns works just fine but inserting new ones makes meta data lose track of their cells the meta data isn t shifted along with the cell values try the following jsfiddle notice that the cell is green and centered because of the cell s classname meta data now remove the kia column and the row the is still green and centered as one would expect insert a new row and a new column where the old ones were removed in step insert row above and insert column to the left of nissan now the number is green and centered instead
| 0
|
126,758
| 26,909,537,172
|
IssuesEvent
|
2023-02-06 22:09:18
|
MetaMask/design-tokens
|
https://api.github.com/repos/MetaMask/design-tokens
|
opened
|
[Ext] Insight Report: Tooltip
|
code design-system
|
### **Description**
Fill out the `Tooltip` insight report from your findings from the audit
The insight report will be part of our decision making framework and is intend to:
- Document all findings from the component audit
- Confirm as many component details as possible
- Mitigate component inconsistencies across Figma, Mobile and Extension
Include your thoughts on component name, description, api and any comments or topic to discuss relating to the component or it's make up. We will review the audit and insight report in our Wednesday technical sync to finalize the details of the component for all platforms.
### **Technical Details**
The insight report should include the following for the component
- name
- description
- variants/props
- requirements (optional)
- discussion/questions (optional)
### **Acceptance Criteria**
- name, description variants/props is filled out and matches with other platforms where possible
### **References**
- [FigJam](https://www.figma.com/file/KP10I7OHiuUsGZ53xAgZTj/Popover-Audit?node-id=0%3A1&t=u3rGtzERyA5zkSPu-1)
- Read exercised `#05 Identify Existing Paradigms in Design and Code` and `#06 IdentifyEmergingandInteresting
Paradigms in Design and Code` in the Design System in 90 Days workbook
|
1.0
|
[Ext] Insight Report: Tooltip - ### **Description**
Fill out the `Tooltip` insight report from your findings from the audit
The insight report will be part of our decision making framework and is intend to:
- Document all findings from the component audit
- Confirm as many component details as possible
- Mitigate component inconsistencies across Figma, Mobile and Extension
Include your thoughts on component name, description, api and any comments or topic to discuss relating to the component or it's make up. We will review the audit and insight report in our Wednesday technical sync to finalize the details of the component for all platforms.
### **Technical Details**
The insight report should include the following for the component
- name
- description
- variants/props
- requirements (optional)
- discussion/questions (optional)
### **Acceptance Criteria**
- name, description variants/props is filled out and matches with other platforms where possible
### **References**
- [FigJam](https://www.figma.com/file/KP10I7OHiuUsGZ53xAgZTj/Popover-Audit?node-id=0%3A1&t=u3rGtzERyA5zkSPu-1)
- Read exercised `#05 Identify Existing Paradigms in Design and Code` and `#06 IdentifyEmergingandInteresting
Paradigms in Design and Code` in the Design System in 90 Days workbook
|
non_process
|
insight report tooltip description fill out the tooltip insight report from your findings from the audit the insight report will be part of our decision making framework and is intend to document all findings from the component audit confirm as many component details as possible mitigate component inconsistencies across figma mobile and extension include your thoughts on component name description api and any comments or topic to discuss relating to the component or it s make up we will review the audit and insight report in our wednesday technical sync to finalize the details of the component for all platforms technical details the insight report should include the following for the component name description variants props requirements optional discussion questions optional acceptance criteria name description variants props is filled out and matches with other platforms where possible references read exercised identify existing paradigms in design and code and identifyemergingandinteresting paradigms in design and code in the design system in days workbook
| 0
|
13,277
| 15,759,440,073
|
IssuesEvent
|
2021-03-31 07:56:01
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
ess-command-output buffer
|
process:command
|
My config is:
- Windows 10
- GNU Emacs 27.1
- ESS 18.10.3
- Org mode 9.4.4
- R 3.6.3
For information I recently changed to this config, so I needed to modify my init.el. Before I used Emacs 24.3.1, Org-mode 7.9.3 and ESS 13.05 without having never faced the issue I describe below.
I use Emacs with org mode and org babel to run R for literate programming to create latex reports. When I run R code from an org file (after having done C-c '), it sends the code to an iESS buffer \*R* and evaluate it. Fine. But after some code evaluations, a buffer called \*ess-command-output* pops up. In it is written:
`ess-output-delimiter75-END`
`The line I evaluated and its result (it is result first and then line evaluated)`
`ess-output-delimiter75-START`
The number 75 can change. If I continue code evaluation, it goes in this buffer but in a messy way. When I go back to the iESS \*R* buffer, it is mentioned "no process" in the mode line. And thus if I close the the \*ess-command-output* buffer and evalaute code from the ESS buffer (where the script is written), it asks me to open a new iESS buffer \*R:2*.
Quite difficult to reproduce this because I cannot identify when it happens. Can be after few evaluations or more.
Any idea of how this issue can be solved? Thanks.
|
1.0
|
ess-command-output buffer - My config is:
- Windows 10
- GNU Emacs 27.1
- ESS 18.10.3
- Org mode 9.4.4
- R 3.6.3
For information I recently changed to this config, so I needed to modify my init.el. Before I used Emacs 24.3.1, Org-mode 7.9.3 and ESS 13.05 without having never faced the issue I describe below.
I use Emacs with org mode and org babel to run R for literate programming to create latex reports. When I run R code from an org file (after having done C-c '), it sends the code to an iESS buffer \*R* and evaluate it. Fine. But after some code evaluations, a buffer called \*ess-command-output* pops up. In it is written:
`ess-output-delimiter75-END`
`The line I evaluated and its result (it is result first and then line evaluated)`
`ess-output-delimiter75-START`
The number 75 can change. If I continue code evaluation, it goes in this buffer but in a messy way. When I go back to the iESS \*R* buffer, it is mentioned "no process" in the mode line. And thus if I close the the \*ess-command-output* buffer and evalaute code from the ESS buffer (where the script is written), it asks me to open a new iESS buffer \*R:2*.
Quite difficult to reproduce this because I cannot identify when it happens. Can be after few evaluations or more.
Any idea of how this issue can be solved? Thanks.
|
process
|
ess command output buffer my config is windows gnu emacs ess org mode r for information i recently changed to this config so i needed to modify my init el before i used emacs org mode and ess without having never faced the issue i describe below i use emacs with org mode and org babel to run r for literate programming to create latex reports when i run r code from an org file after having done c c it sends the code to an iess buffer r and evaluate it fine but after some code evaluations a buffer called ess command output pops up in it is written ess output end the line i evaluated and its result it is result first and then line evaluated ess output start the number can change if i continue code evaluation it goes in this buffer but in a messy way when i go back to the iess r buffer it is mentioned no process in the mode line and thus if i close the the ess command output buffer and evalaute code from the ess buffer where the script is written it asks me to open a new iess buffer r quite difficult to reproduce this because i cannot identify when it happens can be after few evaluations or more any idea of how this issue can be solved thanks
| 1
|
8,016
| 11,205,751,352
|
IssuesEvent
|
2020-01-05 16:21:42
|
luckyos-code/ArgU
|
https://api.github.com/repos/luckyos-code/ArgU
|
closed
|
Preprocessing - Analyse der Qualität der Argumente
|
doing preprocessing
|
Ziel: Welche Art von 'noise' finden wir in den Argumenten?
Dazu sollen 100 - 300 Argumente manuell gelesen und analysiert werden.
Das Resultat sollte in einer Liste gespeichert werden mit den Spalten:
* argument ID
* Textstelle
* Auffälligkeit
* Relevanz [1 - 3]
Relevanz: 1 (unbrauchbar), 2 (muss man bereinigen), 3 (ist gut, wie es ist)
Beispiele:
* 4d3d4471-2019-04-18T11:45:01Z-00002-000
* "I accept."
* zu kurz
* 1
---
* fbe6ad2-2019-04-18T11:12:36Z-00001-000
* "You paid absolutely no attention whatsoever to the verses presented and instead went off into your own la la land"
* beleidigend / ausfallend
* 1
---
* 4d3d4471-2019-04-18T11:45:01Z-00001-000
* "Aussie! Aussie! Aussie! Oi! Oi! Oi!"
* Inhaltslos und provokativ, aber vorher kommen relevante Informationen
* 2
Code, der eine beliebige Menge von Argumenten ausgibt, befindet sich unter [manually_read_arguments_test](tests/manually_read_arguments_test.py)
|
1.0
|
Preprocessing - Analyse der Qualität der Argumente - Ziel: Welche Art von 'noise' finden wir in den Argumenten?
Dazu sollen 100 - 300 Argumente manuell gelesen und analysiert werden.
Das Resultat sollte in einer Liste gespeichert werden mit den Spalten:
* argument ID
* Textstelle
* Auffälligkeit
* Relevanz [1 - 3]
Relevanz: 1 (unbrauchbar), 2 (muss man bereinigen), 3 (ist gut, wie es ist)
Beispiele:
* 4d3d4471-2019-04-18T11:45:01Z-00002-000
* "I accept."
* zu kurz
* 1
---
* fbe6ad2-2019-04-18T11:12:36Z-00001-000
* "You paid absolutely no attention whatsoever to the verses presented and instead went off into your own la la land"
* beleidigend / ausfallend
* 1
---
* 4d3d4471-2019-04-18T11:45:01Z-00001-000
* "Aussie! Aussie! Aussie! Oi! Oi! Oi!"
* Inhaltslos und provokativ, aber vorher kommen relevante Informationen
* 2
Code, der eine beliebige Menge von Argumenten ausgibt, befindet sich unter [manually_read_arguments_test](tests/manually_read_arguments_test.py)
|
process
|
preprocessing analyse der qualität der argumente ziel welche art von noise finden wir in den argumenten dazu sollen argumente manuell gelesen und analysiert werden das resultat sollte in einer liste gespeichert werden mit den spalten argument id textstelle auffälligkeit relevanz relevanz unbrauchbar muss man bereinigen ist gut wie es ist beispiele i accept zu kurz you paid absolutely no attention whatsoever to the verses presented and instead went off into your own la la land beleidigend ausfallend aussie aussie aussie oi oi oi inhaltslos und provokativ aber vorher kommen relevante informationen code der eine beliebige menge von argumenten ausgibt befindet sich unter tests manually read arguments test py
| 1
|
326,963
| 9,963,087,819
|
IssuesEvent
|
2019-07-07 20:10:55
|
roundware/roundware-server
|
https://api.github.com/repos/roundware/roundware-server
|
opened
|
create script for uploading assets including user data
|
enhancement high priority
|
Bulk spreadsheet-based (most likely) method for uploading a lot of assets and write the user data in the asset when uploaded and create new users when necessary.
Flow may be something like:
1. send `POST users` with `device_id` = "firstname_lastname"
2. response will include `user.id` either newly created if doesn't exist or pre-existing based on `device_id`
3. use `user.id` in subsequent `POST assets/` call to create new asset with proper user attached
|
1.0
|
create script for uploading assets including user data - Bulk spreadsheet-based (most likely) method for uploading a lot of assets and write the user data in the asset when uploaded and create new users when necessary.
Flow may be something like:
1. send `POST users` with `device_id` = "firstname_lastname"
2. response will include `user.id` either newly created if doesn't exist or pre-existing based on `device_id`
3. use `user.id` in subsequent `POST assets/` call to create new asset with proper user attached
|
non_process
|
create script for uploading assets including user data bulk spreadsheet based most likely method for uploading a lot of assets and write the user data in the asset when uploaded and create new users when necessary flow may be something like send post users with device id firstname lastname response will include user id either newly created if doesn t exist or pre existing based on device id use user id in subsequent post assets call to create new asset with proper user attached
| 0
|
7,098
| 10,252,569,535
|
IssuesEvent
|
2019-08-21 09:17:53
|
aiidateam/aiida-core
|
https://api.github.com/repos/aiidateam/aiida-core
|
closed
|
Allow `Process.exposed_outputs` to work with `CalcJobNode`s
|
priority/nice-to-have topic/processes type/accepted feature
|
At present the link type filter is set only for `RETURN`:
https://github.com/aiidateam/aiida-core/blob/d79287c3cb9801ff854e9a24c5fcfe55ab433224/aiida/engine/processes/process.py#L843
Should just be a matter of adding `linktype=(LinkType.RETURN, LinkType.CREATE)`?
|
1.0
|
Allow `Process.exposed_outputs` to work with `CalcJobNode`s - At present the link type filter is set only for `RETURN`:
https://github.com/aiidateam/aiida-core/blob/d79287c3cb9801ff854e9a24c5fcfe55ab433224/aiida/engine/processes/process.py#L843
Should just be a matter of adding `linktype=(LinkType.RETURN, LinkType.CREATE)`?
|
process
|
allow process exposed outputs to work with calcjobnode s at present the link type filter is set only for return should just be a matter of adding linktype linktype return linktype create
| 1
|
391,618
| 11,576,297,293
|
IssuesEvent
|
2020-02-21 11:36:42
|
chef/chef
|
https://api.github.com/repos/chef/chef
|
closed
|
knife cookbook upload --all does not upload all cookbooks if any one of the cookbook version is frozen.
|
Aspect: UX Priority: Low
|
<!---
!!!!!! NOTE: CHEF CLIENT BUGS ONLY !!!!!!
This issue tracker is for the code contained within this repo -- `chef-client`, base `knife` functionality (not
plugins), `chef-apply`, `chef-solo`, `chef-client -z`, etc.
* Requests for new or alternative functionality should be made to [feedback.chef.io](https://feedback.chef.io/forums/301644-chef-product-feedback/category/110832-chef-client)
* [Chef Server issues](https://github.com/chef/chef-server/issues/new)
* [ChefDK issues](https://github.com/chef/chef-dk/issues/new)
* Cookbook Issues (see the https://github.com/chef-cookbooks repos or search [Supermarket](https://supermarket.chef.io) or GitHub/Google)
-->
## Description
<!--- Briefly describe the issue -->
``` knife cookbook upload --all ``` does not upload all the cookbooks if anyone of the cookbook's version is frozen. And as per current behaviour knife cookbook upload --alll shows wrong message at the end of command execution i.e. ``` Uploaded all cookbooks. ``` which confuse the user.
We either need to fix the message or the functionality to fix this.
## Chef Version
<!--- Tell us which version of chef-client you are using (see below for Server+ChefDK bugs). -->
```
ChefDK version: 4.6.35
Chef Infra Client version: 15.5.17
Chef InSpec version: 4.18.39
Test Kitchen version: 2.3.4
Foodcritic version: 16.2.0
Cookstyle version: 5.13.7
```
Present in Chef 14 too.
## Platform Version
<!--- Tell us which operating system distribution and version chef-client is running on. -->
## Replication Case
<!--- Tell us what steps to take to replicate your problem. See [How to create a Minimal, Complete, and Verifiable example](https://stackoverflow.com/help/mcve)
for information on how to create a good replication case. -->
I have one of the cookbook in repo whose version is frozen.
Then I try to upload all cookbooks from my cookbook repo using --all option
## Client Output
<!--- The relevant output of the chef-client run or a link to a gist of the entire run, if there is one.
The debug output (chef-client -l debug) may be useful, but please link to a gist, or truncate it. -->
```
knife cookbook upload --all
Uploading berks-test [0.1.0]
Uploading one-more-test [0.1.0]
Uploading one-moret-test [0.1.0]
Uploading starter [1.3.0]
Uploading test-cookbook [0.1.0]
Uploading test-one [0.1.0]
Uploading test-two [0.1.0]
ERROR: Version 0.1.0 of cookbook berks-test is frozen. Use --force to override.
WARNING: Not updating version constraints for some cookbooks in the environment as the cookbook is frozen.
Uploaded all cookbooks.
```
## Stacktrace
<!--- Please include the stacktrace.out output or link to a gist of it, if there is one. -->
|
1.0
|
knife cookbook upload --all does not upload all cookbooks if any one of the cookbook version is frozen. - <!---
!!!!!! NOTE: CHEF CLIENT BUGS ONLY !!!!!!
This issue tracker is for the code contained within this repo -- `chef-client`, base `knife` functionality (not
plugins), `chef-apply`, `chef-solo`, `chef-client -z`, etc.
* Requests for new or alternative functionality should be made to [feedback.chef.io](https://feedback.chef.io/forums/301644-chef-product-feedback/category/110832-chef-client)
* [Chef Server issues](https://github.com/chef/chef-server/issues/new)
* [ChefDK issues](https://github.com/chef/chef-dk/issues/new)
* Cookbook Issues (see the https://github.com/chef-cookbooks repos or search [Supermarket](https://supermarket.chef.io) or GitHub/Google)
-->
## Description
<!--- Briefly describe the issue -->
``` knife cookbook upload --all ``` does not upload all the cookbooks if anyone of the cookbook's version is frozen. And as per current behaviour knife cookbook upload --alll shows wrong message at the end of command execution i.e. ``` Uploaded all cookbooks. ``` which confuse the user.
We either need to fix the message or the functionality to fix this.
## Chef Version
<!--- Tell us which version of chef-client you are using (see below for Server+ChefDK bugs). -->
```
ChefDK version: 4.6.35
Chef Infra Client version: 15.5.17
Chef InSpec version: 4.18.39
Test Kitchen version: 2.3.4
Foodcritic version: 16.2.0
Cookstyle version: 5.13.7
```
Present in Chef 14 too.
## Platform Version
<!--- Tell us which operating system distribution and version chef-client is running on. -->
## Replication Case
<!--- Tell us what steps to take to replicate your problem. See [How to create a Minimal, Complete, and Verifiable example](https://stackoverflow.com/help/mcve)
for information on how to create a good replication case. -->
I have one of the cookbook in repo whose version is frozen.
Then I try to upload all cookbooks from my cookbook repo using --all option
## Client Output
<!--- The relevant output of the chef-client run or a link to a gist of the entire run, if there is one.
The debug output (chef-client -l debug) may be useful, but please link to a gist, or truncate it. -->
```
knife cookbook upload --all
Uploading berks-test [0.1.0]
Uploading one-more-test [0.1.0]
Uploading one-moret-test [0.1.0]
Uploading starter [1.3.0]
Uploading test-cookbook [0.1.0]
Uploading test-one [0.1.0]
Uploading test-two [0.1.0]
ERROR: Version 0.1.0 of cookbook berks-test is frozen. Use --force to override.
WARNING: Not updating version constraints for some cookbooks in the environment as the cookbook is frozen.
Uploaded all cookbooks.
```
## Stacktrace
<!--- Please include the stacktrace.out output or link to a gist of it, if there is one. -->
|
non_process
|
knife cookbook upload all does not upload all cookbooks if any one of the cookbook version is frozen note chef client bugs only this issue tracker is for the code contained within this repo chef client base knife functionality not plugins chef apply chef solo chef client z etc requests for new or alternative functionality should be made to cookbook issues see the repos or search or github google description knife cookbook upload all does not upload all the cookbooks if anyone of the cookbook s version is frozen and as per current behaviour knife cookbook upload alll shows wrong message at the end of command execution i e uploaded all cookbooks which confuse the user we either need to fix the message or the functionality to fix this chef version chefdk version chef infra client version chef inspec version test kitchen version foodcritic version cookstyle version present in chef too platform version replication case tell us what steps to take to replicate your problem see for information on how to create a good replication case i have one of the cookbook in repo whose version is frozen then i try to upload all cookbooks from my cookbook repo using all option client output the relevant output of the chef client run or a link to a gist of the entire run if there is one the debug output chef client l debug may be useful but please link to a gist or truncate it knife cookbook upload all uploading berks test uploading one more test uploading one moret test uploading starter uploading test cookbook uploading test one uploading test two error version of cookbook berks test is frozen use force to override warning not updating version constraints for some cookbooks in the environment as the cookbook is frozen uploaded all cookbooks stacktrace
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.